Christopher

Members
  • Posts

    35
  • Joined

  • Last visited

Everything posted by Christopher

  1. Yes! Thanks for the correction. I updated my original post.
  2. I'm having an issue where my Plex Server no longer being recognized. Plex is seeing my server as a new instance with a message on top saying, "An unclaimed media server has been found on your network. Claim it now". My Libraries are missing because it think it's a new server. Here's what led up to this below... 1) I had several cache drive BTRFS errors in my log. Docker apps to would stop or not respond. 2) Searched the forums and found post with similar issues. I tried a few things to fix it with no success. 3) I decided to recreate my cache drive. I disabled docker and VM and moved all cache folders/files to the array. 5) I cleared my cache configuration and deleted partitions on those my 4 cache SSDs. 6) I setup the 4 SSD drives as cache again and formatted. 7) Copied all files back to the cache array. Restarted server. 9) Deleted my docker.img 10) Enabled docker. Docker created new empty docker.img 11) Installed Plex from previous Apps. Config pointed to original folders 12) Started Plex Any ideas? Is there a way to get Plex to see this as my original instance? If I proceed with this new instance and add existing libraries, will Plex automatically use the existing database/files in my config folder? If I proceed with this how will it effect my Tautulli database? Your help is appreciated!
  3. Thanks for your pgAdmin4 template. Do you have any plans to upgrade from pgAdmin4 version 2.1 to the latest 3.4?
  4. +1. Maybe unRaid can have an option to login with basic http authentication or a forms login page. Sonarr has this option.
  5. I figured it out. I assumed the default user in the container was jovyan so I didn't set it. After reading the docs again, I explicitly set the NB_USER and now it works. Extra Parameters: --user root -e NB_USER="jovyan" -e NB_UID="1000" -e NB_GID="100"
  6. I'm trying to mount a local unRAID appdata folder to the working folder in docker container jupyter/scipy-notebook. The container runs and works fine except my work is not being saved to my volume mount. I believe the problem has to do with incorrect values in the "Extra Parameters" section for the User Id and Group Id. I've tried 99 and 0 as values. Can anyone help me get this working? Your help is appreciated! Thanks! Documentation for volume mounts on this image: https://jupyter-docker-stacks.readthedocs.io/en/latest/using/recipes.html#host-volume-mounts-and-notebook-errors Docker Options for this Image: https://jupyter-docker-stacks.readthedocs.io/en/latest/using/common.html#Docker-Options My Configuration:
  7. http://slickdeals.net/f/7980175-5tb-toshiba-7200rpm-3-5-hdd-ph3500u-1i72-139-99-fs-walmart-toshiba-com-5tb-seagate-enterprise-capacity-7200rpm-hdd-st5000nm0024-159-99-fs-rakuten?v=1&src=SiteSearch
  8. A Friend let me borrow a smaller drive and I was able to complete the process. Thanks for you help!
  9. The parity drive I want to use to restore disk13(3TB) is 4TB. I would like to replace disk13(3TB) with a new drive but my only spares are 5TB. Is there anyway I can make the 5TB appear to be 4TB or 3TB? This is just temporary so I can rebuilt disk13. Thanks, Christopher
  10. Thanks for the info bjp999! Here's the reason why I want to do this: [*]Initial state of my server was perfect. I had just done a parity check and all was well. [*]I upgraded my parity drive from 4GB to 5GB, Unraid started building parity. [*]In the morning I saw disk13 was redballed and parity build did not complete. Nothing has been copied/deleted/edited to the server since. I did not trust my old parity because events happened between steps 1 and 2 that wasn't sure about and I don't recall the details of. So I used reiserfsck to recover disk13 data and it restored about 2/3 of it. I copied the restored data to a local disk on my PC. Now I want to try to rebuild disk13 from my old parity. I realize that it the parity info might be invalid. But I wanted to give it a shot. Once I rebuild it I want to do an md5 comparision with the reiserfsck recovered data. Can you point me to the procedure? Thanks, Christopher
  11. My system currently doesn't have a parity drive. I want to install an old parity drive and have unraid trust it. When I start the array with the old parity drive I want this think everything is normal. Can someone point me to the link for this procedure? Thanks, Christopher
  12. Your right on both points. Thank you very much for your explanation! -Christopher
  13. I'm not having any serious issues I need help with. I just want to understand how the following works: 1) I used reiserfsck to recover data from a disk13 2) reiserfsck recovered 2/3 of the files and created a lost+found folder. 3) I followed the 'Trust My Array' procedure to reset my entire array. 4) I reassigned all the disks to the same slots minus the parity drive. I matched all serials to the old slot and triple checked against a screenshot. Yes, i'm unprotected. 5) I started the array. 6) Now my disk13 shows the original file structure with 100% of the files. I tested a few files that were missing after the reiserfsck rebuild and they work fine. My assumption was that disk13 would still show the reiserfsck rebuilt folder/file structure even after the 'Trust My Array' procedure. How did 100% my original disk13 folder/file structure return after the "reiserfsck --rebuild-tree /dev/md13" rebuilt and created 2/3 the original folder/file structure? Thanks, Christopher
  14. I went ahead and restarted the server and now disk13 is no longer showing as unformatted but still has a red ball. The reiserfsck tool recovered about 2/3 of my files. I've copied the recovered files using terracopy to a drive on my PC. I want to attempt to rebuild the disk13 by putting my old parity drive in. I know that my new parity is not correct. To do this I need to remove the disk13 red ball. Unraid will not let me put back my old parity drive with a red ball. The reason I didn't do a rebuild initially was because the events leading up to disk13 failure were obscure in my mind. Therefore I didn't trust my old parity drive. The reason I want to do a parity(old) rebuilt after a reiserfsck recovery: I want to md5 compare reiserfsck recovered files with parity(old) rebuilt files. If I see that the md5 matches for all or most files, then this give me confidence in the parity(old) rebuilt disk13 data. I found this information below but I wanted to confirm with you before I move forward. Also I have Unraid version 5.0.6. Whats the best way to remove the redball and put my old parity drive back? http://lime-technology.com/wiki/index.php/Make_unRAID_Trust_the_Parity_Drive,_Avoid_Rebuilding_Parity_Unnecessarily Thanks, Christopher
  15. I'm pretty sure I didn't check it but you never know. Here's the df output. md13 is the only one not mounted. I read in another thread that the "unformatted" label in the webgui should really read "unmounted". So the array is in normal mode with all disks mounted except md13. Perhaps reiserfsck didn't complain because the device it was operating on was already unmounted? I've attached a few webgui screenshots. Should I stop the array and then restart the server? Thanks, Christopher
  16. Perhaps I was already in Maintenance mode and I didn't know it? Can Maintenance mode be automatically enable by a redball, parity build or some failure event? I don't see anything regarding Maintenance mode in the Dynamix webgui. Is there cmd I can run in the terminal to show if its in Maintenance mode or not? I did run reiserfsck against md device md13 as shown below. reiserfsck didn't refuse to run or mention anything about Maintenance mode. What's you recommendation moving forward?
  17. No. I have not restarted the array yet since reiserfsck completed. I'm asking if there's anything else I should do before a restart? Also, I did not run reiserfsck in Maintenance mode. My mistake. What's are negative effects of this?
  18. The reiserfsck process just finish and the end of the report is shown below. The web front end still shows the drive as unformatted. The wiki says I should restart the array after reiserfsck completes. Please confirm if there's anything else I should do before a restart? Also what exactly does "Deleted unreachable items 40581" and "Empty lost dirs removed 2" mean from the report? Thanks!
  19. This is going to take a long time. Is there anyway to disable parity while during the reiserfsck process? I can build the parity drive after. I don't think my parity is correct anyways based on the "mdcmd status" results in my original post above. Thanks. Current reiserfsck Status:
  20. Here's the result of "reiserfsck --check /dev/md13": Should I proceed and run "reiserfsck --rebuild-tree /dev/md13"? Thanks, Christopher
  21. I'm running the check command as "reiserfsck --check /dev/md13" and will report the results in the morning. Thanks.
  22. I started by running a parity check on existing 4GB parity drive. Then I followed the steps described on the forum to upgrade my parity drive from 4GB to 5TB. I left it building the parity drive. After a day or so I check to see what the status was. It was complete but disk13 has a red ball with 200K+ errors reported in the Dynamix web gui. I tried accessing the drive and was able to list and open files. When I attempted to get the system log I got some memory error. It was getting late so I decided to shutdown the server pick up in the morning. I turn on the server this morning and disk13 is showing as unformatted. I'm not sure how to interpret the mdcmd results below but I don't think it's good. Please let me know if the mdcmd status is ok and I can rebuild disk13 using the new 5GB parity drive. If not, what other options do I have to recover the data? Also, I still have the old 4GB parity drive. Below is a link to the system log and smart report . The smart report is from a prior test. I didn't run a smart test since this happened. I wasn't sure if it would make the situation worse. Thanks, Christopher Unraid Version: 5.0.6 System log and smart report: https://drive.google.com/file/d/0B00Diiihkv_qSmY5YXpHaFM4VkU/view?usp=sharing /root/mdcmd status | egrep "mdResync|mdState|sbSync" sbSynced=0 sbSyncErrs=0 mdState=STARTED mdResync=0 mdResyncCorr=1 mdResyncPos=0 mdResyncDt=0 mdResyncDb=0 This looks like the error in the system log:
  23. NewEgg has 20% off on all Norco server chassis and $75 off the RPC-4224. Promo code WWHUY685 http://www.newegg.com/Server-Chassis/BrandSubCat/ID-10473-412 -Christopher
  24. Make sure to read the responses on the slickdeals.net forum for this deal. Only 1 year warranty "The drive specs total hours break out that if your computer was on 24/7 that this drive would be dead in 3 months" 37% failure rate http://slickdeals.net/f/5160460-Newegg-Special-3TB-Seagate-HDD-for-124-95?