Duggie264

Members
  • Posts

    42
  • Joined

  • Last visited

Everything posted by Duggie264

  1. Thanks @JorgeB, Just for completeness: I had a couple of disk failures that resulted in a dual Parity swap procedure - Successfull I then recieved some newer replacement HDD that necessitated another dual parity swap procedure. Parity copy went smoothly, prior to the disk rebuild, I accidentally clicked a drop down on the main page. This meant that I would have to redo the parity copy (16-->18TB). I then did a search, and came across this thread (which should be linked from ther Parity Swap procedure page!) Everything would have went well, had I not missed the bit about dual parity and not using the "29" at the end of the cmds... so I now have this: and this Disks 5 and 11 were the locations of the original missing/failed disks, that should have been rebuilt on the two old p[arity disks that you can see occupying thier slots (MDKE anbd QN99) However as I incorrectly entered 29... I am guessing that: Parity 1 Valid Parity 2 Invalid (being rebuilt across disk set as you confirmed above) Disk 5 (original Parity 1) - should be getting overwritten with original emulated Disk 5 data? Disk 11 (original Parity 2) - should be getting overwritten with origianl emulated Disk 11 data? So I suppose the real question is, what will the outcome of this be?
  2. So when starting the Array, would it try and rebuild P2 from the original P2 disk first, or would it just rebuild parity 2 from across the disk set (thus resulting in the loss of data from the failed disk the old P2 is replaceing, I assume?)
  3. @JorgeB, just out of curiosity, if you had dual parity, and left in the numbers "29" after each command, then proceeded to start the array, what is the likely outcome.... wish I was asking for a friend... 🤔🤨😒🙄😔
  4. Awesome, thanks for the help and information! best get swapping 😎
  5. Yeah I am positive - they were brand new drives and were clicking from the get go - one failed to even initialise, and by the time I realised both the other drives had mechanical issues, one had already successfully hounded the array, which subsequently died. Also happy with the parity swap procedure, just wasn’t sure how much more risk there would be swapping out a parity drive when there are already two failed data drives! cheers for your assistance bud, much appreciated!
  6. Thanks @itimpi So I already have dual parity, the issue now is I have dual (array) drive failures. I do have 2 16TB drives, but will I be able to swap out one of the Parity drives, when I already have two disks dead?
  7. BLUF: Can I replace a failed disk with a larger than parity disk (temporarily)? Background: I purchased three disks, one failed on testing, the other two were installed, subsequently one failed. The manufacturer, as part of the returns process has refunded the entire purchase amount to paypal for me, but I am now time limited in and have to return all three disks. All my disks (array and dual parity) are 14TB. I have two brand new 16TB disks. My array now shows two disks missing (as expected) and is running "at risk" Question 1. Will Unraid allow me to (temporarily) replace the missing 14TB drives with 16TB drives (I appreciate how parity works, but do not know A) if the extra two TB on the array disks will cause a parity fail, or B) if they will just be unprotected, or C) if Unraid will even allow the process). I have further 16TB drives on the way that will replace the parity drives in the event of B) 2. Am I correct in assuming that as I am missing two disks, I will be unable to swap a 14TB parity drive out for a 16TB (i.e. the swap-disable process?)
  8. had this a couple of times, was rebuilding so didn't really care as I was due restarts anyway. Just had it whilst chasing what makes my Docker Log fill up so quickly. Specifically I am running Microsoft Edge Version 97.0.1072.55 (Official build) (64-bit), and I opened the terminal from the dashboard, as soon as I entered the command "co=$(docker inspect --format='{{.Name}}' $(docker ps -aq --no-trunc) | sed 's/^.\(.*\)/\1/' | sort); for c_name in $co; do c_size=$(docker inspect --format={{.ID}} $c_name | xargs -I @ sh -c 'ls -hl /var/lib/docker/containers/@/@-json.log' | awk '{print $5 }'); YE='\033[1;33m'; NC='\033[0m'; PI='\033[1;35m'; RE='\033[1;31m'; case "$c_size" in *"K"*) c_size=${YE}$c_size${NC};; *"M"*) p=${c_size%.*}; q=${p%M*}; r=${#q}; if [[ $r -lt 3 ]]; then c_size=${PI}$c_size${NC}; else c_size=${RE}$c_size${NC}; fi ;; esac; echo -e "$c_name $c_size"; done " I started getting the nginx hissy fit Jan 10 10:18:44 TheNewdaleBeast nginx: 2022/01/10 10:18:44 [error] 3871#3871: nchan: Out of shared memory while allocating message of size 10016. Increase nchan_max_reserved_memory. Jan 10 10:18:44 TheNewdaleBeast nginx: 2022/01/10 10:18:44 [error] 3871#3871: *4554856 nchan: error publishing message (HTTP status code 500), client: unix:, server: , request: "POST /pub/update2?buffer_length=1 HTTP/1.1", host: "localhost" Jan 10 10:18:44 TheNewdaleBeast nginx: 2022/01/10 10:18:44 [error] 3871#3871: MEMSTORE:00: can't create shared message for channel /update2) I have 256GB Memory, and was at less than 11% utilisation so that wasn't an obvious issue, the output from df -h shows no issues either. Filesystem Size Used Avail Use% Mounted on rootfs 126G 812M 126G 1% / tmpfs 32M 2.5M 30M 8% /run /dev/sda1 15G 617M 14G 5% /boot overlay 126G 812M 126G 1% /lib/firmware overlay 126G 812M 126G 1% /lib/modules devtmpfs 126G 0 126G 0% /dev tmpfs 126G 0 126G 0% /dev/shm cgroup_root 8.0M 0 8.0M 0% /sys/fs/cgroup tmpfs 128M 28M 101M 22% /var/log tmpfs 1.0M 0 1.0M 0% /mnt/disks tmpfs 1.0M 0 1.0M 0% /mnt/remotes /dev/md1 13T 6.4T 6.4T 51% /mnt/disk1 /dev/md2 13T 9.5T 3.3T 75% /mnt/disk2 /dev/md4 13T 767G 12T 6% /mnt/disk4 /dev/md5 3.7T 3.0T 684G 82% /mnt/disk5 /dev/md8 13T 9.6T 3.2T 76% /mnt/disk8 /dev/md10 13T 9.1T 3.7T 72% /mnt/disk10 /dev/md11 13T 8.9T 3.9T 70% /mnt/disk11 /dev/md12 13T 7.0T 5.8T 55% /mnt/disk12 /dev/md15 13T 5.7T 7.2T 45% /mnt/disk15 /dev/sdq1 1.9T 542G 1.3T 30% /mnt/app-sys-cache /dev/sdh1 1.9T 17M 1.9T 1% /mnt/download-cache /dev/nvme0n1p1 932G 221G 711G 24% /mnt/vm-cache shfs 106T 60T 46T 57% /mnt/user0 shfs 106T 60T 46T 57% /mnt/user /dev/loop2 35G 13G 21G 38% /var/lib/docker /dev/loop3 1.0G 4.2M 905M 1% /etc/libvirt tmpfs 26G 0 26G 0% /run/user/0 I then ran the following commands in order to restore connectivity (after first restart I had no access) root@TheNewdaleBeast:~# /etc/rc.d/rc.nginx restart Checking configuration for correct syntax and then trying to open files referenced in configuration... nginx: the configuration file /etc/nginx/nginx.conf syntax is ok nginx: configuration file /etc/nginx/nginx.conf test is successful Shutdown Nginx gracefully... Found no running processes. Nginx is already running root@TheNewdaleBeast:~# /etc/rc.d/rc.nginx restart Nginx is not running root@TheNewdaleBeast:~# /etc/rc.d/rc.nginx start Starting Nginx server daemon... and normal service was restored Jan 10 10:22:40 TheNewdaleBeast root: Starting unraid-api v2.26.14 Jan 10 10:22:40 TheNewdaleBeast root: Loading the "production" environment. Jan 10 10:22:41 TheNewdaleBeast root: Daemonizing process. Jan 10 10:22:41 TheNewdaleBeast root: Daemonized successfully! Jan 10 10:22:42 TheNewdaleBeast unraid-api[13364]: ✔️ UNRAID API started successfully! Not sure if this is helpful, but to me it appeared that issue was caused (or coincidental with) use of web terminal.
  9. Good news (Phew) rebooted and restarted Array and all is looking good so far!
  10. Yeah I couldn't either - just wanted to ask first this time, rather than waiting until I am deeply mired in a rebooted server with no Array, no Diagnostics and all I have is a sheepish grin!
  11. @rxnelson - Good Point, and most likely why it didn't work for me! I had just moved my 1080 GTX from an i7 6700K based gaming rig to my Dual Xeon E5-2660 Unraid rig, and completely forgot that the Xeons don't have integrated graphics 🙄However I would be curious to know if it will work with an integrated or non-nvidia GPU? Oh and apologies for late reply - I am not getting notifications for some reason!
  12. BLUF: User Shares Missing. in MC mnt/user is missing, mnt/user0 is there and there is a ?user in red. Server has not yet been restarted (only array stop/started) Good evening, I had 13x256GB, 1x1TB SSD and 1x1TB NVME in a giant cache pool. With multiple cache pools now a thing, I wanted to break down my cache and sort out my shares as follows: Get Rid of old cache pool: Disabled Docker and VM Set all my shares to Cache: Yes. Called in Mover to move everything to the array* ** Shut down the array. Removed all drives from the cache pool, then deleted it. * When mover finished there were still some appdata files in Cache and Disk 2 so I spun up MC via ssh and did the following: 1. mnt/disk2/appdata ---> mnt/disk10/ 2. mnt/cache/appdata ---> mnt/disk10/ **When I exited MC and returned to the GUI window I noticed that all my user shares had disappeared (Ahhh F%$£!) I immediately downloaded diagnostics - the one ending 2305 I then spent a bit of time looking for similar errors - wasn't massively concerned as all data is still there. Create New Cache Pools Created the 4 new cache pools. Set the FS to XFS for the single drives and Auto for the actual pools Started the array back up and formatted everything to mount them again. Still no User shares - immediately downloaded diagnostics again - the one ending 2355 on looking in MC I see this: no User on the left, but User0 on the right shows the Shares structure still there. I fully accept that I may have done something wrong/stupid etc, but have three questions: 1. Can someone help? and if so 2. How do I recover my Shares and 3. What did I do that actually screwed it up (as I do like to learn, albeit painfully) Many thanks in anticipation of your assistance! thenewdalebeast-diagnostics-20210715-2305.zip thenewdalebeast-diagnostics-20210715-2355.zip
  13. Allegedly, from 465.xx.onwards nvidia drivers now support passthrough (GPU Passthrough), as passthrough was software hobbled in drivers to "force" upgrades/purchase of Tesla/Quadro. However unable to passthrough EVGA 1080 GTX SC - 8GB successfully to a Win 10 VM - I czn see it in device manager but is a code 43 regardless. Will attempt this script this week at some point if I get time.
  14. Just came across this, nice work! if you are still maintaining/developing it would be great if there was someway to display the mags that reside in a folder structure like: Mag Title/Year/IssueTitle.pdf thanks
  15. Another upvote from me, been having this problem for a couple of years now, and with 15 $TB SAS drives spinning day and night, paid a few quid more than I have needed/wanted to! Currently run 15 14 Seagate ST4000NM0023 (SAS) and 2 ST4000DM005 (SATA) drives through 2 LSI 9207-8i in IT mode. Would love a simple way to put the drives into standby!
  16. Hi, just started getting this crash today, any ideas? 2020-04-19 15:23:48.353687 [info] System information Linux ################# 4.19.107-Unraid #1 SMP Thu Mar 5 13:55:57 PST 2020 x86_64 GNU/Linux 2020-04-19 15:23:48.383161 [info] PUID defined as '99' 2020-04-19 15:23:48.415129 [info] PGID defined as '100' 2020-04-19 15:23:48.918075 [info] UMASK defined as '000' 2020-04-19 15:23:48.945899 [info] Permissions already set for volume mappings 2020-04-19 15:23:48.972882 [warn] TRANS_DIR not defined,(via -e TRANS_DIR), defaulting to '/config/tmp' 2020-04-19 15:23:49.004719 [info] Deleting files in /tmp (non recursive)... 2020-04-19 15:23:49.030933 [info] Starting Supervisor... 2020-04-19 15:23:49,197 INFO Included extra file "/etc/supervisor/conf.d/plexmediaserver.conf" during parsing 2020-04-19 15:23:49,197 INFO Set uid to user 0 succeeded 2020-04-19 15:23:49,199 INFO supervisord started with pid 6 2020-04-19 15:23:49.004719 [info] Deleting files in /tmp (non recursive)... 2020-04-19 15:23:49.030933 [info] Starting Supervisor... 2020-04-19 15:23:49,197 INFO Included extra file "/etc/supervisor/conf.d/plexmediaserver.conf" during parsing 2020-04-19 15:23:49,197 INFO Set uid to user 0 succeeded 2020-04-19 15:23:49,199 INFO supervisord started with pid 6 2020-04-19 15:23:50,202 INFO spawned: 'plexmediaserver' with pid 55 2020-04-19 15:23:50,202 INFO reaped unknown pid 7 2020-04-19 15:23:50,437 DEBG fd 8 closed, stopped monitoring <POutputDispatcher at 139734334410080 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stdout)> 2020-04-19 15:23:50,438 DEBG fd 10 closed, stopped monitoring <POutputDispatcher at 139734334410368 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stderr)> 2020-04-19 15:23:50,438 INFO exited: plexmediaserver (exit status 255; not expected) 2020-04-19 15:23:50,438 DEBG received SIGCHLD indicating a child quit 2020-04-19 15:23:51,441 INFO spawned: 'plexmediaserver' with pid 60 2020-04-19 15:23:51,731 DEBG fd 8 closed, stopped monitoring <POutputDispatcher at 139734335288080 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stdout)> 2020-04-19 15:23:51,731 DEBG fd 10 closed, stopped monitoring <POutputDispatcher at 139734333605152 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stderr)> 2020-04-19 15:23:51,731 INFO exited: plexmediaserver (exit status 255; not expected) 2020-04-19 15:23:51,732 DEBG received SIGCHLD indicating a child quit 2020-04-19 15:23:53,736 INFO spawned: 'plexmediaserver' with pid 65 2020-04-19 15:23:54,021 DEBG fd 8 closed, stopped monitoring <POutputDispatcher at 139734334119840 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stdout)> 2020-04-19 15:23:54,021 DEBG fd 10 closed, stopped monitoring <POutputDispatcher at 139734334410368 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stderr)> 2020-04-19 15:23:54,021 INFO exited: plexmediaserver (exit status 255; not expected) 2020-04-19 15:23:54,021 DEBG received SIGCHLD indicating a child quit 2020-04-19 15:23:54,021 DEBG fd 8 closed, stopped monitoring <POutputDispatcher at 139734334119840 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stdout)> 2020-04-19 15:23:54,021 DEBG fd 10 closed, stopped monitoring <POutputDispatcher at 139734334410368 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stderr)> 2020-04-19 15:23:54,021 INFO exited: plexmediaserver (exit status 255; not expected) 2020-04-19 15:23:54,021 DEBG received SIGCHLD indicating a child quit 2020-04-19 15:23:57,027 INFO spawned: 'plexmediaserver' with pid 70 2020-04-19 15:23:57,323 DEBG fd 8 closed, stopped monitoring <POutputDispatcher at 139734335288080 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stdout)> 2020-04-19 15:23:57,323 DEBG fd 10 closed, stopped monitoring <POutputDispatcher at 139734333605056 for <Subprocess at 139734334410224 with name plexmediaserver in state STARTING> (stderr)> 2020-04-19 15:23:57,324 INFO exited: plexmediaserver (exit status 255; not expected) 2020-04-19 15:23:57,324 DEBG received SIGCHLD indicating a child quit 2020-04-19 15:23:58,325 INFO gave up: plexmediaserver entered FATAL state, too many start retries too quickly 2020-04-19 15:23:58,325 INFO gave up: plexmediaserver entered FATAL state, too many start retries too quickly
  17. Hi I am currently running: Unraid: 6.8.3, Sonarr: Version 2.0.0.5344, Mono Version 5.20.1.34 Ombi: 3.0.4892 I have searched for this issue, but can't seem to find a directly related issue or resolution. I have series that are not marked available. For this example - Friends (1994). Upon pressing the Request radio button, then selecting the Select option, I see that the final episodes from S09 and S10 are marked as missing. I then opened SONARR, and both episodes are there (and watchable), however after a little bit more digging, it appears that the final episodes of both these sesons are range marked episode e.g. S10E17E18. My questions is: Is there a simple way to get Ombi to recognise these episodes, and thus marking the entire series as available? Many Thanks Duggie
  18. trurl - many thanks for your replies, yes, memory is fine, and no I couldn't get any diagnostics as I lost the system completely - You are right about the log server though, I should really get one up and running! Cheers, Duggie
  19. So to add to my misery, after three attempted restarts and three failures, I left to put a load of washing in, came back, restart number 4 and it appears to be good with everything back to normal and currently enjoying a parity check. Grrr... but Yay!
  20. I have been running 6.8 stable for a couple of weeks with no problems. Last weekend, after adding some nzbs to sabz, I noticed that they were not being added. I then restarted the container, however it failed to come back up. Then the dashboard hung and would not load, Main would load only the top half (disk Info) or if I scrolled to bottom of page quickly enough, it would load the reboot, mover options (but no the disk info at the top of the page). I rebooted the server, but the issues remained. I then downgraded to the previous version, and all seemed well. Today I decided to upgrade to the 6.8.1-rc1 version, and everything appeared to go well, for about 2 hours! For that entire time I was in a Ubuntu VM with a passed through KB monitor and mouse, administering the server. Without any indication or notification the monitor screen went blank. I then jumped on seperate PC, but could not open an ssh session nor ping it. The server still appeared to be running (i.e. it hadn't power cycled). I manually fat fingered the server, and it will not reboot. It hangs at the screenshot below (prior to this only other thing is that an unclean shutdown was detected, and dirty bit set, clearing dirty bit, as I would expect on a system crash), then once timeout is complete stalls completely after the pps3 worker is killed. Is there going to be anything retrievable on the flas before I flatten and rebuild (obviously no logs due ot crash and restart/hang)? unraid.mp4 unraid.mp4
  21. So I emailed the maintainers from the SF link you provided, and this was their response: Hi Duggie, a quick glance at the UNRAID changelog it looks like in unRAID 6.6.7 they are using the 4.18.20 kernel. Do you know what kernel is in the newer RC versions? Feel free to submit the diagnostic logs to us. We will take a look. Contrary to their comments, we do maintain the kernel driver as well and have some patches staged to go upstream soon. Thanks, Scott AND ON SENDING MY DIAG LOGS AND THE UNRAID CHANGE-LOGS FOR THE 6.7.0-rcX FAMILY RCVD THIS FROM DON I see a DMAR error logged. That may be what caused the controller lockup, which caused the OS to send down a reset to the drive. Are you able to update the driver and build it for a test? If so, there is a structure member in the scsi_host_template called .max_sectors. It is set to 2048, wondering if you can change it to 1024 for a test? If not, I would have to know what OS I could do the build for you on. Not real sure about unraid. Feb 25 23:03:09 TheNewdaleBeast kernel: DMAR: DRHD: handling fault status reg 2 Feb 25 23:03:09 TheNewdaleBeast kernel: DMAR: [DMA Read] Request device [81:00.0] fault addr fe8c0000 [fault reason 06] PTE Read access is not set Feb 25 23:03:40 TheNewdaleBeast kernel: hpsa 0000:81:00.0: scsi 14:0:7:0: resetting physical Direct-Access SEAGATE ST4000NM0023 PHYS DRV SSDSmartPathCap- En- Exp=1 Feb 25 23:03:57 TheNewdaleBeast avahi-daemon[4764]: Leaving mDNS multicast group on interface br0.IPv6 with address fe80::1085:73ff:fedb:90d4. Feb 25 23:03:57 TheNewdaleBeast avahi-daemon[4764]: Joining mDNS multicast group on interface br0.IPv6 with address fd05:820d:9f35:1:d250:99ff:fec2:52fb. Feb 25 23:03:57 TheNewdaleBeast avahi-daemon[4764]: Registering new address record for fd05:820d:9f35:1:d250:99ff:fec2:52fb on br0.*. Feb 25 23:03:57 TheNewdaleBeast avahi-daemon[4764]: Withdrawing address record for fe80::1085:73ff:fedb:90d4 on br0. Feb 25 23:03:58 TheNewdaleBeast ntpd[3173]: Listen normally on 6 br0 [fd05:820d:9f35:1:d250:99ff:fec2:52fb]:123 Feb 25 23:03:58 TheNewdaleBeast ntpd[3173]: new interface(s) found: waking up resolver Feb 25 23:04:33 TheNewdaleBeast kernel: hpsa 0000:81:00.0: Controller lockup detected: 0x00130000 after 30 Feb 25 23:04:33 TheNewdaleBeast kernel: hpsa 0000:81:00.0: controller lockup detected: LUN:0000000000800601 CDB:01030000000000000000000000000000 Feb 25 23:04:33 TheNewdaleBeast kernel: hpsa 0000:81:00.0: Controller lockup detected during reset wait Feb 25 23:04:33 TheNewdaleBeast kernel: hpsa 0000:81:00.0: scsi 14:0:7:0: reset physical failed Direct-Access SEAGATE ST4000NM0023 PHYS DRV SSDSmartPathCap- En- Exp=1 Feb 25 23:04:33 TheNewdaleBeast kernel: sd 14:0:7:0: Device offlined - not ready after error recovery @limetech would you be able to assist as I am currently about 8000 feet below sea level, with only a snorkel for comfort!
  22. @johnnie.black @limetech Thanks for your response, guess I'll just have to stick with 6.6.7 for the foreseeable, in the meantime I have emailed HP to try and hasten a solution - I will update you if I get a response, Regards Duggie
  23. Hi, I still have the same problem with this update that I have had with every 6.7.0RCx version, namely that once I reboot, multiple drives fail to mount/are disabled. On reversion to 6.6.7 all the drives are fine. I have tried: Install --> reboot Install --> Power down all VMs and containers --> Turn off Docker and VM manager --> Reboot Install --> Power down all VMs and containers --> Turn off Docker and VM manager --> Stop Array --> Reboot Regardless of the process I follow, loads of drives appear failed/corrupt/disabled after reboot and array started. reversion to 6.6.7 returns system to normal operations, except disks 2 and 5 which remain disabled. See attached diagnostics, taken after upgrade and reboot, followed by reversion .thenewdalebeast-diagnostics-20190501-0848.zip If I get some time this weekend, I will do another upgrade, and pull diags at every step.