Lonewolf147

Members
  • Posts

    89
  • Joined

  • Last visited

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

Lonewolf147's Achievements

Apprentice

Apprentice (3/14)

2

Reputation

1

Community Answers

  1. I've been getting tons of those faults all week. But I've also been doing a lot of drive data rebuilds. I did run memtest just this morning to check, and all the tests ran ok.
  2. Yeah, I switched not only the cables, but which port they were each plugged in to. The power cable is a shared split cable from the power supply. I did change which power cables I was using too. I'm still in the rebuild of the one drive, so I'll have to see if it finishes without issue.
  3. That didn't seem to make a difference. I switched out the power cables for those two drives specifically. And they both still failed with write errors after starting the data rebuilds. I did spend the time and did a complete preclear on one of them before putting it back in the system. Still no luck. I removed both drives from the array again, put the one data drive in and started another rebuild, while leaving the second parity drive offline for now. Here's my latest diagnostics too. valaskjalf-diagnostics-20231224-1437.zip
  4. I have an 18tb drive that had a write error, then became disabled. I pulled it, reformatted it, put it back in and the rebuild started. It crashed with another write error a few hours into the build. I had this happen once before a few weeks ago, but it performed the rebuild just fine, and and extended smart test came back with no errors. If there are physical issues with the platters, if I run the preclear app on it, will that mark all the bad sectors if there are any? Here's a sample from my log of the write error occurring. Does this say what is actually wrong? (full diagnostics attached) Dec 21 01:04:26 Valaskjalf kernel: sd 10:0:7:0: [sdy] tag#983 UNKNOWN(0x2003) Result: hostbyte=0x00 driverbyte=DRIVER_OK cmd_age=3s Dec 21 01:04:26 Valaskjalf kernel: sd 10:0:7:0: [sdy] tag#983 Sense Key : 0x2 [current] Dec 21 01:04:26 Valaskjalf kernel: sd 10:0:7:0: [sdy] tag#983 ASC=0x4 ASCQ=0x0 Dec 21 01:04:26 Valaskjalf kernel: sd 10:0:7:0: [sdy] tag#983 CDB: opcode=0x8a 8a 00 00 00 00 00 5a 44 c9 b0 00 00 04 00 00 00 Dec 21 01:04:26 Valaskjalf kernel: blk_print_req_error: 8 callbacks suppressed Dec 21 01:04:26 Valaskjalf kernel: I/O error, dev sdy, sector 1514457520 op 0x1:(WRITE) flags 0x4000 phys_seg 128 prio class 2 Dec 21 01:04:26 Valaskjalf kernel: md: disk17 write error, sector=1514457456 Dec 21 01:04:26 Valaskjalf kernel: md: disk17 write error, sector=1514457464 Dec 21 01:04:26 Valaskjalf kernel: md: disk17 write error, sector=1514457472 . . . Dec 21 01:04:26 Valaskjalf kernel: md: disk17 write error, sector=1514464600 Dec 21 01:04:26 Valaskjalf kernel: md: disk17 write error, sector=1514464608 Dec 21 01:04:26 Valaskjalf kernel: md: disk17 write error, sector=1514464616 Dec 21 01:04:26 Valaskjalf kernel: md: recovery thread: exit status: -4 valaskjalf-diagnostics-20231221-0249.zip
  5. Yeah. I know. It's my current only way of getting the extra drives into the array. I've been running this for over a year now and never had that specific problem. I get that it is risky, and I'm willing to accept that. In the meantime, I'm slowly saving money to get a larger case and a few more HBA's in order to move all the drives to an internal configuration.
  6. Alright, so I'm back up and running now. I don't know why exactly. I shut everything down so I could pull each drive and mark it's location on my spreadsheet. I checked the HBA, it and its cables seemed fine. I went back to my list of 'missing' drives. None of them were on the HBA. All 8 drives were in one external USB enclosure. I'm guessing that when I powered down the enclosure to get at the drive serial numbers, it reset something in it, and let unRaid see them all again. I successfully upgraded back to 6.12.6.
  7. I'm just starting to dig around in the case. I think they may all be on my HBA. I can't verify until I check the serial numbers against the list. But, if that is the problem, then could the update have messed with the HBA drivers? I'm concerned that they didn't come back online when I downgraded, if it is a drivers issue.
  8. I upgraded to 6.12.6 this morning from 6.12.4. When I rebooted, 8 of my drives showed missing. I tried rebooting again, but no luck. I've attached my diagnostics from just after the upgrade. I have downgraded back to 6.12.4 to try and get my system running again. EDIT - Downgrading didn't work. Upon reboot the drives were still missing. valaskjalf-diagnostics-20231204-0712.zip
  9. Question - I have a script that monitors my CPU temp. I have a command built in to have it 'run' once a minute. The script runs just fine, my question is if I set it to run on Array Start, will it automatically be terminated when the array is stopped? If not, is there some command I can add to the script to check the status of the array and to kill the script if the array is stopped?
  10. Just copy/pasting this code into User Scripts I have an error right at the beginning
  11. My bad. I meant to put Raid1, not 0. I've edited my original post. Here's my diagnostics too. I'm looking at my Plex pool drives. valaskjalf-diagnostics-20230918-0808.zip
  12. I currently have a two drive pool in a Raid 1 setup. One drive is 256GB, the other is 1TB. I know that I'm only getting 256 protection. I'm wondering if there is a way to now remove the 256 drive and have the 1tb drive take over the pool at the full 1tb size? Or is the 1tb drive now configured as a 256 in unraid? edit: I originally typed raid 0, I meant to say raid 1, so I changed the wording.
  13. I just started seeing this error in my logs now too. I just upgraded to 6.12.4 yesterday. I usually keep a Firefox tab open with my Dashboard displayed, but for the past day I've actually had a tab open in Edge. When I went to look at it after it had sat for a few hours, all the animations on the dashboard (CPU/RAM/temps, etc) started going super fast, as if there were all fastforwarding to catch up with the current time. Then they all returned to normal speed. When I looked in the logs (checking for other issues) I found all these entries which coincide with the time that I was looking. Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [crit] 12634#12634: ngx_slab_alloc() failed: no memory Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: shpool alloc failed Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: nchan: Out of shared memory while allocating message of size 28129. Increase nchan_max_reserved_memory. Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: *351766 nchan: error publishing message (HTTP status code 500), client: unix:, server: , request: "POST /pub/devices?buffer_length=1 HTTP/1.1", host: "localhost" Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: MEMSTORE:01: can't create shared message for channel /devices Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [crit] 12634#12634: ngx_slab_alloc() failed: no memory Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: shpool alloc failed Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: nchan: Out of shared memory while allocating message of size 16811. Increase nchan_max_reserved_memory. Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: *351769 nchan: error publishing message (HTTP status code 500), client: unix:, server: , request: "POST /pub/disks?buffer_length=1 HTTP/1.1", host: "localhost" Sep 10 01:39:22 Valaskjalf nginx: 2023/09/10 01:39:22 [error] 12634#12634: MEMSTORE:01: can't create shared message for channel /disks Sep 10 01:39:23 Valaskjalf nginx: 2023/09/10 01:39:23 [crit] 12634#12634: ngx_slab_alloc() failed: no memory Sep 10 01:39:23 Valaskjalf nginx: 2023/09/10 01:39:23 [error] 12634#12634: shpool alloc failed Sep 10 01:39:23 Valaskjalf nginx: 2023/09/10 01:39:23 [error] 12634#12634: nchan: Out of shared memory while allocating message of size 28129. Increase nchan_max_reserved_memory. Sep 10 01:39:23 Valaskjalf nginx: 2023/09/10 01:39:23 [error] 12634#12634: *351776 nchan: error publishing message (HTTP status code 500), client: unix:, server: , request: "POST /pub/devices?buffer_length=1 HTTP/1.1", host: "localhost" valaskjalf-diagnostics-20230910-0222.zip
  14. That seemed to do it. Thanks!