Jump to content

JorgeB

Moderators
  • Posts

    67,719
  • Joined

  • Last visited

  • Days Won

    708

Everything posted by JorgeB

  1. v6.9.2 uses kernel 5.10, see if you can find more info if/when that issue was fixed.
  2. Also and if you read what I posted in this thread, I suggested it could be a hardware issue, not that it must be like you mentioned many times since, also suggested that "one thing you can try it to boot the server in safe mode with all docker/VMs disable, let it run as a basic NAS for a few days, if it still crashes it's likely a hardware problem, if it doesn't start turning on the other services one by one.", by doing that you could have already confirmed if it was iGPU related or not.
  3. Unlikely that any mods are seeing the same issue or they would have chimed in by now, mods are just regular users, you can easily test and see if not using the GPU helps or not, it might not even be related to that.
  4. If the disks are outside the server and you don't have enough ports there you can install smartmontools in Windows and run a SMART test.
  5. With what? I don't have any issues with v6.9.2, but I don't use the GPU, you can test and confirm it it helps or not.
  6. Disk has pending sectors, it needs to be replaced, on other disks you can run an extended SMART test to confirm if they are OK or not.
  7. It's logged as a disk problem, but since the SMART test passed after it's OK for now, just keep monitoring.
  8. The btrfs error shouldn't prevent the server from starting, confirmed by the fact that you can boot in safe mode, that suggest a plugin or other configuration problem, you can post the diags after booting in safe mode so we can check the filesystem problem first.
  9. So it's not partition corruption, it's completely missing, easiest way is to rebuild the disk(s), assuming parity is still valid, partition information is outside parity, and Unraid will recreate the partitions, you can easily test by unassigning one of the disks and start the array, if the emulated disk mounts correctly rebuild on top, then repeat for the other one.
  10. The ones listed above as onboard SATA. By the looks of it you should be able to get most data but there could be some data corruption, but you first need to solve the device dropping issue.
  11. Sorry, I must have confused your diags with another one, you have two SASLP controllers, there's no LSI, you can still connect the SSDs to the Intel ports though.
  12. You can look, it's in the lsscsi.txt file in the diags, but it was on the same controller: The other SSD device was also there, so that one is now on the LSI. The SSD was on an SASLP, now is on the other one.
  13. If it's a single backplane it's direct connection, you have devices connected to 3 different controllers: These are on an SASLP [1:0:0:0] disk ATA SanDisk SDSSDH31 70RL /dev/sdj /dev/sg9 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/1:0:0:0 [/sys/devices/pci0000:00/0000:00:01.0/0000:01:00.0/host1/port-1:0/end_device-1:0/target1:0:0/1:0:0:0] [1:0:1:0] disk ATA WDC WD100EMAZ-00 0A83 /dev/sdk /dev/sg10 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/1:0:1:0 [/sys/devices/pci0000:00/0000:00:01.0/0000:01:00.0/host1/port-1:1/end_device-1:1/target1:0:1/1:0:1:0] These on the onboard SATA ports [2:0:0:0] disk ATA ST10000VN0008-2P SC61 /dev/sdb /dev/sg1 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/2:0:0:0 [/sys/devices/pci0000:00/0000:00:11.4/ata1/host2/target2:0:0/2:0:0:0] [3:0:0:0] disk ATA WDC WD60EFRX-68M 0A82 /dev/sdc /dev/sg2 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/3:0:0:0 [/sys/devices/pci0000:00/0000:00:11.4/ata2/host3/target3:0:0/3:0:0:0] [4:0:0:0] disk ATA WDC WD60EFRX-68M 0A82 /dev/sdd /dev/sg3 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/4:0:0:0 [/sys/devices/pci0000:00/0000:00:11.4/ata3/host4/target4:0:0/4:0:0:0] [5:0:0:0] disk ATA ST8000VN0022-2EL SC61 /dev/sde /dev/sg4 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/5:0:0:0 [/sys/devices/pci0000:00/0000:00:11.4/ata4/host5/target5:0:0/5:0:0:0] [6:0:0:0] disk ATA WDC WD60EFRX-68M 0A82 /dev/sdf /dev/sg5 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/6:0:0:0 [/sys/devices/pci0000:00/0000:00:1f.2/ata5/host6/target6:0:0/6:0:0:0] [7:0:0:0] disk ATA WDC WD60EFRX-68M 0A82 /dev/sdg /dev/sg6 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/7:0:0:0 [/sys/devices/pci0000:00/0000:00:1f.2/ata6/host7/target7:0:0/7:0:0:0] [8:0:0:0] disk ATA WDC WD60EFRX-68M 0A82 /dev/sdh /dev/sg7 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/8:0:0:0 [/sys/devices/pci0000:00/0000:00:1f.2/ata7/host8/target8:0:0/8:0:0:0] [9:0:0:0] disk ATA ST10000VN0004-1Z SC60 /dev/sdi /dev/sg8 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/9:0:0:0 [/sys/devices/pci0000:00/0000:00:1f.2/ata8/host9/target9:0:0/9:0:0:0] And these on another SASLP: [12:0:0:0] disk ATA ST10000VN0008-2J SC60 /dev/sdl /dev/sg11 state=running queue_depth=1 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/12:0:0:0 [/sys/devices/pci0000:00/0000:00:01.1/0000:02:00.0/host12/port-12:0/end_device-12:0/target12:0:0/12:0:0:0] [12:0:1:0] disk ATA WDC WD100EMAZ-00 0A83 /dev/sdm /dev/sg12 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/12:0:1:0 [/sys/devices/pci0000:00/0000:00:01.1/0000:02:00.0/host12/port-12:1/end_device-12:1/target12:0:1/12:0:1:0] [12:0:2:0] disk ATA SanDisk Ultra II 00RL /dev/sdn /dev/sg13 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/12:0:2:0 [/sys/devices/pci0000:00/0000:00:01.1/0000:02:00.0/host12/port-12:2/end_device-12:2/target12:0:2/12:0:2:0] [12:0:3:0] disk ATA ST4000DM000-1F21 CC54 /dev/sdo /dev/sg14 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/12:0:3:0 [/sys/devices/pci0000:00/0000:00:01.1/0000:02:00.0/host12/port-12:3/end_device-12:3/target12:0:3/12:0:3:0] [12:0:4:0] disk ATA ST10000VN0008-2P SC61 /dev/sdp /dev/sg15 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/12:0:4:0 [/sys/devices/pci0000:00/0000:00:01.1/0000:02:00.0/host12/port-12:4/end_device-12:4/target12:0:4/12:0:4:0] [12:0:5:0] disk ATA WDC WD60EFRX-68L 0A82 /dev/sdq /dev/sg16 state=running queue_depth=32 scsi_level=6 type=0 device_blocked=0 timeout=30 dir: /sys/bus/scsi/devices/12:0:5:0 [/sys/devices/pci0000:00/0000:00:01.1/0000:02:00.0/host12/port-12:5/end_device-12:5/target12:0:5/12:0:5:0] So I see no reason why you can't connect the SSDs to the Intel ports, it's just a case of swapping things around.
  14. If you can't connect it to the onboard SATA at least connect it to the LSI controller, like the other SSD is, still no trim but at least it's a reliable controller.
  15. Also no issues for me with a ConnectX-3, probably on a old firmware though, since I never updated.
  16. Domains share is set to exclude disk1, and you don't have any other array disks or pools assigned.
  17. These suggest one of the devices dropped offline in the past: Nov 15 16:44:17 Eden kernel: BTRFS info (device sdk1): bdev /dev/sdj1 errs: wr 29362, rd 138395, flush 0, corrupt 4, gen 0 Nov 15 16:44:17 Eden kernel: BTRFS info (device sdk1): bdev /dev/sdk1 errs: wr 0, rd 0, flush 0, corrupt 4, gen 0 And it dropped offline again: Nov 16 00:49:27 Eden kernel: ata11.00: failed to IDENTIFY (I/O error, err_mask=0x1) Nov 16 00:49:27 Eden kernel: ata11.00: revalidation failed (errno=-5) Nov 16 00:49:27 Eden kernel: ata11.00: disabled Try replacing the cables on that device, and ideally don't use the SASLP controller which is not recommended, or at least connect the SSDs to onboard SATA, or trim also won't work, then run a correcting scrub and see if there aren't any uncorrectable errors.
  18. Because the problem is corruption, not failure, in this case looks like LUKS corruption, filesystem corruption can have similar results, i.e., parity won't help, but you can still try what I posted above, it should work.
  19. On second though, that won't work, since the basic flash drive will be blacklisted, I don't know why you just don't use the existing Pro flash drive, my oldest servers are almost 14 years ols and still using the original flash drives, but you'd still need to contact LT to see if what you want is possible. https://unraid.net/contact
  20. Post the output of: fdisk -l /dev/sdX for both disks.
×
×
  • Create New...