garys

Members
  • Posts

    43
  • Joined

  • Last visited

Converted

  • Gender
    Undisclosed

garys's Achievements

Rookie

Rookie (2/14)

0

Reputation

  1. Alright - ran a correcting and fixed 400+ errors, then a non-correcting and came back with 0. I'm considering this issue resolved. Thanks again for all the help - appreciate it greatly!
  2. Ok - I've started the parity check (correcting). I'll post back with how it went. Thanks!
  3. Finally done with the memory tests - got the bad stick isolated (turned out to be four 8GB sticks). Also ran several rounds of testing with the two sticks I plan to use for now, and they've come out unscathed. So, for the next step, would you say another non-correcting parity check, or just go straight to the correcting one? Thanks!
  4. Thanks - I will try testing with 1 stick at a time and see how that goes. I'll post back the results when finished either this evening or in the morning. Here's hoping it's 2 sticks (and not 1 or 4!)
  5. Hey - just an update... You were dead-on with the suggestion run a memtest... it's only been running a few hours and it's already reported 2 errors. Specifically: Tst:9, Pass:0, Failing Address:0042afffb58 - 17071.9MB, Good: c90606c3, Bad: c90606c1, Err-Bits: 00000002, Count: 1, CPU: 0 I don't know much(anything) about memory or memory tests. Have run a handful of tests over the years and have always come out clean in the past. This has me pretty concerned, actually. Knowing that a memory issue is actually causing data issues that are probably not recoverable. Obviously it's why we have backups, but still a bit unnerving. So, the obvious next question is, where to go from here. I'm not 100% sure if it's 2 sticks or 4 (I'm thinking it's 2), so wondering if I should be able to tell which stick has the problem from the error info. If not, I could run the test with only 1 stick in, then switch to the other. I don't think I'm using even close to half of available memory, at least not currently, so i might be able to get by on one stick for awhile. Other than doing the above, I'd be open to other suggestions. Thanks for all the help so far!
  6. Hey - thanks for the replies! So, just to be clear, even though it I'm getting a notification that says Parity check finished (324 errors), there's nothing that needs to be done about that? Here are the log entries I was referring to: May 6 16:08:17 nas1 kernel: REISERFS error (device sde1): vs-5150 search_by_key: invalid format found in block 661592588. Fsck? May 6 16:08:17 nas1 kernel: REISERFS error (device sde1): vs-13070 reiserfs_read_locked_inode: i/o failure occurred trying to find stat data of [1 2 0x0 SD] I thought it was odd too, that it would be referring to parity, but the device mentioned (sde) is my parity disk. What could that mean? Haha, yeah, for some reason I decided to skip that slot because of how the drives were positioned in my 5in3 bays when i did a re-org years ago. No idea why I did that now, but it hasn't seemed to cause any issues, so I've left it. I have 13 data disks + parity. Interesting. Ok - I will run the memtest now and post back the results. Thanks!
  7. Hi All, Been running on Unraid for a long time now (10+ years), and more recently on 6.9.2 on an asus z370/i7-8700k build (14 drives of various sizes, few dockers, couple of VMs). I've been thinking about converting my REISERFS drives (which is all of them, haha) to xfs for quite some time. So, when the urge hit me to start working on that a couple days ago, I decided to do a non-correcting parity check first . I don't have a scheduled one and hadn't done one in almost a year, so it was definite long overdue! Parity check came back with what it called errors (324, but showing up in the log as green?), but also, strangely still says 'Parity is valid' on the Dashboard... I skimmed through the syslog and that's when I noticed a few REISERFS errors that appear to be happening only on the parity drive. Not sure if these are related. The parity errors all look like the below, with different sector #s referenced: Feb 10 06:23:05 nas1 kernel: md: recovery thread: P incorrect, sector=128 Probably unrelated: There were also some BTRFS errors on my cache drive, which I think I have addressed by following some advice in another post. I found the file from the inode referenced, removed it, and ran a scrub (non-correcting) that came back clean after the file removal. Just thought I'd include this bit, as it's responsible for most of the log errors. On the REISERFS/parity check issue, I'm looking for recommended next steps. Diagnostic attached. Thanks! nas1-diagnostics-20230211-0859.zip
  8. I'm having slow write performance also. Can you share the link to the version you rolled back to? Thanks! EDIT: Found the following thread (see first post by Squid on page 85) and followed the steps to roll back to 05-03-2018 version ... fixed! https://lime-technology.com/forums/topic/54648-preclear-plugin/?page=85
  9. Thanks for the info, mbryanr. Your links, plus more searching for the other errors have me feeling a bit better about these (it seems others have experienced similar). It still seems to be a bit of a mystery, but my preclears finished with no additional errors, so I'm taking that as a good sign. At this point, I think I'm going to go ahead and add the new drives (on the new supermicro card) to the array and just monitor the syslog for awhile and see what happens. Thanks to all for the info and advice. -Gary
  10. Hi, icon123. Just wanted to let you know, I just got done w/ preclearing my two new wd20ears drives on the supermicro card and they precleared fine. It did throw a few errors in the syslog during the preclear, but again, looking at other posts, that seems to be somewhat 'normal' with this card. I'm going to add the drives to the array and will definitely post back if I have any issues after that. Hopefully, you'll get to the bottom of your lockup issues. Good luck! -Gary
  11. New stuff in syslog since last post: Mar 27 03:01:05 NAS1 ata_id[29269]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 08:25:54 NAS1 dhcpcd[1445]: sending DHCP_REQUEST for 192.168.1.92 to 192.168.1.1 (Routine) Mar 27 08:25:54 NAS1 dhcpcd[1445]: dhcpIPaddrLeaseTime=86400 in DHCP server response. (Routine) Mar 27 08:25:54 NAS1 dhcpcd[1445]: DHCP_ACK received from (192.168.1.1) (Routine) Mar 27 09:01:13 NAS1 ata_id[6036]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 09:01:35 NAS1 ata_id[6109]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 09:01:45 NAS1 ata_id[6160]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 09:01:52 NAS1 ata_id[6198]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 09:02:03 NAS1 ata_id[6246]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 09:02:08 NAS1 ata_id[6273]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 27 09:02:08 NAS1 kernel: sdc: sdc1 (Drive related) Mar 27 09:02:18 NAS1 kernel: udev: starting version 141 (System) Mar 27 09:30:01 NAS1 kernel: sdb: sdb1 (Drive related) Mar 27 09:30:12 NAS1 kernel: udev: starting version 141 (System) These occurred while doing a full preclear on the two new drives that are attatched to the new supermicro card. Anybody got any clues on these (and the ones in the previous post)? Thanks -Gary
  12. On last boot, the following popped into the syslog about 3 minutes after boot completed. Is this something to be concerned about? Mar 26 20:29:30 NAS1 kernel: sdb: sdb1 (Drive related) Mar 26 20:29:30 NAS1 kernel: ------------[ cut here ]------------ Mar 26 20:29:30 NAS1 kernel: WARNING: at drivers/ata/libata-core.c:5186 ata_qc_issue+0x10b/0x308() (Minor Issues) Mar 26 20:29:30 NAS1 kernel: Hardware name: A760G M2+ Mar 26 20:29:30 NAS1 kernel: Modules linked in: md_mod xor atiixp ahci r8169 mvsas libsas scst scsi_transport_sas (Drive related) Mar 26 20:29:30 NAS1 kernel: Pid: 3864, comm: smartctl Not tainted 2.6.32.9-unRAID #8 (Errors) Mar 26 20:29:30 NAS1 kernel: Call Trace: (Errors) Mar 26 20:29:30 NAS1 kernel: [<c102449e>] warn_slowpath_common+0x60/0x77 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c10244c2>] warn_slowpath_null+0xd/0x10 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11b624d>] ata_qc_issue+0x10b/0x308 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11ac0ba>] ? scsi_init_sgtable+0x5b/0x76 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11ba260>] ata_scsi_translate+0xd1/0xff (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11a816c>] ? scsi_done+0x0/0xd (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11a816c>] ? scsi_done+0x0/0xd (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11baa40>] ata_sas_queuecmd+0x120/0x1d7 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11bc6df>] ? ata_scsi_pass_thru+0x0/0x21d (Errors) Mar 26 20:29:30 NAS1 kernel: [<f843369a>] sas_queuecommand+0x65/0x20d [libsas] (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11a816c>] ? scsi_done+0x0/0xd (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11a82c0>] scsi_dispatch_cmd+0x147/0x181 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11ace4d>] scsi_request_fn+0x351/0x376 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1126798>] __blk_run_queue+0x78/0x10c (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1124446>] elv_insert+0x67/0x153 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11245b8>] __elv_add_request+0x86/0x8b (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1129343>] blk_execute_rq_nowait+0x4f/0x73 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11293dc>] blk_execute_rq+0x75/0x91 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11292cc>] ? blk_end_sync_rq+0x0/0x28 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11295fa>] ? blk_recount_segments+0x16/0x24 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11250c6>] ? blk_rq_bio_prep+0x47/0x78 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1128edb>] ? blk_rq_append_bio+0x14/0x3b (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1129121>] ? blk_rq_map_user+0x12c/0x1be (Errors) Mar 26 20:29:30 NAS1 kernel: [<c112c2bf>] sg_io+0x22d/0x30a (Errors) Mar 26 20:29:30 NAS1 kernel: [<c112471f>] ? elv_queue_empty+0x20/0x22 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c112c5a8>] scsi_cmd_ioctl+0x20c/0x3bc (Errors) Mar 26 20:29:30 NAS1 kernel: [<c104cbfe>] ? get_page_from_freelist+0x2e4/0x344 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11b3257>] sd_ioctl+0x6a/0x8c (Errors) Mar 26 20:29:30 NAS1 kernel: [<c112a420>] __blkdev_driver_ioctl+0x50/0x62 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c112ad1c>] blkdev_ioctl+0x8b0/0x8dc (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1131e2d>] ? kobject_get+0x12/0x17 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c112b0f8>] ? get_disk+0x4a/0x61 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1131e2d>] ? kobject_get+0x12/0x17 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1131e2d>] ? kobject_get+0x12/0x17 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1192208>] ? get_device+0x11/0x18 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c11334a5>] ? radix_tree_lookup_slot+0xd/0xf (Errors) Mar 26 20:29:30 NAS1 kernel: [<c104a179>] ? filemap_fault+0xb8/0x305 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1048c43>] ? unlock_page+0x18/0x1b (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1057c63>] ? __do_fault+0x3a7/0x3da (Errors) Mar 26 20:29:30 NAS1 kernel: [<c10757d4>] ? do_filp_open+0x3d8/0x6d1 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c105985f>] ? handle_mm_fault+0x42d/0x8f1 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c108b6c6>] block_ioctl+0x2a/0x32 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c108b69c>] ? block_ioctl+0x0/0x32 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c10769d5>] vfs_ioctl+0x22/0x67 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1076f33>] do_vfs_ioctl+0x478/0x4ac (Errors) Mar 26 20:29:30 NAS1 kernel: [<c106afbf>] ? fd_install+0x1e/0x43 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1075b02>] ? set_close_on_exec+0x1e/0x38 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1076030>] ? do_fcntl+0x139/0x390 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1076f93>] sys_ioctl+0x2c/0x45 (Errors) Mar 26 20:29:30 NAS1 kernel: [<c1002935>] syscall_call+0x7/0xb (Errors) Mar 26 20:29:30 NAS1 kernel: ---[ end trace 048e91acf2138a64 ]--- Mar 26 20:29:47 NAS1 kernel: sdb: sdb1 (Drive related) Mar 26 20:30:13 NAS1 sshd[4128]: error: Could not get shadow information for root (Errors) Mar 26 20:30:20 NAS1 sshd[4169]: lastlog_filetype: Couldn't stat /var/log/lastlog: No such file or directory Mar 26 20:30:20 NAS1 sshd[4169]: lastlog_openseek: /var/log/lastlog is not a file or directory! Mar 26 20:30:20 NAS1 sshd[4169]: lastlog_filetype: Couldn't stat /var/log/lastlog: No such file or directory Mar 26 20:30:20 NAS1 sshd[4169]: lastlog_openseek: /var/log/lastlog is not a file or directory! Mar 26 20:30:40 NAS1 ata_id[4458]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Mar 26 20:30:40 NAS1 kernel: sdc: sdc1 (Drive related) Mar 26 20:33:22 NAS1 ntpd[1453]: synchronized to 204.62.14.98, stratum 2 Mar 26 21:26:00 NAS1 kernel: mdcmd (18): spindown 0 (Routine) Mar 26 21:26:00 NAS1 kernel: mdcmd (19): spindown 1 (Routine) Mar 26 21:26:00 NAS1 kernel: mdcmd (20): spindown 2 (Routine) Mar 26 21:26:00 NAS1 kernel: mdcmd (21): spindown 3 (Routine) Mar 26 21:26:01 NAS1 kernel: mdcmd (22): spindown 4 (Routine) Mar 26 21:26:01 NAS1 kernel: mdcmd (23): spindown 5 (Routine) Full syslog also attached. Thanks, -Gary syslog.txt
  13. Thanks, Raj. Good to see that another system is having similar drive detection on the SALSP card. I'm not in a hurry to add the drives to the array, so I'll try another preclear on them and see how that comes out. I'll post back the results.
  14. Ok, new round of changes and testing based on posted suggestions (thanks!): 1) Shut down the box and pulled the temp 4-port SIL card I had in there for preclearing the new drives while waiting for the correct cable for the SM card. 1a) Booted into bios and verified settings were correct (see above post), then booted into SM bios and disabled int13. Rebooted. - results: same odd drive types when issuing a preclear -l. /dev/sdb = ata-WDC_WD20EARS-22MVWB0_WD-WCAZA3958401 /dev/sdc = scsi-SATA_WDC_WD20EARS-22_WD-WCAZA3952545 2) Shut down and swapped sata cables 0 and 1 on the drive end of the new forward break-out cable. So, basically I switched which drives the cables were plugged into. - result: same exact ids (in the same order) only serial #s reversed. This means the mis-identification seems to be with the port on the card and not with the drives themselves. /dev/sdb = ata-WDC_WD20EARS-22MVWB0_WD-WCAZA3952545 /dev/sdc = scsi-SATA_WDC_WD20EARS-22_WD-WCAZA3958401 3) Shut down and swapped out cable #1 for cable #2. So, I changed out the first forward break-out cable for the second one (just to verify it's not the cable) - result: same exact behavior as above. 4) Shut down and changed break-out cable from top port on the card to the bottom port on the card. - result: same exact behavior as above. 5) Shut down and unplugged SATA 0 and 1 on drive end of break-out cable and plugged in SATA 2 and 3 instead. - result: same exact behavior as above. So, nothing I've tried so far has changed the fact that it's seeing the drive types strangely. I also re-seated the SM card (again) during one of the shutdowns. I also took copies of the syslog each time after booting up. One piece of good news (maybe?) is that the nasty looking code-dump stuff at the end of the first syslog I posted in this thread never reared its head again. Here's it is, for reference (maybe someone can give an idea of what caused it): Mar 25 00:30:48 NAS1 kernel: WARNING: at drivers/ata/libata-core.c:5186 ata_qc_issue+0x10b/0x308() (Minor Issues) Mar 25 00:30:48 NAS1 kernel: Pid: 3426, comm: smartctl Not tainted 2.6.32.9-unRAID #8 (Errors) Mar 25 00:30:48 NAS1 kernel: Call Trace: (Errors) Mar 25 00:30:48 NAS1 kernel: [<c102449e>] warn_slowpath_common+0x60/0x77 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c10244c2>] warn_slowpath_null+0xd/0x10 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11b624d>] ata_qc_issue+0x10b/0x308 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11ac0ba>] ? scsi_init_sgtable+0x5b/0x76 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11ba260>] ata_scsi_translate+0xd1/0xff (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11a816c>] ? scsi_done+0x0/0xd (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11a816c>] ? scsi_done+0x0/0xd (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11baa40>] ata_sas_queuecmd+0x120/0x1d7 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11bc6df>] ? ata_scsi_pass_thru+0x0/0x21d (Errors) Mar 25 00:30:48 NAS1 kernel: [<f843369a>] sas_queuecommand+0x65/0x20d [libsas] (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11a816c>] ? scsi_done+0x0/0xd (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11a82c0>] scsi_dispatch_cmd+0x147/0x181 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11ace4d>] scsi_request_fn+0x351/0x376 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1126798>] __blk_run_queue+0x78/0x10c (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1124446>] elv_insert+0x67/0x153 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11245b8>] __elv_add_request+0x86/0x8b (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1129343>] blk_execute_rq_nowait+0x4f/0x73 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11293dc>] blk_execute_rq+0x75/0x91 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11292cc>] ? blk_end_sync_rq+0x0/0x28 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11295fa>] ? blk_recount_segments+0x16/0x24 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11250c6>] ? blk_rq_bio_prep+0x47/0x78 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1128edb>] ? blk_rq_append_bio+0x14/0x3b (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1129121>] ? blk_rq_map_user+0x12c/0x1be (Errors) Mar 25 00:30:48 NAS1 kernel: [<c112c2bf>] sg_io+0x22d/0x30a (Errors) Mar 25 00:30:48 NAS1 kernel: [<c112471f>] ? elv_queue_empty+0x20/0x22 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c112c5a8>] scsi_cmd_ioctl+0x20c/0x3bc (Errors) Mar 25 00:30:48 NAS1 kernel: [<c104cbfe>] ? get_page_from_freelist+0x2e4/0x344 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11b3257>] sd_ioctl+0x6a/0x8c (Errors) Mar 25 00:30:48 NAS1 kernel: [<c112a420>] __blkdev_driver_ioctl+0x50/0x62 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c112ad1c>] blkdev_ioctl+0x8b0/0x8dc (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1131e2d>] ? kobject_get+0x12/0x17 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c112b0f8>] ? get_disk+0x4a/0x61 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1131e2d>] ? kobject_get+0x12/0x17 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1131e2d>] ? kobject_get+0x12/0x17 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1192208>] ? get_device+0x11/0x18 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c11334a5>] ? radix_tree_lookup_slot+0xd/0xf (Errors) Mar 25 00:30:48 NAS1 kernel: [<c104a179>] ? filemap_fault+0xb8/0x305 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1048c43>] ? unlock_page+0x18/0x1b (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1057c63>] ? __do_fault+0x3a7/0x3da (Errors) Mar 25 00:30:48 NAS1 kernel: [<c10757d4>] ? do_filp_open+0x3d8/0x6d1 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c105985f>] ? handle_mm_fault+0x42d/0x8f1 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c108b6c6>] block_ioctl+0x2a/0x32 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c108b69c>] ? block_ioctl+0x0/0x32 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c10769d5>] vfs_ioctl+0x22/0x67 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1076f33>] do_vfs_ioctl+0x478/0x4ac (Errors) Mar 25 00:30:48 NAS1 kernel: [<c106afbf>] ? fd_install+0x1e/0x43 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1075b02>] ? set_close_on_exec+0x1e/0x38 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1076030>] ? do_fcntl+0x139/0x390 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1076f93>] sys_ioctl+0x2c/0x45 (Errors) Mar 25 00:30:48 NAS1 kernel: [<c1002935>] syscall_call+0x7/0xb (Errors) Mar 25 00:31:30 NAS1 ata_id[3784]: HDIO_GET_IDENTITY failed for '/dev/block/8:32' (Minor Issues) Here are just the Errors and Minor Issues that I get every time I boot (even with SM card removed): Mar 25 21:55:32 NAS1 kernel: ACPI Error (psargs-0359): [ECEN] Namespace lookup failure, AE_NOT_FOUND (Minor Issues) Mar 25 21:55:32 NAS1 kernel: ACPI Error (psparse-0537): Method parse/execution failed [\] (Node c14760c8), AE_NOT_FOUND (Minor Issues) Mar 25 21:55:32 NAS1 kernel: ACPI Warning: Incorrect checksum in table [OEMB] - 95, should be 88 (20090903/tbutils-314) (Minor Issues) Mar 25 21:55:32 NAS1 kernel: ACPI: I/O resource it87 [0xe85-0xe86] conflicts with ACPI region SENP [0xe85-0xe86] (Minor Issues) Mar 25 21:55:35 NAS1 emhttp: shcmd (29): killall -HUP smbd (Minor Issues) Mar 25 21:55:44 NAS1 apcupsd[1679]: apcupsd FATAL ERROR in linux-usb.c at line 609 Cannot find UPS device -- For a link to detailed USB trouble shooting information, please see <http://www.apcupsd.com/support.html>. (Errors) Mar 25 21:55:44 NAS1 apcupsd[1679]: apcupsd error shutdown completed (Errors) Mar 25 21:57:02 NAS1 sshd[2765]: error: Could not get shadow information for root (Errors) So, here are some of the questions I'm left with: 1) Can I reasonably conclude (probably not, but...) that the nasty syslog messages were caused by a conflict w/ the SIL card I removed before the first reboot on this last round of testing? If not, what other testing can I do to determine the cause? 2) Is it normal to have the first two drives recognized this way on the SM card (does anybody else w/ this card see this?!) ? 3) Are the remaining, consistent syslog issues/errors anything to worry about? 4) Assuming the answers to 1,2,3 are yes,yes & no respectively, should I take the big leap and add these drives to the array? (they passed preclear w/ flying colors, but that was on the SIL card, before I got the new cables) Not sure where to go from here... Thanks in advance for advice/suggestions on this. If needed, I can provide syslogs for any/all of the steps I listed above. I am including the syslog from after step 5 above, for reference. <syslog attached> -Gary edit: added step 1a syslog-20110325-213200.zip
  15. No worries, Raj. For all you do to help people on these forums, you're entitled to a mistake once in awhile! --- I decided not to do a BIOS upgrade just yet - rather exhaust other options since I have a (mostly) working system and I have been burned by unnecessary firmware upgrades in the past - not to say I don't trust your advice, just being extra cautious in this case. Booted into bios and verified all settings - they are default, except for changes recommended in the unRAID wiki specifically for my MB here. Namely, set everything to AHCI, disabled sata/ide combined mode, changed graphics to use smallest amount of memory (32M), changed USB to emulate HD and disabled all boot drives except USB. Done. The SIL3124 card was something I had laying around from my old unraid build. I installed it temporarily to preclear the two new wd20ears drives while I was waiting for the correct cables to arrive (multitasking! ) I simply forgot to remove it when I unhooked the drives from it and hooked up the sm card instead. I agree, I won't be needing it, so it's now removed. My MB only has one suitable (>=4x) pcie slot, so changing slots won't be possible. Ran 8 hours of memtest when i first built the box about 2-3 weeks ago. No problems. So, thanks for the tips! >> see next post for results of new tests done while making above changes.