Neo_x

Members
  • Posts

    116
  • Joined

  • Last visited

Everything posted by Neo_x

  1. And so i am at a loss. 1TB DD test completed - no failures at all - after almost 3 days of continuous writing ... stress test which even a parity check cant match. (see attached results) will try upgrading firmware on my supermicro MV8 controllers now - just to be sure. still outstanding (only ideas i have) Smart tests for the discs and possibly a memory test... md5__1TB.zip
  2. you are copying to a user share , and thus the disk you are copying from is part of the share - and by default the files already exist on it. i previously had to follow these steps(my example disk 2 is red balled) 1. On web management page for configuring user shares, remove disk 2 from inclusion in any shares (or exclude it) (eg. disk1,disk2,disk4 -> disk1,disk4) 2. open up MC , and browse to the /mnt/disk2 folder on both windows. rename all the primary share folders (eg movies/series/music) to movies_ / series_ / Music_ 3. once done, open the renamed folder on the left, and your destination disk / user share on the right, and move the files inside over using F6. Should not encounter any duplicates any more. Repeat as neccesary for all share folders 4. Once all files is moved (verified via main page that the drive is almost empty ( mine had max 30MB on it), you still need to reset the array configuration(utils -> new config), and rebuild it excluding the affected drive (keeping your parity drive exactly the same drive as before...) Do keep in mind - your system is currently at risk, and will only be free from data loss once step 3 is completed
  3. *bump* i am currently running a 1TB DD test with 5 repetitions - problem being that on 22 drives it is pushing the cpu to 100% all the time, thus i am not sure how accurate it is going to be. It is going to take another 48 hours - if power holds .... any other ideas please?
  4. Hi guys unfortunately reverting to stock and re correcting the parity didn't work. :( 2nd check is still going, but it doesn't seem to get better (strange part - "some" (not all) of the error blocks seems to be repeating between the two syslog's.) any recommendations as to how i could find out what the cause of this issue could be?? *panicking here* rc12a_syslog_1.zip rc12a_syslog2.zip
  5. ok done guess you need me to redo parity checks with correction enabled) PS i notice RC12a has released. - should i upgrade just to be safe(although i don't notice anything parity related) *edit* ok restarted stock with rc12a. running a parity check with corrections now- once the second pass has completed, will i submit syslog's again (its already finding errors though.....)
  6. Hi guys i am struggling the last week or two to get a fully stable unraid going, and unfortunately already suffered from a data loss due to a parity issue i am having. Every time i check the parity(with correction) enabled, it finds over 100,000 errors ( all grouped strangely in the first 500MB of the discs.) anyhow - after a drive failing, and i was forced to rebuild from the "corrupt" parity , the new drive had lots of reiserFS issues - which i hope i was able to correct ( time will tell if files was damaged) I thus went on a investigation to find out why the parity could be failing. One possible culprit it seems was cards with Sil3132 chipsets as per the following thread : http://lime-technology.com/forum/index.php?topic=21052.0 In order to avoid time wastage, i thus copied data from the drives on that controller and removed them from the array ( eg reset the configuration from scratch and rebuilt the array excluding the 2 drives) unfortunately, my parity error persisted in the next two parity checks i performed ( see attached syslog's) I followed the steps as prescribed in the following wiki : http://lime-technology.com/wiki/index.php?title=FAQ#How_To_Troubleshoot_Recurring_Parity_Errors but i am stuck unfortunately. script i used ( modified a little bit to allow parameter passing, since 22 scripts didn't make sense ) I based the skip and count on the syslog entries ( although the errors didn't repeat, both sets seems to be below the 1,100,000 range. All the MD5 calculations 25 repetitions (performed simultaneously on the 22 disc's) came out correctly ..... This leads me to believe the disc's is correct. (i also repeated this with a much larger data chunk and could find any fault...) I am running 5.0-rc11 with simplefeatures / Unmenu and i think SNAP as plugins two main controllers is Supermicro PCI Express x4 Low Profile SAS RAID Controller (AOC-SASLP-MV8) , and the 6 other ports i am using from the MB Is there any other tests i can perform in order to determine why my parity is in such a state? Thx Neo_x PS Syslog's / MD5-results and smart reports is attached -just in case i overlooked something. *********************************SOLVED******************************************* took quite a while, but eventually managed pinpoint the issue to s specific two SATA ports on the MB ( sata ports is grouped in two configurations - 4 standard ports, and then these two in some or other RAID capable setup (thus possibly using another chipset). not exactly sure why they only start to act up when i connect drives via my supermicro add-on cards (eg on their own i was able to generate a clean parity). for now i have just simply disconnected them from the setup - willl look at upgrading the setup to a more standardized version when the time requires. - i consolidated data and removed all the old drives form the setup - so for now all is well syslog_parity_check_1.zip syslog_parity_check_2.zip md5.zip smart.zip
  7. Hi guys i have a peculiar set of circumstances. roughly 3 days ago, one of my older 1.5TB drives dropped out of the array, and came back inconsistently after a restart. I then opted to rather take the drive out of the array, and replace it with a 3TB drive i had spare. problem is - i suspect my parity is having issues, since after regenerating the 3TB drive, reiserfck reported alot of errors on it ( so much so that i had to run rebuild-tree). My feeling is the data on the 1.5TB is more trustworthy ( or i would want to compare data with spot checks between the two to be sure). My goal is to copy data from the old 1.5TB to the new 3TB , but i am not having luck in mounting it via unmenu or even SNAP ( i presumed this is because the drive was an UNRAID drive before..?) can you please assist me in manually mounting the drive? i will run a reisrfsck against it(/dev/sda1) just to be sure it is not partition issue causing the mount to fail thx Neo_x *edit* reiserfsck is reporting alot of errors. sighess - will have the checks complete first before proceding *edit2* confirmed - drive was beyond rescue - will try and preclear it at some stage
  8. that was my understanding as well..... It will, only if the "emhttp" process is still active and listening on its input for the event triggered by the power button. It will NOT terminate processes actively having open files or current working directories on the data disks. It will wait "forever" for the disks to be idle, or until your syslog uses all available memory with "attempting to unmount" messages. very interesting point i have now reverted to stock 5.0.RC10 unraid, and will wait for the parity check to finish before trying a shutdown again ( will possibly do 2 checks just to be safe) interestingly enough , when i attempted to backup the flash drive it refused to copy all the files (thus a file system issue on the flash) - after a windows scandisk it was fine - thus it is possible that my flash is n its way to memory-heaven will give feedback once checks is completed (+-3days) IF the flash drive is not writeable on shutdown a parity check will be triggered on startup. I have to gather that the problematic flash drive was a possible cause - after 2 consecutive checks on Rc10, it stayed healthy after a restart. thx for the input guys
  9. that was my understanding as well..... It will, only if the "emhttp" process is still active and listening on its input for the event triggered by the power button. It will NOT terminate processes actively having open files or current working directories on the data disks. It will wait "forever" for the disks to be idle, or until your syslog uses all available memory with "attempting to unmount" messages. very interesting point i have now reverted to stock 5.0.RC10 unraid, and will wait for the parity check to finish before trying a shutdown again ( will possibly do 2 checks just to be safe) interestingly enough , when i attempted to backup the flash drive it refused to copy all the files (thus a file system issue on the flash) - after a windows scandisk it was fine - thus it is possible that my flash is n its way to memory-heaven will give feedback once checks is completed (+-3days)
  10. that was my understanding as well..... anyhoooow as per my first post - the issue occured after a power button power down as well as a GUI power down. 3rd time isn't the charm , even with the power down script from the console i get issues (ie after restart parity regenerates again) link below contains the syslog after the restart http://pastebin.com/MkhW6pKq i am guessing the next recommendation would be to upgrade to latest stock version without add-ons?
  11. Hi guys hope someone can point me in the right direction. i had it happen twice this week already, so i am rather sure i am not losing my marbles due to controller issues, my HDD's doesnt spin down, and thus to safe on electricity costs i normally power down the server in the evenings. issues is - tuesday afternoon ( after powering it down via power button the previous evening, the parity started an automatic check, and in the syslog a line is shown that an unclean shutdown was detected. Parity regen takes roughly 38hiours, and low and behold, thursday evening i power it down via the web gui ( stopping the array and then shutdown), and the same issue repeats(currently running a party check since last night ) i have an external XBMC connected to the array (thus probable sickbeard could have been busy updating scraping information), and the Unraid box also have the folder caching script running. Any recommendations as to why an unclean shutdown could be detected, as i would like to avoid having to regenerate/check the parity every time i power it up ( stresses all the components....) current version : 5.0-rc8a if an before and after syslog is needed, i will test with the power down script via the putty console in an few hours (i noted it makes a back-up before the powerdown) thx for any ideas Neo_X
  12. *update* after previous post with mdcmd seeming to work, i went ahead to enable the default spin down delay to 15minutes (array is very rarely used with only one client currently connected, so i suspect almost all drives to powerdown. in stead i keep seeing(and hearing) the array to be powered up and powered down continuously - not even obeying the 15minute rule.... any thoughts as to why? syslog is attached. thx *edit* noticed in the syslog a reference to Adaptec aacraid driver 1.2-0. on the website i indicated in the OP, its currently already on version 1.5 for Fedora Core versions of Linux. syslog4.zip
  13. Actually, the correct command to spindown disk 1 in the array is: mdcmd spindown 1 I think dgaschk just forgot to change the name of the actual command in his response. Hi Joe thanx for the update. would rather prefer to test on the parity in stead of a data drive. what dis number will the parity be? regards parity disk would be mdcmd spindown 0 if you are not at the "root" home directory, you might have to invoke the "mdcmd" command with the full path: /root/mdcmd spindown 0 ok tried it on drive 0 Dec 8 12:38:31 Storage in.telnetd[17654]: connect from 192.168.0.2 (192.168.0.2) Dec 8 12:38:36 Storage login[17655]: ROOT LOGIN on '/dev/pts/0' from '192.168.0.2' Dec 8 12:43:20 Storage kernel: mdcmd (77): spindown 0 Dec 8 12:43:20 Storage kernel: Dec 8 12:43:29 Storage kernel: sd 1:1:0:0: [sda] Spinning up disk....ready Dec 8 12:43:29 Storage kernel: sd 1:1:0:0: [sda] 5851033600 512-byte logical blocks: (2.99 TB/2.72 TiB) Dec 8 12:46:02 Storage kernel: mdcmd (78): spinup 0 Dec 8 12:46:02 Storage kernel: verified via Unmenu , and the spinning icon did change ( although hdparm-C still showed the particular device as active/idle still) guess next step is to enable spin down on all drives and check again?
  14. Actually, the correct command to spindown disk 1 in the array is: mdcmd spindown 1 I think dgaschk just forgot to change the name of the actual command in his response. Hi Joe thanx for the update. would rather prefer to test on the parity in stead of a data drive. what dis number will the parity be? regards
  15. Hi guys I am having some issues with automated drive spin down on my Unraid box, and would like to know if you can assist me to troubleshoot and possibly resolve it. I initially configured the drives to automatically spin-down via the HTTP interface, but discovered that the system becomes very unstable after a drive spin's down. thus i rather opted to disable automated spin-down for now, in order to troubleshoot it manually before attempting it again. reading up on the forums it seems spindown is achieved/managed currently by the HDPARM command unfortunately even manually trying to sleep the drive doesnt seem to work see output below : *********************************************************************************************** root@Storage:~# hdparm -C /dev/sdt /dev/sdt: drive state is: active/idle root@Storage:~# hdparm -i /dev/sdt /dev/sdt: HDIO_GET_IDENTITY failed: Inappropriate ioctl for device root@Storage:~# hdparm -I /dev/sdt /dev/sdt: ATA device, with non-removable media Model Number: ST31500341AS Serial Number: 9VS2ZFF9 Firmware Revision: CC1H Transport: Serial Standards: Used: unknown (minor revision code 0x0029) Supported: 8 7 6 5 Likely used: 8 Configuration: Logical max current cylinders 16383 16383 heads 16 16 sectors/track 63 63 -- CHS current addressable sectors: 16514064 LBA user addressable sectors: 268435455 LBA48 user addressable sectors: 2930277168 Logical/Physical Sector size: 512 bytes device size with M = 1024*1024: 1430799 MBytes device size with M = 1000*1000: 1500301 MBytes (1500 GB) cache/buffer size = unknown Nominal Media Rotation Rate: 7200 Capabilities: LBA, IORDY(can be disabled) Queue depth: 32 Standby timer values: spec'd by Standard, no device specific minimum R/W multiple sector transfer: Max = 16 Current = ? Recommended acoustic management value: 254, current value: 0 DMA: mdma0 mdma1 mdma2 udma0 udma1 udma2 udma3 udma4 udma5 *udma6 Cycle time: min=120ns recommended=120ns PIO: pio0 pio1 pio2 pio3 pio4 Cycle time: no flow control=120ns IORDY flow control=120ns Commands/features: Enabled Supported: * SMART feature set Security Mode feature set * Power Management feature set * Write cache * Look-ahead * Host Protected Area feature set * WRITE_BUFFER command * READ_BUFFER command * DOWNLOAD_MICROCODE SET_MAX security extension * Automatic Acoustic Management feature set * 48-bit Address feature set * Device Configuration Overlay feature set * Mandatory FLUSH_CACHE * FLUSH_CACHE_EXT * SMART error logging * SMART self-test * General Purpose Logging feature set * WRITE_{DMA|MULTIPLE}_FUA_EXT * 64-bit World wide name Write-Read-Verify feature set * WRITE_UNCORRECTABLE_EXT command * {READ,WRITE}_DMA_EXT_GPL commands * Segmented DOWNLOAD_MICROCODE * Gen1 signaling speed (1.5Gb/s) * Gen2 signaling speed (3.0Gb/s) * Native Command Queueing (NCQ) * Phy event counters Device-initiated interface power management * Software settings preservation * SMART Command Transport (SCT) feature set * SCT Long Sector Access (AC1) * SCT LBA Segment Access (AC2) * SCT Error Recovery Control (AC3) * SCT Features Control (AC4) * SCT Data Tables (AC5) unknown 206[12] (vendor specific) Security: Master password revision code = 65534 supported not enabled not locked not frozen not expired: security count supported: enhanced erase 256min for SECURITY ERASE UNIT. 256min for ENHANCED SECURITY ERASE UNIT. Logical Unit WWN Device Identifier: 5000c50019dc5212 NAA : 5 IEEE OUI : 000c50 Unique ID : 019dc5212 Checksum: correct root@Storage:~# root@Storage:~# hdparm -Y /dev/sdt /dev/sdt: issuing sleep command root@Storage:~# hdparm -C /dev/sdt /dev/sdt: No such device or address root@Storage:~# *********************************************************************************************** Thus as per above, after issuing the hdparm - Y on the device, the drive just seems to disappear. Restarting the unraid box is the only way to restore the array - and even worse - seems to invalidate the parity. syslog is attached just in case I am thinking this is a possible linux driver issue, but i am not knowledgeable enough in order to know if drivers can be updated in order to verify if it is resolved. I found the listed linux drivers in the below link : Adaptec Linux drivers page can be found at : http://www.adaptec.com/en-us/support/raid/sas_raid/sas-52445/ I would definitely prefer rather to fix the functionality on my current controller, as i am stuck in south-Africa, and thus obtaining some of the supported controllers will just be too expensive. Any assistance and/or ideas will be appreciated Regards Neo_X syslog3.txt
  16. Hope you get the issue sorted - sounds like a tricky one may i also recommend that once the issue is solved, that we place the answer on the wiki ( especially since it will be related to V5 - the invalidslot command seems to be not fully supported on V5 anymore) sorry i can't assist tho - must be driving you crazy
  17. really not sure why speed will be slower internally.. but if the drive is USB2.0, you cant expect much more from it (20~25MB/s is roughly the max i have seen).
  18. yup - when it comes to trusting my data - i will almost never again trust MS - they need to come to the party rock solid lost many software RAID-5's before ( yes bad idea... but believe me - unstable / unpredictable doesn't begin to describe it) hope they resolve it though -feature wise - or at least the features that was sold - seems to be very nice indeed
  19. I did consider cases like this when I added that "-d type" option. You are one of the first to report they were able to use it on their hardware. ... .. dev/sdb and /dev/sdg will never be the same disk. You can type: ls -l /dev/disk/by-id/* to see a listing of all your disks and disk partitions by model and serial number. The preclear script was written to not allow you to clear a disk that is assigned to the array, or mounted and in use. Have fun, Joe L. Thx for an informative post Joe yes the controller cost me a pretty penny (more in the region of penny wise pound foolish ). about 2/3 years ago i was a strong follower of RAID5/ RAID6 - and thus decided i needed a good controller that can give me the biggest RAID-6 possible relative to my case capacity at that stage - 16 drives max. problem is - as with any data - i outgrow the server recently, and thus needed to upgrade. Problem is - although its a very nice controller, it was following the same rules of RAID. all drives needs to be the same capacity -and it seems it had another limitation - 16 drives max for a cluster. so yes - i wasn't willing to shell out additional cash upgrading the 16 x 1.5TB drives to 16 x 3TB drives, as this would just be pointless(converting the RAID would have taken days if not weeks - in the region of 30 hours per drive), and then i would be stuck in the same position a few years down the line. so yes - now i am moving over to Unraid - and definitely seems to be the way to go. otherwise - nope are having a "normal" desktop pc in use ( Tri-SLI board with one of the first Quad cores(Q6600) running and i think in the region of 4GB RAM - nothing serious ( home user ). the controller have a duo core cpu and 512MB ram onboard- so i think i am safe performance wise. i am seemingly having some issues with spin-up / spin-down on the controller - will investigate a bit more once i finished migrating data over from backups onto the server. will keep you guys updated
  20. Sorry to resurrect - but i seem to have exactly the same issue - and thus the same request. I had a Series5 Adaptec controller spare after migrating hardware from my old to the Unraid server. and thought it should work nicely in my unraid build(and save some $$$'s in order to find another 24 port capable card) unfortunately - drive temperatures is not showing - only via the smartctl -d sat --all command Spin-down is a bit buggy as well - ie if i click on the spin down all button, i can hear all the drives spin down, but after a few seconds they spin-up again still filling the server with data, so spin-down is not a must as yet, but would love to try and get it working - if possible. any ideas is welcome thx Neo_x *edit* Currently using V5 RC5 syslog.txt
  21. cant remember the exact thread i used, but your steps is correct ( i am currently on RC5). I recently required to move all the data from my unraid array due to a planned server migration (new case ,new drives, same unraid license) since i didn't want to go the route where i add and migrate data one dive at a time on the new server (not trusting my ,linux skills i guess ) , i opted for moving all the data from the disk i want to clear, restoring it with parity rebuilt where neccesary(think i had to do it twice due to capacity issues), and then continuing. moving the data was another mission - but basically had to remove from user share, reboot, rename the folders on the source drive, and then copy via MC back to the user share. its a good idea to keep screenshots handy of drives you want to keep /remove - mistakes can be costly good luck Neo_X
  22. Hi guys nvm - Solved - discovered after studying the usage script that the following command is possible:\ preclear_disk.sh -d sat /dev/sda this instructs preclear to utilize alternate commands when running Smartctl. hope someone can assist - i am running Unraid via an Adaptec controller (Model 52445). It is rather overkill for unraid since it is meant for high levels of RAID, but i didn't want to shell out additional $$$'s to get another controller - currently its performing admirably with roughly 64MB/s on post read clearing 12 drives at the same time everything seems fine - ie i set all the disks up as JBOD - which seems to simluate pass through(not sure on the correct terms) anyhow - so far so good. Unraid picks up the first set of 12 drives i connected (having power issues with connecting more ). Problem i am having - it seems that smartctl doesnt give correct stats on the drive. See sample output below. root@Storage:~# smartctl -a /dev/sdb smartctl 5.40 2010-10-16 r3189 [i486-slackware-linux-gnu] (local build) Copyright (C) 2002-10 by Bruce Allen, http://smartmontools.sourceforge.net Device: ST2000DL003-9VT1 Version: CC3C Serial number: 6YD1RLL6 Device type: disk Transport protocol: SAS Local Time is: Sat Sep 1 08:46:46 2012 SAST Device supports SMART and is Enabled Temperature Warning Disabled or Not Supported SMART Health Status: OK Error Counter logging not supported Device does not support Self Test logging root@Storage:~# as can be expected - this messes up Preclear a bit, since it is unable to read Smart results before ,during and after. (although otherwise - it doesn't crash or halt the preclear in any way - GREAT SCRIPT JOE L ) I managed to find a smartctl command that does give the output for the drive as required Thank you Google root@Storage:~# smartctl -d sat --all /dev/sg1 smartctl 5.40 2010-10-16 r3189 [i486-slackware-linux-gnu] (local build) Copyright (C) 2002-10 by Bruce Allen, http://smartmontools.sourceforge.net === START OF INFORMATION SECTION === Device Model: ST2000DL003-9VT166 Serial Number: 6YD1RLL6 Firmware Version: CC3C User Capacity: 2,000,398,934,016 bytes Device is: Not in smartctl database [for details use: -P showall] ATA Version is: 8 ATA Standard is: ATA-8-ACS revision 4 Local Time is: Sat Sep 1 08:47:26 2012 SAST SMART support is: Available - device has SMART capability. SMART support is: Enabled === START OF READ SMART DATA SECTION === SMART overall-health self-assessment test result: PASSED See vendor-specific Attribute list for marginal Attributes. General SMART Values: Offline data collection status: (0x82) Offline data collection activity was completed without error. Auto Offline Data Collection: Enabled. Self-test execution status: ( 0) The previous self-test routine completed without error or no self-test has ever been run. Total time to complete Offline data collection: ( 612) seconds. Offline data collection capabilities: (0x7b) SMART execute Offline immediate. Auto Offline data collection on/off support. Suspend Offline collection upon new command. Offline surface scan supported. Self-test supported. Conveyance Self-test supported. Selective Self-test supported. SMART capabilities: (0x0003) Saves SMART data before entering power-saving mode. Supports SMART auto save timer. Error logging capability: (0x01) Error logging supported. General Purpose Logging supported. Short self-test routine recommended polling time: ( 1) minutes. Extended self-test routine recommended polling time: ( 255) minutes. Conveyance self-test routine recommended polling time: ( 2) minutes. SCT capabilities: (0x30b7) SCT Status supported. SCT Feature Control supported. SCT Data Table supported. SMART Attributes Data Structure revision number: 10 Vendor Specific SMART Attributes with Thresholds: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 1 Raw_Read_Error_Rate 0x000f 109 099 006 Pre-fail Always - 24419656 3 Spin_Up_Time 0x0003 090 090 000 Pre-fail Always - 0 4 Start_Stop_Count 0x0032 100 100 020 Old_age Always - 286 5 Reallocated_Sector_Ct 0x0033 100 100 036 Pre-fail Always - 0 7 Seek_Error_Rate 0x000f 073 060 030 Pre-fail Always - 4318984658 9 Power_On_Hours 0x0032 096 096 000 Old_age Always - 3906 10 Spin_Retry_Count 0x0013 100 100 097 Pre-fail Always - 0 12 Power_Cycle_Count 0x0032 100 100 020 Old_age Always - 286 183 Runtime_Bad_Block 0x0032 100 100 000 Old_age Always - 0 184 End-to-End_Error 0x0032 100 100 099 Old_age Always - 0 187 Reported_Uncorrect 0x0032 100 100 000 Old_age Always - 0 188 Command_Timeout 0x0032 100 100 000 Old_age Always - 0 189 High_Fly_Writes 0x003a 099 099 000 Old_age Always - 1 190 Airflow_Temperature_Cel 0x0022 059 023 045 Old_age Always In_the_past 41 (75 200 42 25) 191 G-Sense_Error_Rate 0x0032 100 100 000 Old_age Always - 0 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 285 193 Load_Cycle_Count 0x0032 100 100 000 Old_age Always - 286 194 Temperature_Celsius 0x0022 041 077 000 Old_age Always - 41 (0 14 0 0) 195 Hardware_ECC_Recovered 0x001a 036 015 000 Old_age Always - 24419656 197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 0 198 Offline_Uncorrectable 0x0010 100 100 000 Old_age Offline - 0 199 UDMA_CRC_Error_Count 0x003e 200 200 000 Old_age Always - 0 240 Head_Flying_Hours 0x0000 100 253 000 Old_age Offline - 66043712114499 241 Total_LBAs_Written 0x0000 100 253 000 Old_age Offline - 2825939409 242 Total_LBAs_Read 0x0000 100 253 000 Old_age Offline - 3158544765 SMART Error Log Version: 1 No Errors Logged SMART Self-test log structure revision number 1 No self-tests have been logged. [To run self-tests, use: smartctl -t] SMART Selective self-test log data structure revision number 1 SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS 1 0 0 Not_testing 2 0 0 Not_testing 3 0 0 Not_testing 4 0 0 Not_testing 5 0 0 Not_testing Selective self-test flags (0x0): After scanning selected spans, do NOT read-scan remainder of disk. If Selective self-test is pending on power-up, resume after 0 minute delay. So i guess my question is then - how do i go about trusting a drive after a preclear? reading through some of the posts, it seems i need to look for the following : FAILING NOW attributes, 5 Reallocated_Sector_Ct (this should be preferably zero - or else stay a very low number.) 197 Current_Pending_Sector (this should be preferably zero - or else stay a very low number.) also - should i be worried about which "device" i am clearing? (since i gather that SDB and SDG is possibly the same thing...) clear should be completed in about 6 hours - will report on any results i don't understand Thank you Neo_x PS syslog attached just in case syslog.zip
  23. thx for the advise - a spare drives is definitely a better idea it seems i somehow supplied an older syslog the drive is back and working now (new syslog attached - no missing drives) as a safety step i did do a reizerfsck on it, and nothing seems out of place. will keep an eye on it for the time being syslog2.txt
  24. Hi guys I had a stressful morning - with some very odd problems on my Unraid server. I noticed the last few days that i have a suspect data drive( drive works 100% - but the moment i power down the system and up again, the drive have the tendency to be missing from the array, and on a 2nd or 3rd reboot (and alot of cursing / moving/swapping of cables in between ) - it returned. in any event - i completed a preclear and parity check yesterday , and all seems well - but this morning i encountered the same issue again after power down , but this time unfortunately the drive just didn't want to return to the array. thinking that all is lost, i then moved the drive to my "trusty" windows system, from where it was detected immediately - not exactly sure how that could be, since i tried difference SATA ports on the unraid machine to no avail. i even kept the drive powered via the Unraid PSU with just a sata cable to the windows machine - so i was stumped. in any event, i thought it feasible to try and clean the drive - which i did via windows disk management deleting the partition and creating a simple windows partition on it. (note - i didn't do a full format - windows probably only did a quick format..) then i continued on to get a firmware upgrade from seagate(which the drive was a few versions behind). this passed without issues all is well - my expectation then being that once i move the disk back to unraid i will need to restore the data onto the drive. problem being - powering on my Unraid system with the drive connected back onto the array, unraid started without reporting any problems on the array can anyone with some experience on drives please assist me with what i should do now? will a parity check (without correction) suffice, or should i trust the data and the drive (considering the partition action i performed under windows), or should i rather restart the drive from scratch (considering that the drive already contains about 600GB of data...) (i am not even sure on what steps i should follow to do this... - i guess it will involve moving data to other discs, removing from array, preclearing, and then adding to array again?) any assistance / recommendations would be highly appreciated *rather safe than sorry * Neo_x syslog.zip pre-clear-report.zip
  25. no , dont have it enabled. fixed my time zone now just in case did manage to solve (updated my first post) Solved - turns out some files on my source server had a incorrect modified date 2098/01/01 (which got further damaged .modified after copying it to unraid / Linux), causing me to be unable to open them with VLC media player. corrected it with an application called touch (adjusting modified date/time) obtained from http://www.stevemiller.net/downloads/ctb10w32.zip