j5428 Posted July 19, 2012 Share Posted July 19, 2012 Just looking for an opinion on the smart report for this drive. It is the parity drive and has disabled itself 2 times now for the current pending sector error. First time the parity rebuilt ok. Wondering whether to let it rebuild again or just rma the drive. I see there are 81 re-allocated sectors on the drive: === START OF INFORMATION SECTION === Device Model: Hitachi HDS722020ALA330 Serial Number: JK1101Y Firmware Version: JKAOA3EA User Capacity: 2,000,398,934,016 bytes Device is: Not in smartctl database [for details use: -P showall] ATA Version is: 8 ATA Standard is: ATA-8-ACS revision 4 Local Time is: Thu Jul 19 16:09:01 2012 EDT SMART support is: Available - device has SMART capability. SMART support is: Enabled === START OF READ SMART DATA SECTION === SMART overall-health self-assessment test result: PASSED General SMART Values: Offline data collection status: (0x84) Offline data collection activity was suspended by an interrupting command from host. Auto Offline Data Collection: Enabled. Self-test execution status: ( 0) The previous self-test routine completed without error or no self-test has ever been run. Total time to complete Offline data collection: (22330) seconds. Offline data collection capabilities: (0x5b) SMART execute Offline immediate. Auto Offline data collection on/off support. Suspend Offline collection upon new command. Offline surface scan supported. Self-test supported. No Conveyance Self-test supported. Selective Self-test supported. SMART capabilities: (0x0003) Saves SMART data before entering power-saving mode. Supports SMART auto save timer. Error logging capability: (0x01) Error logging supported. General Purpose Logging supported. Short self-test routine recommended polling time: ( 1) minutes. Extended self-test routine recommended polling time: ( 255) minutes. SCT capabilities: (0x003d) SCT Status supported. SCT Feature Control supported. SCT Data Table supported. SMART Attributes Data Structure revision number: 16 Vendor Specific SMART Attributes with Thresholds: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 1 Raw_Read_Error_Rate 0x000b 100 100 016 Pre-fail Always - 0 2 Throughput_Performance 0x0005 133 133 054 Pre-fail Offline - 102 3 Spin_Up_Time 0x0007 115 115 024 Pre-fail Always - 626 (Average 627) 4 Start_Stop_Count 0x0012 100 100 000 Old_age Always - 789 5 Reallocated_Sector_Ct 0x0033 100 100 005 Pre-fail Always - 81 7 Seek_Error_Rate 0x000b 100 100 067 Pre-fail Always - 0 8 Seek_Time_Performance 0x0005 121 121 020 Pre-fail Offline - 35 9 Power_On_Hours 0x0012 099 099 000 Old_age Always - 8761 10 Spin_Retry_Count 0x0013 100 100 060 Pre-fail Always - 0 12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 71 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 1053 193 Load_Cycle_Count 0x0012 100 100 000 Old_age Always - 1053 194 Temperature_Celsius 0x0002 153 153 000 Old_age Always - 39 (Lifetime Min/Max 16/68) 196 Reallocated_Event_Count 0x0032 100 100 000 Old_age Always - 81 197 Current_Pending_Sector 0x0022 100 100 000 Old_age Always - 1 198 Offline_Uncorrectable 0x0008 100 100 000 Old_age Offline - 0 199 UDMA_CRC_Error_Count 0x000a 200 200 000 Old_age Always - 0 SMART Error Log Version: 1 ATA Error Count: 8 (device log contains only the most recent five errors) CR = Command Register [HEX] FR = Features Register [HEX] SC = Sector Count Register [HEX] SN = Sector Number Register [HEX] CL = Cylinder Low Register [HEX] CH = Cylinder High Register [HEX] DH = Device/Head Register [HEX] DC = Device Command Register [HEX] ER = Error register [HEX] ST = Status register [HEX] Powered_Up_Time is measured from power on, and printed as DDd+hh:mm:SS.sss where DD=days, hh=hours, mm=minutes, SS=sec, and sss=millisec. It "wraps" after 49.710 days. Error 8 occurred at disk power-on lifetime: 8760 hours (365 days + 0 hours) When the command that caused the error occurred, the device was active or idle. After command completion occurred, registers were: ER ST SC SN CL CH DH -- -- -- -- -- -- -- 40 51 a1 66 72 3f 04 Commands leading to the command that caused the error were: CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name -- -- -- -- -- -- -- -- ---------------- -------------------- 60 00 00 07 71 3f 40 00 40d+16:50:35.334 READ FPDMA QUEUED 60 00 00 07 6f 3f 40 00 40d+16:50:35.312 READ FPDMA QUEUED 60 00 00 07 6d 3f 40 00 40d+16:50:35.298 READ FPDMA QUEUED 60 00 00 07 6b 3f 40 00 40d+16:50:35.233 READ FPDMA QUEUED 60 00 00 07 69 3f 40 00 40d+16:50:35.221 READ FPDMA QUEUED Error 7 occurred at disk power-on lifetime: 8273 hours (344 days + 17 hours) When the command that caused the error occurred, the device was active or idle. After command completion occurred, registers were: ER ST SC SN CL CH DH -- -- -- -- -- -- -- 40 51 b6 99 78 3f 04 Commands leading to the command that caused the error were: CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name -- -- -- -- -- -- -- -- ---------------- -------------------- 60 00 00 4f 78 3f 40 00 16d+04:58:44.479 READ FPDMA QUEUED 60 10 00 3f 78 3f 40 00 16d+04:58:44.447 READ FPDMA QUEUED 60 00 00 3f 76 3f 40 00 16d+04:58:44.433 READ FPDMA QUEUED 60 40 00 ff 75 3f 40 00 16d+04:58:44.430 READ FPDMA QUEUED 60 00 00 ff 73 3f 40 00 16d+04:58:44.416 READ FPDMA QUEUED Error 6 occurred at disk power-on lifetime: 6901 hours (287 days + 13 hours) When the command that caused the error occurred, the device was active or idle. After command completion occurred, registers were: ER ST SC SN CL CH DH -- -- -- -- -- -- -- 40 51 d9 3e 27 dd 07 Commands leading to the command that caused the error were: CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name -- -- -- -- -- -- -- -- ---------------- -------------------- 60 00 00 17 27 dd 40 00 15d+08:54:05.629 READ FPDMA QUEUED 60 00 00 17 25 dd 40 00 15d+08:54:05.622 READ FPDMA QUEUED 60 00 00 17 23 dd 40 00 15d+08:54:05.615 READ FPDMA QUEUED 60 00 00 17 21 dd 40 00 15d+08:54:05.608 READ FPDMA QUEUED 61 00 00 17 1f dd 40 00 15d+08:54:05.562 WRITE FPDMA QUEUED Error 5 occurred at disk power-on lifetime: 4635 hours (193 days + 3 hours) When the command that caused the error occurred, the device was active or idle. After command completion occurred, registers were: ER ST SC SN CL CH DH -- -- -- -- -- -- -- 40 51 11 76 ad 52 0a Commands leading to the command that caused the error were: CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name -- -- -- -- -- -- -- -- ---------------- -------------------- 60 b8 00 cf ac 52 40 00 35d+10:10:24.192 READ FPDMA QUEUED 60 48 00 87 ab 52 40 00 35d+10:10:24.184 READ FPDMA QUEUED 60 08 00 7f aa 52 40 00 35d+10:10:24.176 READ FPDMA QUEUED 60 f8 00 87 a9 52 40 00 35d+10:10:24.167 READ FPDMA QUEUED 60 78 00 0f a9 52 40 00 35d+10:10:24.159 READ FPDMA QUEUED Error 4 occurred at disk power-on lifetime: 4212 hours (175 days + 12 hours) When the command that caused the error occurred, the device was active or idle. After command completion occurred, registers were: ER ST SC SN CL CH DH -- -- -- -- -- -- -- 40 51 3c 43 a7 52 0a Commands leading to the command that caused the error were: CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name -- -- -- -- -- -- -- -- ---------------- -------------------- 60 00 00 7f a5 52 40 00 27d+07:14:19.329 READ FPDMA QUEUED 60 e8 00 97 a4 52 40 00 27d+07:14:19.318 READ FPDMA QUEUED 60 18 00 7f a3 52 40 00 27d+07:14:19.307 READ FPDMA QUEUED 60 00 00 7f a1 52 40 00 27d+07:14:19.294 READ FPDMA QUEUED 60 00 00 7f 9f 52 40 00 27d+07:14:19.286 READ FPDMA QUEUED SMART Self-test log structure revision number 1 No self-tests have been logged. [To run self-tests, use: smartctl -t] SMART Selective self-test log data structure revision number 1 SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS 1 0 0 Not_testing 2 0 0 Not_testing 3 0 0 Not_testing 4 0 0 Not_testing 5 0 0 Not_testing Selective self-test flags (0x0): After scanning selected spans, do NOT read-scan remainder of disk. If Selective self-test is pending on power-up, resume after 0 minute delay. Link to comment
dgaschk Posted July 19, 2012 Share Posted July 19, 2012 I'd run pre-clear on it at least twice. If the pending goes to zero and the reallocated sectors remains stable then use it, otherwise RMA. Link to comment
tyrindor Posted July 20, 2012 Share Posted July 20, 2012 Personally, I would just RMA a drive that has had that many sectors go bad if it's under warranty. If it's not under warranty then I would preclear it once or twice and see if the numbers increase, if they do, then replace it. Link to comment
Joe L. Posted July 20, 2012 Share Posted July 20, 2012 Personally, I would just RMA a drive that has had that many sectors go bad if it's under warranty. If it's not under warranty then I would preclear it once or twice and see if the numbers increase, if they do, then replace it. I agree. Link to comment
j5428 Posted July 20, 2012 Author Share Posted July 20, 2012 Thats what I was thinking also. RMA being sent out today. I wish hitachi would cross ship. Link to comment
mr-hexen Posted July 20, 2012 Share Posted July 20, 2012 I also agree. 81 sectors in 26 days. RMA it. Link to comment
UhClem Posted July 20, 2012 Share Posted July 20, 2012 I also agree. 81 sectors in 26 days. RMA it. Where do you see that the 81 Reallocated_Sectors occurred within 26 days? Link to comment
mr-hexen Posted July 20, 2012 Share Posted July 20, 2012 3 Spin_Up_Time 0x0007 115 115 024 Pre-fail Always - 626 (Average 627) 626hrs / 24 = 26.xxx days. Link to comment
Joe L. Posted July 21, 2012 Share Posted July 21, 2012 3 Spin_Up_Time 0x0007 115 115 024 Pre-fail Always - 626 (Average 627) 626hrs / 24 = 26.xxx days. How would you then explain these (also in the same report) 9 Power_On_Hours 0x0012 099 099 000 Old_age Always - 8761 Error 8 occurred at disk power-on lifetime: 8760 hours (365 days + 0 hours) The drive has about a years worth of hours spinning... But the history suggests an RMA. Joe L. Link to comment
UhClem Posted July 21, 2012 Share Posted July 21, 2012 3 Spin_Up_Time 0x0007 115 115 024 Pre-fail Always - 626 (Average 627) 626hrs / 24 = 26.xxx days. I was hoping I would be learning a new SMART trick on this one, but ... Spin_Up_Time, whatever its meaning, is: + Not cumulative + Not measured in hours [ I have several of the same drive (7K2000 2TB) and checked some old SMART reports. E.g., a drive with 3 Power_On_Hours had Spin_Up_Time of 477 (w/ no Average) and one with 32 POH had Spin_Up_Time of 587 (Average 621) ] Link to comment
dgaschk Posted July 21, 2012 Share Posted July 21, 2012 Spin up time is a measure of the time it takes for a disk to change from a stopped state to an operational one. The meaning of the value is not documented. Link to comment
Recommended Posts
Archived
This topic is now archived and is closed to further replies.