You bet!
Code:
[root@plexnas] ~# smartctl -a /dev/da11
smartctl 6.3 2014-07-26 r3976 [FreeBSD 9.3-RELEASE-p13 amd64] (local build)
Copyright (C) 2002-14, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: HGST Deskstar NAS
Device Model: HGST HDN724040ALE640
Serial Number: PK2338P4H9LW8C
LU WWN Device Id: 5 000cca 249d275b7
Firmware Version: MJAOA5E0
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 7200 rpm
Form Factor: 3.5 inches
Device is: In smartctl database [for details use: -P show]
ATA Version is: ATA8-ACS T13/1699-D revision 4
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s)
Local Time is: Thu Apr 30 06:12:32 2015 PDT
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART Status not supported: Incomplete response, ATA output registers missing
SMART overall-health self-assessment test result: PASSED
Warning: This result is based on an Attribute check.
General SMART Values:
Offline data collection status: (0x80) Offline data collection activity
was never started.
Auto Offline Data Collection: Enabled.
Self-test execution status: ( 0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: ( 24) seconds.
Offline data collection
capabilities: (0x5b) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
No Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 1) minutes.
Extended self-test routine
recommended polling time: ( 599) minutes.
SCT capabilities: (0x003d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x000b 100 100 016 Pre-fail Always - 0
2 Throughput_Performance 0x0005 136 136 054 Pre-fail Offline - 81
3 Spin_Up_Time 0x0007 130 130 024 Pre-fail Always - 596 (Average 595)
4 Start_Stop_Count 0x0012 100 100 000 Old_age Always - 59
5 Reallocated_Sector_Ct 0x0033 100 100 005 Pre-fail Always - 0
7 Seek_Error_Rate 0x000b 100 100 067 Pre-fail Always - 0
8 Seek_Time_Performance 0x0005 121 121 020 Pre-fail Offline - 34
9 Power_On_Hours 0x0012 100 100 000 Old_age Always - 466
10 Spin_Retry_Count 0x0013 100 100 060 Pre-fail Always - 0
12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 52
192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 69
193 Load_Cycle_Count 0x0012 100 100 000 Old_age Always - 69
194 Temperature_Celsius 0x0002 181 181 000 Old_age Always - 33 (Min/Max 20/46)
196 Reallocated_Event_Count 0x0032 100 100 000 Old_age Always - 0
197 Current_Pending_Sector 0x0022 100 100 000 Old_age Always - 0
198 Offline_Uncorrectable 0x0008 100 100 000 Old_age Offline - 0
199 UDMA_CRC_Error_Count 0x000a 200 200 000 Old_age Always - 0
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
# 1 Extended offline Completed without error 00% 453 -
# 2 Short offline Completed without error 00% 453 -
# 3 Extended offline Completed without error 00% 117 -
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
[root@plexnas] ~#
As a side note, the data on this box (about 11TB now) is a replica of another box, so I decided to play around trying to figure this out - I shut down the box, pulled the drive, reseated it, rebooted and then I see this:
Code:
[root@plexnas] ~# zpool status vol1
pool: vol1
state: DEGRADED
status: One or more devices could not be opened. Sufficient replicas exist for
the pool to continue functioning in a degraded state.
action: Attach the missing device and online it using 'zpool online'.
see: http://illumos.org/msg/ZFS-8000-2Q
scan: none requested
config:
NAME STATE READ WRITE CKSUM
vol1 DEGRADED 0 0 0
raidz2-0 ONLINE 0 0 0
gptid/f46fb4ec-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f69f4e21-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f8cde372-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/faeb3d6d-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/fd087ff0-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/ff28300a-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
raidz2-1 DEGRADED 0 0 0
gptid/013d5491-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/0357b342-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/05811f51-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/079f5f22-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/09b81318-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
475905736839031462 UNAVAIL 0 0 0 was /dev/gptid/0bd7d8b8-ed63-11e4-a956-0cc47a31abcc
errors: No known data errors
So wanting to use the same drive (since the long smart showed no errors again) I did this:
Code:
[root@plexnas] ~# zpool replace vol1 /dev/gptid/0bd7d8b8-ed63-11e4-a956-0cc47a31abcc /dev/da11
and saw this:
Code:
[root@plexnas] ~# zpool status vol1
pool: vol1
state: DEGRADED
status: One or more devices is currently being resilvered. The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Thu Apr 30 06:07:30 2015
28.0G scanned out of 15.6T at 736M/s, 6h10m to go
2.34G resilvered, 0.18% done
config:
NAME STATE READ WRITE CKSUM
vol1 DEGRADED 0 0 0
raidz2-0 ONLINE 0 0 0
gptid/f46fb4ec-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f69f4e21-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f8cde372-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/faeb3d6d-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/fd087ff0-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/ff28300a-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
raidz2-1 DEGRADED 0 0 0
gptid/013d5491-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/0357b342-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/05811f51-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/079f5f22-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/09b81318-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
replacing-5 UNAVAIL 0 0 0
475905736839031462 UNAVAIL 0 0 0 was /dev/gptid/0bd7d8b8-ed63-11e4-a956-0cc47a31abcc
da11 ONLINE 0 0 0 (resilvering)
errors: No known data errors
Then this:
Code:
[root@plexnas] ~# zpool detach vol1 475905736839031462
Which resulted in this:
Code:
[root@plexnas] ~# zpool status vol1
pool: vol1
state: ONLINE
status: One or more devices is currently being resilvered. The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Thu Apr 30 06:08:59 2015
14.6G scanned out of 15.6T at 649M/s, 7h0m to go
1.21G resilvered, 0.09% done
config:
NAME STATE READ WRITE CKSUM
vol1 ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
gptid/f46fb4ec-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f69f4e21-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f8cde372-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/faeb3d6d-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/fd087ff0-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/ff28300a-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
raidz2-1 ONLINE 0 0 0
gptid/013d5491-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/0357b342-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/05811f51-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/079f5f22-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/09b81318-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
da11 ONLINE 0 0 0 (resilvering)
errors: No known data errors
I waited a few minutes and checked again and it appears the resilver is moving right along:
Code:
[root@plexnas] ~# zpool status vol1
pool: vol1
state: ONLINE
status: One or more devices is currently being resilvered. The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Thu Apr 30 06:08:59 2015
1006G scanned out of 15.6T at 1.24G/s, 3h22m to go
83.8G resilvered, 6.28% done
config:
NAME STATE READ WRITE CKSUM
vol1 ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
gptid/f46fb4ec-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f69f4e21-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/f8cde372-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/faeb3d6d-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/fd087ff0-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/ff28300a-ed62-11e4-a956-0cc47a31abcc ONLINE 0 0 0
raidz2-1 ONLINE 0 0 0
gptid/013d5491-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/0357b342-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/05811f51-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/079f5f22-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
gptid/09b81318-ed63-11e4-a956-0cc47a31abcc ONLINE 0 0 0
da11 ONLINE 0 0 0 (resilvering)
errors: No known data errors
So it would appear that my attempt to replace the failed drive with itself has been successful, but I am still wondering why after all of the pre-deployment testing that FreeNas thought the drive was bad. I always stock several brand-new replacement drives and I am wondering if I should just replace it.
My vendor says they run a smart test on any returned drives and if it shows up with no errors they do not take them back, so I would be stuck with a drive that FreeNas does not like (for whatever reason) but when you ask it, it says "I'm just fine"!