- Joined
- Dec 11, 2015
- Messages
- 1,410
Hello!
Im a little bit out of the grooves when it comes to FreeNAS, as I resurrected the box from more or less a year of hibernation, only to be greeted by a degraded volume.
I do remember that a drive acted up several times (particularly when moving or cleaning around the chassis). Presumable a bad SATA connection as smart would rack up errors in "199 UDMA_CRC_ERROR_COUNT". I've not worried, as the degraded volume would normally be fixed by wiggling the sata connector into place and rebooting, running a 0 fix safety scrub.
This time however, more system confusion than normal have ensued. Here is sort of a time line representation:
Shut down the system, wiggled the sata cable, rebooted.
- GUI showed 6/7 drives present. Pool as degraded.
- CLI show 6/7 drives present.
Shut down the system, REPLACED the sata cable, rebooted.
- GUI shows 7/7 drives present. Pool as "healthy" in the Storage->volumes menu. The traffic light shows pool as degraded IIRC.
- CLI shows 7 drives, yet one deviating gptid.
Rebooting:
- GUI shows 7/7 drives present. Pool as "healthy" in the Storage->volumes menu. The traffic light menu: "The volume wd60efrx state is ONLINE: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected."
- CLI :
Reflections:
At this point I recognize there is some sort of resilvering that has begun (did it finish?)
The expected data change from last nights resurrection work could very well correspond to the 4.36GB added, as I did not take notice of the FreeNAS situation immediately.
I'm slightly confused by the system stating "healthy", finding all drives, yet have lost a gptid on a drive in the pool. hm?
Question:
What is the best course of action to fix the broken gptid, and normalize the system?
When searching the forums for similar problems yet seemingly more severe, I find threads dating 2015, saying the drive should be offlined, removed from pool, wiped, then resilvered. Is that necessary for my problem too?
Specs:
FreeNAS 11.1-U6
X11-SSL i3-6100
48GB
9201-16i (20.04.00)
7x WD60EFRX Raidz2
SLOG 120GB SM863
Cheers!
Im a little bit out of the grooves when it comes to FreeNAS, as I resurrected the box from more or less a year of hibernation, only to be greeted by a degraded volume.
I do remember that a drive acted up several times (particularly when moving or cleaning around the chassis). Presumable a bad SATA connection as smart would rack up errors in "199 UDMA_CRC_ERROR_COUNT". I've not worried, as the degraded volume would normally be fixed by wiggling the sata connector into place and rebooting, running a 0 fix safety scrub.
Code:
199 UDMA_CRC_Error_Count 0x0032 200 001 000 Old_age Always - 1452
This time however, more system confusion than normal have ensued. Here is sort of a time line representation:
Shut down the system, wiggled the sata cable, rebooted.
- GUI showed 6/7 drives present. Pool as degraded.
- CLI show 6/7 drives present.
Shut down the system, REPLACED the sata cable, rebooted.
- GUI shows 7/7 drives present. Pool as "healthy" in the Storage->volumes menu. The traffic light shows pool as degraded IIRC.
- CLI shows 7 drives, yet one deviating gptid.
Rebooting:
- GUI shows 7/7 drives present. Pool as "healthy" in the Storage->volumes menu. The traffic light menu: "The volume wd60efrx state is ONLINE: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected."
- CLI :
Code:
zpool status pool: freenas-boot state: ONLINE scan: scrub repaired 0 in 0 days 00:00:16 with 0 errors on Fri Mar 8 03:45:17 2019 config: NAME STATE READ WRITE CKSUM freenas-boot ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 ada0p2 ONLINE 0 0 0 ada1p2 ONLINE 0 0 0 errors: No known data errors pool: wd60efrx state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: http://illumos.org/msg/ZFS-8000-9P scan: resilvered 4.36G in 0 days 00:01:10 with 0 errors on Wed Jul 24 14:08:01 2019 config: NAME STATE READ WRITE CKSUM wd60efrx ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 gptid/14ef1fa6-e0a4-11e5-b134-0cc47ab3208c ONLINE 0 0 0 gptid/15c495ba-e0a4-11e5-b134-0cc47ab3208c ONLINE 0 0 0 gptid/16990bee-e0a4-11e5-b134-0cc47ab3208c ONLINE 0 0 0 gptid/1769399b-e0a4-11e5-b134-0cc47ab3208c ONLINE 0 0 0 gptid/18479def-e0a4-11e5-b134-0cc47ab3208c ONLINE 0 0 0 gptid/1911207e-e0a4-11e5-b134-0cc47ab3208c ONLINE 0 0 0 da6p2 ONLINE 0 0 1 logs gptid/ae50a861-e463-11e8-8177-000c298edbf1 ONLINE 0 0 0 errors: No known data errors
Reflections:
At this point I recognize there is some sort of resilvering that has begun (did it finish?)
The expected data change from last nights resurrection work could very well correspond to the 4.36GB added, as I did not take notice of the FreeNAS situation immediately.
I'm slightly confused by the system stating "healthy", finding all drives, yet have lost a gptid on a drive in the pool. hm?
Question:
What is the best course of action to fix the broken gptid, and normalize the system?
When searching the forums for similar problems yet seemingly more severe, I find threads dating 2015, saying the drive should be offlined, removed from pool, wiped, then resilvered. Is that necessary for my problem too?
Specs:
FreeNAS 11.1-U6
X11-SSL i3-6100
48GB
9201-16i (20.04.00)
7x WD60EFRX Raidz2
SLOG 120GB SM863
Cheers!