mhugelmann
Cadet
- Joined
- Mar 16, 2012
- Messages
- 8
Hey all,
I was running 11.1-U2 for the past month or so without issue. I saw that there were some issues with -U3, so I passed on the update. When -U4 came out (fixing the problems of -U3), I decided to update.
However, upon rebooting the machine, one of my storage pools reported that it was degraded. I checked dmesg, and the system saw the drive, but for some reason didn't allocate it to the pool. (Instead creating a GEOM_ELI device for the disk.)
dmesg | grep ada18:
zpool status -v pool:
I tried rebooting the machine just in case it was something like a timeout error when booting (which has never happened to me before). However, when the system came back up, the error persisted. Same result as the first time.
I thought it was too much of a coincidence that the drive died or that there was some issue with it just as I upgraded the system, so I decided to try downgrading back to -U2. Sure enough, once the system came back up, the pool seemed to be back in a working state. The status of the pool reported that there were 4 checksum errors with the drive that weren't there previously. (I associate this with the upgrade and the system not putting the drive into the pool.)
zpool status -v pool:
As a precaution, I scrubbed the pool just to make sure all the data was intact.
zpool status -v pool:
Since there were 0 errors, I cleared the error.
Has anyone else had this problem with the new update? Could there be any reason that the drive wouldn't work in -U4 but is absolutely fine in -U2? Is this an actual bug and should I be reporting it?
My system's specs are as follows:
Any help is greatly appreciated!
Thank you,
Mike
Edited to use CODE tags instead of SPOILER tags.
I was running 11.1-U2 for the past month or so without issue. I saw that there were some issues with -U3, so I passed on the update. When -U4 came out (fixing the problems of -U3), I decided to update.
However, upon rebooting the machine, one of my storage pools reported that it was degraded. I checked dmesg, and the system saw the drive, but for some reason didn't allocate it to the pool. (Instead creating a GEOM_ELI device for the disk.)
dmesg | grep ada18:
Code:
root@comp:~ # dmesg | grep ada18 ada18 at ata3 bus 0 scbus18 target 0 lun 0 ada18: <WDC WD20EZRX-00D8PB0 80.00A80> ACS-2 ATA SATA 3.x device ada18: Serial Number WD-XXXXXXXXXXXX ada18: 150.000MB/s transfers (SATA 1.x, UDMA5, PIO 8192bytes) ada18: 1907729MB (3907029168 512 byte sectors) ada18: quirks=0x1<4K> GEOM: ada18: the primary GPT table is corrupt or invalid. GEOM: ada18: using the secondary instead -- recovery strongly advised. GEOM_ELI: Device ada18p1.eli created.
zpool status -v pool:
Code:
pool: pool state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Attach the missing device and online it using 'zpool online'. see: http://illumos.org/msg/ZFS-8000-2Q scan: scrub repaired 0 in 0 days 16:13:22 with 0 errors on Fri Mar 23 19:13:23 2018 config: NAME STATE READ WRITE CKSUM pool DEGRADED 0 0 0 raidz1-0 DEGRADED 0 0 0 ada17 ONLINE 0 0 0 14486299466296630956 UNAVAIL 0 0 0 was /dev/ada18 ada19 ONLINE 0 0 0 ada20 ONLINE 0 0 0 cache gpt/pool_cache_1 ONLINE 0 0 0 gpt/pool_cache_2 ONLINE 0 0 0 errors: No known data errors
I tried rebooting the machine just in case it was something like a timeout error when booting (which has never happened to me before). However, when the system came back up, the error persisted. Same result as the first time.
I thought it was too much of a coincidence that the drive died or that there was some issue with it just as I upgraded the system, so I decided to try downgrading back to -U2. Sure enough, once the system came back up, the pool seemed to be back in a working state. The status of the pool reported that there were 4 checksum errors with the drive that weren't there previously. (I associate this with the upgrade and the system not putting the drive into the pool.)
zpool status -v pool:
Code:
pool: pool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: http://illumos.org/msg/ZFS-8000-9P scan: resilvered 156K in 0 days 00:00:01 with 0 errors on Wed Mar 28 00:24:01 2018 config: NAME STATE READ WRITE CKSUM pool ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 ada17 ONLINE 0 0 0 ada18 ONLINE 0 0 4 ada19 ONLINE 0 0 0 ada20 ONLINE 0 0 0 cache gpt/pool_cache_1 ONLINE 0 0 0 gpt/pool_cache_2 ONLINE 0 0 0 errors: No known data errors
As a precaution, I scrubbed the pool just to make sure all the data was intact.
zpool status -v pool:
Code:
pool: pool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was made to correct the error. Applications are unaffected. action: Determine if the device needs to be replaced, and clear the errors using 'zpool clear' or replace the device with 'zpool replace'. see: http://illumos.org/msg/ZFS-8000-9P scan: scrub repaired 0 in 0 days 16:21:40 with 0 errors on Wed Mar 28 16:49:00 2018 config: NAME STATE READ WRITE CKSUM pool ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 ada17 ONLINE 0 0 0 ada18 ONLINE 0 0 4 ada19 ONLINE 0 0 0 ada20 ONLINE 0 0 0 cache gpt/pool_cache_1 ONLINE 0 0 0 gpt/pool_cache_2 ONLINE 0 0 0 errors: No known data errors
Since there were 0 errors, I cleared the error.
Has anyone else had this problem with the new update? Could there be any reason that the drive wouldn't work in -U4 but is absolutely fine in -U2? Is this an actual bug and should I be reporting it?
My system's specs are as follows:
- ECS IC780M-A Motherboard
- AMD Phenom(tm) II X6 1065T Processor
- 16GB DDR2 RAM
- The affected drive is plugged into a Startech PCISATA4R1 SATA expansion card
Any help is greatly appreciated!
Thank you,
Mike
Edited to use CODE tags instead of SPOILER tags.
Last edited: