Hi All,
There's going to be a lot of content here, but I appreciate anyone who can provide some assistance. At this point in time, i've been down for about 2 months. I've been running off my backup, which has been fine but I'm ready to get back to using my primary hardware.
Here is my Previous Post where I had a plethora of errors. I think I've now narrowed it down to a bad IOM and/or bad QSFP cable/connector on my array. I also made a Reddit Post and got some ideas for troubleshooting. At this point I'm still stuck on these GPT corruption issues, and I've seen some people have these with disk failures, and they dd the disk to another disk, and seems like that works things out. What i'd like to know is if I can simply repair the GPT's and get the pool back online?
System Specs:
Dell R720xd
2x E5-2670
384gb RAM
LSI SAS-9207-8e
NetApp DS4246 (2x)
FreeNAS 11.1-U7 AND 11.2-U7
At this moment I'm running in a VM, but the same errors are exhibited when running bare metal. The pool was built and has always been run on bare metal. I was going to build my new system as a VM and it was when I had FreeNAS loaded up, that I saw all vdevs were back online (I'd been struggling with only 2 of them showing up, replacing a cable seemed to to resolve this)
The errors I'm getting during boot are things along these lines:
Using gpart list/status/show, none of the disks in vdev2(raidz3-1) show up.
the disks do show up when doing
but their GPT both primary and secondary seem to be fully gone.
Results of zpool import:
I've seen several ideas.
1) dd the disk to a new disk. (https://forums.freebsd.org/threads/zfs-geom-mfid1-corrupt-or-invalid-gpt-detected.2139/)
2) wipe the first and last part of the disk -
3) give up and start fresh and restore from backup.
I tried option 2, and did not seem to make any difference. I feel like there's got to be a way to rebuild the GPT without having to dd the entire disk? We're talking 12 2TB Disks. I have space, I could DD each disk to a file, and then wipe the disk, and dd back - would that work?
Is this something that is recoverable or should I just let it go and move on?
There's going to be a lot of content here, but I appreciate anyone who can provide some assistance. At this point in time, i've been down for about 2 months. I've been running off my backup, which has been fine but I'm ready to get back to using my primary hardware.
Here is my Previous Post where I had a plethora of errors. I think I've now narrowed it down to a bad IOM and/or bad QSFP cable/connector on my array. I also made a Reddit Post and got some ideas for troubleshooting. At this point I'm still stuck on these GPT corruption issues, and I've seen some people have these with disk failures, and they dd the disk to another disk, and seems like that works things out. What i'd like to know is if I can simply repair the GPT's and get the pool back online?
System Specs:
Dell R720xd
2x E5-2670
384gb RAM
LSI SAS-9207-8e
NetApp DS4246 (2x)
FreeNAS 11.1-U7 AND 11.2-U7
At this moment I'm running in a VM, but the same errors are exhibited when running bare metal. The pool was built and has always been run on bare metal. I was going to build my new system as a VM and it was when I had FreeNAS loaded up, that I saw all vdevs were back online (I'd been struggling with only 2 of them showing up, replacing a cable seemed to to resolve this)
The errors I'm getting during boot are things along these lines:
GEOM: da35: the secondary GPT table is corrupt or invalid.
GEOM: da35: using the primary only -- recovery suggested.
GEOM: multipath/disk1: corrupt or invalid GPT detected.
GEOM: multipath/disk1: GPT rejected -- may not be recoverable.
root@freenas[/var/log]# gpart recover /dev/da25
gpart: arg0 'da25': Invalid argument
Using gpart list/status/show, none of the disks in vdev2(raidz3-1) show up.
the disks do show up when doing
camcontrol devlist all
(or sysctl kern.disks
but their GPT both primary and secondary seem to be fully gone.
Results of zpool import:
Code:
root@freenas[/var/log]# zpool import pool: tank id: 907423887126384953 state: UNAVAIL status: The pool was last accessed by another system. action: The pool cannot be imported due to damaged devices or data. see: http://illumos.org/msg/ZFS-8000-EY config: tank UNAVAIL insufficient replicas raidz3-0 DEGRADED gptid/e6f85169-5136-11e8-b376-bc305bf48148 ONLINE 731087118901044337 UNAVAIL cannot open gptid/181f4f1a-5137-11e8-b376-bc305bf48148 ONLINE gptid/24da255f-5137-11e8-b376-bc305bf48148 ONLINE gptid/325a8e49-5137-11e8-b376-bc305bf48148 ONLINE gptid/4a4df2c1-5137-11e8-b376-bc305bf48148 ONLINE gptid/6a52ee8d-5137-11e8-b376-bc305bf48148 ONLINE gptid/7bc6a747-5137-11e8-b376-bc305bf48148 ONLINE gptid/8d306d6b-5137-11e8-b376-bc305bf48148 ONLINE gptid/a82f0ce2-5137-11e8-b376-bc305bf48148 ONLINE 1014468244676781008 UNAVAIL cannot open gptid/db7bb774-5137-11e8-b376-bc305bf48148 ONLINE raidz3-1 UNAVAIL insufficient replicas 8358001936465582577 UNAVAIL cannot open 13168953014485996713 UNAVAIL cannot open 16674343073354040627 UNAVAIL cannot open 5916167911648451258 UNAVAIL cannot open 10228142690658451320 UNAVAIL cannot open 17309881899809254171 UNAVAIL cannot open 6521923195479449726 UNAVAIL cannot open 16254298423903963076 UNAVAIL cannot open 12519560841692417252 UNAVAIL cannot open 16658654953139949791 UNAVAIL cannot open 11783566184643368972 UNAVAIL cannot open 17155630495564972158 UNAVAIL cannot open raidz3-2 ONLINE gptid/bc7a19f9-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/bd626022-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/bef543fe-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c0e17dc6-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c27fc9e2-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c365fdbb-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c46b5d52-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c60d3684-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c7a6bc30-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/c9416b43-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/ca343c33-c4cb-11e8-b538-bc305bf48148 ONLINE gptid/cb3f2f2c-c4cb-11e8-b538-bc305bf48148 ONLINE
I've seen several ideas.
1) dd the disk to a new disk. (https://forums.freebsd.org/threads/zfs-geom-mfid1-corrupt-or-invalid-gpt-detected.2139/)
2) wipe the first and last part of the disk -
dd if=/dev/zero of=/dev/label/spare1 bs=512 count=1
3) give up and start fresh and restore from backup.
I tried option 2, and did not seem to make any difference. I feel like there's got to be a way to rebuild the GPT without having to dd the entire disk? We're talking 12 2TB Disks. I have space, I could DD each disk to a file, and then wipe the disk, and dd back - would that work?
Is this something that is recoverable or should I just let it go and move on?
Last edited: