Error Expanding Volume

Status
Not open for further replies.

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
I'm having an issue expanding a mirrored volume being used as an iscsi target. I originally created it with 3 pairs of 3TB drives but am trying to add another pair and have the iscsi target expand accordingly.

The drives show up as part of the volume but any attempt to make it larger gets a greater than 80% error.
What am I doing wrong?

x2lvWEO.png


Mvq9A0L.png

PTq5roK.png
 
Last edited:

SweetAndLow

Sweet'NASty
Joined
Nov 6, 2013
Messages
6,421
Simple you can't use that much space for your iscsi volume. You should only fill your pool to 80%at the most. And if you do your performance with iscsi is going to be terrible.
 

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
I have 8X3TB drives (mirrored). The iSCSI volume is only 7.5 TB and I am not able to increase the size at all after adding another 2-drive vdev.
 

Spearfoot

He of the long foot
Moderator
Joined
May 13, 2015
Messages
2,478
I have 8X3TB drives (mirrored). The iSCSI volume is only 7.5 TB and I am not able to increase the size at all after adding another 2-drive vdev.
The values from your Volumes screenshot imply a total of ~11.8TiB, which is about what you'd expect to see from a pool of 4 x 3TB mirrored vdevs. You might try 8TiB instead of 8TB when you edit the zvol. Also, don't forget the 'Pool Available Space Threshold (%)' setting under Sharing->Block(iSCSI).

But it seems your pool is already pretty full (64%), even after adding a vdev. To get enough add'l space, you may have to replace one or more of the mirrored vdevs with larger capacity drives.
 

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
The 64% is the same as it was before I added the additional drives. The problem I'm having is that I haven't been able to increase the size of the iscsi target *at all* since adding the new pair. Maybe I created created another volume by mistake? But all I did was extend the existing volume by adding 2 drives. The only thing I want to do with those 8 drives is serve iscsi to esxi.
 

Spearfoot

He of the long foot
Moderator
Joined
May 13, 2015
Messages
2,478
The 64% is the same as it was before I added the additional drives. The problem I'm having is that I haven't been able to increase the size of the iscsi target *at all* since adding the new pair. Maybe I created created another volume by mistake? But all I did was extend the existing volume by adding 2 drives. The only thing I want to do with those 8 drives is serve iscsi to esxi.
Yikes!

It doesn't look like you created another volume; your output from 'zpool status' shows a pool with 4 vdevs, each a mirror. Still, you might show us the output of 'zfs list' in code tags.

Working in GibiBytes... a 3TB disk is 2794GiB, so your 4-mirror pool should be 4 times that, yielding a capacity of 11,178GiB. Adding up the Used (6.4TiB) and Available (4.5TiB) values for your iZFS pool gives ~11TiB. So the pair you added sure seems to be present in your pool. And therefore you ought to be able to extend the zvol to ~8942GiB (80% of pool size) without FreeNAS fussing about it.

You might try the 'Force size' option...

Do you have good backups? I'd be tempted to destroy and then re-create the pool and zvol; restoring the VMs from backup.

On a related note: block storage on FreeNAS really works best on sparsely populated volumes. Performance will suffer when running it as full as you seem to be doing.
 

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
Here is the output from zfs list. The volume is not actually that full - that's just the size of the iscsi datastore created in esxi - although maybe thats the same as full as far as zfs is concerned. Dunno. I have backups but I just finished moving them onto slower media (which freed up the 2x3TB disks that I then added to the zvol). I hate going in circles.


Code:
NAME USED AVAIL REFER MOUNTPOINT
five 3.13T 1.26T 3.13T /mnt/five
five/jails 96K 1.26T 96K /mnt/five/jails
freenas-boot 704M 57.0G 31K none
freenas-boot/.system 2.10M 57.0G 31K legacy
freenas-boot/.system/configs-0827a2da0c5b428287f6c8d2d52005ef 364K 57.0G 364K legacy
freenas-boot/.system/cores 734K 57.0G 734K legacy
freenas-boot/.system/rrd-0827a2da0c5b428287f6c8d2d52005ef 25K 57.0G 25K legacy
freenas-boot/.system/samba4 91.5K 57.0G 91.5K legacy
freenas-boot/.system/syslog-0827a2da0c5b428287f6c8d2d52005ef 904K 57.0G 904K legacy
freenas-boot/ROOT 695M 57.0G 25K none
freenas-boot/ROOT/Initial-Install 1K 57.0G 594M legacy
freenas-boot/ROOT/Pre-9.10-STABLE-201606270534-460219 1K 57.0G 602M legacy
freenas-boot/ROOT/Wizard-2005-01-03_19:48:32 1K 57.0G 598M legacy
freenas-boot/ROOT/default 695M 57.0G 680M legacy
freenas-boot/grub 6.35M 57.0G 6.33M legacy
iZFS 7.62T 2.92T 96K /mnt/iZFS
iZFS/iZFS 7.62T 4.18T 6.35T -
 

Spearfoot

He of the long foot
Moderator
Joined
May 13, 2015
Messages
2,478
@ColPanic, would you please post your detailed system information per the forum rules? This may help with diagnosing your problem.

@jgreco, @cyberjock : do you guys have any suggestions to make about the iSCSI problem the OP is having? In a nutshell, he's expanded his pool by adding a pair of 3TB disks as the 4th vdev, but can't seem to expand the zvol involved.
 

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
Sure..
FreeNAS-9.10-STABLE-201606270534
VMware 6.0.0
S2600CP4 / 2x Xeon E5-2670 (2 cores to freenas)
64GB ECC RAM (16GB locked to freenas)
LSI 9240-8i (IT mode p20 FW) VT-d passthrough
HP SAS Expander
 

cyberjock

Inactive Account
Joined
Mar 25, 2012
Messages
19,526
I'm going to guess that you expanded the zpool, but the FreeNAS middleware still thinks that the zpool is a smaller size, therefore its not letting you expand. I would try a reboot of your FreeNAS machine.

You are in a VM, which we totally frown on for a boatload of reasons, but I doubt that's actually a factor here.
 

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
Oddly, if I typed in 8.9T or 8.9TiB or TB it wouldnt work. But by typing 8900GB it worked. I'm all set. Thanks!
 

Spearfoot

He of the long foot
Moderator
Joined
May 13, 2015
Messages
2,478
Oddly, if I typed in 8.9T or 8.9TiB or TB it wouldnt work. But by typing 8900GB it worked. I'm all set. Thanks!
Great! Did you have to reboot the system?
 

ColPanic

Dabbler
Joined
Jul 19, 2016
Messages
20
I rebooted it several times but this seemed to solve it.
 
Status
Not open for further replies.
Top