Similar system to iX-3216 WWIX do?

jcl123

Dabbler
Joined
Jul 6, 2019
Messages
23
I was looking at the IX Systems website, I have a setup very similar the iX-3216, same chassis, similar MB, CPU, etc.

So, 16x 3.5" slots, and I do have space for some other 2.5" drives as well as adapters if I want to use some of the 3.5" bays with 2.5" drives.

I would like to follow best practice as closely as I can, or at least see what it is. So, I was wondering, are there sample configurations somewhere that would represent the most typical build-out for one of these if someone ordered one?

Such as:
- Typical setup for data-only, # and type of drives, how they would typically be configured, etc.
- Typical setup for iSCSI

I know the short answer is probably there is "no magic", the top one would likely be a RAID-Z2. But what? All 16 drives setup that way?

For the iSCSI, probably some kind of cache and then mirroring.

-JCL
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
Disclaimer; I don't work for iX. ;)

There's no magic - to build the right solution you'll want to understand not just how much storage you need, but also the bandwidth/IOPS/latency performance you want. If you have an existing environment you're looking to upgrade there there's tools that can give you a good perspective; or if it's a home setup where the only person accountable is you, you can make estimates a little more safely. ;)

Sixteen drives gives you the option to do one or two pools with decent results.

Single pool, sixteen drives, pure capacity storage with absolutely no block (and you guarantee it will never get involved) I would be tempted to do as 2x 8-drive Z2 if you need the 12 drives of usable capacity. Otherwise, I would say do mirrors anyways for the better performance (8 vdevs vs. 2) and 8 drives of usable space. Use a pair of (non)mirrored SSDs for an L2ARC if it's only data - I don't think SLOG would be needed in this use case.

iSCSI config is absolutely mirrors. SSDs for L2ARC if you don't have an all-flash pool, and an NVMe/Optane or NVDIMM SLOG, depending on how far your dollar stretches. If I recall though, you've got some HGST SSDs that would do a pretty decent job of SLOG duties, although they won't handle 10Gbps.

If you're considering having one system pull double duty, then you lean towards which you need more of. 6-drive Z2 for file storage, 5x 2-way mirrors for iSCSI, and a few SSDs for L2ARC/SLOG duties is absolutely valid.

Another side note is that TrueNAS 12 might throw some unexpected wrenches into the works that doing a "future-proof design" for could make sense. I can say that special vdevs for metadata are going to be very popular for certain workloads; "block storage on spindles" will almost certainly be one of them.
 
Last edited:

sretalla

Powered by Neutrality
Moderator
Joined
Jan 1, 2016
Messages
9,703
Use a pair of mirrored SSDs for an L2ARC if it's only data
Why would you mirror L2ARC? what's the risk when it fails? I would have expected striped L2ARC would make much more sense since it's not holding the real data.
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
Why would you mirror L2ARC? what's the risk when it fails? I would have expected striped L2ARC would make much more sense since it's not holding the real data.
"Because I wrote that post late at night."
Striping is the intended configuration.
 

jcl123

Dabbler
Joined
Jul 6, 2019
Messages
23
Disclaimer; I don't work for iX. ;)

There's no magic - to build the right solution you'll want to understand not just how much storage you need, but also the bandwidth/IOPS/latency performance you want. If you have an existing environment you're looking to upgrade there there's tools that can give you a good perspective; or if it's a home setup where the only person accountable is you, you can make estimates a little more safely. ;)

Right, I am trying to wrap my brain around this. In reading about ZFS, there are a few things that I think really make it difficult for me when you put them together: (let me know if any of these are not correct)
- A RAID-Z2 VDEV will have the performance of the slowest single disk, yes? (file based not iSCSI)
- This being said, a few months ago I did a test using 10x of the 4TB 7.2K SAS disks in RAID-Z2​
- Anecdotally, performance seemed OK, it did not take that long to copy 20TB, and read speed over 1Gig was decent​
- You want to restrict the number of disks in a VDEV, seems much past 12 or so and you need to start thinking about it
- Of course, if you add another VDEV, you double the number of parity disks when you add the 2nd RAID-Z2 VDEV​
- But, the two VDEVs would be striped, so you would get the performance of the two slowest disks?​
- The size of a VDEV cannot later be increased by adding disks
- You want an abundance of free space in a VDEV to maintain performance due to COW, etc.
- I do have to be mindful of power, heat, and noise as much as I can

Today, the size of all my data is around 20TB or so, and the growth has slowed down to less than 1TB a year.

One of my solutions to all of this is to have the three servers. When it comes time that I am out of space and need to expand, or performance has eroded, use the three servers to maintain a safe number of copies of the data while one of the servers is re-done with newer / more disks and then rinse and repeat. Only one of the servers would be normally on, the 2nd would be for backup of the first say weekly or monthly (malware or cryptoware can't get to it if it is off, yes I know, snapshots), and the third would be spare, or maybe containing a 3rd copy of the data that is only update every 6 months or more.

All three servers follow as many of the recommendations as possible - ECC RAM, eSSDs, eHDD or NAS drives, hardware that is on the compatibility list, etc.

Sixteen drives gives you the option to do one or two pools with decent results.

Single pool, sixteen drives, pure capacity storage with absolutely no block (and you guarantee it will never get involved) I would be tempted to do as 2x 8-drive Z2 if you need the 12 drives of usable capacity. Otherwise, I would say do mirrors anyways for the better performance (8 vdevs vs. 2) and 8 drives of usable space. Use a pair of (non)mirrored SSDs for an L2ARC if it's only data - I don't think SLOG would be needed in this use case.

- Easy to be sure there will be no block, because you would have to re-format to do that anyways
- So, 2x8x Z2 would be approx 24TB each / 48TB total with four parity drives
- So a mirror would be 8x 2-way mirrors, 32TB usable? I imagine that would be quite fast
- One advantage of this is that it could do block or traditional storage, or maybe both?​
- If they are not mirrored, why do you need two drives for L2ARC?

I have constraints to consider unless I make some changes: (still thinking about it)
- Feel free to suggest alternate configurations
- For the first (primary) server, I was planning to use the 7x 10TB SATA drives (low power and heat)
- I could put these in the regular tower case that can hold 11x drives, because they don't put off as much heat​
- I don't see how else you could configure these other than RAID-Z2​
- I was thinking of filling it up with 4TB drives to see how bad the heat problem is or if it is OK​
- For the 2nd server, this would use the Supermicro 16x chassis (seems a shame to not use it as primary, still thinking)
- This is where I would put the 4TB drives because it is designed for that​
- So this would be to backup the first​
- I was going to buy a 2nd one of these before CV19 hit, still considering that​
- The 3rd server is the other tower case with a capacity of 7x drives
- This is a much lower spec server but still good​
- I do have the JBOD which can handle 10x enterprise or NAS drives, not sure how I am going to use this yet
- Not sure I want an array that consists of internal and external drives​
- Worried what would happen if I accidentally didn't power on the JBOD, or it lost power at some point​
- Thinking whatever I do with it, should be restricted to what can fully fit​
- I can also fit 2x 2.5" drives, the expander can handle way more than that if I need to do something else someday​
- I have two nice LSI RAID controllers on the way to handle VMFS duty, which will be all flash

iSCSI config is absolutely mirrors. SSDs for L2ARC if you don't have an all-flash pool, and an NVMe/Optane or NVDIMM SLOG, depending on how far your dollar stretches. If I recall though, you've got some HGST SSDs that would do a pretty decent job of SLOG duties, although they won't handle 10Gbps.

- I do have some 400Gig NVMe, but they are mixed-use, so not sure they can take that many writes
- Yes, the 400Gig and 200Gig disks are write-intensive, so they are a good choice four sure
- The flash drives could mainly be direct-connect vs. on the expander so that they get the best bandwidth
- Right now I am leaning towards not doing iSCSI for data, if I do maybe I take a couple of disks and just do it as a test
- I have 10Gig cards, but I was planning to only use them to connect servers to each other for backups
- It does increase the speed by 2-4x over 1Gig, limited by the disk speed​

If you're considering having one system pull double duty, then you lean towards which you need more of. 6-drive Z2 for file storage, 5x 2-way mirrors for iSCSI, and a few SSDs for L2ARC/SLOG duties is absolutely valid.

- This is an interesting setup
- In that case, I could just put the 7x 10TB SATA drives in RAID-Z2, and use the remaining 9 slots for mirrors
- This does not include internal non-hot-swap bays, I think I can fit around 8x 2.5" disks this way​
- This would give me 50TB of data and 4x mirrors would give 16TB
- I could use the external jbod to test a mirror setup with iSCSI
- 2nd server with enough data to back this up, buy a new case if it can't handle the heat

Another side note is that TrueNAS 12 might throw some unexpected wrenches into the works that doing a "future-proof design" for could make sense. I can say that special vdevs for metadata are going to be very popular for certain workloads; "block storage on spindles" will almost certainly be one of them.

Ah, I don't know anything about this. I know that new version is coming, but I don't know when and I don't know much about features, I thought they were mostly merging FreeNAS with TrueNAS. But, this again is why I will have three servers, so that I can afford to rebuild from time to time when I need to blow something away to take advantage of something new.

Interesting stuff

-JCL
 

HoneyBadger

actually does care
Administrator
Moderator
iXsystems
Joined
Feb 6, 2014
Messages
5,112
- A RAID-Z2 VDEV will have the performance of the slowest single disk, yes? (file based not iSCSI)
- This being said, a few months ago I did a test using 10x of the 4TB 7.2K SAS disks in RAID-Z2​
- Anecdotally, performance seemed OK, it did not take that long to copy 20TB, and read speed over 1Gig was decent​
"Multiple disks worth of sequential I/O - a single disk worth of random I/O."

As a pool gets filled (and erased, and rewritten) it fragments over time. And ZFS has no defragmentation function. So a brand-new pool will be lightning fast as it has gobs of space to write into, and most of the data stays nice and sequential. A file-based share may stay that way for a long time, especially if it's used as a "write once, read many" type of archival or media storage. Block I/O of course is the polar opposite, and very quickly will cause fragmentation. The copy-on-write nature of ZFS helps mitigate the writes, but it can't stop random seeks on reads. So ZFS gets around this by using ARC and L2ARC - or you use all-flash vdevs and "read fragmentation" becomes much less of a concern.

- You want to restrict the number of disks in a VDEV, seems much past 12 or so and you need to start thinking about it
- Of course, if you add another VDEV, you double the number of parity disks when you add the 2nd RAID-Z2 VDEV​
- But, the two VDEVs would be striped, so you would get the performance of the two slowest disks?​
- The size of a VDEV cannot later be increased by adding disks
- You want an abundance of free space in a VDEV to maintain performance due to COW, etc.
- I do have to be mindful of power, heat, and noise as much as I can

All correct, although I would consider a 12-drive Z2 to be the absolute maximum I'd allow. Assuming 12 drives, I'd be far more comfortable with them as 2x 6-drive Z2. Space is cheap, performance isn't.

As far as power/heat/noise, expect each of your 4TB NL-SAS drives to suck down an average of 10W. Take that into consideration for power/heat/noise.

Today, the size of all my data is around 20TB or so, and the growth has slowed down to less than 1TB a year.

A little too big to handle in all-flash, sadly. ;)

One of my solutions to all of this is to have the three servers. When it comes time that I am out of space and need to expand, or performance has eroded, use the three servers to maintain a safe number of copies of the data while one of the servers is re-done with newer / more disks and then rinse and repeat. Only one of the servers would be normally on, the 2nd would be for backup of the first say weekly or monthly (malware or cryptoware can't get to it if it is off, yes I know, snapshots), and the third would be spare, or maybe containing a 3rd copy of the data that is only update every 6 months or more.

All three servers follow as many of the recommendations as possible - ECC RAM, eSSDs, eHDD or NAS drives, hardware that is on the compatibility list, etc.

Sounds logical, especially with respect to having an offline copy to prevent malware/crypto. I would suggest seeing if you can locate one of them at a trusted friend's house (with a fat network pipe) to satisfy an offsite requirement as well.

- Easy to be sure there will be no block, because you would have to re-format to do that anyways
- So, 2x8x Z2 would be approx 24TB each / 48TB total with four parity drives
- So a mirror would be 8x 2-way mirrors, 32TB usable? I imagine that would be quite fast
- One advantage of this is that it could do block or traditional storage, or maybe both?​
- If they are not mirrored, why do you need two drives for L2ARC?

You can do block off of RAIDZ2, but it won't be fast or efficient. Mirrors will be the fastest solution for random I/O, and as I explained previously, a pool with heavy writes/rewrites over time trends towards random I/O as the limiting factor. If you're confident in your "store and forget" solution then Z2 will do just fine.

As far as dual L2ARC drives, the idea is just that if you're reliant on L2ARC for holding metadata, losing it entirely could cause a bit of performance shock. L2ARC is striped, but losing it isn't catastrophic as it's a read-only copy. Having two drives just spreads the load (and doubles your capacity)

I have constraints to consider unless I make some changes: (still thinking about it)
- Feel free to suggest alternate configurations
- For the first (primary) server, I was planning to use the 7x 10TB SATA drives (low power and heat)
- I could put these in the regular tower case that can hold 11x drives, because they don't put off as much heat​
- I don't see how else you could configure these other than RAID-Z2​
- I was thinking of filling it up with 4TB drives to see how bad the heat problem is or if it is OK​

I like this setup, and yes, RAIDZ2 makes the most sense here. NL-SAS in this chassis will get toasty pretty quickly but don't neglect cooling on your shucked 10T drives.

- For the 2nd server, this would use the Supermicro 16x chassis (seems a shame to not use it as primary, still thinking)
- This is where I would put the 4TB drives because it is designed for that​
- So this would be to backup the first​
- I was going to buy a 2nd one of these before CV19 hit, still considering that​

Can't see this as anything other than 2x 8-drive Z2 really.

- The 3rd server is the other tower case with a capacity of 7x drives
- This is a much lower spec server but still good​
- I do have the JBOD which can handle 10x enterprise or NAS drives, not sure how I am going to use this yet
- Not sure I want an array that consists of internal and external drives​
- Worried what would happen if I accidentally didn't power on the JBOD, or it lost power at some point​
- Thinking whatever I do with it, should be restricted to what can fully fit​
- I can also fit 2x 2.5" drives, the expander can handle way more than that if I need to do something else someday​

You can set failmode=wait on the pool containing internal/external drives, which would pause the system until it's reconnected/repowered. The default in FreeNAS is failmode=continue which wouldn't help since any configuration would just go UNAVAIL immediately with a mix of 7 internal + 10 external.

- I have two nice LSI RAID controllers on the way to handle VMFS duty, which will be all flash

Make sure they pass TRIM/UNMAP properly and use VMFS6 for that tasty automatic space reclamation.

- I do have some 400Gig NVMe, but they are mixed-use, so not sure they can take that many writes
- Yes, the 400Gig and 200Gig disks are write-intensive, so they are a good choice four sure
- The flash drives could mainly be direct-connect vs. on the expander so that they get the best bandwidth
- Right now I am leaning towards not doing iSCSI for data, if I do maybe I take a couple of disks and just do it as a test
- I have 10Gig cards, but I was planning to only use them to connect servers to each other for backups
- It does increase the speed by 2-4x over 1Gig, limited by the disk speed​

The mixed-use Seagate NVMe is rated for 3 DWPD I believe, unless you're going to shove 1.2T per day at them they should be fine. Having a separate system entirely for iSCSI might be worth playing with - if you use all-flash then you don't need to cache quite as aggressively as the penalty for missing ARC is going to NAND vs. spinning disks.

- This is an interesting setup
- In that case, I could just put the 7x 10TB SATA drives in RAID-Z2, and use the remaining 9 slots for mirrors
- This does not include internal non-hot-swap bays, I think I can fit around 8x 2.5" disks this way​
- This would give me 50TB of data and 4x mirrors would give 16TB
- I could use the external jbod to test a mirror setup with iSCSI
- 2nd server with enough data to back this up, buy a new case if it can't handle the heat

You'd want to make sure you've got a lot of cooling if you put it all into the one system, but assuming you mean the 3U SuperMicro that should be already set up with some pretty solid fans that have enough static pressure. Your 8x 2.5" slots could be used for a pair of SLOG devices, and a whole lot of L2ARC depending on where you need it - and then have enough capability to add meta vdevs later once TrueNAS 12 CORE is out/stable enough for beta testing.

Having the external JBOD be purely for iSCSI is a good idea, makes it easy to decommission when you're done. Export the pool and flick the switch.

Ah, I don't know anything about this. I know that new version is coming, but I don't know when and I don't know much about features, I thought they were mostly merging FreeNAS with TrueNAS. But, this again is why I will have three servers, so that I can afford to rebuild from time to time when I need to blow something away to take advantage of something new.

Interesting stuff

-JCL
FreeNAS is becoming "TrueNAS CORE" as of the 12.0 release - the big change has to do with the rebasing of the OpenZFS code on the ZFSonLinux code. That's where we'll get all the fancy new features like the special allocation classes.
 

jcl123

Dabbler
Joined
Jul 6, 2019
Messages
23
Hello, still mulling over the information above, I will follow up later. I wanted to bring two interesting things to your attention:

1. I figured out how to edit the firmware to get the HGST 400Gig SSDs to run at 12Gig instead of 6Gig. It reports that it now supports 12Gig although still only connecting/negotiating at 6Gig. I suspect it's my crappy Amazon cables so I have a genuine LSI cable on the way to see if I can get it to run at 12Gig. I found someone in another forum who did it, I may msg them for help, but I am REALLY close.

2. Are you familiar with the SAS Data Protection information / T10-PI feature? I received my LSI 9361-8i controller, and while setting up the software I noticed a checkbox for it. Upon further reading, the 4TB Seagate drives support this feature. And it appears that all I might need to do is re-format them back to 520-byte sectors (haven't figured out exactly how yet) and then I would be able to use it. It says it has the ability to protect you from silent data corruption on a RAID controller similar to how ZFS is able to. It also mentions that HBAs can support this feature, but I assume at that point you would need an OS/software that supported it as well.
- I doubt FreeNAS supports this feature, it would be cool if it did maybe it could enhance it's capabilities​
- Even if it can't, this makes me feel much better about using the RAID controller for some use cases​

Curious to hear what you and others think.

-JCL
 
Top