I was thinking that splitting up the roles for the bulk data to be on the data vdev while having the special vdev be able to work in tandem at that same time to do the metadata tasks (and some small file offload) might of some help for overall performance?
Like parallel I/O? Maybe in theory that greatly boosts performance, but I would have to defer to someone who has put it into practice and found that it's actually worth it. (As opposed to expanding the capacity of their pool without a narrow scope of metadata/special vdevs; or even without the extra complexity or risk involved of adding more vdevs because you have extra devices at your disposal.)
Increasing RAM (especially fore Core) de facto will boost performance, regardless of
userdata or
metadata. Do they or you notice any slowdowns?
In regards to metadata, see
this thread for a caveat and a solution, plus
check this out going into the future with OpenZFS 2.2.
Would the drives for the metadata setup be better used for a L2ARC or maybe a Dedup vdev (wasn't really planning on needing deduplication).
If you have to ask yourself whether or not using dedup will suit you, then it means you shouldn't use dedup.
Keep in mind that unlike a SLOG or L2ARC, a special vdev is a crucial component of a pool. You can't simply "try it out" and then later decide to casually remove it
Maybe someone can chime in on how realistically beneficial it is to dedicate a set of NVMe drives for a special vdev, even when the pool itself is solely comprised of only NVMe drives.