The Storage section of the graphical interface allows configuration of these options:
- Volumes creates and manages storage volumes.
- Periodic Snapshot Tasks schedules automatic creation of filesystem snapshots.
- Replication Tasks automates the replication of snapshots to a remote system.
- Scrubs schedules scrubs as part of ongoing disk maintenance.
- Snapshots manages local snapshots.
- VMware-Snapshot coordinates ZFS snapshots with a VMware datastore.
The Volumes section of the FreeNAS® graphical interface can be used to format ZFS pools, import a disk to copy its data into an existing pool, or import an existing ZFS pool. It can also be used to create ZFS datasets and zvols and to manage their permissions.
In ZFS terminology, the storage that is managed by ZFS is referred to as a pool. The FreeNAS® graphical interface uses the term volume to refer to a ZFS pool.
Proper storage design is important for any NAS. Please read through this entire chapter before configuring storage disks. All of the features are described to help make it clear which will be the most benefit for your uses, and caveats or caveats or hardware restrictions which could limit their use.
8.1.1. Volume Manager¶
Volume Manager is used to add disks to a ZFS pool. Any old data on added disks is overwritten, so save it elsewhere before reusing a disk. Please see the ZFS Primer for information on ZFS redundancy with multiple disks before using Volume Manager.
Storage → Volumes → Volume Manager opens
a screen like the example shown in
Table 8.1.1 summarizes the configuration options of this screen.
|Volume name||string||ZFS volumes must conform to these
it is recommended to choose a name that will stick out in the logs (e.g. not
|Volume to extend||drop-down menu||used to extend an existing ZFS pool; see Extending a ZFS Volume for instructions|
|Encryption||checkbox||read the section on Encryption before choosing to use encryption|
|Available disks||display||displays the number and size of available disks; hover over show to list the available device names; click the + to add all of the disks to the pool|
|Volume layout||drag and drop||click and drag the icon to select the desired number of disks for a vdev; once at least one disk is selected, the layouts supported by the selected number of disks will be added to the drop-down menu|
|Add Extra Device||button||used to configure multiple vdevs or to add log or cache devices during pool creation|
|Manual setup||button||used to make a non-optimal pool (not recommended); see Manual Setup for details|
Drag the slider to select the desired number of disks. Volume Manager will display the resulting storage capacity, taking swap space into account. To change the layout or the number of disks, use the mouse to drag the slider to the desired volume layout. The Volume layout drop-down menu can also be clicked if a different level of redundancy is required.
For performance and capacity reasons, this screen does not allow creating a volume from disks of differing sizes. While it is not recommended, it is possible to create a volume in this situation by using the Manual setup button and following the instructions in Manual Setup.
Volume Manager only allows choosing a configuration if enough disks have been selected to create that configuration. These layouts are supported:
- Stripe: requires at least one disk
- Mirror: requires at least two disks
- RAIDZ1: requires at least three disks
- RAIDZ2: requires at least four disks
- RAIDZ3: requires at least five disks
- log device: requires at least one dedicated device, SSD is recommended
- cache device: requires at least one dedicated device, SSD is recommended
When more than five disks are being used, consideration must be give to the optimal layout for the best performance and scalability. An overview of the recommended disk group sizes as well as more information about log and cache devices can be found in the ZFS Primer.
The Add Volume button warns that existing data will be cleared. In other words, creating a new volume reformats the selected disks. If the existing data is meant to be preserved, click the Cancel button and refer to Import Disk and Import Volume to see if the existing format is supported. If so, perform that supported action instead. If the current storage format is not supported, it is necessary to back up the data to external media, format the disks, then restore the data to the new volume.
Depending upon the size and number of disks, the type of controller,
and whether encryption is selected, creating the volume may take some
time. After the volume is created, the screen will refresh and the new
volume is listed in the tree under
Storage → Volumes.
Click the + next to the volume name to access its
Change Permissions, Create Dataset, and
Create zvol options.
Beginning with 8.3.1, FreeNAS® supports GELI full disk encryption when creating ZFS volumes. It is important to understand the following when considering whether or not encryption is right for your FreeNAS® system:
- This is not the encryption method used by Oracle’s version of ZFS as that version is not open source and is the property of Oracle.
- This is full disk encryption and not per-filesystem encryption. The underlying drives are first encrypted, then the pool is created on top of the encrypted devices.
- This type of encryption is primarily targeted at users who store sensitive data and want to retain the ability to remove disks from the pool without having to first wipe the disk’s contents.
- This design is only suitable for safe disposal of disks independent of the encryption key. As long as the key and the disks are intact, the system is vulnerable to being decrypted. The key should be protected by a strong passphrase and any backups of the key should be securely stored.
- On the other hand, if the key is lost, the data on the disks is inaccessible. Always back up the key!
- The encryption key is per ZFS volume (pool). If you create multiple pools, each pool has its own encryption key.
- If the system has a lot of disks, there will be a performance hit if the CPU does not support AES-NI or if no crypto hardware is installed. Without hardware acceleration, there will be about a 20% performance hit for a single disk. Performance degradation will continue to increase with more disks. As data is written, it is automatically encrypted and as data is read, it is decrypted on the fly. If the processor does support the AES-NI instruction set, there should be very little, if any, degradation in performance when using encryption. This forum post compares the performance of various CPUs.
- Data in the ARC cache and the contents of RAM are unencrypted.
- Swap is always encrypted, even on unencrypted volumes.
- There is no way to convert an existing, unencrypted volume. Instead, the data must be backed up, the existing pool must be destroyed, a new encrypted volume must be created, and the backup restored to the new volume.
- Hybrid pools are not supported. In other words, newly created vdevs must match the existing encryption scheme. When extending a volume, Volume Manager will automatically encrypt the new vdev being added to the existing encrypted pool.
The encryption facility used by FreeNAS® is designed to protect against physical theft of the disks. It is not designed to protect against unauthorized software access. Ensure that only authorized users have access to the administrative GUI and that proper permissions are set on shares if sensitive data is stored on the system.
To create an encrypted volume, check the Encryption box shown in Figure 8.1.1. A pop-up message reminds you that it is extremely important to make a backup of the key, as without it the data on the disks is inaccessible. Refer to Managing Encrypted Volumes for instructions.
22.214.171.124. Manual Setup¶
The Manual Setup button shown in Figure 8.1.1 can be used to create a non-optimal ZFS volume. While this is not recommended, it can, for example, be used to create a volume containing disks of different sizes.
When using disks of differing sizes, the volume is limited by the size of the smallest disk. For this reason, it is recommended to instead use Volume Manager with same-size disks.
|Volume name||string||ZFS volumes must conform to these
naming conventions ;
it is recommended to choose a name that will stick out in the logs (e.g.
|Encryption||checkbox||read the section on Encryption before choosing to use encryption|
|Member disks||list||highlight desired number of disks from list of available disks|
|Deduplication||drop-down menu||choices are Off, Verify, and On; carefully consider the section on Deduplication before changing this setting|
|ZFS Extra||bullet selection||used to specify if disk is used for storage (None), a log device, a cache device, or a spare|
126.96.36.199. Extending a ZFS Volume¶
The Volume to extend drop-down menu in
Storage → Volumes → Volume Manager,
can be used to add additional disks to an existing ZFS volume. This
drop-down menu will be empty if no ZFS volume exists.
If the existing volume is encrypted, a warning message will remind you that the operation of extending a volume will reset the passphrase and recovery key. After extending the volume, you should immediately recreate both using the instructions in Managing Encrypted Volumes.
After an existing volume has been selected from the drop-down menu, drag and drop the desired disks and select the desired volume layout. For example, disks can be added to increase the capacity of the ZFS pool.
When adding disks to increase the capacity of a volume, ZFS supports the addition of virtual devices, known as vdevs, to an existing ZFS pool. A vdev can be a single disk, a stripe, a mirror, a RAIDZ1, RAIDZ2, or a RAIDZ3. After a vdev is created, more drives cannot be added to that vdev; however, you can stripe a new vdev (and its disks) with another of the same type of existing vdev to increase the overall size of ZFS the pool. In other words, when you extend a ZFS volume, you are really striping similar vdevs. Here are some examples:
- to extend a ZFS stripe, add one or more disks. Since there is no redundancy, you do not have to add the same amount of disks as the existing stripe.
- to extend a ZFS mirror, add the same number of drives. The resulting striped mirror is a RAID 10. For example, if you have 10 drives, you could start by creating a mirror of two drives, extending this mirror by creating another mirror of two drives, and repeating three more times until all 10 drives have been added.
- to extend a three drive RAIDZ1, add three additional drives. The result is a RAIDZ+0, similar to RAID 50 on a hardware controller.
- to extend a RAIDZ2 requires a minimum of four additional drives. The result is a RAIDZ2+0, similar to RAID 60 on a hardware controller.
If you try to add an incorrect number of disks to the existing vdev, an error message will appear, indicating the number of disks that are needed. You will need to select the correct number of disks in order to continue.
188.8.131.52.1. Adding L2ARC or ZIL Devices¶
After the SSDs have been physically installed, click the Volume Manager button and choose the volume from the Volume to extend drop-down menu. Click the + next to the SSD in the Available disks list. In the Volume layout drop-down menu, select Cache (L2ARC) to add a cache device, or Log (ZIL) to add a log device. Finally, click Extend Volume to add the SSD.
8.1.2. Change Permissions¶
Setting permissions is an important aspect of configuring volumes. The graphical administrative interface is meant to set the initial permissions for a volume or dataset in order to make it available as a share. Once a share is available, the client operating system should be used to fine-tune the permissions of the files and directories that are created by the client.
The chapter on Sharing contains configuration examples for several types of permission scenarios. This section provides an overview of the screen that is used to set permissions.
For users and groups to be available, they must either be first created using the instructions in Account or imported from a directory service using the instructions in Directory Services. If more than 50 users or groups are available, the drop-down menus described in this section will automatically truncate their display to 50 for performance reasons. In this case, start to type in the desired user or group name so that the display narrows its search to matching results.
After a volume or dataset is created, it is listed by its mount point
Storage → Volumes → View Volumes.
Clicking the Change Permissions icon for a specific
volume/dataset displays the screen shown in
summarizes the options in this screen.
|Apply Owner (user)||checkbox||uncheck to prevent new permission change from being applied to Owner (user), see Note below|
|Owner (user)||drop-down menu||user to control the volume/dataset; users which were manually created or imported from a directory service will appear in the drop-down menu|
|Apply Owner (group)||checkbox||uncheck to prevent new permission change from being applied to Owner (group), see Note below|
|Owner (group)||drop-down menu||group to control the volume/dataset; groups which were manually created or imported from a directory service will appear in the drop-down menu|
|Apply Mode||checkbox||uncheck to prevent new permission change from being applied to Mode, see Note below|
|Mode||checkboxes||only applies to the Unix or Mac “Permission Type” so will be grayed out if Windows is selected|
|Permission Type||bullet selection||choices are Unix, Mac or Windows; select the type which matches the type of client accessing the volume/dataset|
|Set permission recursively||checkbox||if checked, permissions will also apply to subdirectories of the volume/dataset; if data already exists on the volume/dataset, change the permissions on the client side to prevent a performance lag|
The Apply Owner (user), Apply Owner (group), and Apply Mode checkboxes allow fine-tuning of the change permissions behavior. By default, all boxes are checked and FreeNAS® resets the owner, group, and mode when the Change button is clicked. These checkboxes allow choosing which settings to change. For example, to change just the Owner (group) setting, uncheck the boxes Apply Owner (user) and Apply Mode.
If a mix of operating systems or clients will be accessing the volume/dataset using a non-SMB share, select the Unix Permission Type, as all clients understand them.
The Windows Permission Type augments traditional Unix permissions with ACLs. Use the Windows Permission Type for SMB shares or when the FreeNAS® system is a member of an Active Directory domain.
If you change your mind about the Permission Type, it is not necessary to recreate the volume/dataset, as existing data is not lost. However, changing from Windows to Unix or Mac will remove the extended permissions provided by ACLs from existing files.
When the Windows Permission Type is set, the ACLs are set to what Windows sets on new files and directories by default. The Windows client should then be used to fine-tune the permissions as required.
8.1.3. Create Dataset¶
An existing ZFS volume can be divided into datasets. Permissions, compression, deduplication, and quotas can be set on a per-dataset basis, allowing more granular control over access to storage data. A dataset is similar to a folder in that you can set permissions; it is also similar to a filesystem in that you can set properties such as quotas and compression as well as create snapshots.
ZFS provides thick provisioning using quotas and thin provisioning using reserved space.
Selecting an existing ZFS volume in the tree and clicking Create Dataset shows the screen in Figure 8.1.4.
summarizes the options available when creating a ZFS
dataset. Some settings are only available in
Advanced Mode. To see these settings, either click the
Advanced Mode button, or configure the system to always
display these settings by checking the box
Show advanced fields by default in
System → Advanced.
Most attributes, except for the Dataset Name,
Case Sensitivity, and Record Size, can be
changed after dataset creation by highlighting the dataset name and
clicking its Edit Options button in
Storage → Volumes → View Volumes.
|Dataset Name||string||mandatory; enter a unique name for the dataset|
|Comments||string||short comments or user notes about this dataset|
|Compression Level||drop-down menu||see the section on Compression for a description of the available algorithms|
|Share type||drop-down menu||select the type of share that will be used on the dataset; choices are UNIX for an NFS share, Windows for a SMB share, or Mac for an AFP share|
|Case Sensitivity||drop-down menu||choices are sensitive (default, assumes filenames are case sensitive), insensitive (assumes filenames are not case sensitive), or mixed (understands both types of filenames)|
|Enable atime||Inherit, On, or Off||controls whether the access time for files is updated when they are read; setting this property to Off avoids producing log traffic when reading files and can result in significant performance gains|
|Quota for this dataset||integer||only available in Advanced Mode; default of 0 disables quotas; specifying a value means to use no more than the specified size and is suitable for user datasets to prevent users from hogging available space|
|Quota for this dataset and all children||integer||only available in Advanced Mode; a specified value applies to both this dataset and any child datasets|
|Reserved space for this dataset||integer||only available in Advanced Mode; default of 0 is unlimited; specifying a value means to keep at least this much space free and is suitable for datasets containing logs which could take up all available free space|
|Reserved space for this dataset and all children||integer||only available in Advanced Mode; a specified value applies to both this dataset and any child datasets|
|ZFS Deduplication||drop-down menu||read the section on Deduplication before making a change to this setting|
|Record Size||drop-down menu||only available in Advanced Mode; while ZFS automatically adapts the record size dynamically to adapt to data, if the data has a fixed size (e.g. a database), matching that size may result in better performance|
After a dataset is created, you can click on that dataset and select Create Dataset, thus creating a nested dataset, or a dataset within a dataset. A zvol can also be created within a dataset. When creating datasets, double-check that you are using the Create Dataset option for the intended volume or dataset. If you get confused when creating a dataset on a volume, click all existing datasets to close them–the remaining Create Dataset will be for the volume.
Deduplication is the process of ZFS transparently reusing a single copy of duplicated data to save space. Depending on the amount of duplicate data, deduplicaton can improve storage capacity, as less data is written and stored. However, deduplication is RAM intensive. A general rule of thumb is 5 GB of RAM per terabyte of deduplicated storage. In most cases, compression provides storage gains comparable to deduplication with less impact on performance.
In FreeNAS®, deduplication can be enabled during dataset creation. Be forewarned that there is no way to undedup the data within a dataset once deduplication is enabled, as disabling deduplication has NO EFFECT on existing data. The more data written to a deduplicated dataset, the more RAM it requires. When the system starts storing the DDTs (dedup tables) on disk because they no longer fit into RAM, performance craters. Further, importing an unclean pool can require between 3-5 GB of RAM per terabyte of deduped data, and if the system does not have the needed RAM, it will panic. The only solution is to add more RAM or recreate the pool. Think carefully before enabling dedup! This article provides a good description of the value versus cost considerations for deduplication.
Unless a lot of RAM and a lot of duplicate data is available, do not change the default deduplication setting of “Off”. For performance reasons, consider using compression rather than turning this option on.
If deduplication is changed to On, duplicate data blocks are removed synchronously. The result is that only unique data is stored and common components are shared among files. If deduplication is changed to Verify, ZFS will do a byte-to-byte comparison when two blocks have the same signature to make sure that the block contents are identical. Since hash collisions are extremely rare, Verify is usually not worth the performance hit.
After deduplication is enabled, the only way to disable it
is to use the
zfs set dedup=off dataset_name command
from Shell. However, any data that has already been
deduplicated will not be un-deduplicated. Only newly stored data
after the property change will not be deduplicated. The only way to
remove existing deduplicated data is to copy all of the data off of
the dataset, set the property to off, then copy the data back in
again. Alternately, create a new dataset with
ZFS Deduplication left disabled, copy the data to the
new dataset, and destroy the original dataset.
Deduplication is often considered when using a group of very similar virtual machine images. However, other features of ZFS can provide dedup-like functionality more efficiently. For example, create a dataset for a standard VM, then clone that dataset for other VMs. Only the difference between each created VM and the main dataset are saved, giving the effect of deduplication without the overhead.
When selecting a compression type, you need to balance performance with the amount of disk space saved by compression. Compression is transparent to the client and applications as ZFS automatically compresses data as it is written to a compressed dataset or zvol and automatically decompresses that data as it is read. These compression algorithms are supported:
- lz4: recommended compression method as it allows compressed datasets to operate at near real-time speed. This algorithm only compresses the files that will benefit from compression. By default, ZFS pools made using FreeNAS® 9.2.1 or higher use this compression method, meaning that this algorithm is used if the Compression level is left at Inherit when creating a dataset or zvol.
- gzip: varies from levels 1 to 9 where gzip fastest (level 1) gives the least compression and gzip maximum (level 9) provides the best compression but is discouraged due to its performance impact.
- zle: fast but simple algorithm to eliminate runs of zeroes.
- lzjb: provides decent data compression, but is considered deprecated as lz4 provides much better performance.
If you select Off as the Compression level when creating a dataset or zvol, compression will not be used on the dataset/zvol. This is not recommended as using lz4 has a negligible performance impact and allows for more storage capacity.
8.1.4. Create zvol¶
A zvol is a feature of ZFS that creates a raw block device over ZFS. This allows you to use a zvol as an iSCSI device extent.
To create a zvol, select an existing ZFS volume or dataset from the tree then click Create zvol to open the screen shown in Figure 8.1.5.
The configuration options are described in
Some settings are only available in Advanced Mode. To see
these settings, either click the Advanced Mode button or
configure the system to always display these settings by checking
Show advanced fields by default in
System → Advanced.
|zvol Name||string||mandatory; input a name for the zvol|
|Comments||string||short comments or user notes about this zvol|
|Size for this zvol||integer||specify size and value such as 10Gib; if the size is more than 80% of the available capacity, the creation will fail with an “out of space” error unless Force size is checked|
|Force size||checkbox||by default, the system will not let you create a zvol if that operation will bring the pool to over 80% capacity; while NOT recommended, checking this box will force the creation of the zvol in this situation|
|Compression level||drop-down menu||see the section on Compression for a description of the available algorithms|
|Sparse volume||checkbox||used to provide thin provisioning; use with caution for when this option is selected, writes will fail when the pool is low on space|
|Block size||drop-down menu||only available in Advanced Mode and by default is based on the number of disks in pool; can be set to match the block size of the filesystem which will be formatted onto the iSCSI target|
8.1.5. Import Disk¶
Volume → Import Disk
screen, shown in
is used to import a single disk that has been formatted with the
UFS, NTFS, MSDOS, or EXT2 filesystem. The import is meant to be a
temporary measure to copy the data from a disk to an existing ZFS
dataset. Only one disk can be imported at a time.
Imports of EXT3 or EXT4 filesystems are possible in some cases, although neither is fully supported. EXT3 journaling is not supported, so those filesystems must have an external fsck utility, like the one provided by E2fsprogs utilities, run on them before import. EXT4 filesystems with extended attributes or inodes greater than 128 bytes are not supported. EXT4 filesystems with EXT3 journaling must have an fsck run on them before import, as described above.
Use the drop-down menu to select the disk to import, select the type of filesystem on the disk, and browse to the ZFS dataset that will hold the copied data. When you click Import Volume, the disk is mounted, its contents are copied to the specified ZFS dataset, and the disk is unmounted after the copy operation completes.
8.1.6. Import Volume¶
If you click
Storage → Volumes → Import Volume,
you can configure FreeNAS® to use an existing ZFS pool. This
action is typically performed when an existing FreeNAS® system is
re-installed. Since the operating system is separate from the storage
disks, a new installation does not affect the data on the disks.
However, the new operating system needs to be configured to use the
Figure 8.1.7 shows the initial pop-up window that appears when you import a volume.
If you are importing an unencrypted ZFS pool, select No: Skip to import to open the screen shown in Figure 8.1.8.
Existing volumes should be available for selection from the drop-down menu. In the example shown in Figure 8.1.8, the FreeNAS® system has an existing, unencrypted ZFS pool. Once the volume is selected, click the OK button to import the volume.
If an existing ZFS pool does not show in the drop-down menu, run zpool import from Shell to import the pool.
If you plan to physically install ZFS formatted disks from another system, be sure to export the drives on that system to prevent an “in use by another machine” error during the import.
If you suspect that your hardware is not being detected, run camcontrol devlist from Shell. If the disk does not appear in the output, check to see if the controller driver is supported or if it needs to be loaded using Tunables.
184.108.40.206. Importing an Encrypted Pool¶
Select the disks in the encrypted pool, browse to the location of the saved encryption key, input the passphrase associated with the key, then click OK to decrypt the disks.
The encryption key is required to decrypt the pool. If the pool cannot be decrypted, it cannot be re-imported after a failed upgrade or lost configuration. This means that it is very important to save a copy of the key and to remember the passphrase that was configured for the key. Refer to Managing Encrypted Volumes for instructions on how to manage the keys for encrypted volumes.
Once the pool is decrypted, it will appear in the drop-down menu of Figure 8.1.8. Click the OK button to finish the volume import.
8.1.7. View Disks¶
Storage → Volumes → View Disks
shows all of the disks recognized by the FreeNAS® system. An example is
The current configuration of each device is displayed. Click a disk entry and the Edit button to change its configuration. The configurable options are described in Table 8.1.6.
|Name||string||read-only value showing FreeBSD device name for disk|
|Serial||string||read-only value showing the disk’s serial number|
|HDD Standby||drop-down menu||indicates the time of inactivity (in minutes) before the drive enters standby mode in order to conserve energy; this forum post demonstrates how to determine if a drive has spun down|
|Advanced Power Management||drop-down menu||default is Disabled, can select a power management profile from the menu|
|Acoustic Level||drop-down menu||default is Disabled; can be modified for disks that understand AAM|
|Enable S.M.A.R.T.||checkbox||enabled by default if the disk supports S.M.A.R.T.; unchecking this box will disable any configured S.M.A.R.T. Tests for the disk|
|S.M.A.R.T. extra options||string||additional smartctl(8) options|
If a disk’s serial number is not displayed in this screen, use the smartctl command from Shell. For example, to determine the serial number of disk ada0, type smartctl -a /dev/ada0 | grep Serial.
The Wipe function is provided for when an unused disk is to be discarded.
Make certain that all data has been backed up and that the disk is no longer in use. Triple-check that the correct disk is being selected to be wiped, as recovering data from a wiped disk is usually impossible. If there is any doubt, physically remove the disk, verify that all data is still present on the FreeNAS® system, and wipe the disk in a separate computer.
Clicking Wipe offers several choices. Quick erases only the partitioning information on a disk, making it easy to reuse but without clearing other old data. For more security, Full with zeros overwrites the entire disk with zeros, while Full with random data overwrites the entire disk with random binary data.
Quick wipes take only a few seconds. A Full with zeros wipe of a large disk can take several hours, and a Full with random data takes longer. A progress bar is displayed during the wipe to track status.
8.1.8. View Volumes¶
Storage → Volumes → View Volumes
is used to view and further configure existing ZFS pools, datasets,
and zvols. The example shown in
shows one ZFS pool (volume1) with two datasets (the one
automatically created with the pool, volume1, and dataset1) and
one zvol (zvol1).
Note that in this example, there are two datasets named volume1. The first represents the ZFS pool and its Used and Available entries reflect the total size of the pool, including disk parity. The second represents the implicit or root dataset and its Used and Available entries indicate the amount of disk space available for storage.
Buttons are provided for quick access to Volume Manager, Import Disk, Import Volume, and View Disks. If the system has multipath-capable hardware, an extra button will be added, View Multipaths. For each entry, the columns indicate the Name, how much disk space is Used, how much disk space is Available, the type of Compression, the Compression Ratio, the Status, whether it is mounted as read-only, and any Comments entered for the volume.
Clicking the entry for a pool causes several buttons to appear at the bottom of the screen. The buttons perform these actions:
Detach Volume: allows you to either export the pool or to delete the contents of the pool, depending upon the choice you make in the screen shown in Figure 8.1.12. The Detach Volume screen displays the current used space and indicates if there are any shares, provides checkboxes to Mark the disks as new (destroy data) and to Also delete the share’s configuration, asks if you are sure that you want to do this, and the browser will turn red to alert you that you are about to do something that will make the data inaccessible. If you do not check the box to mark the disks as new, the volume will be exported. This means that the data is not destroyed and the volume can be re-imported at a later time. If you will be moving a ZFS pool from one system to another, perform this export action first as it flushes any unwritten data to disk, writes data to the disk indicating that the export was done, and removes all knowledge of the pool from the system. If you do check the box to mark the disks as new, the pool and all the data in its datasets, zvols, and shares will be destroyed and the underlying disks will be returned to their raw state.
Scrub Volume: scrubs and scheduling them are described in more detail in Scrubs. This button allows manually initiating a scrub. Scrubs are I/O intensive and can negatively impact performance. Avoid initiating a scrub when the system is busy.
A Cancel button is provided to cancel a scrub. When a scrub is cancelled, it is abandoned. The next scrub to run starts from the beginning, not where the cancelled scrub left off.
The status of a running scrub or the statistics from the last completed scrub can be seen by clicking the Volume Status button.
Volume Status: as shown in the example in Figure 8.1.13, this screen shows the device name and status of each disk in the ZFS pool as well as any read, write, or checksum errors. It also indicates the status of the latest ZFS scrub. Clicking the entry for a device causes buttons to appear to edit the device’s options (shown in Figure 8.1.14), offline or online the device, or replace the device (as described in Replacing a Failed Drive).
Upgrade: used to upgrade the pool to the latest ZFS features, as described in Upgrading a ZFS Pool. This button does not appear if the pool is running the latest version of feature flags.
Versions of FreeNAS® prior to 8.3.1 required a reboot to apply changes to the HDD Standby, Advanced Power Management, and Acoustic Level settings. As of 8.3.1, changes to these settings are applied immediately.
Clicking a dataset in
Storage → Volumes → View Volumes
causes buttons to appear at the bottom of the screen, providing these
Change Permissions: edit the dataset’s permissions as described in Change Permissions.
Create Snapshot: create a one-time snapshot. To schedule the regular creation of snapshots, instead use Periodic Snapshot Tasks.
Destroy Dataset: clicking the Destroy Dataset button causes the browser window to turn red to indicate that this is a destructive action. The Destroy Dataset screen forces you to check the box I’m aware this will destroy all child datasets and snapshots within this dataset before it will perform this action.
Edit Options: edit the volume’s properties described in Table 8.1.4. Note that it will not allow changing the dataset’s name.
Create Dataset: used to create a child dataset within this dataset.
Create zvol: create a child zvol within this dataset.
Clicking a zvol in
Storage → Volumes → View Volumes causes
icons to appear at the bottom of the screen:
Create Snapshot, Edit zvol, and
Destroy zvol. Similar to datasets, a zvol’s name cannot be
changed, and destroying a zvol requires confirmation.
220.127.116.11. Managing Encrypted Volumes¶
If the Encryption box is checked during the creation of a
pool, additional buttons appear in the entry for the pool in
Storage → Volumes → View Volumes.
An example is shown in
These additional encryption buttons are used to:
Create/Change Passphrase: click this button to set and confirm the passphrase associated with the GELI encryption key. You will be prompted to enter and repeat the desired passphrase and a red warning reminds you to Remember to add a new recovery key as this action invalidates the previous recovery key. Unlike a password, a passphrase can contain spaces and is typically a series of words. A good passphrase is easy to remember (like the line to a song or piece of literature) but hard to guess (people who know you should not be able to guess the passphrase). Remember this passphrase as you cannot re-import an encrypted volume without it. In other words, if you forget the passphrase, the data on the volume can become inaccessible if you need to re-import the pool. Protect this passphrase as anyone who knows it could re-import your encrypted volume, thwarting the reason for encrypting the disks in the first place.
Once the passphrase is set, the name of this button will change to Change Passphrase. After setting or changing the passphrase, it is important to immediately create a new recovery key by clicking the Add recovery key button. This way, if the passphrase is forgotten, the associated recovery key can be used instead.
Download Key: click this icon to download a backup copy of the GELI encryption key. The encryption key is saved to the client system, not on the FreeNAS® system. You will be prompted to input the password used to access the FreeNAS® administrative GUI before the selecting the directory in which to store the key. Since the GELI encryption key is separate from the FreeNAS® configuration database, it is highly recommended to make a backup of the key. If the key is every lost or destroyed and there is no backup key, the data on the disks is inaccessible.
Encryption Re-key: generates a new GELI encryption key. Typically this is only performed when the administrator suspects that the current key may be compromised. This action also removes the current passphrase.
Add recovery key: generates a new recovery key. This screen prompts for entry of the password used to access the FreeNAS® administrative GUI and then to select the directory in which to save the key. Note that the recovery key is saved to the client system, not on the FreeNAS® system. This recovery key can be used if the passphrase is forgotten. Always immediately add a recovery key whenever the passphrase is changed.
Remove recover key: Typically this is only performed when the administrator suspects that the current recovery key may be compromised. Immediately create a new passphrase and recovery key.
The passphrase, recovery key, and encryption key must be protected. Do not reveal the passphrase to others. On the system containing the downloaded keys, take care that the system and its backups are protected. Anyone who has the keys has the ability to re-import the disks if they are discarded or stolen.
If a re-key fails on a multi-disk system, an alert is generated. Do not ignore this alert as doing so may result in the loss of data.
8.1.9. View Multipaths¶
FreeNAS® uses gmultipath(8) to provide multipath I/O support on systems containing hardware that is capable of multipath. An example would be a dual SAS expander backplane in the chassis or an external JBOD.
Multipath hardware adds fault tolerance to a NAS as the data is still available even if one disk I/O path has a failure.
FreeNAS® automatically detects active/active and active/passive
multipath-capable hardware. Any multipath-capable devices that are
detected will be placed in multipath units with the parent devices
hidden. The configuration will be displayed in
Storage → Volumes → View Multipaths.
Note that this option is not be displayed in the
Storage → Volumes
tree on systems that do not contain multipath-capable hardware.
8.1.10. Replacing a Failed Drive¶
With any form of redundant RAID, failed drives must be replaced as soon as possible to repair the degraded state of the RAID. Depending on the hardware’s capabilities, it might be necessary to reboot to replace the failed drive. Hardware that supports AHCI does not require a reboot.
Striping (RAID0) does not provide redundancy. If a disk in a stripe fails, the volume will be destroyed and must be recreated and the data restored from backup.
If your pool is encrypted with GELI, refer to Replacing an Encrypted Drive before proceeding.
Before physically removing the failed device, go to
Storage → Volumes → View Volumes.
Select the volume’s name. At the bottom of the interface are
several icons, one of which is Volume Status. Click the
Volume Status icon and locate the failed disk. Then
perform these steps:
If the disk is formatted with ZFS, click the disk’s entry then its Offline button in order to change that disk’s status to OFFLINE. This step is needed to properly remove the device from the ZFS pool and to prevent swap issues. If the hardware supports hot-pluggable disks, click the disk’s Offline button, pull the disk, then skip to step 3. If there is no Offline button but only a Replace button, the disk is already offlined and you can safely skip this step.
If the process of changing the disk’s status to OFFLINE fails with a “disk offline failed - no valid replicas” message, the ZFS volume must be scrubbed first with the Scrub Volume button in
Storage → Volumes → View Volumes. After the scrub completes, try to Offline the disk again before proceeding.
If the hardware is not AHCI capable, shut down the system to physically replace the disk. When finished, return to the GUI and locate the OFFLINE disk.
After the disk has been replaced and is showing as OFFLINE, click the disk again and then click its Replace button. Select the replacement disk from the drop-down menu and click the Replace Disk button. After clicking the Replace Disk button, the ZFS pool starts to resilver and the status of the resilver is displayed.
After the drive replacement process is complete, re-add the replaced disk in the S.M.A.R.T. Tests screen.
In the example shown in
a failed disk is being replaced by disk ada5 in the volume named
After the resilver is complete, Volume Status shows a Completed resilver status and indicates any errors. Figure 8.1.17 indicates that the disk replacement was successful in this example.
A disk that is failing but has not completely failed can be replaced in place, without first removing it. Whether this is a good idea depends on the overall condition of the failing disk. A disk with a few newly-bad blocks that is otherwise functional can be left in place during the replacement to provide data redundancy. A drive that is experiencing continuous errors can actually slow down the replacement. In extreme cases, a disk with serious problems might spend so much time retrying failures that it could prevent the replacement resilvering from completing before another drive fails.
18.104.22.168. Replacing an Encrypted Drive¶
If the ZFS pool is encrypted, additional steps are needed when replacing a failed drive.
First, make sure that a passphrase has been set using the instructions in Encryption before attempting to replace the failed drive. Then, follow the steps 1 and 2 as described above. During step 3, you will be prompted to input and confirm the passphrase for the pool. Enter this information then click the Replace Disk button. Wait until the resilvering is complete.
Next, restore the encryption keys to the pool. If the following additional steps are not performed before the next reboot, access to the pool might be permanently lost.
- Highlight the pool that contains the disk that was just replaced and click the Encryption Re-key button in the GUI. Entry of the root password will be required.
- Highlight the pool that contains the disk you just replaced and click Create Passphrase and enter the new passphrase. The old passphrase can be reused if desired.
- Highlight the pool that contains the disk you just replaced and click the Download Key button to save the new encryption key. Since the old key will no longer function, any old keys can be safely discarded.
- Highlight the pool that contains the disk that was just replaced and click the Add Recovery Key button to save the new recovery key. The old recovery key will no longer function, so it can be safely discarded.
22.214.171.124. Removing a Log or Cache Device¶
Added log or cache devices appear in
Storage → Volumes → View Volumes
→ Volume Status.
Clicking the device enables its Replace and
Log and cache devices can be safely removed or replaced with these buttons. Both types of devices improve performance, and throughput can be impacted by their removal.
8.1.11. Replacing Drives to Grow a ZFS Pool¶
The recommended method for expanding the size of a ZFS pool is to pre-plan the number of disks in a vdev and to stripe additional vdevs using Volume Manager as additional capacity is needed.
However, this is not an option if there are no open drive ports and a SAS/SATA HBA card cannot be added. In this case, one disk at a time can be replaced with a larger disk, waiting for the resilvering process to incorporate the new disk into the pool, then repeating with another disk until all of the original disks have been replaced.
The safest way to perform this is to use a spare drive port or an eSATA port and a hard drive dock. The process follows these steps:
- Shut down the system.
- Install one new disk.
- Start up the system.
- Go to
Storage → Volumes, select the pool to expand and click the Volume Status button. Select a disk and click the Replace button. Choose the new disk as the replacement.
- The status of the resilver process can be viewed by running zpool status. When the new disk has resilvered, the old one will be automatically offlined. The system is then shut down to physically remove the replaced disk. One advantage of this approach is that there is no loss of redundancy during the resilver.
If a spare drive port is not available, a drive can be replaced with a larger one using the instructions in Replacing a Failed Drive. This process is slow and places the system in a degraded state. Since a failure at this point could be disastrous, do not attempt this method unless the system has a reliable backup. Replace one drive at a time and wait for the resilver process to complete on the replaced drive before replacing the next drive. After all the drives are replaced and the final resilver completes, the added space will appear in the pool.
8.2. Periodic Snapshot Tasks¶
A periodic snapshot task allows scheduling the creation of read-only versions of ZFS volumes and datasets at a given point in time. Snapshots can be created quickly and, if little data changes, new snapshots take up very little space. For example, a snapshot where no files have changed takes 0 MB of storage, but as changes are made to files, the snapshot size changes to reflect the size of the changes.
Snapshots provide a clever way of keeping a history of files, providing a way to recover an older copy or even a deleted file. For this reason, many administrators take snapshots often (perhaps every fifteen minutes), store them for a period of time (possibly a month), and store them on another system (typically using Replication Tasks). Such a strategy allows the administrator to roll the system back to a specific point in time. If there is a catastrophic loss, an off-site snapshot can be used to restore the system up to the time of the last snapshot.
An existing ZFS volume is required before creating a snapshot. Creating a volume is described in Volume Manager.
If only a one-time snapshot is needed, instead use
Storage → Volumes → View Volumes
and click the Create Snapshot button for the volume or
dataset to snapshot.
|Volume/Dataset||drop-down menu||select an existing ZFS volume, dataset, or zvol|
|Recursive||checkbox||select this box to take separate snapshots of the volume/dataset and each of its child datasets; if unchecked, a single snapshot is taken of only the specified volume/dataset, but not any child datasets|
|Snapshot Lifetime||integer and drop-down menu||length of time to retain the snapshot on this system; if the snapshot is replicated, it is not removed from the receiving system when the lifetime expires|
|Begin||drop-down menu||do not create snapshots before this time of day|
|End||drop-down menu||do not create snapshots after this time of day|
|Interval||drop-down menu||how often to take snapshot between Begin and End times|
|Weekday||checkboxes||which days of the week to take snapshots|
|Enabled||checkbox||uncheck to disable the scheduled snapshot task without deleting it|
If the Recursive box is checked, child datasets of this dataset are included in the snapshot and there is no need to create snapshots for each child dataset. The downside is that there is no way to exclude particular child datasets from a recursive snapshot.
When the OK button is clicked, a snapshot is taken and the task will be repeated according to your settings.
After creating a periodic snapshot task, an entry for the snapshot task will be added to View Periodic Snapshot Tasks. Click an entry to access its Edit and Delete buttons.
8.3. Replication Tasks¶
Replication is the duplication of snapshots from one FreeNAS® system to another computer. When a new snapshot is created on the source computer, it is automatically replicated to the destination computer. Replication is typically used to keep a copy of files on a separate system, with that system sometimes being at a different physical location.
The basic configuration requires a source system with the original data and a destination system where the data will be replicated. The destination system is prepared to receive replicated data, a periodic snapshot of the data on the source system is created, and then a replication task is created. As snapshots are automatically created on the source computer, they are automatically replicated to the destination computer.
Replicated data is not visible on the receiving system until the replication task completes.
8.3.1. Examples: Common Configuration¶
The examples shown here use the same setup of source and destination computers.
126.96.36.199. Alpha (Source)¶
Alpha is the source computer with the data to be replicated. It is at IP address 10.0.0.102. A volume named alphavol has already been created, and a dataset named alphadata has been created on that volume. This dataset contains the files which will be snapshotted and replicated onto Beta.
This new dataset has been created for this example, but a new dataset is not required. Most users will already have datasets containing the data they wish to replicate.
Create a periodic snapshot of the source dataset by selecting
Storage → Volumes.
Click the alphavol/alphadata dataset to highlight it. Create a
periodic snapshot of it by clicking
Periodic Snapshot Tasks, then
Add Periodic Snapshot as shown in
This example creates a snapshot of the alphavol/alphadata dataset every two hours from Monday through Friday between the hours of 9:00 and 18:00 (6:00 PM). Snapshots are automatically deleted after their chosen lifetime of two weeks expires.
188.8.131.52. Beta (Destination)¶
Beta is the destination computer where the replicated data will be copied. It is at IP address 10.0.0.118. A volume named betavol has already been created.
Snapshots are transferred with SSH. To allow incoming connections, this service is enabled on Beta. The service is not required for outgoing connections, and so does not need to be enabled on Alpha.
8.3.2. Example: FreeNAS® to FreeNAS® Semi-Automatic Setup¶
FreeNAS® offers a special semi-automatic setup mode that simplifies setting up replication. Create the replication task on Alpha by clicking Replication Tasks and Add Replication. alphavol/alphadata is selected as the dataset to replicate. betavol is the destination volume where alphadata snapshots are replicated. The Setup mode dropdown is set to Semi-automatic as shown in Figure 8.3.2. The IP address of Beta is entered in the Remote hostname field. A hostname can be entered here if local DNS resolves for that hostname.
If WebGUI HTTP –> HTTPS Redirect has been
System → General
on the destination computer,
Remote HTTP/HTTPS Port must be set to the HTTPS port
(usually 443) and Remote HTTPS must be enabled when
creating the replication on the source computer.
The Remote Auth Token field expects a special token from
the Beta computer. On Beta, choose
Storage → Replication Tasks,
then click Temporary Auth Token. A dialog showing the
temporary authorization token is shown as in
Highlight the temporary authorization token string with the mouse and copy it.
On the Alpha system, paste the copied temporary authorization token string into the Remote Auth Token field as shown in Figure 8.3.4.
Finally, click the OK button to create the replication task. After each periodic snapshot is created, a replication task will copy it to the destination system. See Limiting Replication Times for information about restricting when replication is allowed to run.
The temporary authorization token is only valid for a few minutes. If a Token is invalid message is shown, get a new temporary authorization token from the destination system, clear the Remote Auth Token field, and paste in the new one.
8.3.3. Example: FreeNAS® to FreeNAS® or Other Systems, Manual Setup¶
This example uses the same basic configuration of source and destination computers shown above, but the destination computer is not required to be a FreeNAS® system. Other operating systems can receive the replication if they support SSH, ZFS, and the same features that are in use on the source system. The details of creating volumes and datasets, enabling SSH, and copying encryption keys will vary when the destination computer is not a FreeNAS® system.
184.108.40.206. Encryption Keys¶
A public encryption key must be copied from Alpha to Beta to
allow a secure connection without a password prompt. On Alpha,
Storage → Replication Tasks → View Public Key,
producing the window shown in
Use the mouse to highlight the key data shown in the window, then copy
On Beta, select
Account → Users → View Users. Click the root
account to select it, then click Modify User. Paste the
copied key into the SSH Public Key field and click
OK as shown in
Back on Alpha, create the replication task by clicking Replication Tasks and Add Replication. alphavol/alphadata is selected as the dataset to replicate. The destination volume is betavol. The alphadata dataset and snapshots are replicated there. The IP address of Beta is entered in the Remote hostname field as shown in Figure 8.3.7. A hostname can be entered here if local DNS resolves for that hostname.
Click the SSH Key Scan button to retrieve the SSH host keys from Beta and fill the Remote hostkey field. Finally, click OK to create the replication task. After each periodic snapshot is created, a replication task will copy it to the destination system. See Limiting Replication Times for information about restricting when replication is allowed to run.
8.3.4. Replication Options¶
Table 8.3.1 describes the options in the replication task dialog.
|Volume/Dataset||drop-down menu||ZFS volume or dataset on the source computer containing the snapshots to be replicated; the drop-down menu is empty if a snapshot does not already exist|
|Remote ZFS Volume/Dataset||string||ZFS volume on the remote or destination computer which will store the snapshots; if the destination dataset
is not present, it will be created;
|Recursively replicate child dataset’s snapshots||checkbox||when checked, also replicate snapshots of datasets that are children of the main dataset|
|Delete stale snapshots||checkbox||when checked, delete previous snapshots on the remote or destination computer which are no longer present on the source computer|
|Replication Stream Compression||drop-down menu||choices are lz4 (fastest), pigz (all rounder), plzip (best compression), or Off (no compression); selecting a compression algorithm can reduce the size of the data being replicated|
|Limit (kB/s)||integer||limit replication speed to the specified value in kilobytes/second; default of 0 is unlimited|
|Begin||drop-down menu||replication is not allowed to start before this time; times entered in the Begin and End fields set when replication can occur|
|End||drop-down menu||replication must start by this time; once started, replication will continue until it is finished|
|Enabled||checkbox||uncheck to disable the scheduled replication task without deleting it|
|Setup mode||drop-down menu||Manual or Semi-automatic|
|Remote hostname||string||IP address or DNS name of remote computer where replication is sent|
|Remote port||string||the port used by the SSH server on the remote or destination computer|
|Dedicated User Enabled||checkbox||allow a user account other than root to be used for replication|
|Dedicated User||drop-down menu||only available if Dedicated User Enabled is checked; select the user account to be used for replication|
|Encryption Cipher||drop-down menu||Standard or Fast|
|Remote hostkey||string||use the SSH Key Scan button to retrieve the public host key of the remote or destination computer and populate this field with that key|
The replication task runs after a new periodic snapshot is created. The periodic snapshot and any new manual snapshots of the same dataset are replicated onto the destination computer.
When multiple replications have been created, replication tasks run serially, one after another. Completion time depends on the number and size of snapshots and the bandwidth available between the source and destination computers.
The first time a replication runs, it must duplicate data structures from the source to the destination computer. This can take much longer to complete than subsequent replications, which only send differences in data.
Storage → Replication Tasks displays
Figure 8.3.8, the list of
replication tasks. The Last snapshot sent to remote side
column shows the name of the last snapshot that was successfully
replicated, and Status shows the current status of each
replication task. The display is updated every five seconds, always
showing the latest status.
The encryption key that was copied from the source computer
(Alpha) to the destination computer (Beta) is an RSA public
key located in the
/data/ssh/replication.pub file on the
source computer. The host public key used to identify the
destination computer (Beta) is from the
/etc/ssh/ssh_host_rsa_key.pub file on the destination
8.3.5. Replication Encryption¶
The default Encryption Cipher Standard setting provides good security. Fast is less secure than Standard but can give reasonable transfer rates for devices with limited cryptographic speed. For networks where the entire path between source and destination computers is trusted, the Disabled option can be chosen to send replicated data without encryption.
8.3.6. Limiting Replication Times¶
The Begin and End times in a replication task make it possible to restrict when replication is allowed. These times can be set to only allow replication after business hours, or at other times when disk or network activity will not slow down other operations like snapshots or Scrubs. The default settings allow replication to occur at any time.
These times control when replication task are allowed to start, but will not stop a replication task that is already running. Once a replication task has begun, it will run until finished.
8.3.7. Troubleshooting Replication¶
Replication depends on SSH, disks, network, compression, and encryption to work. A failure or misconfiguration of any of these can prevent successful replication.
SSH must be able to connect from the source system to the destination system with an encryption key. This can be tested from Shell by making an SSH connection from the source system to the destination system. From the previous example, this is a connection from Alpha to Beta at 10.0.0.118. Start the Shell on the source machine (Alpha), then enter this command:
ssh -vv -i /data/ssh/replication 10.0.0.118
On the first connection, the system might say
No matching host key fingerprint found in DNS. Are you sure you want to continue connecting (yes/no)?
Verify that this is the correct destination computer from the
preceeding information on the screen and type
yes. At this
point, an SSH shell connection is open to the destination
If a password is requested, SSH authentication is not working. See
Figure 8.3.5 above. This key
value must be present in the
on Beta, the destination computer. The
file can show diagnostic errors for login problems on the destination
Matching compression and decompression programs must be available on
both the source and destination computers. This is not a problem when
both computers are running FreeNAS®, but other operating systems might
not have lz4, pigz, or plzip compression programs installed by
default. An easy way to diagnose the problem is to set
Replication Stream Compression to Off. If the
replication runs, select the preferred compression method and check
/var/log/debug.log on the FreeNAS® system for errors.
220.127.116.11. Manual Testing¶
On Alpha, the source computer, the
can also show helpful messages to locate the problem.
On the source computer, Alpha, open a Shell and manually send
a single snapshot to the destination computer, Beta. The snapshot
used in this example is named
before, it is located in the alphavol/alphadata dataset. A
@ symbol separates the name of the dataset from the name of
the snapshot in the command.
zfs send firstname.lastname@example.org | ssh -i /data/ssh/replication 10.0.0.118 zfs recv betavol
If a snapshot of that name already exists on the destination computer, the system will refuse to overwrite it with the new snapshot. The existing snapshot on the destination computer can be deleted by opening a Shell on Beta and running this command:
zfs destroy -R email@example.com
Then send the snapshot manually again. Snapshots on the destination
system, Beta, can be listed from the Shell with
zfs list -t snapshot or by going to
Storage → Snapshots.
Error messages here can indicate any remaining problems.
A scrub is the process of ZFS scanning through the data on a volume. Scrubs help to identify data integrity problems, detect silent data corruptions caused by transient hardware issues, and provide early alerts of impending disk failures. FreeNAS® makes it easy to schedule periodic automatic scrubs.
Each volume should be scrubbed at least once a month. Bit errors in critical data can be detected by ZFS, but only when that data is read. Scheduled scrubs can find bit errors in rarely-read data. The amount of time needed for a scrub is proportional to the quantity of data on the volume. Typical scrubs take several hours or longer.
The scrub process is I/O intensive and can negatively impact performance. Schedule scrubs for evenings or weekends to minimize impact to users. Make certain that scrubs and other disk-intensive activity like S.M.A.R.T. Tests are scheduled to run on different days to avoid disk contention and extreme performance impacts.
Scrubs only check used disk space. To check unused disk space, schedule S.M.A.R.T. Tests of Type Long Self-Test to run once or twice a month.
Scrubs are scheduled and managed with
Storage → Scrubs.
When a volume is created, a ZFS scrub is automatically scheduled. An
entry with the same volume name is added to
Storage → Scrubs.
A summary of this entry can be viewed with
Storage → Scrubs → View Scrubs.
displays the default settings for the volume named
this example, the entry has been highlighted and the Edit
button clicked to display the Edit screen.
summarizes the options in this screen.
|Volume||drop-down menu||select ZFS volume to scrub|
|Threshold days||integer||number of days since the last scrub completed before the next scrub can occur, regardless of the calendar schedule; the default is a multiple of 7 which should ensure that the scrub always occurs on the same day of the week|
|Minute||slider or minute selections||if use the slider, scrub occurs every N minutes; if use minute selections, scrub starts at the highlighted minutes|
|Hour||slider or hour selections||if use the slider, scrub occurs every N hours; if use hour selections, scrub occurs at the highlighted hours|
|Day of Month||slider or month selections||if use the slider, scrub occurs every N days; if use month selections, scrub occurs on the highlighted days of the selected months|
|Month||checkboxes||scrub occurs on the selected months|
|Day of week||checkboxes||scrub occurs on the selected days; default is Sunday to least impact users|
|Enabled||checkbox||uncheck to disable the scheduled scrub without deleting it|
Review the default selections and, if necessary, modify them to meet the needs of the environment.
Scrubs can be deleted with the Delete button, but deleting a scrub is not recommended as a scrub provides an early indication of disk issues that could lead to a disk failure. If a scrub is too intensive for the hardware, consider unchecking the Enabled button for the scrub as a temporary measure until the hardware can be upgraded.
The Snapshots tab is used to review the listing of available snapshots. An example is shown in Figure 8.5.1.
If snapshots do not appear, check that the current time
configured in Periodic Snapshot Tasks does not conflict with
the Begin, End, and Interval
settings. If the snapshot was attempted but failed, an entry is
/var/log/messages. This log file can be viewed in
The listing includes the name of the volume or dataset, the name of each snapshot, and the amount of used and referenced data.
Used is the amount of space consumed by this dataset and all of its descendants. This value is checked against the dataset’s quota and reservation. The space used does not include the dataset’s reservation, but does take into account the reservations of any descendent datasets. The amount of space that a dataset consumes from its parent, as well as the amount of space that are freed if this dataset is recursively destroyed, is the greater of its space used and its reservation. When a snapshot is created, the space is initially shared between the snapshot and the filesystem, and possibly with previous snapshots. As the filesystem changes, space that was previously shared becomes unique to the snapshot, and is counted in the snapshot’s space used. Additionally, deleting snapshots can increase the amount of space unique to (and used by) other snapshots. The amount of space used, available, or referenced does not take into account pending changes. While pending changes are generally accounted for within a few seconds, disk changes do not necessarily guarantee that the space usage information is updated immediately.
Space used by individual snapshots can be seen by running
zfs list -t snapshot from Shell.
Refer indicates the amount of data accessible by this dataset, which may or may not be shared with other datasets in the pool. When a snapshot or clone is created, it initially references the same amount of space as the file system or snapshot it was created from, since its contents are identical.
Replication shows whether the snapshot has been replicated to a remote system.
Snapshots have icons on the right side for several actions.
Clone Snapshot prompts for the name of the clone to create. A clone is a writable copy of the snapshot. Since a clone is really a dataset which can be mounted, the clone appears in the Active Volumes tab, instead of the Periodic Snapshots tab, and has the word clone in its name.
Destroy Snapshot a pop-up message asks for confirmation. Child clones must be destroyed before their parent snapshot can be destroyed. While creating a snapshot is instantaneous, deleting a snapshot can be I/O intensive and can take a long time, especially when deduplication is enabled. In order to delete a block in a snapshot, ZFS has to walk all the allocated blocks to see if that block is used anywhere else; if it is not, it can be freed.
The most recent snapshot also has a Rollback Snapshot icon. Clicking the icon asks for confirmation before rolling back to this snapshot state. Confirming by clicking Yes causes any files that have changed since the snapshot was taken to be reverted back to their state at the time of the snapshot.
Rollback is a potentially dangerous operation and causes any configured replication tasks to fail as the replication system uses the existing snapshot when doing an incremental backup. To restore the data within a snapshot, the recommended steps are:
- Clone the desired snapshot.
- Share the clone with the share type or service running on the FreeNAS® system.
- After users have recovered the needed data, destroy the clone in the Active Volumes tab.
This approach does not destroy any on-disk data and has no impact on replication.
A range of snapshots can be selected with the mouse. Click on the
checkbox in the left column of the first snapshot, then press and hold
Shift and click on the checkbox for the end snapshot. This can
be used to select a range of obsolete snapshots to be deleted with the
Destroy icon at the bottom. Be cautious and careful when
deleting ranges of snapshots.
Periodic snapshots can be configured to appear as shadow copies in newer versions of Windows Explorer, as described in Configuring Shadow Copies. Users can access the files in the shadow copy using Explorer without requiring any interaction with the FreeNAS® graphical administrative interface.
The ZFS Snapshots screen allows the creation of filters to view snapshots by selected criteria. To create a filter, click the Define filter icon (near the text No filter applied). When creating a filter:
- select the column or leave the default of Any Column.
- select the condition. Possible conditions are: contains (default), is, starts with, ends with, does not contain, is not, does not start with, does not end with, and is empty.
- enter a value that meets your view criteria.
- click the Filter button to save your filter and exit the define filter screen. Alternately, click the + button to add another filter.
If you create multiple filters, select the filter to use before leaving the define filter screen. Once a filter is selected, the No filter applied text changes to Clear filter. If you click Clear filter, a pop-up message indicates that this removes the filter and all available snapshots are listed.
Storage → VMware-Snapshot
allows you to coordinate ZFS snapshots when using FreeNAS® as a VMware
datastore. Once this type of snapshot is created, FreeNAS® will
automatically snapshot any running VMware virtual machines before
taking a scheduled or manual ZFS snapshot of the dataset or zvol
backing that VMware datastore. The temporary VMware snapshots are then
deleted on the VMware side but still exist in the ZFS snapshot and can
be used as stable resurrection points in that snapshot. These
coordinated snapshots will be listed in Snapshots.
|Hostname||string||IP address or hostname of VMware host; when clustering, this is the vCenter server for the cluster|
|Username||string||user on VMware host with enough permission to snapshot virtual machines|
|Password||string||password associated with Username|
|ZFS Filesystem||drop-down menu||the filesystem to snapshot|
|Datastore||drop-down menu||after entering the Hostname, Username, and Password, click Fetch Datastores to populate the menu and select the datastore with which to synchronize|