Content-type: text/html

<HTML><HEAD><TITLE>Manpage of ZPOOL-FEATURES</TITLE>
</HEAD><BODY>
<H1>ZPOOL-FEATURES</H1>
Section: File Formats (5)<BR>Updated: Jun 8, 2018<BR><A HREF="#index">Index</A>
<A HREF="/man/man2html">Return to Main Contents</A><HR>

<A NAME="lbAB">&nbsp;</A>
<H2>NAME</H2>

zpool-features - ZFS pool feature descriptions
<A NAME="lbAC">&nbsp;</A>
<H2>DESCRIPTION</H2>

<P>
<P>

ZFS pool on-disk format versions are specified via &quot;features&quot; which replace
the old on-disk format numbers (the last supported on-disk format number is
28). To enable a feature on a pool use the <B>upgrade</B> subcommand of the
<A HREF="/man/man2html?8+zpool">zpool</A>(8) command, or set the <B>feature@</B><I>feature_name</I> property
to <B>enabled</B>.
<P>
<P>

The pool format does not affect file system version compatibility or the ability
to send file systems between pools.
<P>
<P>

Since most features can be enabled independently of each other the on-disk
format of the pool is specified by the set of all features marked as
<B>active</B> on the pool. If the pool was created by another software version
this set may include unsupported features.
<A NAME="lbAD">&nbsp;</A>
<H3>Identifying features</H3>

<P>
<P>

Every feature has a GUID of the form <I>com.example:feature_name</I>. The
reverse DNS name ensures that the feature's GUID is unique across all ZFS
implementations. When unsupported features are encountered on a pool they will
be identified by their GUIDs. Refer to the documentation for the ZFS
implementation that created the pool for information about those features.
<P>
<P>

Each supported feature also has a short name. By convention a feature's short
name is the portion of its GUID which follows the ':' (e.g.
<I>com.example:feature_name</I> would have the short name <I>feature_name</I>),
however a feature's short name may differ across ZFS implementations if
following the convention would result in name conflicts.
<A NAME="lbAE">&nbsp;</A>
<H3>Feature states</H3>

<P>
<P>

Features can be in one of three states:
<P>


<B>active</B>

<DL COMPACT><DT><DD>
This feature's on-disk format changes are in effect on the pool. Support for
this feature is required to import the pool in read-write mode. If this
feature is not read-only compatible, support is also required to import the pool
in read-only mode (see &quot;Read-only compatibility&quot;).
</DL>

<P>
<P>


<B>enabled</B>

<DL COMPACT><DT><DD>
An administrator has marked this feature as enabled on the pool, but the
feature's on-disk format changes have not been made yet. The pool can still be
imported by software that does not support this feature, but changes may be made
to the on-disk format at any time which will move the feature to the
<B>active</B> state. Some features may support returning to the <B>enabled</B>
state after becoming <B>active</B>. See feature-specific documentation for
details.
</DL>

<P>
<P>


<B>disabled</B>

<DL COMPACT><DT><DD>
This feature's on-disk format changes have not been made and will not be made
unless an administrator moves the feature to the <B>enabled</B> state. Features
cannot be disabled once they have been enabled.
</DL>

<P>
<P>
<P>

The state of supported features is exposed through pool properties of the form
<I><A HREF="mailto:feature@short_name">feature@short_name</A></I>.
<A NAME="lbAF">&nbsp;</A>
<H3>Read-only compatibility</H3>

<P>
<P>

Some features may make on-disk format changes that do not interfere with other
software's ability to read from the pool. These features are referred to as
&quot;read-only compatible&quot;. If all unsupported features on a pool are read-only
compatible, the pool can be imported in read-only mode by setting the
<B>readonly</B> property during import (see <A HREF="/man/man2html?8+zpool">zpool</A>(8) for details on
importing pools).
<A NAME="lbAG">&nbsp;</A>
<H3>Unsupported features</H3>

<P>
<P>

For each unsupported feature enabled on an imported pool a pool property
named <I><A HREF="mailto:unsupported@feature_name">unsupported@feature_name</A></I> will indicate why the import was allowed
despite the unsupported feature. Possible values for this property are:
<P>
<P>


<B>inactive</B>

<DL COMPACT><DT><DD>
The feature is in the <B>enabled</B> state and therefore the pool's on-disk
format is still compatible with software that does not support this feature.
</DL>

<P>
<P>


<B>readonly</B>

<DL COMPACT><DT><DD>
The feature is read-only compatible and the pool has been imported in
read-only mode.
</DL>

<P>
<A NAME="lbAH">&nbsp;</A>
<H3>Feature dependencies</H3>

<P>
<P>

Some features depend on other features being enabled in order to function
properly. Enabling a feature will automatically enable any features it
depends on.
<A NAME="lbAI">&nbsp;</A>
<H2>FEATURES</H2>

<P>
<P>

The following features are supported on this system:
<P>
<P>


<B>allocation_classes</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.zfsonlinux:allocation_classes<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature enables support for separate allocation classes.
<P>
This feature becomes <B>active</B> when a dedicated allocation class vdev
(dedup or special) is created with the <B>zpool create</B> or <B>zpool add</B>
subcommands. With device removal, it can be returned to the <B>enabled</B>
state if all the dedicated allocation class vdevs are removed.
</DL>

<P>
<P>


<B>async_destroy</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:async_destroy<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
Destroying a file system requires traversing all of its data in order to
return its used space to the pool. Without <B>async_destroy</B> the file system
is not fully removed until all space has been reclaimed. If the destroy
operation is interrupted by a reboot or power outage the next attempt to open
the pool will need to complete the destroy operation synchronously.
<P>
When <B>async_destroy</B> is enabled the file system's data will be reclaimed
by a background process, allowing the destroy operation to complete without
traversing the entire file system. The background process is able to resume
interrupted destroys after the pool has been opened, eliminating the need
to finish interrupted destroys as part of the open operation. The amount
of space remaining to be reclaimed by the background process is available
through the <B>freeing</B> property.
<P>
This feature is only <B>active</B> while <B>freeing</B> is non-zero.
</DL>

<P>
<P>


<B>bookmarks</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:bookmarks<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables use of the <B>zfs bookmark</B> subcommand.
<P>
This feature is <B>active</B> while any bookmarks exist in the pool.
All bookmarks in the pool can be listed by running
<B>zfs list -t bookmark -r </B><I>poolname</I>.
</DL>

<P>
<P>


<B>bookmark_v2</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.datto:bookmark_v2<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>bookmark, extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables the creation and management of larger bookmarks which are
needed for other features in ZFS.
<P>
This feature becomes <B>active</B> when a v2 bookmark is created and will be
returned to the <B>enabled</B> state when all v2 bookmarks are destroyed.
</DL>

<P>
<P>


<B>device_removal</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:device_removal<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature enables the <B>zpool remove</B> subcommand to remove top-level
vdevs, evacuating them to reduce the total size of the pool.
<P>
This feature becomes <B>active</B> when the <B>zpool remove</B> subcommand is used
on a top-level vdev, and will never return to being <B>enabled</B>.
</DL>

<P>
<P>


<B>edonr</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.illumos:edonr<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables the use of the Edon-R hash algorithm for checksum,
including for nopwrite (if compression is also enabled, an overwrite of
a block whose checksum matches the data being written will be ignored).
In an abundance of caution, Edon-R requires verification when used with
dedup: <B>zfs set dedup=edonr,verify</B>.  See <B><A HREF="/man/man2html?8+zfs">zfs</A></B>(8).
<P>
Edon-R is a very high-performance hash algorithm that was part
of the NIST SHA-3 competition. It provides extremely high hash
performance (over 350% faster than SHA-256), but was not selected
because of its unsuitability as a general purpose secure hash algorithm.
This implementation utilizes the new salted checksumming functionality
in ZFS, which means that the checksum is pre-seeded with a secret
256-bit random key (stored on the pool) before being fed the data block
to be checksummed. Thus the produced checksums are unique to a given
pool.
<P>
When the <B>edonr</B> feature is set to <B>enabled</B>, the administrator
can turn on the <B>edonr</B> checksum on any dataset using the
<B>zfs set checksum=edonr</B>. See <A HREF="/man/man2html?8+zfs">zfs</A>(8). This feature becomes
<B>active</B> once a <B>checksum</B> property has been set to <B>edonr</B>,
and will return to being <B>enabled</B> once all filesystems that have
ever had their checksum set to <B>edonr</B> are destroyed.
<P>
The <B>edonr</B> feature is not supported by GRUB and must not be used on
the pool if GRUB needs to access the pool (e.g. for /boot).
</DL>

<P>
<P>


<B>embedded_data</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:embedded_data<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature improves the performance and compression ratio of
highly-compressible blocks.  Blocks whose contents can compress to 112 bytes
or smaller can take advantage of this feature.
<P>
When this feature is enabled, the contents of highly-compressible blocks are
stored in the block &quot;pointer&quot; itself (a misnomer in this case, as it contains
the compressed data, rather than a pointer to its location on disk).  Thus
the space of the block (one sector, typically 512 bytes or 4KB) is saved,
and no additional i/o is needed to read and write the data block.
<P>
This feature becomes <B>active</B> as soon as it is enabled and will
never return to being <B>enabled</B>.
</DL>

<P>
<P>


<B>empty_bpobj</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:empty_bpobj<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature increases the performance of creating and using a large
number of snapshots of a single filesystem or volume, and also reduces
the disk space required.
<P>
When there are many snapshots, each snapshot uses many Block Pointer
Objects (bpobj's) to track blocks associated with that snapshot.
However, in common use cases, most of these bpobj's are empty.  This
feature allows us to create each bpobj on-demand, thus eliminating the
empty bpobjs.
<P>
This feature is <B>active</B> while there are any filesystems, volumes,
or snapshots which were created after enabling this feature.
</DL>

<P>
<P>


<B>enabled_txg</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:enabled_txg<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
Once this feature is enabled ZFS records the transaction group number
in which new features are enabled. This has no user-visible impact,
but other features may depend on this feature.
<P>
This feature becomes <B>active</B> as soon as it is enabled and will
never return to being <B>enabled.
</DL>

<P>
<P>


encryption</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.datto:encryption<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>bookmark_v2, extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables the creation and management of natively encrypted datasets.
<P>
This feature becomes <B>active</B> when an encrypted dataset is created and will
be returned to the <B>enabled</B> state when all datasets that use this feature
are destroyed.
</DL>

<P>
<P>


<B>extensible_dataset</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:extensible_dataset<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature allows more flexible use of internal ZFS data structures,
and exists for other features to depend on.
<P>
This feature will be <B>active</B> when the first dependent feature uses it,
and will be returned to the <B>enabled</B> state when all datasets that use
this feature are destroyed.
</DL>

<P>
<P>


<B>filesystem_limits</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.joyent:filesystem_limits<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables filesystem and snapshot limits. These limits can be used
to control how many filesystems and/or snapshots can be created at the point in
the tree on which the limits are set.
<P>
This feature is <B>active</B> once either of the limit properties has been
set on a dataset. Once activated the feature is never deactivated.
</DL>

<P>
<P>


<B>hole_birth</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:hole_birth<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>enabled_txg<BR></TD></TR>
</TABLE>

<P>
This feature has/had bugs, the result of which is that, if you do a
<B>zfs send -i</B> (or <B>-R</B>, since it uses <B>-i</B>) from an affected
dataset, the receiver will not see any checksum or other errors, but the
resulting destination snapshot will not match the source.  Its use by
<B>zfs send -i</B> has been disabled by default.  See the
<B>send_holes_without_birth_time</B> module parameter in
<A HREF="/man/man2html?5+zfs-module-parameters">zfs-module-parameters</A>(5).
<P>
This feature improves performance of incremental sends (<B>zfs send -i</B>)
and receives for objects with many holes. The most common case of
hole-filled objects is zvols.
<P>
An incremental send stream from snapshot <B>A</B> to snapshot <B>B</B>
contains information about every block that changed between <B>A</B> and
<B>B</B>. Blocks which did not change between those snapshots can be
identified and omitted from the stream using a piece of metadata called
the 'block birth time', but birth times are not recorded for holes (blocks
filled only with zeroes). Since holes created after <B>A</B> cannot be
distinguished from holes created before <B>A</B>, information about every
hole in the entire filesystem or zvol is included in the send stream.
<P>
For workloads where holes are rare this is not a problem. However, when
incrementally replicating filesystems or zvols with many holes (for
example a zvol formatted with another filesystem) a lot of time will
be spent sending and receiving unnecessary information about holes that
already exist on the receiving side.
<P>
Once the <B>hole_birth</B> feature has been enabled the block birth times
of all new holes will be recorded. Incremental sends between snapshots
created after this feature is enabled will use this new metadata to avoid
sending information about holes that already exist on the receiving side.
<P>
This feature becomes <B>active</B> as soon as it is enabled and will
never return to being <B>enabled.
</DL>

<P>
<P>


large_blocks</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.open-zfs:large_blocks<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
The <B>large_block</B> feature allows the record size on a dataset to be
set larger than 128KB.
<P>
This feature becomes <B>active</B> once a dataset contains a file with
a block size larger than 128KB, and will return to being <B>enabled</B> once all
filesystems that have ever had their recordsize larger than 128KB are destroyed.
</DL>

<P>
<P>


<B>large_dnode</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.zfsonlinux:large_dnode<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
The <B>large_dnode</B> feature allows the size of dnodes in a dataset to be
set larger than 512B.
<P>
This feature becomes <B>active</B> once a dataset contains an object with
a dnode larger than 512B, which occurs as a result of setting the
<B>dnodesize</B> dataset property to a value other than <B>legacy</B>. The
feature will return to being <B>enabled</B> once all filesystems that
have ever contained a dnode larger than 512B are destroyed. Large dnodes
allow more data to be stored in the bonus buffer, thus potentially
improving performance by avoiding the use of spill blocks.
</DL>

<P>
<P>


<B>lz4_compress</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.illumos:lz4_compress<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
<B>lz4</B> is a high-performance real-time compression algorithm that
features significantly faster compression and decompression as well as a
higher compression ratio than the older <B>lzjb</B> compression.
Typically, <B>lz4</B> compression is approximately 50% faster on
compressible data and 200% faster on incompressible data than
<B>lzjb</B>. It is also approximately 80% faster on decompression, while
giving approximately 10% better compression ratio.
<P>
When the <B>lz4_compress</B> feature is set to <B>enabled</B>, the
administrator can turn on <B>lz4</B> compression on any dataset on the
pool using the <A HREF="/man/man2html?8+zfs">zfs</A>(8) command. Please note that doing so will
immediately activate the <B>lz4_compress</B> feature on the underlying
pool using the <A HREF="/man/man2html?8+zfs">zfs</A>(8) command. Also, all newly written metadata
will be compressed with <B>lz4</B> algorithm. Since this feature is not
read-only compatible, this operation will render the pool unimportable
on systems without support for the <B>lz4_compress</B> feature.
<P>
Booting off of <B>lz4</B>-compressed root pools is supported.
<P>
This feature becomes <B>active</B> as soon as it is enabled and will
never return to being <B>enabled.
</DL>

<P>
<P>


multi_vdev_crash_dump</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.joyent:multi_vdev_crash_dump<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature allows a dump device to be configured with a pool comprised
of multiple vdevs.  Those vdevs may be arranged in any mirrored or raidz
configuration.
<P>
When the <B>multi_vdev_crash_dump</B> feature is set to <B>enabled</B>,
the administrator can use the <B><A HREF="/man/man2html?1M+dumpadm">dumpadm</A></B>(1M) command to configure a
dump device on a pool comprised of multiple vdevs.
<P>
Under Linux this feature is registered for compatibility but not used.
New pools created under Linux will have the feature <B>enabled</B> but
will never transition to <B>active</B>.  This functionality is not
required in order to support crash dumps under Linux.  Existing pools
where this feature is <B>active</B> can be imported.
</DL>

<P>
<P>


<B>obsolete_counts</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:obsolete_counts<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>device_removal<BR></TD></TR>
</TABLE>

<P>
This feature is an enhancement of device_removal, which will over time
reduce the memory used to track removed devices.  When indirect blocks
are freed or remapped, we note that their part of the indirect mapping
is &quot;obsolete&quot;, i.e. no longer needed.
<P>
This feature becomes <B>active</B> when the <B>zpool remove</B> subcommand is
used on a top-level vdev, and will never return to being <B>enabled</B>.
</DL>

<P>
<P>


<B>project_quota</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.zfsonlinux:project_quota<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature allows administrators to account the spaces and objects usage
information against the project identifier (ID).
<P>
The project ID is new object-based attribute. When upgrading an existing
filesystem, object without project ID attribute will be assigned a zero
project ID. After this feature is enabled, newly created object will inherit
its parent directory's project ID if the parent inherit flag is set (via
<B>chattr +/-P</B> or <B>zfs project [-s|-C]</B>). Otherwise, the new object's
project ID will be set as zero. An object's project ID can be changed at
anytime by the owner (or privileged user) via <B>chattr -p $prjid</B> or
<B>zfs project -p $prjid</B>.
<P>
This feature will become <B>active</B> as soon as it is enabled and will never
return to being <B>disabled</B>. Each filesystem will be upgraded automatically
when remounted or when new file is created under that filesystem. The upgrade
can also be triggered on filesystems via `zfs set version=current &lt;pool/fs&gt;`.
The upgrade process runs in the background and may take a while to complete
for the filesystems containing a large number of files.
</DL>

<P>
<P>


<B>resilver_defer</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.datto:resilver_defer<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature allows zfs to postpone new resilvers if an existing one is already
in progress. Without this feature, any new resilvers will cause the currently
running one to be immediately restarted from the beginning.
<P>
This feature becomes <B>active</B> once a resilver has been deferred, and
returns to being <B>enabled</B> when the deferred resilver begins.
</DL>

<P>
<P>


<B>sha512</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.illumos:sha512<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables the use of the SHA-512/256 truncated hash algorithm
(FIPS 180-4) for checksum and dedup. The native 64-bit arithmetic of
SHA-512 provides an approximate 50% performance boost over SHA-256 on
64-bit hardware and is thus a good minimum-change replacement candidate
for systems where hash performance is important, but these systems
cannot for whatever reason utilize the faster <B>skein</B> and
<B>edonr</B> algorithms.
<P>
When the <B>sha512</B> feature is set to <B>enabled</B>, the administrator
can turn on the <B>sha512</B> checksum on any dataset using
<B>zfs set checksum=sha512</B>. See <A HREF="/man/man2html?8+zfs">zfs</A>(8). This feature becomes
<B>active</B> once a <B>checksum</B> property has been set to <B>sha512</B>,
and will return to being <B>enabled</B> once all filesystems that have
ever had their checksum set to <B>sha512</B> are destroyed.
<P>
The <B>sha512</B> feature is not supported by GRUB and must not be used on
the pool if GRUB needs to access the pool (e.g. for /boot).
</DL>

<P>
<P>


<B>skein</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.illumos:skein<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>no<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature enables the use of the Skein hash algorithm for checksum
and dedup. Skein is a high-performance secure hash algorithm that was a
finalist in the NIST SHA-3 competition. It provides a very high security
margin and high performance on 64-bit hardware (80% faster than
SHA-256). This implementation also utilizes the new salted checksumming
functionality in ZFS, which means that the checksum is pre-seeded with a
secret 256-bit random key (stored on the pool) before being fed the data
block to be checksummed. Thus the produced checksums are unique to a
given pool, preventing hash collision attacks on systems with dedup.
<P>
When the <B>skein</B> feature is set to <B>enabled</B>, the administrator
can turn on the <B>skein</B> checksum on any dataset using
<B>zfs set checksum=skein</B>. See <A HREF="/man/man2html?8+zfs">zfs</A>(8). This feature becomes
<B>active</B> once a <B>checksum</B> property has been set to <B>skein</B>,
and will return to being <B>enabled</B> once all filesystems that have
ever had their checksum set to <B>skein</B> are destroyed.
<P>
The <B>skein</B> feature is not supported by GRUB and must not be used on
the pool if GRUB needs to access the pool (e.g. for /boot).
</DL>

<P>
<P>


<B>spacemap_histogram</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:spacemap_histogram<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This features allows ZFS to maintain more information about how free space
is organized within the pool. If this feature is <B>enabled</B>, ZFS will
set this feature to <B>active</B> when a new space map object is created or
an existing space map is upgraded to the new format. Once the feature is
<B>active</B>, it will remain in that state until the pool is destroyed.
</DL>

<P>
<P>


<B>spacemap_v2</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:spacemap_v2<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature enables the use of the new space map encoding which
consists of two words (instead of one) whenever it is advantageous.
The new encoding allows space maps to represent large regions of
space more efficiently on-disk while also increasing their maximum
addressable offset.
<P>
This feature becomes <B>active</B> once it is <B>enabled</B>, and never
returns back to being <B>enabled</B>.
</DL>

<P>
<P>


<B>userobj_accounting</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>org.zfsonlinux:userobj_accounting<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>extensible_dataset<BR></TD></TR>
</TABLE>

<P>
This feature allows administrators to account the object usage information
by user and group.
<P>
This feature becomes <B>active</B> as soon as it is enabled and will never
return to being <B>enabled</B>. Each filesystem will be upgraded automatically
when remounted, or when new files are created under that filesystem.
The upgrade can also be started manually on filesystems by running
`zfs set version=current &lt;pool/fs&gt;`. The upgrade process runs in the background
and may take a while to complete for filesystems containing a large number of
files.
</DL>

<P>
<P>


<B>zpool_checkpoint</B>

<DL COMPACT><DT><DD>
<TABLE>
<TR VALIGN=top><TD>GUID</TD><TD>com.delphix:zpool_checkpoint<BR></TD></TR>
<TR VALIGN=top><TD>READ-ONLY COMPATIBLE</TD><TD>yes<BR></TD></TR>
<TR VALIGN=top><TD>DEPENDENCIES</TD><TD>none<BR></TD></TR>
</TABLE>

<P>
This feature enables the <B>zpool checkpoint</B> subcommand that can
checkpoint the state of the pool at the time it was issued and later
rewind back to it or discard it.
<P>
This feature becomes <B>active</B> when the <B>zpool checkpoint</B> subcommand
is used to checkpoint the pool.
The feature will only return back to being <B>enabled</B> when the pool
is rewound or the checkpoint has been discarded.
</DL>

<P>
<A NAME="lbAJ">&nbsp;</A>
<H2>SEE ALSO</H2>

<A HREF="/man/man2html?8+zpool">zpool</A>(8)
<P>

<HR>
<A NAME="index">&nbsp;</A><H2>Index</H2>
<DL>
<DT><A HREF="#lbAB">NAME</A><DD>
<DT><A HREF="#lbAC">DESCRIPTION</A><DD>
<DL>
<DT><A HREF="#lbAD">Identifying features</A><DD>
<DT><A HREF="#lbAE">Feature states</A><DD>
<DT><A HREF="#lbAF">Read-only compatibility</A><DD>
<DT><A HREF="#lbAG">Unsupported features</A><DD>
<DT><A HREF="#lbAH">Feature dependencies</A><DD>
</DL>
<DT><A HREF="#lbAI">FEATURES</A><DD>
<DT><A HREF="#lbAJ">SEE ALSO</A><DD>
</DL>
<HR>
This document was created by
<A HREF="/man/man2html">man2html</A>,
using the manual pages.<BR>
Time: 00:49:32 GMT, September 27, 2019
</BODY>
</HTML>
