ZPOOL-EVENTS(8) FreeBSD System Manager's Manual ZPOOL-EVENTS(8)
NAME
zpool-events - list recent events generated by kernel
SYNOPSIS
zpool events [-vHf] [pool]
zpool events -c
DESCRIPTION
Lists all recent events generated by the ZFS kernel modules. These
events are consumed by the zed(8) and used to automate administrative
tasks such as replacing a failed device with a hot spare. For more
information about the subclasses and event payloads that can be generated
see EVENTS and the following sections.
OPTIONS
-c Clear all previous events.
-f Follow mode.
-H Scripted mode. Do not display headers, and separate fields by a
single tab instead of arbitrary space.
-v Print the entire payload for each event.
EVENTS
These are the different event subclasses. The full event name would be
ereport.fs.zfs.SUBCLASS, but only the last part is listed here.
checksum Issued when a checksum error has been detected.
io Issued when there is an I/O error in a vdev in the
pool.
data Issued when there have been data errors in the pool.
deadman Issued when an I/O request is determined to be "hung",
this can be caused by lost completion events due to
flaky hardware or drivers. See zfs_deadman_failmode
in zfs(4) for additional information regarding "hung"
I/O detection and configuration.
delay Issued when a completed I/O request exceeds the
maximum allowed time specified by the zio_slow_io_ms
module parameter. This can be an indicator of
problems with the underlying storage device. The
number of delay events is ratelimited by the
zfs_slow_io_events_per_second module parameter.
config Issued every time a vdev change have been done to the
pool.
zpool Issued when a pool cannot be imported.
zpool.destroy Issued when a pool is destroyed.
zpool.export Issued when a pool is exported.
zpool.import Issued when a pool is imported.
zpool.reguid Issued when a REGUID (new unique identifier for the
pool have been regenerated) have been detected.
vdev.unknown Issued when the vdev is unknown. Such as trying to
clear device errors on a vdev that have failed/been
kicked from the system/pool and is no longer
available.
vdev.open_failed Issued when a vdev could not be opened (because it
didn't exist for example).
vdev.corrupt_data Issued when corrupt data have been detected on a vdev.
vdev.no_replicas Issued when there are no more replicas to sustain the
pool. This would lead to the pool being DEGRADED.
vdev.bad_guid_sum Issued when a missing device in the pool have been
detected.
vdev.too_small Issued when the system (kernel) have removed a device,
and ZFS notices that the device isn't there any more.
This is usually followed by a probe_failure event.
vdev.bad_label Issued when the label is OK but invalid.
vdev.bad_ashift Issued when the ashift alignment requirement has
increased.
vdev.remove Issued when a vdev is detached from a mirror (or a
spare detached from a vdev where it have been used to
replace a failed drive - only works if the original
drive have been re-added).
vdev.clear Issued when clearing device errors in a pool. Such as
running zpool clear on a device in the pool.
vdev.check Issued when a check to see if a given vdev could be
opened is started.
vdev.spare Issued when a spare have kicked in to replace a failed
device.
vdev.autoexpand Issued when a vdev can be automatically expanded.
io_failure Issued when there is an I/O failure in a vdev in the
pool.
probe_failure Issued when a probe fails on a vdev. This would occur
if a vdev have been kicked from the system outside of
ZFS (such as the kernel have removed the device).
log_replay Issued when the intent log cannot be replayed. The
can occur in the case of a missing or damaged log
device.
resilver.start Issued when a resilver is started.
resilver.finish Issued when the running resilver have finished.
scrub.start Issued when a scrub is started on a pool.
scrub.finish Issued when a pool has finished scrubbing.
scrub.abort Issued when a scrub is aborted on a pool.
scrub.resume Issued when a scrub is resumed on a pool.
scrub.paused Issued when a scrub is paused on a pool.
bootfs.vdev.attach
PAYLOADS
This is the payload (data, information) that accompanies an event.
For zed(8), these are set to uppercase and prefixed with ZEVENT_.
pool Pool name.
pool_failmode Failmode - wait, continue, or panic. See the failmode
property in zpoolprops(7) for more information.
pool_guid The GUID of the pool.
pool_context The load state for the pool (0=none, 1=open, 2=import,
3=tryimport, 4=recover 5=error).
vdev_guid The GUID of the vdev in question (the vdev failing or
operated upon with zpool clear, etc.).
vdev_type Type of vdev - disk, file, mirror, etc. See the
Virtual Devices section of zpoolconcepts(7) for more
information on possible values.
vdev_path Full path of the vdev, including any -partX.
vdev_devid ID of vdev (if any).
vdev_fru Physical FRU location.
vdev_state State of vdev (0=uninitialized, 1=closed, 2=offline,
3=removed, 4=failed to open, 5=faulted, 6=degraded,
7=healthy).
vdev_ashift The ashift value of the vdev.
vdev_complete_ts The time the last I/O request completed for the
specified vdev.
vdev_delta_ts The time since the last I/O request completed for the
specified vdev.
vdev_spare_paths List of spares, including full path and any -partX.
vdev_spare_guids GUID(s) of spares.
vdev_read_errors How many read errors that have been detected on the
vdev.
vdev_write_errors How many write errors that have been detected on the
vdev.
vdev_cksum_errors How many checksum errors that have been detected on
the vdev.
parent_guid GUID of the vdev parent.
parent_type Type of parent. See vdev_type.
parent_path Path of the vdev parent (if any).
parent_devid ID of the vdev parent (if any).
zio_objset The object set number for a given I/O request.
zio_object The object number for a given I/O request.
zio_level The indirect level for the block. Level 0 is the
lowest level and includes data blocks. Values > 0
indicate metadata blocks at the appropriate level.
zio_blkid The block ID for a given I/O request.
zio_err The error number for a failure when handling a given
I/O request, compatible with errno(3) with the value
of EBADE used to indicate a ZFS checksum error.
zio_offset The offset in bytes of where to write the I/O request
for the specified vdev.
zio_size The size in bytes of the I/O request.
zio_flags The current flags describing how the I/O request
should be handled. See the I/O FLAGS section for the
full list of I/O flags.
zio_stage The current stage of the I/O in the pipeline. See the
I/O STAGES section for a full list of all the I/O
stages.
zio_pipeline The valid pipeline stages for the I/O. See the I/O
STAGES section for a full list of all the I/O stages.
zio_delay The time elapsed (in nanoseconds) waiting for the
block layer to complete the I/O request. Unlike
zio_delta, this does not include any vdev queuing time
and is therefore solely a measure of the block layer
performance.
zio_timestamp The time when a given I/O request was submitted.
zio_delta The time required to service a given I/O request.
prev_state The previous state of the vdev.
cksum_expected The expected checksum value for the block.
cksum_actual The actual checksum value for an errant block.
cksum_algorithm Checksum algorithm used. See zfsprops(7) for more
information on the available checksum algorithms.
cksum_byteswap Whether or not the data is byteswapped.
bad_ranges [start, end) pairs of corruption offsets. Offsets are
always aligned on a 64-bit boundary, and can include
some gaps of non-corruption. (See bad_ranges_min_gap)
bad_ranges_min_gap
In order to bound the size of the bad_ranges array,
gaps of non-corruption less than or equal to
bad_ranges_min_gap bytes have been merged with
adjacent corruption. Always at least 8 bytes, since
corruption is detected on a 64-bit word basis.
bad_range_sets This array has one element per range in bad_ranges.
Each element contains the count of bits in that range
which were clear in the good data and set in the bad
data.
bad_range_clears This array has one element per range in bad_ranges.
Each element contains the count of bits for that range
which were set in the good data and clear in the bad
data.
bad_set_bits If this field exists, it is an array of (bad data &
~(good data)); that is, the bits set in the bad data
which are cleared in the good data. Each element
corresponds a byte whose offset is in a range in
bad_ranges, and the array is ordered by offset. Thus,
the first element is the first byte in the first
bad_ranges range, and the last element is the last
byte in the last bad_ranges range.
bad_cleared_bits Like bad_set_bits, but contains (good data & ~(bad
data)); that is, the bits set in the good data which
are cleared in the bad data.
bad_set_histogram If this field exists, it is an array of counters.
Each entry counts bits set in a particular bit of a
big-endian uint64 type. The first entry counts bits
set in the high-order bit of the first byte, the 9th
byte, etc, and the last entry counts bits set of the
low-order bit of the 8th byte, the 16th byte, etc.
This information is useful for observing a stuck bit
in a parallel data path, such as IDE or parallel SCSI.
bad_cleared_histogram
If this field exists, it is an array of counters.
Each entry counts bit clears in a particular bit of a
big-endian uint64 type. The first entry counts bits
clears of the high-order bit of the first byte, the
9th byte, etc, and the last entry counts clears of the
low-order bit of the 8th byte, the 16th byte, etc.
This information is useful for observing a stuck bit
in a parallel data path, such as IDE or parallel SCSI.
I/O STAGES
The ZFS I/O pipeline is comprised of various stages which are defined
below. The individual stages are used to construct these basic I/O
operations: Read, Write, Free, Claim, and Ioctl. These stages may be set
on an event to describe the life cycle of a given I/O request.
Stage Bit Mask Operations
--------------------------- ---------- ----------
ZIO_STAGE_OPEN 0x00000001 RWFCI
ZIO_STAGE_READ_BP_INIT 0x00000002 R----
ZIO_STAGE_WRITE_BP_INIT 0x00000004 -W---
ZIO_STAGE_FREE_BP_INIT 0x00000008 --F--
ZIO_STAGE_ISSUE_ASYNC 0x00000010 RWF--
ZIO_STAGE_WRITE_COMPRESS 0x00000020 -W---
ZIO_STAGE_ENCRYPT 0x00000040 -W---
ZIO_STAGE_CHECKSUM_GENERATE 0x00000080 -W---
ZIO_STAGE_NOP_WRITE 0x00000100 -W---
ZIO_STAGE_DDT_READ_START 0x00000200 R----
ZIO_STAGE_DDT_READ_DONE 0x00000400 R----
ZIO_STAGE_DDT_WRITE 0x00000800 -W---
ZIO_STAGE_DDT_FREE 0x00001000 --F--
ZIO_STAGE_GANG_ASSEMBLE 0x00002000 RWFC-
ZIO_STAGE_GANG_ISSUE 0x00004000 RWFC-
ZIO_STAGE_DVA_THROTTLE 0x00008000 -W---
ZIO_STAGE_DVA_ALLOCATE 0x00010000 -W---
ZIO_STAGE_DVA_FREE 0x00020000 --F--
ZIO_STAGE_DVA_CLAIM 0x00040000 ---C-
ZIO_STAGE_READY 0x00080000 RWFCI
ZIO_STAGE_VDEV_IO_START 0x00100000 RW--I
ZIO_STAGE_VDEV_IO_DONE 0x00200000 RW--I
ZIO_STAGE_VDEV_IO_ASSESS 0x00400000 RW--I
ZIO_STAGE_CHECKSUM_VERIFY 0x00800000 R----
ZIO_STAGE_DONE 0x01000000 RWFCI
I/O FLAGS
Every I/O request in the pipeline contains a set of flags which describe
its function and are used to govern its behavior. These flags will be
set in an event as a zio_flags payload entry.
Flag Bit Mask
----------------------- ----------
ZIO_FLAG_DONT_AGGREGATE 0x00000001
ZIO_FLAG_IO_REPAIR 0x00000002
ZIO_FLAG_SELF_HEAL 0x00000004
ZIO_FLAG_RESILVER 0x00000008
ZIO_FLAG_SCRUB 0x00000010
ZIO_FLAG_SCAN_THREAD 0x00000020
ZIO_FLAG_PHYSICAL 0x00000040
ZIO_FLAG_CANFAIL 0x00000080
ZIO_FLAG_SPECULATIVE 0x00000100
ZIO_FLAG_CONFIG_WRITER 0x00000200
ZIO_FLAG_DONT_RETRY 0x00000400
ZIO_FLAG_DONT_CACHE 0x00000800
ZIO_FLAG_NODATA 0x00001000
ZIO_FLAG_INDUCE_DAMAGE 0x00002000
ZIO_FLAG_IO_ALLOCATING 0x00004000
ZIO_FLAG_IO_RETRY 0x00008000
ZIO_FLAG_PROBE 0x00010000
ZIO_FLAG_TRYHARD 0x00020000
ZIO_FLAG_OPTIONAL 0x00040000
ZIO_FLAG_DONT_QUEUE 0x00080000
ZIO_FLAG_DONT_PROPAGATE 0x00100000
ZIO_FLAG_IO_BYPASS 0x00200000
ZIO_FLAG_IO_REWRITE 0x00400000
ZIO_FLAG_RAW_COMPRESS 0x00800000
ZIO_FLAG_RAW_ENCRYPT 0x01000000
ZIO_FLAG_GANG_CHILD 0x02000000
ZIO_FLAG_DDT_CHILD 0x04000000
ZIO_FLAG_GODFATHER 0x08000000
ZIO_FLAG_NOPWRITE 0x10000000
ZIO_FLAG_REEXECUTED 0x20000000
ZIO_FLAG_DELEGATED 0x40000000
ZIO_FLAG_FASTWRITE 0x80000000
SEE ALSO
zfs(4), zed(8), zpool-wait(8)
FreeBSD 13.1-RELEASE-p6 May 27, 2021 FreeBSD 13.1-RELEASE-p6
man2web Home...