table of contents
- bullseye 2.0.3-9
- bullseye-backports 2.1.9-1~bpo11+1
- testing 2.1.9-2
- unstable 2.1.9-3
|ZPOOL-EVENTS(8)||System Manager's Manual||ZPOOL-EVENTS(8)|
list recent events generated by kernel
Lists all recent events generated by the ZFS kernel modules. These events are consumed by the zed(8) and used to automate administrative tasks such as replacing a failed device with a hot spare. For more information about the subclasses and event payloads that can be generated see EVENTS and the following sections.
These are the different event subclasses. The full event name would be ereport.fs.zfs.SUBCLASS, but only the last part is listed here.
- Issued when a checksum error has been detected.
- Issued when there is an I/O error in a vdev in the pool.
- Issued when there have been data errors in the pool.
- Issued when an I/O request is determined to be "hung", this can be caused by lost completion events due to flaky hardware or drivers. See zfs_deadman_failmode in zfs(4) for additional information regarding "hung" I/O detection and configuration.
- Issued when a completed I/O request exceeds the maximum allowed time specified by the zio_slow_io_ms module parameter. This can be an indicator of problems with the underlying storage device. The number of delay events is ratelimited by the zfs_slow_io_events_per_second module parameter.
- Issued every time a vdev change have been done to the pool.
- Issued when a pool cannot be imported.
- Issued when a pool is destroyed.
- Issued when a pool is exported.
- Issued when a pool is imported.
- Issued when a REGUID (new unique identifier for the pool have been regenerated) have been detected.
- Issued when the vdev is unknown. Such as trying to clear device errors on a vdev that have failed/been kicked from the system/pool and is no longer available.
- Issued when a vdev could not be opened (because it didn't exist for example).
- Issued when corrupt data have been detected on a vdev.
- Issued when there are no more replicas to sustain the pool. This would lead to the pool being DEGRADED.
- Issued when a missing device in the pool have been detected.
- Issued when the system (kernel) have removed a device, and ZFS notices that the device isn't there any more. This is usually followed by a probe_failure event.
- Issued when the label is OK but invalid.
- Issued when the ashift alignment requirement has increased.
- Issued when a vdev is detached from a mirror (or a spare detached from a vdev where it have been used to replace a failed drive - only works if the original drive have been re-added).
- Issued when clearing device errors in a pool. Such as running
clearon a device in the pool.
- Issued when a check to see if a given vdev could be opened is started.
- Issued when a spare have kicked in to replace a failed device.
- Issued when a vdev can be automatically expanded.
- Issued when there is an I/O failure in a vdev in the pool.
- Issued when a probe fails on a vdev. This would occur if a vdev have been kicked from the system outside of ZFS (such as the kernel have removed the device).
- Issued when the intent log cannot be replayed. The can occur in the case of a missing or damaged log device.
- Issued when a resilver is started.
- Issued when the running resilver have finished.
- Issued when a scrub is started on a pool.
- Issued when a pool has finished scrubbing.
- Issued when a scrub is aborted on a pool.
- Issued when a scrub is resumed on a pool.
- Issued when a scrub is paused on a pool.
This is the payload (data, information) that accompanies an event.
For zed(8), these are set to uppercase and prefixed with ZEVENT_.
- Pool name.
- Failmode - wait, continue, or panic. See the failmode property in zpoolprops(7) for more information.
- The GUID of the pool.
- The load state for the pool (0=none, 1=open, 2=import, 3=tryimport, 4=recover 5=error).
- The GUID of the vdev in question (the vdev failing or operated upon with
- Type of vdev - disk, file, mirror, etc. See the Virtual Devices section of zpoolconcepts(7) for more information on possible values.
- Full path of the vdev, including any -partX.
- ID of vdev (if any).
- Physical FRU location.
- State of vdev (0=uninitialized, 1=closed, 2=offline, 3=removed, 4=failed to open, 5=faulted, 6=degraded, 7=healthy).
- The ashift value of the vdev.
- The time the last I/O request completed for the specified vdev.
- The time since the last I/O request completed for the specified vdev.
- List of spares, including full path and any -partX.
- GUID(s) of spares.
- How many read errors that have been detected on the vdev.
- How many write errors that have been detected on the vdev.
- How many checksum errors that have been detected on the vdev.
- GUID of the vdev parent.
- Type of parent. See vdev_type.
- Path of the vdev parent (if any).
- ID of the vdev parent (if any).
- The object set number for a given I/O request.
- The object number for a given I/O request.
- The indirect level for the block. Level 0 is the lowest level and includes data blocks. Values > 0 indicate metadata blocks at the appropriate level.
- The block ID for a given I/O request.
- The error number for a failure when handling a given I/O request, compatible with errno(3) with the value of EBADE used to indicate a ZFS checksum error.
- The offset in bytes of where to write the I/O request for the specified vdev.
- The size in bytes of the I/O request.
- The current flags describing how the I/O request should be handled. See the I/O FLAGS section for the full list of I/O flags.
- The current stage of the I/O in the pipeline. See the I/O STAGES section for a full list of all the I/O stages.
- The valid pipeline stages for the I/O. See the I/O STAGES section for a full list of all the I/O stages.
- The time elapsed (in nanoseconds) waiting for the block layer to complete the I/O request. Unlike zio_delta, this does not include any vdev queuing time and is therefore solely a measure of the block layer performance.
- The time when a given I/O request was submitted.
- The time required to service a given I/O request.
- The previous state of the vdev.
- The expected checksum value for the block.
- The actual checksum value for an errant block.
- Checksum algorithm used. See zfsprops(7) for more information on the available checksum algorithms.
- Whether or not the data is byteswapped.
- [start, end) pairs of corruption offsets. Offsets are always aligned on a 64-bit boundary, and can include some gaps of non-corruption. (See bad_ranges_min_gap)
- In order to bound the size of the bad_ranges array, gaps of non-corruption less than or equal to bad_ranges_min_gap bytes have been merged with adjacent corruption. Always at least 8 bytes, since corruption is detected on a 64-bit word basis.
- This array has one element per range in bad_ranges. Each element contains the count of bits in that range which were clear in the good data and set in the bad data.
- This array has one element per range in bad_ranges. Each element contains the count of bits for that range which were set in the good data and clear in the bad data.
- If this field exists, it is an array of (bad data & ~(good data)); that is, the bits set in the bad data which are cleared in the good data. Each element corresponds a byte whose offset is in a range in bad_ranges, and the array is ordered by offset. Thus, the first element is the first byte in the first bad_ranges range, and the last element is the last byte in the last bad_ranges range.
- Like bad_set_bits, but contains (good data & ~(bad data)); that is, the bits set in the good data which are cleared in the bad data.
- If this field exists, it is an array of counters. Each entry counts bits set in a particular bit of a big-endian uint64 type. The first entry counts bits set in the high-order bit of the first byte, the 9th byte, etc, and the last entry counts bits set of the low-order bit of the 8th byte, the 16th byte, etc. This information is useful for observing a stuck bit in a parallel data path, such as IDE or parallel SCSI.
- If this field exists, it is an array of counters. Each entry counts bit clears in a particular bit of a big-endian uint64 type. The first entry counts bits clears of the high-order bit of the first byte, the 9th byte, etc, and the last entry counts clears of the low-order bit of the 8th byte, the 16th byte, etc. This information is useful for observing a stuck bit in a parallel data path, such as IDE or parallel SCSI.
The ZFS I/O pipeline is comprised of various stages which are defined below. The individual stages are used to construct these basic I/O operations: Read, Write, Free, Claim, and Ioctl. These stages may be set on an event to describe the life cycle of a given I/O request.
Every I/O request in the pipeline contains a set of flags which describe its function and are used to govern its behavior. These flags will be set in an event as a zio_flags payload entry.
|May 27, 2021||OpenZFS|