745 lines
27 KiB
Plaintext
745 lines
27 KiB
Plaintext
What: /sys/block/<disk>/alignment_offset
|
||
Date: April 2009
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Storage devices may report a physical block size that is
|
||
bigger than the logical block size (for instance a drive
|
||
with 4KB physical sectors exposing 512-byte logical
|
||
blocks to the operating system). This parameter
|
||
indicates how many bytes the beginning of the device is
|
||
offset from the disk's natural alignment.
|
||
|
||
|
||
What: /sys/block/<disk>/discard_alignment
|
||
Date: May 2011
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Devices that support discard functionality may
|
||
internally allocate space in units that are bigger than
|
||
the exported logical block size. The discard_alignment
|
||
parameter indicates how many bytes the beginning of the
|
||
device is offset from the internal allocation unit's
|
||
natural alignment.
|
||
|
||
|
||
What: /sys/block/<disk>/diskseq
|
||
Date: February 2021
|
||
Contact: Matteo Croce <mcroce@microsoft.com>
|
||
Description:
|
||
The /sys/block/<disk>/diskseq files reports the disk
|
||
sequence number, which is a monotonically increasing
|
||
number assigned to every drive.
|
||
Some devices, like the loop device, refresh such number
|
||
every time the backing file is changed.
|
||
The value type is 64 bit unsigned.
|
||
|
||
|
||
What: /sys/block/<disk>/inflight
|
||
Date: October 2009
|
||
Contact: Jens Axboe <axboe@kernel.dk>, Nikanth Karthikesan <knikanth@suse.de>
|
||
Description:
|
||
Reports the number of I/O requests currently in progress
|
||
(pending / in flight) in a device driver. This can be less
|
||
than the number of requests queued in the block device queue.
|
||
The report contains 2 fields: one for read requests
|
||
and one for write requests.
|
||
The value type is unsigned int.
|
||
Cf. Documentation/block/stat.rst which contains a single value for
|
||
requests in flight.
|
||
This is related to /sys/block/<disk>/queue/nr_requests
|
||
and for SCSI device also its queue_depth.
|
||
|
||
|
||
What: /sys/block/<disk>/integrity/device_is_integrity_capable
|
||
Date: July 2014
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Indicates whether a storage device is capable of storing
|
||
integrity metadata. Set if the device is T10 PI-capable.
|
||
|
||
|
||
What: /sys/block/<disk>/integrity/format
|
||
Date: June 2008
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Metadata format for integrity capable block device.
|
||
E.g. T10-DIF-TYPE1-CRC.
|
||
|
||
|
||
What: /sys/block/<disk>/integrity/protection_interval_bytes
|
||
Date: July 2015
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Describes the number of data bytes which are protected
|
||
by one integrity tuple. Typically the device's logical
|
||
block size.
|
||
|
||
|
||
What: /sys/block/<disk>/integrity/read_verify
|
||
Date: June 2008
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Indicates whether the block layer should verify the
|
||
integrity of read requests serviced by devices that
|
||
support sending integrity metadata.
|
||
|
||
|
||
What: /sys/block/<disk>/integrity/tag_size
|
||
Date: June 2008
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Number of bytes of integrity tag space available per
|
||
512 bytes of data.
|
||
|
||
|
||
What: /sys/block/<disk>/integrity/write_generate
|
||
Date: June 2008
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Indicates whether the block layer should automatically
|
||
generate checksums for write requests bound for
|
||
devices that support receiving integrity metadata.
|
||
|
||
|
||
What: /sys/block/<disk>/<partition>/alignment_offset
|
||
Date: April 2009
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Storage devices may report a physical block size that is
|
||
bigger than the logical block size (for instance a drive
|
||
with 4KB physical sectors exposing 512-byte logical
|
||
blocks to the operating system). This parameter
|
||
indicates how many bytes the beginning of the partition
|
||
is offset from the disk's natural alignment.
|
||
|
||
|
||
What: /sys/block/<disk>/<partition>/discard_alignment
|
||
Date: May 2011
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
Devices that support discard functionality may
|
||
internally allocate space in units that are bigger than
|
||
the exported logical block size. The discard_alignment
|
||
parameter indicates how many bytes the beginning of the
|
||
partition is offset from the internal allocation unit's
|
||
natural alignment.
|
||
|
||
|
||
What: /sys/block/<disk>/<partition>/stat
|
||
Date: February 2008
|
||
Contact: Jerome Marchand <jmarchan@redhat.com>
|
||
Description:
|
||
The /sys/block/<disk>/<partition>/stat files display the
|
||
I/O statistics of partition <partition>. The format is the
|
||
same as the format of /sys/block/<disk>/stat.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/add_random
|
||
Date: June 2010
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This file allows to turn off the disk entropy contribution.
|
||
Default value of this file is '1'(on).
|
||
|
||
|
||
What: /sys/block/<disk>/queue/chunk_sectors
|
||
Date: September 2016
|
||
Contact: Hannes Reinecke <hare@suse.com>
|
||
Description:
|
||
[RO] chunk_sectors has different meaning depending on the type
|
||
of the disk. For a RAID device (dm-raid), chunk_sectors
|
||
indicates the size in 512B sectors of the RAID volume stripe
|
||
segment. For a zoned block device, either host-aware or
|
||
host-managed, chunk_sectors indicates the size in 512B sectors
|
||
of the zones of the device, with the eventual exception of the
|
||
last zone of the device which may be smaller.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/crypto/
|
||
Date: February 2022
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
The presence of this subdirectory of /sys/block/<disk>/queue/
|
||
indicates that the device supports inline encryption. This
|
||
subdirectory contains files which describe the inline encryption
|
||
capabilities of the device. For more information about inline
|
||
encryption, refer to Documentation/block/inline-encryption.rst.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/crypto/max_dun_bits
|
||
Date: February 2022
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This file shows the maximum length, in bits, of data unit
|
||
numbers accepted by the device in inline encryption requests.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/crypto/modes/<mode>
|
||
Date: February 2022
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] For each crypto mode (i.e., encryption/decryption
|
||
algorithm) the device supports with inline encryption, a file
|
||
will exist at this location. It will contain a hexadecimal
|
||
number that is a bitmask of the supported data unit sizes, in
|
||
bytes, for that crypto mode.
|
||
|
||
Currently, the crypto modes that may be supported are:
|
||
|
||
* AES-256-XTS
|
||
* AES-128-CBC-ESSIV
|
||
* Adiantum
|
||
|
||
For example, if a device supports AES-256-XTS inline encryption
|
||
with data unit sizes of 512 and 4096 bytes, the file
|
||
/sys/block/<disk>/queue/crypto/modes/AES-256-XTS will exist and
|
||
will contain "0x1200".
|
||
|
||
|
||
What: /sys/block/<disk>/queue/crypto/num_keyslots
|
||
Date: February 2022
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This file shows the number of keyslots the device has for
|
||
use with inline encryption.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/dax
|
||
Date: June 2016
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This file indicates whether the device supports Direct
|
||
Access (DAX), used by CPU-addressable storage to bypass the
|
||
pagecache. It shows '1' if true, '0' if not.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/discard_granularity
|
||
Date: May 2011
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] Devices that support discard functionality may internally
|
||
allocate space using units that are bigger than the logical
|
||
block size. The discard_granularity parameter indicates the size
|
||
of the internal allocation unit in bytes if reported by the
|
||
device. Otherwise the discard_granularity will be set to match
|
||
the device's physical block size. A discard_granularity of 0
|
||
means that the device does not support discard functionality.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/discard_max_bytes
|
||
Date: May 2011
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RW] While discard_max_hw_bytes is the hardware limit for the
|
||
device, this setting is the software limit. Some devices exhibit
|
||
large latencies when large discards are issued, setting this
|
||
value lower will make Linux issue smaller discards and
|
||
potentially help reduce latencies induced by large discard
|
||
operations.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/discard_max_hw_bytes
|
||
Date: July 2015
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] Devices that support discard functionality may have
|
||
internal limits on the number of bytes that can be trimmed or
|
||
unmapped in a single operation. The `discard_max_hw_bytes`
|
||
parameter is set by the device driver to the maximum number of
|
||
bytes that can be discarded in a single operation. Discard
|
||
requests issued to the device must not exceed this limit. A
|
||
`discard_max_hw_bytes` value of 0 means that the device does not
|
||
support discard functionality.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/discard_zeroes_data
|
||
Date: May 2011
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] Will always return 0. Don't rely on any specific behavior
|
||
for discards, and don't read this file.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/dma_alignment
|
||
Date: May 2022
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
Reports the alignment that user space addresses must have to be
|
||
used for raw block device access with O_DIRECT and other driver
|
||
specific passthrough mechanisms.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/fua
|
||
Date: May 2018
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] Whether or not the block driver supports the FUA flag for
|
||
write requests. FUA stands for Force Unit Access. If the FUA
|
||
flag is set that means that write requests must bypass the
|
||
volatile cache of the storage device.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/hw_sector_size
|
||
Date: January 2008
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This is the hardware sector size of the device, in bytes.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/independent_access_ranges/
|
||
Date: October 2021
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] The presence of this sub-directory of the
|
||
/sys/block/xxx/queue/ directory indicates that the device is
|
||
capable of executing requests targeting different sector ranges
|
||
in parallel. For instance, single LUN multi-actuator hard-disks
|
||
will have an independent_access_ranges directory if the device
|
||
correctly advertizes the sector ranges of its actuators.
|
||
|
||
The independent_access_ranges directory contains one directory
|
||
per access range, with each range described using the sector
|
||
(RO) attribute file to indicate the first sector of the range
|
||
and the nr_sectors (RO) attribute file to indicate the total
|
||
number of sectors in the range starting from the first sector of
|
||
the range. For example, a dual-actuator hard-disk will have the
|
||
following independent_access_ranges entries.::
|
||
|
||
$ tree /sys/block/<disk>/queue/independent_access_ranges/
|
||
/sys/block/<disk>/queue/independent_access_ranges/
|
||
|-- 0
|
||
| |-- nr_sectors
|
||
| `-- sector
|
||
`-- 1
|
||
|-- nr_sectors
|
||
`-- sector
|
||
|
||
The sector and nr_sectors attributes use 512B sector unit,
|
||
regardless of the actual block size of the device. Independent
|
||
access ranges do not overlap and include all sectors within the
|
||
device capacity. The access ranges are numbered in increasing
|
||
order of the range start sector, that is, the sector attribute
|
||
of range 0 always has the value 0.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/io_poll
|
||
Date: November 2015
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] When read, this file shows whether polling is enabled (1)
|
||
or disabled (0). Writing '0' to this file will disable polling
|
||
for this device. Writing any non-zero value will enable this
|
||
feature.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/io_poll_delay
|
||
Date: November 2016
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] If polling is enabled, this controls what kind of polling
|
||
will be performed. It defaults to -1, which is classic polling.
|
||
In this mode, the CPU will repeatedly ask for completions
|
||
without giving up any time. If set to 0, a hybrid polling mode
|
||
is used, where the kernel will attempt to make an educated guess
|
||
at when the IO will complete. Based on this guess, the kernel
|
||
will put the process issuing IO to sleep for an amount of time,
|
||
before entering a classic poll loop. This mode might be a little
|
||
slower than pure classic polling, but it will be more efficient.
|
||
If set to a value larger than 0, the kernel will put the process
|
||
issuing IO to sleep for this amount of microseconds before
|
||
entering classic polling.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/io_timeout
|
||
Date: November 2018
|
||
Contact: Weiping Zhang <zhangweiping@didiglobal.com>
|
||
Description:
|
||
[RW] io_timeout is the request timeout in milliseconds. If a
|
||
request does not complete in this time then the block driver
|
||
timeout handler is invoked. That timeout handler can decide to
|
||
retry the request, to fail it or to start a device recovery
|
||
strategy.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/iostats
|
||
Date: January 2009
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This file is used to control (on/off) the iostats
|
||
accounting of the disk.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/logical_block_size
|
||
Date: May 2009
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] This is the smallest unit the storage device can address.
|
||
It is typically 512 bytes.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_active_zones
|
||
Date: July 2020
|
||
Contact: Niklas Cassel <niklas.cassel@wdc.com>
|
||
Description:
|
||
[RO] For zoned block devices (zoned attribute indicating
|
||
"host-managed" or "host-aware"), the sum of zones belonging to
|
||
any of the zone states: EXPLICIT OPEN, IMPLICIT OPEN or CLOSED,
|
||
is limited by this value. If this value is 0, there is no limit.
|
||
|
||
If the host attempts to exceed this limit, the driver should
|
||
report this error with BLK_STS_ZONE_ACTIVE_RESOURCE, which user
|
||
space may see as the EOVERFLOW errno.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_discard_segments
|
||
Date: February 2017
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] The maximum number of DMA scatter/gather entries in a
|
||
discard request.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_hw_sectors_kb
|
||
Date: September 2004
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This is the maximum number of kilobytes supported in a
|
||
single data transfer.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_integrity_segments
|
||
Date: September 2010
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] Maximum number of elements in a DMA scatter/gather list
|
||
with integrity data that will be submitted by the block layer
|
||
core to the associated block driver.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_open_zones
|
||
Date: July 2020
|
||
Contact: Niklas Cassel <niklas.cassel@wdc.com>
|
||
Description:
|
||
[RO] For zoned block devices (zoned attribute indicating
|
||
"host-managed" or "host-aware"), the sum of zones belonging to
|
||
any of the zone states: EXPLICIT OPEN or IMPLICIT OPEN, is
|
||
limited by this value. If this value is 0, there is no limit.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_sectors_kb
|
||
Date: September 2004
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This is the maximum number of kilobytes that the block
|
||
layer will allow for a filesystem request. Must be smaller than
|
||
or equal to the maximum size allowed by the hardware. Write 0
|
||
to use default kernel settings.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_segment_size
|
||
Date: March 2010
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] Maximum size in bytes of a single element in a DMA
|
||
scatter/gather list.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/max_segments
|
||
Date: March 2010
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] Maximum number of elements in a DMA scatter/gather list
|
||
that is submitted to the associated block driver.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/minimum_io_size
|
||
Date: April 2009
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] Storage devices may report a granularity or preferred
|
||
minimum I/O size which is the smallest request the device can
|
||
perform without incurring a performance penalty. For disk
|
||
drives this is often the physical block size. For RAID arrays
|
||
it is often the stripe chunk size. A properly aligned multiple
|
||
of minimum_io_size is the preferred request size for workloads
|
||
where a high number of I/O operations is desired.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/nomerges
|
||
Date: January 2010
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] Standard I/O elevator operations include attempts to merge
|
||
contiguous I/Os. For known random I/O loads these attempts will
|
||
always fail and result in extra cycles being spent in the
|
||
kernel. This allows one to turn off this behavior on one of two
|
||
ways: When set to 1, complex merge checks are disabled, but the
|
||
simple one-shot merges with the previous I/O request are
|
||
enabled. When set to 2, all merge tries are disabled. The
|
||
default value is 0 - which enables all types of merge tries.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/nr_requests
|
||
Date: July 2003
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This controls how many requests may be allocated in the
|
||
block layer for read or write requests. Note that the total
|
||
allocated number may be twice this amount, since it applies only
|
||
to reads or writes (not the accumulated sum).
|
||
|
||
To avoid priority inversion through request starvation, a
|
||
request queue maintains a separate request pool per each cgroup
|
||
when CONFIG_BLK_CGROUP is enabled, and this parameter applies to
|
||
each such per-block-cgroup request pool. IOW, if there are N
|
||
block cgroups, each request queue may have up to N request
|
||
pools, each independently regulated by nr_requests.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/nr_zones
|
||
Date: November 2018
|
||
Contact: Damien Le Moal <damien.lemoal@wdc.com>
|
||
Description:
|
||
[RO] nr_zones indicates the total number of zones of a zoned
|
||
block device ("host-aware" or "host-managed" zone model). For
|
||
regular block devices, the value is always 0.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/optimal_io_size
|
||
Date: April 2009
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] Storage devices may report an optimal I/O size, which is
|
||
the device's preferred unit for sustained I/O. This is rarely
|
||
reported for disk drives. For RAID arrays it is usually the
|
||
stripe width or the internal track size. A properly aligned
|
||
multiple of optimal_io_size is the preferred request size for
|
||
workloads where sustained throughput is desired. If no optimal
|
||
I/O size is reported this file contains 0.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/physical_block_size
|
||
Date: May 2009
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] This is the smallest unit a physical storage device can
|
||
write atomically. It is usually the same as the logical block
|
||
size but may be bigger. One example is SATA drives with 4KB
|
||
sectors that expose a 512-byte logical block size to the
|
||
operating system. For stacked block devices the
|
||
physical_block_size variable contains the maximum
|
||
physical_block_size of the component devices.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/read_ahead_kb
|
||
Date: May 2004
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] Maximum number of kilobytes to read-ahead for filesystems
|
||
on this block device.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/rotational
|
||
Date: January 2009
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This file is used to stat if the device is of rotational
|
||
type or non-rotational type.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/rq_affinity
|
||
Date: September 2008
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] If this option is '1', the block layer will migrate request
|
||
completions to the cpu "group" that originally submitted the
|
||
request. For some workloads this provides a significant
|
||
reduction in CPU cycles due to caching effects.
|
||
|
||
For storage configurations that need to maximize distribution of
|
||
completion processing setting this option to '2' forces the
|
||
completion to run on the requesting cpu (bypassing the "group"
|
||
aggregation logic).
|
||
|
||
|
||
What: /sys/block/<disk>/queue/scheduler
|
||
Date: October 2004
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] When read, this file will display the current and available
|
||
IO schedulers for this block device. The currently active IO
|
||
scheduler will be enclosed in [] brackets. Writing an IO
|
||
scheduler name to this file will switch control of this block
|
||
device to that new IO scheduler. Note that writing an IO
|
||
scheduler name to this file will attempt to load that IO
|
||
scheduler module, if it isn't already present in the system.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/stable_writes
|
||
Date: September 2020
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This file will contain '1' if memory must not be modified
|
||
while it is being used in a write request to this device. When
|
||
this is the case and the kernel is performing writeback of a
|
||
page, the kernel will wait for writeback to complete before
|
||
allowing the page to be modified again, rather than allowing
|
||
immediate modification as is normally the case. This
|
||
restriction arises when the device accesses the memory multiple
|
||
times where the same data must be seen every time -- for
|
||
example, once to calculate a checksum and once to actually write
|
||
the data. If no such restriction exists, this file will contain
|
||
'0'. This file is writable for testing purposes.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/throttle_sample_time
|
||
Date: March 2017
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] This is the time window that blk-throttle samples data, in
|
||
millisecond. blk-throttle makes decision based on the
|
||
samplings. Lower time means cgroups have more smooth throughput,
|
||
but higher CPU overhead. This exists only when
|
||
CONFIG_BLK_DEV_THROTTLING_LOW is enabled.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/virt_boundary_mask
|
||
Date: April 2021
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This file shows the I/O segment memory alignment mask for
|
||
the block device. I/O requests to this device will be split
|
||
between segments wherever either the memory address of the end
|
||
of the previous segment or the memory address of the beginning
|
||
of the current segment is not aligned to virt_boundary_mask + 1
|
||
bytes.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/wbt_lat_usec
|
||
Date: November 2016
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] If the device is registered for writeback throttling, then
|
||
this file shows the target minimum read latency. If this latency
|
||
is exceeded in a given window of time (see wb_window_usec), then
|
||
the writeback throttling will start scaling back writes. Writing
|
||
a value of '0' to this file disables the feature. Writing a
|
||
value of '-1' to this file resets the value to the default
|
||
setting.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/write_cache
|
||
Date: April 2016
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RW] When read, this file will display whether the device has
|
||
write back caching enabled or not. It will return "write back"
|
||
for the former case, and "write through" for the latter. Writing
|
||
to this file can change the kernels view of the device, but it
|
||
doesn't alter the device state. This means that it might not be
|
||
safe to toggle the setting from "write back" to "write through",
|
||
since that will also eliminate cache flushes issued by the
|
||
kernel.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/write_same_max_bytes
|
||
Date: January 2012
|
||
Contact: Martin K. Petersen <martin.petersen@oracle.com>
|
||
Description:
|
||
[RO] Some devices support a write same operation in which a
|
||
single data block can be written to a range of several
|
||
contiguous blocks on storage. This can be used to wipe areas on
|
||
disk or to initialize drives in a RAID configuration.
|
||
write_same_max_bytes indicates how many bytes can be written in
|
||
a single write same command. If write_same_max_bytes is 0, write
|
||
same is not supported by the device.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/write_zeroes_max_bytes
|
||
Date: November 2016
|
||
Contact: Chaitanya Kulkarni <chaitanya.kulkarni@wdc.com>
|
||
Description:
|
||
[RO] Devices that support write zeroes operation in which a
|
||
single request can be issued to zero out the range of contiguous
|
||
blocks on storage without having any payload in the request.
|
||
This can be used to optimize writing zeroes to the devices.
|
||
write_zeroes_max_bytes indicates how many bytes can be written
|
||
in a single write zeroes command. If write_zeroes_max_bytes is
|
||
0, write zeroes is not supported by the device.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/zone_append_max_bytes
|
||
Date: May 2020
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This is the maximum number of bytes that can be written to
|
||
a sequential zone of a zoned block device using a zone append
|
||
write operation (REQ_OP_ZONE_APPEND). This value is always 0 for
|
||
regular block devices.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/zone_write_granularity
|
||
Date: January 2021
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] This indicates the alignment constraint, in bytes, for
|
||
write operations in sequential zones of zoned block devices
|
||
(devices with a zoned attributed that reports "host-managed" or
|
||
"host-aware"). This value is always 0 for regular block devices.
|
||
|
||
|
||
What: /sys/block/<disk>/queue/zoned
|
||
Date: September 2016
|
||
Contact: Damien Le Moal <damien.lemoal@wdc.com>
|
||
Description:
|
||
[RO] zoned indicates if the device is a zoned block device and
|
||
the zone model of the device if it is indeed zoned. The
|
||
possible values indicated by zoned are "none" for regular block
|
||
devices and "host-aware" or "host-managed" for zoned block
|
||
devices. The characteristics of host-aware and host-managed
|
||
zoned block devices are described in the ZBC (Zoned Block
|
||
Commands) and ZAC (Zoned Device ATA Command Set) standards.
|
||
These standards also define the "drive-managed" zone model.
|
||
However, since drive-managed zoned block devices do not support
|
||
zone commands, they will be treated as regular block devices and
|
||
zoned will report "none".
|
||
|
||
|
||
What: /sys/block/<disk>/hidden
|
||
Date: March 2023
|
||
Contact: linux-block@vger.kernel.org
|
||
Description:
|
||
[RO] the block device is hidden. it doesn’t produce events, and
|
||
can’t be opened from userspace or using blkdev_get*.
|
||
Used for the underlying components of multipath devices.
|
||
|
||
|
||
What: /sys/block/<disk>/stat
|
||
Date: February 2008
|
||
Contact: Jerome Marchand <jmarchan@redhat.com>
|
||
Description:
|
||
The /sys/block/<disk>/stat files displays the I/O
|
||
statistics of disk <disk>. They contain 11 fields:
|
||
|
||
== ==============================================
|
||
1 reads completed successfully
|
||
2 reads merged
|
||
3 sectors read
|
||
4 time spent reading (ms)
|
||
5 writes completed
|
||
6 writes merged
|
||
7 sectors written
|
||
8 time spent writing (ms)
|
||
9 I/Os currently in progress
|
||
10 time spent doing I/Os (ms)
|
||
11 weighted time spent doing I/Os (ms)
|
||
12 discards completed
|
||
13 discards merged
|
||
14 sectors discarded
|
||
15 time spent discarding (ms)
|
||
16 flush requests completed
|
||
17 time spent flushing (ms)
|
||
== ==============================================
|
||
|
||
For more details refer Documentation/admin-guide/iostats.rst
|