1 # SPDX-License-Identifier: GPL-2.0-only
3 # Block device driver configuration
7 bool "Multiple devices driver support (RAID and LVM)"
10 Support multiple physical spindles through a single logical device.
11 Required for RAID and logical volume management.
16 tristate "RAID support"
17 select BLOCK_HOLDER_DEPRECATED if SYSFS
19 # BLOCK_LEGACY_AUTOLOAD requirement should be removed
20 # after relevant mdadm enhancements - to make "names=yes"
21 # the default - are widely available.
22 select BLOCK_LEGACY_AUTOLOAD
24 This driver lets you combine several hard disk partitions into one
25 logical block device. This can be used to simply append one
26 partition to another one or to combine several redundant hard disks
27 into a RAID1/4/5 device so as to provide protection against hard
28 disk failures. This is called "Software RAID" since the combining of
29 the partitions is done by the kernel. "Hardware RAID" means that the
30 combining is done by a dedicated controller; if you have such a
31 controller, you do not need to say Y here.
33 More information about Software RAID on Linux is contained in the
34 Software RAID mini-HOWTO, available from
35 <https://www.tldp.org/docs.html#howto>. There you will also learn
36 where to get the supporting user space utilities raidtools.
41 bool "Autodetect RAID arrays during kernel boot"
42 depends on BLK_DEV_MD=y
45 If you say Y here, then the kernel will try to autodetect raid
46 arrays as part of its boot process.
48 If you don't use raid and say Y, this autodetection can cause
49 a several-second delay in the boot time due to various
50 synchronisation steps that are part of this step.
55 bool "MD bitmap file support (deprecated)"
58 If you say Y here, support for write intent bitmaps in files on an
59 external file system is enabled. This is an alternative to the internal
60 bitmaps near the MD superblock, and very problematic code that abuses
61 various kernel APIs and can only work with files on a file system not
62 actually sitting on the MD device.
65 tristate "Linear (append) mode (deprecated)"
68 If you say Y here, then your multiple devices driver will be able to
69 use the so-called linear mode, i.e. it will combine the hard disk
70 partitions by simply appending one to the other.
72 To compile this as a module, choose M here: the module
73 will be called linear.
78 tristate "RAID-0 (striping) mode"
81 If you say Y here, then your multiple devices driver will be able to
82 use the so-called raid0 mode, i.e. it will combine the hard disk
83 partitions into one logical device in such a fashion as to fill them
84 up evenly, one chunk here and one chunk there. This will increase
85 the throughput rate if the partitions reside on distinct disks.
87 Information about Software RAID on Linux is contained in the
88 Software-RAID mini-HOWTO, available from
89 <https://www.tldp.org/docs.html#howto>. There you will also
90 learn where to get the supporting user space utilities raidtools.
92 To compile this as a module, choose M here: the module
98 tristate "RAID-1 (mirroring) mode"
101 A RAID-1 set consists of several disk drives which are exact copies
102 of each other. In the event of a mirror failure, the RAID driver
103 will continue to use the operational mirrors in the set, providing
104 an error free MD (multiple device) to the higher levels of the
105 kernel. In a set with N drives, the available space is the capacity
106 of a single drive, and the set protects against a failure of (N - 1)
109 Information about Software RAID on Linux is contained in the
110 Software-RAID mini-HOWTO, available from
111 <https://www.tldp.org/docs.html#howto>. There you will also
112 learn where to get the supporting user space utilities raidtools.
114 If you want to use such a RAID-1 set, say Y. To compile this code
115 as a module, choose M here: the module will be called raid1.
120 tristate "RAID-10 (mirrored striping) mode"
121 depends on BLK_DEV_MD
123 RAID-10 provides a combination of striping (RAID-0) and
124 mirroring (RAID-1) with easier configuration and more flexible
126 Unlike RAID-0, but like RAID-1, RAID-10 requires all devices to
127 be the same size (or at least, only as much as the smallest device
129 RAID-10 provides a variety of layouts that provide different levels
130 of redundancy and performance.
132 RAID-10 requires mdadm-1.7.0 or later, available at:
134 https://www.kernel.org/pub/linux/utils/raid/mdadm/
139 tristate "RAID-4/RAID-5/RAID-6 mode"
140 depends on BLK_DEV_MD
146 select ASYNC_RAID6_RECOV
148 A RAID-5 set of N drives with a capacity of C MB per drive provides
149 the capacity of C * (N - 1) MB, and protects against a failure
150 of a single drive. For a given sector (row) number, (N - 1) drives
151 contain data sectors, and one drive contains the parity protection.
152 For a RAID-4 set, the parity blocks are present on a single drive,
153 while a RAID-5 set distributes the parity across the drives in one
154 of the available parity distribution methods.
156 A RAID-6 set of N drives with a capacity of C MB per drive
157 provides the capacity of C * (N - 2) MB, and protects
158 against a failure of any two drives. For a given sector
159 (row) number, (N - 2) drives contain data sectors, and two
160 drives contains two independent redundancy syndromes. Like
161 RAID-5, RAID-6 distributes the syndromes across the drives
162 in one of the available parity distribution methods.
164 Information about Software RAID on Linux is contained in the
165 Software-RAID mini-HOWTO, available from
166 <https://www.tldp.org/docs.html#howto>. There you will also
167 learn where to get the supporting user space utilities raidtools.
169 If you want to use such a RAID-4/RAID-5/RAID-6 set, say Y. To
170 compile this code as a module, choose M here: the module
171 will be called raid456.
176 tristate "Multipath I/O support (deprecated)"
177 depends on BLK_DEV_MD
179 MD_MULTIPATH provides a simple multi-path personality for use
180 the MD framework. It is not under active development. New
181 projects should consider using DM_MULTIPATH which has more
182 features and more testing.
187 tristate "Faulty test module for MD (deprecated)"
188 depends on BLK_DEV_MD
190 The "faulty" module allows for a block device that occasionally returns
191 read or write errors. It is useful for testing.
197 tristate "Cluster Support for MD"
198 depends on BLK_DEV_MD
202 Clustering support for MD devices. This enables locking and
203 synchronization across multiple systems on the cluster, so all
204 nodes in the cluster can access the MD devices simultaneously.
206 This brings the redundancy (and uptime) of RAID levels across the
207 nodes of the cluster. Currently, it can work with raid1 and raid10
212 source "drivers/md/bcache/Kconfig"
214 config BLK_DEV_DM_BUILTIN
218 tristate "Device mapper support"
219 select BLOCK_HOLDER_DEPRECATED if SYSFS
220 select BLK_DEV_DM_BUILTIN
221 select BLK_MQ_STACKING
222 depends on DAX || DAX=n
224 Device-mapper is a low level volume manager. It works by allowing
225 people to specify mappings for ranges of logical sectors. Various
226 mapping types are available, in addition people may write their own
227 modules containing custom mappings if they wish.
229 Higher level volume managers such as LVM2 use this driver.
231 To compile this as a module, choose M here: the module will be
237 bool "Device mapper debugging support"
238 depends on BLK_DEV_DM
240 Enable this for messages that may help debug device-mapper problems.
246 depends on BLK_DEV_DM
248 This interface allows you to do buffered I/O on a device and acts
249 as a cache, holding recently-read blocks in memory and performing
252 config DM_DEBUG_BLOCK_MANAGER_LOCKING
253 bool "Block manager locking"
256 Block manager locking can catch various metadata corruption issues.
260 config DM_DEBUG_BLOCK_STACK_TRACING
261 bool "Keep stack trace of persistent data block lock holders"
262 depends on STACKTRACE_SUPPORT && DM_DEBUG_BLOCK_MANAGER_LOCKING
265 Enable this for messages that may help debug problems with the
266 block manager locking used by thin provisioning and caching.
272 depends on BLK_DEV_DM
274 Some bio locking schemes used by other device-mapper targets
275 including thin provisioning.
277 source "drivers/md/persistent-data/Kconfig"
280 tristate "Unstriped target"
281 depends on BLK_DEV_DM
283 Unstripes I/O so it is issued solely on a single drive in a HW
284 RAID0 or dm-striped target.
287 tristate "Crypt target support"
288 depends on BLK_DEV_DM
289 depends on (ENCRYPTED_KEYS || ENCRYPTED_KEYS=n)
290 depends on (TRUSTED_KEYS || TRUSTED_KEYS=n)
295 This device-mapper target allows you to create a device that
296 transparently encrypts the data on it. You'll need to activate
297 the ciphers you're going to use in the cryptoapi configuration.
299 For further information on dm-crypt and userspace tools see:
300 <https://gitlab.com/cryptsetup/cryptsetup/wikis/DMCrypt>
302 To compile this code as a module, choose M here: the module will
308 tristate "Snapshot target"
309 depends on BLK_DEV_DM
312 Allow volume managers to take writable snapshots of a device.
314 config DM_THIN_PROVISIONING
315 tristate "Thin provisioning target"
316 depends on BLK_DEV_DM
317 select DM_PERSISTENT_DATA
320 Provides thin provisioning and snapshots that share a data store.
323 tristate "Cache target (EXPERIMENTAL)"
324 depends on BLK_DEV_DM
326 select DM_PERSISTENT_DATA
329 dm-cache attempts to improve performance of a block device by
330 moving frequently used data to a smaller, higher performance
331 device. Different 'policy' plugins can be used to change the
332 algorithms used to select which blocks are promoted, demoted,
333 cleaned etc. It supports writeback and writethrough modes.
336 tristate "Stochastic MQ Cache Policy (EXPERIMENTAL)"
340 A cache policy that uses a multiqueue ordered by recent hits
341 to select which blocks should be promoted and demoted.
342 This is meant to be a general purpose policy. It prioritises
343 reads over writes. This SMQ policy (vs MQ) offers the promise
344 of less memory utilization, improved performance and increased
345 adaptability in the face of changing workloads.
348 tristate "Writecache target"
349 depends on BLK_DEV_DM
351 The writecache target caches writes on persistent memory or SSD.
352 It is intended for databases or other programs that need extremely
355 The writecache target doesn't cache reads because reads are supposed
356 to be cached in standard RAM.
359 tristate "Emulated block size target (EXPERIMENTAL)"
360 depends on BLK_DEV_DM && !HIGHMEM
363 dm-ebs emulates smaller logical block size on backing devices
364 with larger ones (e.g. 512 byte sectors on 4K native disks).
367 tristate "Era target (EXPERIMENTAL)"
368 depends on BLK_DEV_DM
370 select DM_PERSISTENT_DATA
373 dm-era tracks which parts of a block device are written to
374 over time. Useful for maintaining cache coherency when using
378 tristate "Clone target (EXPERIMENTAL)"
379 depends on BLK_DEV_DM
381 select DM_PERSISTENT_DATA
383 dm-clone produces a one-to-one copy of an existing, read-only source
384 device into a writable destination device. The cloned device is
385 visible/mountable immediately and the copy of the source device to the
386 destination device happens in the background, in parallel with user
392 tristate "Mirror target"
393 depends on BLK_DEV_DM
395 Allow volume managers to mirror logical volumes, also
396 needed for live data migration tools such as 'pvmove'.
398 config DM_LOG_USERSPACE
399 tristate "Mirror userspace logging"
400 depends on DM_MIRROR && NET
403 The userspace logging module provides a mechanism for
404 relaying the dm-dirty-log API to userspace. Log designs
405 which are more suited to userspace implementation (e.g.
406 shared storage logs) or experimental logs can be implemented
407 by leveraging this framework.
410 tristate "RAID 1/4/5/6/10 target"
411 depends on BLK_DEV_DM
418 A dm target that supports RAID1, RAID10, RAID4, RAID5 and RAID6 mappings
420 A RAID-5 set of N drives with a capacity of C MB per drive provides
421 the capacity of C * (N - 1) MB, and protects against a failure
422 of a single drive. For a given sector (row) number, (N - 1) drives
423 contain data sectors, and one drive contains the parity protection.
424 For a RAID-4 set, the parity blocks are present on a single drive,
425 while a RAID-5 set distributes the parity across the drives in one
426 of the available parity distribution methods.
428 A RAID-6 set of N drives with a capacity of C MB per drive
429 provides the capacity of C * (N - 2) MB, and protects
430 against a failure of any two drives. For a given sector
431 (row) number, (N - 2) drives contain data sectors, and two
432 drives contains two independent redundancy syndromes. Like
433 RAID-5, RAID-6 distributes the syndromes across the drives
434 in one of the available parity distribution methods.
437 tristate "Zero target"
438 depends on BLK_DEV_DM
440 A target that discards writes, and returns all zeroes for
441 reads. Useful in some recovery situations.
444 tristate "Multipath target"
445 depends on BLK_DEV_DM
446 # nasty syntax but means make DM_MULTIPATH independent
447 # of SCSI_DH if the latter isn't defined but if
448 # it is, DM_MULTIPATH must depend on it. We get a build
449 # error if SCSI_DH=m and DM_MULTIPATH=y
450 depends on !SCSI_DH || SCSI
452 Allow volume managers to support multipath hardware.
454 config DM_MULTIPATH_QL
455 tristate "I/O Path Selector based on the number of in-flight I/Os"
456 depends on DM_MULTIPATH
458 This path selector is a dynamic load balancer which selects
459 the path with the least number of in-flight I/Os.
463 config DM_MULTIPATH_ST
464 tristate "I/O Path Selector based on the service time"
465 depends on DM_MULTIPATH
467 This path selector is a dynamic load balancer which selects
468 the path expected to complete the incoming I/O in the shortest
473 config DM_MULTIPATH_HST
474 tristate "I/O Path Selector based on historical service time"
475 depends on DM_MULTIPATH
477 This path selector is a dynamic load balancer which selects
478 the path expected to complete the incoming I/O in the shortest
479 time by comparing estimated service time (based on historical
484 config DM_MULTIPATH_IOA
485 tristate "I/O Path Selector based on CPU submission"
486 depends on DM_MULTIPATH
488 This path selector selects the path based on the CPU the IO is
489 executed on and the CPU to path mapping setup at path addition time.
494 tristate "I/O delaying target"
495 depends on BLK_DEV_DM
497 A target that delays reads and/or writes and can send
498 them to different devices. Useful for testing.
503 tristate "Bad sector simulation target"
504 depends on BLK_DEV_DM
506 A target that simulates bad sector behavior.
512 bool "DM \"dm-mod.create=\" parameter support"
513 depends on BLK_DEV_DM=y
515 Enable "dm-mod.create=" parameter to create mapped devices at init time.
516 This option is useful to allow mounting rootfs without requiring an
518 See Documentation/admin-guide/device-mapper/dm-init.rst for dm-mod.create="..."
525 depends on BLK_DEV_DM
527 Generate udev events for DM events.
530 tristate "Flakey target"
531 depends on BLK_DEV_DM
533 A target that intermittently fails I/O for debugging purposes.
536 tristate "Verity target support"
537 depends on BLK_DEV_DM
542 This device-mapper target creates a read-only device that
543 transparently validates the data on one underlying device against
544 a pre-generated tree of cryptographic checksums stored on a second
547 You'll need to activate the digests you're going to use in the
548 cryptoapi configuration.
550 To compile this code as a module, choose M here: the module will
555 config DM_VERITY_VERIFY_ROOTHASH_SIG
557 bool "Verity data device root hash signature verification support"
559 select SYSTEM_DATA_VERIFICATION
561 Add ability for dm-verity device to be validated if the
562 pre-generated tree of cryptographic checksums passed has a pkcs#7
563 signature file that can validate the roothash of the tree.
565 By default, rely on the builtin trusted keyring.
569 config DM_VERITY_VERIFY_ROOTHASH_SIG_SECONDARY_KEYRING
570 bool "Verity data device root hash signature verification with secondary keyring"
571 depends on DM_VERITY_VERIFY_ROOTHASH_SIG
572 depends on SECONDARY_TRUSTED_KEYRING
574 Rely on the secondary trusted keyring to verify dm-verity signatures.
579 bool "Verity forward error correction support"
582 select REED_SOLOMON_DEC8
584 Add forward error correction support to dm-verity. This option
585 makes it possible to use pre-generated error correction data to
586 recover from corrupted blocks.
591 tristate "Switch target support (EXPERIMENTAL)"
592 depends on BLK_DEV_DM
594 This device-mapper target creates a device that supports an arbitrary
595 mapping of fixed-size regions of I/O across a fixed set of paths.
596 The path used for any specific region can be switched dynamically
597 by sending the target a message.
599 To compile this code as a module, choose M here: the module will
605 tristate "Log writes target support"
606 depends on BLK_DEV_DM
608 This device-mapper target takes two devices, one device to use
609 normally, one to log all write operations done to the first device.
610 This is for use by file system developers wishing to verify that
611 their fs is writing a consistent file system at all times by allowing
612 them to replay the log in a variety of ways and to check the
615 To compile this code as a module, choose M here: the module will
616 be called dm-log-writes.
621 tristate "Integrity target support"
622 depends on BLK_DEV_DM
623 select BLK_DEV_INTEGRITY
626 select CRYPTO_SKCIPHER
628 select DM_AUDIT if AUDIT
630 This device-mapper target emulates a block device that has
631 additional per-sector tags that can be used for storing
632 integrity information.
634 This integrity target is used with the dm-crypt target to
635 provide authenticated disk encryption or it can be used
638 To compile this code as a module, choose M here: the module will
639 be called dm-integrity.
642 tristate "Drive-managed zoned block device target support"
643 depends on BLK_DEV_DM
644 depends on BLK_DEV_ZONED
647 This device-mapper target takes a host-managed or host-aware zoned
648 block device and exposes most of its capacity as a regular block
649 device (drive-managed zoned block device) without any write
650 constraints. This is mainly intended for use with file systems that
651 do not natively support zoned block devices but still want to
652 benefit from the increased capacity offered by SMR disks. Other uses
653 by applications using raw block devices (for example object stores)
656 To compile this code as a module, choose M here: the module will
662 bool "DM audit events"
665 Generate audit events for device-mapper.
667 Enables audit logging of several security relevant events in the
668 particular device-mapper targets, especially the integrity target.