NAME | SYNOPSIS | DESCRIPTION | OPTIONS | Examples | SEE ALSO | COLOPHON

LVCREATE(8)                System Manager's Manual               LVCREATE(8)

NAME         top

       lvcreate - create a logical volume in an existing volume group

SYNOPSIS         top

       lvcreate [-a|--activate [a][e|l|s]{y|n}] [--addtag Tag] [--alloc
       AllocationPolicy] [-A|--autobackup {y|n}] [-H|--cache] [--cachemode
       {passthrough|writeback|writethrough}] [--cachepolicy Policy]
       [--cachepool CachePoolLogicalVolume] [--cachesettings Key=Value]
       [-c|--chunksize ChunkSize] [--commandprofile ProfileName]
       [-C|--contiguous {y|n}] [-d|--debug] [--discards
       {ignore|nopassdown|passdown}] [--errorwhenfull {y|n}] [{-l|--extents
       LogicalExtentsNumber[%{FREE|PVS|VG}] | -L|--size LogicalVolumeSize}
       [-i|--stripes Stripes [-I|--stripesize StripeSize]]] [-h|-?|--help]
       [-K|--ignoreactivationskip] [--ignoremonitoring] [--minor Minor
       [-j|--major Major]] [--metadataprofile ProfileName] [-m|--mirrors
       Mirrors [--corelog|--mirrorlog {disk|core|mirrored}] [--nosync]
       [-R|--regionsize MirrorLogRegionSize]] [--monitor {y|n}] [-n|--name
       LogicalVolume] [--noudevsync] [-p|--permission {r|rw}]
       [-M|--persistent {y|n}] [--poolmetadatasize MetadataVolumeSize]
       [--poolmetadataspare {y|n}] [--[raid]maxrecoveryrate Rate]
       [--[raid]minrecoveryrate Rate] [-r|--readahead
       {ReadAheadSectors|auto|none}] [--reportformat {basic|json}]
       [-k|--setactivationskip {y|n}] [-s|--snapshot] [-V|--virtualsize
       VirtualSize] [-t|--test] [-T|--thin] [--thinpool
       ThinPoolLogicalVolume] [--type SegmentType] [-v|--verbose]
       [-W|--wipesignatures {y|n}] [-Z|--zero {y|n}] [VolumeGroup |
       {ExternalOrigin|Origin|Pool}LogicalVolume
       [PhysicalVolumePath[:PE[-PE]]...]]

       lvcreate [-l|--extents LogicalExtentsNumber[%{FREE|ORIGIN|PVS|VG}] |
       -L|--size LogicalVolumeSize] [-c|--chunksize ChunkSize]
       [--commandprofile ProfileName] [--noudevsync] [--ignoremonitoring]
       [--metadataprofile ProfileName] [--monitor {y|n}] [-n|--name
       SnapshotLogicalVolumeName] [--reportformat {basic|json}]
       -s|--snapshot|-H|--cache {[VolumeGroup/]OriginalLogicalVolume
       [-V|--virtualsize VirtualSize]}

DESCRIPTION         top

       lvcreate creates a new logical volume in a volume group (see
       vgcreate(8), vgchange(8)) by allocating logical extents from the free
       physical extent pool of that volume group.  If there are not enough
       free physical extents then the volume group can be extended (see
       vgextend(8)) with other physical volumes or by reducing existing
       logical volumes of this volume group in size (see lvreduce(8)).  If
       you specify one or more PhysicalVolumes, allocation of physical
       extents will be restricted to these volumes.
       The second form supports the creation of snapshot logical volumes
       which keep the contents of the original logical volume for backup
       purposes.

OPTIONS         top

       See lvm(8) for common options.

       -a|--activate [a][l|e|s]{y|n}
              Controls the availability of the Logical Volumes for immediate
              use after the command finishes running.  By default, new
              Logical Volumes are activated (-ay).  If it is possible
              technically, -an will leave the new Logical Volume inactive.
              But for example, snapshots of active origin can only be
              created in the active state so -an cannot be used with --type
              snapshot. This does not apply to thin volume snapshots, which
              are by default created with flag to skip their activation
              (-ky).  Normally the --zero n argument has to be supplied too
              because zeroing (the default behaviour) also requires
              activation.  If autoactivation option is used (-aay), the
              logical volume is activated only if it matches an item in the
              activation/auto_activation_volume_list set in lvm.conf(5).
              For autoactivated logical volumes, --zero n and
              --wipesignatures n is always assumed and it can't be
              overridden. If the clustered locking is enabled, -aey will
              activate exclusively on one node and -a{a|l}y will activate
              only on the local node.

       -H|--cache
              Creates cache or cache pool logical volume.  Specifying the
              optional argument --extents or --size will cause the creation
              of the cache logical volume.  When the Volume group name is
              specified together with existing logical volume name which is
              NOT a cache pool name, such volume is treated as cache origin
              volume and cache pool is created. In this case the --extents
              or --size is used to specify size of cache pool volume.  See
              lvmcache(7) for more info about caching support.  Note that
              the cache segment type requires a dm-cache kernel module
              version 1.3.0 or greater.

       --cachemode {passthrough|writeback|writethrough}
              Specifying a cache mode determines when the writes to a cache
              LV are considered complete.  When writeback is specified, a
              write is considered complete as soon as it is stored in the
              cache pool LV.  If writethough is specified, a write is
              considered complete only when it has been stored in the cache
              pool LV and on the origin LV.  While writethrough may be
              slower for writes, it is more resilient if something should
              happen to a device associated with the cache pool LV. With
              passthrough mode, all reads are served from origin LV (all
              reads miss the cache) and all writes are forwarded to the
              origin LV; additionally, write hits cause cache block
              invalidates. See lvmcache(7) for more details.

       --cachepolicy Policy
              Only applicable to cached LVs; see also lvmcache(7). Sets the
              cache policy. mq is the basic policy name. smq is more
              advanced version available in newer kernels.

       --cachepool CachePoolLogicalVolume{Name|Path}
              Specifies the name of cache pool volume name. The other way to
              specify pool name is to append name to Volume group name
              argument.

       --cachesettings Key=Value
              Only applicable to cached LVs; see also lvmcache(7). Sets the
              cache tunable settings. In most use-cases, default values
              should be adequate.  Special string value default switches
              setting back to its default kernel value and removes it from
              the list of settings stored in lvm2 metadata.

       -c|--chunksize ChunkSize[b|B|s|S|k|K|m|M|g|G]
              Gives the size of chunk for snapshot, cache pool and thin pool
              logical volumes.  Default unit is in kilobytes.
              For snapshots the value must be power of 2 between 4KiB and
              512KiB and the default value is 4KiB.
              For cache pools the value must a multiple of 32KiB between
              32KiB and 1GiB. The default is 64KiB.  When the size is
              specified with volume caching, it may not be smaller than
              cache pool creation chunk size was.
              For thin pools the value must be a multiple of 64KiB between
              64KiB and 1GiB.  Default value starts with 64KiB and grows up
              to fit the pool metadata size within 128MiB, if the pool
              metadata size is not specified.  See lvm.conf(5) setting
              allocation/thin_pool_chunk_size_policy to select different
              calculation policy.  Thin pool target version <1.4 requires
              this value to be a power of 2.  For target version <1.5
              discard is not supported for non power of 2 values.

       -C|--contiguous {y|n}
              Sets or resets the contiguous allocation policy for logical
              volumes. Default is no contiguous allocation based on a next
              free principle.

       --corelog
              This is shortcut for option --mirrorlog core.

       --discards {ignore|nopassdown|passdown}
              Sets discards behavior for thin pool.  Default is passdown.

       --errorwhenfull {y|n}
              Configures thin pool behaviour when data space is exhausted.
              Default is no.  Device will queue I/O operations until target
              timeout (see dm-thin-pool kernel module option
              no_space_timeout) expires. Thus configured system has a time
              to i.e. extend the size of thin pool data device.  When set to
              yes, the I/O operation is immeditelly errored.

       -K|--ignoreactivationskip
              Ignore the flag to skip Logical Volumes during activation.
              Use --setactivationskip option to set or reset activation
              skipping flag persistently for logical volume.

       --ignoremonitoring
              Make no attempt to interact with dmeventd unless --monitor is
              specified.

       -l|--extents LogicalExtentsNumber[%{VG|PVS|FREE|ORIGIN}]
              Specifies the size of the new LV in logical extents.  The
              number of physical extents allocated may be different, and
              depends on the LV type.  Certain LV types require more
              physical extents for data redundancy or metadata.  An
              alternate syntax allows the size to be determined indirectly
              as a percentage of the size of a related VG, LV, or set of
              PVs.  The suffix %VG denotes the total size of the VG, the
              suffix %FREE the remaining free space in the VG, and the
              suffix %PVS the free space in the specified Physical Volumes.
              For a snapshot, the size can be expressed as a percentage of
              the total size of the Origin Logical Volume with the suffix
              %ORIGIN (100%ORIGIN provides space for the whole origin).
              When expressed as a percentage, the size defines an upper
              limit for the number of logical extents in the new LV. The
              precise number of logical extents in the new LV is not
              determined until the command has completed.

       -j|--major Major
              Sets the major number.  Major numbers are not supported with
              pool volumes.  This option is supported only on older systems
              (kernel version 2.4) and is ignored on modern Linux systems
              where major numbers are dynamically assigned.

       --metadataprofile ProfileName
              Uses and attaches the ProfileName configuration profile to the
              logical volume metadata. Whenever the logical volume is
              processed next time, the profile is automatically applied. If
              the volume group has another profile attached, the logical
              volume profile is preferred.  See lvm.conf(5) for more
              information about metadata profiles.

       --minor Minor
              Sets the minor number.  Minor numbers are not supported with
              pool volumes.

       -m|--mirrors mirrors
              Creates a mirrored logical volume with mirrors copies.  For
              example, specifying -m 1 would result in a mirror with two-
              sides; that is, a linear volume plus one copy.

              Specifying the optional argument --nosync will cause the
              creation of the mirror LV to skip the initial
              resynchronization.  Any data written afterwards will be
              mirrored, but the original contents will not be copied.

              This is useful for skipping a potentially long and resource
              intensive initial sync of an empty mirrored RaidLV.

              There are two implementations of mirroring which can be used
              and correspond to the "raid1" and "mirror" segment types.  The
              default is "raid1".  See the --type option for more
              information if you would like to use the legacy "mirror"
              segment type.  See lvm.conf(5) settings
              global/mirror_segtype_default and
              global/raid10_segtype_default to configure default mirror
              segment type.  The options --mirrorlog and --corelog apply to
              the legacy "mirror" segment type only.

              Note the current maxima for mirrors are 7 for "mirror"
              providing 8 mirror legs and 9 for "raid1" providing 10 legs.

       --mirrorlog {disk|core|mirrored}
              Specifies the type of log to be used for logical volumes
              utilizing the legacy "mirror" segment type.
              The default is disk, which is persistent and requires a small
              amount of storage space, usually on a separate device from the
              data being mirrored.
              Using core means the mirror is regenerated by copying the data
              from the first device each time the logical volume is
              activated, like after every reboot.
              Using mirrored will create a persistent log that is itself
              mirrored.

       --monitor {y|n}
              Starts or avoids monitoring a mirrored, snapshot or thin pool
              logical volume with dmeventd, if it is installed.  If a device
              used by a monitored mirror reports an I/O error, the failure
              is handled according to activation/mirror_image_fault_policy
              and activation/mirror_log_fault_policy set in lvm.conf(5).

       -n|--name LogicalVolume{Name|Path}
              Sets the name for the new logical volume.
              Without this option a default name of "lvol#" will be
              generated where # is the LVM internal number of the logical
              volume.

       --nosync
              Causes the creation of mirror, raid1, raid4, raid5 and raid10
              to skip the initial resynchronization.  In case of mirror,
              raid1 and raid10, any data written afterwards will be
              mirrored, but the original contents will not be copied.  In
              case of raid4 and raid5, no parity blocks will be written,
              though any data written afterwards will cause parity blocks to
              be stored.
              This is useful for skipping a potentially long and resource
              intensive initial sync of an empty mirror/raid1/raid4/raid5
              and raid10 LV.
              This option is not valid for raid6, because raid6 relies on
              proper parity (P and Q Syndromes) being created during initial
              synchronization in order to reconstruct proper user date in
              case of device failures.

              raid0 and raid0_meta don't provide any data copies or parity
              support and thus don't support initial resynchronization.

       --noudevsync
              Disables udev synchronisation. The process will not wait for
              notification from udev.  It will continue irrespective of any
              possible udev processing in the background.  You should only
              use this if udev is not running or has rules that ignore the
              devices LVM2 creates.

       -p|--permission {r|rw}
              Sets access permissions to read only (r) or read and write
              (rw).
              Default is read and write.

       -M|--persistent {y|n}
              Set to y to make the minor number specified persistent.  Pool
              volumes cannot have persistent major and minor numbers.
              Defaults to yes only when major or minor number is specified.
              Otherwise it is no.

       --poolmetadatasize MetadataVolumeSize[b|B|s|S|k|K|m|M|g|G]
              Sets the size of pool's metadata logical volume.  Supported
              values are in range between 2MiB and 16GiB for thin pool, and
              upto 16GiB for cache pool. The minimum value is computed from
              pool's data size.  Default value for thin pool is
              (Pool_LV_size / Pool_LV_chunk_size * 64b).  To work with a
              thin pool, there should be at least 25% of free space when the
              size of metadata is smaller then 16MiB, or at least 4MiB of
              free space otherwise.  Default unit is megabytes.

       --poolmetadataspare {y|n}
              Controls creation and maintanence of pool metadata spare
              logical volume that will be used for automated pool recovery.
              Only one such volume is maintained within a volume group with
              the size of the biggest pool metadata volume.  Default is yes.

       --[raid]maxrecoveryrate Rate[b|B|s|S|k|K|m|M|g|G]
              Sets the maximum recovery rate for a RAID logical volume.
              Rate is specified as an amount per second for each device in
              the array.  If no suffix is given, then KiB/sec/device is
              assumed.  Setting the recovery rate to 0 means it will be
              unbounded.

       --[raid]minrecoveryrate Rate[b|B|s|S|k|K|m|M|g|G]
              Sets the minimum recovery rate for a RAID logical volume.
              Rate is specified as an amount per second for each device in
              the array.  If no suffix is given, then KiB/sec/device is
              assumed.  Setting the recovery rate to 0 means it will be
              unbounded.

       -r|--readahead {ReadAheadSectors|auto|none}
              Sets read ahead sector count of this logical volume.  For
              volume groups with metadata in lvm1 format, this must be a
              value between 2 and 120.  The default value is auto which
              allows the kernel to choose a suitable value automatically.
              none is equivalent to specifying zero.

       -R|--regionsize MirrorLogRegionSize[b|B|s|S|k|K|m|M|g|G]
              A mirror is divided into regions of this size (in MiB), and
              the mirror log uses this granularity to track which regions
              are in sync.

       -k|--setactivationskip {y|n}
              Controls whether Logical Volumes are persistently flagged to
              be skipped during activation. By default, thin snapshot
              volumes are flagged for activation skip.  See lvm.conf(5)
              activation/auto_set_activation_skip how to change its default
              behaviour.  To activate such volumes, an extra
              --ignoreactivationskip option must be used. The flag is not
              applied during deactivation. Use lvchange --setactivationskip
              command to change the skip flag for existing volumes.  To see
              whether the flag is attached, use lvs command where the state
              of the flag is reported within lv_attr bits.

       -L|--size LogicalVolumeSize[b|B|s|S|k|K|m|M|g|G|t|T|p|P|e|E]
              Gives the size to allocate for the new logical volume.  A size
              suffix of B for bytes, S for sectors as 512 bytes, K for
              kilobytes, M for megabytes, G for gigabytes, T for terabytes,
              P for petabytes or E for exabytes is optional.
              Default unit is megabytes.

       -s|--snapshot OriginalLogicalVolume{Name|Path}
              Creates a snapshot logical volume (or snapshot) for an
              existing, so called original logical volume (or origin).
              Snapshots provide a 'frozen image' of the contents of the
              origin while the origin can still be updated. They enable
              consistent backups and online recovery of removed/overwritten
              data/files.
              Thin snapshot is created when the origin is a thin volume and
              the size IS NOT specified. Thin snapshot shares same blocks
              within the thin pool volume.  The non thin volume snapshot
              with the specified size does not need the same amount of
              storage the origin has. In a typical scenario, 15-20% might be
              enough. In case the snapshot runs out of storage, use
              lvextend(8) to grow it. Shrinking a snapshot is supported by
              lvreduce(8) as well. Run lvs(8) on the snapshot in order to
              check how much data is allocated to it.  Note: a small amount
              of the space you allocate to the snapshot is used to track the
              locations of the chunks of data, so you should allocate
              slightly more space than you actually need and monitor
              (--monitor) the rate at which the snapshot data is growing so
              you can avoid running out of space.  If --thinpool is
              specified, thin volume is created that will use given original
              logical volume as an external origin that serves unprovisioned
              blocks.  Only read-only volumes can be used as external
              origins.  To make the volume external origin, lvm expects the
              volume to be inactive.  External origin volume can be
              used/shared for many thin volumes even from different thin
              pools. See lvconvert(8) for online conversion to thin volumes
              with external origin.

       -i|--stripes Stripes
              Gives the number of stripes.  This is equal to the number of
              physical volumes to scatter the logical volume data.  When
              creating a RAID 4/5/6 logical volume, the extra devices which
              are necessary for parity are internally accounted for.
              Specifying -i 3 would cause 3 devices for striped and RAID 0
              logical volumes, 4 devices for RAID 4/5, 5 devices for RAID 6
              and 6 devices for RAID 10.  Alternatively, RAID 0 will stripe
              across 2 devices, RAID 4/5 across 3 PVs, RAID 6 across 5 PVs
              and RAID 10 across 4 PVs in the volume group if the -i
              argument is omitted.  In order to stripe across all PVs of the
              VG if the -i argument is omitted, set
              raid_stripe_all_devices=1 in the allocation section of
              lvm.conf (5) or add
              --config allocation/raid_stripe_all_devices=1
              to the command.

              Note the current maxima for stripes depend on the created RAID
              type.  For raid10, the maximum of stripes is 32, for raid0, it
              is 64, for raid4/5, it is 63 and for raid6 it is 62.

              See the --nosync option to optionally avoid initial
              syncrhonization of RaidLVs.

              Two implementations of basic striping are available in the
              kernel.  The original device-mapper implementation is the
              default and should normally be used.  The alternative
              implementation using MD, available since version 1.7 of the
              RAID device-mapper kernel target (kernel version 4.2) is
              provided to facilitate the development of new RAID features.
              It may be accessed with --type raid0[_meta], but is best
              avoided at present because of assorted restrictions on
              resizing and converting such devices.

       -I|--stripesize StripeSize
              Gives the number of kilobytes for the granularity of the
              stripes.
              StripeSize must be 2^n (n = 2 to 9) for metadata in LVM1
              format.  For metadata in LVM2 format, the stripe size may be a
              larger power of 2 but must not exceed the physical extent
              size.

       -T|--thin
              Creates thin pool or thin logical volume or both.  Specifying
              the optional argument --size or --extents will cause the
              creation of the thin pool logical volume.  Specifying the
              optional argument --virtualsize will cause the creation of the
              thin logical volume from given thin pool volume.  Specifying
              both arguments will cause the creation of both thin pool and
              thin volume using this pool.  See lvmthin(7) for more info
              about thin provisioning support.  Thin provisioning requires
              device mapper kernel driver from kernel 3.2 or greater.

       --thinpool ThinPoolLogicalVolume{Name|Path}
              Specifies the name of thin pool volume name. The other way to
              specify pool name is to append name to Volume group name
              argument.

       --type SegmentType
              Creates a logical volume with the specified segment type.
              Supported types are: cache, cache-pool, error, linear, mirror,
              raid0, raid1, raid4, raid5_la, raid5_ls (= raid5), raid5_ra,
              raid5_rs, raid6_nc, raid6_nr, raid6_zr (= raid6), raid10,
              snapshot, striped, thin, thin-pool or zero.  Segment type may
              have a commandline switch alias that will enable its use.
              When the type is not explicitly specified an implicit type is
              selected from combination of options: -H|--cache|--cachepool
              (cache or cachepool), -T|--thin|--thinpool (thin or thinpool),
              -m|--mirrors (raid1 or mirror), -s|--snapshot|-V|--virtualsize
              (snapshot or thin), -i|--stripes (striped).  The default
              segment type is linear.

       -V|--virtualsize VirtualSize[b|B|s|S|k|K|m|M|g|G|t|T|p|P|e|E]
              Creates a thinly provisioned device or a sparse device of the
              given size (in MiB by default).  See lvm.conf(5) settings
              global/sparse_segtype_default to configure default sparse
              segment type.  See lvmthin(7) for more info about thin
              provisioning support.  Anything written to a sparse snapshot
              will be returned when reading from it.  Reading from other
              areas of the device will return blocks of zeros.  Virtual
              snapshot (sparse snapshot) is implemented by creating a hidden
              virtual device of the requested size using the zero target.  A
              suffix of _vorigin is used for this device.  Note: using
              sparse snapshots is not efficient for larger device sizes
              (GiB), thin provisioning should be used for this case.

       -W|--wipesignatures {y|n}
              Controls detection and subsequent wiping of signatures on
              newly created Logical Volume. There's a prompt for each
              signature detected to confirm its wiping (unless --yes is used
              where LVM assumes 'yes' answer for each prompt automatically).
              If this option is not specified, then by default -W |
              --wipesignatures y is assumed each time the zeroing is done
              (-Z | --zero y). This default behaviour can be controlled by
              allocation/wipe_signatures_when_zeroing_new_lvs setting found
              in lvm.conf(5).
              If blkid wiping is used (allocation/use_blkid_wiping setting
              in lvm.conf(5)) and LVM2 is compiled with blkid wiping
              support, then blkid(8) library is used to detect the
              signatures (use blkid -k command to list the signatures that
              are recognized).  Otherwise, native LVM2 code is used to
              detect signatures (MD RAID, swap and LUKS signatures are
              detected only in this case).
              Logical volume is not wiped if the read only flag is set.

       -Z|--zero {y|n}
              Controls zeroing of the first 4KiB of data in the new logical
              volume.  Default is yes.  Snapshot COW volumes are always
              zeroed.  Logical volume is not zeroed if the read only flag is
              set.
              Warning: trying to mount an unzeroed logical volume can cause
              the system to hang.

Examples         top

       Creates a striped logical volume with 3 stripes, a stripe size of
       8KiB and a size of 100MiB in the volume group named vg00.  The
       logical volume name will be chosen by lvcreate:

       lvcreate -i 3 -I 8 -L 100M vg00

       Creates a mirror logical volume with 2 sides with a useable size of
       500 MiB.  This operation would require 3 devices (or option --alloc
       anywhere) - two for the mirror devices and one for the disk log:

       lvcreate -m1 -L 500M vg00

       Creates a mirror logical volume with 2 sides with a useable size of
       500 MiB.  This operation would require 2 devices - the log is "in-
       memory":

       lvcreate -m1 --mirrorlog core -L 500M vg00

       Creates a snapshot logical volume named "vg00/snap" which has access
       to the contents of the original logical volume named "vg00/lvol1" at
       snapshot logical volume creation time. If the original logical volume
       contains a file system, you can mount the snapshot logical volume on
       an arbitrary directory in order to access the contents of the
       filesystem to run a backup while the original filesystem continues to
       get updated:

       lvcreate --size 100m --snapshot --name snap /dev/vg00/lvol1

       Creates a snapshot logical volume named "vg00/snap" with size for
       overwriting 20% of the original logical volume named "vg00/lvol1".:

       lvcreate -s -l 20%ORIGIN --name snap vg00/lvol1

       Creates a sparse device named /dev/vg1/sparse of size 1TiB with space
       for just under 100MiB of actual data on it:

       lvcreate --virtualsize 1T --size 100M --snapshot --name sparse vg1

       Creates a linear logical volume "vg00/lvol1" using physical extents
       /dev/sda:0-7 and /dev/sdb:0-7 for allocation of extents:

       lvcreate -L 64M -n lvol1 vg00 /dev/sda:0-7 /dev/sdb:0-7

       Creates a 5GiB RAID5 logical volume "vg00/my_lv", with 3 stripes
       (plus a parity drive for a total of 4 devices) and a stripesize of
       64KiB:

       lvcreate --type raid5 -L 5G -i 3 -I 64 -n my_lv vg00

       Creates a RAID5 logical volume "vg00/my_lv", using all of the free
       space in the VG and spanning all the PVs in the VG (note that the
       command will fail if there's more than 8 PVs in the VG in which case
       -i 7 has to be used to get to the currently possible maximum of 8
       devices including parity for RaidLVs):

       lvcreate --config allocation/raid_stripe_all_devices=1 --type raid5
       -l 100%FREE -n my_lv vg00

       Creates a 5GiB RAID10 logical volume "vg00/my_lv", with 2 stripes on
       2 2-way mirrors.  Note that the -i and -m arguments behave
       differently.  The -i specifies the number of stripes.  The -m
       specifies the number of additional copies:

       lvcreate --type raid10 -L 5G -i 2 -m 1 -n my_lv vg00

       Creates 100MiB pool logical volume for thin provisioning build with 2
       stripes 64KiB and chunk size 256KiB together with 1TiB thin
       provisioned logical volume "vg00/thin_lv":

       lvcreate -i 2 -I 64 -c 256 -L100M -T vg00/pool -V 1T --name thin_lv

       Creates a thin snapshot volume "thinsnap" of thin volume "thinvol"
       that will share the same blocks within the thin pool.  Note: the size
       MUST NOT be specified, otherwise the non-thin snapshot is created
       instead:

       lvcreate -s vg00/thinvol --name thinsnap

       Creates a thin snapshot volume of read-only inactive volume "origin"
       which then becomes the thin external origin for the thin snapshot
       volume in vg00 that will use an existing thin pool "vg00/pool":

       lvcreate -s --thinpool vg00/pool origin

       Create a cache pool LV that can later be used to cache one logical
       volume.

       lvcreate --type cache-pool -L 1G -n my_lv_cachepool vg /dev/fast1

       If there is an existing cache pool LV, create the large slow device
       (i.e. the origin LV) and link it to the supplied cache pool LV,
       creating a cache LV.

       lvcreate --cache -L 100G -n my_lv vg/my_lv_cachepool /dev/slow1

       If there is an existing logical volume, create the small and fast
       cache pool LV and link it to the supplied existing logical volume
       (i.e. the origin LV), creating a cache LV.

       lvcreate --type cache -L 1G -n my_lv_cachepool vg/my_lv /dev/fast1

SEE ALSO         top

       lvm(8), lvm.conf(5), lvmcache(7), lvmthin(7), lvconvert(8),
       lvchange(8), lvextend(8), lvreduce(8), lvremove(8), lvrename(8)
       lvs(8), lvscan(8), vgcreate(8), blkid(8)

COLOPHON         top

       This page is part of the lvm2 (Logical Volume Manager 2) project.
       Information about the project can be found at 
       ⟨http://www.sourceware.org/lvm2/⟩.  If you have a bug report for this
       manual page, send it to linux-lvm@redhat.com.  This page was obtained
       from the project's upstream Git repository 
       ⟨https://git.fedorahosted.org/git/lvm2.git⟩ on 2017-03-13.  If you
       discover any rendering problems in this HTML version of the page, or
       you believe there is a better or more up-to-date source for the page,
       or you have corrections or improvements to the information in this
       COLOPHON (which is not part of the original manual page), send a mail
       to man-pages@man7.org

Sistina Software UKLVM TOOLS 2.02.169(2)-git (2016-11-30)        LVCREATE(8)