Original author(s) Operating system | Written in C License GNU GPL | |
![]() | ||
Stable release 2.02.160 / 6 July 2016; 7 months ago (2016-07-06) Website sources.redhat.com/lvm2/ |
In Linux, Logical Volume Manager (LVM) is a device mapper target that provides logical volume management for the Linux kernel. Most modern Linux distributions are LVM-aware to the point of being able to have their root file systems on a logical volume.
Contents
- Common uses
- Basic functionality
- Advanced functionality
- RAID
- High availability
- Volume group allocation policy
- Implementation
- Caveats
- References
Heinz Mauelshagen wrote the original LVM code in 1998, taking its primary design guidelines from the HP-UX's volume manager.
Common uses
LVM is used for the following purposes:
LVM can be considered as a thin software layer on top of the hard disks and partitions, which creates an abstraction of continuity and ease-of-use for managing hard drive replacement, repartitioning and backup.
The Ganeti solution stack relies on the Linux Logical Volume Manager.
Basic functionality
lvmetad
daemon.Advanced functionality
devices/multipath_component_detection=1
is set in lvm.conf
. This prevents LVM from activating volumes on an individual path instead of the multipath device.RAID
lvchange --raidmaxrecoveryrate
and lvchange --raidminrecoveryrate
to maintain acceptable I/O performance while rebuilding a LV that includes RAID functionality.High availability
The LVM also works in a shared-storage cluster in which disks holding the PVs are shared between multiple host computers, but can require an additional daemon to mediate metadata access via a form of locking.
clvmd
, which is in constant contact with other clvmd
daemons in the cluster and can communicate a desire to get a lock on a particular set of objects.clvmd
by making the locking of LVM objects transparent to the rest of LVM, without relying on a distributed lock manager.The above described mechanisms only resolve the issues with LVM's access to the storage. The file system selected to be on top of such LVs must either support clustering by itself (such as GFS2 or VxFS) or it must only be mounted by a single cluster node at any time (such as in an active-passive configuration).
Volume group allocation policy
LVM VGs must contain a default allocation policy for new volumes created from it. This can later be changed for each LV using the lvconvert -A
command, or on the VG itself via vgchange --alloc
. To minimize fragmentation, LVM will attempt the strictest policy (contiguous) first and then progress toward the most liberal policy defined for the LVM object until allocation finally succeeds.
In RAID configurations, almost all policies are applied to each leg in isolation. For example, even if a LV has a policy of cling, expanding the file system will not result in LVM using a PV if it is already used by one of the other legs in the RAID setup. LVs with RAID functionality will put each leg on different PVs, making the other PVs unavailable to any other given leg. If this was the only option available, expansion of the LV would fail. In this sense, the logic behind cling will only apply to expanding each of the individual legs of the array.
Available allocation policies are:
Implementation
Typically, the first megabyte of each physical volume contains a mostly ASCII-encoded structure referred to as an "LVM header" or "LVM head". Originally, the LVM head used to be written in the first and last megabyte of each PV for redundancy (in case of a partial hardware failure); however, this was later changed to only the first megabyte. Each PV's header is a complete copy of the entire volume group's layout, including the UUIDs of all other PVs and of LVs, and allocation map of PEs to LEs. This simplifies data recovery if a PV is lost.
In the 2.6-series of the Linux Kernel, the LVM is implemented in terms of the device mapper, a simple block-level scheme for creating virtual block devices and mapping their contents onto other block devices. This minimizes the amount of relatively hard-to-debug kernel code needed to implement the LVM. It also allows its I/O redirection services to be shared with other volume managers (such as EVMS). Any LVM-specific code is pushed out into its user-space tools, which merely manipulate these mappings and reconstruct their state from on-disk metadata upon each invocation.
To bring a volume group online, the "vgchange" tool:
- Searches for PVs in all available block devices.
- Parses the metadata header in each PV found.
- Computes the layouts of all visible volume groups.
- Loops over each logical volume in the volume group to be brought online and:
- Checks if the logical volume to be brought online has all its PVs visible.
- Creates a new, empty device mapping.
- Maps it (with the "linear" target) onto the data areas of the PVs the logical volume belongs to.
To move an online logical volume between PVs on the same Volume Group, use the "pvmove" tool:
- Creates a new, empty device mapping for the destination.
- Applies the "mirror" target to the original and destination maps. The kernel will start the mirror in "degraded" mode and begin copying data from the original to the destination to bring it into sync.
- Replaces the original mapping with the destination when the mirror comes into sync, then destroys the original.
These device mapper operations take place transparently, without applications or file systems being aware that their underlying storage is moving.
Caveats
pvmove
command.vgconvert --pvmetadatacopies
. If the LVM can not read a proper header using the first copy, it will check the end of the volume for a backup header. Most Linux distributions keep a running backup in /etc/lvm/backup
, which enables manual rewriting of a corrupted LVM head using the vgcfgrestore
command.