LVM に Arch Linux をインストールする

提供: ArchWiki
2022年9月1日 (木) 13:16時点におけるAshMyzk (トーク | 投稿記録)による版 (ページの作成:「Category:Arch の入手とインストール en:Install Arch Linux on LVM pt:Install Arch Linux on LVM {{Related articles start}} {{Related|LVM}} {{Related|…」)
(差分) ← 古い版 | 最新版 (差分) | 新しい版 → (差分)
ナビゲーションに移動 検索に移動

関連記事

LVM ボリュームは、インストール手順パーティショニングフォーマットの間で作成する必要があります。パーティションを直接フォーマットしてルートファイルシステムにするのではなく、ファイルシステムを論理ボリューム (LV) 内で作成します。

概要:

  • Install the required packages. (refer to LVM#Getting started)
  • Create partition(s) where your physical volumes (PVs) will reside.
  • Create your PVs. If you have one disk it is best to just create one PV in one large partition. If you have multiple disks you can create partitions on each of them and create a PV on each partition.
  • Create your volume group (VG) and add all PVs to it.
  • Create logical volumes (LVs) inside that VG.
  • Continue with Installation guide#Format the partitions.
  • When you reach the “Create initial ramdisk environment” step in the Installation guide, add the lvm2 hook to /etc/mkinitcpio.conf (see below for details).
警告: /boot cannot reside in LVM when using a boot loader which does not support LVM; you must create a separate /boot partition and format it directly. Only GRUB is known to support LVM.

インストール

You will follow along with the installation guide until you come to Installation guide#Partition the disks. At this point you will diverge and doing all your partitioning with LVM in mind.

パーティションを作成する

First, partition your disks as required before configuring LVM.

Create the partitions:

  • If you use Master Boot Record partition table, set the partition type ID to 8e (partition type Linux LVM in fdisk).
  • If you use GUID Partition Table, set the partition type GUID to E6D6D379-F507-44C2-A23C-238F2A3DF928 (partition type Linux LVM in fdisk and 8e00 in gdisk).

物理ボリュームを作成する

To list all your devices capable of being used as a physical volume:

# lvmdiskscan
警告: Make sure you target the correct device, or below commands will result in data loss!

Create a physical volume on them:

# pvcreate DEVICE

This command creates a header on each device so it can be used for LVM. As defined in LVM#LVM building blocks, DEVICE can be any block device, e.g. a disk /dev/sda, a partition /dev/sda2 or a loop back device. For example:

# pvcreate /dev/sda2

You can track created physical volumes with:

# pvdisplay

You can also get summary information on physical volumes with:

# pvscan
ヒント: If you run into trouble with a pre-existing disk signature, you can delete it using wipefs.
ノート: If using a SSD without partitioning it first, use pvcreate --dataalignment 1m /dev/sda (for erase block size < 1 MiB), see e.g. here

ボリュームグループを作成・拡張する

First you need to create a volume group on any one of the physical volumes:

# vgcreate <volume_group> <physical_volume>

For example:

# vgcreate VolGroup00 /dev/sda2

See lvm(8) for a list of valid characters for volume group names.

Extending the volume group is just as easy:

# vgextend <volume_group> <physical_volume>

For example, to add both sdb1 and sdc to your volume group:

# vgextend VolGroup00 /dev/sdb1
# vgextend VolGroup00 /dev/sdc

You can track how your volume group grows with:

# vgdisplay

This is also what you would do if you wanted to add a disk to a RAID or mirror group with failed disks.

ノート: You can create more than one volume group if you need to, but then you will not have all your storage presented as a single disk.

物理ボリュームとボリュームグループを同時に作成する

LVM allows you to combine the creation of a volume group and the physical volumes in one easy step. For example, to create the group VolGroup00 with the three devices mentioned above, you can run:

# vgcreate VolGroup00 /dev/sda2 /dev/sdb1 /dev/sdc

This command will first set up the three partitions as physical volumes (if necessary) and then create the volume group with the three volumes. The command will warn you if it detects an existing filesystem on any devices.

論理ボリュームを作成する

ヒント:
  • If you wish to use snapshots, logical volume caching, thin provisioned logical volumes or RAID see LVM#Logical volumes.
  • If a logical volume will be formatted with ext4, leave at least 256 MiB free space in the volume group to allow using e2scrub(8).

Now we need to create logical volumes on this volume group. You create a logical volume with the next command by specifying the new volume's name and size, and the volume group it will reside on:

# lvcreate -L <size> <volume_group> -n <logical_volume>

For example:

# lvcreate -L 10G VolGroup00 -n lvolhome

This will create a logical volume that you can access later with /dev/VolGroup00/lvolhome. Just like volume groups, you can use any name you want for your logical volume when creating it besides a few exceptions listed in lvm(8) § VALID_NAMES.

You can also specify one or more physical volumes to restrict where LVM allocates the data. For example, you may wish to create a logical volume for the root filesystem on your small SSD, and your home volume on a slower mechanical drive. Simply add the physical volume devices to the command line, for example:

# lvcreate -L 10G VolGroup00 -n lvolhome /dev/sdc1

To use all the free space left in a volume group, use the next command:

# lvcreate -l 100%FREE  <volume_group> -n <logical_volume>

You can track created logical volumes with:

# lvdisplay
ノート: You may need to load the device-mapper kernel module (modprobe dm_mod) for the above commands to succeed.
ヒント: You can start out with relatively small logical volumes and expand them later if needed. For simplicity, leave some free space in the volume group so there is room for expansion.

論理ボリュームをフォーマット・マウントする

Your logical volumes should now be located in /dev/YourVolumeGroupName/. If you cannot find them, use the next commands to bring up the module for creating device nodes and to make volume groups available:

# modprobe dm_mod
# vgscan
# vgchange -ay

Now you can format your logical volumes and mount them as normal partitions (see mount a file system for additional details):

# mkfs.<fstype> /dev/<volume_group>/<logical_volume>
# mount /dev/<volume_group>/<logical_volume> /<mountpoint>

For example:

# mkfs.ext4 /dev/VolGroup00/lvolhome
# mount /dev/VolGroup00/lvolhome /home
警告: When choosing mountpoints, just select your newly created logical volumes (use: /dev/Volgroup00/lvolhome). Do not select the actual partitions on which logical volumes were created (do not use: /dev/sda2).

システムを設定する

mkinitcpio フックを追加する

In case your root filesystem is on LVM, you will need to enable the appropriate mkinitcpio hooks, otherwise your system might not boot. Enable:

  • udev and lvm2 for the default busybox-based initramfs
  • systemd and lvm2 for systemd-based initramfs

udev is there by default. Edit the file and insert lvm2 between block and filesystems like so:

/etc/mkinitcpio.conf
HOOKS=(base udev ... block lvm2 filesystems)

For systemd based initramfs:

/etc/mkinitcpio.conf
HOOKS=(base systemd ... block lvm2 filesystems)

Afterwards, you can continue in normal installation instructions with the create an initial ramdisk step.

ヒント:
  • The lvm2 hook is installed by lvm2, not mkinitcpio. If you are running mkinitcpio in an arch-chroot for a new installation, lvm2 must be installed inside the arch-chroot for mkinitcpio to find the lvm2 hook. If lvm2 only exists outside the arch-chroot, mkinitcpio will output Error: Hook 'lvm2' cannot be found.
  • If your root filesystem is on LVM RAID see #Configure mkinitcpio for RAID.

mkinitcpio を RAID 用に設定する

If your root filesystem is on LVM RAID additionally to the lvm2 hook, you need to add dm-raid and the appropriate RAID modules (e.g. raid0, raid1, raid10 and/or raid456) to the MODULES array in mkinitcpio.conf. Also dm_integrity module is needed if you created RAID with integrity checksums (--raidintegrity option in lvcreate).

For busybox based initramfs:

/etc/mkinitcpio.conf
MODULES=(dm-raid dm_integrity raid0 raid1 raid10 raid456)
HOOKS=(base udev ... block lvm2 filesystems)

For systemd based initramfs:

/etc/mkinitcpio.conf
MODULES=(dm-raid dm_integrity raid0 raid1 raid10 raid456)
HOOKS=(base systemd ... block lvm2 filesystems)

For systems using LVM thin volumes the location of systemd between udev and block causes the malfunction of the thin volume. just with the parameters (base udev autodetect modconf block lvm2 filesystems keyboard fsck) works: root FS inside a lvm logical volume, and other LVM's (normal and thin) at other locations.

カーネルブートオプション

If the root file system resides in a logical volume, the root= kernel parameter must be pointed to the mapped device, e.g /dev/vg-name/lv-name.