Mondo Rescue LVM on RAID How-to Gerhard Gappmeier 2004-01-15 Abstract Mondo Rescue is great to save all your data and restore it later to a different disk geometry. But when changing to LVM on RAID you have to do more then Mondo can do. At first you have to change the boot loader, because GRUB which is normally used in actual SuSE distributions can’t boot RAID systems. When using LVM and RAID for your root file system you need the according modules already during the boot process so that the kernel is able to mount it. All this changes can lead to some troubles which costed me one and a half day to solve it. You will find this also interesting if you only want to use one of LVM or RAID. Contents 1 Prerequisites 2 2 Boot mindi 3 3 Making new partitions using parted 3 4 Starting RAID 4 5 Format the boot partition 5 6 Starting LVM 5 7 Create LVM physical volume 5 8 Create LVM volume group 5 9 Create LVM logical volumes 6 10 Create and initialize the swap space 6 11 Create file systems 7 1 12 Edit mountlist.txt 7 13 Restore all data to new partitions 7 14 Mounting the partitions 8 15 Edit fstab 8 16 Edit lilo.conf 8 17 Create initrd 9 18 Install LILO boot loader 11 19 Troubleshooting 11 1 Prerequisites To understand this document you should already be familiar with the fundamentals of Mondo Rescue, SOFT-RAID (software based Redundant Array of Independent Disks), LVM (Logical Volume Manager) and initrd (initial RAM disk). This document describes the procedure based on a normal SuSE 8.2 installation which uses a modular kernel 2.4.20 but it should also work for most other systems with a kernel that supports LVM. If you compile the drivers (reiserfs, lvm-mod, raid1, ...) into the kernel you don’t have to make an initrd as described later in chapter 17 on page 9. But if you compile your kernel yourself you should already know that ;-). For the following sections I assume that you have an installed system which you have backed up with mondo rescue via NFS (Network File System). The term “target system” will be used for your system you want to change and “backup system” for your system where you are storing the backups. You can also make a backup direct to CD, but then you have to adapt the following steps. I use NFS because I don’t have a CD writer in my server and because it’s very convenient to use NFS. Check your backup if you can restore it, because the next steps will destroy all data on your system. So if you encounter any problems you can at least restore your old system. In my example I’m using two IDE disks where each disk is master on one of the two channels of my IDE controller. So the device names are /dev/hda and /dev/hdc. Change this names to your needs. This two disks are used to make the RAID1 (mirroring) system. Burn the mindi.iso image to CD to get your system up again. With mindi you can run mondorestore and all other commands you need to configure the system. In my case pico (a unix text editor) which comes with mindi produced a segmentation fault and vi (another text editor) also didn’t work so I copied all files via NFS, that I wanted to edit and used KWrite (graphical editor in KDE) to comfortable edit it on another machine. 2 2 Boot mindi After booting from the mindi CD mondorestore gets automatically started and the first dialog appears. Select Interactive -¿ NFS then hit enter without entering a path. This will mount the NFS share you have used for the backup automatically on /tmp/isodir. Now you can cancel the dialog to return to the command line. Before you can start mondorestore again you have to change the partition layout and edit mondo’s mountlist.txt. 3 Making new partitions using parted For booting you need just a small partition with 25MB which uses also RAID. This will be /dev/md0 and gets mounted on /boot. Then you make a second RAID partition which uses the rest of the disk space. This will be /dev/md1 and will be the physical device for LVM. Start parted for the first disk: parted /dev/hda If you don’t know the parted commands use ? to list it. View your partitions using the command print Delete all existing partitions using rm MINOR Create the boot partition with 25 MB mkpart primary 0 25 Mark it as RAID partition (this sets the partition type FD) set 1 raid on Create the second partition with maximum size. You can view your disk geometry with print again. mkpart primary 25 YOUR DISK SIZE Mark it as RAID partition set 2 raid on Now use the same procedure for the second disk /dev/hdc. 3 4 Starting RAID Create the following raidtab and copy it to your NFS share so you don’t need to write it on your mindi system. Then copy it to /etc via NFS (cp /tmp/isodir/raidtab /etc). /etc is on your RAM disk at the moment which mindi has created for you so this will be lost if you are rebooting. So it is always a good idea to keep this and the following configuration files on your NFS. raidtab: # sample raiddev configuration file # # ’persistent’ RAID1 setup, with no spare disks: # raiddev /dev/md0 raid-level 1 nr-raid-disks 1 nr-spare-disks 0 persistent-superblock 1 chunk-size 4 device /dev/hda1 raid-disk 0 device /dev/hdc1 raid-disk 1 # # persistent RAID1 array with 0 spare disk. # raiddev /dev/md1 raid-level 1 nr-raid-disks 1 nr-spare-disks 0 persistent-superblock 1 chunk-size 4 device /dev/hda2 raid-disk 0 device /dev/hdc2 raid-disk 1 Load the RAID module modprobe raid1 Create the RAID partitions mkraid /dev/md0 mkraid /dev/md1 mkraid also starts the RAID system. If you want to start an existing RAID partition use raidstart /dev/md0. You can check the RAID status with cat /proc/mdstat. Now you are ready to format the boot partition and to set up LVM. 4 5 Format the boot partition For /boot we have already created 25MB RAID partition which is now running as /dev/md0. Till now this partition has no file system. For this little partition we create now an extended2 linux file system. mke2fs /dev/md0 That’s all, you could already mount /dev/md0 now and access it. 6 Starting LVM Before creating or accessing any physical volumes you have to start the LVM sub system. This is done by vgscan This command also recognizes existing volumes and is creating /etc/lvmtab and /etc/lvmtab.d . 7 Create LVM physical volume Now we create our LVM physical volume on /dev/md1. This physical volume will be used for our volume group. A volume group can consist of more physical devices, but we have only one in our case. pvcreate /dev/md1 You should see the following output: pvcreate -- physical volume "/dev/md1" successfully created 8 Create LVM volume group This volume group will use all data of our physical RAID1 partition /dev/md1. Check if /dev/md1 is a symbolic link. The following command does not work with symbolic links. I had to use vgcreate -A n vg00 /dev/md/1 This creates the volume group vg00. -A n deactivates autobackup of the configuration which does not make sense because we’re actually working on a RAM disk. You should see the following output: vgcreate -- INFO: using default physical extent size 4.00 MB vgcreate -- INFO: maximum logical volume size is 255.99 Gigabyte vgcreate -- WARNING: you don’t have an automatic backup of "vg00" vgcreate -- volume group "vg00" successfully created and activated 5 9 Create LVM logical volumes LVM logical volumes can be used similar as normal partitions. They can contain any file system and data and are also mounted the same way. The big advantage is that with LVM you can resize them while they are in use and you can make snapshots for consistent data backup etc. I advise to use ReiserFS as file system, because after resizing a LVM logical volume you must also resize the file system (or before when making it smaller) and at the moment this is only possible with ReiserFS while the disk is in use (hot resizing). The partition layout depends on your needs, but I advise to make own partitions for /var, /tmp, /usr and /home. /var contains spooling data and log files and can consume all your disk space when something does not work correctly. And if you run out of disk space the system can hang. If it is on an own partition the root file system is not affected. The same can occur with temporary files in /tmp and user data in /usr and /home so it’s always a good idea to separate them. I’m using the following partition layout for LVM: device /dev/vg00/root /dev/vg00/home /dev/vg00/usr /dev/vg00/opt/ /dev/vg00/tmp /dev/vg00/var /dev/vg00/swap mount point / /home /usr /opt /tmp /var swap size 4 GB 4 GB 4 GB 4 GB 512 MB 1 GB 512 MB Create this logical volumes this way lvcreate -A n -L 4096 -n root vg00 lvcreate -A n -L 4096 -n home vg00 lvcreate -A n -L 4096 -n usr vg00 lvcreate -A n -L 4096 -n opt vg00 lvcreate -A n -L 512 -n tmp vg00 lvcreate -A n -L 1024 -n var vg00 lvcreate -A n -L 512 -n swap vg00 For each lvcreate command you should see the following output lvcreate -- WARNING: you don’t have an automatic backup of vg00 lvcreate -- logical volume ‘‘/dev/vg00/root’’ successfully created 10 Create and initialize the swap space Create swap space on logical volume /dev/vg00/swap. mkswap /dev/vg00/swap Activate it. swapon /dev/vg00/swap 6 11 Create file systems We are creating ReiserFS on our volumes. mkreiserfs /dev/vg00/root mkreiserfs /dev/vg00/home mkreiserfs /dev/vg00/usr mkreiserfs /dev/vg00/opt mkreiserfs /dev/vg00/tmp mkreiserfs /dev/vg00/var 12 Edit mountlist.txt Mondo has created a file /tmp/mountlist.txt which is reflecting you old partitions which have been backed up. Copy this file to /tmp/isodir (NFS) so that you can edit it on another machine. mountlist.txt: # device mount point file system /dev/md0 /boot ext2 25600 /dev/vg00/swap swap swap 524288 /dev/vg00/root / reiserfs 4194304 /dev/vg00/home /home reiserfs 4194304 /dev/vg00/opt /opt reiserfs 4194304 /dev/vg00/tmp /tmp reiserfs 131072 /dev/vg00/usr /usr reiserfs 4194304 /dev/vg00/var /var reiserfs 1048576 size[KB] Now copy it back to /tmp so that we can start the restore procedure. cp /tmp/isodir/mountlist.txt /tmp 13 Restore all data to new partitions Now we can use the mondorestore program to restore the data to our new partition layout. mondorestore --interactive You’ll see the same dialog as after booting mindi. Choose again Interactive-¿NFS and hit enter without entering a path. Now you see the next dialog with your new partition layout. Well then you can proceed to restore the data the normal way. If your partitions are big enough to hold the data this should work without any errors. 7 14 Mounting the partitions After restoring the data and returning to the command line again you can mount all partitions defined in mountlist.txt using the script: mount-me With unmount-me you can unmount all this partitions again but before we have to change some configuration files. The new system is mounted in /tmp/RESTORING but it contains still you old fstab and boot configuration. We need to change this so that you can boot it. 15 Edit fstab At first we have to change your /mnt/RESTORING/etc/fstab to your new partition layout. fstab: /dev/vg00/swap swap swap pri=42 0 0 /dev/md0 /boot ext2 defaults 0 0 /dev/vg00/root / reiserfs defaults 0 0 /dev/vg00/usr /usr reiserfs defaults 0 0 /dev/vg00/tmp /tmp reiserfs defaults 0 0 /dev/vg00/opt /opt reiserfs defaults 0 0 /dev/vg00/var /var reiserfs defaults 0 0 /dev/vg00/home /home reiserfs defaults 0 0 devpts /dev/pts devpts mode=0620,gid=5 0 0 proc /proc proc defaults 0 0 usbdevfs /proc/bus/usb usbdevfs noauto 0 0 /dev/cdrom /media/cdrom auto ro,noauto,user,exec 0 0 16 Edit lilo.conf Now we have to configure LILO to boot our RAID system. Change or create the file /mnt/RESTORING/etc/lilo.conf if it does not exist. LILO must be configured to use a initrd which contains the modules for RAID, LVM and ReiserFS. Otherwise the root file system can not be mounted. The creation of the initrd is shown in the next step. lilo.conf: # boot device is our first raid device, the boot sector is written onto both disks in the parti boot = /dev/md0 # extra boot to write the MBR onto both disks raid-extra-boot=/dev/hda,/dev/hdc map=/boot/map default = linux 8 disk=/dev/hda bios=0x80 disk=/dev/hdc bios=0x81 lba32 menu-scheme = Wg:kw:Wg:Wg prompt read-only timeout = 80 image = /boot/vmlinuz label = linux append = "splash=silent" initrd = /boot/initrd-lvm-2.4.20-4GB.gz root = /dev/vg00/root vga = 0x31a image = /boot/vmlinuz.shipped label = failsafe append = "ide=nodma apm=off acpi=off vga=normal nosmp noapic maxcpus=0 3" initrd = /boot/initrd-lvm-2.4.20-4GB.gz root = /dev/vg00/root vga = 0x31a 17 Create initrd There exists a script lvmcreate initrd1 which comes with LVM. It creates an initrd for your kernel and contains the necessary LVM module. But this alone is not enough because the RAID module is missing, so we have to modify the resulting initrd. We must call the script chrooted because we want create it on our target system not on our RAM disk. chroot /mnt/RESTORING lvmcreate initrd You can run it in verbose mode when adding the -v switch. So you can see which files are added. This creates a file /mnt/RESTORING/boot/initrd-lvm-KERNELVERSION.gz. Go for sure that you are using the same name in your lilo.conf. After decompressing the file you can mount it via the loopback device as any other device to your desired mount point. gunzip initrd-lvm-2.4.20-4GB.gz mount -o loop initrd-lvm-2.4.20-4GB /mnt/initrd This initrd contains a mini linux with all modules and commands to setup LVM. We’ll extend this now to also support RAID and ReiserFS. A ls -l should then show the following contents: 1 I would have expected the the ReiserFS module is added by this script because I was using ReiserFS already in my old system. But the module was missing in the created initrd so it seems like only LVM is added and all other modules you need you have to add manually. 9 drwxr-xr-x 8 root root 1024 2004-01-14 13:56 ./ drwxr-xr-x 3 root root 320 2004-01-14 16:55 ../ drwx------ 2 root root 1024 2004-01-14 13:08 bin/ drwxr-xr-x 10 root root 10240 2004-01-14 13:32 dev/ drwxr-xr-x 2 root root 1024 2004-01-14 16:16 etc/ drwxr-xr-x 3 root root 1024 2004-01-14 13:32 lib/ -r-xr-xr-x 1 root root 248 2004-01-14 13:56 linuxrc* drwxr-xr-x 2 root root 1024 2004-01-14 13:32 proc/ drwx------ 2 root root 1024 2004-01-14 13:14 sbin/ During the boot process the initrd is loaded into memory and the containing file linuxrc gets executed. We have to change this linuxrc to load all needed kernel modules and setup RAID and LVM. linuxrc: #!/bin/sh export PATH=/bin:/sbin echo ’Loading module reiserfs ...’ insmod reiserfs echo ’Loading module raid1 ...’ insmod raid1 echo "raidautorun ..." raidautorun echo "done ..." echo ’Loading module lvm-mod ...’ insmod lvm-mod mount /proc echo ’Starting LVM ...’ vgscan vgchange -a y echo ’done ...’ umount /proc The commands insmod, modprobe, vgscan, vgchange, mount and umount are already added by the lvmcreate initrd script. The problem is loading the kernel modules do not work: “Kernel requires old modprobe, but couldn’t run modprobe.old: No such file or directory” So we have to add modprobe.old and insmod.old too. Just copy it to /sbin in the initrd. cp /sbin/insmod.old /mnt/initrd/sbin cp /sbin/modprobe.old /mnt/initrd/sbin Furthermore we need raidautorun to setup the RAID system. cp /sbin/raidautorun /mnt/initrd/sbin The kernel module lvm-mod.o should already be there in /lib/modules/2.4.20-4GB/kernel/drivers/md but we also need raid1.o and reiserfs.o. Create the necessary directories in /lib/modules and add the modules to the initrd. 10 initrd = /mnt/initrd/lib/modules/2.4.20-4GB/kernel # to make the path shorter cp /lib/modules/2.4.20-4GB/kernel/drivers/md/raid1.o $initrd/drivers/md cp /lib/modules/2.4.20-4GB/kernel/drivers/md/lvm-mod.o $initrd/drivers/md cp /lib/modules/2.4.20-4GB/kernel/fs/reiserfs/reiserfs.o $initrd/fs/reiserfs Now all needed files are in the initrd. Unmount it, compress it again and copy it to the target system. umount /mnt/initrd gzip initrd-lvm-2.4.20-4GB cp /tmp/isodir/initrd-lvm-2.4.20-4GB.gz /mnt/RESTORING/boot 18 Install LILO boot loader Lilo must also be executed chrooted so that it can read and write the directories of our target system which is mounted in /mnt/RESTORING and not of the currently running mindi system. Lilo needs also access to /dev so we have to move it into our chrooted environment. This is done by: mount --bind /dev /mnt/RESTORING/dev Now we can run lilo to install the boot loader. chroot /mnt/RESTORING lilo Congratulations. Your are ready to boot your new LVM on RAID linux. Remove the mindi CD and press CTRL-ALT-DEL to reboot. Watch the kernel boot messages to find possible problems. 19 Troubleshooting If you have to reboot mindi because of some problems you always have to do the same steps to access your already created partitions. To make this easier I’ve written this script so after configuring NFS and quitting mondorestore you only have to call /tmp/isodir/setup.sh. setup.sh: #!/bin/sh cp /tmp/isodir/mountlist.txt /tmp cp /tmp/isodir/raidtab /etc raidstart --all modprobe lvm-mod vgscan vgchange -a y vg00 mount-me mount --bind /dev /mnt/RESTORING/dev 11
© Copyright 2024