[[PageOutline]] * verbose log of [../] = mdadm -D /dev/md0 = * mitty@ubuntu-haa:~$ sudo mdadm -D /dev/md0 {{{ /dev/md0: Version : 00.90 Creation Time : Thu Mar 24 17:31:36 2011 Raid Level : raid1 Array Size : 2047936 (2000.27 MiB 2097.09 MB) Used Dev Size : 2047936 (2000.27 MiB 2097.09 MB) Raid Devices : 2 Total Devices : 2 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Thu Mar 24 17:31:47 2011 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 UUID : 32578407:6f33f50b:893cf340:745f5dce (local to host ubuntu-haa) Events : 0.34 Number Major Minor RaidDevice State 0 8 17 0 active sync /dev/sdb1 1 8 33 1 active sync /dev/sdc1 }}} = dpkg-reconfigure mdadm = * see also wiki:TipAndDoc/RAID#dpkg-reconfiguremdadm * mitty@ubuntu-haa:~$ sudo dpkg-reconfigure mdadm {{{ ┌──────────────────────────┤ Configuring mdadm ├─ │ │ │ If the kernel supports it (versions greater than 2.6.14), mdadm can │ │ periodically check the redundancy of MD arrays (RAIDs). This may be a │ │ resource-intensive process, depending on the local setup, but it could │ │ help prevent rare cases of data loss. Note that this is a read-only │ │ check unless errors are found; if errors are found, mdadm will try to │ │ correct them, which may result in write access to the media. │ │ │ │ The default, if turned on, is to check on the first Sunday of every │ │ month at 01:06. │ │ │ │ Should mdadm run monthly redundancy checks of the MD arrays? │ │ │ │ [[]] │ │ │ └────────────────────────────────────── (snip) ┌──────────────────────────┤ Configuring mdadm ├─ │ │ │ The MD (RAID) monitor daemon sends email notifications in response to │ │ important MD events (such as a disk failure). │ │ │ │ Enabling this option is recommended. │ │ │ │ Do you want to start the MD monitoring daemon? │ │ │ │ [[]] │ │ │ └────────────────────────────────────── (snip) ┌─────────────────────────┤ Configuring mdadm ├── │ Please enter the email address of the user who should get the email │ │ notifications for important MD events. │ │ │ │ Recipient for email notifications: │ │ │ │ root_________________________________________________________________ │ │ │ │ │ │ │ └─────────────────────────────────────── (snip) ┌───────────────────────────┤ Configuring mdadm ├ │ │ │ If your root filesystem is on a RAID, and a disk is missing at boot, it │ │ can either boot with the degraded array, or hold the system at a │ │ recovery shell. │ │ │ │ Running a system with a degraded RAID could result in permanent data │ │ loss if it suffers another hardware fault. │ │ │ │ If you do not have access to the server console to use the recovery │ │ shell, you might answer "yes" to enable the system to boot unattended. │ │ │ │ Do you want to boot your system if your RAID becomes degraded? │ │ │ │ [[]] │ │ │ └─────────────────────────────────────── }}} = pvcreate /dev/md0 = * mitty@ubuntu-haa:~$ sudo pvcreate /dev/md0 {{{ Physical volume "/dev/md0" successfully created }}} * mitty@ubuntu-haa:~$ sudo pvdisplay -C {{{ PV VG Fmt Attr PSize PFree /dev/md0 lvm2 -- 1.95g 1.95g }}} * mitty@ubuntu-haa:~$ sudo pvdisplay {{{ "/dev/md0" is a new physical volume of "1.95 GiB" --- NEW Physical volume --- PV Name /dev/md0 VG Name PV Size 1.95 GiB Allocatable NO PE Size 0 Total PE 0 Free PE 0 Allocated PE 0 PV UUID Z2JXRP-fa5g-SYS5-xzMs-Lq8C-1Jbh-QPKihr }}} = vgcreate -s 32 vgnfs /dev/md0 = * mitty@ubuntu-haa:~$ sudo vgcreate -s 32 vgnfs /dev/md0 {{{ Volume group "vgnfs" successfully created }}} * mitty@ubuntu-haa:~$ sudo vgdisplay -C {{{ VG #PV #LV #SN Attr VSize VFree vgnfs 1 0 0 wz--n- 1.94g 1.94g }}} * mitty@ubuntu-haa:~$ sudo vgdisplay {{{ --- Volume group --- VG Name vgnfs System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 1 VG Access read/write VG Status resizable MAX LV 0 Cur LV 0 Open LV 0 Max PV 0 Cur PV 1 Act PV 1 VG Size 1.94 GiB PE Size 32.00 MiB Total PE 62 Alloc PE / Size 0 / 0 Free PE / Size 62 / 1.94 GiB VG UUID I6vVoh-6gCJ-9uvA-v2MV-Fyva-7J8v-Cvftfi }}} * mitty@ubuntu-haa:~$ sudo pvdisplay -C {{{ PV VG Fmt Attr PSize PFree /dev/md0 vgnfs lvm2 a- 1.94g 1.94g }}} * mitty@ubuntu-haa:~$ sudo pvdisplay {{{ --- Physical volume --- PV Name /dev/md0 VG Name vgnfs PV Size 1.95 GiB / not usable 15.94 MiB Allocatable yes PE Size 32.00 MiB Total PE 62 Free PE 62 Allocated PE 0 PV UUID Z2JXRP-fa5g-SYS5-xzMs-Lq8C-1Jbh-QPKihr }}} = lvcreate --size 1.5G -n drbd vgnfs = * mitty@ubuntu-haa:~$ sudo lvcreate --size 1.5G -n drbd vgnfs {{{ Logical volume "drbd" created }}} * mitty@ubuntu-haa:~$ sudo pvdisplay -C {{{ PV VG Fmt Attr PSize PFree /dev/md0 vgnfs lvm2 a- 1.94g 448.00m }}} * mitty@ubuntu-haa:~$ sudo pvdisplay {{{ --- Physical volume --- PV Name /dev/md0 VG Name vgnfs PV Size 1.95 GiB / not usable 15.94 MiB Allocatable yes PE Size 32.00 MiB Total PE 62 Free PE 14 Allocated PE 48 PV UUID Z2JXRP-fa5g-SYS5-xzMs-Lq8C-1Jbh-QPKihr }}} * mitty@ubuntu-haa:~$ sudo vgdisplay -v {{{ Finding all volume groups Finding volume group "vgnfs" --- Volume group --- VG Name vgnfs System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 2 VG Access read/write VG Status resizable MAX LV 0 Cur LV 1 Open LV 0 Max PV 0 Cur PV 1 Act PV 1 VG Size 1.94 GiB PE Size 32.00 MiB Total PE 62 Alloc PE / Size 48 / 1.50 GiB Free PE / Size 14 / 448.00 MiB VG UUID I6vVoh-6gCJ-9uvA-v2MV-Fyva-7J8v-Cvftfi --- Logical volume --- LV Name /dev/vgnfs/drbd VG Name vgnfs LV UUID dNxdNj-hZCZ-GMrC-woMk-0hA2-f3oR-sBsrHI LV Write Access read/write LV Status available # open 0 LV Size 1.50 GiB Current LE 48 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 251:0 --- Physical volumes --- PV Name /dev/md0 PV UUID Z2JXRP-fa5g-SYS5-xzMs-Lq8C-1Jbh-QPKihr PV Status allocatable Total PE / Free PE 62 / 14 }}} * mitty@ubuntu-haa:~$ sudo pvscan {{{ PV /dev/md0 VG vgnfs lvm2 [1.94 GiB / 448.00 MiB free] Total: 1 [1.94 GiB] / in use: 1 [1.94 GiB] / in no VG: 0 [0 ] }}} * mitty@ubuntu-haa:~$ sudo vgscan {{{ Reading all physical volumes. This may take a while... Found volume group "vgnfs" using metadata type lvm2 }}} * mitty@ubuntu-haa:~$ sudo lvscan {{{ ACTIVE '/dev/vgnfs/drbd' [1.50 GiB] inherit }}}