mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-22 17:35:59 +03:00
d9e8895a96
dm-integrity stores checksums of the data written to an LV, and returns an error if data read from the LV does not match the previously saved checksum. When used on raid images, dm-raid will correct the error by reading the block from another image, and the device user sees no error. The integrity metadata (checksums) are stored on an internal LV allocated by lvm for each linear image. The internal LV is allocated on the same PV as the image. Create a raid LV with an integrity layer over each raid image (for raid levels 1,4,5,6,10): lvcreate --type raidN --raidintegrity y [options] Add an integrity layer to images of an existing raid LV: lvconvert --raidintegrity y LV Remove the integrity layer from images of a raid LV: lvconvert --raidintegrity n LV Settings Use --raidintegritymode journal|bitmap (journal is default) to configure the method used by dm-integrity to ensure crash consistency. Initialization When integrity is added to an LV, the kernel needs to initialize the integrity metadata/checksums for all blocks in the LV. The data corruption checking performed by dm-integrity will only operate on areas of the LV that are already initialized. The progress of integrity initialization is reported by the "syncpercent" LV reporting field (and under the Cpy%Sync lvs column.) Example: create a raid1 LV with integrity: $ lvcreate --type raid1 -m1 --raidintegrity y -n rr -L1G foo Creating integrity metadata LV rr_rimage_0_imeta with size 12.00 MiB. Logical volume "rr_rimage_0_imeta" created. Creating integrity metadata LV rr_rimage_1_imeta with size 12.00 MiB. Logical volume "rr_rimage_1_imeta" created. Logical volume "rr" created. $ lvs -a foo LV VG Attr LSize Origin Cpy%Sync rr foo rwi-a-r--- 1.00g 4.93 [rr_rimage_0] foo gwi-aor--- 1.00g [rr_rimage_0_iorig] 41.02 [rr_rimage_0_imeta] foo ewi-ao---- 12.00m [rr_rimage_0_iorig] foo -wi-ao---- 1.00g [rr_rimage_1] foo gwi-aor--- 1.00g [rr_rimage_1_iorig] 39.45 [rr_rimage_1_imeta] foo ewi-ao---- 12.00m [rr_rimage_1_iorig] foo -wi-ao---- 1.00g [rr_rmeta_0] foo ewi-aor--- 4.00m [rr_rmeta_1] foo ewi-aor--- 4.00m
736 lines
17 KiB
Bash
736 lines
17 KiB
Bash
#!/usr/bin/env bash
|
|
|
|
# Copyright (C) 2018 Red Hat, Inc. All rights reserved.
|
|
#
|
|
# This copyrighted material is made available to anyone wishing to use,
|
|
# modify, copy, or redistribute it subject to the terms and conditions
|
|
# of the GNU General Public License v.2.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program; if not, write to the Free Software Foundation,
|
|
# Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
SKIP_WITH_LVMPOLLD=1
|
|
|
|
. lib/inittest
|
|
|
|
aux have_integrity 1 5 0 || skip
|
|
which mkfs.xfs || skip
|
|
which xfs_growfs || skip
|
|
|
|
mnt="mnt"
|
|
mkdir -p $mnt
|
|
|
|
aux prepare_devs 5 64
|
|
|
|
for i in `seq 1 16384`; do echo -n "A" >> fileA; done
|
|
for i in `seq 1 16384`; do echo -n "B" >> fileB; done
|
|
for i in `seq 1 16384`; do echo -n "C" >> fileC; done
|
|
|
|
# generate random data
|
|
dd if=/dev/urandom of=randA bs=512K count=2
|
|
dd if=/dev/urandom of=randB bs=512K count=3
|
|
dd if=/dev/urandom of=randC bs=512K count=4
|
|
|
|
_prepare_vg() {
|
|
# zero devs so we are sure to find the correct file data
|
|
# on the underlying devs when corrupting it
|
|
dd if=/dev/zero of="$dev1" || true
|
|
dd if=/dev/zero of="$dev2" || true
|
|
dd if=/dev/zero of="$dev3" || true
|
|
dd if=/dev/zero of="$dev4" || true
|
|
dd if=/dev/zero of="$dev5" || true
|
|
vgcreate $SHARED $vg "$dev1" "$dev2" "$dev3" "$dev4" "$dev5"
|
|
pvs
|
|
}
|
|
|
|
_test_fs_with_error() {
|
|
mkfs.xfs -f "$DM_DEV_DIR/$vg/$lv1"
|
|
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# add original data
|
|
cp fileA $mnt
|
|
cp fileB $mnt
|
|
cp fileC $mnt
|
|
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
|
|
# corrupt the original data on the underying dev
|
|
# flip one bit in fileB, changing a 0x42 to 0x43
|
|
# the bit is changed in the last 4096 byte block
|
|
# of the file, so when reading back the file we
|
|
# will get the first three 4096 byte blocks, for
|
|
# a total of 12288 bytes before getting an error
|
|
# on the last 4096 byte block.
|
|
xxd "$dev1" > dev1.txt
|
|
tac dev1.txt > dev1.rev
|
|
sed -e '0,/4242 4242 4242 4242 4242 4242 4242 4242/ s/4242 4242 4242 4242 4242 4242 4242 4242/4242 4242 4242 4242 4242 4242 4242 4243/' dev1.rev > dev1.rev.bad
|
|
tac dev1.rev.bad > dev1.bad
|
|
xxd -r dev1.bad > "$dev1"
|
|
rm dev1.txt dev1.rev dev1.rev.bad dev1.bad
|
|
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# read complete fileA which was not corrupted
|
|
dd if=$mnt/fileA of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp
|
|
diff fileA tmp
|
|
rm tmp
|
|
|
|
# read partial fileB which was corrupted
|
|
not dd if=$mnt/fileB of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp | grep 12288
|
|
not diff fileB tmp
|
|
rm tmp
|
|
|
|
umount $mnt
|
|
}
|
|
|
|
_test_fs_with_raid() {
|
|
mkfs.xfs -f "$DM_DEV_DIR/$vg/$lv1"
|
|
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# add original data
|
|
cp fileA $mnt
|
|
cp fileB $mnt
|
|
cp fileC $mnt
|
|
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
|
|
xxd "$dev1" > dev1.txt
|
|
tac dev1.txt > dev1.rev
|
|
sed -e '0,/4242 4242 4242 4242 4242 4242 4242 4242/ s/4242 4242 4242 4242 4242 4242 4242 4242/4242 4242 4242 4242 4242 4242 4242 4243/' dev1.rev > dev1.rev.bad
|
|
tac dev1.rev.bad > dev1.bad
|
|
xxd -r dev1.bad > "$dev1"
|
|
rm dev1.txt dev1.rev dev1.rev.bad dev1.bad
|
|
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# read complete fileA which was not corrupted
|
|
dd if=$mnt/fileA of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp | grep 16384
|
|
diff fileA tmp
|
|
rm tmp
|
|
|
|
# read complete fileB, corruption is corrected by raid
|
|
dd if=$mnt/fileB of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp | grep 16384
|
|
diff fileB tmp
|
|
rm tmp
|
|
|
|
umount $mnt
|
|
}
|
|
|
|
_add_new_data_to_mnt() {
|
|
mkfs.xfs -f "$DM_DEV_DIR/$vg/$lv1"
|
|
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# add original data
|
|
cp randA $mnt
|
|
cp randB $mnt
|
|
cp randC $mnt
|
|
mkdir $mnt/1
|
|
cp fileA $mnt/1
|
|
cp fileB $mnt/1
|
|
cp fileC $mnt/1
|
|
mkdir $mnt/2
|
|
cp fileA $mnt/2
|
|
cp fileB $mnt/2
|
|
cp fileC $mnt/2
|
|
}
|
|
|
|
_add_more_data_to_mnt() {
|
|
mkdir $mnt/more
|
|
cp fileA $mnt/more
|
|
cp fileB $mnt/more
|
|
cp fileC $mnt/more
|
|
cp randA $mnt/more
|
|
cp randB $mnt/more
|
|
cp randC $mnt/more
|
|
}
|
|
|
|
_verify_data_on_mnt() {
|
|
diff randA $mnt/randA
|
|
diff randB $mnt/randB
|
|
diff randC $mnt/randC
|
|
diff fileA $mnt/1/fileA
|
|
diff fileB $mnt/1/fileB
|
|
diff fileC $mnt/1/fileC
|
|
diff fileA $mnt/2/fileA
|
|
diff fileB $mnt/2/fileB
|
|
diff fileC $mnt/2/fileC
|
|
}
|
|
|
|
_verify_data_on_lv() {
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
_verify_data_on_mnt
|
|
rm $mnt/randA
|
|
rm $mnt/randB
|
|
rm $mnt/randC
|
|
rm -rf $mnt/1
|
|
rm -rf $mnt/2
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
}
|
|
|
|
_sync_percent() {
|
|
local checklv=$1
|
|
get lv_field "$checklv" sync_percent | cut -d. -f1
|
|
}
|
|
|
|
_wait_recalc() {
|
|
local checklv=$1
|
|
|
|
for i in $(seq 1 10) ; do
|
|
sync=$(_sync_percent "$checklv")
|
|
echo "sync_percent is $sync"
|
|
|
|
if test "$sync" = "100"; then
|
|
return
|
|
fi
|
|
|
|
sleep 1
|
|
done
|
|
|
|
echo "timeout waiting for recalc"
|
|
return 1
|
|
}
|
|
|
|
# Test corrupting data on an image and verifying that
|
|
# it is detected by integrity and corrected by raid.
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m2 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid4 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test removing integrity from an active LV
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid4 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test adding integrity to an active LV
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid4 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test lvextend while inactive
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
lvextend -l 16 $vg/$lv1
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
lvextend -l 16 $vg/$lv1
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test lvextend while active
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test adding image to raid1
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert -y -m+1 $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test removing image from raid1
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m2 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert -y -m-1 $vg/$lv1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test disallowed operations on raid+integrity
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
not lvconvert -y -m-1 $vg/$lv1
|
|
not lvconvert --splitmirrors 1 -n tmp -y $vg/$lv1
|
|
not lvconvert --splitmirrors 1 --trackchanges -y $vg/$lv1
|
|
not lvchange --syncaction check $vg/$lv1
|
|
not lvchange --syncaction repair $vg/$lv1
|
|
not lvreduce -L4M $vg/$lv1
|
|
not lvcreate -s -n snap -L4M $vg/$lv1
|
|
not pvmove -n $vg/$lv1 "$dev1"
|
|
not pvmove "$dev1"
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Repeat many of the tests above using bitmap mode
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# remove from active lv
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# add to active lv
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y --raidintegritymode bitmap $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# lvextend active
|
|
_prepare_vg
|
|
lvcreate --type raid1 --raidintegrity y --raidintegritymode bitmap -m1 -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
xfs_growfs $mnt
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# add image to raid1
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert -y -m+1 $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test that raid+integrity cannot be a sublv
|
|
# part1: cannot add integrity to a raid LV that is already a sublv
|
|
|
|
_prepare_vg
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvconvert -y --type thin-pool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_tdata
|
|
not lvconvert --raidintegrity y $vg/${lv1}_tmeta
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvconvert -y --type cache-pool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_cdata
|
|
not lvconvert --raidintegrity y $vg/${lv1}_cmeta
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type cache-pool -n cpool -l 8 $vg
|
|
lvconvert -y --type cache --cachepool cpool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type raid1 -m1 -n cvol -l 8 $vg
|
|
lvconvert -y --type cache --cachevol cvol $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
not lvconvert --raidintegrity y $vg/cvol
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate -n cvol -l 8 $vg
|
|
lvchange -an $vg
|
|
lvconvert -y --type writecache --cachevol cvol $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_wcorig
|
|
lvremove -y $vg/$lv1
|
|
|
|
# Test that raid+integrity cannot be a sublv
|
|
# part2: cannot convert an existing raid+integrity LV into a sublv
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvconvert -y --type thin-pool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_tdata
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type raid1 -m1 -n $lv2 -l 8 $vg
|
|
lvconvert -y --type cache --cachevol $lv2 $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
not lvconvert --raidintegrity y $vg/${lv2}_vol
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type raid1 -m1 -n $lv2 -l 8 $vg
|
|
lvconvert -y --type cache --cachepool $lv2 $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
not lvconvert --raidintegrity y $vg/${lv2}_cpool_cdata
|
|
lvremove -y $vg/$lv1
|
|
|
|
# cannot add integrity to raid that has a snapshot
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate -s -n $lv2 -l 8 $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
lvremove -y $vg/$lv1
|
|
|
|
vgremove -ff $vg
|