mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-22 17:35:59 +03:00
7b04ed07ba
The test was using a raid+integrity LV without first waiting for the integrity sync, which could cause the test to fail (depending on init speed) where it depends on integrity to work in uninitialized areas. Also use cmp instead of diff.
764 lines
18 KiB
Bash
764 lines
18 KiB
Bash
#!/usr/bin/env bash
|
|
|
|
# Copyright (C) 2018 Red Hat, Inc. All rights reserved.
|
|
#
|
|
# This copyrighted material is made available to anyone wishing to use,
|
|
# modify, copy, or redistribute it subject to the terms and conditions
|
|
# of the GNU General Public License v.2.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program; if not, write to the Free Software Foundation,
|
|
# Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
SKIP_WITH_LVMPOLLD=1
|
|
|
|
. lib/inittest
|
|
|
|
aux have_integrity 1 5 0 || skip
|
|
which mkfs.xfs || skip
|
|
which xfs_growfs || skip
|
|
|
|
mnt="mnt"
|
|
mkdir -p $mnt
|
|
|
|
aux prepare_devs 5 64
|
|
|
|
printf "%0.sA" {1..16384} >> fileA
|
|
printf "%0.sB" {1..16384} >> fileB
|
|
printf "%0.sC" {1..16384} >> fileC
|
|
|
|
# generate random data
|
|
dd if=/dev/urandom of=randA bs=512K count=2
|
|
dd if=/dev/urandom of=randB bs=512K count=3
|
|
dd if=/dev/urandom of=randC bs=512K count=4
|
|
|
|
_prepare_vg() {
|
|
# zero devs so we are sure to find the correct file data
|
|
# on the underlying devs when corrupting it
|
|
dd if=/dev/zero of="$dev1" || true
|
|
dd if=/dev/zero of="$dev2" || true
|
|
dd if=/dev/zero of="$dev3" || true
|
|
dd if=/dev/zero of="$dev4" || true
|
|
dd if=/dev/zero of="$dev5" || true
|
|
vgcreate $SHARED $vg "$dev1" "$dev2" "$dev3" "$dev4" "$dev5"
|
|
pvs
|
|
}
|
|
|
|
_test_fs_with_error() {
|
|
mkfs.xfs -f "$DM_DEV_DIR/$vg/$lv1"
|
|
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# add original data
|
|
cp fileA $mnt
|
|
cp fileB $mnt
|
|
cp fileC $mnt
|
|
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
|
|
# corrupt the original data on the underying dev
|
|
# flip one bit in fileB, changing a 0x42 to 0x43
|
|
# the bit is changed in the last 4096 byte block
|
|
# of the file, so when reading back the file we
|
|
# will get the first three 4096 byte blocks, for
|
|
# a total of 12288 bytes before getting an error
|
|
# on the last 4096 byte block.
|
|
xxd "$dev1" > dev1.txt
|
|
tac dev1.txt > dev1.rev
|
|
sed -e '0,/4242 4242 4242 4242 4242 4242 4242 4242/ s/4242 4242 4242 4242 4242 4242 4242 4242/4242 4242 4242 4242 4242 4242 4242 4243/' dev1.rev > dev1.rev.bad
|
|
tac dev1.rev.bad > dev1.bad
|
|
xxd -r dev1.bad > "$dev1"
|
|
rm dev1.txt dev1.rev dev1.rev.bad dev1.bad
|
|
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# read complete fileA which was not corrupted
|
|
dd if=$mnt/fileA of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp
|
|
cmp -b fileA tmp
|
|
rm tmp
|
|
|
|
# read partial fileB which was corrupted
|
|
not dd if=$mnt/fileB of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp | grep 12288
|
|
not cmp -b fileB tmp
|
|
rm tmp
|
|
|
|
umount $mnt
|
|
}
|
|
|
|
_test_fs_with_raid() {
|
|
mkfs.xfs -f "$DM_DEV_DIR/$vg/$lv1"
|
|
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# add original data
|
|
cp fileA $mnt
|
|
cp fileB $mnt
|
|
cp fileC $mnt
|
|
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
|
|
xxd "$dev1" > dev1.txt
|
|
tac dev1.txt > dev1.rev
|
|
sed -e '0,/4242 4242 4242 4242 4242 4242 4242 4242/ s/4242 4242 4242 4242 4242 4242 4242 4242/4242 4242 4242 4242 4242 4242 4242 4243/' dev1.rev > dev1.rev.bad
|
|
tac dev1.rev.bad > dev1.bad
|
|
xxd -r dev1.bad > "$dev1"
|
|
rm dev1.txt dev1.rev dev1.rev.bad dev1.bad
|
|
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# read complete fileA which was not corrupted
|
|
dd if=$mnt/fileA of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp | grep 16384
|
|
cmp -b fileA tmp
|
|
rm tmp
|
|
|
|
# read complete fileB, corruption is corrected by raid
|
|
dd if=$mnt/fileB of=tmp bs=1k
|
|
ls -l tmp
|
|
stat -c %s tmp | grep 16384
|
|
cmp -b fileB tmp
|
|
rm tmp
|
|
|
|
umount $mnt
|
|
}
|
|
|
|
_add_new_data_to_mnt() {
|
|
mkfs.xfs -f "$DM_DEV_DIR/$vg/$lv1"
|
|
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
|
|
# add original data
|
|
cp randA $mnt
|
|
cp randB $mnt
|
|
cp randC $mnt
|
|
mkdir $mnt/1
|
|
cp fileA $mnt/1
|
|
cp fileB $mnt/1
|
|
cp fileC $mnt/1
|
|
mkdir $mnt/2
|
|
cp fileA $mnt/2
|
|
cp fileB $mnt/2
|
|
cp fileC $mnt/2
|
|
}
|
|
|
|
_add_more_data_to_mnt() {
|
|
mkdir $mnt/more
|
|
cp fileA $mnt/more
|
|
cp fileB $mnt/more
|
|
cp fileC $mnt/more
|
|
cp randA $mnt/more
|
|
cp randB $mnt/more
|
|
cp randC $mnt/more
|
|
}
|
|
|
|
_verify_data_on_mnt() {
|
|
cmp -b randA $mnt/randA
|
|
cmp -b randB $mnt/randB
|
|
cmp -b randC $mnt/randC
|
|
cmp -b fileA $mnt/1/fileA
|
|
cmp -b fileB $mnt/1/fileB
|
|
cmp -b fileC $mnt/1/fileC
|
|
cmp -b fileA $mnt/2/fileA
|
|
cmp -b fileB $mnt/2/fileB
|
|
cmp -b fileC $mnt/2/fileC
|
|
}
|
|
|
|
_verify_data_on_lv() {
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
_verify_data_on_mnt
|
|
rm $mnt/randA
|
|
rm $mnt/randB
|
|
rm $mnt/randC
|
|
rm -rf $mnt/1
|
|
rm -rf $mnt/2
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
}
|
|
|
|
_sync_percent() {
|
|
local checklv=$1
|
|
get lv_field "$checklv" sync_percent | cut -d. -f1
|
|
}
|
|
|
|
_wait_recalc() {
|
|
local checklv=$1
|
|
|
|
for i in $(seq 1 10) ; do
|
|
sync=$(_sync_percent "$checklv")
|
|
echo "sync_percent is $sync"
|
|
|
|
if test "$sync" = "100"; then
|
|
return
|
|
fi
|
|
|
|
sleep 1
|
|
done
|
|
|
|
# TODO: There is some strange bug, first leg of RAID with integrity
|
|
# enabled never gets in sync. I saw this in BB, but not when executing
|
|
# the commands manually
|
|
if test -z "$sync"; then
|
|
echo "TEST WARNING: Resync of dm-integrity device '$checklv' failed"
|
|
dmsetup status "$DM_DEV_DIR/mapper/${checklv/\//-}"
|
|
exit
|
|
fi
|
|
echo "timeout waiting for recalc"
|
|
return 1
|
|
}
|
|
|
|
# Test corrupting data on an image and verifying that
|
|
# it is detected by integrity and corrected by raid.
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m2 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid4 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
_wait_recalc $vg/${lv1}_rimage_3
|
|
_wait_recalc $vg/${lv1}_rimage_4
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
_wait_recalc $vg/${lv1}_rimage_3
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test removing integrity from an active LV
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid4 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test adding integrity to an active LV
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid4 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test lvextend while inactive
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
lvextend -l 16 $vg/$lv1
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
lvextend -l 16 $vg/$lv1
|
|
lvchange -ay $vg/$lv1
|
|
mount "$DM_DEV_DIR/$vg/$lv1" $mnt
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test lvextend while active
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid5 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid10 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
xfs_growfs $mnt
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test adding image to raid1
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert -y -m+1 $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test removing image from raid1
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m2 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert -y -m-1 $vg/$lv1
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test disallowed operations on raid+integrity
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
not lvconvert -y -m-1 $vg/$lv1
|
|
not lvconvert --splitmirrors 1 -n tmp -y $vg/$lv1
|
|
not lvconvert --splitmirrors 1 --trackchanges -y $vg/$lv1
|
|
not lvchange --syncaction check $vg/$lv1
|
|
not lvchange --syncaction repair $vg/$lv1
|
|
not lvreduce -L4M $vg/$lv1
|
|
not lvcreate -s -n snap -L4M $vg/$lv1
|
|
not pvmove -n $vg/$lv1 "$dev1"
|
|
not pvmove "$dev1"
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Repeat many of the tests above using bitmap mode
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
_prepare_vg
|
|
lvcreate --type raid6 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_test_fs_with_raid
|
|
lvchange -an $vg/$lv1
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# remove from active lv
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity n $vg/$lv1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# add to active lv
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert --raidintegrity y --raidintegritymode bitmap $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# lvextend active
|
|
_prepare_vg
|
|
lvcreate --type raid1 --raidintegrity y --raidintegritymode bitmap -m1 -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvextend -l 16 $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
xfs_growfs $mnt
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# add image to raid1
|
|
_prepare_vg
|
|
lvcreate --type raid1 -m1 --raidintegrity y --raidintegritymode bitmap -n $lv1 -l 8 $vg
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
lvs -a -o+devices $vg
|
|
_add_new_data_to_mnt
|
|
lvconvert -y -m+1 $vg/$lv1
|
|
_wait_recalc $vg/${lv1}_rimage_0
|
|
_wait_recalc $vg/${lv1}_rimage_1
|
|
_wait_recalc $vg/${lv1}_rimage_2
|
|
lvs -a -o+devices $vg
|
|
_add_more_data_to_mnt
|
|
_verify_data_on_mnt
|
|
umount $mnt
|
|
lvchange -an $vg/$lv1
|
|
_verify_data_on_lv
|
|
lvremove $vg/$lv1
|
|
vgremove -ff $vg
|
|
|
|
# Test that raid+integrity cannot be a sublv
|
|
# part1: cannot add integrity to a raid LV that is already a sublv
|
|
|
|
_prepare_vg
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvconvert -y --type thin-pool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_tdata
|
|
not lvconvert --raidintegrity y $vg/${lv1}_tmeta
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvconvert -y --type cache-pool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_cdata
|
|
not lvconvert --raidintegrity y $vg/${lv1}_cmeta
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type cache-pool -n cpool -l 8 $vg
|
|
lvconvert -y --type cache --cachepool cpool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type raid1 -m1 -n cvol -l 8 $vg
|
|
lvconvert -y --type cache --cachevol cvol $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
not lvconvert --raidintegrity y $vg/cvol
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate -n cvol -l 8 $vg
|
|
lvchange -an $vg
|
|
lvconvert -y --type writecache --cachevol cvol $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_wcorig
|
|
lvremove -y $vg/$lv1
|
|
|
|
# Test that raid+integrity cannot be a sublv
|
|
# part2: cannot convert an existing raid+integrity LV into a sublv
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvconvert -y --type thin-pool $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_tdata
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type raid1 -m1 -n $lv2 -l 8 $vg
|
|
lvconvert -y --type cache --cachevol $lv2 $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
not lvconvert --raidintegrity y $vg/${lv2}_vol
|
|
lvremove -y $vg/$lv1
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate --type raid1 -m1 -n $lv2 -l 8 $vg
|
|
lvconvert -y --type cache --cachepool $lv2 $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/${lv1}_corig
|
|
not lvconvert --raidintegrity y $vg/${lv2}_cpool_cdata
|
|
lvremove -y $vg/$lv1
|
|
|
|
# cannot add integrity to raid that has a snapshot
|
|
|
|
lvcreate --type raid1 -m1 -n $lv1 -l 8 $vg
|
|
lvcreate -s -n $lv2 -l 8 $vg/$lv1
|
|
not lvconvert --raidintegrity y $vg/$lv1
|
|
lvremove -y $vg/$lv1
|
|
|
|
vgremove -ff $vg
|