1
0
mirror of https://github.com/samba-team/samba.git synced 2024-12-25 23:21:54 +03:00
samba-mirror/ctdb/tests/simple/78_ctdb_large_db_recovery.sh
Amitay Isaacs a7b8ee87fe ctdb-tests: Add a test for recovery of large databases
Signed-off-by: Amitay Isaacs <amitay@gmail.com>
Reviewed-by: Martin Schwenke <martin@meltin.net>
2016-03-25 03:26:15 +01:00

108 lines
2.7 KiB
Bash
Executable File

#!/bin/bash
test_info()
{
cat <<EOF
Older style of recovery using PULL_DB and PUSH_DB controls tries to
construct a single large marshall buffer for all the records in the
database. However, this approach is problematic as talloc restricts the
maximum size of buffer to 256M. Also, trying to construct and send large
buffers is inefficient and can cause CTDB daemon to be tied up for long
periods of time.
Instead new style recovery is introduced using DB_PULL and
DB_PUSH_START/DB_PUSH_CONFIRM controls. This sends the records in
batches of ~RecBufferSizeLimit in size at a time.
Expected results:
* The recovery should complete successfully
EOF
}
. "${TEST_SCRIPTS_DIR}/integration.bash"
ctdb_test_init "$@"
set -e
cluster_is_healthy
# Reset configuration
ctdb_restart_when_done
#
# Main test
#
TEST1DB="large_persistent_db.tdb"
TEST2DB="large_volatile_db.tdb"
RECDATA=$(onnode 0 mktemp)
# Create a persistent database to test
echo "create persistent test database $TEST1DB"
try_command_on_node 0 $CTDB attach $TEST1DB persistent
# Wipe Test database
echo "wipe test database $TEST1DB"
try_command_on_node 0 $CTDB wipedb $TEST1DB
# Create dummy record data
echo "creating dummy record data"
onnode 0 dd if=/dev/urandom of=$RECDATA bs=10K count=1
# Add 345 records
echo "Adding 345 records"
for i in $(seq 1 345) ; do
try_command_on_node 0 $CTDB pstore $TEST1DB record$i $RECDATA || exit 1
done
# Create a volatile database to test
echo "create volatile test database $TEST2DB"
try_command_on_node 0 $CTDB attach $TEST2DB
# Wipe Test database
echo "wipe test database $TEST2DB"
try_command_on_node 0 $CTDB wipedb $TEST2DB
# Create dummy record data
v1="1234567890"
v2="$v1$v1$v1$v1$v1$v1$v1$v1$v1$v1"
v3="$v2$v2$v2$v2$v2$v2$v2$v2$v2$v2"
# Add 1234 records
echo "Adding 1234 records"
for i in $(seq 1 1234) ; do
try_command_on_node 0 $CTDB writekey $TEST2DB record$i $v3 || exit 1
done
echo "Find out which node is recmaster"
try_command_on_node 0 $CTDB recmaster
recmaster="$out"
# Set RecBufferSizeLimit to 10000
try_command_on_node $recmaster $CTDB setvar RecBufferSizeLimit 10000
# Do a recovery
echo "force recovery"
try_command_on_node 0 $CTDB recover
wait_until_node_has_status 0 recovered 30
# check that there are correct number of records
num_records=$(db_ctdb_cattdb_count_records 0 $TEST1DB)
if [ $num_records = "345" ] ; then
echo "OK: persistent database recovered correctly"
else
echo "BAD: persistent database ended up with $num_records of 345 records"
exit 1
fi
num_records=$(db_ctdb_cattdb_count_records 0 $TEST2DB)
if [ $num_records = "1234" ] ; then
echo "OK: volatile database recovered correctly"
else
echo "BAD: volatile database ended up with $num_records of 1234 records"
exit 1
fi