mgmt/glusterd: make sure quota enforcer has established connection with quotad before marking quota as enabled.

without this patch there is a window of time when quota is marked as
enabled in quota-enforcer, but connection to quotad wouldn't have been
established. Any checklimit done during this period can result in a
failed fop because of unavailability of quotad.

Change-Id: I0d509fabc434dd55ce9ec59157123524197fcc80
Signed-off-by: Raghavendra G <rgowdapp@redhat.com>
BUG: 969461
Signed-off-by: Raghavendra G <rgowdapp@redhat.com>
Reviewed-on: http://review.gluster.org/6572
Tested-by: Gluster Build System <jenkins@build.gluster.com>
Reviewed-by: Vijay Bellur <vbellur@redhat.com>
This commit is contained in:
Raghavendra G 2013-12-23 14:54:05 +05:30 committed by Vijay Bellur
parent a10100a4b8
commit f05607fe05
10 changed files with 181 additions and 73 deletions

View File

@ -3341,6 +3341,34 @@ reconfigure (rpc_transport_t *this, dict_t *options)
priv->windowsize = (int)windowsize;
if (dict_get (this->options, "non-blocking-io")) {
optstr = data_to_str (dict_get (this->options,
"non-blocking-io"));
if (gf_string2boolean (optstr, &tmp_bool) == -1) {
gf_log (this->name, GF_LOG_ERROR,
"'non-blocking-io' takes only boolean options,"
" not taking any action");
tmp_bool = 1;
}
if (!tmp_bool) {
priv->bio = 1;
gf_log (this->name, GF_LOG_WARNING,
"disabling non-blocking IO");
}
}
if (!priv->bio) {
ret = __socket_nonblock (priv->sock);
if (ret == -1) {
gf_log (this->name, GF_LOG_WARNING,
"NBIO on %d failed (%s)",
priv->sock, strerror (errno));
goto out;
}
}
ret = 0;
out:
return ret;

84
tests/basic/quota-anon-fd-nfs.t Executable file
View File

@ -0,0 +1,84 @@
#!/bin/bash
. $(dirname $0)/../include.rc
. $(dirname $0)/../fileio.rc
cleanup;
TESTS_EXPECTED_IN_LOOP=16
TEST glusterd
TEST pidof glusterd
TEST $CLI volume info;
TEST $CLI volume create $V0 $H0:$B0/brick1;
EXPECT 'Created' volinfo_field $V0 'Status';
# The test makes use of inode-lru-limit to hit a scenario, where we
# find an inode whose ancestry is not there. Following is the
# hypothesis (which is confirmed by seeing logs indicating that
# codepath has been executed, but not through a good understanding of
# NFS internals).
# At the end of an fop, the reference count of an inode would be
# zero. The inode (and its ancestry) persists in memory only
# because of non-zero lookup count. These looked up inodes are put
# in an lru queue of size 1 (here). So, there can be at most one
# such inode in memory.
# NFS Server makes use of anonymous fds. So, if it cannot find
# valid fd, it does a nameless lookup. This gives us an inode
# whose ancestry is NULL. When a write happens on this inode,
# quota-enforcer/marker finds a NULL ancestry and asks
# storage/posix to build it.
TEST $CLI volume set $V0 network.inode-lru-limit 1
TEST $CLI volume set $V0 performance.nfs.write-behind off
TEST $CLI volume start $V0;
EXPECT 'Started' volinfo_field $V0 'Status';
TEST $CLI volume quota $V0 enable
TEST $CLI volume quota $V0 limit-usage / 1
TEST mount -t nfs -o noac,soft,nolock,vers=3 $H0:/$V0 $N0
deep=/0/1/2/3/4/5/6/7/8/9
TEST mkdir -p $N0/$deep
TEST touch $N0/$deep/file1 $N0/$deep/file2 $N0/$deep/file3 $N0/$deep/file4
TEST fd_open 3 'w' "$N0/$deep/file1"
TEST fd_open 4 'w' "$N0/$deep/file2"
TEST fd_open 5 'w' "$N0/$deep/file3"
TEST fd_open 6 'w' "$N0/$deep/file4"
# consume all quota
TEST ! dd if=/dev/zero of="$N0/$deep/file" bs=1MB count=1
# At the end of each fop in server, reference count of the
# inode associated with each of the file above drops to zero and hence
# put into lru queue. Since lru-limit is set to 1, an fop next file
# will displace the current inode from itable. This will ensure that
# when writes happens on same fd, fd resolution results in
# nameless lookup from server and quota_writev encounters an fd
# associated with an inode whose parent is not present in itable.
for j in $(seq 1 2); do
for i in $(seq 3 6); do
# failing writes indicate that we are enforcing quota set on /
# even with anonymous fds.
TEST_IN_LOOP ! fd_write $i "content"
TEST_IN_LOOP sync
done
done
exec 3>&-
exec 4>&-
exec 5>&-
exec 6>&-
$CLI volume statedump $V0 all
TEST umount -l $N0
cleanup;

View File

@ -1,46 +0,0 @@
#!/bin/bash
. $(dirname $0)/../include.rc
. $(dirname $0)/../volume.rc
cleanup;
TEST glusterd
TEST $CLI volume create $V0 $H0:$B0/${V0}{1}
function volinfo_field()
{
local vol=$1;
local field=$2;
$CLI volume info $vol | grep "^$field: " | sed 's/.*: //';
}
## Verify volume is is created
EXPECT "$V0" volinfo_field $V0 'Volume Name';
EXPECT 'Created' volinfo_field $V0 'Status';
## Start volume and verify
TEST $CLI volume start $V0;
EXPECT 'Started' volinfo_field $V0 'Status';
TEST $CLI volume quota $V0 enable;
## Mount NFS
TEST mount -t nfs -o nolock,soft,intr $H0:/$V0 $N0;
mkdir -p $N0/0/1
TEST $CLI volume quota $V0 limit-usage /0/1 1GB 75%;
deep=/0/1/2/3/4/5/6/7/8/9
mkdir -p $N0/$deep
dd if=/dev/zero of=$N0/$deep/file bs=1M count=502 &
kill_brick $V0 $H0 $B0/${V0}{1}
kill -TERM $(get_nfs_pid)
$CLI volume start $V0 force;
cleanup;

View File

@ -21,7 +21,7 @@ TEST $CLI volume set $V0 performance.read-ahead off
TEST $CLI volume set $V0 background-self-heal-count 0
TEST $CLI volume set $V0 self-heal-daemon off
TEST $CLI volume quota $V0 enable
sleep 5 # wait for brick to connect to quotad
TEST kill_brick $V0 $H0 $B0/${V0}0
TEST glusterfs --volfile-id=/$V0 --volfile-server=$H0 $M0 --attribute-timeout=0 --entry-timeout=0
cd $M0

View File

@ -19,7 +19,6 @@ TEST $CLI volume quota $V0 enable;
TEST MOUNTDIR="/tmp/$RANDOM"
TEST mkdir $MOUNTDIR
TEST glusterfs -s $H0 --volfile-id=$V0 $MOUNTDIR
sleep 10
function set_quota(){
mkdir "$MOUNTDIR/$name"

View File

@ -22,7 +22,6 @@ function __init()
TEST glusterfs --volfile-id=$V0 --volfile-server=$H0 $M0
TEST $CLI volume quota $V0 enable
sleep 15
}
#CASE-1

View File

@ -295,6 +295,37 @@ quota_enforcer_notify (struct rpc_clnt *rpc, void *mydata,
return ret;
}
int
quota_enforcer_blocking_connect (rpc_clnt_t *rpc)
{
dict_t *options = NULL;
int ret = -1;
options = dict_new ();
if (options == NULL)
goto out;
ret = dict_set_str (options, "non-blocking-io", "no");
if (ret)
goto out;
rpc->conn.trans->reconfigure (rpc->conn.trans, options);
rpc_clnt_start (rpc);
ret = dict_set_str (options, "non-blocking-io", "yes");
if (ret)
goto out;
rpc->conn.trans->reconfigure (rpc->conn.trans, options);
ret = 0;
out:
dict_unref (options);
return ret;
}
//Returns a started rpc_clnt. Creates a new rpc_clnt if quota_priv doesn't have
//one already
struct rpc_clnt *
@ -309,9 +340,13 @@ quota_enforcer_init (xlator_t *this, dict_t *options)
gf_log (this->name, GF_LOG_TRACE, "quota enforcer clnt already "
"inited");
//Turns out to be a NOP if the clnt is already connected.
rpc_clnt_start (priv->rpc_clnt);
ret = quota_enforcer_blocking_connect (priv->rpc_clnt);
if (ret)
goto out;
return priv->rpc_clnt;
}
priv->quota_enforcer = &quota_enforcer_clnt;
ret = dict_set_str (options, "transport.address-family", "unix");
@ -339,7 +374,11 @@ quota_enforcer_init (xlator_t *this, dict_t *options)
goto out;
}
rpc_clnt_start (rpc);
ret = quota_enforcer_blocking_connect (rpc);
if (ret)
goto out;
ret = 0;
out:
if (ret) {
if (rpc)

View File

@ -4177,14 +4177,15 @@ err:
int
reconfigure (xlator_t *this, dict_t *options)
{
int32_t ret = -1;
quota_priv_t *priv = NULL;
int32_t ret = -1;
quota_priv_t *priv = NULL;
gf_boolean_t quota_on = _gf_false;
priv = this->private;
GF_OPTION_RECONF ("deem-statfs", priv->consider_statfs, options, bool,
out);
GF_OPTION_RECONF ("server-quota", priv->is_quota_on, options, bool,
GF_OPTION_RECONF ("server-quota", quota_on, options, bool,
out);
GF_OPTION_RECONF ("default-soft-limit", priv->default_soft_lim,
options, percent, out);
@ -4195,7 +4196,7 @@ reconfigure (xlator_t *this, dict_t *options)
GF_OPTION_RECONF ("hard-timeout", priv->hard_timeout, options,
time, out);
if (priv->is_quota_on) {
if (quota_on) {
priv->rpc_clnt = quota_enforcer_init (this,
this->options);
if (priv->rpc_clnt == NULL) {
@ -4216,6 +4217,8 @@ reconfigure (xlator_t *this, dict_t *options)
}
}
priv->is_quota_on = quota_on;
ret = 0;
out:
return ret;

View File

@ -996,14 +996,6 @@ glusterd_quotad_op (int opcode)
ret = glusterd_check_generate_start_quotad ();
break;
case GF_QUOTA_OPTION_TYPE_DEFAULT_SOFT_LIMIT:
case GF_QUOTA_OPTION_TYPE_HARD_TIMEOUT:
case GF_QUOTA_OPTION_TYPE_SOFT_TIMEOUT:
case GF_QUOTA_OPTION_TYPE_ALERT_TIME:
ret = glusterd_reconfigure_quotad ();
break;
default:
ret = 0;
break;
@ -1131,6 +1123,12 @@ glusterd_op_quota (dict_t *dict, char **op_errstr, dict_t *rsp_dict)
goto out;
}
if (priv->op_version > GD_OP_VERSION_MIN) {
ret = glusterd_quotad_op (type);
if (ret)
goto out;
}
ret = glusterd_create_volfiles_and_notify_services (volinfo);
if (ret) {
gf_log (this->name, GF_LOG_ERROR, "Unable to re-create "
@ -1151,11 +1149,6 @@ glusterd_op_quota (dict_t *dict, char **op_errstr, dict_t *rsp_dict)
if (rsp_dict && start_crawl == _gf_true)
glusterd_quota_initiate_fs_crawl (priv, volname, type);
if (priv->op_version > GD_OP_VERSION_MIN) {
ret = glusterd_quotad_op (type);
if (ret)
goto out;
}
ret = 0;
out:
return ret;

View File

@ -3965,7 +3965,7 @@ glusterd_nodesvc_disconnect (char *server)
}
int32_t
glusterd_nodesvc_start (char *server)
glusterd_nodesvc_start (char *server, gf_boolean_t wait)
{
int32_t ret = -1;
xlator_t *this = NULL;
@ -4051,7 +4051,16 @@ glusterd_nodesvc_start (char *server)
runner_log (&runner, "", GF_LOG_DEBUG,
"Starting the nfs/glustershd services");
ret = runner_run_nowait (&runner);
if (!wait) {
ret = runner_run_nowait (&runner);
} else {
synclock_unlock (&priv->big_lock);
{
ret = runner_run (&runner);
}
synclock_lock (&priv->big_lock);
}
if (ret == 0) {
glusterd_nodesvc_connect (server, sockfpath);
}
@ -4062,19 +4071,19 @@ out:
int
glusterd_nfs_server_start ()
{
return glusterd_nodesvc_start ("nfs");
return glusterd_nodesvc_start ("nfs", _gf_false);
}
int
glusterd_shd_start ()
{
return glusterd_nodesvc_start ("glustershd");
return glusterd_nodesvc_start ("glustershd", _gf_false);
}
int
glusterd_quotad_start ()
{
return glusterd_nodesvc_start ("quotad");
return glusterd_nodesvc_start ("quotad", _gf_true);
}
gf_boolean_t