1
0
mirror of https://github.com/samba-team/samba.git synced 2024-12-24 21:34:56 +03:00
samba-mirror/source4/torture/raw/lockbench.c
Jeremy Allison b2de5a81bf s4: popt: Global replace of cmdline_credentials -> popt_get_cmdline_credentials().
Add one use of popt_set_cmdline_credentials().
Fix 80 column limits when cmdline_credentials changes
to popt_get_cmdline_credentials().

Signed-off-by: Jeremy Allison <jra@samba.org>
Reviewed-by: Andreas Schneider <asn@samba.org>
2017-05-11 20:30:13 +02:00

448 lines
12 KiB
C

/*
Unix SMB/CIFS implementation.
locking benchmark
Copyright (C) Andrew Tridgell 2006
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
#include "includes.h"
#include "libcli/raw/libcliraw.h"
#include "libcli/raw/raw_proto.h"
#include "system/time.h"
#include "system/filesys.h"
#include "libcli/libcli.h"
#include "torture/util.h"
#include "lib/events/events.h"
#include "lib/cmdline/popt_common.h"
#include "libcli/composite/composite.h"
#include "libcli/smb_composite/smb_composite.h"
#include "libcli/resolve/resolve.h"
#include "param/param.h"
#include "torture/raw/proto.h"
#include "libcli/smb/smbXcli_base.h"
#include "../lib/util/util_net.h"
#define BASEDIR "\\benchlock"
#define FNAME BASEDIR "\\lock.dat"
static int nprocs;
static int lock_failed;
static int num_connected;
enum lock_stage {LOCK_INITIAL, LOCK_LOCK, LOCK_UNLOCK};
struct benchlock_state {
struct torture_context *tctx;
struct tevent_context *ev;
struct smbcli_tree *tree;
TALLOC_CTX *mem_ctx;
int client_num;
int fnum;
enum lock_stage stage;
int lock_offset;
int unlock_offset;
int count;
int lastcount;
struct smbcli_request *req;
struct smb_composite_connect reconnect;
struct tevent_timer *te;
/* these are used for reconnections */
const char **dest_ports;
const char *dest_host;
const char *called_name;
const char *service_type;
};
static void lock_completion(struct smbcli_request *);
/*
send the next lock request
*/
static void lock_send(struct benchlock_state *state)
{
union smb_lock io;
struct smb_lock_entry lock;
switch (state->stage) {
case LOCK_INITIAL:
io.lockx.in.ulock_cnt = 0;
io.lockx.in.lock_cnt = 1;
state->lock_offset = 0;
state->unlock_offset = 0;
lock.offset = state->lock_offset;
break;
case LOCK_LOCK:
io.lockx.in.ulock_cnt = 0;
io.lockx.in.lock_cnt = 1;
state->lock_offset = (state->lock_offset+1)%(nprocs+1);
lock.offset = state->lock_offset;
break;
case LOCK_UNLOCK:
io.lockx.in.ulock_cnt = 1;
io.lockx.in.lock_cnt = 0;
lock.offset = state->unlock_offset;
state->unlock_offset = (state->unlock_offset+1)%(nprocs+1);
break;
}
lock.count = 1;
lock.pid = state->tree->session->pid;
io.lockx.level = RAW_LOCK_LOCKX;
io.lockx.in.mode = LOCKING_ANDX_LARGE_FILES;
io.lockx.in.timeout = 100000;
io.lockx.in.locks = &lock;
io.lockx.in.file.fnum = state->fnum;
state->req = smb_raw_lock_send(state->tree, &io);
if (state->req == NULL) {
DEBUG(0,("Failed to setup lock\n"));
lock_failed++;
}
state->req->async.private_data = state;
state->req->async.fn = lock_completion;
}
static void reopen_connection(struct tevent_context *ev, struct tevent_timer *te,
struct timeval t, void *private_data);
static void reopen_file(struct tevent_context *ev, struct tevent_timer *te,
struct timeval t, void *private_data)
{
struct benchlock_state *state = (struct benchlock_state *)private_data;
/* reestablish our open file */
state->fnum = smbcli_open(state->tree, FNAME, O_RDWR|O_CREAT, DENY_NONE);
if (state->fnum == -1) {
printf("Failed to open %s on connection %d\n", FNAME, state->client_num);
exit(1);
}
num_connected++;
DEBUG(0,("reconnect to %s finished (%u connected)\n", state->dest_host,
num_connected));
state->stage = LOCK_INITIAL;
lock_send(state);
}
/*
complete an async reconnect
*/
static void reopen_connection_complete(struct composite_context *ctx)
{
struct benchlock_state *state = (struct benchlock_state *)ctx->async.private_data;
NTSTATUS status;
struct smb_composite_connect *io = &state->reconnect;
status = smb_composite_connect_recv(ctx, state->mem_ctx);
if (!NT_STATUS_IS_OK(status)) {
talloc_free(state->te);
state->te = tevent_add_timer(state->ev, state->mem_ctx,
timeval_current_ofs(1,0),
reopen_connection, state);
return;
}
talloc_free(state->tree);
state->tree = io->out.tree;
/* do the reopen as a separate event */
tevent_add_timer(state->ev, state->mem_ctx, timeval_zero(), reopen_file, state);
}
/*
reopen a connection
*/
static void reopen_connection(struct tevent_context *ev, struct tevent_timer *te,
struct timeval t, void *private_data)
{
struct benchlock_state *state = (struct benchlock_state *)private_data;
struct composite_context *ctx;
struct smb_composite_connect *io = &state->reconnect;
char *host, *share;
state->te = NULL;
if (!torture_get_conn_index(state->client_num, state->mem_ctx, state->tctx, &host, &share)) {
DEBUG(0,("Can't find host/share for reconnect?!\n"));
exit(1);
}
io->in.dest_host = state->dest_host;
io->in.dest_ports = state->dest_ports;
io->in.gensec_settings = lpcfg_gensec_settings(state->mem_ctx, state->tctx->lp_ctx);
io->in.socket_options = lpcfg_socket_options(state->tctx->lp_ctx);
io->in.called_name = state->called_name;
io->in.service = share;
io->in.service_type = state->service_type;
io->in.credentials = popt_get_cmdline_credentials();
io->in.fallback_to_anonymous = false;
io->in.workgroup = lpcfg_workgroup(state->tctx->lp_ctx);
lpcfg_smbcli_options(state->tctx->lp_ctx, &io->in.options);
lpcfg_smbcli_session_options(state->tctx->lp_ctx, &io->in.session_options);
/* kill off the remnants of the old connection */
talloc_free(state->tree);
state->tree = NULL;
ctx = smb_composite_connect_send(io, state->mem_ctx,
lpcfg_resolve_context(state->tctx->lp_ctx),
state->ev);
if (ctx == NULL) {
DEBUG(0,("Failed to setup async reconnect\n"));
exit(1);
}
ctx->async.fn = reopen_connection_complete;
ctx->async.private_data = state;
}
/*
called when a lock completes
*/
static void lock_completion(struct smbcli_request *req)
{
struct benchlock_state *state = (struct benchlock_state *)req->async.private_data;
NTSTATUS status = smbcli_request_simple_recv(req);
state->req = NULL;
if (!NT_STATUS_IS_OK(status)) {
if (NT_STATUS_EQUAL(status, NT_STATUS_END_OF_FILE) ||
NT_STATUS_EQUAL(status, NT_STATUS_LOCAL_DISCONNECT) ||
NT_STATUS_EQUAL(status, NT_STATUS_CONNECTION_RESET)) {
talloc_free(state->tree);
state->tree = NULL;
num_connected--;
DEBUG(0,("reopening connection to %s\n", state->dest_host));
talloc_free(state->te);
state->te = tevent_add_timer(state->ev, state->mem_ctx,
timeval_current_ofs(1,0),
reopen_connection, state);
} else {
DEBUG(0,("Lock failed - %s\n", nt_errstr(status)));
lock_failed++;
}
return;
}
switch (state->stage) {
case LOCK_INITIAL:
state->stage = LOCK_LOCK;
break;
case LOCK_LOCK:
state->stage = LOCK_UNLOCK;
break;
case LOCK_UNLOCK:
state->stage = LOCK_LOCK;
break;
}
state->count++;
lock_send(state);
}
static void echo_completion(struct smbcli_request *req)
{
struct benchlock_state *state = (struct benchlock_state *)req->async.private_data;
NTSTATUS status = smbcli_request_simple_recv(req);
if (NT_STATUS_EQUAL(status, NT_STATUS_END_OF_FILE) ||
NT_STATUS_EQUAL(status, NT_STATUS_LOCAL_DISCONNECT) ||
NT_STATUS_EQUAL(status, NT_STATUS_CONNECTION_RESET)) {
talloc_free(state->tree);
state->tree = NULL;
num_connected--;
DEBUG(0,("reopening connection to %s\n", state->dest_host));
talloc_free(state->te);
state->te = tevent_add_timer(state->ev, state->mem_ctx,
timeval_current_ofs(1,0),
reopen_connection, state);
}
}
static void report_rate(struct tevent_context *ev, struct tevent_timer *te,
struct timeval t, void *private_data)
{
struct benchlock_state *state = talloc_get_type(private_data,
struct benchlock_state);
int i;
for (i=0;i<nprocs;i++) {
printf("%5u ", (unsigned)(state[i].count - state[i].lastcount));
state[i].lastcount = state[i].count;
}
printf("\r");
fflush(stdout);
tevent_add_timer(ev, state, timeval_current_ofs(1, 0), report_rate, state);
/* send an echo on each interface to ensure it stays alive - this helps
with IP takeover */
for (i=0;i<nprocs;i++) {
struct smb_echo p;
struct smbcli_request *req;
if (!state[i].tree) {
continue;
}
p.in.repeat_count = 1;
p.in.size = 0;
p.in.data = NULL;
req = smb_raw_echo_send(state[i].tree->session->transport, &p);
req->async.private_data = &state[i];
req->async.fn = echo_completion;
}
}
/*
benchmark locking calls
*/
bool torture_bench_lock(struct torture_context *torture)
{
bool ret = true;
TALLOC_CTX *mem_ctx = talloc_new(torture);
int i, j;
int timelimit = torture_setting_int(torture, "timelimit", 10);
struct timeval tv;
struct benchlock_state *state;
int total = 0, minops=0;
struct smbcli_state *cli;
bool progress;
off_t offset;
int initial_locks = torture_setting_int(torture, "initial_locks", 0);
progress = torture_setting_bool(torture, "progress", true);
nprocs = torture_setting_int(torture, "nprocs", 4);
state = talloc_zero_array(mem_ctx, struct benchlock_state, nprocs);
printf("Opening %d connections\n", nprocs);
for (i=0;i<nprocs;i++) {
const struct sockaddr_storage *dest_ss;
char addrstr[INET6_ADDRSTRLEN];
const char *dest_str;
uint16_t dest_port;
state[i].tctx = torture;
state[i].mem_ctx = talloc_new(state);
state[i].client_num = i;
state[i].ev = torture->ev;
if (!torture_open_connection_ev(&cli, i, torture, torture->ev)) {
return false;
}
talloc_steal(state[i].mem_ctx, cli);
state[i].tree = cli->tree;
dest_ss = smbXcli_conn_remote_sockaddr(
state[i].tree->session->transport->conn);
dest_str = print_sockaddr(addrstr, sizeof(addrstr), dest_ss);
dest_port = get_sockaddr_port(dest_ss);
state[i].dest_host = talloc_strdup(state[i].mem_ctx, dest_str);
state[i].dest_ports = talloc_array(state[i].mem_ctx,
const char *, 2);
state[i].dest_ports[0] = talloc_asprintf(state[i].dest_ports,
"%u", dest_port);
state[i].dest_ports[1] = NULL;
state[i].called_name = talloc_strdup(state[i].mem_ctx,
smbXcli_conn_remote_name(cli->tree->session->transport->conn));
state[i].service_type = talloc_strdup(state[i].mem_ctx, "?????");
}
num_connected = i;
if (!torture_setup_dir(cli, BASEDIR)) {
goto failed;
}
for (i=0;i<nprocs;i++) {
state[i].fnum = smbcli_open(state[i].tree,
FNAME,
O_RDWR|O_CREAT, DENY_NONE);
if (state[i].fnum == -1) {
printf("Failed to open %s on connection %d\n", FNAME, i);
goto failed;
}
/* Optionally, lock initial_locks for each proc beforehand. */
if (i == 0 && initial_locks > 0) {
printf("Initializing %d locks on each proc.\n",
initial_locks);
}
for (j = 0; j < initial_locks; j++) {
offset = (0xFFFFFED8LLU * (i+2)) + j;
if (!NT_STATUS_IS_OK(smbcli_lock64(state[i].tree,
state[i].fnum, offset, 1, 0, WRITE_LOCK))) {
printf("Failed initializing, lock=%d\n", j);
goto failed;
}
}
state[i].stage = LOCK_INITIAL;
lock_send(&state[i]);
}
tv = timeval_current();
if (progress) {
tevent_add_timer(torture->ev, state, timeval_current_ofs(1, 0), report_rate, state);
}
printf("Running for %d seconds\n", timelimit);
while (timeval_elapsed(&tv) < timelimit) {
tevent_loop_once(torture->ev);
if (lock_failed) {
DEBUG(0,("locking failed\n"));
goto failed;
}
}
printf("%.2f ops/second\n", total/timeval_elapsed(&tv));
minops = state[0].count;
for (i=0;i<nprocs;i++) {
printf("[%d] %u ops\n", i, state[i].count);
if (state[i].count < minops) minops = state[i].count;
}
if (minops < 0.5*total/nprocs) {
printf("Failed: unbalanced locking\n");
goto failed;
}
for (i=0;i<nprocs;i++) {
talloc_free(state[i].req);
smb_raw_exit(state[i].tree->session);
}
smbcli_deltree(state[0].tree, BASEDIR);
talloc_free(mem_ctx);
printf("\n");
return ret;
failed:
smbcli_deltree(state[0].tree, BASEDIR);
talloc_free(mem_ctx);
return false;
}