mirror of
https://github.com/samba-team/samba.git
synced 2024-12-24 21:34:56 +03:00
b2de5a81bf
Add one use of popt_set_cmdline_credentials(). Fix 80 column limits when cmdline_credentials changes to popt_get_cmdline_credentials(). Signed-off-by: Jeremy Allison <jra@samba.org> Reviewed-by: Andreas Schneider <asn@samba.org>
448 lines
12 KiB
C
448 lines
12 KiB
C
/*
|
|
Unix SMB/CIFS implementation.
|
|
|
|
locking benchmark
|
|
|
|
Copyright (C) Andrew Tridgell 2006
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 3 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "includes.h"
|
|
#include "libcli/raw/libcliraw.h"
|
|
#include "libcli/raw/raw_proto.h"
|
|
#include "system/time.h"
|
|
#include "system/filesys.h"
|
|
#include "libcli/libcli.h"
|
|
#include "torture/util.h"
|
|
#include "lib/events/events.h"
|
|
#include "lib/cmdline/popt_common.h"
|
|
#include "libcli/composite/composite.h"
|
|
#include "libcli/smb_composite/smb_composite.h"
|
|
#include "libcli/resolve/resolve.h"
|
|
#include "param/param.h"
|
|
#include "torture/raw/proto.h"
|
|
#include "libcli/smb/smbXcli_base.h"
|
|
#include "../lib/util/util_net.h"
|
|
|
|
#define BASEDIR "\\benchlock"
|
|
#define FNAME BASEDIR "\\lock.dat"
|
|
|
|
static int nprocs;
|
|
static int lock_failed;
|
|
static int num_connected;
|
|
|
|
enum lock_stage {LOCK_INITIAL, LOCK_LOCK, LOCK_UNLOCK};
|
|
|
|
struct benchlock_state {
|
|
struct torture_context *tctx;
|
|
struct tevent_context *ev;
|
|
struct smbcli_tree *tree;
|
|
TALLOC_CTX *mem_ctx;
|
|
int client_num;
|
|
int fnum;
|
|
enum lock_stage stage;
|
|
int lock_offset;
|
|
int unlock_offset;
|
|
int count;
|
|
int lastcount;
|
|
struct smbcli_request *req;
|
|
struct smb_composite_connect reconnect;
|
|
struct tevent_timer *te;
|
|
|
|
/* these are used for reconnections */
|
|
const char **dest_ports;
|
|
const char *dest_host;
|
|
const char *called_name;
|
|
const char *service_type;
|
|
};
|
|
|
|
static void lock_completion(struct smbcli_request *);
|
|
|
|
/*
|
|
send the next lock request
|
|
*/
|
|
static void lock_send(struct benchlock_state *state)
|
|
{
|
|
union smb_lock io;
|
|
struct smb_lock_entry lock;
|
|
|
|
switch (state->stage) {
|
|
case LOCK_INITIAL:
|
|
io.lockx.in.ulock_cnt = 0;
|
|
io.lockx.in.lock_cnt = 1;
|
|
state->lock_offset = 0;
|
|
state->unlock_offset = 0;
|
|
lock.offset = state->lock_offset;
|
|
break;
|
|
case LOCK_LOCK:
|
|
io.lockx.in.ulock_cnt = 0;
|
|
io.lockx.in.lock_cnt = 1;
|
|
state->lock_offset = (state->lock_offset+1)%(nprocs+1);
|
|
lock.offset = state->lock_offset;
|
|
break;
|
|
case LOCK_UNLOCK:
|
|
io.lockx.in.ulock_cnt = 1;
|
|
io.lockx.in.lock_cnt = 0;
|
|
lock.offset = state->unlock_offset;
|
|
state->unlock_offset = (state->unlock_offset+1)%(nprocs+1);
|
|
break;
|
|
}
|
|
|
|
lock.count = 1;
|
|
lock.pid = state->tree->session->pid;
|
|
|
|
io.lockx.level = RAW_LOCK_LOCKX;
|
|
io.lockx.in.mode = LOCKING_ANDX_LARGE_FILES;
|
|
io.lockx.in.timeout = 100000;
|
|
io.lockx.in.locks = &lock;
|
|
io.lockx.in.file.fnum = state->fnum;
|
|
|
|
state->req = smb_raw_lock_send(state->tree, &io);
|
|
if (state->req == NULL) {
|
|
DEBUG(0,("Failed to setup lock\n"));
|
|
lock_failed++;
|
|
}
|
|
state->req->async.private_data = state;
|
|
state->req->async.fn = lock_completion;
|
|
}
|
|
|
|
static void reopen_connection(struct tevent_context *ev, struct tevent_timer *te,
|
|
struct timeval t, void *private_data);
|
|
|
|
|
|
static void reopen_file(struct tevent_context *ev, struct tevent_timer *te,
|
|
struct timeval t, void *private_data)
|
|
{
|
|
struct benchlock_state *state = (struct benchlock_state *)private_data;
|
|
|
|
/* reestablish our open file */
|
|
state->fnum = smbcli_open(state->tree, FNAME, O_RDWR|O_CREAT, DENY_NONE);
|
|
if (state->fnum == -1) {
|
|
printf("Failed to open %s on connection %d\n", FNAME, state->client_num);
|
|
exit(1);
|
|
}
|
|
|
|
num_connected++;
|
|
|
|
DEBUG(0,("reconnect to %s finished (%u connected)\n", state->dest_host,
|
|
num_connected));
|
|
|
|
state->stage = LOCK_INITIAL;
|
|
lock_send(state);
|
|
}
|
|
|
|
/*
|
|
complete an async reconnect
|
|
*/
|
|
static void reopen_connection_complete(struct composite_context *ctx)
|
|
{
|
|
struct benchlock_state *state = (struct benchlock_state *)ctx->async.private_data;
|
|
NTSTATUS status;
|
|
struct smb_composite_connect *io = &state->reconnect;
|
|
|
|
status = smb_composite_connect_recv(ctx, state->mem_ctx);
|
|
if (!NT_STATUS_IS_OK(status)) {
|
|
talloc_free(state->te);
|
|
state->te = tevent_add_timer(state->ev, state->mem_ctx,
|
|
timeval_current_ofs(1,0),
|
|
reopen_connection, state);
|
|
return;
|
|
}
|
|
|
|
talloc_free(state->tree);
|
|
state->tree = io->out.tree;
|
|
|
|
/* do the reopen as a separate event */
|
|
tevent_add_timer(state->ev, state->mem_ctx, timeval_zero(), reopen_file, state);
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
reopen a connection
|
|
*/
|
|
static void reopen_connection(struct tevent_context *ev, struct tevent_timer *te,
|
|
struct timeval t, void *private_data)
|
|
{
|
|
struct benchlock_state *state = (struct benchlock_state *)private_data;
|
|
struct composite_context *ctx;
|
|
struct smb_composite_connect *io = &state->reconnect;
|
|
char *host, *share;
|
|
|
|
state->te = NULL;
|
|
|
|
if (!torture_get_conn_index(state->client_num, state->mem_ctx, state->tctx, &host, &share)) {
|
|
DEBUG(0,("Can't find host/share for reconnect?!\n"));
|
|
exit(1);
|
|
}
|
|
|
|
io->in.dest_host = state->dest_host;
|
|
io->in.dest_ports = state->dest_ports;
|
|
io->in.gensec_settings = lpcfg_gensec_settings(state->mem_ctx, state->tctx->lp_ctx);
|
|
io->in.socket_options = lpcfg_socket_options(state->tctx->lp_ctx);
|
|
io->in.called_name = state->called_name;
|
|
io->in.service = share;
|
|
io->in.service_type = state->service_type;
|
|
io->in.credentials = popt_get_cmdline_credentials();
|
|
io->in.fallback_to_anonymous = false;
|
|
io->in.workgroup = lpcfg_workgroup(state->tctx->lp_ctx);
|
|
lpcfg_smbcli_options(state->tctx->lp_ctx, &io->in.options);
|
|
lpcfg_smbcli_session_options(state->tctx->lp_ctx, &io->in.session_options);
|
|
|
|
/* kill off the remnants of the old connection */
|
|
talloc_free(state->tree);
|
|
state->tree = NULL;
|
|
|
|
ctx = smb_composite_connect_send(io, state->mem_ctx,
|
|
lpcfg_resolve_context(state->tctx->lp_ctx),
|
|
state->ev);
|
|
if (ctx == NULL) {
|
|
DEBUG(0,("Failed to setup async reconnect\n"));
|
|
exit(1);
|
|
}
|
|
|
|
ctx->async.fn = reopen_connection_complete;
|
|
ctx->async.private_data = state;
|
|
}
|
|
|
|
|
|
/*
|
|
called when a lock completes
|
|
*/
|
|
static void lock_completion(struct smbcli_request *req)
|
|
{
|
|
struct benchlock_state *state = (struct benchlock_state *)req->async.private_data;
|
|
NTSTATUS status = smbcli_request_simple_recv(req);
|
|
state->req = NULL;
|
|
if (!NT_STATUS_IS_OK(status)) {
|
|
if (NT_STATUS_EQUAL(status, NT_STATUS_END_OF_FILE) ||
|
|
NT_STATUS_EQUAL(status, NT_STATUS_LOCAL_DISCONNECT) ||
|
|
NT_STATUS_EQUAL(status, NT_STATUS_CONNECTION_RESET)) {
|
|
talloc_free(state->tree);
|
|
state->tree = NULL;
|
|
num_connected--;
|
|
DEBUG(0,("reopening connection to %s\n", state->dest_host));
|
|
talloc_free(state->te);
|
|
state->te = tevent_add_timer(state->ev, state->mem_ctx,
|
|
timeval_current_ofs(1,0),
|
|
reopen_connection, state);
|
|
} else {
|
|
DEBUG(0,("Lock failed - %s\n", nt_errstr(status)));
|
|
lock_failed++;
|
|
}
|
|
return;
|
|
}
|
|
|
|
switch (state->stage) {
|
|
case LOCK_INITIAL:
|
|
state->stage = LOCK_LOCK;
|
|
break;
|
|
case LOCK_LOCK:
|
|
state->stage = LOCK_UNLOCK;
|
|
break;
|
|
case LOCK_UNLOCK:
|
|
state->stage = LOCK_LOCK;
|
|
break;
|
|
}
|
|
|
|
state->count++;
|
|
lock_send(state);
|
|
}
|
|
|
|
|
|
static void echo_completion(struct smbcli_request *req)
|
|
{
|
|
struct benchlock_state *state = (struct benchlock_state *)req->async.private_data;
|
|
NTSTATUS status = smbcli_request_simple_recv(req);
|
|
if (NT_STATUS_EQUAL(status, NT_STATUS_END_OF_FILE) ||
|
|
NT_STATUS_EQUAL(status, NT_STATUS_LOCAL_DISCONNECT) ||
|
|
NT_STATUS_EQUAL(status, NT_STATUS_CONNECTION_RESET)) {
|
|
talloc_free(state->tree);
|
|
state->tree = NULL;
|
|
num_connected--;
|
|
DEBUG(0,("reopening connection to %s\n", state->dest_host));
|
|
talloc_free(state->te);
|
|
state->te = tevent_add_timer(state->ev, state->mem_ctx,
|
|
timeval_current_ofs(1,0),
|
|
reopen_connection, state);
|
|
}
|
|
}
|
|
|
|
static void report_rate(struct tevent_context *ev, struct tevent_timer *te,
|
|
struct timeval t, void *private_data)
|
|
{
|
|
struct benchlock_state *state = talloc_get_type(private_data,
|
|
struct benchlock_state);
|
|
int i;
|
|
for (i=0;i<nprocs;i++) {
|
|
printf("%5u ", (unsigned)(state[i].count - state[i].lastcount));
|
|
state[i].lastcount = state[i].count;
|
|
}
|
|
printf("\r");
|
|
fflush(stdout);
|
|
tevent_add_timer(ev, state, timeval_current_ofs(1, 0), report_rate, state);
|
|
|
|
/* send an echo on each interface to ensure it stays alive - this helps
|
|
with IP takeover */
|
|
for (i=0;i<nprocs;i++) {
|
|
struct smb_echo p;
|
|
struct smbcli_request *req;
|
|
|
|
if (!state[i].tree) {
|
|
continue;
|
|
}
|
|
|
|
p.in.repeat_count = 1;
|
|
p.in.size = 0;
|
|
p.in.data = NULL;
|
|
req = smb_raw_echo_send(state[i].tree->session->transport, &p);
|
|
req->async.private_data = &state[i];
|
|
req->async.fn = echo_completion;
|
|
}
|
|
}
|
|
|
|
/*
|
|
benchmark locking calls
|
|
*/
|
|
bool torture_bench_lock(struct torture_context *torture)
|
|
{
|
|
bool ret = true;
|
|
TALLOC_CTX *mem_ctx = talloc_new(torture);
|
|
int i, j;
|
|
int timelimit = torture_setting_int(torture, "timelimit", 10);
|
|
struct timeval tv;
|
|
struct benchlock_state *state;
|
|
int total = 0, minops=0;
|
|
struct smbcli_state *cli;
|
|
bool progress;
|
|
off_t offset;
|
|
int initial_locks = torture_setting_int(torture, "initial_locks", 0);
|
|
|
|
progress = torture_setting_bool(torture, "progress", true);
|
|
|
|
nprocs = torture_setting_int(torture, "nprocs", 4);
|
|
|
|
state = talloc_zero_array(mem_ctx, struct benchlock_state, nprocs);
|
|
|
|
printf("Opening %d connections\n", nprocs);
|
|
for (i=0;i<nprocs;i++) {
|
|
const struct sockaddr_storage *dest_ss;
|
|
char addrstr[INET6_ADDRSTRLEN];
|
|
const char *dest_str;
|
|
uint16_t dest_port;
|
|
|
|
state[i].tctx = torture;
|
|
state[i].mem_ctx = talloc_new(state);
|
|
state[i].client_num = i;
|
|
state[i].ev = torture->ev;
|
|
if (!torture_open_connection_ev(&cli, i, torture, torture->ev)) {
|
|
return false;
|
|
}
|
|
talloc_steal(state[i].mem_ctx, cli);
|
|
state[i].tree = cli->tree;
|
|
|
|
dest_ss = smbXcli_conn_remote_sockaddr(
|
|
state[i].tree->session->transport->conn);
|
|
dest_str = print_sockaddr(addrstr, sizeof(addrstr), dest_ss);
|
|
dest_port = get_sockaddr_port(dest_ss);
|
|
|
|
state[i].dest_host = talloc_strdup(state[i].mem_ctx, dest_str);
|
|
state[i].dest_ports = talloc_array(state[i].mem_ctx,
|
|
const char *, 2);
|
|
state[i].dest_ports[0] = talloc_asprintf(state[i].dest_ports,
|
|
"%u", dest_port);
|
|
state[i].dest_ports[1] = NULL;
|
|
state[i].called_name = talloc_strdup(state[i].mem_ctx,
|
|
smbXcli_conn_remote_name(cli->tree->session->transport->conn));
|
|
state[i].service_type = talloc_strdup(state[i].mem_ctx, "?????");
|
|
}
|
|
|
|
num_connected = i;
|
|
|
|
if (!torture_setup_dir(cli, BASEDIR)) {
|
|
goto failed;
|
|
}
|
|
|
|
for (i=0;i<nprocs;i++) {
|
|
state[i].fnum = smbcli_open(state[i].tree,
|
|
FNAME,
|
|
O_RDWR|O_CREAT, DENY_NONE);
|
|
if (state[i].fnum == -1) {
|
|
printf("Failed to open %s on connection %d\n", FNAME, i);
|
|
goto failed;
|
|
}
|
|
|
|
/* Optionally, lock initial_locks for each proc beforehand. */
|
|
if (i == 0 && initial_locks > 0) {
|
|
printf("Initializing %d locks on each proc.\n",
|
|
initial_locks);
|
|
}
|
|
|
|
for (j = 0; j < initial_locks; j++) {
|
|
offset = (0xFFFFFED8LLU * (i+2)) + j;
|
|
if (!NT_STATUS_IS_OK(smbcli_lock64(state[i].tree,
|
|
state[i].fnum, offset, 1, 0, WRITE_LOCK))) {
|
|
printf("Failed initializing, lock=%d\n", j);
|
|
goto failed;
|
|
}
|
|
}
|
|
|
|
state[i].stage = LOCK_INITIAL;
|
|
lock_send(&state[i]);
|
|
}
|
|
|
|
tv = timeval_current();
|
|
|
|
if (progress) {
|
|
tevent_add_timer(torture->ev, state, timeval_current_ofs(1, 0), report_rate, state);
|
|
}
|
|
|
|
printf("Running for %d seconds\n", timelimit);
|
|
while (timeval_elapsed(&tv) < timelimit) {
|
|
tevent_loop_once(torture->ev);
|
|
|
|
if (lock_failed) {
|
|
DEBUG(0,("locking failed\n"));
|
|
goto failed;
|
|
}
|
|
}
|
|
|
|
printf("%.2f ops/second\n", total/timeval_elapsed(&tv));
|
|
minops = state[0].count;
|
|
for (i=0;i<nprocs;i++) {
|
|
printf("[%d] %u ops\n", i, state[i].count);
|
|
if (state[i].count < minops) minops = state[i].count;
|
|
}
|
|
if (minops < 0.5*total/nprocs) {
|
|
printf("Failed: unbalanced locking\n");
|
|
goto failed;
|
|
}
|
|
|
|
for (i=0;i<nprocs;i++) {
|
|
talloc_free(state[i].req);
|
|
smb_raw_exit(state[i].tree->session);
|
|
}
|
|
|
|
smbcli_deltree(state[0].tree, BASEDIR);
|
|
talloc_free(mem_ctx);
|
|
printf("\n");
|
|
return ret;
|
|
|
|
failed:
|
|
smbcli_deltree(state[0].tree, BASEDIR);
|
|
talloc_free(mem_ctx);
|
|
return false;
|
|
}
|