1
0
mirror of https://github.com/samba-team/samba.git synced 2024-12-23 17:34:34 +03:00
samba-mirror/source3/smbd/smb2_lock.c
Volker Lendecke 246cb1961f smbd: Fix CID 1273096 Dereference before null check
Signed-off-by: Volker Lendecke <vl@samba.org>
Reviewed-by: Jeremy Allison <jra@samba.org>
2015-06-23 22:12:09 +02:00

908 lines
23 KiB
C

/*
Unix SMB/CIFS implementation.
Core SMB2 server
Copyright (C) Stefan Metzmacher 2009
Copyright (C) Jeremy Allison 2010
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
#include "includes.h"
#include "smbd/smbd.h"
#include "smbd/globals.h"
#include "../libcli/smb/smb_common.h"
#include "../lib/util/tevent_ntstatus.h"
#include "messages.h"
struct smbd_smb2_lock_element {
uint64_t offset;
uint64_t length;
uint32_t flags;
};
struct smbd_smb2_lock_state {
struct smbd_smb2_request *smb2req;
struct smb_request *smb1req;
struct blocking_lock_record *blr;
uint16_t lock_count;
struct smbd_lock_element *locks;
};
static void remove_pending_lock(struct smbd_smb2_lock_state *state,
struct blocking_lock_record *blr);
static struct tevent_req *smbd_smb2_lock_send(TALLOC_CTX *mem_ctx,
struct tevent_context *ev,
struct smbd_smb2_request *smb2req,
struct files_struct *in_fsp,
uint16_t in_lock_count,
struct smbd_smb2_lock_element *in_locks);
static NTSTATUS smbd_smb2_lock_recv(struct tevent_req *req);
static void smbd_smb2_request_lock_done(struct tevent_req *subreq);
NTSTATUS smbd_smb2_request_process_lock(struct smbd_smb2_request *req)
{
const uint8_t *inbody;
uint16_t in_lock_count;
uint64_t in_file_id_persistent;
uint64_t in_file_id_volatile;
struct files_struct *in_fsp;
struct smbd_smb2_lock_element *in_locks;
struct tevent_req *subreq;
const uint8_t *lock_buffer;
uint16_t l;
NTSTATUS status;
status = smbd_smb2_request_verify_sizes(req, 0x30);
if (!NT_STATUS_IS_OK(status)) {
return smbd_smb2_request_error(req, status);
}
inbody = SMBD_SMB2_IN_BODY_PTR(req);
in_lock_count = CVAL(inbody, 0x02);
/* 0x04 - 4 bytes reserved */
in_file_id_persistent = BVAL(inbody, 0x08);
in_file_id_volatile = BVAL(inbody, 0x10);
if (in_lock_count < 1) {
return smbd_smb2_request_error(req, NT_STATUS_INVALID_PARAMETER);
}
if (((in_lock_count - 1) * 0x18) > SMBD_SMB2_IN_DYN_LEN(req)) {
return smbd_smb2_request_error(req, NT_STATUS_INVALID_PARAMETER);
}
in_locks = talloc_array(req, struct smbd_smb2_lock_element,
in_lock_count);
if (in_locks == NULL) {
return smbd_smb2_request_error(req, NT_STATUS_NO_MEMORY);
}
l = 0;
lock_buffer = inbody + 0x18;
in_locks[l].offset = BVAL(lock_buffer, 0x00);
in_locks[l].length = BVAL(lock_buffer, 0x08);
in_locks[l].flags = IVAL(lock_buffer, 0x10);
/* 0x14 - 4 reserved bytes */
lock_buffer = SMBD_SMB2_IN_DYN_PTR(req);
for (l=1; l < in_lock_count; l++) {
in_locks[l].offset = BVAL(lock_buffer, 0x00);
in_locks[l].length = BVAL(lock_buffer, 0x08);
in_locks[l].flags = IVAL(lock_buffer, 0x10);
/* 0x14 - 4 reserved bytes */
lock_buffer += 0x18;
}
in_fsp = file_fsp_smb2(req, in_file_id_persistent, in_file_id_volatile);
if (in_fsp == NULL) {
return smbd_smb2_request_error(req, NT_STATUS_FILE_CLOSED);
}
subreq = smbd_smb2_lock_send(req, req->sconn->ev_ctx,
req, in_fsp,
in_lock_count,
in_locks);
if (subreq == NULL) {
return smbd_smb2_request_error(req, NT_STATUS_NO_MEMORY);
}
tevent_req_set_callback(subreq, smbd_smb2_request_lock_done, req);
return smbd_smb2_request_pending_queue(req, subreq, 500);
}
static void smbd_smb2_request_lock_done(struct tevent_req *subreq)
{
struct smbd_smb2_request *smb2req = tevent_req_callback_data(subreq,
struct smbd_smb2_request);
DATA_BLOB outbody;
NTSTATUS status;
NTSTATUS error; /* transport error */
status = smbd_smb2_lock_recv(subreq);
TALLOC_FREE(subreq);
if (!NT_STATUS_IS_OK(status)) {
error = smbd_smb2_request_error(smb2req, status);
if (!NT_STATUS_IS_OK(error)) {
smbd_server_connection_terminate(smb2req->xconn,
nt_errstr(error));
return;
}
return;
}
outbody = smbd_smb2_generate_outbody(smb2req, 0x04);
if (outbody.data == NULL) {
error = smbd_smb2_request_error(smb2req, NT_STATUS_NO_MEMORY);
if (!NT_STATUS_IS_OK(error)) {
smbd_server_connection_terminate(smb2req->xconn,
nt_errstr(error));
return;
}
return;
}
SSVAL(outbody.data, 0x00, 0x04); /* struct size */
SSVAL(outbody.data, 0x02, 0); /* reserved */
error = smbd_smb2_request_done(smb2req, outbody, NULL);
if (!NT_STATUS_IS_OK(error)) {
smbd_server_connection_terminate(smb2req->xconn,
nt_errstr(error));
return;
}
}
static struct tevent_req *smbd_smb2_lock_send(TALLOC_CTX *mem_ctx,
struct tevent_context *ev,
struct smbd_smb2_request *smb2req,
struct files_struct *fsp,
uint16_t in_lock_count,
struct smbd_smb2_lock_element *in_locks)
{
struct tevent_req *req;
struct smbd_smb2_lock_state *state;
struct smb_request *smb1req;
int32_t timeout = -1;
bool isunlock = false;
uint16_t i;
struct smbd_lock_element *locks;
NTSTATUS status;
bool async = false;
req = tevent_req_create(mem_ctx, &state,
struct smbd_smb2_lock_state);
if (req == NULL) {
return NULL;
}
state->smb2req = smb2req;
smb2req->subreq = req; /* So we can find this when going async. */
smb1req = smbd_smb2_fake_smb_request(smb2req);
if (tevent_req_nomem(smb1req, req)) {
return tevent_req_post(req, ev);
}
state->smb1req = smb1req;
DEBUG(10,("smbd_smb2_lock_send: %s - %s\n",
fsp_str_dbg(fsp), fsp_fnum_dbg(fsp)));
locks = talloc_array(state, struct smbd_lock_element, in_lock_count);
if (locks == NULL) {
tevent_req_nterror(req, NT_STATUS_NO_MEMORY);
return tevent_req_post(req, ev);
}
switch (in_locks[0].flags) {
case SMB2_LOCK_FLAG_SHARED:
case SMB2_LOCK_FLAG_EXCLUSIVE:
if (in_lock_count > 1) {
tevent_req_nterror(req, NT_STATUS_INVALID_PARAMETER);
return tevent_req_post(req, ev);
}
timeout = -1;
break;
case SMB2_LOCK_FLAG_SHARED|SMB2_LOCK_FLAG_FAIL_IMMEDIATELY:
case SMB2_LOCK_FLAG_EXCLUSIVE|SMB2_LOCK_FLAG_FAIL_IMMEDIATELY:
timeout = 0;
break;
case SMB2_LOCK_FLAG_UNLOCK:
/* only the first lock gives the UNLOCK bit - see
MS-SMB2 3.3.5.14 */
isunlock = true;
timeout = 0;
break;
default:
tevent_req_nterror(req, NT_STATUS_INVALID_PARAMETER);
return tevent_req_post(req, ev);
}
if (!isunlock && (in_lock_count > 1)) {
/*
* 3.3.5.14.2 says we SHOULD fail with INVALID_PARAMETER if we
* have more than one lock and one of those is blocking.
*/
for (i=0; i<in_lock_count; i++) {
uint32_t flags = in_locks[i].flags;
if ((flags & SMB2_LOCK_FLAG_FAIL_IMMEDIATELY) == 0) {
tevent_req_nterror(
req, NT_STATUS_INVALID_PARAMETER);
return tevent_req_post(req, ev);
}
}
}
for (i=0; i<in_lock_count; i++) {
bool invalid = false;
switch (in_locks[i].flags) {
case SMB2_LOCK_FLAG_SHARED:
case SMB2_LOCK_FLAG_EXCLUSIVE:
if (isunlock) {
invalid = true;
break;
}
break;
case SMB2_LOCK_FLAG_SHARED|SMB2_LOCK_FLAG_FAIL_IMMEDIATELY:
case SMB2_LOCK_FLAG_EXCLUSIVE|SMB2_LOCK_FLAG_FAIL_IMMEDIATELY:
if (isunlock) {
invalid = true;
}
break;
case SMB2_LOCK_FLAG_UNLOCK:
if (!isunlock) {
tevent_req_nterror(req,
NT_STATUS_INVALID_PARAMETER);
return tevent_req_post(req, ev);
}
break;
default:
if (isunlock) {
/*
* If the first element was a UNLOCK
* we need to defer the error response
* to the backend, because we need to process
* all unlock elements before
*/
invalid = true;
break;
}
tevent_req_nterror(req, NT_STATUS_INVALID_PARAMETER);
return tevent_req_post(req, ev);
}
locks[i].smblctx = fsp->op->global->open_persistent_id;
locks[i].offset = in_locks[i].offset;
locks[i].count = in_locks[i].length;
if (in_locks[i].flags & SMB2_LOCK_FLAG_EXCLUSIVE) {
locks[i].brltype = WRITE_LOCK;
} else if (in_locks[i].flags & SMB2_LOCK_FLAG_SHARED) {
locks[i].brltype = READ_LOCK;
} else if (invalid) {
/*
* this is an invalid UNLOCK element
* and the backend needs to test for
* brltype != UNLOCK_LOCK and return
* NT_STATUS_INVALID_PARAMETER
*/
locks[i].brltype = READ_LOCK;
} else {
locks[i].brltype = UNLOCK_LOCK;
}
DEBUG(10,("smbd_smb2_lock_send: index %d offset=%llu, count=%llu, "
"smblctx = %llu type %d\n",
i,
(unsigned long long)locks[i].offset,
(unsigned long long)locks[i].count,
(unsigned long long)locks[i].smblctx,
(int)locks[i].brltype ));
}
state->locks = locks;
state->lock_count = in_lock_count;
if (isunlock) {
status = smbd_do_unlocking(smb1req, fsp,
in_lock_count, locks);
async = false;
} else {
status = smbd_do_locking(smb1req, fsp,
0,
timeout,
in_lock_count,
locks,
&async);
}
if (!NT_STATUS_IS_OK(status)) {
if (NT_STATUS_EQUAL(status, NT_STATUS_FILE_LOCK_CONFLICT)) {
status = NT_STATUS_LOCK_NOT_GRANTED;
}
tevent_req_nterror(req, status);
return tevent_req_post(req, ev);
}
if (async) {
tevent_req_defer_callback(req, smb2req->sconn->ev_ctx);
SMBPROFILE_IOBYTES_ASYNC_SET_IDLE(smb2req->profile);
return req;
}
tevent_req_done(req);
return tevent_req_post(req, ev);
}
static NTSTATUS smbd_smb2_lock_recv(struct tevent_req *req)
{
NTSTATUS status;
if (tevent_req_is_nterror(req, &status)) {
tevent_req_received(req);
return status;
}
tevent_req_received(req);
return NT_STATUS_OK;
}
/****************************************************************
Cancel an outstanding blocking lock request.
*****************************************************************/
static bool smbd_smb2_lock_cancel(struct tevent_req *req)
{
struct smbd_smb2_request *smb2req = NULL;
struct smbd_smb2_lock_state *state = tevent_req_data(req,
struct smbd_smb2_lock_state);
if (!state) {
return false;
}
if (!state->smb2req) {
return false;
}
smb2req = state->smb2req;
remove_pending_lock(state, state->blr);
/*
* If the request is canceled because of logoff, tdis or close
* the status is NT_STATUS_RANGE_NOT_LOCKED instead of
* NT_STATUS_CANCELLED.
*
* Note that the close case is handled in
* cancel_pending_lock_requests_by_fid_smb2(SHUTDOWN_CLOSE)
* for now.
*/
if (!NT_STATUS_IS_OK(smb2req->session->status)) {
tevent_req_nterror(req, NT_STATUS_RANGE_NOT_LOCKED);
return true;
}
if (!NT_STATUS_IS_OK(smb2req->tcon->status)) {
tevent_req_nterror(req, NT_STATUS_RANGE_NOT_LOCKED);
return true;
}
tevent_req_nterror(req, NT_STATUS_CANCELLED);
return true;
}
/****************************************************************
Got a message saying someone unlocked a file. Re-schedule all
blocking lock requests as we don't know if anything overlapped.
*****************************************************************/
static void received_unlock_msg(struct messaging_context *msg,
void *private_data,
uint32_t msg_type,
struct server_id server_id,
DATA_BLOB *data)
{
struct smbd_server_connection *sconn =
talloc_get_type_abort(private_data,
struct smbd_server_connection);
DEBUG(10,("received_unlock_msg (SMB2)\n"));
process_blocking_lock_queue_smb2(sconn, timeval_current());
}
/****************************************************************
Function to get the blr on a pending record.
*****************************************************************/
struct blocking_lock_record *get_pending_smb2req_blr(struct smbd_smb2_request *smb2req)
{
struct smbd_smb2_lock_state *state = NULL;
const uint8_t *inhdr;
if (!smb2req) {
return NULL;
}
if (smb2req->subreq == NULL) {
return NULL;
}
if (!tevent_req_is_in_progress(smb2req->subreq)) {
return NULL;
}
inhdr = SMBD_SMB2_IN_HDR_PTR(smb2req);
if (SVAL(inhdr, SMB2_HDR_OPCODE) != SMB2_OP_LOCK) {
return NULL;
}
state = tevent_req_data(smb2req->subreq,
struct smbd_smb2_lock_state);
if (!state) {
return NULL;
}
return state->blr;
}
/****************************************************************
Set up the next brl timeout.
*****************************************************************/
static bool recalc_smb2_brl_timeout(struct smbd_server_connection *sconn)
{
struct smbXsrv_connection *xconn = NULL;
struct timeval next_timeout = timeval_zero();
int max_brl_timeout = lp_parm_int(-1, "brl", "recalctime", 5);
TALLOC_FREE(sconn->smb2.locks.brl_timeout);
if (sconn->client != NULL) {
xconn = sconn->client->connections;
}
for (; xconn != NULL; xconn = xconn->next) {
struct smbd_smb2_request *smb2req, *nextreq;
for (smb2req = xconn->smb2.requests; smb2req; smb2req = nextreq) {
struct blocking_lock_record *blr =
get_pending_smb2req_blr(smb2req);
nextreq = smb2req->next;
if (blr == NULL) {
continue;
}
if (!timeval_is_zero(&blr->expire_time)) {
next_timeout = timeval_brl_min(&next_timeout,
&blr->expire_time);
continue;
}
/*
* If we're blocked on pid 0xFFFFFFFFFFFFFFFFLL this is
* a POSIX lock, so calculate a timeout of
* 10 seconds into the future.
*/
if (blr->blocking_smblctx == 0xFFFFFFFFFFFFFFFFLL) {
struct timeval psx_to;
psx_to = timeval_current_ofs(10, 0);
next_timeout = timeval_brl_min(&next_timeout,
&psx_to);
}
}
}
if (timeval_is_zero(&next_timeout)) {
DEBUG(10, ("recalc_smb2_brl_timeout:Next "
"timeout = Infinite.\n"));
return true;
}
/*
* To account for unclean shutdowns by clients we need a
* maximum timeout that we use for checking pending locks. If
* we have any pending locks at all, then check if the pending
* lock can continue at least every brl:recalctime seconds
* (default 5 seconds).
*
* This saves us needing to do a message_send_all() in the
* SIGCHLD handler in the parent daemon. That
* message_send_all() caused O(n^2) work to be done when IP
* failovers happened in clustered Samba, which could make the
* entire system unusable for many minutes.
*/
if (max_brl_timeout > 0) {
struct timeval min_to = timeval_current_ofs(max_brl_timeout, 0);
next_timeout = timeval_brl_min(&next_timeout, &min_to);
}
if (DEBUGLVL(10)) {
struct timeval cur, from_now;
cur = timeval_current();
from_now = timeval_until(&cur, &next_timeout);
DEBUG(10, ("recalc_smb2_brl_timeout: Next "
"timeout = %d.%d seconds from now.\n",
(int)from_now.tv_sec, (int)from_now.tv_usec));
}
sconn->smb2.locks.brl_timeout = tevent_add_timer(
sconn->ev_ctx,
NULL,
next_timeout,
brl_timeout_fn,
sconn);
if (!sconn->smb2.locks.brl_timeout) {
return false;
}
return true;
}
/****************************************************************
Get an SMB2 lock request to go async. lock_timeout should
always be -1 here.
*****************************************************************/
bool push_blocking_lock_request_smb2( struct byte_range_lock *br_lck,
struct smb_request *smb1req,
files_struct *fsp,
int lock_timeout,
int lock_num,
uint64_t smblctx,
enum brl_type lock_type,
enum brl_flavour lock_flav,
uint64_t offset,
uint64_t count,
uint64_t blocking_smblctx)
{
struct smbd_server_connection *sconn = smb1req->sconn;
struct smbd_smb2_request *smb2req = smb1req->smb2req;
struct tevent_req *req = NULL;
struct smbd_smb2_lock_state *state = NULL;
struct blocking_lock_record *blr = NULL;
NTSTATUS status = NT_STATUS_OK;
if (!smb2req) {
return false;
}
req = smb2req->subreq;
if (!req) {
return false;
}
if (!tevent_req_is_in_progress(smb2req->subreq)) {
return false;
}
state = tevent_req_data(req, struct smbd_smb2_lock_state);
if (!state) {
return false;
}
blr = talloc_zero(state, struct blocking_lock_record);
if (!blr) {
return false;
}
blr->fsp = fsp;
if (lock_timeout == -1) {
blr->expire_time.tv_sec = 0;
blr->expire_time.tv_usec = 0; /* Never expire. */
} else {
blr->expire_time = timeval_current_ofs_msec(lock_timeout);
}
blr->lock_num = lock_num;
blr->smblctx = smblctx;
blr->blocking_smblctx = blocking_smblctx;
blr->lock_flav = lock_flav;
blr->lock_type = lock_type;
blr->offset = offset;
blr->count = count;
/* Specific brl_lock() implementations can fill this in. */
blr->blr_private = NULL;
/* Add a pending lock record for this. */
status = brl_lock(sconn->msg_ctx,
br_lck,
smblctx,
messaging_server_id(sconn->msg_ctx),
offset,
count,
lock_type == READ_LOCK ? PENDING_READ_LOCK : PENDING_WRITE_LOCK,
blr->lock_flav,
true,
NULL);
if (!NT_STATUS_IS_OK(status)) {
DEBUG(0,("push_blocking_lock_request_smb2: "
"failed to add PENDING_LOCK record.\n"));
TALLOC_FREE(blr);
return false;
}
state->blr = blr;
DEBUG(10,("push_blocking_lock_request_smb2: file %s timeout %d\n",
fsp_str_dbg(fsp),
lock_timeout ));
recalc_smb2_brl_timeout(sconn);
/* Ensure we'll receive messages when this is unlocked. */
if (!sconn->smb2.locks.blocking_lock_unlock_state) {
messaging_register(sconn->msg_ctx, sconn,
MSG_SMB_UNLOCK, received_unlock_msg);
sconn->smb2.locks.blocking_lock_unlock_state = true;
}
/* allow this request to be canceled */
tevent_req_set_cancel_fn(req, smbd_smb2_lock_cancel);
return true;
}
/****************************************************************
Remove a pending lock record under lock.
*****************************************************************/
static void remove_pending_lock(struct smbd_smb2_lock_state *state,
struct blocking_lock_record *blr)
{
struct byte_range_lock *br_lck = brl_get_locks(
state, blr->fsp);
DEBUG(10, ("remove_pending_lock: BLR = %p\n", blr));
if (br_lck) {
brl_lock_cancel(br_lck,
blr->smblctx,
messaging_server_id(blr->fsp->conn->sconn->msg_ctx),
blr->offset,
blr->count,
blr->lock_flav);
TALLOC_FREE(br_lck);
}
}
/****************************************************************
Re-proccess a blocking lock request.
This is equivalent to process_lockingX() inside smbd/blocking.c
*****************************************************************/
static void reprocess_blocked_smb2_lock(struct smbd_smb2_request *smb2req,
struct timeval tv_curr)
{
NTSTATUS status = NT_STATUS_UNSUCCESSFUL;
struct blocking_lock_record *blr = NULL;
struct smbd_smb2_lock_state *state = NULL;
struct byte_range_lock *br_lck = NULL;
struct smbd_lock_element *e = NULL;
files_struct *fsp = NULL;
if (!smb2req->subreq) {
return;
}
SMBPROFILE_IOBYTES_ASYNC_SET_BUSY(smb2req->profile);
state = tevent_req_data(smb2req->subreq, struct smbd_smb2_lock_state);
if (!state) {
return;
}
blr = state->blr;
fsp = blr->fsp;
/* We can only have one blocked lock in SMB2. */
SMB_ASSERT(state->lock_count == 1);
SMB_ASSERT(blr->lock_num == 0);
/* Try and get the outstanding lock. */
e = &state->locks[blr->lock_num];
br_lck = do_lock(fsp->conn->sconn->msg_ctx,
fsp,
e->smblctx,
e->count,
e->offset,
e->brltype,
WINDOWS_LOCK,
true,
&status,
&blr->blocking_smblctx);
TALLOC_FREE(br_lck);
if (NT_STATUS_IS_OK(status)) {
/*
* Success - we got the lock.
*/
DEBUG(3,("reprocess_blocked_smb2_lock SUCCESS file = %s, "
"%s, num_locks=%d\n",
fsp_str_dbg(fsp),
fsp_fnum_dbg(fsp),
(int)state->lock_count));
remove_pending_lock(state, blr);
tevent_req_done(smb2req->subreq);
return;
}
if (!NT_STATUS_EQUAL(status,NT_STATUS_LOCK_NOT_GRANTED) &&
!NT_STATUS_EQUAL(status,NT_STATUS_FILE_LOCK_CONFLICT)) {
/*
* We have other than a "can't get lock"
* error. Return an error.
*/
remove_pending_lock(state, blr);
tevent_req_nterror(smb2req->subreq, status);
return;
}
/*
* We couldn't get the lock for this record.
* If the time has expired, return a lock error.
*/
if (!timeval_is_zero(&blr->expire_time) &&
timeval_compare(&blr->expire_time, &tv_curr) <= 0) {
remove_pending_lock(state, blr);
tevent_req_nterror(smb2req->subreq, NT_STATUS_LOCK_NOT_GRANTED);
return;
}
/*
* Still can't get the lock - keep waiting.
*/
DEBUG(10,("reprocess_blocked_smb2_lock: failed to get lock "
"for file %s, %s. Still waiting....\n",
fsp_str_dbg(fsp),
fsp_fnum_dbg(fsp)));
SMBPROFILE_IOBYTES_ASYNC_SET_IDLE(smb2req->profile);
return;
}
/****************************************************************
Attempt to proccess all outstanding blocking locks pending on
the request queue.
*****************************************************************/
void process_blocking_lock_queue_smb2(
struct smbd_server_connection *sconn, struct timeval tv_curr)
{
struct smbXsrv_connection *xconn = NULL;
if (sconn != NULL && sconn->client != NULL) {
xconn = sconn->client->connections;
}
for (; xconn != NULL; xconn = xconn->next) {
struct smbd_smb2_request *smb2req, *nextreq;
for (smb2req = xconn->smb2.requests; smb2req; smb2req = nextreq) {
const uint8_t *inhdr;
nextreq = smb2req->next;
if (smb2req->subreq == NULL) {
/* This message has been processed. */
continue;
}
if (!tevent_req_is_in_progress(smb2req->subreq)) {
/* This message has been processed. */
continue;
}
inhdr = SMBD_SMB2_IN_HDR_PTR(smb2req);
if (SVAL(inhdr, SMB2_HDR_OPCODE) == SMB2_OP_LOCK) {
reprocess_blocked_smb2_lock(smb2req, tv_curr);
}
}
}
recalc_smb2_brl_timeout(sconn);
}
/****************************************************************************
Remove any locks on this fd. Called from file_close().
****************************************************************************/
void cancel_pending_lock_requests_by_fid_smb2(files_struct *fsp,
struct byte_range_lock *br_lck,
enum file_close_type close_type)
{
struct smbd_server_connection *sconn = fsp->conn->sconn;
struct smbXsrv_connection *xconn = NULL;
if (sconn != NULL && sconn->client != NULL) {
xconn = sconn->client->connections;
}
for (; xconn != NULL; xconn = xconn->next) {
struct smbd_smb2_request *smb2req, *nextreq;
for (smb2req = xconn->smb2.requests; smb2req; smb2req = nextreq) {
struct smbd_smb2_lock_state *state = NULL;
files_struct *fsp_curr = NULL;
struct blocking_lock_record *blr = NULL;
const uint8_t *inhdr;
nextreq = smb2req->next;
if (smb2req->subreq == NULL) {
/* This message has been processed. */
continue;
}
if (!tevent_req_is_in_progress(smb2req->subreq)) {
/* This message has been processed. */
continue;
}
inhdr = SMBD_SMB2_IN_HDR_PTR(smb2req);
if (SVAL(inhdr, SMB2_HDR_OPCODE) != SMB2_OP_LOCK) {
/* Not a lock call. */
continue;
}
state = tevent_req_data(smb2req->subreq,
struct smbd_smb2_lock_state);
if (!state) {
/* Strange - is this even possible ? */
continue;
}
fsp_curr = smb2req->compat_chain_fsp;
if (fsp_curr == NULL) {
/* Strange - is this even possible ? */
continue;
}
if (fsp_curr != fsp) {
/* It's not our fid */
continue;
}
blr = state->blr;
/* Remove the entries from the lock db. */
brl_lock_cancel(br_lck,
blr->smblctx,
messaging_server_id(sconn->msg_ctx),
blr->offset,
blr->count,
blr->lock_flav);
/* Finally end the request. */
if (close_type == SHUTDOWN_CLOSE) {
tevent_req_done(smb2req->subreq);
} else {
tevent_req_nterror(smb2req->subreq,
NT_STATUS_RANGE_NOT_LOCKED);
}
}
}
}