1998-08-19 05:49:57 +04:00
/*
2002-01-30 09:08:46 +03:00
Unix SMB / CIFS implementation .
1998-08-19 05:49:57 +04:00
Blocking Locking functions
2003-02-27 04:04:34 +03:00
Copyright ( C ) Jeremy Allison 1998 - 2003
2008-11-03 19:56:55 +03:00
1998-08-19 05:49:57 +04:00
This program is free software ; you can redistribute it and / or modify
it under the terms of the GNU General Public License as published by
2007-07-09 23:25:36 +04:00
the Free Software Foundation ; either version 3 of the License , or
1998-08-19 05:49:57 +04:00
( at your option ) any later version .
2008-11-03 19:56:55 +03:00
1998-08-19 05:49:57 +04:00
This program is distributed in the hope that it will be useful ,
but WITHOUT ANY WARRANTY ; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE . See the
GNU General Public License for more details .
2008-11-03 19:56:55 +03:00
1998-08-19 05:49:57 +04:00
You should have received a copy of the GNU General Public License
2007-07-10 04:52:41 +04:00
along with this program . If not , see < http : //www.gnu.org/licenses/>.
1998-08-19 05:49:57 +04:00
*/
# include "includes.h"
2020-10-28 14:09:39 +03:00
# include "locking/share_mode_lock.h"
2011-03-22 18:57:01 +03:00
# include "smbd/smbd.h"
2009-01-08 14:03:45 +03:00
# include "smbd/globals.h"
2011-03-24 17:31:06 +03:00
# include "messages.h"
2019-06-13 13:38:57 +03:00
# include "lib/util/tevent_ntstatus.h"
# include "lib/dbwrap/dbwrap_watch.h"
# include "librpc/gen_ndr/ndr_open_files.h"
2009-01-08 14:03:45 +03:00
2006-07-18 01:09:02 +04:00
# undef DBGC_CLASS
# define DBGC_CLASS DBGC_LOCKING
2001-10-02 08:29:50 +04:00
2019-06-13 13:38:57 +03:00
NTSTATUS smbd_do_locks_try (
struct files_struct * fsp ,
uint16_t num_locks ,
struct smbd_lock_element * locks ,
uint16_t * blocker_idx ,
struct server_id * blocking_pid ,
uint64_t * blocking_smblctx )
{
NTSTATUS status = NT_STATUS_OK ;
uint16_t i ;
for ( i = 0 ; i < num_locks ; i + + ) {
struct smbd_lock_element * e = & locks [ i ] ;
2019-07-01 15:30:15 +03:00
status = do_lock (
2019-06-13 13:38:57 +03:00
fsp ,
2019-08-08 20:26:28 +03:00
locks , /* req_mem_ctx */
& e - > req_guid ,
2019-06-13 13:38:57 +03:00
e - > smblctx ,
e - > count ,
e - > offset ,
e - > brltype ,
2021-11-17 02:06:59 +03:00
e - > lock_flav ,
2019-06-13 13:38:57 +03:00
blocking_pid ,
blocking_smblctx ) ;
if ( ! NT_STATUS_IS_OK ( status ) ) {
break ;
}
}
if ( NT_STATUS_IS_OK ( status ) ) {
return NT_STATUS_OK ;
}
* blocker_idx = i ;
/*
* Undo the locks we successfully got
*/
for ( i = i - 1 ; i ! = UINT16_MAX ; i - - ) {
struct smbd_lock_element * e = & locks [ i ] ;
2019-07-01 16:33:58 +03:00
do_unlock ( fsp ,
2019-06-13 13:38:57 +03:00
e - > smblctx ,
e - > count ,
e - > offset ,
2021-11-17 02:06:59 +03:00
e - > lock_flav ) ;
2019-06-13 13:38:57 +03:00
}
return status ;
}
static bool smbd_smb1_fsp_add_blocked_lock_req (
struct files_struct * fsp , struct tevent_req * req )
{
size_t num_reqs = talloc_array_length ( fsp - > blocked_smb1_lock_reqs ) ;
struct tevent_req * * tmp = NULL ;
tmp = talloc_realloc (
fsp ,
fsp - > blocked_smb1_lock_reqs ,
struct tevent_req * ,
num_reqs + 1 ) ;
if ( tmp = = NULL ) {
return false ;
}
fsp - > blocked_smb1_lock_reqs = tmp ;
fsp - > blocked_smb1_lock_reqs [ num_reqs ] = req ;
return true ;
}
struct smbd_smb1_do_locks_state {
struct tevent_context * ev ;
struct smb_request * smbreq ;
struct files_struct * fsp ;
2019-08-13 19:34:36 +03:00
uint32_t timeout ;
2019-08-15 18:26:43 +03:00
uint32_t polling_msecs ;
2019-08-19 16:29:32 +03:00
uint32_t retry_msecs ;
2019-06-13 13:38:57 +03:00
struct timeval endtime ;
bool large_offset ; /* required for correct cancel */
uint16_t num_locks ;
struct smbd_lock_element * locks ;
uint16_t blocker ;
2019-08-19 13:04:43 +03:00
NTSTATUS deny_status ;
2019-06-13 13:38:57 +03:00
} ;
2019-08-13 19:34:36 +03:00
static void smbd_smb1_do_locks_try ( struct tevent_req * req ) ;
2019-06-13 13:38:57 +03:00
static void smbd_smb1_do_locks_retry ( struct tevent_req * subreq ) ;
static void smbd_smb1_blocked_locks_cleanup (
struct tevent_req * req , enum tevent_req_state req_state ) ;
2019-08-16 15:55:13 +03:00
static NTSTATUS smbd_smb1_do_locks_check (
struct files_struct * fsp ,
uint16_t num_locks ,
struct smbd_lock_element * locks ,
uint16_t * blocker_idx ,
struct server_id * blocking_pid ,
uint64_t * blocking_smblctx ) ;
2019-06-13 13:38:57 +03:00
2019-08-15 19:02:57 +03:00
static void smbd_smb1_do_locks_setup_timeout (
struct smbd_smb1_do_locks_state * state ,
const struct smbd_lock_element * blocker )
{
struct files_struct * fsp = state - > fsp ;
if ( ! timeval_is_zero ( & state - > endtime ) ) {
/*
* already done
*/
return ;
}
if ( ( state - > timeout ! = 0 ) & & ( state - > timeout ! = UINT32_MAX ) ) {
/*
* Windows internal resolution for blocking locks
* seems to be about 200 ms . . . Don ' t wait for less than
* that . JRA .
*/
state - > timeout = MAX ( state - > timeout , lp_lock_spin_time ( ) ) ;
}
if ( state - > timeout ! = 0 ) {
goto set_endtime ;
}
if ( blocker = = NULL ) {
goto set_endtime ;
}
if ( ( blocker - > offset > = 0xEF000000 ) & &
( ( blocker - > offset > > 63 ) = = 0 ) ) {
/*
* This must be an optimization of an ancient
* application bug . . .
*/
state - > timeout = lp_lock_spin_time ( ) ;
}
2020-04-03 10:48:58 +03:00
if ( fsp - > fsp_flags . lock_failure_seen & &
2019-08-15 19:02:57 +03:00
( blocker - > offset = = fsp - > lock_failure_offset ) ) {
/*
* Delay repeated lock attempts on the same
* lock . Maybe a more advanced version of the
* above check ?
*/
DBG_DEBUG ( " Delaying lock request due to previous "
" failure \n " ) ;
state - > timeout = lp_lock_spin_time ( ) ;
}
set_endtime :
/*
* Note state - > timeout might still 0 ,
* but that ' s ok , as we don ' t want to retry
* in that case .
*/
state - > endtime = timeval_add ( & state - > smbreq - > request_time ,
state - > timeout / 1000 ,
( state - > timeout % 1000 ) * 1000 ) ;
}
2019-08-19 16:29:32 +03:00
static void smbd_smb1_do_locks_update_retry_msecs (
struct smbd_smb1_do_locks_state * state )
{
/*
* The default lp_lock_spin_time ( ) is 200 ms ,
* we just use half of it to trigger the first retry .
*
* v_min is in the range of 0.001 to 10 secs
* ( 0.1 secs by default )
*
* v_max is in the range of 0.01 to 100 secs
* ( 1.0 secs by default )
*
* The typical steps are :
* 0.1 , 0.2 , 0.3 , 0.4 , . . . 1.0
*/
uint32_t v_min = MAX ( 2 , MIN ( 20000 , lp_lock_spin_time ( ) ) ) / 2 ;
uint32_t v_max = 10 * v_min ;
if ( state - > retry_msecs > = v_max ) {
state - > retry_msecs = v_max ;
return ;
}
state - > retry_msecs + = v_min ;
}
2019-08-15 18:26:43 +03:00
static void smbd_smb1_do_locks_update_polling_msecs (
struct smbd_smb1_do_locks_state * state )
{
/*
* The default lp_lock_spin_time ( ) is 200 ms .
*
* v_min is in the range of 0.002 to 20 secs
* ( 0.2 secs by default )
*
* v_max is in the range of 0.02 to 200 secs
* ( 2.0 secs by default )
*
* The typical steps are :
* 0.2 , 0.4 , 0.6 , 0.8 , . . . 2.0
*/
uint32_t v_min = MAX ( 2 , MIN ( 20000 , lp_lock_spin_time ( ) ) ) ;
uint32_t v_max = 10 * v_min ;
if ( state - > polling_msecs > = v_max ) {
state - > polling_msecs = v_max ;
return ;
}
state - > polling_msecs + = v_min ;
}
2019-06-13 13:38:57 +03:00
struct tevent_req * smbd_smb1_do_locks_send (
TALLOC_CTX * mem_ctx ,
struct tevent_context * ev ,
struct smb_request * * smbreq , /* talloc_move()d into our state */
struct files_struct * fsp ,
2019-08-13 19:34:36 +03:00
uint32_t lock_timeout ,
2019-06-13 13:38:57 +03:00
bool large_offset ,
uint16_t num_locks ,
struct smbd_lock_element * locks )
{
2019-08-15 21:09:55 +03:00
struct tevent_req * req = NULL ;
2019-06-13 13:38:57 +03:00
struct smbd_smb1_do_locks_state * state = NULL ;
bool ok ;
req = tevent_req_create (
mem_ctx , & state , struct smbd_smb1_do_locks_state ) ;
if ( req = = NULL ) {
return NULL ;
}
state - > ev = ev ;
state - > smbreq = talloc_move ( state , smbreq ) ;
state - > fsp = fsp ;
2019-08-13 19:34:36 +03:00
state - > timeout = lock_timeout ;
2019-06-13 13:38:57 +03:00
state - > large_offset = large_offset ;
state - > num_locks = num_locks ;
state - > locks = locks ;
2021-11-17 03:15:54 +03:00
state - > deny_status = NT_STATUS_LOCK_NOT_GRANTED ;
DBG_DEBUG ( " state=%p, state->smbreq=%p \n " , state , state - > smbreq ) ;
if ( num_locks = = 0 | | locks = = NULL ) {
DBG_DEBUG ( " no locks \n " ) ;
tevent_req_done ( req ) ;
return tevent_req_post ( req , ev ) ;
}
2019-06-13 13:38:57 +03:00
2021-11-17 03:41:09 +03:00
if ( state - > locks [ 0 ] . lock_flav = = POSIX_LOCK ) {
2019-08-19 13:04:43 +03:00
/*
* SMB1 posix locks always use
* NT_STATUS_FILE_LOCK_CONFLICT .
*/
state - > deny_status = NT_STATUS_FILE_LOCK_CONFLICT ;
2019-06-13 13:38:57 +03:00
}
2019-08-15 21:09:55 +03:00
smbd_smb1_do_locks_try ( req ) ;
if ( ! tevent_req_is_in_progress ( req ) ) {
2019-06-13 13:38:57 +03:00
return tevent_req_post ( req , ev ) ;
}
ok = smbd_smb1_fsp_add_blocked_lock_req ( fsp , req ) ;
if ( ! ok ) {
tevent_req_oom ( req ) ;
2019-08-15 21:09:55 +03:00
return tevent_req_post ( req , ev ) ;
2019-06-13 13:38:57 +03:00
}
tevent_req_set_cleanup_fn ( req , smbd_smb1_blocked_locks_cleanup ) ;
return req ;
}
static void smbd_smb1_blocked_locks_cleanup (
struct tevent_req * req , enum tevent_req_state req_state )
{
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
struct files_struct * fsp = state - > fsp ;
struct tevent_req * * blocked = fsp - > blocked_smb1_lock_reqs ;
size_t num_blocked = talloc_array_length ( blocked ) ;
2020-03-26 12:31:47 +03:00
size_t i ;
2019-06-13 13:38:57 +03:00
DBG_DEBUG ( " req=%p, state=%p, req_state=%d \n " ,
req ,
state ,
( int ) req_state ) ;
if ( req_state = = TEVENT_REQ_RECEIVED ) {
DBG_DEBUG ( " already received \n " ) ;
return ;
}
for ( i = 0 ; i < num_blocked ; i + + ) {
if ( blocked [ i ] = = req ) {
break ;
}
}
SMB_ASSERT ( i < num_blocked ) ;
2020-03-26 12:31:47 +03:00
ARRAY_DEL_ELEMENT ( blocked , i , num_blocked ) ;
2019-06-13 13:38:57 +03:00
fsp - > blocked_smb1_lock_reqs = talloc_realloc (
fsp , blocked , struct tevent_req * , num_blocked - 1 ) ;
}
2019-08-16 15:55:13 +03:00
static NTSTATUS smbd_smb1_do_locks_check_blocked (
uint16_t num_blocked ,
struct smbd_lock_element * blocked ,
uint16_t num_locks ,
struct smbd_lock_element * locks ,
uint16_t * blocker_idx ,
uint64_t * blocking_smblctx )
{
uint16_t li ;
for ( li = 0 ; li < num_locks ; li + + ) {
struct smbd_lock_element * l = & locks [ li ] ;
uint16_t bi ;
bool valid ;
valid = byte_range_valid ( l - > offset , l - > count ) ;
if ( ! valid ) {
return NT_STATUS_INVALID_LOCK_RANGE ;
}
for ( bi = 0 ; bi < num_blocked ; bi + + ) {
struct smbd_lock_element * b = & blocked [ li ] ;
bool overlap ;
/* Read locks never conflict. */
if ( l - > brltype = = READ_LOCK & & b - > brltype = = READ_LOCK ) {
continue ;
}
overlap = byte_range_overlap ( l - > offset ,
l - > count ,
b - > offset ,
b - > count ) ;
if ( ! overlap ) {
continue ;
}
* blocker_idx = li ;
* blocking_smblctx = b - > smblctx ;
return NT_STATUS_LOCK_NOT_GRANTED ;
}
}
return NT_STATUS_OK ;
}
static NTSTATUS smbd_smb1_do_locks_check (
struct files_struct * fsp ,
uint16_t num_locks ,
struct smbd_lock_element * locks ,
uint16_t * blocker_idx ,
struct server_id * blocking_pid ,
uint64_t * blocking_smblctx )
{
struct tevent_req * * blocked = fsp - > blocked_smb1_lock_reqs ;
size_t num_blocked = talloc_array_length ( blocked ) ;
NTSTATUS status ;
size_t bi ;
/*
* We check the pending / blocked requests
* from the oldest to the youngest request .
*
* Note due to the retry logic the current request
* might already be in the list .
*/
for ( bi = 0 ; bi < num_blocked ; bi + + ) {
struct smbd_smb1_do_locks_state * blocked_state =
tevent_req_data ( blocked [ bi ] ,
struct smbd_smb1_do_locks_state ) ;
if ( blocked_state - > locks = = locks ) {
SMB_ASSERT ( blocked_state - > num_locks = = num_locks ) ;
/*
* We found ourself . . .
*/
break ;
}
status = smbd_smb1_do_locks_check_blocked (
blocked_state - > num_locks ,
blocked_state - > locks ,
num_locks ,
locks ,
blocker_idx ,
blocking_smblctx ) ;
if ( ! NT_STATUS_IS_OK ( status ) ) {
* blocking_pid = messaging_server_id (
fsp - > conn - > sconn - > msg_ctx ) ;
return status ;
}
}
status = smbd_do_locks_try (
fsp ,
num_locks ,
locks ,
blocker_idx ,
blocking_pid ,
blocking_smblctx ) ;
if ( ! NT_STATUS_IS_OK ( status ) ) {
return status ;
}
return NT_STATUS_OK ;
}
2019-08-13 19:34:36 +03:00
static void smbd_smb1_do_locks_try ( struct tevent_req * req )
2019-06-13 13:38:57 +03:00
{
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
struct files_struct * fsp = state - > fsp ;
struct share_mode_lock * lck ;
2019-08-15 19:17:47 +03:00
struct timeval endtime = { 0 } ;
2019-06-13 13:38:57 +03:00
struct server_id blocking_pid = { 0 } ;
uint64_t blocking_smblctx = 0 ;
2019-08-13 19:34:36 +03:00
struct tevent_req * subreq = NULL ;
2019-06-13 13:38:57 +03:00
NTSTATUS status ;
bool ok ;
2019-08-15 19:17:47 +03:00
bool expired ;
2019-06-13 13:38:57 +03:00
lck = get_existing_share_mode_lock ( state , fsp - > file_id ) ;
if ( tevent_req_nomem ( lck , req ) ) {
DBG_DEBUG ( " Could not get share mode lock \n " ) ;
return ;
}
2019-08-16 15:55:13 +03:00
status = smbd_smb1_do_locks_check (
2019-06-13 13:38:57 +03:00
fsp ,
2019-08-16 15:55:13 +03:00
state - > num_locks ,
state - > locks ,
2019-06-13 13:38:57 +03:00
& state - > blocker ,
& blocking_pid ,
& blocking_smblctx ) ;
if ( NT_STATUS_IS_OK ( status ) ) {
goto done ;
}
2019-08-19 16:29:32 +03:00
if ( NT_STATUS_EQUAL ( status , NT_STATUS_RETRY ) ) {
/*
* We got NT_STATUS_RETRY ,
* we reset polling_msecs so that
* that the retries based on LOCK_NOT_GRANTED
* will later start with small intervalls again .
*/
state - > polling_msecs = 0 ;
/*
* The backend wasn ' t able to decide yet .
* We need to wait even for non - blocking
* locks .
*
* The backend uses blocking_smblctx = = UINT64_MAX
* to indicate that we should use retry timers .
*
* It uses blocking_smblctx = = 0 to indicate
* it will use share_mode_wakeup_waiters ( )
* to wake us . Note that unrelated changes in
* locking . tdb may cause retries .
*/
if ( blocking_smblctx ! = UINT64_MAX ) {
SMB_ASSERT ( blocking_smblctx = = 0 ) ;
goto setup_retry ;
}
smbd_smb1_do_locks_update_retry_msecs ( state ) ;
DBG_DEBUG ( " Waiting for a backend decision. "
" Retry in % " PRIu32 " msecs \n " ,
state - > retry_msecs ) ;
/*
* We completely ignore state - > endtime here
* we we ' ll wait for a backend decision forever .
* If the backend is smart enough to implement
* some NT_STATUS_RETRY logic , it has to
* switch to any other status after in order
* to avoid waiting forever .
*/
endtime = timeval_current_ofs_msec ( state - > retry_msecs ) ;
goto setup_retry ;
}
2019-06-13 13:38:57 +03:00
if ( ! ERROR_WAS_LOCK_DENIED ( status ) ) {
goto done ;
}
2019-08-19 16:29:32 +03:00
/*
* We got LOCK_NOT_GRANTED , make sure
* a following STATUS_RETRY will start
* with short intervalls again .
*/
state - > retry_msecs = 0 ;
2019-06-13 13:38:57 +03:00
2019-08-15 19:18:50 +03:00
smbd_smb1_do_locks_setup_timeout ( state , & state - > locks [ state - > blocker ] ) ;
DBG_DEBUG ( " timeout=% " PRIu32 " , blocking_smblctx=% " PRIu64 " \n " ,
state - > timeout ,
blocking_smblctx ) ;
2019-08-15 15:21:38 +03:00
/*
2019-08-15 19:17:47 +03:00
* The client specified timeout expired
2019-08-15 15:21:38 +03:00
* avoid further retries .
*
* Otherwise keep waiting either waiting
* for changes in locking . tdb or the polling
* mode timers waiting for posix locks .
2019-08-19 13:04:43 +03:00
*
* If the endtime is not expired yet ,
* it means we ' ll retry after a timeout .
* In that case we ' ll have to return
* NT_STATUS_FILE_LOCK_CONFLICT
* instead of NT_STATUS_LOCK_NOT_GRANTED .
2019-08-15 15:21:38 +03:00
*/
2019-08-15 19:17:47 +03:00
expired = timeval_expired ( & state - > endtime ) ;
if ( expired ) {
2019-08-19 13:04:43 +03:00
status = state - > deny_status ;
2019-08-15 15:21:38 +03:00
goto done ;
}
2019-08-19 13:04:43 +03:00
state - > deny_status = NT_STATUS_FILE_LOCK_CONFLICT ;
2019-08-15 15:21:38 +03:00
2019-06-13 13:38:57 +03:00
endtime = state - > endtime ;
if ( blocking_smblctx = = UINT64_MAX ) {
struct timeval tmp ;
2019-08-15 18:26:43 +03:00
smbd_smb1_do_locks_update_polling_msecs ( state ) ;
DBG_DEBUG ( " Blocked on a posix lock. Retry in % " PRIu32 " msecs \n " ,
state - > polling_msecs ) ;
2019-06-13 13:38:57 +03:00
2019-08-15 18:26:43 +03:00
tmp = timeval_current_ofs_msec ( state - > polling_msecs ) ;
2019-06-13 13:38:57 +03:00
endtime = timeval_min ( & endtime , & tmp ) ;
}
2019-08-19 16:29:32 +03:00
setup_retry :
2019-11-04 15:06:20 +03:00
subreq = share_mode_watch_send (
2020-11-03 19:36:08 +03:00
state , state - > ev , lck , blocking_pid ) ;
2019-08-19 16:21:50 +03:00
if ( tevent_req_nomem ( subreq , req ) ) {
goto done ;
}
TALLOC_FREE ( lck ) ;
tevent_req_set_callback ( subreq , smbd_smb1_do_locks_retry , req ) ;
2019-08-19 16:29:32 +03:00
if ( timeval_is_zero ( & endtime ) ) {
return ;
}
2019-06-13 13:38:57 +03:00
ok = tevent_req_set_endtime ( subreq , state - > ev , endtime ) ;
if ( ! ok ) {
status = NT_STATUS_NO_MEMORY ;
goto done ;
}
return ;
done :
TALLOC_FREE ( lck ) ;
smbd_smb1_brl_finish_by_req ( req , status ) ;
}
2019-08-13 19:34:36 +03:00
static void smbd_smb1_do_locks_retry ( struct tevent_req * subreq )
{
struct tevent_req * req = tevent_req_callback_data (
subreq , struct tevent_req ) ;
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
NTSTATUS status ;
bool ok ;
/*
* Make sure we run as the user again
*/
2019-07-13 17:20:11 +03:00
ok = change_to_user_and_service_by_fsp ( state - > fsp ) ;
2019-08-13 19:34:36 +03:00
if ( ! ok ) {
tevent_req_nterror ( req , NT_STATUS_ACCESS_DENIED ) ;
return ;
}
2019-11-04 15:06:20 +03:00
status = share_mode_watch_recv ( subreq , NULL , NULL ) ;
2019-08-13 19:34:36 +03:00
TALLOC_FREE ( subreq ) ;
2019-11-04 15:06:20 +03:00
DBG_DEBUG ( " share_mode_watch_recv returned %s \n " ,
2019-08-13 19:34:36 +03:00
nt_errstr ( status ) ) ;
2019-08-15 15:21:38 +03:00
/*
* We ignore any errors here , it ' s most likely
* we just get NT_STATUS_OK or NT_STATUS_IO_TIMEOUT .
*
* In any case we can just give it a retry .
*/
2019-08-13 19:34:36 +03:00
smbd_smb1_do_locks_try ( req ) ;
}
2019-06-13 13:38:57 +03:00
NTSTATUS smbd_smb1_do_locks_recv ( struct tevent_req * req )
{
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
NTSTATUS status = NT_STATUS_OK ;
bool err ;
err = tevent_req_is_nterror ( req , & status ) ;
DBG_DEBUG ( " err=%d, status=%s \n " , ( int ) err , nt_errstr ( status ) ) ;
if ( tevent_req_is_nterror ( req , & status ) ) {
struct files_struct * fsp = state - > fsp ;
struct smbd_lock_element * blocker =
& state - > locks [ state - > blocker ] ;
DBG_DEBUG ( " Setting lock_failure_offset=% " PRIu64 " \n " ,
blocker - > offset ) ;
2020-04-03 10:48:58 +03:00
fsp - > fsp_flags . lock_failure_seen = true ;
2019-06-13 13:38:57 +03:00
fsp - > lock_failure_offset = blocker - > offset ;
return status ;
}
tevent_req_received ( req ) ;
return NT_STATUS_OK ;
}
bool smbd_smb1_do_locks_extract_smbreq (
struct tevent_req * req ,
TALLOC_CTX * mem_ctx ,
struct smb_request * * psmbreq )
{
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
DBG_DEBUG ( " req=%p, state=%p, state->smbreq=%p \n " ,
req ,
state ,
state - > smbreq ) ;
if ( state - > smbreq = = NULL ) {
return false ;
}
* psmbreq = talloc_move ( mem_ctx , & state - > smbreq ) ;
return true ;
}
void smbd_smb1_brl_finish_by_req ( struct tevent_req * req , NTSTATUS status )
{
DBG_DEBUG ( " req=%p, status=%s \n " , req , nt_errstr ( status ) ) ;
if ( NT_STATUS_IS_OK ( status ) ) {
tevent_req_done ( req ) ;
} else {
tevent_req_nterror ( req , status ) ;
}
}
bool smbd_smb1_brl_finish_by_lock (
struct files_struct * fsp ,
bool large_offset ,
enum brl_flavour lock_flav ,
struct smbd_lock_element lock ,
NTSTATUS finish_status )
{
struct tevent_req * * blocked = fsp - > blocked_smb1_lock_reqs ;
size_t num_blocked = talloc_array_length ( blocked ) ;
size_t i ;
DBG_DEBUG ( " num_blocked=%zu \n " , num_blocked ) ;
for ( i = 0 ; i < num_blocked ; i + + ) {
struct tevent_req * req = blocked [ i ] ;
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
uint16_t j ;
DBG_DEBUG ( " i=%zu, req=%p \n " , i , req ) ;
if ( ( state - > large_offset ! = large_offset ) | |
2021-11-17 03:41:09 +03:00
( state - > locks [ 0 ] . lock_flav ! = lock_flav ) ) {
2019-06-13 13:38:57 +03:00
continue ;
}
for ( j = 0 ; j < state - > num_locks ; j + + ) {
struct smbd_lock_element * l = & state - > locks [ j ] ;
if ( ( lock . smblctx = = l - > smblctx ) & &
( lock . offset = = l - > offset ) & &
( lock . count = = l - > count ) ) {
smbd_smb1_brl_finish_by_req (
req , finish_status ) ;
return true ;
}
}
}
return false ;
}
static struct files_struct * smbd_smb1_brl_finish_by_mid_fn (
struct files_struct * fsp , void * private_data )
{
struct tevent_req * * blocked = fsp - > blocked_smb1_lock_reqs ;
size_t num_blocked = talloc_array_length ( blocked ) ;
uint64_t mid = * ( ( uint64_t * ) private_data ) ;
size_t i ;
DBG_DEBUG ( " fsp=%p, num_blocked=%zu \n " , fsp , num_blocked ) ;
for ( i = 0 ; i < num_blocked ; i + + ) {
struct tevent_req * req = blocked [ i ] ;
struct smbd_smb1_do_locks_state * state = tevent_req_data (
req , struct smbd_smb1_do_locks_state ) ;
struct smb_request * smbreq = state - > smbreq ;
if ( smbreq - > mid = = mid ) {
tevent_req_nterror ( req , NT_STATUS_FILE_LOCK_CONFLICT ) ;
return fsp ;
}
}
return NULL ;
}
/*
* This walks the list of fsps , we store the blocked reqs attached to
* them . It can be expensive , but this is legacy SMB1 and trying to
* remember looking at traces I don ' t reall many of those calls .
*/
bool smbd_smb1_brl_finish_by_mid (
struct smbd_server_connection * sconn , uint64_t mid )
{
struct files_struct * found = files_forall (
sconn , smbd_smb1_brl_finish_by_mid_fn , & mid ) ;
return ( found ! = NULL ) ;
}