2019-05-19 15:08:55 +03:00
// SPDX-License-Identifier: GPL-2.0-only
2005-04-17 02:20:36 +04:00
/*
2005-08-12 00:25:47 +04:00
* linux / net / sunrpc / clnt . c
2005-04-17 02:20:36 +04:00
*
* This file contains the high - level RPC interface .
* It is modeled as a finite state machine to support both synchronous
* and asynchronous requests .
*
* - RPC header generation and argument serialization .
* - Credential refresh .
* - TCP connect handling .
* - Retry of operation when it is suspected the operation failed because
* of uid squashing on the server , or when the credentials were stale
* and need to be refreshed , or when a packet was damaged in transit .
* This may be have to be moved to the VFS layer .
*
* Copyright ( C ) 1992 , 1993 Rick Sladkey < jrs @ world . std . com >
* Copyright ( C ) 1995 , 1996 Olaf Kirch < okir @ monad . swb . de >
*/
# include <linux/module.h>
# include <linux/types.h>
2008-05-22 01:09:41 +04:00
# include <linux/kallsyms.h>
2005-04-17 02:20:36 +04:00
# include <linux/mm.h>
2009-08-09 23:14:25 +04:00
# include <linux/namei.h>
# include <linux/mount.h>
2005-04-17 02:20:36 +04:00
# include <linux/slab.h>
2013-10-17 22:12:23 +04:00
# include <linux/rcupdate.h>
2005-04-17 02:20:36 +04:00
# include <linux/utsname.h>
2006-03-20 21:44:22 +03:00
# include <linux/workqueue.h>
2011-05-09 23:22:44 +04:00
# include <linux/in.h>
2007-12-10 22:56:24 +03:00
# include <linux/in6.h>
2011-05-09 23:22:44 +04:00
# include <linux/un.h>
2005-04-17 02:20:36 +04:00
# include <linux/sunrpc/clnt.h>
2013-02-04 21:50:00 +04:00
# include <linux/sunrpc/addr.h>
2005-04-17 02:20:36 +04:00
# include <linux/sunrpc/rpc_pipe_fs.h>
2006-03-20 21:44:22 +03:00
# include <linux/sunrpc/metrics.h>
2009-04-01 17:23:03 +04:00
# include <linux/sunrpc/bc_xprt.h>
2012-02-06 19:08:08 +04:00
# include <trace/events/sunrpc.h>
2005-04-17 02:20:36 +04:00
2009-04-01 17:23:03 +04:00
# include "sunrpc.h"
2012-01-12 22:07:51 +04:00
# include "netns.h"
2005-04-17 02:20:36 +04:00
2014-11-18 00:58:04 +03:00
# if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
2005-04-17 02:20:36 +04:00
# define RPCDBG_FACILITY RPCDBG_CALL
# endif
2007-06-16 22:18:40 +04:00
/*
* All RPC clients are linked into this list
*/
2005-04-17 02:20:36 +04:00
static DECLARE_WAIT_QUEUE_HEAD ( destroy_wait ) ;
static void call_start ( struct rpc_task * task ) ;
static void call_reserve ( struct rpc_task * task ) ;
static void call_reserveresult ( struct rpc_task * task ) ;
static void call_allocate ( struct rpc_task * task ) ;
2018-08-24 23:28:28 +03:00
static void call_encode ( struct rpc_task * task ) ;
2005-04-17 02:20:36 +04:00
static void call_decode ( struct rpc_task * task ) ;
static void call_bind ( struct rpc_task * task ) ;
2005-08-12 00:25:11 +04:00
static void call_bind_status ( struct rpc_task * task ) ;
2005-04-17 02:20:36 +04:00
static void call_transmit ( struct rpc_task * task ) ;
static void call_status ( struct rpc_task * task ) ;
2005-11-10 05:45:24 +03:00
static void call_transmit_status ( struct rpc_task * task ) ;
2005-04-17 02:20:36 +04:00
static void call_refresh ( struct rpc_task * task ) ;
static void call_refreshresult ( struct rpc_task * task ) ;
static void call_connect ( struct rpc_task * task ) ;
static void call_connect_status ( struct rpc_task * task ) ;
2019-02-11 19:24:48 +03:00
static int rpc_encode_header ( struct rpc_task * task ,
struct xdr_stream * xdr ) ;
2019-02-11 19:24:58 +03:00
static int rpc_decode_header ( struct rpc_task * task ,
struct xdr_stream * xdr ) ;
2009-12-03 23:58:56 +03:00
static int rpc_ping ( struct rpc_clnt * clnt ) ;
2019-03-07 22:10:32 +03:00
static void rpc_check_timeout ( struct rpc_task * task ) ;
2007-06-23 18:17:16 +04:00
2007-06-16 22:18:40 +04:00
static void rpc_register_client ( struct rpc_clnt * clnt )
{
2012-03-02 02:00:56 +04:00
struct net * net = rpc_net_ns ( clnt ) ;
struct sunrpc_net * sn = net_generic ( net , sunrpc_net_id ) ;
2012-01-12 22:07:51 +04:00
spin_lock ( & sn - > rpc_client_lock ) ;
list_add ( & clnt - > cl_clients , & sn - > all_clients ) ;
spin_unlock ( & sn - > rpc_client_lock ) ;
2007-06-16 22:18:40 +04:00
}
static void rpc_unregister_client ( struct rpc_clnt * clnt )
{
2012-03-02 02:00:56 +04:00
struct net * net = rpc_net_ns ( clnt ) ;
struct sunrpc_net * sn = net_generic ( net , sunrpc_net_id ) ;
2012-01-12 22:07:51 +04:00
spin_lock ( & sn - > rpc_client_lock ) ;
2007-06-16 22:18:40 +04:00
list_del ( & clnt - > cl_clients ) ;
2012-01-12 22:07:51 +04:00
spin_unlock ( & sn - > rpc_client_lock ) ;
2007-06-16 22:18:40 +04:00
}
2005-04-17 02:20:36 +04:00
2012-01-11 19:18:01 +04:00
static void __rpc_clnt_remove_pipedir ( struct rpc_clnt * clnt )
{
2013-08-27 01:44:26 +04:00
rpc_remove_client_dir ( clnt ) ;
2012-01-11 19:18:01 +04:00
}
static void rpc_clnt_remove_pipedir ( struct rpc_clnt * clnt )
{
2012-03-02 02:00:56 +04:00
struct net * net = rpc_net_ns ( clnt ) ;
2012-01-11 19:18:01 +04:00
struct super_block * pipefs_sb ;
2012-03-02 02:00:56 +04:00
pipefs_sb = rpc_get_sb_net ( net ) ;
2012-01-11 19:18:01 +04:00
if ( pipefs_sb ) {
__rpc_clnt_remove_pipedir ( clnt ) ;
2012-03-02 02:00:56 +04:00
rpc_put_sb_net ( net ) ;
2012-01-11 19:18:01 +04:00
}
}
static struct dentry * rpc_setup_pipedir_sb ( struct super_block * sb ,
2013-08-23 21:03:50 +04:00
struct rpc_clnt * clnt )
2005-04-17 02:20:36 +04:00
{
2005-09-23 19:08:25 +04:00
static uint32_t clntid ;
2013-08-23 21:03:50 +04:00
const char * dir_name = clnt - > cl_program - > pipe_dir_name ;
2009-08-09 23:14:25 +04:00
char name [ 15 ] ;
2012-01-11 19:18:01 +04:00
struct dentry * dir , * dentry ;
2005-04-17 02:20:36 +04:00
2012-01-11 19:18:01 +04:00
dir = rpc_d_lookup_sb ( sb , dir_name ) ;
2012-10-23 18:43:25 +04:00
if ( dir = = NULL ) {
pr_info ( " RPC: pipefs directory doesn't exist: %s \n " , dir_name ) ;
2012-01-11 19:18:01 +04:00
return dir ;
2012-10-23 18:43:25 +04:00
}
2005-09-23 19:08:25 +04:00
for ( ; ; ) {
2013-07-14 16:43:54 +04:00
snprintf ( name , sizeof ( name ) , " clnt%x " , ( unsigned int ) clntid + + ) ;
2009-08-09 23:14:25 +04:00
name [ sizeof ( name ) - 1 ] = ' \0 ' ;
2013-07-14 16:43:54 +04:00
dentry = rpc_create_client_dir ( dir , name , clnt ) ;
2012-01-11 19:18:01 +04:00
if ( ! IS_ERR ( dentry ) )
2009-08-09 23:14:25 +04:00
break ;
2013-07-14 16:43:54 +04:00
if ( dentry = = ERR_PTR ( - EEXIST ) )
continue ;
printk ( KERN_INFO " RPC: Couldn't create pipefs entry "
" %s/%s, error %ld \n " ,
dir_name , name , PTR_ERR ( dentry ) ) ;
break ;
2005-04-17 02:20:36 +04:00
}
2012-01-11 19:18:01 +04:00
dput ( dir ) ;
return dentry ;
}
static int
2013-08-23 21:03:50 +04:00
rpc_setup_pipedir ( struct super_block * pipefs_sb , struct rpc_clnt * clnt )
2012-01-11 19:18:01 +04:00
{
2012-01-11 19:18:42 +04:00
struct dentry * dentry ;
2012-01-11 19:18:01 +04:00
2013-08-27 01:44:26 +04:00
if ( clnt - > cl_program - > pipe_dir_name ! = NULL ) {
dentry = rpc_setup_pipedir_sb ( pipefs_sb , clnt ) ;
if ( IS_ERR ( dentry ) )
return PTR_ERR ( dentry ) ;
}
2009-08-09 23:14:25 +04:00
return 0 ;
2005-04-17 02:20:36 +04:00
}
2013-08-23 21:03:50 +04:00
static int rpc_clnt_skip_event ( struct rpc_clnt * clnt , unsigned long event )
2012-04-27 13:00:17 +04:00
{
2013-08-23 21:03:50 +04:00
if ( clnt - > cl_program - > pipe_dir_name = = NULL )
2013-06-24 11:52:59 +04:00
return 1 ;
2013-08-23 21:03:50 +04:00
2013-08-27 01:44:26 +04:00
switch ( event ) {
case RPC_PIPEFS_MOUNT :
if ( clnt - > cl_pipedir_objects . pdh_dentry ! = NULL )
return 1 ;
if ( atomic_read ( & clnt - > cl_count ) = = 0 )
return 1 ;
break ;
case RPC_PIPEFS_UMOUNT :
if ( clnt - > cl_pipedir_objects . pdh_dentry = = NULL )
return 1 ;
break ;
}
2012-04-27 13:00:17 +04:00
return 0 ;
}
static int __rpc_clnt_handle_event ( struct rpc_clnt * clnt , unsigned long event ,
struct super_block * sb )
2012-01-11 19:18:17 +04:00
{
struct dentry * dentry ;
switch ( event ) {
case RPC_PIPEFS_MOUNT :
2013-08-23 21:03:50 +04:00
dentry = rpc_setup_pipedir_sb ( sb , clnt ) ;
2012-10-23 18:43:25 +04:00
if ( ! dentry )
return - ENOENT ;
2012-01-11 19:18:17 +04:00
if ( IS_ERR ( dentry ) )
return PTR_ERR ( dentry ) ;
break ;
case RPC_PIPEFS_UMOUNT :
__rpc_clnt_remove_pipedir ( clnt ) ;
break ;
default :
printk ( KERN_ERR " %s: unknown event: %ld \n " , __func__ , event ) ;
return - ENOTSUPP ;
}
2016-08-08 12:13:49 +03:00
return 0 ;
2012-01-11 19:18:17 +04:00
}
2012-04-27 13:00:17 +04:00
static int __rpc_pipefs_event ( struct rpc_clnt * clnt , unsigned long event ,
struct super_block * sb )
{
int error = 0 ;
for ( ; ; clnt = clnt - > cl_parent ) {
if ( ! rpc_clnt_skip_event ( clnt , event ) )
error = __rpc_clnt_handle_event ( clnt , event , sb ) ;
if ( error | | clnt = = clnt - > cl_parent )
break ;
}
return error ;
}
2012-02-27 22:05:29 +04:00
static struct rpc_clnt * rpc_get_client_for_event ( struct net * net , int event )
{
struct sunrpc_net * sn = net_generic ( net , sunrpc_net_id ) ;
struct rpc_clnt * clnt ;
spin_lock ( & sn - > rpc_client_lock ) ;
list_for_each_entry ( clnt , & sn - > all_clients , cl_clients ) {
2012-04-27 13:00:17 +04:00
if ( rpc_clnt_skip_event ( clnt , event ) )
2012-02-27 22:05:29 +04:00
continue ;
spin_unlock ( & sn - > rpc_client_lock ) ;
return clnt ;
}
spin_unlock ( & sn - > rpc_client_lock ) ;
return NULL ;
}
2012-01-11 19:18:17 +04:00
static int rpc_pipefs_event ( struct notifier_block * nb , unsigned long event ,
void * ptr )
{
struct super_block * sb = ptr ;
struct rpc_clnt * clnt ;
int error = 0 ;
2012-02-27 22:05:29 +04:00
while ( ( clnt = rpc_get_client_for_event ( sb - > s_fs_info , event ) ) ) {
2012-01-11 19:18:17 +04:00
error = __rpc_pipefs_event ( clnt , event , sb ) ;
if ( error )
break ;
}
return error ;
}
static struct notifier_block rpc_clients_block = {
. notifier_call = rpc_pipefs_event ,
2012-01-10 16:13:19 +04:00
. priority = SUNRPC_PIPEFS_RPC_PRIO ,
2012-01-11 19:18:17 +04:00
} ;
int rpc_clients_notifier_register ( void )
{
return rpc_pipefs_notifier_register ( & rpc_clients_block ) ;
}
void rpc_clients_notifier_unregister ( void )
{
return rpc_pipefs_notifier_unregister ( & rpc_clients_block ) ;
}
2013-10-17 22:12:23 +04:00
static struct rpc_xprt * rpc_clnt_set_transport ( struct rpc_clnt * clnt ,
struct rpc_xprt * xprt ,
const struct rpc_timeout * timeout )
{
struct rpc_xprt * old ;
spin_lock ( & clnt - > cl_lock ) ;
2013-10-29 00:42:44 +04:00
old = rcu_dereference_protected ( clnt - > cl_xprt ,
lockdep_is_held ( & clnt - > cl_lock ) ) ;
2013-10-17 22:12:23 +04:00
if ( ! xprt_bound ( xprt ) )
clnt - > cl_autobind = 1 ;
clnt - > cl_timeout = timeout ;
rcu_assign_pointer ( clnt - > cl_xprt , xprt ) ;
spin_unlock ( & clnt - > cl_lock ) ;
return old ;
}
2012-04-30 19:52:40 +04:00
static void rpc_clnt_set_nodename ( struct rpc_clnt * clnt , const char * nodename )
{
2015-01-31 02:12:28 +03:00
clnt - > cl_nodelen = strlcpy ( clnt - > cl_nodename ,
nodename , sizeof ( clnt - > cl_nodename ) ) ;
2012-04-30 19:52:40 +04:00
}
2013-10-17 22:12:17 +04:00
static int rpc_client_register ( struct rpc_clnt * clnt ,
rpc_authflavor_t pseudoflavor ,
const char * client_name )
2013-06-24 11:52:52 +04:00
{
2013-08-27 03:23:04 +04:00
struct rpc_auth_create_args auth_args = {
2013-10-17 22:12:17 +04:00
. pseudoflavor = pseudoflavor ,
. target_name = client_name ,
2013-08-27 03:23:04 +04:00
} ;
2013-06-24 11:52:52 +04:00
struct rpc_auth * auth ;
struct net * net = rpc_net_ns ( clnt ) ;
struct super_block * pipefs_sb ;
2013-07-10 23:33:01 +04:00
int err ;
2013-06-24 11:52:52 +04:00
2015-03-31 19:03:28 +03:00
rpc_clnt_debugfs_register ( clnt ) ;
2014-11-26 22:44:43 +03:00
2013-06-24 11:52:52 +04:00
pipefs_sb = rpc_get_sb_net ( net ) ;
if ( pipefs_sb ) {
2013-08-23 21:03:50 +04:00
err = rpc_setup_pipedir ( pipefs_sb , clnt ) ;
2013-06-24 11:52:52 +04:00
if ( err )
goto out ;
}
2013-07-10 23:33:01 +04:00
rpc_register_client ( clnt ) ;
if ( pipefs_sb )
rpc_put_sb_net ( net ) ;
2013-08-27 03:23:04 +04:00
auth = rpcauth_create ( & auth_args , clnt ) ;
2013-06-24 11:52:52 +04:00
if ( IS_ERR ( auth ) ) {
dprintk ( " RPC: Couldn't create auth handle (flavor %u) \n " ,
2013-10-17 22:12:17 +04:00
pseudoflavor ) ;
2013-06-24 11:52:52 +04:00
err = PTR_ERR ( auth ) ;
goto err_auth ;
}
2013-07-10 23:33:01 +04:00
return 0 ;
err_auth :
pipefs_sb = rpc_get_sb_net ( net ) ;
2013-07-15 06:57:50 +04:00
rpc_unregister_client ( clnt ) ;
2013-07-10 23:33:01 +04:00
__rpc_clnt_remove_pipedir ( clnt ) ;
2013-06-24 11:52:52 +04:00
out :
if ( pipefs_sb )
rpc_put_sb_net ( net ) ;
2014-11-26 22:44:43 +03:00
rpc_clnt_debugfs_unregister ( clnt ) ;
2013-06-24 11:52:52 +04:00
return err ;
}
2013-09-05 05:51:44 +04:00
static DEFINE_IDA ( rpc_clids ) ;
2017-01-20 11:48:39 +03:00
void rpc_cleanup_clids ( void )
{
ida_destroy ( & rpc_clids ) ;
}
2013-09-05 05:51:44 +04:00
static int rpc_alloc_clid ( struct rpc_clnt * clnt )
{
int clid ;
clid = ida_simple_get ( & rpc_clids , 0 , 0 , GFP_KERNEL ) ;
if ( clid < 0 )
return clid ;
clnt - > cl_clid = clid ;
return 0 ;
}
static void rpc_free_clid ( struct rpc_clnt * clnt )
{
ida_simple_remove ( & rpc_clids , clnt - > cl_clid ) ;
}
2013-09-02 21:28:04 +04:00
static struct rpc_clnt * rpc_new_client ( const struct rpc_create_args * args ,
2016-01-30 22:17:26 +03:00
struct rpc_xprt_switch * xps ,
2013-09-02 21:28:04 +04:00
struct rpc_xprt * xprt ,
struct rpc_clnt * parent )
2005-04-17 02:20:36 +04:00
{
2012-01-20 22:53:56 +04:00
const struct rpc_program * program = args - > program ;
const struct rpc_version * version ;
2013-10-17 22:12:23 +04:00
struct rpc_clnt * clnt = NULL ;
const struct rpc_timeout * timeout ;
2015-01-31 02:12:28 +03:00
const char * nodename = args - > nodename ;
2005-04-17 02:20:36 +04:00
int err ;
2007-09-12 02:00:20 +04:00
2007-06-15 01:26:17 +04:00
err = rpciod_up ( ) ;
if ( err )
goto out_no_rpciod ;
2007-12-21 00:03:53 +03:00
2013-04-05 22:13:21 +04:00
err = - EINVAL ;
2007-12-21 00:03:53 +03:00
if ( args - > version > = program - > nrvers )
goto out_err ;
version = program - > version [ args - > version ] ;
if ( version = = NULL )
2005-04-17 02:20:36 +04:00
goto out_err ;
err = - ENOMEM ;
2006-07-22 01:51:30 +04:00
clnt = kzalloc ( sizeof ( * clnt ) , GFP_KERNEL ) ;
2005-04-17 02:20:36 +04:00
if ( ! clnt )
goto out_err ;
2013-09-02 21:28:04 +04:00
clnt - > cl_parent = parent ? : clnt ;
2005-04-17 02:20:36 +04:00
2013-09-05 05:51:44 +04:00
err = rpc_alloc_clid ( clnt ) ;
if ( err )
goto out_no_clid ;
2005-04-17 02:20:36 +04:00
2019-04-25 00:46:42 +03:00
clnt - > cl_cred = get_cred ( args - > cred ) ;
2005-04-17 02:20:36 +04:00
clnt - > cl_procinfo = version - > procs ;
clnt - > cl_maxproc = version - > nrprocs ;
2008-09-28 10:21:26 +04:00
clnt - > cl_prog = args - > prognumber ? : program - > number ;
2005-04-17 02:20:36 +04:00
clnt - > cl_vers = version - > number ;
clnt - > cl_stats = program - > stats ;
2006-03-20 21:44:22 +03:00
clnt - > cl_metrics = rpc_alloc_iostats ( clnt ) ;
2013-08-26 23:38:11 +04:00
rpc_init_pipe_dir_head ( & clnt - > cl_pipedir_objects ) ;
2006-11-21 18:40:23 +03:00
err = - ENOMEM ;
if ( clnt - > cl_metrics = = NULL )
goto out_no_stats ;
2006-11-16 19:37:27 +03:00
clnt - > cl_program = program ;
2007-06-15 00:40:14 +04:00
INIT_LIST_HEAD ( & clnt - > cl_tasks ) ;
2007-06-16 22:17:01 +04:00
spin_lock_init ( & clnt - > cl_lock ) ;
2005-04-17 02:20:36 +04:00
2013-10-17 22:12:23 +04:00
timeout = xprt - > timeout ;
2007-12-21 00:03:55 +03:00
if ( args - > timeout ! = NULL ) {
memcpy ( & clnt - > cl_timeout_default , args - > timeout ,
sizeof ( clnt - > cl_timeout_default ) ) ;
2013-10-17 22:12:23 +04:00
timeout = & clnt - > cl_timeout_default ;
2007-12-21 00:03:55 +03:00
}
2013-10-17 22:12:23 +04:00
rpc_clnt_set_transport ( clnt , xprt , timeout ) ;
2016-01-30 22:17:26 +03:00
xprt_iter_init ( & clnt - > cl_xpi , xps ) ;
xprt_switch_put ( xps ) ;
2013-10-17 22:12:23 +04:00
2005-04-17 02:20:36 +04:00
clnt - > cl_rtt = & clnt - > cl_rtt_default ;
2007-12-21 00:03:55 +03:00
rpc_init_rtt ( & clnt - > cl_rtt_default , clnt - > cl_timeout - > to_initval ) ;
2005-04-17 02:20:36 +04:00
2010-09-13 03:55:25 +04:00
atomic_set ( & clnt - > cl_count , 1 ) ;
2007-06-15 00:40:31 +04:00
2015-01-31 02:12:28 +03:00
if ( nodename = = NULL )
nodename = utsname ( ) - > nodename ;
2005-04-17 02:20:36 +04:00
/* save the nodename */
2015-01-31 02:12:28 +03:00
rpc_clnt_set_nodename ( clnt , nodename ) ;
2013-06-24 11:52:52 +04:00
2013-10-17 22:12:17 +04:00
err = rpc_client_register ( clnt , args - > authflavor , args - > client_name ) ;
2013-06-24 11:52:52 +04:00
if ( err )
goto out_no_path ;
2013-09-02 21:28:04 +04:00
if ( parent )
atomic_inc ( & parent - > cl_count ) ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_new ( clnt , xprt , program - > name , args - > servername ) ;
2005-04-17 02:20:36 +04:00
return clnt ;
out_no_path :
2006-11-21 18:40:23 +03:00
rpc_free_iostats ( clnt - > cl_metrics ) ;
out_no_stats :
2019-04-25 00:46:42 +03:00
put_cred ( clnt - > cl_cred ) ;
2013-09-05 05:51:44 +04:00
rpc_free_clid ( clnt ) ;
out_no_clid :
2005-04-17 02:20:36 +04:00
kfree ( clnt ) ;
out_err :
2007-06-15 01:26:17 +04:00
rpciod_down ( ) ;
out_no_rpciod :
2016-01-30 22:17:26 +03:00
xprt_switch_put ( xps ) ;
2013-04-05 22:13:21 +04:00
xprt_put ( xprt ) ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_new_err ( program - > name , args - > servername , err ) ;
2005-04-17 02:20:36 +04:00
return ERR_PTR ( err ) ;
}
2016-05-17 00:03:42 +03:00
static struct rpc_clnt * rpc_create_xprt ( struct rpc_create_args * args ,
2014-03-24 07:58:59 +04:00
struct rpc_xprt * xprt )
{
struct rpc_clnt * clnt = NULL ;
2016-01-30 22:17:26 +03:00
struct rpc_xprt_switch * xps ;
2014-03-24 07:58:59 +04:00
2016-05-17 19:38:21 +03:00
if ( args - > bc_xprt & & args - > bc_xprt - > xpt_bc_xps ) {
2016-08-22 21:57:42 +03:00
WARN_ON_ONCE ( ! ( args - > protocol & XPRT_TRANSPORT_BC ) ) ;
2016-05-17 19:38:21 +03:00
xps = args - > bc_xprt - > xpt_bc_xps ;
xprt_switch_get ( xps ) ;
} else {
xps = xprt_switch_alloc ( xprt , GFP_KERNEL ) ;
if ( xps = = NULL ) {
xprt_put ( xprt ) ;
return ERR_PTR ( - ENOMEM ) ;
}
if ( xprt - > bc_xprt ) {
xprt_switch_get ( xps ) ;
xprt - > bc_xprt - > xpt_bc_xps = xps ;
}
2016-05-21 00:07:17 +03:00
}
2016-01-30 22:17:26 +03:00
clnt = rpc_new_client ( args , xps , xprt , NULL ) ;
2014-03-24 07:58:59 +04:00
if ( IS_ERR ( clnt ) )
return clnt ;
if ( ! ( args - > flags & RPC_CLNT_CREATE_NOPING ) ) {
int err = rpc_ping ( clnt ) ;
if ( err ! = 0 ) {
rpc_shutdown_client ( clnt ) ;
return ERR_PTR ( err ) ;
}
}
clnt - > cl_softrtry = 1 ;
2019-04-07 20:58:58 +03:00
if ( args - > flags & ( RPC_CLNT_CREATE_HARDRTRY | RPC_CLNT_CREATE_SOFTERR ) ) {
2014-03-24 07:58:59 +04:00
clnt - > cl_softrtry = 0 ;
2019-04-07 20:58:58 +03:00
if ( args - > flags & RPC_CLNT_CREATE_SOFTERR )
clnt - > cl_softerr = 1 ;
}
2014-03-24 07:58:59 +04:00
if ( args - > flags & RPC_CLNT_CREATE_AUTOBIND )
clnt - > cl_autobind = 1 ;
2014-09-25 06:35:58 +04:00
if ( args - > flags & RPC_CLNT_CREATE_NO_RETRANS_TIMEOUT )
clnt - > cl_noretranstimeo = 1 ;
2014-03-24 07:58:59 +04:00
if ( args - > flags & RPC_CLNT_CREATE_DISCRTRY )
clnt - > cl_discrtry = 1 ;
if ( ! ( args - > flags & RPC_CLNT_CREATE_QUIET ) )
clnt - > cl_chatty = 1 ;
return clnt ;
}
2012-07-10 14:55:09 +04:00
/**
2006-08-23 04:06:20 +04:00
* rpc_create - create an RPC client and transport with one call
* @ args : rpc_clnt create argument structure
*
* Creates and initializes an RPC transport and an RPC client .
*
* It can ping the server in order to determine if it is up , and to see if
* it supports this program and version . RPC_CLNT_CREATE_NOPING disables
* this behavior so asynchronous tasks can also use rpc_create .
*/
struct rpc_clnt * rpc_create ( struct rpc_create_args * args )
{
struct rpc_xprt * xprt ;
2007-09-10 21:47:07 +04:00
struct xprt_create xprtargs = {
2010-09-29 16:05:12 +04:00
. net = args - > net ,
2007-09-10 21:47:57 +04:00
. ident = args - > protocol ,
2007-07-10 00:23:35 +04:00
. srcaddr = args - > saddress ,
2007-07-08 15:08:54 +04:00
. dstaddr = args - > address ,
. addrlen = args - > addrsize ,
2012-03-02 02:01:05 +04:00
. servername = args - > servername ,
2009-09-10 18:33:30 +04:00
. bc_xprt = args - > bc_xprt ,
2007-07-08 15:08:54 +04:00
} ;
2007-12-10 22:56:24 +03:00
char servername [ 48 ] ;
2017-04-27 15:50:51 +03:00
struct rpc_clnt * clnt ;
int i ;
2006-08-23 04:06:20 +04:00
2016-05-17 00:03:42 +03:00
if ( args - > bc_xprt ) {
2016-08-22 21:57:42 +03:00
WARN_ON_ONCE ( ! ( args - > protocol & XPRT_TRANSPORT_BC ) ) ;
2016-05-17 00:03:42 +03:00
xprt = args - > bc_xprt - > xpt_bc_xprt ;
if ( xprt ) {
xprt_get ( xprt ) ;
return rpc_create_xprt ( args , xprt ) ;
}
}
2013-04-14 19:42:00 +04:00
if ( args - > flags & RPC_CLNT_CREATE_INFINITE_SLOTS )
xprtargs . flags | = XPRT_CREATE_INFINITE_SLOTS ;
2013-04-11 23:06:36 +04:00
if ( args - > flags & RPC_CLNT_CREATE_NO_IDLE_TIMEOUT )
xprtargs . flags | = XPRT_CREATE_NO_IDLE_TIMEOUT ;
2007-07-01 20:13:22 +04:00
/*
* If the caller chooses not to specify a hostname , whip
* up a string representation of the passed - in address .
*/
2012-03-02 02:01:05 +04:00
if ( xprtargs . servername = = NULL ) {
2011-05-09 23:22:44 +04:00
struct sockaddr_un * sun =
( struct sockaddr_un * ) args - > address ;
2011-05-09 23:22:25 +04:00
struct sockaddr_in * sin =
( struct sockaddr_in * ) args - > address ;
struct sockaddr_in6 * sin6 =
( struct sockaddr_in6 * ) args - > address ;
2007-12-10 22:56:24 +03:00
servername [ 0 ] = ' \0 ' ;
switch ( args - > address - > sa_family ) {
2011-05-09 23:22:44 +04:00
case AF_LOCAL :
snprintf ( servername , sizeof ( servername ) , " %s " ,
sun - > sun_path ) ;
break ;
2011-05-09 23:22:25 +04:00
case AF_INET :
2008-10-31 10:54:56 +03:00
snprintf ( servername , sizeof ( servername ) , " %pI4 " ,
& sin - > sin_addr . s_addr ) ;
2007-12-10 22:56:24 +03:00
break ;
2011-05-09 23:22:25 +04:00
case AF_INET6 :
2008-10-29 22:52:50 +03:00
snprintf ( servername , sizeof ( servername ) , " %pI6 " ,
2011-05-09 23:22:25 +04:00
& sin6 - > sin6_addr ) ;
2007-12-10 22:56:24 +03:00
break ;
default :
/* caller wants default server name, but
* address family isn ' t recognized . */
return ERR_PTR ( - EINVAL ) ;
}
2012-03-02 02:01:05 +04:00
xprtargs . servername = servername ;
2007-07-01 20:13:22 +04:00
}
2007-12-10 22:56:24 +03:00
xprt = xprt_create_transport ( & xprtargs ) ;
if ( IS_ERR ( xprt ) )
return ( struct rpc_clnt * ) xprt ;
2006-08-23 04:06:20 +04:00
/*
* By default , kernel RPC client connects from a reserved port .
* CAP_NET_BIND_SERVICE will not be set for unprivileged requesters ,
* but it is always enabled for rpciod , which handles the connect
* operation .
*/
xprt - > resvport = 1 ;
if ( args - > flags & RPC_CLNT_CREATE_NONPRIVPORT )
xprt - > resvport = 0 ;
2019-10-17 18:13:54 +03:00
xprt - > reuseport = 0 ;
if ( args - > flags & RPC_CLNT_CREATE_REUSEPORT )
xprt - > reuseport = 1 ;
2006-08-23 04:06:20 +04:00
2017-04-27 15:50:51 +03:00
clnt = rpc_create_xprt ( args , xprt ) ;
if ( IS_ERR ( clnt ) | | args - > nconnect < = 1 )
return clnt ;
for ( i = 0 ; i < args - > nconnect - 1 ; i + + ) {
if ( rpc_clnt_add_xprt ( clnt , & xprtargs , NULL , NULL ) < 0 )
break ;
}
return clnt ;
2006-08-23 04:06:20 +04:00
}
2006-08-23 04:06:22 +04:00
EXPORT_SYMBOL_GPL ( rpc_create ) ;
2006-08-23 04:06:20 +04:00
2005-04-17 02:20:36 +04:00
/*
* This function clones the RPC client structure . It allows us to share the
* same transport while varying parameters such as the authentication
* flavour .
*/
2012-09-15 01:23:52 +04:00
static struct rpc_clnt * __rpc_clone_client ( struct rpc_create_args * args ,
struct rpc_clnt * clnt )
2005-04-17 02:20:36 +04:00
{
2016-01-30 22:17:26 +03:00
struct rpc_xprt_switch * xps ;
2012-03-02 02:00:56 +04:00
struct rpc_xprt * xprt ;
2012-09-15 01:23:52 +04:00
struct rpc_clnt * new ;
int err ;
2005-04-17 02:20:36 +04:00
2012-09-15 01:23:52 +04:00
err = - ENOMEM ;
2012-03-02 02:00:56 +04:00
rcu_read_lock ( ) ;
xprt = xprt_get ( rcu_dereference ( clnt - > cl_xprt ) ) ;
2016-01-30 22:17:26 +03:00
xps = xprt_switch_get ( rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ) ;
2012-03-02 02:00:56 +04:00
rcu_read_unlock ( ) ;
2016-01-30 22:17:26 +03:00
if ( xprt = = NULL | | xps = = NULL ) {
xprt_put ( xprt ) ;
xprt_switch_put ( xps ) ;
2012-09-15 01:23:52 +04:00
goto out_err ;
2016-01-30 22:17:26 +03:00
}
2012-09-15 01:23:52 +04:00
args - > servername = xprt - > servername ;
2015-01-31 02:12:28 +03:00
args - > nodename = clnt - > cl_nodename ;
2012-09-15 01:23:52 +04:00
2016-01-30 22:17:26 +03:00
new = rpc_new_client ( args , xps , xprt , clnt ) ;
2020-05-13 00:13:39 +03:00
if ( IS_ERR ( new ) )
return new ;
2012-09-15 01:23:52 +04:00
/* Turn off autobind on clones */
new - > cl_autobind = 0 ;
new - > cl_softrtry = clnt - > cl_softrtry ;
2019-04-07 20:58:58 +03:00
new - > cl_softerr = clnt - > cl_softerr ;
2014-09-25 06:35:58 +04:00
new - > cl_noretranstimeo = clnt - > cl_noretranstimeo ;
2012-09-15 01:23:52 +04:00
new - > cl_discrtry = clnt - > cl_discrtry ;
new - > cl_chatty = clnt - > cl_chatty ;
2018-12-03 03:30:30 +03:00
new - > cl_principal = clnt - > cl_principal ;
2005-04-17 02:20:36 +04:00
return new ;
2012-09-15 01:23:52 +04:00
out_err :
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_clone_err ( clnt , err ) ;
2006-11-16 19:37:27 +03:00
return ERR_PTR ( err ) ;
2005-04-17 02:20:36 +04:00
}
2012-09-15 01:23:52 +04:00
/**
* rpc_clone_client - Clone an RPC client structure
*
* @ clnt : RPC client whose parameters are copied
*
* Returns a fresh RPC client or an ERR_PTR .
*/
struct rpc_clnt * rpc_clone_client ( struct rpc_clnt * clnt )
{
struct rpc_create_args args = {
. program = clnt - > cl_program ,
. prognumber = clnt - > cl_prog ,
. version = clnt - > cl_vers ,
. authflavor = clnt - > cl_auth - > au_flavor ,
2019-04-25 00:46:42 +03:00
. cred = clnt - > cl_cred ,
2012-09-15 01:23:52 +04:00
} ;
return __rpc_clone_client ( & args , clnt ) ;
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_clone_client ) ;
2005-04-17 02:20:36 +04:00
2012-09-15 01:24:02 +04:00
/**
* rpc_clone_client_set_auth - Clone an RPC client structure and set its auth
*
* @ clnt : RPC client whose parameters are copied
2013-01-10 05:12:35 +04:00
* @ flavor : security flavor for new client
2012-09-15 01:24:02 +04:00
*
* Returns a fresh RPC client or an ERR_PTR .
*/
struct rpc_clnt *
rpc_clone_client_set_auth ( struct rpc_clnt * clnt , rpc_authflavor_t flavor )
{
struct rpc_create_args args = {
. program = clnt - > cl_program ,
. prognumber = clnt - > cl_prog ,
. version = clnt - > cl_vers ,
. authflavor = flavor ,
2019-04-25 00:46:42 +03:00
. cred = clnt - > cl_cred ,
2012-09-15 01:24:02 +04:00
} ;
return __rpc_clone_client ( & args , clnt ) ;
}
EXPORT_SYMBOL_GPL ( rpc_clone_client_set_auth ) ;
2013-10-17 22:12:23 +04:00
/**
* rpc_switch_client_transport : switch the RPC transport on the fly
* @ clnt : pointer to a struct rpc_clnt
* @ args : pointer to the new transport arguments
* @ timeout : pointer to the new timeout parameters
*
* This function allows the caller to switch the RPC transport for the
* rpc_clnt structure ' clnt ' to allow it to connect to a mirrored NFS
* server , for instance . It assumes that the caller has ensured that
* there are no active RPC tasks by using some form of locking .
*
* Returns zero if " clnt " is now using the new xprt . Otherwise a
* negative errno is returned , and " clnt " continues to use the old
* xprt .
*/
int rpc_switch_client_transport ( struct rpc_clnt * clnt ,
struct xprt_create * args ,
const struct rpc_timeout * timeout )
{
const struct rpc_timeout * old_timeo ;
rpc_authflavor_t pseudoflavor ;
2016-01-30 22:17:26 +03:00
struct rpc_xprt_switch * xps , * oldxps ;
2013-10-17 22:12:23 +04:00
struct rpc_xprt * xprt , * old ;
struct rpc_clnt * parent ;
int err ;
xprt = xprt_create_transport ( args ) ;
2020-05-13 00:13:39 +03:00
if ( IS_ERR ( xprt ) )
2013-10-17 22:12:23 +04:00
return PTR_ERR ( xprt ) ;
2016-01-30 22:17:26 +03:00
xps = xprt_switch_alloc ( xprt , GFP_KERNEL ) ;
if ( xps = = NULL ) {
xprt_put ( xprt ) ;
return - ENOMEM ;
}
2013-10-17 22:12:23 +04:00
pseudoflavor = clnt - > cl_auth - > au_flavor ;
old_timeo = clnt - > cl_timeout ;
old = rpc_clnt_set_transport ( clnt , xprt , timeout ) ;
2016-01-30 22:17:26 +03:00
oldxps = xprt_iter_xchg_switch ( & clnt - > cl_xpi , xps ) ;
2013-10-17 22:12:23 +04:00
rpc_unregister_client ( clnt ) ;
__rpc_clnt_remove_pipedir ( clnt ) ;
2014-11-26 22:44:43 +03:00
rpc_clnt_debugfs_unregister ( clnt ) ;
2013-10-17 22:12:23 +04:00
/*
* A new transport was created . " clnt " therefore
* becomes the root of a new cl_parent tree . clnt ' s
* children , if it has any , still point to the old xprt .
*/
parent = clnt - > cl_parent ;
clnt - > cl_parent = clnt ;
/*
* The old rpc_auth cache cannot be re - used . GSS
* contexts in particular are between a single
* client and server .
*/
err = rpc_client_register ( clnt , pseudoflavor , NULL ) ;
if ( err )
goto out_revert ;
synchronize_rcu ( ) ;
if ( parent ! = clnt )
rpc_release_client ( parent ) ;
2016-01-30 22:17:26 +03:00
xprt_switch_put ( oldxps ) ;
2013-10-17 22:12:23 +04:00
xprt_put ( old ) ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_replace_xprt ( clnt ) ;
2013-10-17 22:12:23 +04:00
return 0 ;
out_revert :
2016-01-30 22:17:26 +03:00
xps = xprt_iter_xchg_switch ( & clnt - > cl_xpi , oldxps ) ;
2013-10-17 22:12:23 +04:00
rpc_clnt_set_transport ( clnt , old , old_timeo ) ;
clnt - > cl_parent = parent ;
rpc_client_register ( clnt , pseudoflavor , NULL ) ;
2016-01-30 22:17:26 +03:00
xprt_switch_put ( xps ) ;
2013-10-17 22:12:23 +04:00
xprt_put ( xprt ) ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_replace_xprt_err ( clnt ) ;
2013-10-17 22:12:23 +04:00
return err ;
}
EXPORT_SYMBOL_GPL ( rpc_switch_client_transport ) ;
2016-01-31 04:39:19 +03:00
static
int rpc_clnt_xprt_iter_init ( struct rpc_clnt * clnt , struct rpc_xprt_iter * xpi )
{
struct rpc_xprt_switch * xps ;
rcu_read_lock ( ) ;
xps = xprt_switch_get ( rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ) ;
rcu_read_unlock ( ) ;
if ( xps = = NULL )
return - EAGAIN ;
xprt_iter_init_listall ( xpi , xps ) ;
xprt_switch_put ( xps ) ;
return 0 ;
}
/**
* rpc_clnt_iterate_for_each_xprt - Apply a function to all transports
* @ clnt : pointer to client
* @ fn : function to apply
* @ data : void pointer to function data
*
* Iterates through the list of RPC transports currently attached to the
* client and applies the function fn ( clnt , xprt , data ) .
*
* On error , the iteration stops , and the function returns the error value .
*/
int rpc_clnt_iterate_for_each_xprt ( struct rpc_clnt * clnt ,
int ( * fn ) ( struct rpc_clnt * , struct rpc_xprt * , void * ) ,
void * data )
{
struct rpc_xprt_iter xpi ;
int ret ;
ret = rpc_clnt_xprt_iter_init ( clnt , & xpi ) ;
if ( ret )
return ret ;
for ( ; ; ) {
struct rpc_xprt * xprt = xprt_iter_get_next ( & xpi ) ;
if ( ! xprt )
break ;
ret = fn ( clnt , xprt , data ) ;
xprt_put ( xprt ) ;
if ( ret < 0 )
break ;
}
xprt_iter_destroy ( & xpi ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( rpc_clnt_iterate_for_each_xprt ) ;
2010-07-31 22:29:08 +04:00
/*
* Kill all tasks for the given client .
* XXX : kill their descendants as well ?
*/
void rpc_killall_tasks ( struct rpc_clnt * clnt )
{
struct rpc_task * rovr ;
if ( list_empty ( & clnt - > cl_tasks ) )
return ;
2020-05-13 00:13:39 +03:00
2010-07-31 22:29:08 +04:00
/*
* Spin lock all_tasks to prevent changes . . .
*/
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_killall ( clnt ) ;
2010-07-31 22:29:08 +04:00
spin_lock ( & clnt - > cl_lock ) ;
2019-04-07 20:58:44 +03:00
list_for_each_entry ( rovr , & clnt - > cl_tasks , tk_task )
rpc_signal_task ( rovr ) ;
2010-07-31 22:29:08 +04:00
spin_unlock ( & clnt - > cl_lock ) ;
}
EXPORT_SYMBOL_GPL ( rpc_killall_tasks ) ;
2005-04-17 02:20:36 +04:00
/*
* Properly shut down an RPC client , terminating all outstanding
2007-06-10 03:49:36 +04:00
* requests .
2005-04-17 02:20:36 +04:00
*/
2007-06-15 00:40:32 +04:00
void rpc_shutdown_client ( struct rpc_clnt * clnt )
2005-04-17 02:20:36 +04:00
{
2012-10-31 01:01:40 +04:00
might_sleep ( ) ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_shutdown ( clnt ) ;
2005-04-17 02:20:36 +04:00
2007-06-15 00:40:31 +04:00
while ( ! list_empty ( & clnt - > cl_tasks ) ) {
2005-04-17 02:20:36 +04:00
rpc_killall_tasks ( clnt ) ;
2006-01-10 07:52:53 +03:00
wait_event_timeout ( destroy_wait ,
2007-06-15 00:40:31 +04:00
list_empty ( & clnt - > cl_tasks ) , 1 * HZ ) ;
2005-04-17 02:20:36 +04:00
}
2007-06-15 00:40:32 +04:00
rpc_release_client ( clnt ) ;
2005-04-17 02:20:36 +04:00
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_shutdown_client ) ;
2005-04-17 02:20:36 +04:00
/*
2007-06-15 00:40:31 +04:00
* Free an RPC client
2005-04-17 02:20:36 +04:00
*/
2020-04-03 06:33:41 +03:00
static void rpc_free_client_work ( struct work_struct * work )
{
struct rpc_clnt * clnt = container_of ( work , struct rpc_clnt , cl_work ) ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_free ( clnt ) ;
2020-04-03 06:33:41 +03:00
/* These might block on processes that might allocate memory,
* so they cannot be called in rpciod , so they are handled separately
* here .
*/
rpc_clnt_debugfs_unregister ( clnt ) ;
2020-05-14 01:38:40 +03:00
rpc_free_clid ( clnt ) ;
2020-04-03 06:33:41 +03:00
rpc_clnt_remove_pipedir ( clnt ) ;
2020-05-09 00:25:11 +03:00
xprt_put ( rcu_dereference_raw ( clnt - > cl_xprt ) ) ;
2020-04-03 06:33:41 +03:00
kfree ( clnt ) ;
rpciod_down ( ) ;
}
2013-11-13 02:24:36 +04:00
static struct rpc_clnt *
2010-09-13 03:55:25 +04:00
rpc_free_client ( struct rpc_clnt * clnt )
2005-04-17 02:20:36 +04:00
{
2013-11-13 02:24:36 +04:00
struct rpc_clnt * parent = NULL ;
2020-05-13 00:13:39 +03:00
trace_rpc_clnt_release ( clnt ) ;
2012-01-20 22:53:37 +04:00
if ( clnt - > cl_parent ! = clnt )
2013-11-13 02:24:36 +04:00
parent = clnt - > cl_parent ;
2013-06-26 10:15:14 +04:00
rpc_unregister_client ( clnt ) ;
2006-03-20 21:44:22 +03:00
rpc_free_iostats ( clnt - > cl_metrics ) ;
clnt - > cl_metrics = NULL ;
2016-01-30 22:17:26 +03:00
xprt_iter_destroy ( & clnt - > cl_xpi ) ;
2019-04-25 00:46:42 +03:00
put_cred ( clnt - > cl_cred ) ;
2020-04-03 06:33:41 +03:00
INIT_WORK ( & clnt - > cl_work , rpc_free_client_work ) ;
schedule_work ( & clnt - > cl_work ) ;
2013-11-13 02:24:36 +04:00
return parent ;
2005-04-17 02:20:36 +04:00
}
2007-06-27 00:57:41 +04:00
/*
* Free an RPC client
*/
2018-07-24 22:29:15 +03:00
static struct rpc_clnt *
2010-09-13 03:55:25 +04:00
rpc_free_auth ( struct rpc_clnt * clnt )
2007-06-27 00:57:41 +04:00
{
2013-11-13 02:24:36 +04:00
if ( clnt - > cl_auth = = NULL )
return rpc_free_client ( clnt ) ;
2007-06-27 00:57:41 +04:00
/*
* Note : RPCSEC_GSS may need to send NULL RPC calls in order to
* release remaining GSS contexts . This mechanism ensures
* that it can do so safely .
*/
2010-09-13 03:55:25 +04:00
atomic_inc ( & clnt - > cl_count ) ;
2007-06-27 00:57:41 +04:00
rpcauth_release ( clnt - > cl_auth ) ;
clnt - > cl_auth = NULL ;
2010-09-13 03:55:25 +04:00
if ( atomic_dec_and_test ( & clnt - > cl_count ) )
2013-11-13 02:24:36 +04:00
return rpc_free_client ( clnt ) ;
return NULL ;
2007-06-27 00:57:41 +04:00
}
2005-04-17 02:20:36 +04:00
/*
2007-06-15 00:40:31 +04:00
* Release reference to the RPC client
2005-04-17 02:20:36 +04:00
*/
void
rpc_release_client ( struct rpc_clnt * clnt )
{
2013-11-13 02:24:36 +04:00
do {
if ( list_empty ( & clnt - > cl_tasks ) )
wake_up ( & destroy_wait ) ;
if ( ! atomic_dec_and_test ( & clnt - > cl_count ) )
break ;
clnt = rpc_free_auth ( clnt ) ;
} while ( clnt ! = NULL ) ;
2007-06-15 00:40:31 +04:00
}
2012-05-26 02:09:55 +04:00
EXPORT_SYMBOL_GPL ( rpc_release_client ) ;
2007-06-15 00:40:31 +04:00
2005-06-22 21:16:23 +04:00
/**
* rpc_bind_new_program - bind a new RPC program to an existing client
2008-02-14 02:03:23 +03:00
* @ old : old rpc_client
* @ program : rpc program to set
* @ vers : rpc program version
2005-06-22 21:16:23 +04:00
*
* Clones the rpc client and sets up a new RPC program . This is mainly
* of use for enabling different RPC programs to share the same transport .
* The Sun NFSv2 / v3 ACL protocol can do this .
*/
struct rpc_clnt * rpc_bind_new_program ( struct rpc_clnt * old ,
2012-01-20 22:53:56 +04:00
const struct rpc_program * program ,
2007-09-12 02:00:09 +04:00
u32 vers )
2005-06-22 21:16:23 +04:00
{
2012-11-01 20:14:14 +04:00
struct rpc_create_args args = {
. program = program ,
. prognumber = program - > number ,
. version = vers ,
. authflavor = old - > cl_auth - > au_flavor ,
2019-04-25 00:46:42 +03:00
. cred = old - > cl_cred ,
2012-11-01 20:14:14 +04:00
} ;
2005-06-22 21:16:23 +04:00
struct rpc_clnt * clnt ;
int err ;
2012-11-01 20:14:14 +04:00
clnt = __rpc_clone_client ( & args , old ) ;
2005-06-22 21:16:23 +04:00
if ( IS_ERR ( clnt ) )
goto out ;
2009-12-03 23:58:56 +03:00
err = rpc_ping ( clnt ) ;
2005-06-22 21:16:23 +04:00
if ( err ! = 0 ) {
rpc_shutdown_client ( clnt ) ;
clnt = ERR_PTR ( err ) ;
}
2007-02-10 02:38:13 +03:00
out :
2005-06-22 21:16:23 +04:00
return clnt ;
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_bind_new_program ) ;
2005-06-22 21:16:23 +04:00
2019-07-11 23:33:12 +03:00
struct rpc_xprt *
rpc_task_get_xprt ( struct rpc_clnt * clnt , struct rpc_xprt * xprt )
2017-04-28 17:52:42 +03:00
{
struct rpc_xprt_switch * xps ;
if ( ! xprt )
return NULL ;
rcu_read_lock ( ) ;
xps = rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ;
atomic_long_inc ( & xps - > xps_queuelen ) ;
rcu_read_unlock ( ) ;
atomic_long_inc ( & xprt - > queuelen ) ;
return xprt ;
}
static void
rpc_task_release_xprt ( struct rpc_clnt * clnt , struct rpc_xprt * xprt )
{
struct rpc_xprt_switch * xps ;
atomic_long_dec ( & xprt - > queuelen ) ;
rcu_read_lock ( ) ;
xps = rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ;
atomic_long_dec ( & xps - > xps_queuelen ) ;
rcu_read_unlock ( ) ;
xprt_put ( xprt ) ;
}
2018-06-20 00:24:58 +03:00
void rpc_task_release_transport ( struct rpc_task * task )
{
struct rpc_xprt * xprt = task - > tk_xprt ;
if ( xprt ) {
task - > tk_xprt = NULL ;
2017-04-28 17:52:42 +03:00
if ( task - > tk_client )
rpc_task_release_xprt ( task - > tk_client , xprt ) ;
else
xprt_put ( xprt ) ;
2018-06-20 00:24:58 +03:00
}
}
EXPORT_SYMBOL_GPL ( rpc_task_release_transport ) ;
2010-07-31 22:29:08 +04:00
void rpc_task_release_client ( struct rpc_task * task )
{
struct rpc_clnt * clnt = task - > tk_client ;
2017-04-28 17:52:42 +03:00
rpc_task_release_transport ( task ) ;
2010-07-31 22:29:08 +04:00
if ( clnt ! = NULL ) {
/* Remove from client task list */
spin_lock ( & clnt - > cl_lock ) ;
list_del ( & task - > tk_task ) ;
spin_unlock ( & clnt - > cl_lock ) ;
task - > tk_client = NULL ;
rpc_release_client ( clnt ) ;
}
2018-06-20 00:24:58 +03:00
}
2016-01-31 00:39:26 +03:00
2019-07-11 23:33:12 +03:00
static struct rpc_xprt *
rpc_task_get_first_xprt ( struct rpc_clnt * clnt )
{
struct rpc_xprt * xprt ;
rcu_read_lock ( ) ;
xprt = xprt_get ( rcu_dereference ( clnt - > cl_xprt ) ) ;
rcu_read_unlock ( ) ;
return rpc_task_get_xprt ( clnt , xprt ) ;
}
static struct rpc_xprt *
rpc_task_get_next_xprt ( struct rpc_clnt * clnt )
{
return rpc_task_get_xprt ( clnt , xprt_iter_get_next ( & clnt - > cl_xpi ) ) ;
}
2018-06-20 00:24:58 +03:00
static
void rpc_task_set_transport ( struct rpc_task * task , struct rpc_clnt * clnt )
{
2019-05-30 03:41:28 +03:00
if ( task - > tk_xprt )
return ;
if ( task - > tk_flags & RPC_TASK_NO_ROUND_ROBIN )
task - > tk_xprt = rpc_task_get_first_xprt ( clnt ) ;
else
2019-07-11 23:33:12 +03:00
task - > tk_xprt = rpc_task_get_next_xprt ( clnt ) ;
2010-07-31 22:29:08 +04:00
}
static
void rpc_task_set_client ( struct rpc_task * task , struct rpc_clnt * clnt )
{
2016-01-31 00:39:26 +03:00
2010-07-31 22:29:08 +04:00
if ( clnt ! = NULL ) {
2018-06-20 00:24:58 +03:00
rpc_task_set_transport ( task , clnt ) ;
2010-07-31 22:29:08 +04:00
task - > tk_client = clnt ;
2010-09-13 03:55:25 +04:00
atomic_inc ( & clnt - > cl_count ) ;
2010-07-31 22:29:08 +04:00
if ( clnt - > cl_softrtry )
task - > tk_flags | = RPC_TASK_SOFT ;
2019-04-07 20:58:58 +03:00
if ( clnt - > cl_softerr )
task - > tk_flags | = RPC_TASK_TIMEOUT ;
2013-09-24 20:00:27 +04:00
if ( clnt - > cl_noretranstimeo )
task - > tk_flags | = RPC_TASK_NO_RETRANS_TIMEOUT ;
2015-06-03 23:14:25 +03:00
if ( atomic_read ( & clnt - > cl_swapper ) )
task - > tk_flags | = RPC_TASK_SWAPPER ;
2010-07-31 22:29:08 +04:00
/* Add to the client's list of all tasks */
spin_lock ( & clnt - > cl_lock ) ;
list_add_tail ( & task - > tk_task , & clnt - > cl_tasks ) ;
spin_unlock ( & clnt - > cl_lock ) ;
}
}
static void
rpc_task_set_rpc_message ( struct rpc_task * task , const struct rpc_message * msg )
{
if ( msg ! = NULL ) {
task - > tk_msg . rpc_proc = msg - > rpc_proc ;
task - > tk_msg . rpc_argp = msg - > rpc_argp ;
task - > tk_msg . rpc_resp = msg - > rpc_resp ;
2020-02-08 03:11:12 +03:00
task - > tk_msg . rpc_cred = msg - > rpc_cred ;
if ( ! ( task - > tk_flags & RPC_TASK_CRED_NOREF ) )
get_cred ( task - > tk_msg . rpc_cred ) ;
2010-07-31 22:29:08 +04:00
}
}
2005-04-17 02:20:36 +04:00
/*
* Default callback for async RPC calls
*/
static void
2006-01-03 11:55:04 +03:00
rpc_default_callback ( struct rpc_task * task , void * data )
2005-04-17 02:20:36 +04:00
{
}
2006-01-03 11:55:04 +03:00
static const struct rpc_call_ops rpc_default_ops = {
. rpc_call_done = rpc_default_callback ,
} ;
2007-07-14 23:39:59 +04:00
/**
* rpc_run_task - Allocate a new RPC task , then run rpc_execute against it
* @ task_setup_data : pointer to task initialisation data
*/
struct rpc_task * rpc_run_task ( const struct rpc_task_setup * task_setup_data )
2007-06-12 18:02:37 +04:00
{
2010-04-17 00:41:10 +04:00
struct rpc_task * task ;
2007-06-12 18:02:37 +04:00
2007-07-14 23:39:59 +04:00
task = rpc_new_task ( task_setup_data ) ;
2007-06-12 18:02:37 +04:00
2020-02-08 03:16:34 +03:00
if ( ! RPC_IS_ASYNC ( task ) )
task - > tk_flags | = RPC_TASK_CRED_NOREF ;
2010-07-31 22:29:08 +04:00
rpc_task_set_client ( task , task_setup_data - > rpc_client ) ;
rpc_task_set_rpc_message ( task , task_setup_data - > rpc_message ) ;
if ( task - > tk_action = = NULL )
rpc_call_start ( task ) ;
2007-06-12 18:02:37 +04:00
atomic_inc ( & task - > tk_count ) ;
rpc_execute ( task ) ;
2010-04-17 00:41:10 +04:00
return task ;
2007-06-12 18:02:37 +04:00
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_run_task ) ;
2007-06-12 18:02:37 +04:00
/**
* rpc_call_sync - Perform a synchronous RPC call
* @ clnt : pointer to RPC client
* @ msg : RPC call parameters
* @ flags : RPC call flags
2005-04-17 02:20:36 +04:00
*/
2008-02-14 19:11:30 +03:00
int rpc_call_sync ( struct rpc_clnt * clnt , const struct rpc_message * msg , int flags )
2005-04-17 02:20:36 +04:00
{
struct rpc_task * task ;
2007-07-14 23:39:59 +04:00
struct rpc_task_setup task_setup_data = {
. rpc_client = clnt ,
. rpc_message = msg ,
. callback_ops = & rpc_default_ops ,
. flags = flags ,
} ;
2007-06-12 18:02:37 +04:00
int status ;
2005-04-17 02:20:36 +04:00
2012-11-02 00:04:40 +04:00
WARN_ON_ONCE ( flags & RPC_TASK_ASYNC ) ;
if ( flags & RPC_TASK_ASYNC ) {
rpc_release_calldata ( task_setup_data . callback_ops ,
task_setup_data . callback_data ) ;
return - EINVAL ;
}
2005-04-17 02:20:36 +04:00
2007-07-14 23:39:59 +04:00
task = rpc_run_task ( & task_setup_data ) ;
2007-06-12 18:02:37 +04:00
if ( IS_ERR ( task ) )
return PTR_ERR ( task ) ;
2006-01-03 11:55:10 +03:00
status = task - > tk_status ;
2007-01-24 22:54:53 +03:00
rpc_put_task ( task ) ;
2005-04-17 02:20:36 +04:00
return status ;
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_call_sync ) ;
2005-04-17 02:20:36 +04:00
2007-06-12 18:02:37 +04:00
/**
* rpc_call_async - Perform an asynchronous RPC call
* @ clnt : pointer to RPC client
* @ msg : RPC call parameters
* @ flags : RPC call flags
2008-02-14 02:03:23 +03:00
* @ tk_ops : RPC call ops
2007-06-12 18:02:37 +04:00
* @ data : user call data
2005-04-17 02:20:36 +04:00
*/
int
2008-02-14 19:11:30 +03:00
rpc_call_async ( struct rpc_clnt * clnt , const struct rpc_message * msg , int flags ,
2006-01-03 11:55:04 +03:00
const struct rpc_call_ops * tk_ops , void * data )
2005-04-17 02:20:36 +04:00
{
struct rpc_task * task ;
2007-07-14 23:39:59 +04:00
struct rpc_task_setup task_setup_data = {
. rpc_client = clnt ,
. rpc_message = msg ,
. callback_ops = tk_ops ,
. callback_data = data ,
. flags = flags | RPC_TASK_ASYNC ,
} ;
2005-04-17 02:20:36 +04:00
2007-07-14 23:39:59 +04:00
task = rpc_run_task ( & task_setup_data ) ;
2007-06-12 18:02:37 +04:00
if ( IS_ERR ( task ) )
return PTR_ERR ( task ) ;
rpc_put_task ( task ) ;
return 0 ;
2005-04-17 02:20:36 +04:00
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_call_async ) ;
2005-04-17 02:20:36 +04:00
2011-07-14 03:20:49 +04:00
# if defined(CONFIG_SUNRPC_BACKCHANNEL)
2019-03-05 15:30:48 +03:00
static void call_bc_encode ( struct rpc_task * task ) ;
2009-04-01 17:23:03 +04:00
/**
* rpc_run_bc_task - Allocate a new RPC task for backchannel use , then run
* rpc_execute against it
2009-09-24 22:58:42 +04:00
* @ req : RPC request
2009-04-01 17:23:03 +04:00
*/
2015-06-02 05:59:08 +03:00
struct rpc_task * rpc_run_bc_task ( struct rpc_rqst * req )
2009-04-01 17:23:03 +04:00
{
struct rpc_task * task ;
struct rpc_task_setup task_setup_data = {
2015-06-02 05:59:08 +03:00
. callback_ops = & rpc_default_ops ,
2018-08-24 23:28:28 +03:00
. flags = RPC_TASK_SOFTCONN |
RPC_TASK_NO_RETRANS_TIMEOUT ,
2009-04-01 17:23:03 +04:00
} ;
dprintk ( " RPC: rpc_run_bc_task req= %p \n " , req ) ;
/*
* Create an rpc_task to send the data
*/
task = rpc_new_task ( & task_setup_data ) ;
2018-09-02 00:21:01 +03:00
xprt_init_bc_request ( req , task ) ;
2009-04-01 17:23:03 +04:00
2019-03-05 15:30:48 +03:00
task - > tk_action = call_bc_encode ;
2009-04-01 17:23:03 +04:00
atomic_inc ( & task - > tk_count ) ;
2012-10-23 18:43:27 +04:00
WARN_ON_ONCE ( atomic_read ( & task - > tk_count ) ! = 2 ) ;
2009-04-01 17:23:03 +04:00
rpc_execute ( task ) ;
dprintk ( " RPC: rpc_run_bc_task: task= %p \n " , task ) ;
return task ;
}
2011-07-14 03:20:49 +04:00
# endif /* CONFIG_SUNRPC_BACKCHANNEL */
2009-04-01 17:23:03 +04:00
2019-02-11 19:25:20 +03:00
/**
* rpc_prepare_reply_pages - Prepare to receive a reply data payload into pages
* @ req : RPC request to prepare
* @ pages : vector of struct page pointers
* @ base : offset in first page where receive should start , in bytes
* @ len : expected size of the upper layer data payload , in bytes
* @ hdrsize : expected size of upper layer reply header , in XDR words
*
*/
void rpc_prepare_reply_pages ( struct rpc_rqst * req , struct page * * pages ,
unsigned int base , unsigned int len ,
unsigned int hdrsize )
{
2020-11-22 04:46:18 +03:00
hdrsize + = RPC_REPHDRSIZE + req - > rq_cred - > cr_auth - > au_ralign ;
2019-02-11 19:25:25 +03:00
2019-02-11 19:25:20 +03:00
xdr_inline_pages ( & req - > rq_rcv_buf , hdrsize < < 2 , pages , base , len ) ;
2020-05-13 00:13:28 +03:00
trace_rpc_xdr_reply_pages ( req - > rq_task , & req - > rq_rcv_buf ) ;
2019-02-11 19:25:20 +03:00
}
EXPORT_SYMBOL_GPL ( rpc_prepare_reply_pages ) ;
2007-10-26 02:40:21 +04:00
void
rpc_call_start ( struct rpc_task * task )
{
task - > tk_action = call_start ;
}
EXPORT_SYMBOL_GPL ( rpc_call_start ) ;
2006-08-23 04:06:17 +04:00
/**
* rpc_peeraddr - extract remote peer address from clnt ' s xprt
* @ clnt : RPC client structure
* @ buf : target buffer
2008-02-14 02:03:23 +03:00
* @ bufsize : length of target buffer
2006-08-23 04:06:17 +04:00
*
* Returns the number of bytes that are actually in the stored address .
*/
size_t rpc_peeraddr ( struct rpc_clnt * clnt , struct sockaddr * buf , size_t bufsize )
{
size_t bytes ;
2012-03-02 02:00:56 +04:00
struct rpc_xprt * xprt ;
rcu_read_lock ( ) ;
xprt = rcu_dereference ( clnt - > cl_xprt ) ;
2006-08-23 04:06:17 +04:00
2012-03-02 02:00:56 +04:00
bytes = xprt - > addrlen ;
2006-08-23 04:06:17 +04:00
if ( bytes > bufsize )
bytes = bufsize ;
2012-03-02 02:00:56 +04:00
memcpy ( buf , & xprt - > addr , bytes ) ;
rcu_read_unlock ( ) ;
return bytes ;
2006-08-23 04:06:17 +04:00
}
2006-08-23 04:06:22 +04:00
EXPORT_SYMBOL_GPL ( rpc_peeraddr ) ;
2006-08-23 04:06:17 +04:00
2006-08-23 04:06:18 +04:00
/**
* rpc_peeraddr2str - return remote peer address in printable format
* @ clnt : RPC client structure
* @ format : address format
*
2012-03-02 02:00:56 +04:00
* NB : the lifetime of the memory referenced by the returned pointer is
* the same as the rpc_xprt itself . As long as the caller uses this
* pointer , it must hold the RCU read lock .
2006-08-23 04:06:18 +04:00
*/
2008-01-08 02:34:48 +03:00
const char * rpc_peeraddr2str ( struct rpc_clnt * clnt ,
enum rpc_display_format_t format )
2006-08-23 04:06:18 +04:00
{
2012-03-02 02:00:56 +04:00
struct rpc_xprt * xprt ;
xprt = rcu_dereference ( clnt - > cl_xprt ) ;
2006-12-06 00:35:37 +03:00
if ( xprt - > address_strings [ format ] ! = NULL )
return xprt - > address_strings [ format ] ;
else
return " unprintable " ;
2006-08-23 04:06:18 +04:00
}
2006-08-23 04:06:22 +04:00
EXPORT_SYMBOL_GPL ( rpc_peeraddr2str ) ;
2006-08-23 04:06:18 +04:00
2012-03-02 02:01:14 +04:00
static const struct sockaddr_in rpc_inaddr_loopback = {
. sin_family = AF_INET ,
. sin_addr . s_addr = htonl ( INADDR_ANY ) ,
} ;
static const struct sockaddr_in6 rpc_in6addr_loopback = {
. sin6_family = AF_INET6 ,
. sin6_addr = IN6ADDR_ANY_INIT ,
} ;
/*
* Try a getsockname ( ) on a connected datagram socket . Using a
* connected datagram socket prevents leaving a socket in TIME_WAIT .
* This conserves the ephemeral port number space .
*
* Returns zero and fills in " buf " if successful ; otherwise , a
* negative errno is returned .
*/
static int rpc_sockname ( struct net * net , struct sockaddr * sap , size_t salen ,
2018-02-12 22:00:20 +03:00
struct sockaddr * buf )
2012-03-02 02:01:14 +04:00
{
struct socket * sock ;
int err ;
err = __sock_create ( net , sap - > sa_family ,
SOCK_DGRAM , IPPROTO_UDP , & sock , 1 ) ;
if ( err < 0 ) {
dprintk ( " RPC: can't create UDP socket (%d) \n " , err ) ;
goto out ;
}
switch ( sap - > sa_family ) {
case AF_INET :
err = kernel_bind ( sock ,
( struct sockaddr * ) & rpc_inaddr_loopback ,
sizeof ( rpc_inaddr_loopback ) ) ;
break ;
case AF_INET6 :
err = kernel_bind ( sock ,
( struct sockaddr * ) & rpc_in6addr_loopback ,
sizeof ( rpc_in6addr_loopback ) ) ;
break ;
default :
err = - EAFNOSUPPORT ;
goto out ;
}
if ( err < 0 ) {
dprintk ( " RPC: can't bind UDP socket (%d) \n " , err ) ;
goto out_release ;
}
err = kernel_connect ( sock , sap , salen , 0 ) ;
if ( err < 0 ) {
dprintk ( " RPC: can't connect UDP socket (%d) \n " , err ) ;
goto out_release ;
}
2018-02-12 22:00:20 +03:00
err = kernel_getsockname ( sock , buf ) ;
2012-03-02 02:01:14 +04:00
if ( err < 0 ) {
dprintk ( " RPC: getsockname failed (%d) \n " , err ) ;
goto out_release ;
}
err = 0 ;
if ( buf - > sa_family = = AF_INET6 ) {
struct sockaddr_in6 * sin6 = ( struct sockaddr_in6 * ) buf ;
sin6 - > sin6_scope_id = 0 ;
}
dprintk ( " RPC: %s succeeded \n " , __func__ ) ;
out_release :
sock_release ( sock ) ;
out :
return err ;
}
/*
* Scraping a connected socket failed , so we don ' t have a useable
* local address . Fallback : generate an address that will prevent
* the server from calling us back .
*
* Returns zero and fills in " buf " if successful ; otherwise , a
* negative errno is returned .
*/
static int rpc_anyaddr ( int family , struct sockaddr * buf , size_t buflen )
{
switch ( family ) {
case AF_INET :
if ( buflen < sizeof ( rpc_inaddr_loopback ) )
return - EINVAL ;
memcpy ( buf , & rpc_inaddr_loopback ,
sizeof ( rpc_inaddr_loopback ) ) ;
break ;
case AF_INET6 :
if ( buflen < sizeof ( rpc_in6addr_loopback ) )
return - EINVAL ;
memcpy ( buf , & rpc_in6addr_loopback ,
sizeof ( rpc_in6addr_loopback ) ) ;
2015-12-31 02:14:06 +03:00
break ;
2012-03-02 02:01:14 +04:00
default :
dprintk ( " RPC: %s: address family not supported \n " ,
__func__ ) ;
return - EAFNOSUPPORT ;
}
dprintk ( " RPC: %s: succeeded \n " , __func__ ) ;
return 0 ;
}
/**
* rpc_localaddr - discover local endpoint address for an RPC client
* @ clnt : RPC client structure
* @ buf : target buffer
* @ buflen : size of target buffer , in bytes
*
* Returns zero and fills in " buf " and " buflen " if successful ;
* otherwise , a negative errno is returned .
*
* This works even if the underlying transport is not currently connected ,
* or if the upper layer never previously provided a source address .
*
* The result of this function call is transient : multiple calls in
* succession may give different results , depending on how local
* networking configuration changes over time .
*/
int rpc_localaddr ( struct rpc_clnt * clnt , struct sockaddr * buf , size_t buflen )
{
struct sockaddr_storage address ;
struct sockaddr * sap = ( struct sockaddr * ) & address ;
struct rpc_xprt * xprt ;
struct net * net ;
size_t salen ;
int err ;
rcu_read_lock ( ) ;
xprt = rcu_dereference ( clnt - > cl_xprt ) ;
salen = xprt - > addrlen ;
memcpy ( sap , & xprt - > addr , salen ) ;
net = get_net ( xprt - > xprt_net ) ;
rcu_read_unlock ( ) ;
rpc_set_port ( sap , 0 ) ;
2018-02-12 22:00:20 +03:00
err = rpc_sockname ( net , sap , salen , buf ) ;
2012-03-02 02:01:14 +04:00
put_net ( net ) ;
if ( err ! = 0 )
/* Couldn't discover local address, return ANYADDR */
return rpc_anyaddr ( sap - > sa_family , buf , buflen ) ;
return 0 ;
}
EXPORT_SYMBOL_GPL ( rpc_localaddr ) ;
2005-04-17 02:20:36 +04:00
void
rpc_setbufsize ( struct rpc_clnt * clnt , unsigned int sndsize , unsigned int rcvsize )
{
2012-03-02 02:00:56 +04:00
struct rpc_xprt * xprt ;
rcu_read_lock ( ) ;
xprt = rcu_dereference ( clnt - > cl_xprt ) ;
2005-08-26 03:25:56 +04:00
if ( xprt - > ops - > set_buffer_size )
xprt - > ops - > set_buffer_size ( xprt , sndsize , rcvsize ) ;
2012-03-02 02:00:56 +04:00
rcu_read_unlock ( ) ;
2005-04-17 02:20:36 +04:00
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_setbufsize ) ;
2005-04-17 02:20:36 +04:00
2012-03-02 02:00:56 +04:00
/**
* rpc_net_ns - Get the network namespace for this RPC client
* @ clnt : RPC client to query
*
*/
struct net * rpc_net_ns ( struct rpc_clnt * clnt )
{
struct net * ret ;
rcu_read_lock ( ) ;
ret = rcu_dereference ( clnt - > cl_xprt ) - > xprt_net ;
rcu_read_unlock ( ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( rpc_net_ns ) ;
/**
* rpc_max_payload - Get maximum payload size for a transport , in bytes
* @ clnt : RPC client to query
2005-04-17 02:20:36 +04:00
*
* For stream transports , this is one RPC record fragment ( see RFC
* 1831 ) , as we don ' t support multi - record requests yet . For datagram
* transports , this is the size of an IP packet minus the IP , UDP , and
* RPC header sizes .
*/
size_t rpc_max_payload ( struct rpc_clnt * clnt )
{
2012-03-02 02:00:56 +04:00
size_t ret ;
rcu_read_lock ( ) ;
ret = rcu_dereference ( clnt - > cl_xprt ) - > max_payload ;
rcu_read_unlock ( ) ;
return ret ;
2005-04-17 02:20:36 +04:00
}
2006-08-23 04:06:22 +04:00
EXPORT_SYMBOL_GPL ( rpc_max_payload ) ;
2005-04-17 02:20:36 +04:00
2016-05-02 21:40:40 +03:00
/**
* rpc_max_bc_payload - Get maximum backchannel payload size , in bytes
* @ clnt : RPC client to query
*/
size_t rpc_max_bc_payload ( struct rpc_clnt * clnt )
{
struct rpc_xprt * xprt ;
size_t ret ;
rcu_read_lock ( ) ;
xprt = rcu_dereference ( clnt - > cl_xprt ) ;
ret = xprt - > ops - > bc_maxpayload ( xprt ) ;
rcu_read_unlock ( ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( rpc_max_bc_payload ) ;
2019-07-16 20:51:29 +03:00
unsigned int rpc_num_bc_slots ( struct rpc_clnt * clnt )
{
struct rpc_xprt * xprt ;
unsigned int ret ;
rcu_read_lock ( ) ;
xprt = rcu_dereference ( clnt - > cl_xprt ) ;
ret = xprt - > ops - > bc_num_slots ( xprt ) ;
rcu_read_unlock ( ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( rpc_num_bc_slots ) ;
2006-01-03 11:55:50 +03:00
/**
* rpc_force_rebind - force transport to check that remote port is unchanged
* @ clnt : client to rebind
*
*/
void rpc_force_rebind ( struct rpc_clnt * clnt )
{
2012-03-02 02:00:56 +04:00
if ( clnt - > cl_autobind ) {
rcu_read_lock ( ) ;
xprt_clear_bound ( rcu_dereference ( clnt - > cl_xprt ) ) ;
rcu_read_unlock ( ) ;
}
2006-01-03 11:55:50 +03:00
}
2006-08-23 04:06:22 +04:00
EXPORT_SYMBOL_GPL ( rpc_force_rebind ) ;
2006-01-03 11:55:50 +03:00
2019-04-07 20:58:45 +03:00
static int
__rpc_restart_call ( struct rpc_task * task , void ( * action ) ( struct rpc_task * ) )
2009-04-01 17:22:40 +04:00
{
2014-03-20 20:59:09 +04:00
task - > tk_status = 0 ;
2019-04-07 20:58:54 +03:00
task - > tk_rpc_status = 0 ;
2019-04-07 20:58:45 +03:00
task - > tk_action = action ;
2010-07-31 22:29:07 +04:00
return 1 ;
2009-04-01 17:22:40 +04:00
}
2005-04-17 02:20:36 +04:00
/*
* Restart an ( async ) RPC call . Usually called from within the
* exit handler .
*/
2010-07-31 22:29:07 +04:00
int
2005-04-17 02:20:36 +04:00
rpc_restart_call ( struct rpc_task * task )
{
2019-04-07 20:58:45 +03:00
return __rpc_restart_call ( task , call_start ) ;
2005-04-17 02:20:36 +04:00
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_restart_call ) ;
2005-04-17 02:20:36 +04:00
2019-04-07 20:58:45 +03:00
/*
* Restart an ( async ) RPC call from the call_prepare state .
* Usually called from within the exit handler .
*/
int
rpc_restart_call_prepare ( struct rpc_task * task )
{
if ( task - > tk_ops - > rpc_call_prepare ! = NULL )
return __rpc_restart_call ( task , rpc_prepare_task ) ;
return rpc_restart_call ( task ) ;
}
EXPORT_SYMBOL_GPL ( rpc_restart_call_prepare ) ;
2014-11-26 22:44:43 +03:00
const char
* rpc_proc_name ( const struct rpc_task * task )
2008-05-22 01:09:12 +04:00
{
const struct rpc_procinfo * proc = task - > tk_msg . rpc_proc ;
if ( proc ) {
if ( proc - > p_name )
return proc - > p_name ;
else
return " NULL " ;
} else
return " no proc " ;
}
2019-04-07 20:58:54 +03:00
static void
__rpc_call_rpcerror ( struct rpc_task * task , int tk_status , int rpc_status )
{
2020-05-13 00:13:23 +03:00
trace_rpc_call_rpcerror ( task , tk_status , rpc_status ) ;
2019-04-07 20:58:54 +03:00
task - > tk_rpc_status = rpc_status ;
rpc_exit ( task , tk_status ) ;
}
static void
rpc_call_rpcerror ( struct rpc_task * task , int status )
{
__rpc_call_rpcerror ( task , status , status ) ;
}
2005-04-17 02:20:36 +04:00
/*
* 0. Initial state
*
* Other FSM states can be visited zero or more times , but
* this state is visited exactly once for each RPC .
*/
static void
call_start ( struct rpc_task * task )
{
struct rpc_clnt * clnt = task - > tk_client ;
2017-05-09 00:27:10 +03:00
int idx = task - > tk_msg . rpc_proc - > p_statidx ;
2005-04-17 02:20:36 +04:00
2017-11-03 20:46:14 +03:00
trace_rpc_request ( task ) ;
2005-04-17 02:20:36 +04:00
2017-05-09 00:27:10 +03:00
/* Increment call count (version might not be valid for ping) */
if ( clnt - > cl_program - > version [ clnt - > cl_vers ] )
clnt - > cl_program - > version [ clnt - > cl_vers ] - > counts [ idx ] + + ;
2005-04-17 02:20:36 +04:00
clnt - > cl_stats - > rpccnt + + ;
task - > tk_action = call_reserve ;
2018-06-20 00:24:58 +03:00
rpc_task_set_transport ( task , clnt ) ;
2005-04-17 02:20:36 +04:00
}
/*
* 1. Reserve an RPC call slot
*/
static void
call_reserve ( struct rpc_task * task )
{
task - > tk_status = 0 ;
task - > tk_action = call_reserveresult ;
xprt_reserve ( task ) ;
}
2013-04-14 18:49:37 +04:00
static void call_retry_reserve ( struct rpc_task * task ) ;
2005-04-17 02:20:36 +04:00
/*
* 1 b . Grok the result of xprt_reserve ( )
*/
static void
call_reserveresult ( struct rpc_task * task )
{
int status = task - > tk_status ;
/*
* After a call to xprt_reserve ( ) , we must have either
* a request slot or else an error status .
*/
task - > tk_status = 0 ;
if ( status > = 0 ) {
if ( task - > tk_rqstp ) {
2010-09-13 03:55:25 +04:00
task - > tk_action = call_refresh ;
2005-04-17 02:20:36 +04:00
return ;
}
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , - EIO ) ;
2005-04-17 02:20:36 +04:00
return ;
}
/*
* Even though there was an error , we may have acquired
* a request slot somehow . Make sure not to leak it .
*/
2019-10-09 19:58:08 +03:00
if ( task - > tk_rqstp )
2005-04-17 02:20:36 +04:00
xprt_release ( task ) ;
switch ( status ) {
2012-05-19 20:12:53 +04:00
case - ENOMEM :
rpc_delay ( task , HZ > > 2 ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2005-04-17 02:20:36 +04:00
case - EAGAIN : /* woken up; retry */
2013-04-14 18:49:37 +04:00
task - > tk_action = call_retry_reserve ;
2005-04-17 02:20:36 +04:00
return ;
default :
2019-10-09 19:58:08 +03:00
rpc_call_rpcerror ( task , status ) ;
2005-04-17 02:20:36 +04:00
}
}
2013-04-14 18:49:37 +04:00
/*
* 1 c . Retry reserving an RPC call slot
*/
static void
call_retry_reserve ( struct rpc_task * task )
{
task - > tk_status = 0 ;
task - > tk_action = call_reserveresult ;
xprt_retry_reserve ( task ) ;
}
2005-04-17 02:20:36 +04:00
/*
2010-09-13 03:55:25 +04:00
* 2. Bind and / or refresh the credentials
*/
static void
call_refresh ( struct rpc_task * task )
{
task - > tk_action = call_refreshresult ;
task - > tk_status = 0 ;
task - > tk_client - > cl_stats - > rpcauthrefresh + + ;
rpcauth_refreshcred ( task ) ;
}
/*
* 2 a . Process the results of a credential refresh
*/
static void
call_refreshresult ( struct rpc_task * task )
{
int status = task - > tk_status ;
task - > tk_status = 0 ;
2010-11-20 19:13:31 +03:00
task - > tk_action = call_refresh ;
2010-09-13 03:55:25 +04:00
switch ( status ) {
2010-11-20 19:13:31 +03:00
case 0 :
2013-12-17 21:16:11 +04:00
if ( rpcauth_uptodatecred ( task ) ) {
2010-11-20 19:13:31 +03:00
task - > tk_action = call_allocate ;
2013-12-17 21:16:11 +04:00
return ;
}
/* Use rate-limiting and a max number of retries if refresh
* had status 0 but failed to update the cred .
*/
2020-08-24 01:36:59 +03:00
fallthrough ;
2010-09-13 03:55:25 +04:00
case - ETIMEDOUT :
rpc_delay ( task , 3 * HZ ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2010-11-20 19:13:31 +03:00
case - EAGAIN :
status = - EACCES ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2013-08-14 19:59:13 +04:00
case - EKEYEXPIRED :
2010-11-20 19:13:31 +03:00
if ( ! task - > tk_cred_retry )
break ;
task - > tk_cred_retry - - ;
2020-07-08 23:10:03 +03:00
trace_rpc_retry_refresh_status ( task ) ;
2010-11-20 19:13:31 +03:00
return ;
2010-09-13 03:55:25 +04:00
}
2020-07-08 23:10:03 +03:00
trace_rpc_refresh_status ( task ) ;
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , status ) ;
2010-09-13 03:55:25 +04:00
}
/*
* 2 b . Allocate the buffer . For details , see sched . c : rpc_malloc .
2006-01-03 11:55:49 +03:00
* ( Note : buffer memory is freed in xprt_release ) .
2005-04-17 02:20:36 +04:00
*/
static void
call_allocate ( struct rpc_task * task )
{
2019-02-11 19:25:41 +03:00
const struct rpc_auth * auth = task - > tk_rqstp - > rq_cred - > cr_auth ;
2006-01-03 11:55:49 +03:00
struct rpc_rqst * req = task - > tk_rqstp ;
2013-01-08 18:10:21 +04:00
struct rpc_xprt * xprt = req - > rq_xprt ;
2017-05-12 16:36:49 +03:00
const struct rpc_procinfo * proc = task - > tk_msg . rpc_proc ;
2016-09-15 17:55:20 +03:00
int status ;
2005-04-17 02:20:36 +04:00
2007-03-30 00:47:53 +04:00
task - > tk_status = 0 ;
2018-08-24 23:28:28 +03:00
task - > tk_action = call_encode ;
2007-03-30 00:47:53 +04:00
2019-04-11 22:16:52 +03:00
if ( req - > rq_buffer )
2005-04-17 02:20:36 +04:00
return ;
2007-03-30 00:47:53 +04:00
if ( proc - > p_proc ! = 0 ) {
BUG_ON ( proc - > p_arglen = = 0 ) ;
if ( proc - > p_decode ! = NULL )
BUG_ON ( proc - > p_replen = = 0 ) ;
}
2005-04-17 02:20:36 +04:00
2007-03-30 00:47:53 +04:00
/*
* Calculate the size ( in quads ) of the RPC call
* and reply headers , and convert both values
* to byte sizes .
*/
2019-02-11 19:25:41 +03:00
req - > rq_callsize = RPC_CALLHDRSIZE + ( auth - > au_cslack < < 1 ) +
proc - > p_arglen ;
2007-03-30 00:47:53 +04:00
req - > rq_callsize < < = 2 ;
2019-03-15 19:55:59 +03:00
/*
* Note : the reply buffer must at minimum allocate enough space
* for the ' struct accepted_reply ' from RFC5531 .
*/
req - > rq_rcvsize = RPC_REPHDRSIZE + auth - > au_rslack + \
max_t ( size_t , proc - > p_replen , 2 ) ;
2007-03-30 00:47:53 +04:00
req - > rq_rcvsize < < = 2 ;
2016-09-15 17:55:20 +03:00
status = xprt - > ops - > buf_alloc ( task ) ;
2020-07-08 23:09:11 +03:00
trace_rpc_buf_alloc ( task , status ) ;
2015-05-11 21:02:25 +03:00
xprt_inject_disconnect ( xprt ) ;
2019-04-11 22:16:52 +03:00
if ( status = = 0 )
2016-09-15 17:55:20 +03:00
return ;
if ( status ! = - ENOMEM ) {
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , status ) ;
2016-09-15 17:55:20 +03:00
return ;
}
2007-01-31 20:14:08 +03:00
2011-06-17 18:14:59 +04:00
if ( RPC_IS_ASYNC ( task ) | | ! fatal_signal_pending ( current ) ) {
2007-10-01 20:06:44 +04:00
task - > tk_action = call_allocate ;
2005-04-17 02:20:36 +04:00
rpc_delay ( task , HZ > > 4 ) ;
return ;
}
2019-09-12 15:06:51 +03:00
rpc_call_rpcerror ( task , - ERESTARTSYS ) ;
2005-04-17 02:20:36 +04:00
}
2018-08-28 16:00:27 +03:00
static int
2005-11-10 05:45:24 +03:00
rpc_task_need_encode ( struct rpc_task * task )
{
2018-08-24 23:28:28 +03:00
return test_bit ( RPC_TASK_NEED_XMIT , & task - > tk_runstate ) = = 0 & &
( ! ( task - > tk_flags & RPC_TASK_SENT ) | |
! ( task - > tk_flags & RPC_TASK_NO_RETRANS_TIMEOUT ) | |
xprt_request_need_retransmit ( task ) ) ;
2005-11-10 05:45:24 +03:00
}
2005-04-17 02:20:36 +04:00
static void
2008-05-22 01:09:19 +04:00
rpc_xdr_encode ( struct rpc_task * task )
2005-04-17 02:20:36 +04:00
{
struct rpc_rqst * req = task - > tk_rqstp ;
2019-02-11 19:24:48 +03:00
struct xdr_stream xdr ;
2005-04-17 02:20:36 +04:00
2016-09-15 17:55:12 +03:00
xdr_buf_init ( & req - > rq_snd_buf ,
req - > rq_buffer ,
req - > rq_callsize ) ;
xdr_buf_init ( & req - > rq_rcv_buf ,
2016-09-15 17:55:37 +03:00
req - > rq_rbuffer ,
2016-09-15 17:55:12 +03:00
req - > rq_rcvsize ) ;
2005-04-17 02:20:36 +04:00
2019-09-10 20:01:35 +03:00
req - > rq_reply_bytes_recvd = 0 ;
2019-02-11 19:24:48 +03:00
req - > rq_snd_buf . head [ 0 ] . iov_len = 0 ;
xdr_init_encode ( & xdr , & req - > rq_snd_buf ,
req - > rq_snd_buf . head [ 0 ] . iov_base , req ) ;
2019-07-18 04:22:38 +03:00
xdr_free_bvec ( & req - > rq_snd_buf ) ;
2019-02-11 19:24:48 +03:00
if ( rpc_encode_header ( task , & xdr ) )
2005-04-17 02:20:36 +04:00
return ;
2008-05-22 01:09:19 +04:00
2019-02-11 19:24:48 +03:00
task - > tk_status = rpcauth_wrap_req ( task , & xdr ) ;
2005-04-17 02:20:36 +04:00
}
2018-08-24 23:28:28 +03:00
/*
* 3. Encode arguments of an RPC call
*/
static void
call_encode ( struct rpc_task * task )
{
if ( ! rpc_task_need_encode ( task ) )
goto out ;
2020-07-08 23:09:37 +03:00
2019-09-10 20:01:35 +03:00
/* Dequeue task from the receive queue while we're encoding */
xprt_request_dequeue_xprt ( task ) ;
2018-08-24 23:28:28 +03:00
/* Encode here so that rpcsec_gss can use correct sequence number. */
rpc_xdr_encode ( task ) ;
/* Did the encode result in an error condition? */
if ( task - > tk_status ! = 0 ) {
/* Was the error nonfatal? */
2019-01-03 01:53:13 +03:00
switch ( task - > tk_status ) {
case - EAGAIN :
case - ENOMEM :
2018-08-24 23:28:28 +03:00
rpc_delay ( task , HZ > > 4 ) ;
2019-01-03 01:53:13 +03:00
break ;
case - EKEYEXPIRED :
2019-04-29 12:32:31 +03:00
if ( ! task - > tk_cred_retry ) {
rpc_exit ( task , task - > tk_status ) ;
} else {
task - > tk_action = call_refresh ;
task - > tk_cred_retry - - ;
2020-07-08 23:10:03 +03:00
trace_rpc_retry_refresh_status ( task ) ;
2019-04-29 12:32:31 +03:00
}
2019-01-03 01:53:13 +03:00
break ;
default :
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , task - > tk_status ) ;
2019-01-03 01:53:13 +03:00
}
2018-08-24 23:28:28 +03:00
return ;
}
/* Add task to reply queue before transmission to avoid races */
if ( rpc_reply_expected ( task ) )
xprt_request_enqueue_receive ( task ) ;
xprt_request_enqueue_transmit ( task ) ;
out :
2019-04-11 22:16:52 +03:00
task - > tk_action = call_transmit ;
/* Check that the connection is OK */
if ( ! xprt_bound ( task - > tk_xprt ) )
task - > tk_action = call_bind ;
else if ( ! xprt_connected ( task - > tk_xprt ) )
task - > tk_action = call_connect ;
2018-08-24 23:28:28 +03:00
}
2019-03-10 18:17:13 +03:00
/*
* Helpers to check if the task was already transmitted , and
* to take action when that is the case .
*/
static bool
rpc_task_transmitted ( struct rpc_task * task )
{
return ! test_bit ( RPC_TASK_NEED_XMIT , & task - > tk_runstate ) ;
}
static void
rpc_task_handle_transmitted ( struct rpc_task * task )
{
xprt_end_transmit ( task ) ;
task - > tk_action = call_transmit_status ;
}
2005-04-17 02:20:36 +04:00
/*
* 4. Get the server port number if not yet set
*/
static void
call_bind ( struct rpc_task * task )
{
2013-01-08 19:08:33 +04:00
struct rpc_xprt * xprt = task - > tk_rqstp - > rq_xprt ;
2005-04-17 02:20:36 +04:00
2019-03-10 18:17:13 +03:00
if ( rpc_task_transmitted ( task ) ) {
rpc_task_handle_transmitted ( task ) ;
return ;
}
2019-03-09 20:07:17 +03:00
if ( xprt_bound ( xprt ) ) {
task - > tk_action = call_connect ;
return ;
}
task - > tk_action = call_bind_status ;
2019-03-10 19:22:39 +03:00
if ( ! xprt_prepare_transmit ( task ) )
return ;
2019-03-09 20:07:17 +03:00
xprt - > ops - > rpcbind ( task ) ;
2005-04-17 02:20:36 +04:00
}
/*
2005-08-12 00:25:11 +04:00
* 4 a . Sort out bind result
*/
static void
call_bind_status ( struct rpc_task * task )
{
2019-08-16 00:26:17 +03:00
struct rpc_xprt * xprt = task - > tk_rqstp - > rq_xprt ;
2007-09-12 02:00:47 +04:00
int status = - EIO ;
2005-08-12 00:25:11 +04:00
2019-03-10 18:17:13 +03:00
if ( rpc_task_transmitted ( task ) ) {
rpc_task_handle_transmitted ( task ) ;
return ;
}
2019-08-16 00:26:17 +03:00
if ( task - > tk_status > = 0 )
goto out_next ;
if ( xprt_bound ( xprt ) ) {
2005-08-12 00:25:11 +04:00
task - > tk_status = 0 ;
2019-08-16 00:26:17 +03:00
goto out_next ;
2005-08-12 00:25:11 +04:00
}
switch ( task - > tk_status ) {
2008-07-07 20:18:53 +04:00
case - ENOMEM :
rpc_delay ( task , HZ > > 2 ) ;
2007-09-12 02:00:41 +04:00
goto retry_timeout ;
2005-08-12 00:25:11 +04:00
case - EACCES :
2020-07-08 23:10:14 +03:00
trace_rpcb_prog_unavail_err ( task ) ;
2007-09-12 02:00:52 +04:00
/* fail immediately if this is an RPC ping */
if ( task - > tk_msg . rpc_proc - > p_proc = = 0 ) {
status = - EOPNOTSUPP ;
break ;
}
2011-05-31 23:15:34 +04:00
if ( task - > tk_rebind_retry = = 0 )
break ;
task - > tk_rebind_retry - - ;
2005-10-07 07:12:58 +04:00
rpc_delay ( task , 3 * HZ ) ;
2006-08-31 23:44:52 +04:00
goto retry_timeout ;
2019-08-16 05:55:19 +03:00
case - ENOBUFS :
rpc_delay ( task , HZ > > 2 ) ;
goto retry_timeout ;
2019-03-10 19:22:39 +03:00
case - EAGAIN :
goto retry_timeout ;
2005-08-12 00:25:11 +04:00
case - ETIMEDOUT :
2020-07-08 23:10:14 +03:00
trace_rpcb_timeout_err ( task ) ;
2006-08-31 23:44:52 +04:00
goto retry_timeout ;
2005-08-12 00:25:11 +04:00
case - EPFNOSUPPORT :
2007-09-12 02:00:47 +04:00
/* server doesn't support any rpcbind version we know of */
2020-07-08 23:10:14 +03:00
trace_rpcb_bind_version_err ( task ) ;
2005-08-12 00:25:11 +04:00
break ;
case - EPROTONOSUPPORT :
2020-07-08 23:10:14 +03:00
trace_rpcb_bind_version_err ( task ) ;
2014-03-17 20:57:31 +04:00
goto retry_timeout ;
2009-12-03 23:58:56 +03:00
case - ECONNREFUSED : /* connection problems */
case - ECONNRESET :
2013-12-31 22:11:43 +04:00
case - ECONNABORTED :
2009-12-03 23:58:56 +03:00
case - ENOTCONN :
case - EHOSTDOWN :
2017-11-30 15:21:33 +03:00
case - ENETDOWN :
2009-12-03 23:58:56 +03:00
case - EHOSTUNREACH :
case - ENETUNREACH :
case - EPIPE :
2020-07-08 23:10:14 +03:00
trace_rpcb_unreachable_err ( task ) ;
2009-12-03 23:58:56 +03:00
if ( ! RPC_IS_SOFTCONN ( task ) ) {
rpc_delay ( task , 5 * HZ ) ;
goto retry_timeout ;
}
status = task - > tk_status ;
break ;
2005-08-12 00:25:11 +04:00
default :
2020-07-08 23:10:14 +03:00
trace_rpcb_unrecognized_err ( task ) ;
2005-08-12 00:25:11 +04:00
}
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , status ) ;
2005-08-12 00:25:11 +04:00
return ;
2019-08-16 00:26:17 +03:00
out_next :
task - > tk_action = call_connect ;
return ;
2006-08-31 23:44:52 +04:00
retry_timeout :
2014-03-17 20:57:31 +04:00
task - > tk_status = 0 ;
2019-03-10 19:22:39 +03:00
task - > tk_action = call_bind ;
2019-03-10 00:06:47 +03:00
rpc_check_timeout ( task ) ;
2005-08-12 00:25:11 +04:00
}
/*
* 4 b . Connect to the RPC server
2005-04-17 02:20:36 +04:00
*/
static void
call_connect ( struct rpc_task * task )
{
2013-01-08 19:08:33 +04:00
struct rpc_xprt * xprt = task - > tk_rqstp - > rq_xprt ;
2005-04-17 02:20:36 +04:00
2019-03-10 18:17:13 +03:00
if ( rpc_task_transmitted ( task ) ) {
rpc_task_handle_transmitted ( task ) ;
return ;
}
2019-03-09 20:07:17 +03:00
if ( xprt_connected ( xprt ) ) {
task - > tk_action = call_transmit ;
return ;
}
task - > tk_action = call_connect_status ;
if ( task - > tk_status < 0 )
return ;
if ( task - > tk_flags & RPC_TASK_NOCONNECT ) {
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , - ENOTCONN ) ;
2019-03-09 20:07:17 +03:00
return ;
2005-04-17 02:20:36 +04:00
}
2019-03-10 19:22:39 +03:00
if ( ! xprt_prepare_transmit ( task ) )
return ;
2019-03-09 20:07:17 +03:00
xprt_connect ( task ) ;
2005-04-17 02:20:36 +04:00
}
/*
2005-08-12 00:25:11 +04:00
* 4 c . Sort out connect result
2005-04-17 02:20:36 +04:00
*/
static void
call_connect_status ( struct rpc_task * task )
{
2019-08-16 00:26:17 +03:00
struct rpc_xprt * xprt = task - > tk_rqstp - > rq_xprt ;
2005-04-17 02:20:36 +04:00
struct rpc_clnt * clnt = task - > tk_client ;
int status = task - > tk_status ;
2019-03-10 18:17:13 +03:00
if ( rpc_task_transmitted ( task ) ) {
rpc_task_handle_transmitted ( task ) ;
2018-11-30 20:48:47 +03:00
return ;
}
2018-03-16 17:33:44 +03:00
trace_rpc_connect_status ( task ) ;
2019-08-16 00:26:17 +03:00
if ( task - > tk_status = = 0 ) {
clnt - > cl_stats - > netreconn + + ;
goto out_next ;
}
if ( xprt_connected ( xprt ) ) {
task - > tk_status = 0 ;
goto out_next ;
}
2013-09-26 23:22:45 +04:00
task - > tk_status = 0 ;
2005-04-17 02:20:36 +04:00
switch ( status ) {
2013-03-05 02:29:33 +04:00
case - ECONNREFUSED :
2017-08-18 10:12:51 +03:00
/* A positive refusal suggests a rebind is needed. */
if ( RPC_IS_SOFTCONN ( task ) )
break ;
if ( clnt - > cl_autobind ) {
rpc_force_rebind ( clnt ) ;
2019-03-07 22:10:32 +03:00
goto out_retry ;
2017-08-18 10:12:51 +03:00
}
2020-08-24 01:36:59 +03:00
fallthrough ;
2013-03-05 02:29:33 +04:00
case - ECONNRESET :
2013-12-31 22:11:43 +04:00
case - ECONNABORTED :
2017-11-30 15:21:33 +03:00
case - ENETDOWN :
2013-03-05 02:29:33 +04:00
case - ENETUNREACH :
2013-12-31 22:11:43 +04:00
case - EHOSTUNREACH :
2014-07-03 08:02:57 +04:00
case - EPIPE :
2019-12-23 18:28:44 +03:00
case - EPROTO :
sunrpc: Don't engage exponential backoff when connection attempt is rejected.
xs_connect() contains an exponential backoff mechanism so the repeated
connection attempts are delayed by longer and longer amounts.
This is appropriate when the connection failed due to a timeout, but
it not appropriate when a definitive "no" answer is received. In such
cases, call_connect_status() imposes a minimum 3-second back-off, so
not having the exponetial back-off will never result in immediate
retries.
The current situation is a problem when the NFS server tries to
register with rpcbind but rpcbind isn't running. All connection
attempts are made on the same "xprt" and as the connection is never
"closed", the exponential back delays successive attempts to register,
or de-register, different protocols. This results in a multi-minute
delay with no benefit.
So, when call_connect_status() receives a definitive "no", use
xprt_conditional_disconnect() to cancel the previous connection attempt.
This will set XPRT_CLOSE_WAIT so that xprt->ops->close() calls xs_close()
which resets the reestablish_timeout.
To ensure xprt_conditional_disconnect() does the right thing, we
ensure that rq_connect_cookie is set before a connection attempt, and
allow xprt_conditional_disconnect() to complete even when the
transport is not fully connected.
Signed-off-by: NeilBrown <neilb@suse.com>
Signed-off-by: Trond Myklebust <trond.myklebust@primarydata.com>
2016-11-23 06:44:58 +03:00
xprt_conditional_disconnect ( task - > tk_rqstp - > rq_xprt ,
task - > tk_rqstp - > rq_connect_cookie ) ;
2013-03-05 02:29:33 +04:00
if ( RPC_IS_SOFTCONN ( task ) )
break ;
2014-03-20 19:23:03 +04:00
/* retry with existing socket, after a delay */
rpc_delay ( task , 3 * HZ ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2019-08-16 05:55:19 +03:00
case - EADDRINUSE :
2018-12-17 21:34:59 +03:00
case - ENOTCONN :
2013-03-05 02:29:33 +04:00
case - EAGAIN :
2014-03-17 20:51:44 +04:00
case - ETIMEDOUT :
2019-03-07 22:10:32 +03:00
goto out_retry ;
2019-08-16 05:55:19 +03:00
case - ENOBUFS :
rpc_delay ( task , HZ > > 2 ) ;
goto out_retry ;
2005-04-17 02:20:36 +04:00
}
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , status ) ;
2019-03-07 22:10:32 +03:00
return ;
2019-08-16 00:26:17 +03:00
out_next :
task - > tk_action = call_transmit ;
return ;
2019-03-07 22:10:32 +03:00
out_retry :
/* Check for timeouts before looping back to call_bind */
task - > tk_action = call_bind ;
rpc_check_timeout ( task ) ;
2005-04-17 02:20:36 +04:00
}
/*
* 5. Transmit the RPC request , and wait for reply
*/
static void
call_transmit ( struct rpc_task * task )
{
2019-03-10 18:17:13 +03:00
if ( rpc_task_transmitted ( task ) ) {
rpc_task_handle_transmitted ( task ) ;
return ;
}
2019-03-04 22:19:31 +03:00
task - > tk_action = call_transmit_status ;
2019-03-09 20:07:17 +03:00
if ( ! xprt_prepare_transmit ( task ) )
return ;
task - > tk_status = 0 ;
2018-09-04 06:39:27 +03:00
if ( test_bit ( RPC_TASK_NEED_XMIT , & task - > tk_runstate ) ) {
2019-03-09 20:07:17 +03:00
if ( ! xprt_connected ( task - > tk_xprt ) ) {
task - > tk_status = - ENOTCONN ;
2018-09-04 06:39:27 +03:00
return ;
2019-03-04 22:19:31 +03:00
}
2019-03-09 20:07:17 +03:00
xprt_transmit ( task ) ;
2018-09-04 06:39:27 +03:00
}
xprt_end_transmit ( task ) ;
2006-07-28 01:22:50 +04:00
}
/*
* 5 a . Handle cleanup after a transmission
*/
static void
call_transmit_status ( struct rpc_task * task )
{
task - > tk_action = call_status ;
2009-12-03 23:58:56 +03:00
/*
* Common case : success . Force the compiler to put this
* test first .
*/
2019-03-09 20:07:17 +03:00
if ( rpc_task_transmitted ( task ) ) {
2019-04-15 18:54:13 +03:00
task - > tk_status = 0 ;
xprt_request_wait_receive ( task ) ;
2009-12-03 23:58:56 +03:00
return ;
}
2009-03-11 21:37:57 +03:00
switch ( task - > tk_status ) {
default :
2018-09-04 00:37:36 +03:00
break ;
2018-08-28 16:20:10 +03:00
case - EBADMSG :
2018-08-24 23:28:28 +03:00
task - > tk_status = 0 ;
task - > tk_action = call_encode ;
2018-08-28 16:20:10 +03:00
break ;
2009-03-11 21:37:57 +03:00
/*
* Special cases : if we ' ve been waiting on the
* socket ' s write_space ( ) callback , or if the
* socket just returned a connection error ,
* then hold onto the transport lock .
*/
2018-08-28 16:20:10 +03:00
case - ENOBUFS :
rpc_delay ( task , HZ > > 2 ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2018-09-04 06:39:27 +03:00
case - EBADSLT :
2018-08-28 16:20:10 +03:00
case - EAGAIN :
task - > tk_action = call_transmit ;
task - > tk_status = 0 ;
break ;
2009-03-11 21:37:57 +03:00
case - ECONNREFUSED :
case - EHOSTDOWN :
2017-11-30 15:21:33 +03:00
case - ENETDOWN :
2009-03-11 21:37:57 +03:00
case - EHOSTUNREACH :
case - ENETUNREACH :
rpc: Add -EPERM processing for xs_udp_send_request()
If an iptables drop rule is added for an nfs server, the client can end up in
a softlockup. Because of the way that xs_sendpages() is structured, the -EPERM
is ignored since the prior bits of the packet may have been successfully queued
and thus xs_sendpages() returns a non-zero value. Then, xs_udp_send_request()
thinks that because some bits were queued it should return -EAGAIN. We then try
the request again and again, resulting in cpu spinning. Reproducer:
1) open a file on the nfs server '/nfs/foo' (mounted using udp)
2) iptables -A OUTPUT -d <nfs server ip> -j DROP
3) write to /nfs/foo
4) close /nfs/foo
5) iptables -D OUTPUT -d <nfs server ip> -j DROP
The softlockup occurs in step 4 above.
The previous patch, allows xs_sendpages() to return both a sent count and
any error values that may have occurred. Thus, if we get an -EPERM, return
that to the higher level code.
With this patch in place we can successfully abort the above sequence and
avoid the softlockup.
I also tried the above test case on an nfs mount on tcp and although the system
does not softlockup, I still ended up with the 'hung_task' firing after 120
seconds, due to the i/o being stuck. The tcp case appears a bit harder to fix,
since -EPERM appears to get ignored much lower down in the stack and does not
propogate up to xs_sendpages(). This case is not quite as insidious as the
softlockup and it is not addressed here.
Reported-by: Yigong Lou <ylou@akamai.com>
Signed-off-by: Jason Baron <jbaron@akamai.com>
Signed-off-by: Trond Myklebust <trond.myklebust@primarydata.com>
2014-09-24 22:08:04 +04:00
case - EPERM :
SUNRPC: Allow RPCs to fail quickly if the server is unreachable
The kernel sometimes makes RPC calls to services that aren't running.
Because the kernel's RPC client always assumes the hard retry semantic
when reconnecting a connection-oriented RPC transport, the underlying
reconnect logic takes a long while to time out, even though the remote
may have responded immediately with ECONNREFUSED.
In certain cases, like upcalls to our local rpcbind daemon, or for NFS
mount requests, we'd like the kernel to fail immediately if the remote
service isn't reachable. This allows another transport to be tried
immediately, or the pending request can be abandoned quickly.
Introduce a per-request flag which controls how call_transmit_status()
behaves when request transmission fails because the server cannot be
reached.
We don't want soft connection semantics to apply to other errors. The
default case of the switch statement in call_transmit_status() no
longer falls through; the fall through code is copied to the default
case, and a "break;" is added.
The transport's connection re-establishment timeout is also ignored for
such requests. We want the request to fail immediately, so the
reconnect delay is skipped. Additionally, we don't want a connect
failure here to further increase the reconnect timeout value, since
this request will not be retried.
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Trond Myklebust <Trond.Myklebust@netapp.com>
2009-12-03 23:58:56 +03:00
if ( RPC_IS_SOFTCONN ( task ) ) {
2018-03-16 17:33:55 +03:00
if ( ! task - > tk_msg . rpc_proc - > p_proc )
trace_xprt_ping ( task - > tk_xprt ,
task - > tk_status ) ;
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , task - > tk_status ) ;
2019-03-07 22:10:32 +03:00
return ;
SUNRPC: Allow RPCs to fail quickly if the server is unreachable
The kernel sometimes makes RPC calls to services that aren't running.
Because the kernel's RPC client always assumes the hard retry semantic
when reconnecting a connection-oriented RPC transport, the underlying
reconnect logic takes a long while to time out, even though the remote
may have responded immediately with ECONNREFUSED.
In certain cases, like upcalls to our local rpcbind daemon, or for NFS
mount requests, we'd like the kernel to fail immediately if the remote
service isn't reachable. This allows another transport to be tried
immediately, or the pending request can be abandoned quickly.
Introduce a per-request flag which controls how call_transmit_status()
behaves when request transmission fails because the server cannot be
reached.
We don't want soft connection semantics to apply to other errors. The
default case of the switch statement in call_transmit_status() no
longer falls through; the fall through code is copied to the default
case, and a "break;" is added.
The transport's connection re-establishment timeout is also ignored for
such requests. We want the request to fail immediately, so the
reconnect delay is skipped. Additionally, we don't want a connect
failure here to further increase the reconnect timeout value, since
this request will not be retried.
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Trond Myklebust <Trond.Myklebust@netapp.com>
2009-12-03 23:58:56 +03:00
}
2020-08-24 01:36:59 +03:00
fallthrough ;
SUNRPC: Allow RPCs to fail quickly if the server is unreachable
The kernel sometimes makes RPC calls to services that aren't running.
Because the kernel's RPC client always assumes the hard retry semantic
when reconnecting a connection-oriented RPC transport, the underlying
reconnect logic takes a long while to time out, even though the remote
may have responded immediately with ECONNREFUSED.
In certain cases, like upcalls to our local rpcbind daemon, or for NFS
mount requests, we'd like the kernel to fail immediately if the remote
service isn't reachable. This allows another transport to be tried
immediately, or the pending request can be abandoned quickly.
Introduce a per-request flag which controls how call_transmit_status()
behaves when request transmission fails because the server cannot be
reached.
We don't want soft connection semantics to apply to other errors. The
default case of the switch statement in call_transmit_status() no
longer falls through; the fall through code is copied to the default
case, and a "break;" is added.
The transport's connection re-establishment timeout is also ignored for
such requests. We want the request to fail immediately, so the
reconnect delay is skipped. Additionally, we don't want a connect
failure here to further increase the reconnect timeout value, since
this request will not be retried.
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Trond Myklebust <Trond.Myklebust@netapp.com>
2009-12-03 23:58:56 +03:00
case - ECONNRESET :
2013-12-31 22:11:43 +04:00
case - ECONNABORTED :
2015-02-09 05:44:04 +03:00
case - EADDRINUSE :
SUNRPC: Allow RPCs to fail quickly if the server is unreachable
The kernel sometimes makes RPC calls to services that aren't running.
Because the kernel's RPC client always assumes the hard retry semantic
when reconnecting a connection-oriented RPC transport, the underlying
reconnect logic takes a long while to time out, even though the remote
may have responded immediately with ECONNREFUSED.
In certain cases, like upcalls to our local rpcbind daemon, or for NFS
mount requests, we'd like the kernel to fail immediately if the remote
service isn't reachable. This allows another transport to be tried
immediately, or the pending request can be abandoned quickly.
Introduce a per-request flag which controls how call_transmit_status()
behaves when request transmission fails because the server cannot be
reached.
We don't want soft connection semantics to apply to other errors. The
default case of the switch statement in call_transmit_status() no
longer falls through; the fall through code is copied to the default
case, and a "break;" is added.
The transport's connection re-establishment timeout is also ignored for
such requests. We want the request to fail immediately, so the
reconnect delay is skipped. Additionally, we don't want a connect
failure here to further increase the reconnect timeout value, since
this request will not be retried.
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Trond Myklebust <Trond.Myklebust@netapp.com>
2009-12-03 23:58:56 +03:00
case - ENOTCONN :
2009-03-11 21:37:59 +03:00
case - EPIPE :
2019-03-04 22:19:31 +03:00
task - > tk_action = call_bind ;
task - > tk_status = 0 ;
2018-08-28 16:00:27 +03:00
break ;
2009-03-11 21:37:57 +03:00
}
2019-03-07 22:10:32 +03:00
rpc_check_timeout ( task ) ;
2005-04-17 02:20:36 +04:00
}
2011-07-14 03:20:49 +04:00
# if defined(CONFIG_SUNRPC_BACKCHANNEL)
2019-03-05 15:30:48 +03:00
static void call_bc_transmit ( struct rpc_task * task ) ;
static void call_bc_transmit_status ( struct rpc_task * task ) ;
static void
call_bc_encode ( struct rpc_task * task )
{
xprt_request_enqueue_transmit ( task ) ;
task - > tk_action = call_bc_transmit ;
}
2009-04-01 17:23:03 +04:00
/*
* 5 b . Send the backchannel RPC reply . On error , drop the reply . In
* addition , disconnect on connectivity errors .
*/
static void
call_bc_transmit ( struct rpc_task * task )
{
2019-03-05 15:30:48 +03:00
task - > tk_action = call_bc_transmit_status ;
if ( test_bit ( RPC_TASK_NEED_XMIT , & task - > tk_runstate ) ) {
if ( ! xprt_prepare_transmit ( task ) )
return ;
task - > tk_status = 0 ;
xprt_transmit ( task ) ;
2009-04-01 17:23:03 +04:00
}
2019-03-05 15:30:48 +03:00
xprt_end_transmit ( task ) ;
}
2009-04-01 17:23:03 +04:00
2019-03-05 15:30:48 +03:00
static void
call_bc_transmit_status ( struct rpc_task * task )
{
struct rpc_rqst * req = task - > tk_rqstp ;
2015-06-02 18:53:21 +03:00
2019-04-15 18:54:13 +03:00
if ( rpc_task_transmitted ( task ) )
task - > tk_status = 0 ;
2009-04-01 17:23:03 +04:00
switch ( task - > tk_status ) {
case 0 :
/* Success */
2017-11-30 15:21:33 +03:00
case - ENETDOWN :
2009-04-01 17:23:03 +04:00
case - EHOSTDOWN :
case - EHOSTUNREACH :
case - ENETUNREACH :
2015-06-19 20:04:13 +03:00
case - ECONNRESET :
case - ECONNREFUSED :
case - EADDRINUSE :
case - ENOTCONN :
case - EPIPE :
break ;
2019-03-05 15:30:48 +03:00
case - ENOBUFS :
rpc_delay ( task , HZ > > 2 ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2019-03-05 15:30:48 +03:00
case - EBADSLT :
2018-09-04 06:39:27 +03:00
case - EAGAIN :
2019-03-05 15:30:48 +03:00
task - > tk_status = 0 ;
task - > tk_action = call_bc_transmit ;
return ;
2009-04-01 17:23:03 +04:00
case - ETIMEDOUT :
/*
* Problem reaching the server . Disconnect and let the
* forechannel reestablish the connection . The server will
* have to retransmit the backchannel request and we ' ll
* reprocess it . Since these ops are idempotent , there ' s no
* need to cache our reply at this time .
*/
printk ( KERN_NOTICE " RPC: Could not send backchannel reply "
" error: %d \n " , task - > tk_status ) ;
2013-01-08 18:10:21 +04:00
xprt_conditional_disconnect ( req - > rq_xprt ,
2009-04-01 17:23:03 +04:00
req - > rq_connect_cookie ) ;
break ;
default :
/*
* We were unable to reply and will have to drop the
* request . The server should reconnect and retransmit .
*/
printk ( KERN_NOTICE " RPC: Could not send backchannel reply "
" error: %d \n " , task - > tk_status ) ;
break ;
}
2015-06-02 18:53:21 +03:00
task - > tk_action = rpc_exit_task ;
2009-04-01 17:23:03 +04:00
}
2011-07-14 03:20:49 +04:00
# endif /* CONFIG_SUNRPC_BACKCHANNEL */
2009-04-01 17:23:03 +04:00
2005-04-17 02:20:36 +04:00
/*
* 6. Sort out the RPC call status
*/
static void
call_status ( struct rpc_task * task )
{
struct rpc_clnt * clnt = task - > tk_client ;
int status ;
2018-03-16 17:33:55 +03:00
if ( ! task - > tk_msg . rpc_proc - > p_proc )
trace_xprt_ping ( task - > tk_xprt , task - > tk_status ) ;
2005-04-17 02:20:36 +04:00
status = task - > tk_status ;
if ( status > = 0 ) {
task - > tk_action = call_decode ;
return ;
}
2012-02-06 19:08:08 +04:00
trace_rpc_call_status ( task ) ;
2005-04-17 02:20:36 +04:00
task - > tk_status = 0 ;
switch ( status ) {
2006-08-30 22:32:49 +04:00
case - EHOSTDOWN :
2017-11-30 15:21:33 +03:00
case - ENETDOWN :
2006-08-30 22:32:49 +04:00
case - EHOSTUNREACH :
case - ENETUNREACH :
rpc: Add -EPERM processing for xs_udp_send_request()
If an iptables drop rule is added for an nfs server, the client can end up in
a softlockup. Because of the way that xs_sendpages() is structured, the -EPERM
is ignored since the prior bits of the packet may have been successfully queued
and thus xs_sendpages() returns a non-zero value. Then, xs_udp_send_request()
thinks that because some bits were queued it should return -EAGAIN. We then try
the request again and again, resulting in cpu spinning. Reproducer:
1) open a file on the nfs server '/nfs/foo' (mounted using udp)
2) iptables -A OUTPUT -d <nfs server ip> -j DROP
3) write to /nfs/foo
4) close /nfs/foo
5) iptables -D OUTPUT -d <nfs server ip> -j DROP
The softlockup occurs in step 4 above.
The previous patch, allows xs_sendpages() to return both a sent count and
any error values that may have occurred. Thus, if we get an -EPERM, return
that to the higher level code.
With this patch in place we can successfully abort the above sequence and
avoid the softlockup.
I also tried the above test case on an nfs mount on tcp and although the system
does not softlockup, I still ended up with the 'hung_task' firing after 120
seconds, due to the i/o being stuck. The tcp case appears a bit harder to fix,
since -EPERM appears to get ignored much lower down in the stack and does not
propogate up to xs_sendpages(). This case is not quite as insidious as the
softlockup and it is not addressed here.
Reported-by: Yigong Lou <ylou@akamai.com>
Signed-off-by: Jason Baron <jbaron@akamai.com>
Signed-off-by: Trond Myklebust <trond.myklebust@primarydata.com>
2014-09-24 22:08:04 +04:00
case - EPERM :
2019-03-10 00:06:47 +03:00
if ( RPC_IS_SOFTCONN ( task ) )
goto out_exit ;
2006-08-30 22:32:49 +04:00
/*
* Delay any retries for 3 seconds , then handle as if it
* were a timeout .
*/
rpc_delay ( task , 3 * HZ ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2005-04-17 02:20:36 +04:00
case - ETIMEDOUT :
break ;
case - ECONNREFUSED :
2013-12-31 22:11:43 +04:00
case - ECONNRESET :
case - ECONNABORTED :
2019-05-29 17:46:00 +03:00
case - ENOTCONN :
2006-01-03 11:55:50 +03:00
rpc_force_rebind ( clnt ) ;
2019-08-16 15:58:48 +03:00
break ;
2015-02-09 05:44:04 +03:00
case - EADDRINUSE :
2009-03-11 21:37:59 +03:00
rpc_delay ( task , 3 * HZ ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2009-03-11 21:37:59 +03:00
case - EPIPE :
2005-04-17 02:20:36 +04:00
case - EAGAIN :
break ;
case - EIO :
/* shutdown or soft timeout */
2019-03-10 00:06:47 +03:00
goto out_exit ;
2005-04-17 02:20:36 +04:00
default :
2008-06-10 00:51:31 +04:00
if ( clnt - > cl_chatty )
printk ( " %s: RPC call returned error %d \n " ,
2013-08-23 19:48:15 +04:00
clnt - > cl_program - > name , - status ) ;
2019-03-10 00:06:47 +03:00
goto out_exit ;
2005-04-17 02:20:36 +04:00
}
2019-03-10 00:06:47 +03:00
task - > tk_action = call_encode ;
2020-06-23 18:24:09 +03:00
if ( status ! = - ECONNRESET & & status ! = - ECONNABORTED )
rpc_check_timeout ( task ) ;
2019-03-10 00:06:47 +03:00
return ;
out_exit :
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , status ) ;
2005-04-17 02:20:36 +04:00
}
2019-03-19 18:24:54 +03:00
static bool
rpc_check_connected ( const struct rpc_rqst * req )
{
/* No allocated request or transport? return true */
if ( ! req | | ! req - > rq_xprt )
return true ;
return xprt_connected ( req - > rq_xprt ) ;
}
2005-04-17 02:20:36 +04:00
static void
2019-03-07 22:10:32 +03:00
rpc_check_timeout ( struct rpc_task * task )
2005-04-17 02:20:36 +04:00
{
struct rpc_clnt * clnt = task - > tk_client ;
2020-05-09 21:07:13 +03:00
if ( RPC_SIGNALLED ( task ) ) {
rpc_call_rpcerror ( task , - ERESTARTSYS ) ;
return ;
}
2019-03-07 22:10:32 +03:00
if ( xprt_adjust_timeout ( task - > tk_rqstp ) = = 0 )
return ;
2005-04-17 02:20:36 +04:00
2020-07-08 23:09:58 +03:00
trace_rpc_timeout_status ( task ) ;
2006-03-20 21:44:17 +03:00
task - > tk_timeouts + + ;
2019-03-19 18:24:54 +03:00
if ( RPC_IS_SOFTCONN ( task ) & & ! rpc_check_connected ( task - > tk_rqstp ) ) {
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , - ETIMEDOUT ) ;
2009-12-03 23:58:56 +03:00
return ;
}
2019-03-19 18:24:54 +03:00
2005-04-17 02:20:36 +04:00
if ( RPC_IS_SOFT ( task ) ) {
2019-04-07 20:58:55 +03:00
/*
* Once a " no retrans timeout " soft tasks ( a . k . a NFSv4 ) has
* been sent , it should time out only if the transport
* connection gets terminally broken .
*/
if ( ( task - > tk_flags & RPC_TASK_NO_RETRANS_TIMEOUT ) & &
rpc_check_connected ( task - > tk_rqstp ) )
return ;
2012-07-18 22:17:11 +04:00
if ( clnt - > cl_chatty ) {
2019-04-07 20:58:57 +03:00
pr_notice_ratelimited (
" %s: server %s not responding, timed out \n " ,
2013-08-23 19:48:15 +04:00
clnt - > cl_program - > name ,
2016-01-31 00:39:26 +03:00
task - > tk_xprt - > servername ) ;
2012-07-18 22:17:11 +04:00
}
2011-04-24 22:28:45 +04:00
if ( task - > tk_flags & RPC_TASK_TIMEOUT )
2019-04-07 20:58:54 +03:00
rpc_call_rpcerror ( task , - ETIMEDOUT ) ;
2011-04-24 22:28:45 +04:00
else
2019-04-07 20:58:54 +03:00
__rpc_call_rpcerror ( task , - EIO , - ETIMEDOUT ) ;
2005-04-17 02:20:36 +04:00
return ;
}
2006-01-03 11:55:52 +03:00
if ( ! ( task - > tk_flags & RPC_CALL_MAJORSEEN ) ) {
2005-04-17 02:20:36 +04:00
task - > tk_flags | = RPC_CALL_MAJORSEEN ;
2012-03-02 02:01:05 +04:00
if ( clnt - > cl_chatty ) {
2019-04-07 20:58:57 +03:00
pr_notice_ratelimited (
" %s: server %s not responding, still trying \n " ,
clnt - > cl_program - > name ,
task - > tk_xprt - > servername ) ;
2012-03-02 02:01:05 +04:00
}
2005-04-17 02:20:36 +04:00
}
2006-01-03 11:55:50 +03:00
rpc_force_rebind ( clnt ) ;
2008-04-23 00:47:55 +04:00
/*
* Did our request time out due to an RPCSEC_GSS out - of - sequence
* event ? RFC2203 requires the server to drop all such requests .
*/
rpcauth_invalcred ( task ) ;
2019-03-07 22:10:32 +03:00
}
2005-04-17 02:20:36 +04:00
/*
* 7. Decode the RPC reply
*/
static void
call_decode ( struct rpc_task * task )
{
struct rpc_clnt * clnt = task - > tk_client ;
struct rpc_rqst * req = task - > tk_rqstp ;
2019-02-11 19:24:58 +03:00
struct xdr_stream xdr ;
2019-09-16 16:12:19 +03:00
int err ;
2005-04-17 02:20:36 +04:00
2019-02-11 19:24:58 +03:00
if ( ! task - > tk_msg . rpc_proc - > p_decode ) {
2018-08-28 23:27:31 +03:00
task - > tk_action = rpc_exit_task ;
return ;
}
2006-01-03 11:55:52 +03:00
if ( task - > tk_flags & RPC_CALL_MAJORSEEN ) {
2012-03-02 02:01:05 +04:00
if ( clnt - > cl_chatty ) {
2019-04-07 20:58:57 +03:00
pr_notice_ratelimited ( " %s: server %s OK \n " ,
2013-08-23 19:48:15 +04:00
clnt - > cl_program - > name ,
2016-01-31 00:39:26 +03:00
task - > tk_xprt - > servername ) ;
2012-03-02 02:01:05 +04:00
}
2005-04-17 02:20:36 +04:00
task - > tk_flags & = ~ RPC_CALL_MAJORSEEN ;
}
2006-03-20 21:44:51 +03:00
/*
* Ensure that we see all writes made by xprt_complete_rqst ( )
2009-04-01 17:23:28 +04:00
* before it changed req - > rq_reply_bytes_recvd .
2006-03-20 21:44:51 +03:00
*/
smp_rmb ( ) ;
2019-09-16 16:12:19 +03:00
/*
* Did we ever call xprt_complete_rqst ( ) ? If not , we should assume
* the message is incomplete .
*/
err = - EAGAIN ;
if ( ! req - > rq_reply_bytes_recvd )
goto out ;
2005-04-17 02:20:36 +04:00
req - > rq_rcv_buf . len = req - > rq_private_buf . len ;
2020-05-13 00:13:28 +03:00
trace_rpc_xdr_recvfrom ( task , & req - > rq_rcv_buf ) ;
2005-04-17 02:20:36 +04:00
/* Check that the softirq receive buffer is valid */
WARN_ON ( memcmp ( & req - > rq_rcv_buf , & req - > rq_private_buf ,
sizeof ( req - > rq_rcv_buf ) ) ! = 0 ) ;
2019-02-11 19:24:58 +03:00
xdr_init_decode ( & xdr , & req - > rq_rcv_buf ,
req - > rq_rcv_buf . head [ 0 ] . iov_base , req ) ;
2019-09-16 16:12:19 +03:00
err = rpc_decode_header ( task , & xdr ) ;
out :
switch ( err ) {
2019-02-11 19:24:58 +03:00
case 0 :
task - > tk_action = rpc_exit_task ;
task - > tk_status = rpcauth_unwrap_resp ( task , & xdr ) ;
2006-01-03 11:55:03 +03:00
return ;
2019-02-11 19:24:58 +03:00
case - EAGAIN :
task - > tk_status = 0 ;
2019-05-29 19:49:52 +03:00
if ( task - > tk_client - > cl_discrtry )
xprt_conditional_disconnect ( req - > rq_xprt ,
req - > rq_connect_cookie ) ;
2019-03-10 00:06:47 +03:00
task - > tk_action = call_encode ;
rpc_check_timeout ( task ) ;
2019-05-29 19:49:52 +03:00
break ;
case - EKEYREJECTED :
task - > tk_action = call_reserve ;
rpc_check_timeout ( task ) ;
rpcauth_invalcred ( task ) ;
/* Ensure we obtain a new XID if we retry! */
xprt_release ( task ) ;
2008-04-19 21:15:47 +04:00
}
2005-04-17 02:20:36 +04:00
}
2019-02-11 19:24:48 +03:00
static int
rpc_encode_header ( struct rpc_task * task , struct xdr_stream * xdr )
2005-04-17 02:20:36 +04:00
{
struct rpc_clnt * clnt = task - > tk_client ;
struct rpc_rqst * req = task - > tk_rqstp ;
2019-02-11 19:24:48 +03:00
__be32 * p ;
int error ;
error = - EMSGSIZE ;
p = xdr_reserve_space ( xdr , RPC_CALLHDRSIZE < < 2 ) ;
if ( ! p )
goto out_fail ;
* p + + = req - > rq_xid ;
* p + + = rpc_call ;
* p + + = cpu_to_be32 ( RPC_VERSION ) ;
* p + + = cpu_to_be32 ( clnt - > cl_prog ) ;
* p + + = cpu_to_be32 ( clnt - > cl_vers ) ;
* p = cpu_to_be32 ( task - > tk_msg . rpc_proc - > p_proc ) ;
error = rpcauth_marshcred ( task , xdr ) ;
if ( error < 0 )
goto out_fail ;
return 0 ;
out_fail :
trace_rpc_bad_callhdr ( task ) ;
2019-09-12 15:06:51 +03:00
rpc_call_rpcerror ( task , error ) ;
2019-02-11 19:24:48 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
2019-02-11 19:24:58 +03:00
static noinline int
rpc_decode_header ( struct rpc_task * task , struct xdr_stream * xdr )
2005-04-17 02:20:36 +04:00
{
2012-03-02 02:01:05 +04:00
struct rpc_clnt * clnt = task - > tk_client ;
2019-03-15 16:29:00 +03:00
int error ;
2019-02-11 19:24:58 +03:00
__be32 * p ;
2005-04-17 02:20:36 +04:00
2019-02-11 19:24:53 +03:00
/* RFC-1014 says that the representation of XDR data must be a
* multiple of four bytes
* - if it isn ' t pointer subtraction in the NFS client may give
* undefined results
*/
if ( task - > tk_rqstp - > rq_rcv_buf . len & 3 )
2019-03-15 16:29:00 +03:00
goto out_unparsable ;
2005-04-17 02:20:36 +04:00
2019-02-11 19:24:58 +03:00
p = xdr_inline_decode ( xdr , 3 * sizeof ( * p ) ) ;
if ( ! p )
goto out_unparsable ;
2019-02-11 19:24:53 +03:00
p + + ; /* skip XID */
if ( * p + + ! = rpc_reply )
goto out_unparsable ;
if ( * p + + ! = rpc_msg_accepted )
goto out_msg_denied ;
2009-04-01 17:22:54 +04:00
2019-02-11 19:24:58 +03:00
error = rpcauth_checkverf ( task , xdr ) ;
if ( error )
2019-02-11 19:24:53 +03:00
goto out_verifier ;
2019-02-11 19:24:58 +03:00
p = xdr_inline_decode ( xdr , sizeof ( * p ) ) ;
if ( ! p )
2019-02-11 19:24:53 +03:00
goto out_unparsable ;
2019-02-11 19:24:58 +03:00
switch ( * p ) {
2019-02-11 19:24:53 +03:00
case rpc_success :
2019-02-11 19:24:58 +03:00
return 0 ;
2019-02-11 19:24:53 +03:00
case rpc_prog_unavail :
trace_rpc__prog_unavail ( task ) ;
2005-06-22 21:16:23 +04:00
error = - EPFNOSUPPORT ;
goto out_err ;
2019-02-11 19:24:53 +03:00
case rpc_prog_mismatch :
trace_rpc__prog_mismatch ( task ) ;
2005-06-22 21:16:23 +04:00
error = - EPROTONOSUPPORT ;
goto out_err ;
2019-02-11 19:24:53 +03:00
case rpc_proc_unavail :
trace_rpc__proc_unavail ( task ) ;
2005-06-22 21:16:23 +04:00
error = - EOPNOTSUPP ;
goto out_err ;
2019-02-11 19:24:53 +03:00
case rpc_garbage_args :
2019-03-15 17:12:30 +03:00
case rpc_system_err :
2019-02-11 19:24:53 +03:00
trace_rpc__garbage_args ( task ) ;
2019-03-15 16:29:00 +03:00
error = - EIO ;
2019-02-11 19:24:53 +03:00
break ;
2005-04-17 02:20:36 +04:00
default :
2019-03-15 16:29:00 +03:00
goto out_unparsable ;
2005-04-17 02:20:36 +04:00
}
2006-01-03 11:55:03 +03:00
out_garbage :
2012-03-02 02:01:05 +04:00
clnt - > cl_stats - > rpcgarbage + + ;
2005-04-17 02:20:36 +04:00
if ( task - > tk_garb_retry ) {
task - > tk_garb_retry - - ;
2018-08-24 23:28:28 +03:00
task - > tk_action = call_encode ;
2019-02-11 19:24:58 +03:00
return - EAGAIN ;
2005-04-17 02:20:36 +04:00
}
out_err :
2019-09-12 15:06:51 +03:00
rpc_call_rpcerror ( task , error ) ;
2019-02-11 19:24:58 +03:00
return error ;
2019-02-11 19:24:53 +03:00
out_unparsable :
trace_rpc__unparsable ( task ) ;
error = - EIO ;
2006-01-03 11:55:03 +03:00
goto out_garbage ;
2019-02-11 19:24:53 +03:00
out_verifier :
trace_rpc_bad_verifier ( task ) ;
goto out_garbage ;
out_msg_denied :
2019-03-15 16:29:00 +03:00
error = - EACCES ;
2019-02-11 19:24:58 +03:00
p = xdr_inline_decode ( xdr , sizeof ( * p ) ) ;
if ( ! p )
goto out_unparsable ;
2019-02-11 19:24:53 +03:00
switch ( * p + + ) {
case rpc_auth_error :
break ;
case rpc_mismatch :
trace_rpc__mismatch ( task ) ;
error = - EPROTONOSUPPORT ;
goto out_err ;
default :
2019-03-15 16:29:00 +03:00
goto out_unparsable ;
2019-02-11 19:24:53 +03:00
}
2019-02-11 19:24:58 +03:00
p = xdr_inline_decode ( xdr , sizeof ( * p ) ) ;
if ( ! p )
goto out_unparsable ;
2019-02-11 19:24:53 +03:00
switch ( * p + + ) {
case rpc_autherr_rejectedcred :
case rpc_autherr_rejectedverf :
case rpcsec_gsserr_credproblem :
case rpcsec_gsserr_ctxproblem :
if ( ! task - > tk_cred_retry )
break ;
task - > tk_cred_retry - - ;
trace_rpc__stale_creds ( task ) ;
2019-05-29 19:49:52 +03:00
return - EKEYREJECTED ;
2019-02-11 19:24:53 +03:00
case rpc_autherr_badcred :
case rpc_autherr_badverf :
/* possibly garbled cred/verf? */
if ( ! task - > tk_garb_retry )
break ;
task - > tk_garb_retry - - ;
trace_rpc__bad_creds ( task ) ;
task - > tk_action = call_encode ;
2019-02-11 19:24:58 +03:00
return - EAGAIN ;
2019-02-11 19:24:53 +03:00
case rpc_autherr_tooweak :
trace_rpc__auth_tooweak ( task ) ;
pr_warn ( " RPC: server %s requires stronger authentication. \n " ,
task - > tk_xprt - > servername ) ;
break ;
default :
2019-03-15 16:29:00 +03:00
goto out_unparsable ;
2019-02-11 19:24:53 +03:00
}
goto out_err ;
2005-04-17 02:20:36 +04:00
}
2005-06-22 21:16:20 +04:00
2017-05-08 10:31:19 +03:00
static void rpcproc_encode_null ( struct rpc_rqst * rqstp , struct xdr_stream * xdr ,
const void * obj )
2005-06-22 21:16:20 +04:00
{
}
2017-05-08 15:58:11 +03:00
static int rpcproc_decode_null ( struct rpc_rqst * rqstp , struct xdr_stream * xdr ,
void * obj )
2005-06-22 21:16:20 +04:00
{
return 0 ;
}
2017-05-12 16:36:49 +03:00
static const struct rpc_procinfo rpcproc_null = {
2005-06-22 21:16:20 +04:00
. p_encode = rpcproc_encode_null ,
. p_decode = rpcproc_decode_null ,
} ;
2009-12-03 23:58:56 +03:00
static int rpc_ping ( struct rpc_clnt * clnt )
2005-06-22 21:16:20 +04:00
{
struct rpc_message msg = {
. rpc_proc = & rpcproc_null ,
} ;
int err ;
2018-12-03 03:30:30 +03:00
err = rpc_call_sync ( clnt , & msg , RPC_TASK_SOFT | RPC_TASK_SOFTCONN |
RPC_TASK_NULLCREDS ) ;
2005-06-22 21:16:20 +04:00
return err ;
}
2007-06-16 22:18:40 +04:00
2016-01-31 07:43:35 +03:00
static
struct rpc_task * rpc_call_null_helper ( struct rpc_clnt * clnt ,
struct rpc_xprt * xprt , struct rpc_cred * cred , int flags ,
const struct rpc_call_ops * ops , void * data )
2007-06-23 18:17:16 +04:00
{
struct rpc_message msg = {
. rpc_proc = & rpcproc_null ,
} ;
2007-07-14 23:39:59 +04:00
struct rpc_task_setup task_setup_data = {
. rpc_client = clnt ,
2016-01-31 07:43:35 +03:00
. rpc_xprt = xprt ,
2007-07-14 23:39:59 +04:00
. rpc_message = & msg ,
2018-12-03 03:30:30 +03:00
. rpc_op_cred = cred ,
2016-01-31 07:43:35 +03:00
. callback_ops = ( ops ! = NULL ) ? ops : & rpc_default_ops ,
. callback_data = data ,
2020-05-13 00:13:55 +03:00
. flags = flags | RPC_TASK_SOFT | RPC_TASK_SOFTCONN |
RPC_TASK_NULLCREDS ,
2007-07-14 23:39:59 +04:00
} ;
2016-01-31 07:43:35 +03:00
2007-07-14 23:39:59 +04:00
return rpc_run_task ( & task_setup_data ) ;
2007-06-23 18:17:16 +04:00
}
2016-01-31 07:43:35 +03:00
struct rpc_task * rpc_call_null ( struct rpc_clnt * clnt , struct rpc_cred * cred , int flags )
{
return rpc_call_null_helper ( clnt , NULL , cred , flags , NULL , NULL ) ;
}
2007-07-14 23:39:59 +04:00
EXPORT_SYMBOL_GPL ( rpc_call_null ) ;
2007-06-23 18:17:16 +04:00
2016-01-31 07:43:35 +03:00
struct rpc_cb_add_xprt_calldata {
struct rpc_xprt_switch * xps ;
struct rpc_xprt * xprt ;
} ;
static void rpc_cb_add_xprt_done ( struct rpc_task * task , void * calldata )
{
struct rpc_cb_add_xprt_calldata * data = calldata ;
if ( task - > tk_status = = 0 )
rpc_xprt_switch_add_xprt ( data - > xps , data - > xprt ) ;
}
static void rpc_cb_add_xprt_release ( void * calldata )
{
struct rpc_cb_add_xprt_calldata * data = calldata ;
xprt_put ( data - > xprt ) ;
xprt_switch_put ( data - > xps ) ;
kfree ( data ) ;
}
2016-07-25 00:06:28 +03:00
static const struct rpc_call_ops rpc_cb_add_xprt_call_ops = {
2016-01-31 07:43:35 +03:00
. rpc_call_done = rpc_cb_add_xprt_done ,
. rpc_release = rpc_cb_add_xprt_release ,
} ;
/**
* rpc_clnt_test_and_add_xprt - Test and add a new transport to a rpc_clnt
* @ clnt : pointer to struct rpc_clnt
* @ xps : pointer to struct rpc_xprt_switch ,
* @ xprt : pointer struct rpc_xprt
* @ dummy : unused
*/
int rpc_clnt_test_and_add_xprt ( struct rpc_clnt * clnt ,
struct rpc_xprt_switch * xps , struct rpc_xprt * xprt ,
void * dummy )
{
struct rpc_cb_add_xprt_calldata * data ;
struct rpc_task * task ;
data = kmalloc ( sizeof ( * data ) , GFP_NOFS ) ;
if ( ! data )
return - ENOMEM ;
data - > xps = xprt_switch_get ( xps ) ;
data - > xprt = xprt_get ( xprt ) ;
2017-04-27 15:50:51 +03:00
if ( rpc_xprt_switch_has_addr ( data - > xps , ( struct sockaddr * ) & xprt - > addr ) ) {
rpc_cb_add_xprt_release ( data ) ;
goto success ;
}
2016-01-31 07:43:35 +03:00
2020-05-13 00:13:55 +03:00
task = rpc_call_null_helper ( clnt , xprt , NULL , RPC_TASK_ASYNC ,
2016-01-31 07:43:35 +03:00
& rpc_cb_add_xprt_call_ops , data ) ;
2020-04-20 18:14:19 +03:00
2016-01-31 07:43:35 +03:00
rpc_put_task ( task ) ;
2017-04-27 15:50:51 +03:00
success :
2016-01-31 07:43:35 +03:00
return 1 ;
}
EXPORT_SYMBOL_GPL ( rpc_clnt_test_and_add_xprt ) ;
2016-09-09 16:22:26 +03:00
/**
* rpc_clnt_setup_test_and_add_xprt ( )
*
* This is an rpc_clnt_add_xprt setup ( ) function which returns 1 so :
* 1 ) caller of the test function must dereference the rpc_xprt_switch
* and the rpc_xprt .
* 2 ) test function must call rpc_xprt_switch_add_xprt , usually in
* the rpc_call_done routine .
*
* Upon success ( return of 1 ) , the test function adds the new
* transport to the rpc_clnt xprt switch
*
* @ clnt : struct rpc_clnt to get the new transport
* @ xps : the rpc_xprt_switch to hold the new transport
* @ xprt : the rpc_xprt to test
* @ data : a struct rpc_add_xprt_test pointer that holds the test function
* and test function call data
*/
int rpc_clnt_setup_test_and_add_xprt ( struct rpc_clnt * clnt ,
struct rpc_xprt_switch * xps ,
struct rpc_xprt * xprt ,
void * data )
{
struct rpc_task * task ;
struct rpc_add_xprt_test * xtest = ( struct rpc_add_xprt_test * ) data ;
int status = - EADDRINUSE ;
xprt = xprt_get ( xprt ) ;
xprt_switch_get ( xps ) ;
if ( rpc_xprt_switch_has_addr ( xps , ( struct sockaddr * ) & xprt - > addr ) )
goto out_err ;
/* Test the connection */
2020-05-13 00:13:55 +03:00
task = rpc_call_null_helper ( clnt , xprt , NULL , 0 , NULL , NULL ) ;
2016-09-09 16:22:26 +03:00
if ( IS_ERR ( task ) ) {
status = PTR_ERR ( task ) ;
goto out_err ;
}
status = task - > tk_status ;
rpc_put_task ( task ) ;
if ( status < 0 )
goto out_err ;
/* rpc_xprt_switch and rpc_xprt are deferrenced by add_xprt_test() */
xtest - > add_xprt_test ( clnt , xprt , xtest - > data ) ;
2018-12-19 09:59:57 +03:00
xprt_put ( xprt ) ;
xprt_switch_put ( xps ) ;
2016-09-09 16:22:26 +03:00
/* so that rpc_clnt_add_xprt does not call rpc_xprt_switch_add_xprt */
return 1 ;
out_err :
xprt_put ( xprt ) ;
xprt_switch_put ( xps ) ;
pr_info ( " RPC: rpc_clnt_test_xprt failed: %d addr %s not added \n " ,
status , xprt - > address_strings [ RPC_DISPLAY_ADDR ] ) ;
return status ;
}
EXPORT_SYMBOL_GPL ( rpc_clnt_setup_test_and_add_xprt ) ;
2016-01-31 07:43:35 +03:00
/**
* rpc_clnt_add_xprt - Add a new transport to a rpc_clnt
* @ clnt : pointer to struct rpc_clnt
* @ xprtargs : pointer to struct xprt_create
* @ setup : callback to test and / or set up the connection
* @ data : pointer to setup function data
*
* Creates a new transport using the parameters set in args and
* adds it to clnt .
* If ping is set , then test that connectivity succeeds before
* adding the new transport .
*
*/
int rpc_clnt_add_xprt ( struct rpc_clnt * clnt ,
struct xprt_create * xprtargs ,
int ( * setup ) ( struct rpc_clnt * ,
struct rpc_xprt_switch * ,
struct rpc_xprt * ,
void * ) ,
void * data )
{
struct rpc_xprt_switch * xps ;
struct rpc_xprt * xprt ;
2017-02-08 19:17:54 +03:00
unsigned long connect_timeout ;
2016-08-04 07:08:45 +03:00
unsigned long reconnect_timeout ;
2019-10-17 18:13:54 +03:00
unsigned char resvport , reuseport ;
2016-01-31 07:43:35 +03:00
int ret = 0 ;
rcu_read_lock ( ) ;
xps = xprt_switch_get ( rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ) ;
xprt = xprt_iter_xprt ( & clnt - > cl_xpi ) ;
if ( xps = = NULL | | xprt = = NULL ) {
rcu_read_unlock ( ) ;
2019-06-10 05:16:56 +03:00
xprt_switch_put ( xps ) ;
2016-01-31 07:43:35 +03:00
return - EAGAIN ;
}
resvport = xprt - > resvport ;
2019-10-17 18:13:54 +03:00
reuseport = xprt - > reuseport ;
2017-02-08 19:17:54 +03:00
connect_timeout = xprt - > connect_timeout ;
2016-08-04 07:08:45 +03:00
reconnect_timeout = xprt - > max_reconnect_timeout ;
2016-01-31 07:43:35 +03:00
rcu_read_unlock ( ) ;
xprt = xprt_create_transport ( xprtargs ) ;
if ( IS_ERR ( xprt ) ) {
ret = PTR_ERR ( xprt ) ;
goto out_put_switch ;
}
xprt - > resvport = resvport ;
2019-10-17 18:13:54 +03:00
xprt - > reuseport = reuseport ;
2017-02-08 19:17:54 +03:00
if ( xprt - > ops - > set_connect_timeout ! = NULL )
xprt - > ops - > set_connect_timeout ( xprt ,
connect_timeout ,
reconnect_timeout ) ;
2016-01-31 07:43:35 +03:00
rpc_xprt_switch_set_roundrobin ( xps ) ;
if ( setup ) {
ret = setup ( clnt , xps , xprt , data ) ;
if ( ret ! = 0 )
goto out_put_xprt ;
}
rpc_xprt_switch_add_xprt ( xps , xprt ) ;
out_put_xprt :
xprt_put ( xprt ) ;
out_put_switch :
xprt_switch_put ( xps ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( rpc_clnt_add_xprt ) ;
2017-02-08 19:17:54 +03:00
struct connect_timeout_data {
unsigned long connect_timeout ;
unsigned long reconnect_timeout ;
} ;
2016-08-06 02:03:31 +03:00
static int
2017-02-08 19:17:54 +03:00
rpc_xprt_set_connect_timeout ( struct rpc_clnt * clnt ,
2016-08-06 02:03:31 +03:00
struct rpc_xprt * xprt ,
void * data )
{
2017-02-08 19:17:54 +03:00
struct connect_timeout_data * timeo = data ;
2016-08-06 02:03:31 +03:00
2017-02-08 19:17:54 +03:00
if ( xprt - > ops - > set_connect_timeout )
xprt - > ops - > set_connect_timeout ( xprt ,
timeo - > connect_timeout ,
timeo - > reconnect_timeout ) ;
2016-08-06 02:03:31 +03:00
return 0 ;
}
void
2017-02-08 19:17:55 +03:00
rpc_set_connect_timeout ( struct rpc_clnt * clnt ,
unsigned long connect_timeout ,
unsigned long reconnect_timeout )
2016-08-06 02:03:31 +03:00
{
2017-02-08 19:17:54 +03:00
struct connect_timeout_data timeout = {
2017-02-08 19:17:55 +03:00
. connect_timeout = connect_timeout ,
. reconnect_timeout = reconnect_timeout ,
2017-02-08 19:17:54 +03:00
} ;
2016-08-06 02:03:31 +03:00
rpc_clnt_iterate_for_each_xprt ( clnt ,
2017-02-08 19:17:54 +03:00
rpc_xprt_set_connect_timeout ,
& timeout ) ;
2016-08-06 02:03:31 +03:00
}
2017-02-08 19:17:55 +03:00
EXPORT_SYMBOL_GPL ( rpc_set_connect_timeout ) ;
2016-08-06 02:03:31 +03:00
2016-09-09 16:22:23 +03:00
void rpc_clnt_xprt_switch_put ( struct rpc_clnt * clnt )
{
2016-10-26 17:33:31 +03:00
rcu_read_lock ( ) ;
2016-09-09 16:22:23 +03:00
xprt_switch_put ( rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ) ;
2016-10-26 17:33:31 +03:00
rcu_read_unlock ( ) ;
2016-09-09 16:22:23 +03:00
}
EXPORT_SYMBOL_GPL ( rpc_clnt_xprt_switch_put ) ;
2016-09-09 16:22:24 +03:00
void rpc_clnt_xprt_switch_add_xprt ( struct rpc_clnt * clnt , struct rpc_xprt * xprt )
{
2016-10-26 17:33:31 +03:00
rcu_read_lock ( ) ;
2016-09-09 16:22:24 +03:00
rpc_xprt_switch_add_xprt ( rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ,
xprt ) ;
2016-10-26 17:33:31 +03:00
rcu_read_unlock ( ) ;
2016-09-09 16:22:24 +03:00
}
EXPORT_SYMBOL_GPL ( rpc_clnt_xprt_switch_add_xprt ) ;
2016-09-09 16:22:25 +03:00
bool rpc_clnt_xprt_switch_has_addr ( struct rpc_clnt * clnt ,
const struct sockaddr * sap )
{
struct rpc_xprt_switch * xps ;
bool ret ;
rcu_read_lock ( ) ;
2016-10-26 17:33:31 +03:00
xps = rcu_dereference ( clnt - > cl_xpi . xpi_xpswitch ) ;
2016-09-09 16:22:25 +03:00
ret = rpc_xprt_switch_has_addr ( xps , sap ) ;
rcu_read_unlock ( ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( rpc_clnt_xprt_switch_has_addr ) ;
2014-11-18 00:58:04 +03:00
# if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
2008-05-22 01:09:26 +04:00
static void rpc_show_header ( void )
{
2008-05-22 01:09:41 +04:00
printk ( KERN_INFO " -pid- flgs status -client- --rqstp- "
" -timeout ---ops-- \n " ) ;
2008-05-22 01:09:26 +04:00
}
2008-05-22 01:09:33 +04:00
static void rpc_show_task ( const struct rpc_clnt * clnt ,
const struct rpc_task * task )
{
const char * rpc_waitq = " none " ;
if ( RPC_IS_QUEUED ( task ) )
rpc_waitq = rpc_qname ( task - > tk_waitqueue ) ;
2010-12-21 18:52:24 +03:00
printk ( KERN_INFO " %5u %04x %6d %8p %8p %8ld %8p %sv%u %s a:%ps q:%s \n " ,
2008-05-22 01:09:41 +04:00
task - > tk_pid , task - > tk_flags , task - > tk_status ,
2019-04-07 20:58:50 +03:00
clnt , task - > tk_rqstp , rpc_task_timeout ( task ) , task - > tk_ops ,
2013-08-23 19:48:15 +04:00
clnt - > cl_program - > name , clnt - > cl_vers , rpc_proc_name ( task ) ,
2010-12-21 18:52:24 +03:00
task - > tk_action , rpc_waitq ) ;
2008-05-22 01:09:33 +04:00
}
2012-01-12 22:07:51 +04:00
void rpc_show_tasks ( struct net * net )
2007-06-16 22:18:40 +04:00
{
struct rpc_clnt * clnt ;
2008-05-22 01:09:33 +04:00
struct rpc_task * task ;
2008-05-22 01:09:26 +04:00
int header = 0 ;
2012-01-12 22:07:51 +04:00
struct sunrpc_net * sn = net_generic ( net , sunrpc_net_id ) ;
2007-06-16 22:18:40 +04:00
2012-01-12 22:07:51 +04:00
spin_lock ( & sn - > rpc_client_lock ) ;
list_for_each_entry ( clnt , & sn - > all_clients , cl_clients ) {
2007-06-16 22:18:40 +04:00
spin_lock ( & clnt - > cl_lock ) ;
2008-05-22 01:09:33 +04:00
list_for_each_entry ( task , & clnt - > cl_tasks , tk_task ) {
2008-05-22 01:09:26 +04:00
if ( ! header ) {
rpc_show_header ( ) ;
header + + ;
}
2008-05-22 01:09:33 +04:00
rpc_show_task ( clnt , task ) ;
2007-06-16 22:18:40 +04:00
}
spin_unlock ( & clnt - > cl_lock ) ;
}
2012-01-12 22:07:51 +04:00
spin_unlock ( & sn - > rpc_client_lock ) ;
2007-06-16 22:18:40 +04:00
}
# endif
2015-06-03 23:14:25 +03:00
# if IS_ENABLED(CONFIG_SUNRPC_SWAP)
2016-01-31 04:05:34 +03:00
static int
rpc_clnt_swap_activate_callback ( struct rpc_clnt * clnt ,
struct rpc_xprt * xprt ,
void * dummy )
{
return xprt_enable_swap ( xprt ) ;
}
2015-06-03 23:14:25 +03:00
int
rpc_clnt_swap_activate ( struct rpc_clnt * clnt )
{
2016-01-31 04:05:34 +03:00
if ( atomic_inc_return ( & clnt - > cl_swapper ) = = 1 )
return rpc_clnt_iterate_for_each_xprt ( clnt ,
rpc_clnt_swap_activate_callback , NULL ) ;
return 0 ;
2015-06-03 23:14:25 +03:00
}
EXPORT_SYMBOL_GPL ( rpc_clnt_swap_activate ) ;
2016-01-31 04:05:34 +03:00
static int
rpc_clnt_swap_deactivate_callback ( struct rpc_clnt * clnt ,
struct rpc_xprt * xprt ,
void * dummy )
{
xprt_disable_swap ( xprt ) ;
return 0 ;
}
2015-06-03 23:14:25 +03:00
void
rpc_clnt_swap_deactivate ( struct rpc_clnt * clnt )
{
2016-01-31 04:05:34 +03:00
if ( atomic_dec_if_positive ( & clnt - > cl_swapper ) = = 0 )
rpc_clnt_iterate_for_each_xprt ( clnt ,
rpc_clnt_swap_deactivate_callback , NULL ) ;
2015-06-03 23:14:25 +03:00
}
EXPORT_SYMBOL_GPL ( rpc_clnt_swap_deactivate ) ;
# endif /* CONFIG_SUNRPC_SWAP */