2012-05-16 23:58:40 +04:00
# define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
2007-09-12 13:50:50 +04:00
# include <linux/workqueue.h>
# include <linux/rtnetlink.h>
# include <linux/cache.h>
# include <linux/slab.h>
# include <linux/list.h>
# include <linux/delay.h>
2007-09-27 09:04:26 +04:00
# include <linux/sched.h>
2008-04-15 11:35:23 +04:00
# include <linux/idr.h>
2009-07-10 13:51:33 +04:00
# include <linux/rculist.h>
2009-07-10 13:51:35 +04:00
# include <linux/nsproxy.h>
2013-04-12 04:50:06 +04:00
# include <linux/fs.h>
# include <linux/proc_ns.h>
2011-05-05 04:51:50 +04:00
# include <linux/file.h>
2011-07-15 19:47:34 +04:00
# include <linux/export.h>
2012-06-14 13:31:10 +04:00
# include <linux/user_namespace.h>
2015-01-15 17:11:15 +03:00
# include <linux/net_namespace.h>
# include <linux/rtnetlink.h>
# include <net/sock.h>
# include <net/netlink.h>
2007-09-12 13:50:50 +04:00
# include <net/net_namespace.h>
2008-04-15 11:36:08 +04:00
# include <net/netns/generic.h>
2007-09-12 13:50:50 +04:00
/*
* Our network namespace constructor / destructor lists
*/
static LIST_HEAD ( pernet_list ) ;
static struct list_head * first_device = & pernet_list ;
2014-05-13 02:11:20 +04:00
DEFINE_MUTEX ( net_mutex ) ;
2007-09-12 13:50:50 +04:00
LIST_HEAD ( net_namespace_list ) ;
2008-10-08 13:35:06 +04:00
EXPORT_SYMBOL_GPL ( net_namespace_list ) ;
2007-09-12 13:50:50 +04:00
2012-07-18 13:06:07 +04:00
struct net init_net = {
. dev_base_head = LIST_HEAD_INIT ( init_net . dev_base_head ) ,
} ;
2008-01-23 09:05:33 +03:00
EXPORT_SYMBOL ( init_net ) ;
2007-09-12 13:50:50 +04:00
2008-04-15 11:36:08 +04:00
# define INITIAL_NET_GEN_PTRS 13 /* +1 for len +2 for rcu_head */
2012-01-26 04:41:38 +04:00
static unsigned int max_gen_ptrs = INITIAL_NET_GEN_PTRS ;
static struct net_generic * net_alloc_generic ( void )
{
struct net_generic * ng ;
size_t generic_size = offsetof ( struct net_generic , ptr [ max_gen_ptrs ] ) ;
ng = kzalloc ( generic_size , GFP_KERNEL ) ;
if ( ng )
ng - > len = max_gen_ptrs ;
return ng ;
}
2010-04-23 05:40:47 +04:00
static int net_assign_generic ( struct net * net , int id , void * data )
{
struct net_generic * ng , * old_ng ;
BUG_ON ( ! mutex_is_locked ( & net_mutex ) ) ;
BUG_ON ( id = = 0 ) ;
2010-10-25 07:20:11 +04:00
old_ng = rcu_dereference_protected ( net - > gen ,
lockdep_is_held ( & net_mutex ) ) ;
ng = old_ng ;
2010-04-23 05:40:47 +04:00
if ( old_ng - > len > = id )
goto assign ;
2012-01-26 04:41:38 +04:00
ng = net_alloc_generic ( ) ;
2010-04-23 05:40:47 +04:00
if ( ng = = NULL )
return - ENOMEM ;
/*
* Some synchronisation notes :
*
* The net_generic explores the net - > gen array inside rcu
* read section . Besides once set the net - > gen - > ptr [ x ]
* pointer never changes ( see rules in netns / generic . h ) .
*
* That said , we simply duplicate this array and schedule
* the old copy for kfree after a grace period .
*/
memcpy ( & ng - > ptr , & old_ng - > ptr , old_ng - > len * sizeof ( void * ) ) ;
rcu_assign_pointer ( net - > gen , ng ) ;
2011-03-18 07:06:32 +03:00
kfree_rcu ( old_ng , rcu ) ;
2010-04-23 05:40:47 +04:00
assign :
ng - > ptr [ id - 1 ] = data ;
return 0 ;
}
2009-11-30 01:25:28 +03:00
static int ops_init ( const struct pernet_operations * ops , struct net * net )
{
2012-04-16 08:43:15 +04:00
int err = - ENOMEM ;
void * data = NULL ;
2009-11-30 01:25:28 +03:00
if ( ops - > id & & ops - > size ) {
2012-04-16 08:43:15 +04:00
data = kzalloc ( ops - > size , GFP_KERNEL ) ;
2009-11-30 01:25:28 +03:00
if ( ! data )
2012-04-16 08:43:15 +04:00
goto out ;
2009-11-30 01:25:28 +03:00
err = net_assign_generic ( net , * ops - > id , data ) ;
2012-04-16 08:43:15 +04:00
if ( err )
goto cleanup ;
2009-11-30 01:25:28 +03:00
}
2012-04-16 08:43:15 +04:00
err = 0 ;
2009-11-30 01:25:28 +03:00
if ( ops - > init )
2012-04-16 08:43:15 +04:00
err = ops - > init ( net ) ;
if ( ! err )
return 0 ;
cleanup :
kfree ( data ) ;
out :
return err ;
2009-11-30 01:25:28 +03:00
}
static void ops_free ( const struct pernet_operations * ops , struct net * net )
{
if ( ops - > id & & ops - > size ) {
int id = * ops - > id ;
kfree ( net_generic ( net , id ) ) ;
}
}
2009-12-03 05:29:03 +03:00
static void ops_exit_list ( const struct pernet_operations * ops ,
struct list_head * net_exit_list )
{
struct net * net ;
if ( ops - > exit ) {
list_for_each_entry ( net , net_exit_list , exit_list )
ops - > exit ( net ) ;
}
if ( ops - > exit_batch )
ops - > exit_batch ( net_exit_list ) ;
}
static void ops_free_list ( const struct pernet_operations * ops ,
struct list_head * net_exit_list )
{
struct net * net ;
if ( ops - > size & & ops - > id ) {
list_for_each_entry ( net , net_exit_list , exit_list )
ops_free ( ops , net ) ;
}
}
2015-01-15 17:11:15 +03:00
static int alloc_netid ( struct net * net , struct net * peer , int reqid )
{
int min = 0 , max = 0 ;
ASSERT_RTNL ( ) ;
if ( reqid > = 0 ) {
min = reqid ;
max = reqid + 1 ;
}
return idr_alloc ( & net - > netns_ids , peer , min , max , GFP_KERNEL ) ;
}
/* This function is used by idr_for_each(). If net is equal to peer, the
* function returns the id so that idr_for_each ( ) stops . Because we cannot
* returns the id 0 ( idr_for_each ( ) will not stop ) , we return the magic value
* NET_ID_ZERO ( - 1 ) for it .
*/
# define NET_ID_ZERO -1
static int net_eq_idr ( int id , void * net , void * peer )
{
if ( net_eq ( net , peer ) )
return id ? : NET_ID_ZERO ;
return 0 ;
}
static int __peernet2id ( struct net * net , struct net * peer , bool alloc )
{
int id = idr_for_each ( & net - > netns_ids , net_eq_idr , peer ) ;
ASSERT_RTNL ( ) ;
/* Magic value for id 0. */
if ( id = = NET_ID_ZERO )
return 0 ;
if ( id > 0 )
return id ;
if ( alloc )
return alloc_netid ( net , peer , - 1 ) ;
return - ENOENT ;
}
/* This function returns the id of a peer netns. If no id is assigned, one will
* be allocated and returned .
*/
int peernet2id ( struct net * net , struct net * peer )
{
2015-04-03 13:02:37 +03:00
bool alloc = atomic_read ( & peer - > count ) = = 0 ? false : true ;
int id ;
2015-01-15 17:11:15 +03:00
2015-04-03 13:02:37 +03:00
id = __peernet2id ( net , peer , alloc ) ;
2015-01-15 17:11:15 +03:00
return id > = 0 ? id : NETNSA_NSID_NOT_ASSIGNED ;
}
2015-01-20 17:15:47 +03:00
EXPORT_SYMBOL ( peernet2id ) ;
2015-01-15 17:11:15 +03:00
struct net * get_net_ns_by_id ( struct net * net , int id )
{
struct net * peer ;
if ( id < 0 )
return NULL ;
rcu_read_lock ( ) ;
peer = idr_find ( & net - > netns_ids , id ) ;
if ( peer )
get_net ( peer ) ;
rcu_read_unlock ( ) ;
return peer ;
}
2007-09-12 13:50:50 +04:00
/*
* setup_net runs the initializers for the network namespace object .
*/
2012-06-14 13:31:10 +04:00
static __net_init int setup_net ( struct net * net , struct user_namespace * user_ns )
2007-09-12 13:50:50 +04:00
{
/* Must be called with net_mutex held */
2009-11-30 01:25:28 +03:00
const struct pernet_operations * ops , * saved_ops ;
2009-02-22 11:07:53 +03:00
int error = 0 ;
2009-12-03 05:29:03 +03:00
LIST_HEAD ( net_exit_list ) ;
2007-09-12 13:50:50 +04:00
atomic_set ( & net - > count , 1 ) ;
2011-06-09 05:13:01 +04:00
atomic_set ( & net - > passive , 1 ) ;
2011-06-21 07:11:20 +04:00
net - > dev_base_seq = 1 ;
2012-06-14 13:31:10 +04:00
net - > user_ns = user_ns ;
2015-01-15 17:11:15 +03:00
idr_init ( & net - > netns_ids ) ;
2009-02-22 11:07:53 +03:00
2007-09-19 00:20:41 +04:00
list_for_each_entry ( ops , & pernet_list , list ) {
2009-11-30 01:25:28 +03:00
error = ops_init ( ops , net ) ;
if ( error < 0 )
goto out_undo ;
2007-09-12 13:50:50 +04:00
}
out :
return error ;
2007-09-19 00:20:41 +04:00
2007-09-12 13:50:50 +04:00
out_undo :
/* Walk through the list backwards calling the exit functions
* for the pernet modules whose init functions did not fail .
*/
2009-12-03 05:29:03 +03:00
list_add ( & net - > exit_list , & net_exit_list ) ;
2009-11-30 01:25:28 +03:00
saved_ops = ops ;
2009-12-03 05:29:03 +03:00
list_for_each_entry_continue_reverse ( ops , & pernet_list , list )
ops_exit_list ( ops , & net_exit_list ) ;
2009-11-30 01:25:28 +03:00
ops = saved_ops ;
list_for_each_entry_continue_reverse ( ops , & pernet_list , list )
2009-12-03 05:29:03 +03:00
ops_free_list ( ops , & net_exit_list ) ;
2007-10-31 01:38:57 +03:00
rcu_barrier ( ) ;
2007-09-12 13:50:50 +04:00
goto out ;
}
2007-11-01 10:44:50 +03:00
2009-02-24 02:37:35 +03:00
# ifdef CONFIG_NET_NS
static struct kmem_cache * net_cachep ;
static struct workqueue_struct * netns_wq ;
2009-02-22 11:07:53 +03:00
static struct net * net_alloc ( void )
2007-11-07 12:30:30 +03:00
{
2009-02-22 11:07:53 +03:00
struct net * net = NULL ;
struct net_generic * ng ;
ng = net_alloc_generic ( ) ;
if ( ! ng )
goto out ;
net = kmem_cache_zalloc ( net_cachep , GFP_KERNEL ) ;
2007-11-07 12:30:30 +03:00
if ( ! net )
2009-02-22 11:07:53 +03:00
goto out_free ;
2007-11-07 12:30:30 +03:00
2009-02-22 11:07:53 +03:00
rcu_assign_pointer ( net - > gen , ng ) ;
out :
return net ;
out_free :
kfree ( ng ) ;
goto out ;
}
static void net_free ( struct net * net )
{
2014-09-09 19:24:53 +04:00
kfree ( rcu_access_pointer ( net - > gen ) ) ;
2007-11-07 12:30:30 +03:00
kmem_cache_free ( net_cachep , net ) ;
}
2011-06-09 05:13:01 +04:00
void net_drop_ns ( void * p )
{
struct net * ns = p ;
if ( ns & & atomic_dec_and_test ( & ns - > passive ) )
net_free ( ns ) ;
}
2012-06-14 13:31:10 +04:00
struct net * copy_net_ns ( unsigned long flags ,
struct user_namespace * user_ns , struct net * old_net )
2007-09-27 09:04:26 +04:00
{
2009-05-04 22:12:14 +04:00
struct net * net ;
int rv ;
2007-09-27 09:04:26 +04:00
2011-04-15 06:26:25 +04:00
if ( ! ( flags & CLONE_NEWNET ) )
return get_net ( old_net ) ;
2009-05-04 22:12:14 +04:00
net = net_alloc ( ) ;
if ( ! net )
return ERR_PTR ( - ENOMEM ) ;
2012-06-14 13:31:10 +04:00
get_user_ns ( user_ns ) ;
2007-09-27 09:04:26 +04:00
mutex_lock ( & net_mutex ) ;
2012-06-14 13:31:10 +04:00
rv = setup_net ( net , user_ns ) ;
2009-05-04 22:12:14 +04:00
if ( rv = = 0 ) {
2009-02-22 11:07:53 +03:00
rtnl_lock ( ) ;
2009-07-10 13:51:33 +04:00
list_add_tail_rcu ( & net - > list , & net_namespace_list ) ;
2009-02-22 11:07:53 +03:00
rtnl_unlock ( ) ;
}
2007-09-27 09:04:26 +04:00
mutex_unlock ( & net_mutex ) ;
2009-05-04 22:12:14 +04:00
if ( rv < 0 ) {
2012-06-14 13:31:10 +04:00
put_user_ns ( user_ns ) ;
2011-06-09 05:13:01 +04:00
net_drop_ns ( net ) ;
2009-05-04 22:12:14 +04:00
return ERR_PTR ( rv ) ;
}
return net ;
}
2009-02-22 11:07:53 +03:00
2009-11-30 01:25:27 +03:00
static DEFINE_SPINLOCK ( cleanup_list_lock ) ;
static LIST_HEAD ( cleanup_list ) ; /* Must hold cleanup_list_lock to touch */
2007-11-01 10:44:50 +03:00
static void cleanup_net ( struct work_struct * work )
{
2009-11-30 01:25:28 +03:00
const struct pernet_operations * ops ;
2015-04-03 13:02:36 +03:00
struct net * net , * tmp ;
2014-04-25 04:50:54 +04:00
struct list_head net_kill_list ;
2009-12-03 05:29:03 +03:00
LIST_HEAD ( net_exit_list ) ;
2007-11-01 10:44:50 +03:00
2009-11-30 01:25:27 +03:00
/* Atomically snapshot the list of namespaces to cleanup */
spin_lock_irq ( & cleanup_list_lock ) ;
list_replace_init ( & cleanup_list , & net_kill_list ) ;
spin_unlock_irq ( & cleanup_list_lock ) ;
2007-11-01 10:44:50 +03:00
mutex_lock ( & net_mutex ) ;
/* Don't let anyone else find us. */
rtnl_lock ( ) ;
2009-12-03 05:29:03 +03:00
list_for_each_entry ( net , & net_kill_list , cleanup_list ) {
2009-11-30 01:25:27 +03:00
list_del_rcu ( & net - > list ) ;
2009-12-03 05:29:03 +03:00
list_add_tail ( & net - > exit_list , & net_exit_list ) ;
2015-04-03 13:02:36 +03:00
for_each_net ( tmp ) {
int id = __peernet2id ( tmp , net , false ) ;
if ( id > = 0 )
idr_remove ( & tmp - > netns_ids , id ) ;
}
idr_destroy ( & net - > netns_ids ) ;
2009-12-03 05:29:03 +03:00
}
2007-11-01 10:44:50 +03:00
rtnl_unlock ( ) ;
2009-07-10 13:51:33 +04:00
/*
* Another CPU might be rcu - iterating the list , wait for it .
* This needs to be before calling the exit ( ) notifiers , so
* the rcu_barrier ( ) below isn ' t sufficient alone .
*/
synchronize_rcu ( ) ;
2007-11-01 10:44:50 +03:00
/* Run all of the network namespace exit methods */
2009-12-03 05:29:03 +03:00
list_for_each_entry_reverse ( ops , & pernet_list , list )
ops_exit_list ( ops , & net_exit_list ) ;
2009-11-30 01:25:28 +03:00
/* Free the net generic variables */
2009-12-03 05:29:03 +03:00
list_for_each_entry_reverse ( ops , & pernet_list , list )
ops_free_list ( ops , & net_exit_list ) ;
2007-11-01 10:44:50 +03:00
mutex_unlock ( & net_mutex ) ;
/* Ensure there are no outstanding rcu callbacks using this
* network namespace .
*/
rcu_barrier ( ) ;
/* Finally it is safe to free my network namespace structure */
2009-12-03 05:29:03 +03:00
list_for_each_entry_safe ( net , tmp , & net_exit_list , exit_list ) {
list_del_init ( & net - > exit_list ) ;
2012-06-14 13:31:10 +04:00
put_user_ns ( net - > user_ns ) ;
2011-06-09 05:13:01 +04:00
net_drop_ns ( net ) ;
2009-11-30 01:25:27 +03:00
}
2007-11-01 10:44:50 +03:00
}
2009-11-30 01:25:27 +03:00
static DECLARE_WORK ( net_cleanup_work , cleanup_net ) ;
2007-11-01 10:44:50 +03:00
void __put_net ( struct net * net )
{
/* Cleanup the network namespace in process context */
2009-11-30 01:25:27 +03:00
unsigned long flags ;
spin_lock_irqsave ( & cleanup_list_lock , flags ) ;
list_add ( & net - > cleanup_list , & cleanup_list ) ;
spin_unlock_irqrestore ( & cleanup_list_lock , flags ) ;
queue_work ( netns_wq , & net_cleanup_work ) ;
2007-11-01 10:44:50 +03:00
}
EXPORT_SYMBOL_GPL ( __put_net ) ;
2011-05-12 07:51:13 +04:00
struct net * get_net_ns_by_fd ( int fd )
{
struct file * file ;
2014-11-01 09:32:53 +03:00
struct ns_common * ns ;
2011-05-12 07:51:13 +04:00
struct net * net ;
file = proc_ns_fget ( fd ) ;
2011-06-05 04:37:35 +04:00
if ( IS_ERR ( file ) )
return ERR_CAST ( file ) ;
2011-05-12 07:51:13 +04:00
2014-11-01 10:13:17 +03:00
ns = get_proc_ns ( file_inode ( file ) ) ;
2014-11-01 09:32:53 +03:00
if ( ns - > ops = = & netns_operations )
net = get_net ( container_of ( ns , struct net , ns ) ) ;
2011-06-05 04:37:35 +04:00
else
net = ERR_PTR ( - EINVAL ) ;
2011-05-12 07:51:13 +04:00
2011-06-05 04:37:35 +04:00
fput ( file ) ;
2011-05-12 07:51:13 +04:00
return net ;
}
2007-11-01 10:44:50 +03:00
# else
2011-05-12 07:51:13 +04:00
struct net * get_net_ns_by_fd ( int fd )
{
return ERR_PTR ( - EINVAL ) ;
}
2007-11-01 10:44:50 +03:00
# endif
2015-01-12 17:34:05 +03:00
EXPORT_SYMBOL_GPL ( get_net_ns_by_fd ) ;
2007-11-01 10:44:50 +03:00
2009-07-10 13:51:35 +04:00
struct net * get_net_ns_by_pid ( pid_t pid )
{
struct task_struct * tsk ;
struct net * net ;
/* Lookup the network namespace */
net = ERR_PTR ( - ESRCH ) ;
rcu_read_lock ( ) ;
tsk = find_task_by_vpid ( pid ) ;
if ( tsk ) {
struct nsproxy * nsproxy ;
2014-02-04 07:13:49 +04:00
task_lock ( tsk ) ;
nsproxy = tsk - > nsproxy ;
2009-07-10 13:51:35 +04:00
if ( nsproxy )
net = get_net ( nsproxy - > net_ns ) ;
2014-02-04 07:13:49 +04:00
task_unlock ( tsk ) ;
2009-07-10 13:51:35 +04:00
}
rcu_read_unlock ( ) ;
return net ;
}
EXPORT_SYMBOL_GPL ( get_net_ns_by_pid ) ;
2011-06-15 21:21:48 +04:00
static __net_init int net_ns_net_init ( struct net * net )
{
2014-11-01 09:32:53 +03:00
# ifdef CONFIG_NET_NS
net - > ns . ops = & netns_operations ;
# endif
2014-11-01 07:45:45 +03:00
return ns_alloc_inum ( & net - > ns ) ;
2011-06-15 21:21:48 +04:00
}
static __net_exit void net_ns_net_exit ( struct net * net )
{
2014-11-01 07:45:45 +03:00
ns_free_inum ( & net - > ns ) ;
2011-06-15 21:21:48 +04:00
}
static struct pernet_operations __net_initdata net_ns_ops = {
. init = net_ns_net_init ,
. exit = net_ns_net_exit ,
} ;
2015-01-15 17:11:15 +03:00
static struct nla_policy rtnl_net_policy [ NETNSA_MAX + 1 ] = {
[ NETNSA_NONE ] = { . type = NLA_UNSPEC } ,
[ NETNSA_NSID ] = { . type = NLA_S32 } ,
[ NETNSA_PID ] = { . type = NLA_U32 } ,
[ NETNSA_FD ] = { . type = NLA_U32 } ,
} ;
static int rtnl_net_newid ( struct sk_buff * skb , struct nlmsghdr * nlh )
{
struct net * net = sock_net ( skb - > sk ) ;
struct nlattr * tb [ NETNSA_MAX + 1 ] ;
struct net * peer ;
int nsid , err ;
err = nlmsg_parse ( nlh , sizeof ( struct rtgenmsg ) , tb , NETNSA_MAX ,
rtnl_net_policy ) ;
if ( err < 0 )
return err ;
if ( ! tb [ NETNSA_NSID ] )
return - EINVAL ;
nsid = nla_get_s32 ( tb [ NETNSA_NSID ] ) ;
if ( tb [ NETNSA_PID ] )
peer = get_net_ns_by_pid ( nla_get_u32 ( tb [ NETNSA_PID ] ) ) ;
else if ( tb [ NETNSA_FD ] )
peer = get_net_ns_by_fd ( nla_get_u32 ( tb [ NETNSA_FD ] ) ) ;
else
return - EINVAL ;
if ( IS_ERR ( peer ) )
return PTR_ERR ( peer ) ;
if ( __peernet2id ( net , peer , false ) > = 0 ) {
err = - EEXIST ;
goto out ;
}
err = alloc_netid ( net , peer , nsid ) ;
if ( err > 0 )
err = 0 ;
out :
put_net ( peer ) ;
return err ;
}
static int rtnl_net_get_size ( void )
{
return NLMSG_ALIGN ( sizeof ( struct rtgenmsg ) )
+ nla_total_size ( sizeof ( s32 ) ) /* NETNSA_NSID */
;
}
static int rtnl_net_fill ( struct sk_buff * skb , u32 portid , u32 seq , int flags ,
int cmd , struct net * net , struct net * peer )
{
struct nlmsghdr * nlh ;
struct rtgenmsg * rth ;
int id ;
ASSERT_RTNL ( ) ;
nlh = nlmsg_put ( skb , portid , seq , cmd , sizeof ( * rth ) , flags ) ;
if ( ! nlh )
return - EMSGSIZE ;
rth = nlmsg_data ( nlh ) ;
rth - > rtgen_family = AF_UNSPEC ;
id = __peernet2id ( net , peer , false ) ;
if ( id < 0 )
id = NETNSA_NSID_NOT_ASSIGNED ;
if ( nla_put_s32 ( skb , NETNSA_NSID , id ) )
goto nla_put_failure ;
nlmsg_end ( skb , nlh ) ;
return 0 ;
nla_put_failure :
nlmsg_cancel ( skb , nlh ) ;
return - EMSGSIZE ;
}
static int rtnl_net_getid ( struct sk_buff * skb , struct nlmsghdr * nlh )
{
struct net * net = sock_net ( skb - > sk ) ;
struct nlattr * tb [ NETNSA_MAX + 1 ] ;
struct sk_buff * msg ;
int err = - ENOBUFS ;
struct net * peer ;
err = nlmsg_parse ( nlh , sizeof ( struct rtgenmsg ) , tb , NETNSA_MAX ,
rtnl_net_policy ) ;
if ( err < 0 )
return err ;
if ( tb [ NETNSA_PID ] )
peer = get_net_ns_by_pid ( nla_get_u32 ( tb [ NETNSA_PID ] ) ) ;
else if ( tb [ NETNSA_FD ] )
peer = get_net_ns_by_fd ( nla_get_u32 ( tb [ NETNSA_FD ] ) ) ;
else
return - EINVAL ;
if ( IS_ERR ( peer ) )
return PTR_ERR ( peer ) ;
msg = nlmsg_new ( rtnl_net_get_size ( ) , GFP_KERNEL ) ;
if ( ! msg ) {
err = - ENOMEM ;
goto out ;
}
err = rtnl_net_fill ( msg , NETLINK_CB ( skb ) . portid , nlh - > nlmsg_seq , 0 ,
RTM_GETNSID , net , peer ) ;
if ( err < 0 )
goto err_out ;
err = rtnl_unicast ( msg , net , NETLINK_CB ( skb ) . portid ) ;
goto out ;
err_out :
nlmsg_free ( msg ) ;
out :
put_net ( peer ) ;
return err ;
}
2007-09-12 13:50:50 +04:00
static int __init net_ns_init ( void )
{
2009-02-22 11:07:53 +03:00
struct net_generic * ng ;
2007-09-12 13:50:50 +04:00
2007-11-01 10:46:50 +03:00
# ifdef CONFIG_NET_NS
2007-09-12 13:50:50 +04:00
net_cachep = kmem_cache_create ( " net_namespace " , sizeof ( struct net ) ,
SMP_CACHE_BYTES ,
SLAB_PANIC , NULL ) ;
2007-11-20 10:18:16 +03:00
/* Create workqueue for cleanup */
netns_wq = create_singlethread_workqueue ( " netns " ) ;
if ( ! netns_wq )
panic ( " Could not create netns workq " ) ;
2007-11-01 10:46:50 +03:00
# endif
2007-11-20 10:18:16 +03:00
2009-02-22 11:07:53 +03:00
ng = net_alloc_generic ( ) ;
if ( ! ng )
panic ( " Could not allocate generic netns " ) ;
rcu_assign_pointer ( init_net . gen , ng ) ;
2007-09-12 13:50:50 +04:00
mutex_lock ( & net_mutex ) ;
2012-06-14 13:31:10 +04:00
if ( setup_net ( & init_net , & init_user_ns ) )
2009-05-22 02:10:31 +04:00
panic ( " Could not setup the initial network namespace " ) ;
2007-09-12 13:50:50 +04:00
2007-09-27 09:40:08 +04:00
rtnl_lock ( ) ;
2009-07-10 13:51:33 +04:00
list_add_tail_rcu ( & init_net . list , & net_namespace_list ) ;
2007-09-27 09:40:08 +04:00
rtnl_unlock ( ) ;
2007-09-12 13:50:50 +04:00
mutex_unlock ( & net_mutex ) ;
2011-06-15 21:21:48 +04:00
register_pernet_subsys ( & net_ns_ops ) ;
2015-01-15 17:11:15 +03:00
rtnl_register ( PF_UNSPEC , RTM_NEWNSID , rtnl_net_newid , NULL , NULL ) ;
rtnl_register ( PF_UNSPEC , RTM_GETNSID , rtnl_net_getid , NULL , NULL ) ;
2007-09-12 13:50:50 +04:00
return 0 ;
}
pure_initcall ( net_ns_init ) ;
2007-11-13 14:23:21 +03:00
# ifdef CONFIG_NET_NS
2009-11-30 01:25:28 +03:00
static int __register_pernet_operations ( struct list_head * list ,
struct pernet_operations * ops )
2007-09-12 13:50:50 +04:00
{
2009-12-03 05:29:03 +03:00
struct net * net ;
2007-09-12 13:50:50 +04:00
int error ;
2009-12-03 05:29:03 +03:00
LIST_HEAD ( net_exit_list ) ;
2007-09-12 13:50:50 +04:00
list_add_tail ( & ops - > list , list ) ;
2009-11-30 01:25:28 +03:00
if ( ops - > init | | ( ops - > id & & ops - > size ) ) {
2007-11-01 10:42:43 +03:00
for_each_net ( net ) {
2009-11-30 01:25:28 +03:00
error = ops_init ( ops , net ) ;
2007-09-12 13:50:50 +04:00
if ( error )
goto out_undo ;
2009-12-03 05:29:03 +03:00
list_add_tail ( & net - > exit_list , & net_exit_list ) ;
2007-09-12 13:50:50 +04:00
}
}
2007-11-01 10:42:43 +03:00
return 0 ;
2007-09-12 13:50:50 +04:00
out_undo :
/* If I have an error cleanup all namespaces I initialized */
list_del ( & ops - > list ) ;
2009-12-03 05:29:03 +03:00
ops_exit_list ( ops , & net_exit_list ) ;
ops_free_list ( ops , & net_exit_list ) ;
2007-11-01 10:42:43 +03:00
return error ;
2007-09-12 13:50:50 +04:00
}
2009-11-30 01:25:28 +03:00
static void __unregister_pernet_operations ( struct pernet_operations * ops )
2007-09-12 13:50:50 +04:00
{
struct net * net ;
2009-12-03 05:29:03 +03:00
LIST_HEAD ( net_exit_list ) ;
2007-09-12 13:50:50 +04:00
list_del ( & ops - > list ) ;
2009-12-03 05:29:03 +03:00
for_each_net ( net )
list_add_tail ( & net - > exit_list , & net_exit_list ) ;
ops_exit_list ( ops , & net_exit_list ) ;
ops_free_list ( ops , & net_exit_list ) ;
2007-09-12 13:50:50 +04:00
}
2007-11-13 14:23:21 +03:00
# else
2009-11-30 01:25:28 +03:00
static int __register_pernet_operations ( struct list_head * list ,
struct pernet_operations * ops )
2007-11-13 14:23:21 +03:00
{
2012-04-16 08:43:15 +04:00
return ops_init ( ops , & init_net ) ;
2007-11-13 14:23:21 +03:00
}
2009-11-30 01:25:28 +03:00
static void __unregister_pernet_operations ( struct pernet_operations * ops )
2007-11-13 14:23:21 +03:00
{
2009-12-03 05:29:03 +03:00
LIST_HEAD ( net_exit_list ) ;
list_add ( & init_net . exit_list , & net_exit_list ) ;
ops_exit_list ( ops , & net_exit_list ) ;
ops_free_list ( ops , & net_exit_list ) ;
2007-11-13 14:23:21 +03:00
}
2009-11-30 01:25:28 +03:00
# endif /* CONFIG_NET_NS */
2007-11-13 14:23:21 +03:00
2008-04-15 11:35:23 +04:00
static DEFINE_IDA ( net_generic_ids ) ;
2009-11-30 01:25:28 +03:00
static int register_pernet_operations ( struct list_head * list ,
struct pernet_operations * ops )
{
int error ;
if ( ops - > id ) {
again :
error = ida_get_new_above ( & net_generic_ids , 1 , ops - > id ) ;
if ( error < 0 ) {
if ( error = = - EAGAIN ) {
ida_pre_get ( & net_generic_ids , GFP_KERNEL ) ;
goto again ;
}
return error ;
}
2012-01-26 04:41:38 +04:00
max_gen_ptrs = max_t ( unsigned int , max_gen_ptrs , * ops - > id ) ;
2009-11-30 01:25:28 +03:00
}
error = __register_pernet_operations ( list , ops ) ;
2009-12-03 05:29:06 +03:00
if ( error ) {
rcu_barrier ( ) ;
if ( ops - > id )
ida_remove ( & net_generic_ids , * ops - > id ) ;
}
2009-11-30 01:25:28 +03:00
return error ;
}
static void unregister_pernet_operations ( struct pernet_operations * ops )
{
__unregister_pernet_operations ( ops ) ;
2009-12-03 05:29:06 +03:00
rcu_barrier ( ) ;
2009-11-30 01:25:28 +03:00
if ( ops - > id )
ida_remove ( & net_generic_ids , * ops - > id ) ;
}
2007-09-12 13:50:50 +04:00
/**
* register_pernet_subsys - register a network namespace subsystem
* @ ops : pernet operations structure for the subsystem
*
* Register a subsystem which has init and exit functions
* that are called when network namespaces are created and
* destroyed respectively .
*
* When registered all network namespace init functions are
* called for every existing network namespace . Allowing kernel
* modules to have a race free view of the set of network namespaces .
*
* When a new network namespace is created all of the init
* methods are called in the order in which they were registered .
*
* When a network namespace is destroyed all of the exit methods
* are called in the reverse of the order with which they were
* registered .
*/
int register_pernet_subsys ( struct pernet_operations * ops )
{
int error ;
mutex_lock ( & net_mutex ) ;
error = register_pernet_operations ( first_device , ops ) ;
mutex_unlock ( & net_mutex ) ;
return error ;
}
EXPORT_SYMBOL_GPL ( register_pernet_subsys ) ;
/**
* unregister_pernet_subsys - unregister a network namespace subsystem
* @ ops : pernet operations structure to manipulate
*
* Remove the pernet operations structure from the list to be
2008-02-03 18:56:48 +03:00
* used when network namespaces are created or destroyed . In
2007-09-12 13:50:50 +04:00
* addition run the exit method for all existing network
* namespaces .
*/
2010-04-25 11:49:56 +04:00
void unregister_pernet_subsys ( struct pernet_operations * ops )
2007-09-12 13:50:50 +04:00
{
mutex_lock ( & net_mutex ) ;
2010-04-25 11:49:56 +04:00
unregister_pernet_operations ( ops ) ;
2007-09-12 13:50:50 +04:00
mutex_unlock ( & net_mutex ) ;
}
EXPORT_SYMBOL_GPL ( unregister_pernet_subsys ) ;
/**
* register_pernet_device - register a network namespace device
* @ ops : pernet operations structure for the subsystem
*
* Register a device which has init and exit functions
* that are called when network namespaces are created and
* destroyed respectively .
*
* When registered all network namespace init functions are
* called for every existing network namespace . Allowing kernel
* modules to have a race free view of the set of network namespaces .
*
* When a new network namespace is created all of the init
* methods are called in the order in which they were registered .
*
* When a network namespace is destroyed all of the exit methods
* are called in the reverse of the order with which they were
* registered .
*/
int register_pernet_device ( struct pernet_operations * ops )
{
int error ;
mutex_lock ( & net_mutex ) ;
error = register_pernet_operations ( & pernet_list , ops ) ;
if ( ! error & & ( first_device = = & pernet_list ) )
first_device = & ops - > list ;
mutex_unlock ( & net_mutex ) ;
return error ;
}
EXPORT_SYMBOL_GPL ( register_pernet_device ) ;
/**
* unregister_pernet_device - unregister a network namespace netdevice
* @ ops : pernet operations structure to manipulate
*
* Remove the pernet operations structure from the list to be
2008-02-03 18:56:48 +03:00
* used when network namespaces are created or destroyed . In
2007-09-12 13:50:50 +04:00
* addition run the exit method for all existing network
* namespaces .
*/
void unregister_pernet_device ( struct pernet_operations * ops )
{
mutex_lock ( & net_mutex ) ;
if ( & ops - > list = = first_device )
first_device = first_device - > next ;
unregister_pernet_operations ( ops ) ;
mutex_unlock ( & net_mutex ) ;
}
EXPORT_SYMBOL_GPL ( unregister_pernet_device ) ;
2010-03-08 05:14:23 +03:00
# ifdef CONFIG_NET_NS
2014-11-01 07:37:32 +03:00
static struct ns_common * netns_get ( struct task_struct * task )
2010-03-08 05:14:23 +03:00
{
2011-05-05 04:51:50 +04:00
struct net * net = NULL ;
struct nsproxy * nsproxy ;
2014-02-04 07:13:49 +04:00
task_lock ( task ) ;
nsproxy = task - > nsproxy ;
2011-05-05 04:51:50 +04:00
if ( nsproxy )
net = get_net ( nsproxy - > net_ns ) ;
2014-02-04 07:13:49 +04:00
task_unlock ( task ) ;
2011-05-05 04:51:50 +04:00
2014-11-01 07:10:50 +03:00
return net ? & net - > ns : NULL ;
}
static inline struct net * to_net_ns ( struct ns_common * ns )
{
return container_of ( ns , struct net , ns ) ;
2010-03-08 05:14:23 +03:00
}
2014-11-01 07:37:32 +03:00
static void netns_put ( struct ns_common * ns )
2010-03-08 05:14:23 +03:00
{
2014-11-01 07:10:50 +03:00
put_net ( to_net_ns ( ns ) ) ;
2010-03-08 05:14:23 +03:00
}
2014-11-01 07:37:32 +03:00
static int netns_install ( struct nsproxy * nsproxy , struct ns_common * ns )
2010-03-08 05:14:23 +03:00
{
2014-11-01 07:10:50 +03:00
struct net * net = to_net_ns ( ns ) ;
2012-07-26 12:13:20 +04:00
2012-12-14 19:55:36 +04:00
if ( ! ns_capable ( net - > user_ns , CAP_SYS_ADMIN ) | |
2013-03-20 23:49:49 +04:00
! ns_capable ( current_user_ns ( ) , CAP_SYS_ADMIN ) )
2012-07-26 12:13:20 +04:00
return - EPERM ;
2010-03-08 05:14:23 +03:00
put_net ( nsproxy - > net_ns ) ;
2012-07-26 12:13:20 +04:00
nsproxy - > net_ns = get_net ( net ) ;
2010-03-08 05:14:23 +03:00
return 0 ;
}
const struct proc_ns_operations netns_operations = {
. name = " net " ,
. type = CLONE_NEWNET ,
. get = netns_get ,
. put = netns_put ,
. install = netns_install ,
} ;
# endif