2010-02-27 22:41:45 +03:00
/*
* Bridge multicast support .
*
* Copyright ( c ) 2010 Herbert Xu < herbert @ gondor . apana . org . au >
*
* This program is free software ; you can redistribute it and / or modify it
* under the terms of the GNU General Public License as published by the Free
* Software Foundation ; either version 2 of the License , or ( at your option )
* any later version .
*
*/
# include <linux/err.h>
2014-06-07 20:26:28 +04:00
# include <linux/export.h>
2010-02-27 22:41:45 +03:00
# include <linux/if_ether.h>
# include <linux/igmp.h>
# include <linux/jhash.h>
# include <linux/kernel.h>
2010-02-27 22:41:51 +03:00
# include <linux/log2.h>
2010-02-27 22:41:45 +03:00
# include <linux/netdevice.h>
# include <linux/netfilter_bridge.h>
# include <linux/random.h>
# include <linux/rculist.h>
# include <linux/skbuff.h>
# include <linux/slab.h>
# include <linux/timer.h>
2013-05-22 01:52:54 +04:00
# include <linux/inetdevice.h>
2010-02-27 22:41:45 +03:00
# include <net/ip.h>
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
# include <net/ipv6.h>
# include <net/mld.h>
2010-04-27 21:16:54 +04:00
# include <net/ip6_checksum.h>
2013-09-04 04:13:39 +04:00
# include <net/addrconf.h>
2010-04-22 20:54:22 +04:00
# endif
2010-02-27 22:41:45 +03:00
# include "br_private.h"
2013-08-30 19:28:17 +04:00
static void br_multicast_start_querier ( struct net_bridge * br ,
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * query ) ;
2015-06-19 11:22:57 +03:00
static void br_multicast_add_router ( struct net_bridge * br ,
struct net_bridge_port * port ) ;
2015-07-13 15:28:37 +03:00
static void br_ip4_multicast_leave_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
__be32 group ,
__u16 vid ) ;
# if IS_ENABLED(CONFIG_IPV6)
static void br_ip6_multicast_leave_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
const struct in6_addr * group ,
__u16 vid ) ;
# endif
2012-12-10 06:15:35 +04:00
unsigned int br_mdb_rehash_seq ;
2012-04-13 06:37:42 +04:00
2010-04-18 07:42:07 +04:00
static inline int br_ip_equal ( const struct br_ip * a , const struct br_ip * b )
{
if ( a - > proto ! = b - > proto )
return 0 ;
2013-02-13 16:00:17 +04:00
if ( a - > vid ! = b - > vid )
return 0 ;
2010-04-18 07:42:07 +04:00
switch ( a - > proto ) {
case htons ( ETH_P_IP ) :
return a - > u . ip4 = = b - > u . ip4 ;
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
case htons ( ETH_P_IPV6 ) :
return ipv6_addr_equal ( & a - > u . ip6 , & b - > u . ip6 ) ;
# endif
2010-04-18 07:42:07 +04:00
}
return 0 ;
}
2013-02-13 16:00:17 +04:00
static inline int __br_ip4_hash ( struct net_bridge_mdb_htable * mdb , __be32 ip ,
__u16 vid )
2010-02-27 22:41:45 +03:00
{
2013-02-13 16:00:17 +04:00
return jhash_2words ( ( __force u32 ) ip , vid , mdb - > secret ) & ( mdb - > max - 1 ) ;
2010-02-27 22:41:45 +03:00
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static inline int __br_ip6_hash ( struct net_bridge_mdb_htable * mdb ,
2013-02-13 16:00:17 +04:00
const struct in6_addr * ip ,
__u16 vid )
2010-04-22 20:54:22 +04:00
{
2013-02-13 16:00:17 +04:00
return jhash_2words ( ipv6_addr_hash ( ip ) , vid ,
mdb - > secret ) & ( mdb - > max - 1 ) ;
2010-04-22 20:54:22 +04:00
}
# endif
2010-04-18 07:42:07 +04:00
static inline int br_ip_hash ( struct net_bridge_mdb_htable * mdb ,
struct br_ip * ip )
{
switch ( ip - > proto ) {
case htons ( ETH_P_IP ) :
2013-02-13 16:00:17 +04:00
return __br_ip4_hash ( mdb , ip - > u . ip4 , ip - > vid ) ;
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
case htons ( ETH_P_IPV6 ) :
2013-02-13 16:00:17 +04:00
return __br_ip6_hash ( mdb , & ip - > u . ip6 , ip - > vid ) ;
2010-04-22 20:54:22 +04:00
# endif
2010-04-18 07:42:07 +04:00
}
return 0 ;
2010-02-27 22:41:45 +03:00
}
static struct net_bridge_mdb_entry * __br_mdb_ip_get (
2010-04-18 07:42:07 +04:00
struct net_bridge_mdb_htable * mdb , struct br_ip * dst , int hash )
2010-02-27 22:41:45 +03:00
{
struct net_bridge_mdb_entry * mp ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry_rcu ( mp , & mdb - > mhash [ hash ] , hlist [ mdb - > ver ] ) {
2010-04-18 07:42:07 +04:00
if ( br_ip_equal ( & mp - > addr , dst ) )
2010-02-27 22:41:45 +03:00
return mp ;
}
return NULL ;
}
2012-12-12 02:23:08 +04:00
struct net_bridge_mdb_entry * br_mdb_ip_get ( struct net_bridge_mdb_htable * mdb ,
struct br_ip * dst )
2010-07-05 18:50:08 +04:00
{
if ( ! mdb )
return NULL ;
return __br_mdb_ip_get ( mdb , dst , br_ip_hash ( mdb , dst ) ) ;
}
2010-04-18 07:42:07 +04:00
static struct net_bridge_mdb_entry * br_mdb_ip4_get (
2013-02-13 16:00:17 +04:00
struct net_bridge_mdb_htable * mdb , __be32 dst , __u16 vid )
2010-02-27 22:41:45 +03:00
{
2010-04-18 07:42:07 +04:00
struct br_ip br_dst ;
br_dst . u . ip4 = dst ;
br_dst . proto = htons ( ETH_P_IP ) ;
2013-02-13 16:00:17 +04:00
br_dst . vid = vid ;
2010-03-16 06:38:25 +03:00
2010-07-05 18:50:08 +04:00
return br_mdb_ip_get ( mdb , & br_dst ) ;
2010-04-18 07:42:07 +04:00
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static struct net_bridge_mdb_entry * br_mdb_ip6_get (
2013-02-13 16:00:17 +04:00
struct net_bridge_mdb_htable * mdb , const struct in6_addr * dst ,
__u16 vid )
2010-04-22 20:54:22 +04:00
{
struct br_ip br_dst ;
2010-03-16 06:38:25 +03:00
2011-11-21 07:39:03 +04:00
br_dst . u . ip6 = * dst ;
2010-04-22 20:54:22 +04:00
br_dst . proto = htons ( ETH_P_IPV6 ) ;
2013-02-13 16:00:17 +04:00
br_dst . vid = vid ;
2010-04-22 20:54:22 +04:00
2010-07-05 18:50:08 +04:00
return br_mdb_ip_get ( mdb , & br_dst ) ;
2010-04-22 20:54:22 +04:00
}
# endif
2010-02-27 22:41:45 +03:00
struct net_bridge_mdb_entry * br_mdb_get ( struct net_bridge * br ,
2013-03-07 07:05:33 +04:00
struct sk_buff * skb , u16 vid )
2010-02-27 22:41:45 +03:00
{
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * mdb = rcu_dereference ( br - > mdb ) ;
2010-04-18 07:42:07 +04:00
struct br_ip ip ;
2010-07-05 18:50:08 +04:00
if ( br - > multicast_disabled )
2010-02-27 22:41:45 +03:00
return NULL ;
2010-04-18 07:42:07 +04:00
if ( BR_INPUT_SKB_CB ( skb ) - > igmp )
2010-02-27 22:41:45 +03:00
return NULL ;
2010-04-18 07:42:07 +04:00
ip . proto = skb - > protocol ;
2013-03-07 07:05:33 +04:00
ip . vid = vid ;
2010-04-18 07:42:07 +04:00
2010-02-27 22:41:45 +03:00
switch ( skb - > protocol ) {
case htons ( ETH_P_IP ) :
2010-04-18 07:42:07 +04:00
ip . u . ip4 = ip_hdr ( skb ) - > daddr ;
break ;
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
case htons ( ETH_P_IPV6 ) :
2011-11-21 07:39:03 +04:00
ip . u . ip6 = ipv6_hdr ( skb ) - > daddr ;
2010-04-22 20:54:22 +04:00
break ;
# endif
2010-04-18 07:42:07 +04:00
default :
return NULL ;
2010-02-27 22:41:45 +03:00
}
2010-04-18 07:42:07 +04:00
return br_mdb_ip_get ( mdb , & ip ) ;
2010-02-27 22:41:45 +03:00
}
static void br_mdb_free ( struct rcu_head * head )
{
struct net_bridge_mdb_htable * mdb =
container_of ( head , struct net_bridge_mdb_htable , rcu ) ;
struct net_bridge_mdb_htable * old = mdb - > old ;
mdb - > old = NULL ;
kfree ( old - > mhash ) ;
kfree ( old ) ;
}
static int br_mdb_copy ( struct net_bridge_mdb_htable * new ,
struct net_bridge_mdb_htable * old ,
int elasticity )
{
struct net_bridge_mdb_entry * mp ;
int maxlen ;
int len ;
int i ;
for ( i = 0 ; i < old - > max ; i + + )
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry ( mp , & old - > mhash [ i ] , hlist [ old - > ver ] )
2010-02-27 22:41:45 +03:00
hlist_add_head ( & mp - > hlist [ new - > ver ] ,
2010-04-18 07:42:07 +04:00
& new - > mhash [ br_ip_hash ( new , & mp - > addr ) ] ) ;
2010-02-27 22:41:45 +03:00
if ( ! elasticity )
return 0 ;
maxlen = 0 ;
for ( i = 0 ; i < new - > max ; i + + ) {
len = 0 ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry ( mp , & new - > mhash [ i ] , hlist [ new - > ver ] )
2010-02-27 22:41:45 +03:00
len + + ;
if ( len > maxlen )
maxlen = len ;
}
return maxlen > elasticity ? - EINVAL : 0 ;
}
2012-12-12 02:23:08 +04:00
void br_multicast_free_pg ( struct rcu_head * head )
2010-02-27 22:41:45 +03:00
{
struct net_bridge_port_group * p =
container_of ( head , struct net_bridge_port_group , rcu ) ;
kfree ( p ) ;
}
static void br_multicast_free_group ( struct rcu_head * head )
{
struct net_bridge_mdb_entry * mp =
container_of ( head , struct net_bridge_mdb_entry , rcu ) ;
kfree ( mp ) ;
}
static void br_multicast_group_expired ( unsigned long data )
{
struct net_bridge_mdb_entry * mp = ( void * ) data ;
struct net_bridge * br = mp - > br ;
struct net_bridge_mdb_htable * mdb ;
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | | timer_pending ( & mp - > timer ) )
goto out ;
2011-02-12 12:05:42 +03:00
mp - > mglist = false ;
2010-02-27 22:41:45 +03:00
if ( mp - > ports )
goto out ;
2010-11-15 09:38:10 +03:00
mdb = mlock_dereference ( br - > mdb , br ) ;
2010-02-27 22:41:45 +03:00
hlist_del_rcu ( & mp - > hlist [ mdb - > ver ] ) ;
mdb - > size - - ;
call_rcu_bh ( & mp - > rcu , br_multicast_free_group ) ;
out :
spin_unlock ( & br - > multicast_lock ) ;
}
static void br_multicast_del_pg ( struct net_bridge * br ,
struct net_bridge_port_group * pg )
{
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * mdb ;
2010-02-27 22:41:45 +03:00
struct net_bridge_mdb_entry * mp ;
struct net_bridge_port_group * p ;
2010-11-15 09:38:10 +03:00
struct net_bridge_port_group __rcu * * pp ;
mdb = mlock_dereference ( br - > mdb , br ) ;
2010-02-27 22:41:45 +03:00
2010-04-18 07:42:07 +04:00
mp = br_mdb_ip_get ( mdb , & pg - > addr ) ;
2010-02-27 22:41:45 +03:00
if ( WARN_ON ( ! mp ) )
return ;
2010-11-15 09:38:10 +03:00
for ( pp = & mp - > ports ;
( p = mlock_dereference ( * pp , br ) ) ! = NULL ;
pp = & p - > next ) {
2010-02-27 22:41:45 +03:00
if ( p ! = pg )
continue ;
2010-04-27 19:01:06 +04:00
rcu_assign_pointer ( * pp , p - > next ) ;
2010-02-27 22:41:45 +03:00
hlist_del_init ( & p - > mglist ) ;
del_timer ( & p - > timer ) ;
2015-07-15 17:16:50 +03:00
br_mdb_notify ( br - > dev , p - > port , & pg - > addr , RTM_DELMDB ,
p - > state ) ;
2010-02-27 22:41:45 +03:00
call_rcu_bh ( & p - > rcu , br_multicast_free_pg ) ;
2013-10-20 02:58:57 +04:00
if ( ! mp - > ports & & ! mp - > mglist & &
2010-02-27 22:41:45 +03:00
netif_running ( br - > dev ) )
mod_timer ( & mp - > timer , jiffies ) ;
return ;
}
WARN_ON ( 1 ) ;
}
static void br_multicast_port_group_expired ( unsigned long data )
{
struct net_bridge_port_group * pg = ( void * ) data ;
struct net_bridge * br = pg - > port - > br ;
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | | timer_pending ( & pg - > timer ) | |
2012-12-15 02:09:51 +04:00
hlist_unhashed ( & pg - > mglist ) | | pg - > state & MDB_PERMANENT )
2010-02-27 22:41:45 +03:00
goto out ;
br_multicast_del_pg ( br , pg ) ;
out :
spin_unlock ( & br - > multicast_lock ) ;
}
2010-11-15 09:38:10 +03:00
static int br_mdb_rehash ( struct net_bridge_mdb_htable __rcu * * mdbp , int max ,
2010-02-27 22:41:45 +03:00
int elasticity )
{
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * old = rcu_dereference_protected ( * mdbp , 1 ) ;
2010-02-27 22:41:45 +03:00
struct net_bridge_mdb_htable * mdb ;
int err ;
mdb = kmalloc ( sizeof ( * mdb ) , GFP_ATOMIC ) ;
if ( ! mdb )
return - ENOMEM ;
mdb - > max = max ;
mdb - > old = old ;
mdb - > mhash = kzalloc ( max * sizeof ( * mdb - > mhash ) , GFP_ATOMIC ) ;
if ( ! mdb - > mhash ) {
kfree ( mdb ) ;
return - ENOMEM ;
}
mdb - > size = old ? old - > size : 0 ;
mdb - > ver = old ? old - > ver ^ 1 : 0 ;
if ( ! old | | elasticity )
get_random_bytes ( & mdb - > secret , sizeof ( mdb - > secret ) ) ;
else
mdb - > secret = old - > secret ;
if ( ! old )
goto out ;
err = br_mdb_copy ( mdb , old , elasticity ) ;
if ( err ) {
kfree ( mdb - > mhash ) ;
kfree ( mdb ) ;
return err ;
}
2012-12-10 06:15:35 +04:00
br_mdb_rehash_seq + + ;
2010-02-27 22:41:45 +03:00
call_rcu_bh ( & mdb - > rcu , br_mdb_free ) ;
out :
rcu_assign_pointer ( * mdbp , mdb ) ;
return 0 ;
}
2010-04-18 07:42:07 +04:00
static struct sk_buff * br_ip4_multicast_alloc_query ( struct net_bridge * br ,
__be32 group )
2010-02-27 22:41:45 +03:00
{
struct sk_buff * skb ;
struct igmphdr * ih ;
struct ethhdr * eth ;
struct iphdr * iph ;
skb = netdev_alloc_skb_ip_align ( br - > dev , sizeof ( * eth ) + sizeof ( * iph ) +
sizeof ( * ih ) + 4 ) ;
if ( ! skb )
goto out ;
skb - > protocol = htons ( ETH_P_IP ) ;
skb_reset_mac_header ( skb ) ;
eth = eth_hdr ( skb ) ;
2014-02-23 12:05:25 +04:00
ether_addr_copy ( eth - > h_source , br - > dev - > dev_addr ) ;
2010-02-27 22:41:45 +03:00
eth - > h_dest [ 0 ] = 1 ;
eth - > h_dest [ 1 ] = 0 ;
eth - > h_dest [ 2 ] = 0x5e ;
eth - > h_dest [ 3 ] = 0 ;
eth - > h_dest [ 4 ] = 0 ;
eth - > h_dest [ 5 ] = 1 ;
eth - > h_proto = htons ( ETH_P_IP ) ;
skb_put ( skb , sizeof ( * eth ) ) ;
skb_set_network_header ( skb , skb - > len ) ;
iph = ip_hdr ( skb ) ;
iph - > version = 4 ;
iph - > ihl = 6 ;
iph - > tos = 0xc0 ;
iph - > tot_len = htons ( sizeof ( * iph ) + sizeof ( * ih ) + 4 ) ;
iph - > id = 0 ;
iph - > frag_off = htons ( IP_DF ) ;
iph - > ttl = 1 ;
iph - > protocol = IPPROTO_IGMP ;
2013-05-22 01:52:54 +04:00
iph - > saddr = br - > multicast_query_use_ifaddr ?
inet_select_addr ( br - > dev , 0 , RT_SCOPE_LINK ) : 0 ;
2010-02-27 22:41:45 +03:00
iph - > daddr = htonl ( INADDR_ALLHOSTS_GROUP ) ;
( ( u8 * ) & iph [ 1 ] ) [ 0 ] = IPOPT_RA ;
( ( u8 * ) & iph [ 1 ] ) [ 1 ] = 4 ;
( ( u8 * ) & iph [ 1 ] ) [ 2 ] = 0 ;
( ( u8 * ) & iph [ 1 ] ) [ 3 ] = 0 ;
ip_send_check ( iph ) ;
skb_put ( skb , 24 ) ;
skb_set_transport_header ( skb , skb - > len ) ;
ih = igmp_hdr ( skb ) ;
ih - > type = IGMP_HOST_MEMBERSHIP_QUERY ;
ih - > code = ( group ? br - > multicast_last_member_interval :
br - > multicast_query_response_interval ) /
( HZ / IGMP_TIMER_SCALE ) ;
ih - > group = group ;
ih - > csum = 0 ;
ih - > csum = ip_compute_csum ( ( void * ) ih , sizeof ( struct igmphdr ) ) ;
skb_put ( skb , sizeof ( * ih ) ) ;
__skb_pull ( skb , sizeof ( * eth ) ) ;
out :
return skb ;
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static struct sk_buff * br_ip6_multicast_alloc_query ( struct net_bridge * br ,
2011-04-22 08:53:02 +04:00
const struct in6_addr * group )
2010-04-22 20:54:22 +04:00
{
struct sk_buff * skb ;
struct ipv6hdr * ip6h ;
struct mld_msg * mldq ;
struct ethhdr * eth ;
u8 * hopopt ;
unsigned long interval ;
skb = netdev_alloc_skb_ip_align ( br - > dev , sizeof ( * eth ) + sizeof ( * ip6h ) +
8 + sizeof ( * mldq ) ) ;
if ( ! skb )
goto out ;
skb - > protocol = htons ( ETH_P_IPV6 ) ;
/* Ethernet header */
skb_reset_mac_header ( skb ) ;
eth = eth_hdr ( skb ) ;
2014-02-23 12:05:25 +04:00
ether_addr_copy ( eth - > h_source , br - > dev - > dev_addr ) ;
2010-04-22 20:54:22 +04:00
eth - > h_proto = htons ( ETH_P_IPV6 ) ;
skb_put ( skb , sizeof ( * eth ) ) ;
/* IPv6 header + HbH option */
skb_set_network_header ( skb , skb - > len ) ;
ip6h = ipv6_hdr ( skb ) ;
* ( __force __be32 * ) ip6h = htonl ( 0x60000000 ) ;
2010-12-14 11:42:16 +03:00
ip6h - > payload_len = htons ( 8 + sizeof ( * mldq ) ) ;
2010-04-22 20:54:22 +04:00
ip6h - > nexthdr = IPPROTO_HOPOPTS ;
ip6h - > hop_limit = 1 ;
2011-03-22 14:40:32 +03:00
ipv6_addr_set ( & ip6h - > daddr , htonl ( 0xff020000 ) , 0 , 0 , htonl ( 1 ) ) ;
2012-03-05 08:52:44 +04:00
if ( ipv6_dev_get_saddr ( dev_net ( br - > dev ) , br - > dev , & ip6h - > daddr , 0 ,
& ip6h - > saddr ) ) {
kfree_skb ( skb ) ;
return NULL ;
}
2011-02-17 11:17:51 +03:00
ipv6_eth_mc_map ( & ip6h - > daddr , eth - > h_dest ) ;
2010-04-22 20:54:22 +04:00
hopopt = ( u8 * ) ( ip6h + 1 ) ;
hopopt [ 0 ] = IPPROTO_ICMPV6 ; /* next hdr */
hopopt [ 1 ] = 0 ; /* length of HbH */
hopopt [ 2 ] = IPV6_TLV_ROUTERALERT ; /* Router Alert */
hopopt [ 3 ] = 2 ; /* Length of RA Option */
hopopt [ 4 ] = 0 ; /* Type = 0x0000 (MLD) */
hopopt [ 5 ] = 0 ;
2012-05-17 10:00:25 +04:00
hopopt [ 6 ] = IPV6_TLV_PAD1 ; /* Pad1 */
hopopt [ 7 ] = IPV6_TLV_PAD1 ; /* Pad1 */
2010-04-22 20:54:22 +04:00
skb_put ( skb , sizeof ( * ip6h ) + 8 ) ;
/* ICMPv6 */
skb_set_transport_header ( skb , skb - > len ) ;
mldq = ( struct mld_msg * ) icmp6_hdr ( skb ) ;
2013-06-17 01:20:34 +04:00
interval = ipv6_addr_any ( group ) ?
br - > multicast_query_response_interval :
br - > multicast_last_member_interval ;
2010-04-22 20:54:22 +04:00
mldq - > mld_type = ICMPV6_MGM_QUERY ;
mldq - > mld_code = 0 ;
mldq - > mld_cksum = 0 ;
mldq - > mld_maxdelay = htons ( ( u16 ) jiffies_to_msecs ( interval ) ) ;
mldq - > mld_reserved = 0 ;
2011-11-21 07:39:03 +04:00
mldq - > mld_mca = * group ;
2010-04-22 20:54:22 +04:00
/* checksum */
mldq - > mld_cksum = csum_ipv6_magic ( & ip6h - > saddr , & ip6h - > daddr ,
sizeof ( * mldq ) , IPPROTO_ICMPV6 ,
csum_partial ( mldq ,
sizeof ( * mldq ) , 0 ) ) ;
skb_put ( skb , sizeof ( * mldq ) ) ;
__skb_pull ( skb , sizeof ( * eth ) ) ;
out :
return skb ;
}
# endif
2010-04-18 07:42:07 +04:00
static struct sk_buff * br_multicast_alloc_query ( struct net_bridge * br ,
struct br_ip * addr )
{
switch ( addr - > proto ) {
case htons ( ETH_P_IP ) :
return br_ip4_multicast_alloc_query ( br , addr - > u . ip4 ) ;
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
case htons ( ETH_P_IPV6 ) :
return br_ip6_multicast_alloc_query ( br , & addr - > u . ip6 ) ;
# endif
2010-04-18 07:42:07 +04:00
}
return NULL ;
}
2010-02-27 22:41:45 +03:00
static struct net_bridge_mdb_entry * br_multicast_get_group (
2010-04-18 07:42:07 +04:00
struct net_bridge * br , struct net_bridge_port * port ,
struct br_ip * group , int hash )
2010-02-27 22:41:45 +03:00
{
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * mdb ;
2010-02-27 22:41:45 +03:00
struct net_bridge_mdb_entry * mp ;
2012-04-15 09:58:06 +04:00
unsigned int count = 0 ;
unsigned int max ;
2010-02-27 22:41:45 +03:00
int elasticity ;
int err ;
2010-11-15 09:38:10 +03:00
mdb = rcu_dereference_protected ( br - > mdb , 1 ) ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry ( mp , & mdb - > mhash [ hash ] , hlist [ mdb - > ver ] ) {
2010-02-27 22:41:45 +03:00
count + + ;
2010-04-18 07:42:07 +04:00
if ( unlikely ( br_ip_equal ( group , & mp - > addr ) ) )
2010-02-27 22:41:45 +03:00
return mp ;
}
elasticity = 0 ;
max = mdb - > max ;
if ( unlikely ( count > br - > hash_elasticity & & count ) ) {
if ( net_ratelimit ( ) )
2010-05-10 13:31:09 +04:00
br_info ( br , " Multicast hash table "
" chain limit reached: %s \n " ,
port ? port - > dev - > name : br - > dev - > name ) ;
2010-02-27 22:41:45 +03:00
elasticity = br - > hash_elasticity ;
}
if ( mdb - > size > = max ) {
max * = 2 ;
2012-07-11 02:29:19 +04:00
if ( unlikely ( max > br - > hash_max ) ) {
br_warn ( br , " Multicast hash table maximum of %d "
" reached, disabling snooping: %s \n " ,
br - > hash_max ,
port ? port - > dev - > name : br - > dev - > name ) ;
2010-02-27 22:41:45 +03:00
err = - E2BIG ;
disable :
br - > multicast_disabled = 1 ;
goto err ;
}
}
if ( max > mdb - > max | | elasticity ) {
if ( mdb - > old ) {
if ( net_ratelimit ( ) )
2010-05-10 13:31:09 +04:00
br_info ( br , " Multicast hash table "
" on fire: %s \n " ,
port ? port - > dev - > name : br - > dev - > name ) ;
2010-02-27 22:41:45 +03:00
err = - EEXIST ;
goto err ;
}
err = br_mdb_rehash ( & br - > mdb , max , elasticity ) ;
if ( err ) {
2010-05-10 13:31:09 +04:00
br_warn ( br , " Cannot rehash multicast "
" hash table, disabling snooping: %s, %d, %d \n " ,
port ? port - > dev - > name : br - > dev - > name ,
mdb - > size , err ) ;
2010-02-27 22:41:45 +03:00
goto disable ;
}
err = - EAGAIN ;
goto err ;
}
return NULL ;
err :
mp = ERR_PTR ( err ) ;
return mp ;
}
2012-12-12 02:23:08 +04:00
struct net_bridge_mdb_entry * br_multicast_new_group ( struct net_bridge * br ,
struct net_bridge_port * port , struct br_ip * group )
2010-02-27 22:41:45 +03:00
{
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * mdb ;
2010-02-27 22:41:45 +03:00
struct net_bridge_mdb_entry * mp ;
int hash ;
2010-12-10 06:18:04 +03:00
int err ;
2010-02-27 22:41:45 +03:00
2010-11-15 09:38:10 +03:00
mdb = rcu_dereference_protected ( br - > mdb , 1 ) ;
2010-02-27 22:41:45 +03:00
if ( ! mdb ) {
2010-12-10 06:18:04 +03:00
err = br_mdb_rehash ( & br - > mdb , BR_HASH_SIZE , 0 ) ;
if ( err )
return ERR_PTR ( err ) ;
2010-02-27 22:41:45 +03:00
goto rehash ;
}
hash = br_ip_hash ( mdb , group ) ;
mp = br_multicast_get_group ( br , port , group , hash ) ;
switch ( PTR_ERR ( mp ) ) {
case 0 :
break ;
case - EAGAIN :
rehash :
2010-11-15 09:38:10 +03:00
mdb = rcu_dereference_protected ( br - > mdb , 1 ) ;
2010-02-27 22:41:45 +03:00
hash = br_ip_hash ( mdb , group ) ;
break ;
default :
goto out ;
}
mp = kzalloc ( sizeof ( * mp ) , GFP_ATOMIC ) ;
if ( unlikely ( ! mp ) )
2010-12-10 06:18:04 +03:00
return ERR_PTR ( - ENOMEM ) ;
2010-02-27 22:41:45 +03:00
mp - > br = br ;
2010-04-18 07:42:07 +04:00
mp - > addr = * group ;
2013-07-20 07:07:16 +04:00
setup_timer ( & mp - > timer , br_multicast_group_expired ,
( unsigned long ) mp ) ;
2010-02-27 22:41:45 +03:00
hlist_add_head_rcu ( & mp - > hlist [ mdb - > ver ] , & mdb - > mhash [ hash ] ) ;
mdb - > size + + ;
out :
return mp ;
}
2012-12-12 02:23:08 +04:00
struct net_bridge_port_group * br_multicast_new_port_group (
struct net_bridge_port * port ,
struct br_ip * group ,
2012-12-15 02:09:51 +04:00
struct net_bridge_port_group __rcu * next ,
unsigned char state )
2012-12-12 02:23:08 +04:00
{
struct net_bridge_port_group * p ;
p = kzalloc ( sizeof ( * p ) , GFP_ATOMIC ) ;
if ( unlikely ( ! p ) )
return NULL ;
p - > addr = * group ;
p - > port = port ;
2012-12-15 02:09:51 +04:00
p - > state = state ;
2012-12-13 10:51:28 +04:00
rcu_assign_pointer ( p - > next , next ) ;
2012-12-12 02:23:08 +04:00
hlist_add_head ( & p - > mglist , & port - > mglist ) ;
setup_timer ( & p - > timer , br_multicast_port_group_expired ,
( unsigned long ) p ) ;
return p ;
}
2010-02-27 22:41:45 +03:00
static int br_multicast_add_group ( struct net_bridge * br ,
2010-04-18 07:42:07 +04:00
struct net_bridge_port * port ,
struct br_ip * group )
2010-02-27 22:41:45 +03:00
{
struct net_bridge_mdb_entry * mp ;
struct net_bridge_port_group * p ;
2010-11-15 09:38:10 +03:00
struct net_bridge_port_group __rcu * * pp ;
2013-10-20 02:58:57 +04:00
unsigned long now = jiffies ;
2010-02-27 22:41:45 +03:00
int err ;
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | |
( port & & port - > state = = BR_STATE_DISABLED ) )
goto out ;
mp = br_multicast_new_group ( br , port , group ) ;
err = PTR_ERR ( mp ) ;
2010-12-10 06:18:04 +03:00
if ( IS_ERR ( mp ) )
2010-02-27 22:41:45 +03:00
goto err ;
if ( ! port ) {
2011-02-12 12:05:42 +03:00
mp - > mglist = true ;
2013-10-20 02:58:57 +04:00
mod_timer ( & mp - > timer , now + br - > multicast_membership_interval ) ;
2010-02-27 22:41:45 +03:00
goto out ;
}
2010-11-15 09:38:10 +03:00
for ( pp = & mp - > ports ;
( p = mlock_dereference ( * pp , br ) ) ! = NULL ;
pp = & p - > next ) {
2010-02-27 22:41:45 +03:00
if ( p - > port = = port )
2013-10-20 02:58:57 +04:00
goto found ;
2010-02-27 22:41:45 +03:00
if ( ( unsigned long ) p - > port < ( unsigned long ) port )
break ;
}
2012-12-15 02:09:51 +04:00
p = br_multicast_new_port_group ( port , group , * pp , MDB_TEMPORARY ) ;
2010-02-27 22:41:45 +03:00
if ( unlikely ( ! p ) )
goto err ;
rcu_assign_pointer ( * pp , p ) ;
2015-07-09 13:11:10 +03:00
br_mdb_notify ( br - > dev , port , group , RTM_NEWMDB , MDB_TEMPORARY ) ;
2010-02-27 22:41:45 +03:00
2013-10-20 02:58:57 +04:00
found :
mod_timer ( & p - > timer , now + br - > multicast_membership_interval ) ;
2010-02-27 22:41:45 +03:00
out :
err = 0 ;
err :
spin_unlock ( & br - > multicast_lock ) ;
return err ;
}
2010-04-18 07:42:07 +04:00
static int br_ip4_multicast_add_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-02-13 16:00:17 +04:00
__be32 group ,
__u16 vid )
2010-04-18 07:42:07 +04:00
{
struct br_ip br_group ;
if ( ipv4_is_local_multicast ( group ) )
return 0 ;
br_group . u . ip4 = group ;
br_group . proto = htons ( ETH_P_IP ) ;
2013-02-13 16:00:17 +04:00
br_group . vid = vid ;
2010-04-18 07:42:07 +04:00
return br_multicast_add_group ( br , port , & br_group ) ;
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static int br_ip6_multicast_add_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-02-13 16:00:17 +04:00
const struct in6_addr * group ,
__u16 vid )
2010-04-22 20:54:22 +04:00
{
struct br_ip br_group ;
2013-09-04 04:13:39 +04:00
if ( ipv6_addr_is_ll_all_nodes ( group ) )
2010-04-22 20:54:22 +04:00
return 0 ;
2011-11-21 07:39:03 +04:00
br_group . u . ip6 = * group ;
2011-02-15 16:19:17 +03:00
br_group . proto = htons ( ETH_P_IPV6 ) ;
2013-02-13 16:00:17 +04:00
br_group . vid = vid ;
2010-04-22 20:54:22 +04:00
return br_multicast_add_group ( br , port , & br_group ) ;
}
# endif
2010-02-27 22:41:45 +03:00
static void br_multicast_router_expired ( unsigned long data )
{
struct net_bridge_port * port = ( void * ) data ;
struct net_bridge * br = port - > br ;
spin_lock ( & br - > multicast_lock ) ;
if ( port - > multicast_router ! = 1 | |
timer_pending ( & port - > multicast_router_timer ) | |
hlist_unhashed ( & port - > rlist ) )
goto out ;
hlist_del_init_rcu ( & port - > rlist ) ;
out :
spin_unlock ( & br - > multicast_lock ) ;
}
static void br_multicast_local_router_expired ( unsigned long data )
{
}
2013-08-30 19:28:17 +04:00
static void br_multicast_querier_expired ( struct net_bridge * br ,
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * query )
2012-04-13 06:37:42 +04:00
{
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | | br - > multicast_disabled )
goto out ;
2013-08-30 19:28:17 +04:00
br_multicast_start_querier ( br , query ) ;
2012-04-13 06:37:42 +04:00
out :
spin_unlock ( & br - > multicast_lock ) ;
}
2013-08-30 19:28:17 +04:00
static void br_ip4_multicast_querier_expired ( unsigned long data )
{
struct net_bridge * br = ( void * ) data ;
2014-06-07 20:26:26 +04:00
br_multicast_querier_expired ( br , & br - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
}
# if IS_ENABLED(CONFIG_IPV6)
static void br_ip6_multicast_querier_expired ( unsigned long data )
{
struct net_bridge * br = ( void * ) data ;
2014-06-07 20:26:26 +04:00
br_multicast_querier_expired ( br , & br - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
}
# endif
2014-06-07 20:26:27 +04:00
static void br_multicast_select_own_querier ( struct net_bridge * br ,
struct br_ip * ip ,
struct sk_buff * skb )
{
if ( ip - > proto = = htons ( ETH_P_IP ) )
br - > ip4_querier . addr . u . ip4 = ip_hdr ( skb ) - > saddr ;
# if IS_ENABLED(CONFIG_IPV6)
else
br - > ip6_querier . addr . u . ip6 = ipv6_hdr ( skb ) - > saddr ;
# endif
}
2010-04-18 07:42:07 +04:00
static void __br_multicast_send_query ( struct net_bridge * br ,
struct net_bridge_port * port ,
struct br_ip * ip )
2010-02-27 22:41:45 +03:00
{
struct sk_buff * skb ;
2010-04-18 07:42:07 +04:00
skb = br_multicast_alloc_query ( br , ip ) ;
2010-02-27 22:41:45 +03:00
if ( ! skb )
2010-04-18 07:42:07 +04:00
return ;
2010-02-27 22:41:45 +03:00
if ( port ) {
skb - > dev = port - > dev ;
2015-04-06 05:19:04 +03:00
NF_HOOK ( NFPROTO_BRIDGE , NF_BR_LOCAL_OUT , NULL , skb ,
NULL , skb - > dev ,
2014-11-17 14:20:28 +03:00
br_dev_queue_push_xmit ) ;
2014-06-07 20:26:27 +04:00
} else {
br_multicast_select_own_querier ( br , ip , skb ) ;
2010-02-27 22:41:45 +03:00
netif_rx ( skb ) ;
2014-06-07 20:26:27 +04:00
}
2010-04-18 07:42:07 +04:00
}
static void br_multicast_send_query ( struct net_bridge * br ,
2013-08-30 19:28:17 +04:00
struct net_bridge_port * port ,
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * own_query )
2010-04-18 07:42:07 +04:00
{
unsigned long time ;
struct br_ip br_group ;
2014-06-07 20:26:26 +04:00
struct bridge_mcast_other_query * other_query = NULL ;
2010-04-18 07:42:07 +04:00
if ( ! netif_running ( br - > dev ) | | br - > multicast_disabled | |
2013-08-30 19:28:17 +04:00
! br - > multicast_querier )
2010-04-18 07:42:07 +04:00
return ;
2010-04-22 20:54:22 +04:00
memset ( & br_group . u , 0 , sizeof ( br_group . u ) ) ;
2014-06-07 20:26:26 +04:00
if ( port ? ( own_query = = & port - > ip4_own_query ) :
( own_query = = & br - > ip4_own_query ) ) {
other_query = & br - > ip4_other_query ;
2013-08-30 19:28:17 +04:00
br_group . proto = htons ( ETH_P_IP ) ;
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2013-08-30 19:28:17 +04:00
} else {
2014-06-07 20:26:26 +04:00
other_query = & br - > ip6_other_query ;
2013-08-30 19:28:17 +04:00
br_group . proto = htons ( ETH_P_IPV6 ) ;
2010-04-22 20:54:22 +04:00
# endif
2013-08-30 19:28:17 +04:00
}
2014-06-07 20:26:26 +04:00
if ( ! other_query | | timer_pending ( & other_query - > timer ) )
2013-08-30 19:28:17 +04:00
return ;
__br_multicast_send_query ( br , port , & br_group ) ;
2010-02-27 22:41:45 +03:00
time = jiffies ;
2014-06-07 20:26:26 +04:00
time + = own_query - > startup_sent < br - > multicast_startup_query_count ?
2010-02-27 22:41:45 +03:00
br - > multicast_startup_query_interval :
br - > multicast_query_interval ;
2014-06-07 20:26:26 +04:00
mod_timer ( & own_query - > timer , time ) ;
2010-02-27 22:41:45 +03:00
}
2014-06-07 20:26:26 +04:00
static void
br_multicast_port_query_expired ( struct net_bridge_port * port ,
struct bridge_mcast_own_query * query )
2010-02-27 22:41:45 +03:00
{
struct net_bridge * br = port - > br ;
spin_lock ( & br - > multicast_lock ) ;
2010-03-06 04:14:09 +03:00
if ( port - > state = = BR_STATE_DISABLED | |
port - > state = = BR_STATE_BLOCKING )
2010-02-27 22:41:45 +03:00
goto out ;
2013-08-30 19:28:17 +04:00
if ( query - > startup_sent < br - > multicast_startup_query_count )
query - > startup_sent + + ;
2010-02-27 22:41:45 +03:00
2013-08-30 19:28:17 +04:00
br_multicast_send_query ( port - > br , port , query ) ;
2010-02-27 22:41:45 +03:00
out :
spin_unlock ( & br - > multicast_lock ) ;
}
2013-08-30 19:28:17 +04:00
static void br_ip4_multicast_port_query_expired ( unsigned long data )
{
struct net_bridge_port * port = ( void * ) data ;
2014-06-07 20:26:26 +04:00
br_multicast_port_query_expired ( port , & port - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
}
# if IS_ENABLED(CONFIG_IPV6)
static void br_ip6_multicast_port_query_expired ( unsigned long data )
{
struct net_bridge_port * port = ( void * ) data ;
2014-06-07 20:26:26 +04:00
br_multicast_port_query_expired ( port , & port - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
}
# endif
2010-02-27 22:41:45 +03:00
void br_multicast_add_port ( struct net_bridge_port * port )
{
port - > multicast_router = 1 ;
setup_timer ( & port - > multicast_router_timer , br_multicast_router_expired ,
( unsigned long ) port ) ;
2014-06-07 20:26:26 +04:00
setup_timer ( & port - > ip4_own_query . timer ,
br_ip4_multicast_port_query_expired , ( unsigned long ) port ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
setup_timer ( & port - > ip6_own_query . timer ,
br_ip6_multicast_port_query_expired , ( unsigned long ) port ) ;
2013-08-30 19:28:17 +04:00
# endif
2010-02-27 22:41:45 +03:00
}
void br_multicast_del_port ( struct net_bridge_port * port )
{
2015-07-15 17:16:51 +03:00
struct net_bridge * br = port - > br ;
struct net_bridge_port_group * pg ;
struct hlist_node * n ;
/* Take care of the remaining groups, only perm ones should be left */
spin_lock_bh ( & br - > multicast_lock ) ;
hlist_for_each_entry_safe ( pg , n , & port - > mglist , mglist )
br_multicast_del_pg ( br , pg ) ;
spin_unlock_bh ( & br - > multicast_lock ) ;
2010-02-27 22:41:45 +03:00
del_timer_sync ( & port - > multicast_router_timer ) ;
}
2014-06-07 20:26:26 +04:00
static void br_multicast_enable ( struct bridge_mcast_own_query * query )
2010-02-27 22:41:50 +03:00
{
2013-08-30 19:28:17 +04:00
query - > startup_sent = 0 ;
2010-02-27 22:41:50 +03:00
2013-08-30 19:28:17 +04:00
if ( try_to_del_timer_sync ( & query - > timer ) > = 0 | |
del_timer ( & query - > timer ) )
mod_timer ( & query - > timer , jiffies ) ;
2010-02-27 22:41:50 +03:00
}
2010-02-27 22:41:45 +03:00
void br_multicast_enable_port ( struct net_bridge_port * port )
{
struct net_bridge * br = port - > br ;
spin_lock ( & br - > multicast_lock ) ;
if ( br - > multicast_disabled | | ! netif_running ( br - > dev ) )
goto out ;
2014-06-07 20:26:26 +04:00
br_multicast_enable ( & port - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
br_multicast_enable ( & port - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
# endif
2015-06-19 11:22:57 +03:00
if ( port - > multicast_router = = 2 & & hlist_unhashed ( & port - > rlist ) )
br_multicast_add_router ( br , port ) ;
2010-02-27 22:41:45 +03:00
out :
spin_unlock ( & br - > multicast_lock ) ;
}
void br_multicast_disable_port ( struct net_bridge_port * port )
{
struct net_bridge * br = port - > br ;
struct net_bridge_port_group * pg ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
struct hlist_node * n ;
2010-02-27 22:41:45 +03:00
spin_lock ( & br - > multicast_lock ) ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry_safe ( pg , n , & port - > mglist , mglist )
2015-07-15 17:16:51 +03:00
if ( pg - > state = = MDB_TEMPORARY )
br_multicast_del_pg ( br , pg ) ;
2010-02-27 22:41:45 +03:00
if ( ! hlist_unhashed ( & port - > rlist ) )
hlist_del_init_rcu ( & port - > rlist ) ;
del_timer ( & port - > multicast_router_timer ) ;
2014-06-07 20:26:26 +04:00
del_timer ( & port - > ip4_own_query . timer ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
del_timer ( & port - > ip6_own_query . timer ) ;
2013-08-30 19:28:17 +04:00
# endif
2010-02-27 22:41:45 +03:00
spin_unlock ( & br - > multicast_lock ) ;
}
2010-04-18 07:42:07 +04:00
static int br_ip4_multicast_igmp3_report ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb ,
u16 vid )
2010-02-27 22:41:45 +03:00
{
struct igmpv3_report * ih ;
struct igmpv3_grec * grec ;
int i ;
int len ;
int num ;
int type ;
int err = 0 ;
__be32 group ;
ih = igmpv3_report_hdr ( skb ) ;
num = ntohs ( ih - > ngrec ) ;
len = sizeof ( * ih ) ;
for ( i = 0 ; i < num ; i + + ) {
len + = sizeof ( * grec ) ;
if ( ! pskb_may_pull ( skb , len ) )
return - EINVAL ;
2010-04-08 08:20:47 +04:00
grec = ( void * ) ( skb - > data + len - sizeof ( * grec ) ) ;
2010-02-27 22:41:45 +03:00
group = grec - > grec_mca ;
type = grec - > grec_type ;
2010-04-20 07:20:05 +04:00
len + = ntohs ( grec - > grec_nsrcs ) * 4 ;
2010-02-27 22:41:45 +03:00
if ( ! pskb_may_pull ( skb , len ) )
return - EINVAL ;
/* We treat this as an IGMPv2 report for now. */
switch ( type ) {
case IGMPV3_MODE_IS_INCLUDE :
case IGMPV3_MODE_IS_EXCLUDE :
case IGMPV3_CHANGE_TO_INCLUDE :
case IGMPV3_CHANGE_TO_EXCLUDE :
case IGMPV3_ALLOW_NEW_SOURCES :
case IGMPV3_BLOCK_OLD_SOURCES :
break ;
default :
continue ;
}
2015-07-13 15:28:37 +03:00
if ( ( type = = IGMPV3_CHANGE_TO_INCLUDE | |
type = = IGMPV3_MODE_IS_INCLUDE ) & &
ntohs ( grec - > grec_nsrcs ) = = 0 ) {
br_ip4_multicast_leave_group ( br , port , group , vid ) ;
} else {
err = br_ip4_multicast_add_group ( br , port , group , vid ) ;
if ( err )
break ;
}
2010-02-27 22:41:45 +03:00
}
return err ;
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static int br_ip6_multicast_mld2_report ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb ,
u16 vid )
2010-04-22 20:54:22 +04:00
{
struct icmp6hdr * icmp6h ;
struct mld2_grec * grec ;
int i ;
int len ;
int num ;
int err = 0 ;
if ( ! pskb_may_pull ( skb , sizeof ( * icmp6h ) ) )
return - EINVAL ;
icmp6h = icmp6_hdr ( skb ) ;
num = ntohs ( icmp6h - > icmp6_dataun . un_data16 [ 1 ] ) ;
len = sizeof ( * icmp6h ) ;
for ( i = 0 ; i < num ; i + + ) {
__be16 * nsrcs , _nsrcs ;
nsrcs = skb_header_pointer ( skb ,
len + offsetof ( struct mld2_grec ,
2011-02-15 16:19:18 +03:00
grec_nsrcs ) ,
2010-04-22 20:54:22 +04:00
sizeof ( _nsrcs ) , & _nsrcs ) ;
if ( ! nsrcs )
return - EINVAL ;
if ( ! pskb_may_pull ( skb ,
len + sizeof ( * grec ) +
2011-02-15 16:19:19 +03:00
sizeof ( struct in6_addr ) * ntohs ( * nsrcs ) ) )
2010-04-22 20:54:22 +04:00
return - EINVAL ;
grec = ( struct mld2_grec * ) ( skb - > data + len ) ;
2011-02-15 16:19:19 +03:00
len + = sizeof ( * grec ) +
sizeof ( struct in6_addr ) * ntohs ( * nsrcs ) ;
2010-04-22 20:54:22 +04:00
/* We treat these as MLDv1 reports for now. */
switch ( grec - > grec_type ) {
case MLD2_MODE_IS_INCLUDE :
case MLD2_MODE_IS_EXCLUDE :
case MLD2_CHANGE_TO_INCLUDE :
case MLD2_CHANGE_TO_EXCLUDE :
case MLD2_ALLOW_NEW_SOURCES :
case MLD2_BLOCK_OLD_SOURCES :
break ;
default :
continue ;
}
2015-07-13 15:28:37 +03:00
if ( ( grec - > grec_type = = MLD2_CHANGE_TO_INCLUDE | |
grec - > grec_type = = MLD2_MODE_IS_INCLUDE ) & &
ntohs ( * nsrcs ) = = 0 ) {
br_ip6_multicast_leave_group ( br , port , & grec - > grec_mca ,
vid ) ;
} else {
err = br_ip6_multicast_add_group ( br , port ,
& grec - > grec_mca , vid ) ;
if ( ! err )
break ;
}
2010-04-22 20:54:22 +04:00
}
return err ;
}
# endif
2014-06-07 20:26:27 +04:00
static bool br_ip4_multicast_select_querier ( struct net_bridge * br ,
2014-06-07 20:26:29 +04:00
struct net_bridge_port * port ,
2014-06-07 20:26:27 +04:00
__be32 saddr )
{
if ( ! timer_pending ( & br - > ip4_own_query . timer ) & &
! timer_pending ( & br - > ip4_other_query . timer ) )
goto update ;
if ( ! br - > ip4_querier . addr . u . ip4 )
goto update ;
if ( ntohl ( saddr ) < = ntohl ( br - > ip4_querier . addr . u . ip4 ) )
goto update ;
return false ;
update :
br - > ip4_querier . addr . u . ip4 = saddr ;
2014-06-07 20:26:29 +04:00
/* update protected by general multicast_lock by caller */
rcu_assign_pointer ( br - > ip4_querier . port , port ) ;
2014-06-07 20:26:27 +04:00
return true ;
}
# if IS_ENABLED(CONFIG_IPV6)
static bool br_ip6_multicast_select_querier ( struct net_bridge * br ,
2014-06-07 20:26:29 +04:00
struct net_bridge_port * port ,
2014-06-07 20:26:27 +04:00
struct in6_addr * saddr )
{
if ( ! timer_pending ( & br - > ip6_own_query . timer ) & &
! timer_pending ( & br - > ip6_other_query . timer ) )
goto update ;
if ( ipv6_addr_cmp ( saddr , & br - > ip6_querier . addr . u . ip6 ) < = 0 )
goto update ;
return false ;
update :
br - > ip6_querier . addr . u . ip6 = * saddr ;
2014-06-07 20:26:29 +04:00
/* update protected by general multicast_lock by caller */
rcu_assign_pointer ( br - > ip6_querier . port , port ) ;
2014-06-07 20:26:27 +04:00
return true ;
}
# endif
static bool br_multicast_select_querier ( struct net_bridge * br ,
2014-06-07 20:26:29 +04:00
struct net_bridge_port * port ,
2014-06-07 20:26:27 +04:00
struct br_ip * saddr )
{
switch ( saddr - > proto ) {
case htons ( ETH_P_IP ) :
2014-06-07 20:26:29 +04:00
return br_ip4_multicast_select_querier ( br , port , saddr - > u . ip4 ) ;
2014-06-07 20:26:27 +04:00
# if IS_ENABLED(CONFIG_IPV6)
case htons ( ETH_P_IPV6 ) :
2014-06-07 20:26:29 +04:00
return br_ip6_multicast_select_querier ( br , port , & saddr - > u . ip6 ) ;
2014-06-07 20:26:27 +04:00
# endif
}
return false ;
}
2013-08-30 19:28:17 +04:00
static void
2014-06-07 20:26:26 +04:00
br_multicast_update_query_timer ( struct net_bridge * br ,
struct bridge_mcast_other_query * query ,
unsigned long max_delay )
2013-08-01 03:06:20 +04:00
{
2014-06-07 20:26:26 +04:00
if ( ! timer_pending ( & query - > timer ) )
query - > delay_time = jiffies + max_delay ;
2013-08-01 03:06:20 +04:00
2014-06-07 20:26:26 +04:00
mod_timer ( & query - > timer , jiffies + br - > multicast_querier_interval ) ;
2013-08-01 03:06:20 +04:00
}
2010-04-27 19:01:04 +04:00
/*
2013-06-21 11:37:25 +04:00
* Add port to router_list
2010-04-27 19:01:04 +04:00
* list is maintained ordered by pointer value
* and locked by br - > multicast_lock and RCU
*/
2010-02-27 22:41:49 +03:00
static void br_multicast_add_router ( struct net_bridge * br ,
struct net_bridge_port * port )
{
2010-04-27 11:13:11 +04:00
struct net_bridge_port * p ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
struct hlist_node * slot = NULL ;
2010-04-27 11:13:11 +04:00
2015-06-09 20:23:57 +03:00
if ( ! hlist_unhashed ( & port - > rlist ) )
return ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry ( p , & br - > router_list , rlist ) {
2010-04-27 19:01:04 +04:00
if ( ( unsigned long ) port > = ( unsigned long ) p )
break ;
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
slot = & p - > rlist ;
2010-04-27 11:13:11 +04:00
}
2010-04-27 19:01:04 +04:00
if ( slot )
2014-08-07 03:09:16 +04:00
hlist_add_behind_rcu ( & port - > rlist , slot ) ;
2010-04-27 11:13:11 +04:00
else
hlist_add_head_rcu ( & port - > rlist , & br - > router_list ) ;
2010-02-27 22:41:49 +03:00
}
2010-02-27 22:41:45 +03:00
static void br_multicast_mark_router ( struct net_bridge * br ,
struct net_bridge_port * port )
{
unsigned long now = jiffies ;
if ( ! port ) {
if ( br - > multicast_router = = 1 )
mod_timer ( & br - > multicast_router_timer ,
now + br - > multicast_querier_interval ) ;
return ;
}
if ( port - > multicast_router ! = 1 )
return ;
2010-02-27 22:41:49 +03:00
br_multicast_add_router ( br , port ) ;
2010-02-27 22:41:45 +03:00
mod_timer ( & port - > multicast_router_timer ,
now + br - > multicast_querier_interval ) ;
}
static void br_multicast_query_received ( struct net_bridge * br ,
struct net_bridge_port * port ,
2014-06-07 20:26:26 +04:00
struct bridge_mcast_other_query * query ,
2014-06-07 20:26:27 +04:00
struct br_ip * saddr ,
2013-08-01 03:06:20 +04:00
unsigned long max_delay )
2010-02-27 22:41:45 +03:00
{
2014-06-07 20:26:29 +04:00
if ( ! br_multicast_select_querier ( br , port , saddr ) )
2010-02-27 22:41:45 +03:00
return ;
2014-06-07 20:26:27 +04:00
br_multicast_update_query_timer ( br , query , max_delay ) ;
2010-02-27 22:41:45 +03:00
br_multicast_mark_router ( br , port ) ;
}
2010-04-18 07:42:07 +04:00
static int br_ip4_multicast_query ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb ,
u16 vid )
2010-02-27 22:41:45 +03:00
{
2011-04-22 08:53:02 +04:00
const struct iphdr * iph = ip_hdr ( skb ) ;
2010-02-27 22:41:45 +03:00
struct igmphdr * ih = igmp_hdr ( skb ) ;
struct net_bridge_mdb_entry * mp ;
struct igmpv3_query * ih3 ;
struct net_bridge_port_group * p ;
2010-11-15 09:38:10 +03:00
struct net_bridge_port_group __rcu * * pp ;
2014-06-07 20:26:27 +04:00
struct br_ip saddr ;
2010-02-27 22:41:45 +03:00
unsigned long max_delay ;
unsigned long now = jiffies ;
__be32 group ;
2010-03-13 23:27:21 +03:00
int err = 0 ;
2010-02-27 22:41:45 +03:00
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | |
( port & & port - > state = = BR_STATE_DISABLED ) )
goto out ;
group = ih - > group ;
if ( skb - > len = = sizeof ( * ih ) ) {
max_delay = ih - > code * ( HZ / IGMP_TIMER_SCALE ) ;
if ( ! max_delay ) {
max_delay = 10 * HZ ;
group = 0 ;
}
2015-05-02 15:01:07 +03:00
} else if ( skb - > len > = sizeof ( * ih3 ) ) {
2010-02-27 22:41:45 +03:00
ih3 = igmpv3_query_hdr ( skb ) ;
if ( ih3 - > nsrcs )
2010-03-13 23:27:21 +03:00
goto out ;
2010-02-27 22:41:45 +03:00
2010-03-15 22:27:00 +03:00
max_delay = ih3 - > code ?
IGMPV3_MRC ( ih3 - > code ) * ( HZ / IGMP_TIMER_SCALE ) : 1 ;
2015-05-02 15:01:07 +03:00
} else {
2014-03-11 01:25:24 +04:00
goto out ;
}
2014-06-07 20:26:27 +04:00
if ( ! group ) {
saddr . proto = htons ( ETH_P_IP ) ;
saddr . u . ip4 = iph - > saddr ;
2013-08-01 03:06:20 +04:00
2014-06-07 20:26:27 +04:00
br_multicast_query_received ( br , port , & br - > ip4_other_query ,
& saddr , max_delay ) ;
2010-02-27 22:41:45 +03:00
goto out ;
2014-06-07 20:26:27 +04:00
}
2010-02-27 22:41:45 +03:00
2013-02-13 16:00:17 +04:00
mp = br_mdb_ip4_get ( mlock_dereference ( br - > mdb , br ) , group , vid ) ;
2010-02-27 22:41:45 +03:00
if ( ! mp )
goto out ;
max_delay * = br - > multicast_last_member_count ;
2011-02-12 12:05:42 +03:00
if ( mp - > mglist & &
2010-02-27 22:41:45 +03:00
( timer_pending ( & mp - > timer ) ?
time_after ( mp - > timer . expires , now + max_delay ) :
try_to_del_timer_sync ( & mp - > timer ) > = 0 ) )
mod_timer ( & mp - > timer , now + max_delay ) ;
2010-11-15 09:38:10 +03:00
for ( pp = & mp - > ports ;
( p = mlock_dereference ( * pp , br ) ) ! = NULL ;
pp = & p - > next ) {
2010-02-27 22:41:45 +03:00
if ( timer_pending ( & p - > timer ) ?
time_after ( p - > timer . expires , now + max_delay ) :
try_to_del_timer_sync ( & p - > timer ) > = 0 )
2011-02-11 15:42:07 +03:00
mod_timer ( & p - > timer , now + max_delay ) ;
2010-02-27 22:41:45 +03:00
}
out :
spin_unlock ( & br - > multicast_lock ) ;
2010-03-13 23:27:21 +03:00
return err ;
2010-02-27 22:41:45 +03:00
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static int br_ip6_multicast_query ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb ,
u16 vid )
2010-04-22 20:54:22 +04:00
{
2011-04-22 08:53:02 +04:00
const struct ipv6hdr * ip6h = ipv6_hdr ( skb ) ;
2012-12-13 10:51:28 +04:00
struct mld_msg * mld ;
2010-04-22 20:54:22 +04:00
struct net_bridge_mdb_entry * mp ;
struct mld2_query * mld2q ;
2010-11-15 09:38:10 +03:00
struct net_bridge_port_group * p ;
struct net_bridge_port_group __rcu * * pp ;
2014-06-07 20:26:27 +04:00
struct br_ip saddr ;
2010-04-22 20:54:22 +04:00
unsigned long max_delay ;
unsigned long now = jiffies ;
2011-04-22 08:53:02 +04:00
const struct in6_addr * group = NULL ;
2014-03-11 01:25:24 +04:00
bool is_general_query ;
2010-04-22 20:54:22 +04:00
int err = 0 ;
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | |
( port & & port - > state = = BR_STATE_DISABLED ) )
goto out ;
if ( skb - > len = = sizeof ( * mld ) ) {
if ( ! pskb_may_pull ( skb , sizeof ( * mld ) ) ) {
err = - EINVAL ;
goto out ;
}
mld = ( struct mld_msg * ) icmp6_hdr ( skb ) ;
2012-07-10 03:56:12 +04:00
max_delay = msecs_to_jiffies ( ntohs ( mld - > mld_maxdelay ) ) ;
2010-04-22 20:54:22 +04:00
if ( max_delay )
group = & mld - > mld_mca ;
2013-08-06 02:32:05 +04:00
} else {
2010-04-22 20:54:22 +04:00
if ( ! pskb_may_pull ( skb , sizeof ( * mld2q ) ) ) {
err = - EINVAL ;
goto out ;
}
mld2q = ( struct mld2_query * ) icmp6_hdr ( skb ) ;
if ( ! mld2q - > mld2q_nsrcs )
group = & mld2q - > mld2q_mca ;
2013-09-04 02:19:39 +04:00
max_delay = max ( msecs_to_jiffies ( mldv2_mrc ( mld2q ) ) , 1UL ) ;
2010-04-22 20:54:22 +04:00
}
2014-03-11 01:25:24 +04:00
is_general_query = group & & ipv6_addr_any ( group ) ;
2014-06-07 20:26:27 +04:00
if ( is_general_query ) {
saddr . proto = htons ( ETH_P_IPV6 ) ;
saddr . u . ip6 = ip6h - > saddr ;
2013-08-01 03:06:20 +04:00
2014-06-07 20:26:27 +04:00
br_multicast_query_received ( br , port , & br - > ip6_other_query ,
& saddr , max_delay ) ;
2010-04-22 20:54:22 +04:00
goto out ;
2014-06-12 03:41:23 +04:00
} else if ( ! group ) {
goto out ;
2014-06-07 20:26:27 +04:00
}
2010-04-22 20:54:22 +04:00
2013-02-13 16:00:17 +04:00
mp = br_mdb_ip6_get ( mlock_dereference ( br - > mdb , br ) , group , vid ) ;
2010-04-22 20:54:22 +04:00
if ( ! mp )
goto out ;
max_delay * = br - > multicast_last_member_count ;
2011-02-12 12:05:42 +03:00
if ( mp - > mglist & &
2010-04-22 20:54:22 +04:00
( timer_pending ( & mp - > timer ) ?
time_after ( mp - > timer . expires , now + max_delay ) :
try_to_del_timer_sync ( & mp - > timer ) > = 0 ) )
mod_timer ( & mp - > timer , now + max_delay ) ;
2010-11-15 09:38:10 +03:00
for ( pp = & mp - > ports ;
( p = mlock_dereference ( * pp , br ) ) ! = NULL ;
pp = & p - > next ) {
2010-04-22 20:54:22 +04:00
if ( timer_pending ( & p - > timer ) ?
time_after ( p - > timer . expires , now + max_delay ) :
try_to_del_timer_sync ( & p - > timer ) > = 0 )
2011-02-11 15:42:07 +03:00
mod_timer ( & p - > timer , now + max_delay ) ;
2010-04-22 20:54:22 +04:00
}
out :
spin_unlock ( & br - > multicast_lock ) ;
return err ;
}
# endif
2014-06-07 20:26:26 +04:00
static void
br_multicast_leave_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
struct br_ip * group ,
struct bridge_mcast_other_query * other_query ,
struct bridge_mcast_own_query * own_query )
2010-02-27 22:41:45 +03:00
{
struct net_bridge_mdb_htable * mdb ;
struct net_bridge_mdb_entry * mp ;
struct net_bridge_port_group * p ;
unsigned long now ;
unsigned long time ;
spin_lock ( & br - > multicast_lock ) ;
if ( ! netif_running ( br - > dev ) | |
( port & & port - > state = = BR_STATE_DISABLED ) | |
2014-06-07 20:26:26 +04:00
timer_pending ( & other_query - > timer ) )
2010-02-27 22:41:45 +03:00
goto out ;
2010-11-15 09:38:10 +03:00
mdb = mlock_dereference ( br - > mdb , br ) ;
2010-02-27 22:41:45 +03:00
mp = br_mdb_ip_get ( mdb , group ) ;
if ( ! mp )
goto out ;
2013-08-30 19:28:17 +04:00
if ( br - > multicast_querier ) {
2013-05-22 01:52:56 +04:00
__br_multicast_send_query ( br , port , & mp - > addr ) ;
time = jiffies + br - > multicast_last_member_count *
br - > multicast_last_member_interval ;
2013-08-30 19:28:17 +04:00
2014-06-07 20:26:26 +04:00
mod_timer ( & own_query - > timer , time ) ;
2013-05-22 01:52:56 +04:00
for ( p = mlock_dereference ( mp - > ports , br ) ;
p ! = NULL ;
p = mlock_dereference ( p - > next , br ) ) {
if ( p - > port ! = port )
continue ;
if ( ! hlist_unhashed ( & p - > mglist ) & &
( timer_pending ( & p - > timer ) ?
time_after ( p - > timer . expires , time ) :
try_to_del_timer_sync ( & p - > timer ) > = 0 ) ) {
mod_timer ( & p - > timer , time ) ;
}
break ;
}
}
2012-12-06 01:24:45 +04:00
if ( port & & ( port - > flags & BR_MULTICAST_FAST_LEAVE ) ) {
2012-12-04 03:56:40 +04:00
struct net_bridge_port_group __rcu * * pp ;
for ( pp = & mp - > ports ;
( p = mlock_dereference ( * pp , br ) ) ! = NULL ;
pp = & p - > next ) {
if ( p - > port ! = port )
continue ;
rcu_assign_pointer ( * pp , p - > next ) ;
hlist_del_init ( & p - > mglist ) ;
del_timer ( & p - > timer ) ;
2015-07-09 13:11:10 +03:00
br_mdb_notify ( br - > dev , port , group , RTM_DELMDB ,
p - > state ) ;
2012-12-04 03:56:40 +04:00
call_rcu_bh ( & p - > rcu , br_multicast_free_pg ) ;
2013-10-20 02:58:57 +04:00
if ( ! mp - > ports & & ! mp - > mglist & &
2012-12-04 03:56:40 +04:00
netif_running ( br - > dev ) )
mod_timer ( & mp - > timer , jiffies ) ;
}
goto out ;
}
2010-02-27 22:41:45 +03:00
now = jiffies ;
time = now + br - > multicast_last_member_count *
br - > multicast_last_member_interval ;
if ( ! port ) {
2013-10-20 02:58:57 +04:00
if ( mp - > mglist & &
2010-02-27 22:41:45 +03:00
( timer_pending ( & mp - > timer ) ?
time_after ( mp - > timer . expires , time ) :
try_to_del_timer_sync ( & mp - > timer ) > = 0 ) ) {
mod_timer ( & mp - > timer , time ) ;
}
2013-10-20 02:58:57 +04:00
goto out ;
}
for ( p = mlock_dereference ( mp - > ports , br ) ;
p ! = NULL ;
p = mlock_dereference ( p - > next , br ) ) {
if ( p - > port ! = port )
continue ;
if ( ! hlist_unhashed ( & p - > mglist ) & &
( timer_pending ( & p - > timer ) ?
time_after ( p - > timer . expires , time ) :
try_to_del_timer_sync ( & p - > timer ) > = 0 ) ) {
mod_timer ( & p - > timer , time ) ;
}
break ;
2010-02-27 22:41:45 +03:00
}
out :
spin_unlock ( & br - > multicast_lock ) ;
}
2010-04-18 07:42:07 +04:00
static void br_ip4_multicast_leave_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-02-13 16:00:17 +04:00
__be32 group ,
__u16 vid )
2010-04-18 07:42:07 +04:00
{
struct br_ip br_group ;
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * own_query ;
2010-04-18 07:42:07 +04:00
if ( ipv4_is_local_multicast ( group ) )
return ;
2014-06-07 20:26:26 +04:00
own_query = port ? & port - > ip4_own_query : & br - > ip4_own_query ;
2010-04-18 07:42:07 +04:00
br_group . u . ip4 = group ;
br_group . proto = htons ( ETH_P_IP ) ;
2013-02-13 16:00:17 +04:00
br_group . vid = vid ;
2010-04-18 07:42:07 +04:00
2014-06-07 20:26:26 +04:00
br_multicast_leave_group ( br , port , & br_group , & br - > ip4_other_query ,
own_query ) ;
2010-04-18 07:42:07 +04:00
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static void br_ip6_multicast_leave_group ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-02-13 16:00:17 +04:00
const struct in6_addr * group ,
__u16 vid )
2010-04-22 20:54:22 +04:00
{
struct br_ip br_group ;
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * own_query ;
2010-04-22 20:54:22 +04:00
2013-09-04 04:13:39 +04:00
if ( ipv6_addr_is_ll_all_nodes ( group ) )
2010-04-22 20:54:22 +04:00
return ;
2014-06-07 20:26:26 +04:00
own_query = port ? & port - > ip6_own_query : & br - > ip6_own_query ;
2011-11-21 07:39:03 +04:00
br_group . u . ip6 = * group ;
2010-04-22 20:54:22 +04:00
br_group . proto = htons ( ETH_P_IPV6 ) ;
2013-02-13 16:00:17 +04:00
br_group . vid = vid ;
2010-04-22 20:54:22 +04:00
2014-06-07 20:26:26 +04:00
br_multicast_leave_group ( br , port , & br_group , & br - > ip6_other_query ,
own_query ) ;
2010-04-22 20:54:22 +04:00
}
# endif
2010-04-18 07:42:07 +04:00
2010-02-27 22:41:45 +03:00
static int br_multicast_ipv4_rcv ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb ,
u16 vid )
2010-02-27 22:41:45 +03:00
{
2015-05-02 15:01:07 +03:00
struct sk_buff * skb_trimmed = NULL ;
2010-02-27 22:41:45 +03:00
struct igmphdr * ih ;
int err ;
2015-05-02 15:01:07 +03:00
err = ip_mc_check_igmp ( skb , & skb_trimmed ) ;
2010-02-27 22:41:45 +03:00
2015-05-02 15:01:07 +03:00
if ( err = = - ENOMSG ) {
if ( ! ipv4_is_local_multicast ( ip_hdr ( skb ) - > daddr ) )
2011-06-23 06:39:12 +04:00
BR_INPUT_SKB_CB ( skb ) - > mrouters_only = 1 ;
2010-02-27 22:41:45 +03:00
return 0 ;
2015-05-02 15:01:07 +03:00
} else if ( err < 0 ) {
return err ;
2011-06-23 06:39:12 +04:00
}
2010-02-27 22:41:45 +03:00
BR_INPUT_SKB_CB ( skb ) - > igmp = 1 ;
2015-05-02 15:01:07 +03:00
ih = igmp_hdr ( skb ) ;
2010-02-27 22:41:45 +03:00
switch ( ih - > type ) {
case IGMP_HOST_MEMBERSHIP_REPORT :
case IGMPV2_HOST_MEMBERSHIP_REPORT :
2011-06-13 19:04:43 +04:00
BR_INPUT_SKB_CB ( skb ) - > mrouters_only = 1 ;
2013-02-13 16:00:17 +04:00
err = br_ip4_multicast_add_group ( br , port , ih - > group , vid ) ;
2010-02-27 22:41:45 +03:00
break ;
case IGMPV3_HOST_MEMBERSHIP_REPORT :
2015-05-02 15:01:07 +03:00
err = br_ip4_multicast_igmp3_report ( br , port , skb_trimmed , vid ) ;
2010-02-27 22:41:45 +03:00
break ;
case IGMP_HOST_MEMBERSHIP_QUERY :
2015-05-02 15:01:07 +03:00
err = br_ip4_multicast_query ( br , port , skb_trimmed , vid ) ;
2010-02-27 22:41:45 +03:00
break ;
case IGMP_HOST_LEAVE_MESSAGE :
2013-02-13 16:00:17 +04:00
br_ip4_multicast_leave_group ( br , port , ih - > group , vid ) ;
2010-02-27 22:41:45 +03:00
break ;
}
2015-05-02 15:01:07 +03:00
if ( skb_trimmed )
kfree_skb ( skb_trimmed ) ;
2010-02-27 22:41:45 +03:00
return err ;
}
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
static int br_multicast_ipv6_rcv ( struct net_bridge * br ,
struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb ,
u16 vid )
2010-04-22 20:54:22 +04:00
{
2015-05-02 15:01:07 +03:00
struct sk_buff * skb_trimmed = NULL ;
struct mld_msg * mld ;
2010-04-22 20:54:22 +04:00
int err ;
2015-05-02 15:01:07 +03:00
err = ipv6_mc_check_mld ( skb , & skb_trimmed ) ;
2010-04-22 20:54:22 +04:00
2015-05-02 15:01:07 +03:00
if ( err = = - ENOMSG ) {
if ( ! ipv6_addr_is_ll_all_nodes ( & ipv6_hdr ( skb ) - > daddr ) )
BR_INPUT_SKB_CB ( skb ) - > mrouters_only = 1 ;
2010-04-22 20:54:22 +04:00
return 0 ;
2015-05-02 15:01:07 +03:00
} else if ( err < 0 ) {
return err ;
2010-04-22 20:54:22 +04:00
}
BR_INPUT_SKB_CB ( skb ) - > igmp = 1 ;
2015-05-02 15:01:07 +03:00
mld = ( struct mld_msg * ) skb_transport_header ( skb ) ;
2010-04-22 20:54:22 +04:00
2015-05-02 15:01:07 +03:00
switch ( mld - > mld_type ) {
2010-04-22 20:54:22 +04:00
case ICMPV6_MGM_REPORT :
2011-06-13 19:06:58 +04:00
BR_INPUT_SKB_CB ( skb ) - > mrouters_only = 1 ;
2013-02-13 16:00:17 +04:00
err = br_ip6_multicast_add_group ( br , port , & mld - > mld_mca , vid ) ;
2010-04-22 20:54:22 +04:00
break ;
case ICMPV6_MLD2_REPORT :
2015-05-02 15:01:07 +03:00
err = br_ip6_multicast_mld2_report ( br , port , skb_trimmed , vid ) ;
2010-04-22 20:54:22 +04:00
break ;
case ICMPV6_MGM_QUERY :
2015-05-02 15:01:07 +03:00
err = br_ip6_multicast_query ( br , port , skb_trimmed , vid ) ;
2010-04-22 20:54:22 +04:00
break ;
case ICMPV6_MGM_REDUCTION :
2013-02-13 16:00:17 +04:00
br_ip6_multicast_leave_group ( br , port , & mld - > mld_mca , vid ) ;
2015-05-02 15:01:07 +03:00
break ;
2010-04-22 20:54:22 +04:00
}
2015-05-02 15:01:07 +03:00
if ( skb_trimmed )
kfree_skb ( skb_trimmed ) ;
2010-04-22 20:54:22 +04:00
return err ;
}
# endif
2010-02-27 22:41:45 +03:00
int br_multicast_rcv ( struct net_bridge * br , struct net_bridge_port * port ,
2013-10-28 23:45:07 +04:00
struct sk_buff * skb , u16 vid )
2010-02-27 22:41:45 +03:00
{
2010-04-25 12:06:40 +04:00
BR_INPUT_SKB_CB ( skb ) - > igmp = 0 ;
BR_INPUT_SKB_CB ( skb ) - > mrouters_only = 0 ;
2010-02-27 22:41:45 +03:00
if ( br - > multicast_disabled )
return 0 ;
switch ( skb - > protocol ) {
case htons ( ETH_P_IP ) :
2013-10-28 23:45:07 +04:00
return br_multicast_ipv4_rcv ( br , port , skb , vid ) ;
2011-12-10 13:48:31 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2010-04-22 20:54:22 +04:00
case htons ( ETH_P_IPV6 ) :
2013-10-28 23:45:07 +04:00
return br_multicast_ipv6_rcv ( br , port , skb , vid ) ;
2010-04-22 20:54:22 +04:00
# endif
2010-02-27 22:41:45 +03:00
}
return 0 ;
}
2013-08-30 19:28:17 +04:00
static void br_multicast_query_expired ( struct net_bridge * br ,
2014-06-07 20:26:29 +04:00
struct bridge_mcast_own_query * query ,
struct bridge_mcast_querier * querier )
2013-08-30 19:28:17 +04:00
{
spin_lock ( & br - > multicast_lock ) ;
if ( query - > startup_sent < br - > multicast_startup_query_count )
query - > startup_sent + + ;
2015-05-28 14:42:54 +03:00
RCU_INIT_POINTER ( querier - > port , NULL ) ;
2013-08-30 19:28:17 +04:00
br_multicast_send_query ( br , NULL , query ) ;
spin_unlock ( & br - > multicast_lock ) ;
}
static void br_ip4_multicast_query_expired ( unsigned long data )
2010-02-27 22:41:45 +03:00
{
struct net_bridge * br = ( void * ) data ;
2014-06-07 20:26:29 +04:00
br_multicast_query_expired ( br , & br - > ip4_own_query , & br - > ip4_querier ) ;
2013-08-30 19:28:17 +04:00
}
2010-02-27 22:41:45 +03:00
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
static void br_ip6_multicast_query_expired ( unsigned long data )
{
struct net_bridge * br = ( void * ) data ;
2010-02-27 22:41:45 +03:00
2014-06-07 20:26:29 +04:00
br_multicast_query_expired ( br , & br - > ip6_own_query , & br - > ip6_querier ) ;
2010-02-27 22:41:45 +03:00
}
2013-08-30 19:28:17 +04:00
# endif
2010-02-27 22:41:45 +03:00
void br_multicast_init ( struct net_bridge * br )
{
br - > hash_elasticity = 4 ;
br - > hash_max = 512 ;
br - > multicast_router = 1 ;
2012-04-13 06:37:42 +04:00
br - > multicast_querier = 0 ;
2013-05-22 01:52:54 +04:00
br - > multicast_query_use_ifaddr = 0 ;
2010-02-27 22:41:45 +03:00
br - > multicast_last_member_count = 2 ;
br - > multicast_startup_query_count = 2 ;
br - > multicast_last_member_interval = HZ ;
br - > multicast_query_response_interval = 10 * HZ ;
br - > multicast_startup_query_interval = 125 * HZ / 4 ;
br - > multicast_query_interval = 125 * HZ ;
br - > multicast_querier_interval = 255 * HZ ;
br - > multicast_membership_interval = 260 * HZ ;
2014-06-07 20:26:26 +04:00
br - > ip4_other_query . delay_time = 0 ;
2014-06-07 20:26:29 +04:00
br - > ip4_querier . port = NULL ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
br - > ip6_other_query . delay_time = 0 ;
2014-06-07 20:26:29 +04:00
br - > ip6_querier . port = NULL ;
2013-08-30 19:28:17 +04:00
# endif
2013-08-01 03:06:20 +04:00
2010-02-27 22:41:45 +03:00
spin_lock_init ( & br - > multicast_lock ) ;
setup_timer ( & br - > multicast_router_timer ,
br_multicast_local_router_expired , 0 ) ;
2014-06-07 20:26:26 +04:00
setup_timer ( & br - > ip4_other_query . timer ,
br_ip4_multicast_querier_expired , ( unsigned long ) br ) ;
setup_timer ( & br - > ip4_own_query . timer , br_ip4_multicast_query_expired ,
2010-02-27 22:41:45 +03:00
( unsigned long ) br ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
setup_timer ( & br - > ip6_other_query . timer ,
br_ip6_multicast_querier_expired , ( unsigned long ) br ) ;
setup_timer ( & br - > ip6_own_query . timer , br_ip6_multicast_query_expired ,
2013-08-30 19:28:17 +04:00
( unsigned long ) br ) ;
# endif
2010-02-27 22:41:45 +03:00
}
2013-08-30 19:28:17 +04:00
static void __br_multicast_open ( struct net_bridge * br ,
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * query )
2010-02-27 22:41:45 +03:00
{
2013-08-30 19:28:17 +04:00
query - > startup_sent = 0 ;
2010-02-27 22:41:45 +03:00
if ( br - > multicast_disabled )
return ;
2013-08-30 19:28:17 +04:00
mod_timer ( & query - > timer , jiffies ) ;
}
void br_multicast_open ( struct net_bridge * br )
{
2014-06-07 20:26:26 +04:00
__br_multicast_open ( br , & br - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
__br_multicast_open ( br , & br - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
# endif
2010-02-27 22:41:45 +03:00
}
void br_multicast_stop ( struct net_bridge * br )
{
del_timer_sync ( & br - > multicast_router_timer ) ;
2014-06-07 20:26:26 +04:00
del_timer_sync ( & br - > ip4_other_query . timer ) ;
del_timer_sync ( & br - > ip4_own_query . timer ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
del_timer_sync ( & br - > ip6_other_query . timer ) ;
del_timer_sync ( & br - > ip6_own_query . timer ) ;
2013-08-30 19:28:17 +04:00
# endif
2015-07-15 17:16:51 +03:00
}
void br_multicast_dev_del ( struct net_bridge * br )
{
struct net_bridge_mdb_htable * mdb ;
struct net_bridge_mdb_entry * mp ;
struct hlist_node * n ;
u32 ver ;
int i ;
2010-02-27 22:41:45 +03:00
spin_lock_bh ( & br - > multicast_lock ) ;
2010-11-15 09:38:10 +03:00
mdb = mlock_dereference ( br - > mdb , br ) ;
2010-02-27 22:41:45 +03:00
if ( ! mdb )
goto out ;
br - > mdb = NULL ;
ver = mdb - > ver ;
for ( i = 0 ; i < mdb - > max ; i + + ) {
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 05:06:00 +04:00
hlist_for_each_entry_safe ( mp , n , & mdb - > mhash [ i ] ,
2010-02-27 22:41:45 +03:00
hlist [ ver ] ) {
del_timer ( & mp - > timer ) ;
call_rcu_bh ( & mp - > rcu , br_multicast_free_group ) ;
}
}
if ( mdb - > old ) {
spin_unlock_bh ( & br - > multicast_lock ) ;
2010-03-06 00:03:35 +03:00
rcu_barrier_bh ( ) ;
2010-02-27 22:41:45 +03:00
spin_lock_bh ( & br - > multicast_lock ) ;
WARN_ON ( mdb - > old ) ;
}
mdb - > old = mdb ;
call_rcu_bh ( & mdb - > rcu , br_mdb_free ) ;
out :
spin_unlock_bh ( & br - > multicast_lock ) ;
}
2010-02-27 22:41:49 +03:00
int br_multicast_set_router ( struct net_bridge * br , unsigned long val )
{
2015-05-23 04:12:34 +03:00
int err = - EINVAL ;
2010-02-27 22:41:49 +03:00
spin_lock_bh ( & br - > multicast_lock ) ;
switch ( val ) {
case 0 :
case 2 :
del_timer ( & br - > multicast_router_timer ) ;
/* fall through */
case 1 :
br - > multicast_router = val ;
err = 0 ;
break ;
}
spin_unlock_bh ( & br - > multicast_lock ) ;
return err ;
}
int br_multicast_set_port_router ( struct net_bridge_port * p , unsigned long val )
{
struct net_bridge * br = p - > br ;
2015-05-23 04:12:34 +03:00
int err = - EINVAL ;
2010-02-27 22:41:49 +03:00
spin_lock ( & br - > multicast_lock ) ;
switch ( val ) {
case 0 :
case 1 :
case 2 :
p - > multicast_router = val ;
err = 0 ;
if ( val < 2 & & ! hlist_unhashed ( & p - > rlist ) )
hlist_del_init_rcu ( & p - > rlist ) ;
if ( val = = 1 )
break ;
del_timer ( & p - > multicast_router_timer ) ;
if ( val = = 0 )
break ;
br_multicast_add_router ( br , p ) ;
break ;
}
spin_unlock ( & br - > multicast_lock ) ;
return err ;
}
2010-02-27 22:41:50 +03:00
2013-08-30 19:28:17 +04:00
static void br_multicast_start_querier ( struct net_bridge * br ,
2014-06-07 20:26:26 +04:00
struct bridge_mcast_own_query * query )
2010-02-27 22:41:50 +03:00
{
struct net_bridge_port * port ;
2012-04-13 06:37:42 +04:00
2013-08-30 19:28:17 +04:00
__br_multicast_open ( br , query ) ;
2012-04-13 06:37:42 +04:00
list_for_each_entry ( port , & br - > port_list , list ) {
if ( port - > state = = BR_STATE_DISABLED | |
port - > state = = BR_STATE_BLOCKING )
continue ;
2014-06-07 20:26:26 +04:00
if ( query = = & br - > ip4_own_query )
br_multicast_enable ( & port - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
else
2014-06-07 20:26:26 +04:00
br_multicast_enable ( & port - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
# endif
2012-04-13 06:37:42 +04:00
}
}
int br_multicast_toggle ( struct net_bridge * br , unsigned long val )
{
2010-07-29 04:45:30 +04:00
int err = 0 ;
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * mdb ;
2010-02-27 22:41:50 +03:00
2011-11-10 09:48:03 +04:00
spin_lock_bh ( & br - > multicast_lock ) ;
2010-02-27 22:41:50 +03:00
if ( br - > multicast_disabled = = ! val )
goto unlock ;
br - > multicast_disabled = ! val ;
if ( br - > multicast_disabled )
goto unlock ;
2010-07-29 04:45:30 +04:00
if ( ! netif_running ( br - > dev ) )
goto unlock ;
2010-11-15 09:38:10 +03:00
mdb = mlock_dereference ( br - > mdb , br ) ;
if ( mdb ) {
if ( mdb - > old ) {
2010-02-27 22:41:50 +03:00
err = - EEXIST ;
rollback :
br - > multicast_disabled = ! ! val ;
goto unlock ;
}
2010-11-15 09:38:10 +03:00
err = br_mdb_rehash ( & br - > mdb , mdb - > max ,
2010-02-27 22:41:50 +03:00
br - > hash_elasticity ) ;
if ( err )
goto rollback ;
}
2014-06-07 20:26:26 +04:00
br_multicast_start_querier ( br , & br - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
br_multicast_start_querier ( br , & br - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
# endif
2010-02-27 22:41:50 +03:00
unlock :
2011-11-10 09:48:03 +04:00
spin_unlock_bh ( & br - > multicast_lock ) ;
2010-02-27 22:41:50 +03:00
return err ;
}
2010-02-27 22:41:51 +03:00
2012-04-13 06:37:42 +04:00
int br_multicast_set_querier ( struct net_bridge * br , unsigned long val )
{
2013-08-01 03:06:20 +04:00
unsigned long max_delay ;
2012-04-13 06:37:42 +04:00
val = ! ! val ;
spin_lock_bh ( & br - > multicast_lock ) ;
if ( br - > multicast_querier = = val )
goto unlock ;
br - > multicast_querier = val ;
2013-08-01 03:06:20 +04:00
if ( ! val )
goto unlock ;
max_delay = br - > multicast_query_response_interval ;
2014-06-07 20:26:26 +04:00
if ( ! timer_pending ( & br - > ip4_other_query . timer ) )
br - > ip4_other_query . delay_time = jiffies + max_delay ;
2013-08-30 19:28:17 +04:00
2014-06-07 20:26:26 +04:00
br_multicast_start_querier ( br , & br - > ip4_own_query ) ;
2013-08-30 19:28:17 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:26 +04:00
if ( ! timer_pending ( & br - > ip6_other_query . timer ) )
br - > ip6_other_query . delay_time = jiffies + max_delay ;
2013-08-30 19:28:17 +04:00
2014-06-07 20:26:26 +04:00
br_multicast_start_querier ( br , & br - > ip6_own_query ) ;
2013-08-30 19:28:17 +04:00
# endif
2012-04-13 06:37:42 +04:00
unlock :
spin_unlock_bh ( & br - > multicast_lock ) ;
return 0 ;
}
2010-02-27 22:41:51 +03:00
int br_multicast_set_hash_max ( struct net_bridge * br , unsigned long val )
{
2015-05-23 04:12:34 +03:00
int err = - EINVAL ;
2010-02-27 22:41:51 +03:00
u32 old ;
2010-11-15 09:38:10 +03:00
struct net_bridge_mdb_htable * mdb ;
2010-02-27 22:41:51 +03:00
2014-01-06 23:00:32 +04:00
spin_lock_bh ( & br - > multicast_lock ) ;
2010-02-27 22:41:51 +03:00
if ( ! is_power_of_2 ( val ) )
goto unlock ;
2010-11-15 09:38:10 +03:00
mdb = mlock_dereference ( br - > mdb , br ) ;
if ( mdb & & val < mdb - > size )
2010-02-27 22:41:51 +03:00
goto unlock ;
err = 0 ;
old = br - > hash_max ;
br - > hash_max = val ;
2010-11-15 09:38:10 +03:00
if ( mdb ) {
if ( mdb - > old ) {
2010-02-27 22:41:51 +03:00
err = - EEXIST ;
rollback :
br - > hash_max = old ;
goto unlock ;
}
err = br_mdb_rehash ( & br - > mdb , br - > hash_max ,
br - > hash_elasticity ) ;
if ( err )
goto rollback ;
}
unlock :
2014-01-06 23:00:32 +04:00
spin_unlock_bh ( & br - > multicast_lock ) ;
2010-02-27 22:41:51 +03:00
return err ;
}
2014-06-07 20:26:28 +04:00
/**
* br_multicast_list_adjacent - Returns snooped multicast addresses
* @ dev : The bridge port adjacent to which to retrieve addresses
* @ br_ip_list : The list to store found , snooped multicast IP addresses in
*
* Creates a list of IP addresses ( struct br_ip_list ) sensed by the multicast
* snooping feature on all bridge ports of dev ' s bridge device , excluding
* the addresses from dev itself .
*
* Returns the number of items added to br_ip_list .
*
* Notes :
* - br_ip_list needs to be initialized by caller
* - br_ip_list might contain duplicates in the end
* ( needs to be taken care of by caller )
* - br_ip_list needs to be freed by caller
*/
int br_multicast_list_adjacent ( struct net_device * dev ,
struct list_head * br_ip_list )
{
struct net_bridge * br ;
struct net_bridge_port * port ;
struct net_bridge_port_group * group ;
struct br_ip_list * entry ;
int count = 0 ;
rcu_read_lock ( ) ;
if ( ! br_ip_list | | ! br_port_exists ( dev ) )
goto unlock ;
port = br_port_get_rcu ( dev ) ;
if ( ! port | | ! port - > br )
goto unlock ;
br = port - > br ;
list_for_each_entry_rcu ( port , & br - > port_list , list ) {
if ( ! port - > dev | | port - > dev = = dev )
continue ;
hlist_for_each_entry_rcu ( group , & port - > mglist , mglist ) {
entry = kmalloc ( sizeof ( * entry ) , GFP_ATOMIC ) ;
if ( ! entry )
goto unlock ;
entry - > addr = group - > addr ;
list_add ( & entry - > list , br_ip_list ) ;
count + + ;
}
}
unlock :
rcu_read_unlock ( ) ;
return count ;
}
EXPORT_SYMBOL_GPL ( br_multicast_list_adjacent ) ;
2014-06-07 20:26:29 +04:00
2014-07-07 07:41:17 +04:00
/**
* br_multicast_has_querier_anywhere - Checks for a querier on a bridge
* @ dev : The bridge port providing the bridge on which to check for a querier
* @ proto : The protocol family to check for : IGMP - > ETH_P_IP , MLD - > ETH_P_IPV6
*
* Checks whether the given interface has a bridge on top and if so returns
* true if a valid querier exists anywhere on the bridged link layer .
* Otherwise returns false .
*/
bool br_multicast_has_querier_anywhere ( struct net_device * dev , int proto )
{
struct net_bridge * br ;
struct net_bridge_port * port ;
struct ethhdr eth ;
bool ret = false ;
rcu_read_lock ( ) ;
if ( ! br_port_exists ( dev ) )
goto unlock ;
port = br_port_get_rcu ( dev ) ;
if ( ! port | | ! port - > br )
goto unlock ;
br = port - > br ;
memset ( & eth , 0 , sizeof ( eth ) ) ;
eth . h_proto = htons ( proto ) ;
ret = br_multicast_querier_exists ( br , & eth ) ;
unlock :
rcu_read_unlock ( ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( br_multicast_has_querier_anywhere ) ;
2014-06-07 20:26:29 +04:00
/**
* br_multicast_has_querier_adjacent - Checks for a querier behind a bridge port
* @ dev : The bridge port adjacent to which to check for a querier
* @ proto : The protocol family to check for : IGMP - > ETH_P_IP , MLD - > ETH_P_IPV6
*
* Checks whether the given interface has a bridge on top and if so returns
* true if a selected querier is behind one of the other ports of this
* bridge . Otherwise returns false .
*/
bool br_multicast_has_querier_adjacent ( struct net_device * dev , int proto )
{
struct net_bridge * br ;
struct net_bridge_port * port ;
bool ret = false ;
rcu_read_lock ( ) ;
if ( ! br_port_exists ( dev ) )
goto unlock ;
port = br_port_get_rcu ( dev ) ;
if ( ! port | | ! port - > br )
goto unlock ;
br = port - > br ;
switch ( proto ) {
case ETH_P_IP :
if ( ! timer_pending ( & br - > ip4_other_query . timer ) | |
rcu_dereference ( br - > ip4_querier . port ) = = port )
goto unlock ;
break ;
2014-06-12 03:41:24 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-06-07 20:26:29 +04:00
case ETH_P_IPV6 :
if ( ! timer_pending ( & br - > ip6_other_query . timer ) | |
rcu_dereference ( br - > ip6_querier . port ) = = port )
goto unlock ;
break ;
2014-06-12 03:41:24 +04:00
# endif
2014-06-07 20:26:29 +04:00
default :
goto unlock ;
}
ret = true ;
unlock :
rcu_read_unlock ( ) ;
return ret ;
}
EXPORT_SYMBOL_GPL ( br_multicast_has_querier_adjacent ) ;