Merge branch 'code-movement-to-br_switchdev-c'
Vladimir Oltean says: ==================== Code movement to br_switchdev.c This is one more refactoring patch set for the Linux bridge, where more logic that is specific to switchdev is moved into br_switchdev.c, which is compiled out when CONFIG_NET_SWITCHDEV is disabled. ==================== Link: https://lore.kernel.org/r/20211027162119.2496321-1-vladimir.oltean@nxp.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
commit
a812a046c2
@ -552,252 +552,16 @@ out:
|
||||
return nlmsg_size;
|
||||
}
|
||||
|
||||
struct br_mdb_complete_info {
|
||||
struct net_bridge_port *port;
|
||||
struct br_ip ip;
|
||||
};
|
||||
|
||||
static void br_mdb_complete(struct net_device *dev, int err, void *priv)
|
||||
{
|
||||
struct br_mdb_complete_info *data = priv;
|
||||
struct net_bridge_port_group __rcu **pp;
|
||||
struct net_bridge_port_group *p;
|
||||
struct net_bridge_mdb_entry *mp;
|
||||
struct net_bridge_port *port = data->port;
|
||||
struct net_bridge *br = port->br;
|
||||
|
||||
if (err)
|
||||
goto err;
|
||||
|
||||
spin_lock_bh(&br->multicast_lock);
|
||||
mp = br_mdb_ip_get(br, &data->ip);
|
||||
if (!mp)
|
||||
goto out;
|
||||
for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
|
||||
pp = &p->next) {
|
||||
if (p->key.port != port)
|
||||
continue;
|
||||
p->flags |= MDB_PG_FLAGS_OFFLOAD;
|
||||
}
|
||||
out:
|
||||
spin_unlock_bh(&br->multicast_lock);
|
||||
err:
|
||||
kfree(priv);
|
||||
}
|
||||
|
||||
static void br_switchdev_mdb_populate(struct switchdev_obj_port_mdb *mdb,
|
||||
const struct net_bridge_mdb_entry *mp)
|
||||
{
|
||||
if (mp->addr.proto == htons(ETH_P_IP))
|
||||
ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
|
||||
#if IS_ENABLED(CONFIG_IPV6)
|
||||
else if (mp->addr.proto == htons(ETH_P_IPV6))
|
||||
ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
|
||||
#endif
|
||||
else
|
||||
ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
|
||||
|
||||
mdb->vid = mp->addr.vid;
|
||||
}
|
||||
|
||||
static int br_mdb_replay_one(struct notifier_block *nb, struct net_device *dev,
|
||||
const struct switchdev_obj_port_mdb *mdb,
|
||||
unsigned long action, const void *ctx,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct switchdev_notifier_port_obj_info obj_info = {
|
||||
.info = {
|
||||
.dev = dev,
|
||||
.extack = extack,
|
||||
.ctx = ctx,
|
||||
},
|
||||
.obj = &mdb->obj,
|
||||
};
|
||||
int err;
|
||||
|
||||
err = nb->notifier_call(nb, action, &obj_info);
|
||||
return notifier_to_errno(err);
|
||||
}
|
||||
|
||||
static int br_mdb_queue_one(struct list_head *mdb_list,
|
||||
enum switchdev_obj_id id,
|
||||
const struct net_bridge_mdb_entry *mp,
|
||||
struct net_device *orig_dev)
|
||||
{
|
||||
struct switchdev_obj_port_mdb *mdb;
|
||||
|
||||
mdb = kzalloc(sizeof(*mdb), GFP_ATOMIC);
|
||||
if (!mdb)
|
||||
return -ENOMEM;
|
||||
|
||||
mdb->obj.id = id;
|
||||
mdb->obj.orig_dev = orig_dev;
|
||||
br_switchdev_mdb_populate(mdb, mp);
|
||||
list_add_tail(&mdb->obj.list, mdb_list);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int br_mdb_replay(struct net_device *br_dev, struct net_device *dev,
|
||||
const void *ctx, bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
const struct net_bridge_mdb_entry *mp;
|
||||
struct switchdev_obj *obj, *tmp;
|
||||
struct net_bridge *br;
|
||||
unsigned long action;
|
||||
LIST_HEAD(mdb_list);
|
||||
int err = 0;
|
||||
|
||||
ASSERT_RTNL();
|
||||
|
||||
if (!nb)
|
||||
return 0;
|
||||
|
||||
if (!netif_is_bridge_master(br_dev) || !netif_is_bridge_port(dev))
|
||||
return -EINVAL;
|
||||
|
||||
br = netdev_priv(br_dev);
|
||||
|
||||
if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
|
||||
return 0;
|
||||
|
||||
/* We cannot walk over br->mdb_list protected just by the rtnl_mutex,
|
||||
* because the write-side protection is br->multicast_lock. But we
|
||||
* need to emulate the [ blocking ] calling context of a regular
|
||||
* switchdev event, so since both br->multicast_lock and RCU read side
|
||||
* critical sections are atomic, we have no choice but to pick the RCU
|
||||
* read side lock, queue up all our events, leave the critical section
|
||||
* and notify switchdev from blocking context.
|
||||
*/
|
||||
rcu_read_lock();
|
||||
|
||||
hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
|
||||
struct net_bridge_port_group __rcu * const *pp;
|
||||
const struct net_bridge_port_group *p;
|
||||
|
||||
if (mp->host_joined) {
|
||||
err = br_mdb_queue_one(&mdb_list,
|
||||
SWITCHDEV_OBJ_ID_HOST_MDB,
|
||||
mp, br_dev);
|
||||
if (err) {
|
||||
rcu_read_unlock();
|
||||
goto out_free_mdb;
|
||||
}
|
||||
}
|
||||
|
||||
for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
|
||||
pp = &p->next) {
|
||||
if (p->key.port->dev != dev)
|
||||
continue;
|
||||
|
||||
err = br_mdb_queue_one(&mdb_list,
|
||||
SWITCHDEV_OBJ_ID_PORT_MDB,
|
||||
mp, dev);
|
||||
if (err) {
|
||||
rcu_read_unlock();
|
||||
goto out_free_mdb;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
rcu_read_unlock();
|
||||
|
||||
if (adding)
|
||||
action = SWITCHDEV_PORT_OBJ_ADD;
|
||||
else
|
||||
action = SWITCHDEV_PORT_OBJ_DEL;
|
||||
|
||||
list_for_each_entry(obj, &mdb_list, list) {
|
||||
err = br_mdb_replay_one(nb, dev, SWITCHDEV_OBJ_PORT_MDB(obj),
|
||||
action, ctx, extack);
|
||||
if (err)
|
||||
goto out_free_mdb;
|
||||
}
|
||||
|
||||
out_free_mdb:
|
||||
list_for_each_entry_safe(obj, tmp, &mdb_list, list) {
|
||||
list_del(&obj->list);
|
||||
kfree(SWITCHDEV_OBJ_PORT_MDB(obj));
|
||||
}
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
static void br_mdb_switchdev_host_port(struct net_device *dev,
|
||||
struct net_device *lower_dev,
|
||||
struct net_bridge_mdb_entry *mp,
|
||||
int type)
|
||||
{
|
||||
struct switchdev_obj_port_mdb mdb = {
|
||||
.obj = {
|
||||
.id = SWITCHDEV_OBJ_ID_HOST_MDB,
|
||||
.flags = SWITCHDEV_F_DEFER,
|
||||
.orig_dev = dev,
|
||||
},
|
||||
};
|
||||
|
||||
br_switchdev_mdb_populate(&mdb, mp);
|
||||
|
||||
switch (type) {
|
||||
case RTM_NEWMDB:
|
||||
switchdev_port_obj_add(lower_dev, &mdb.obj, NULL);
|
||||
break;
|
||||
case RTM_DELMDB:
|
||||
switchdev_port_obj_del(lower_dev, &mdb.obj);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
static void br_mdb_switchdev_host(struct net_device *dev,
|
||||
struct net_bridge_mdb_entry *mp, int type)
|
||||
{
|
||||
struct net_device *lower_dev;
|
||||
struct list_head *iter;
|
||||
|
||||
netdev_for_each_lower_dev(dev, lower_dev, iter)
|
||||
br_mdb_switchdev_host_port(dev, lower_dev, mp, type);
|
||||
}
|
||||
|
||||
void br_mdb_notify(struct net_device *dev,
|
||||
struct net_bridge_mdb_entry *mp,
|
||||
struct net_bridge_port_group *pg,
|
||||
int type)
|
||||
{
|
||||
struct br_mdb_complete_info *complete_info;
|
||||
struct switchdev_obj_port_mdb mdb = {
|
||||
.obj = {
|
||||
.id = SWITCHDEV_OBJ_ID_PORT_MDB,
|
||||
.flags = SWITCHDEV_F_DEFER,
|
||||
},
|
||||
};
|
||||
struct net *net = dev_net(dev);
|
||||
struct sk_buff *skb;
|
||||
int err = -ENOBUFS;
|
||||
|
||||
if (pg) {
|
||||
br_switchdev_mdb_populate(&mdb, mp);
|
||||
|
||||
mdb.obj.orig_dev = pg->key.port->dev;
|
||||
switch (type) {
|
||||
case RTM_NEWMDB:
|
||||
complete_info = kmalloc(sizeof(*complete_info), GFP_ATOMIC);
|
||||
if (!complete_info)
|
||||
break;
|
||||
complete_info->port = pg->key.port;
|
||||
complete_info->ip = mp->addr;
|
||||
mdb.obj.complete_priv = complete_info;
|
||||
mdb.obj.complete = br_mdb_complete;
|
||||
if (switchdev_port_obj_add(pg->key.port->dev, &mdb.obj, NULL))
|
||||
kfree(complete_info);
|
||||
break;
|
||||
case RTM_DELMDB:
|
||||
switchdev_port_obj_del(pg->key.port->dev, &mdb.obj);
|
||||
break;
|
||||
}
|
||||
} else {
|
||||
br_mdb_switchdev_host(dev, mp, type);
|
||||
}
|
||||
br_switchdev_mdb_notify(dev, mp, pg, type);
|
||||
|
||||
skb = nlmsg_new(rtnl_mdb_nlmsg_size(pg), GFP_ATOMIC);
|
||||
if (!skb)
|
||||
|
@ -956,9 +956,11 @@ int br_multicast_toggle_vlan_snooping(struct net_bridge *br, bool on,
|
||||
struct netlink_ext_ack *extack);
|
||||
bool br_multicast_toggle_global_vlan(struct net_bridge_vlan *vlan, bool on);
|
||||
|
||||
int br_mdb_replay(struct net_device *br_dev, struct net_device *dev,
|
||||
const void *ctx, bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack);
|
||||
void br_switchdev_mdb_notify(struct net_device *dev,
|
||||
struct net_bridge_mdb_entry *mp,
|
||||
struct net_bridge_port_group *pg,
|
||||
int type);
|
||||
|
||||
int br_rports_fill_info(struct sk_buff *skb,
|
||||
const struct net_bridge_mcast *brmctx);
|
||||
int br_multicast_dump_querier_state(struct sk_buff *skb,
|
||||
@ -1394,12 +1396,11 @@ static inline bool br_multicast_toggle_global_vlan(struct net_bridge_vlan *vlan,
|
||||
return false;
|
||||
}
|
||||
|
||||
static inline int br_mdb_replay(struct net_device *br_dev,
|
||||
struct net_device *dev, const void *ctx,
|
||||
bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack)
|
||||
static inline void br_switchdev_mdb_notify(struct net_device *dev,
|
||||
struct net_bridge_mdb_entry *mp,
|
||||
struct net_bridge_port_group *pg,
|
||||
int type)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
static inline bool
|
||||
@ -1459,9 +1460,6 @@ void br_vlan_notify(const struct net_bridge *br,
|
||||
const struct net_bridge_port *p,
|
||||
u16 vid, u16 vid_range,
|
||||
int cmd);
|
||||
int br_vlan_replay(struct net_device *br_dev, struct net_device *dev,
|
||||
const void *ctx, bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack);
|
||||
bool br_vlan_can_enter_range(const struct net_bridge_vlan *v_curr,
|
||||
const struct net_bridge_vlan *range_end);
|
||||
|
||||
@ -1708,13 +1706,11 @@ static inline bool br_vlan_can_enter_range(const struct net_bridge_vlan *v_curr,
|
||||
return true;
|
||||
}
|
||||
|
||||
static inline int br_vlan_replay(struct net_device *br_dev,
|
||||
struct net_device *dev, const void *ctx,
|
||||
bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack)
|
||||
static inline u16 br_vlan_flags(const struct net_bridge_vlan *v, u16 pvid)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
return 0;
|
||||
}
|
||||
|
||||
#endif
|
||||
|
||||
/* br_vlan_options.c */
|
||||
|
@ -4,6 +4,7 @@
|
||||
#include <linux/netdevice.h>
|
||||
#include <linux/rtnetlink.h>
|
||||
#include <linux/skbuff.h>
|
||||
#include <net/ip.h>
|
||||
#include <net/switchdev.h>
|
||||
|
||||
#include "br_private.h"
|
||||
@ -280,9 +281,10 @@ static void nbp_switchdev_del(struct net_bridge_port *p)
|
||||
}
|
||||
}
|
||||
|
||||
static int br_fdb_replay_one(struct net_bridge *br, struct notifier_block *nb,
|
||||
const struct net_bridge_fdb_entry *fdb,
|
||||
unsigned long action, const void *ctx)
|
||||
static int
|
||||
br_switchdev_fdb_replay_one(struct net_bridge *br, struct notifier_block *nb,
|
||||
const struct net_bridge_fdb_entry *fdb,
|
||||
unsigned long action, const void *ctx)
|
||||
{
|
||||
struct switchdev_notifier_fdb_info item;
|
||||
int err;
|
||||
@ -293,8 +295,9 @@ static int br_fdb_replay_one(struct net_bridge *br, struct notifier_block *nb,
|
||||
return notifier_to_errno(err);
|
||||
}
|
||||
|
||||
static int br_fdb_replay(const struct net_device *br_dev, const void *ctx,
|
||||
bool adding, struct notifier_block *nb)
|
||||
static int
|
||||
br_switchdev_fdb_replay(const struct net_device *br_dev, const void *ctx,
|
||||
bool adding, struct notifier_block *nb)
|
||||
{
|
||||
struct net_bridge_fdb_entry *fdb;
|
||||
struct net_bridge *br;
|
||||
@ -317,7 +320,7 @@ static int br_fdb_replay(const struct net_device *br_dev, const void *ctx,
|
||||
rcu_read_lock();
|
||||
|
||||
hlist_for_each_entry_rcu(fdb, &br->fdb_list, fdb_node) {
|
||||
err = br_fdb_replay_one(br, nb, fdb, action, ctx);
|
||||
err = br_switchdev_fdb_replay_one(br, nb, fdb, action, ctx);
|
||||
if (err)
|
||||
break;
|
||||
}
|
||||
@ -327,6 +330,348 @@ static int br_fdb_replay(const struct net_device *br_dev, const void *ctx,
|
||||
return err;
|
||||
}
|
||||
|
||||
static int
|
||||
br_switchdev_vlan_replay_one(struct notifier_block *nb,
|
||||
struct net_device *dev,
|
||||
struct switchdev_obj_port_vlan *vlan,
|
||||
const void *ctx, unsigned long action,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct switchdev_notifier_port_obj_info obj_info = {
|
||||
.info = {
|
||||
.dev = dev,
|
||||
.extack = extack,
|
||||
.ctx = ctx,
|
||||
},
|
||||
.obj = &vlan->obj,
|
||||
};
|
||||
int err;
|
||||
|
||||
err = nb->notifier_call(nb, action, &obj_info);
|
||||
return notifier_to_errno(err);
|
||||
}
|
||||
|
||||
static int br_switchdev_vlan_replay(struct net_device *br_dev,
|
||||
struct net_device *dev,
|
||||
const void *ctx, bool adding,
|
||||
struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct net_bridge_vlan_group *vg;
|
||||
struct net_bridge_vlan *v;
|
||||
struct net_bridge_port *p;
|
||||
struct net_bridge *br;
|
||||
unsigned long action;
|
||||
int err = 0;
|
||||
u16 pvid;
|
||||
|
||||
ASSERT_RTNL();
|
||||
|
||||
if (!nb)
|
||||
return 0;
|
||||
|
||||
if (!netif_is_bridge_master(br_dev))
|
||||
return -EINVAL;
|
||||
|
||||
if (!netif_is_bridge_master(dev) && !netif_is_bridge_port(dev))
|
||||
return -EINVAL;
|
||||
|
||||
if (netif_is_bridge_master(dev)) {
|
||||
br = netdev_priv(dev);
|
||||
vg = br_vlan_group(br);
|
||||
p = NULL;
|
||||
} else {
|
||||
p = br_port_get_rtnl(dev);
|
||||
if (WARN_ON(!p))
|
||||
return -EINVAL;
|
||||
vg = nbp_vlan_group(p);
|
||||
br = p->br;
|
||||
}
|
||||
|
||||
if (!vg)
|
||||
return 0;
|
||||
|
||||
if (adding)
|
||||
action = SWITCHDEV_PORT_OBJ_ADD;
|
||||
else
|
||||
action = SWITCHDEV_PORT_OBJ_DEL;
|
||||
|
||||
pvid = br_get_pvid(vg);
|
||||
|
||||
list_for_each_entry(v, &vg->vlan_list, vlist) {
|
||||
struct switchdev_obj_port_vlan vlan = {
|
||||
.obj.orig_dev = dev,
|
||||
.obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
|
||||
.flags = br_vlan_flags(v, pvid),
|
||||
.vid = v->vid,
|
||||
};
|
||||
|
||||
if (!br_vlan_should_use(v))
|
||||
continue;
|
||||
|
||||
err = br_switchdev_vlan_replay_one(nb, dev, &vlan, ctx,
|
||||
action, extack);
|
||||
if (err)
|
||||
return err;
|
||||
}
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_BRIDGE_IGMP_SNOOPING
|
||||
struct br_switchdev_mdb_complete_info {
|
||||
struct net_bridge_port *port;
|
||||
struct br_ip ip;
|
||||
};
|
||||
|
||||
static void br_switchdev_mdb_complete(struct net_device *dev, int err, void *priv)
|
||||
{
|
||||
struct br_switchdev_mdb_complete_info *data = priv;
|
||||
struct net_bridge_port_group __rcu **pp;
|
||||
struct net_bridge_port_group *p;
|
||||
struct net_bridge_mdb_entry *mp;
|
||||
struct net_bridge_port *port = data->port;
|
||||
struct net_bridge *br = port->br;
|
||||
|
||||
if (err)
|
||||
goto err;
|
||||
|
||||
spin_lock_bh(&br->multicast_lock);
|
||||
mp = br_mdb_ip_get(br, &data->ip);
|
||||
if (!mp)
|
||||
goto out;
|
||||
for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
|
||||
pp = &p->next) {
|
||||
if (p->key.port != port)
|
||||
continue;
|
||||
p->flags |= MDB_PG_FLAGS_OFFLOAD;
|
||||
}
|
||||
out:
|
||||
spin_unlock_bh(&br->multicast_lock);
|
||||
err:
|
||||
kfree(priv);
|
||||
}
|
||||
|
||||
static void br_switchdev_mdb_populate(struct switchdev_obj_port_mdb *mdb,
|
||||
const struct net_bridge_mdb_entry *mp)
|
||||
{
|
||||
if (mp->addr.proto == htons(ETH_P_IP))
|
||||
ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
|
||||
#if IS_ENABLED(CONFIG_IPV6)
|
||||
else if (mp->addr.proto == htons(ETH_P_IPV6))
|
||||
ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
|
||||
#endif
|
||||
else
|
||||
ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
|
||||
|
||||
mdb->vid = mp->addr.vid;
|
||||
}
|
||||
|
||||
static void br_switchdev_host_mdb_one(struct net_device *dev,
|
||||
struct net_device *lower_dev,
|
||||
struct net_bridge_mdb_entry *mp,
|
||||
int type)
|
||||
{
|
||||
struct switchdev_obj_port_mdb mdb = {
|
||||
.obj = {
|
||||
.id = SWITCHDEV_OBJ_ID_HOST_MDB,
|
||||
.flags = SWITCHDEV_F_DEFER,
|
||||
.orig_dev = dev,
|
||||
},
|
||||
};
|
||||
|
||||
br_switchdev_mdb_populate(&mdb, mp);
|
||||
|
||||
switch (type) {
|
||||
case RTM_NEWMDB:
|
||||
switchdev_port_obj_add(lower_dev, &mdb.obj, NULL);
|
||||
break;
|
||||
case RTM_DELMDB:
|
||||
switchdev_port_obj_del(lower_dev, &mdb.obj);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
static void br_switchdev_host_mdb(struct net_device *dev,
|
||||
struct net_bridge_mdb_entry *mp, int type)
|
||||
{
|
||||
struct net_device *lower_dev;
|
||||
struct list_head *iter;
|
||||
|
||||
netdev_for_each_lower_dev(dev, lower_dev, iter)
|
||||
br_switchdev_host_mdb_one(dev, lower_dev, mp, type);
|
||||
}
|
||||
|
||||
static int
|
||||
br_switchdev_mdb_replay_one(struct notifier_block *nb, struct net_device *dev,
|
||||
const struct switchdev_obj_port_mdb *mdb,
|
||||
unsigned long action, const void *ctx,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct switchdev_notifier_port_obj_info obj_info = {
|
||||
.info = {
|
||||
.dev = dev,
|
||||
.extack = extack,
|
||||
.ctx = ctx,
|
||||
},
|
||||
.obj = &mdb->obj,
|
||||
};
|
||||
int err;
|
||||
|
||||
err = nb->notifier_call(nb, action, &obj_info);
|
||||
return notifier_to_errno(err);
|
||||
}
|
||||
|
||||
static int br_switchdev_mdb_queue_one(struct list_head *mdb_list,
|
||||
enum switchdev_obj_id id,
|
||||
const struct net_bridge_mdb_entry *mp,
|
||||
struct net_device *orig_dev)
|
||||
{
|
||||
struct switchdev_obj_port_mdb *mdb;
|
||||
|
||||
mdb = kzalloc(sizeof(*mdb), GFP_ATOMIC);
|
||||
if (!mdb)
|
||||
return -ENOMEM;
|
||||
|
||||
mdb->obj.id = id;
|
||||
mdb->obj.orig_dev = orig_dev;
|
||||
br_switchdev_mdb_populate(mdb, mp);
|
||||
list_add_tail(&mdb->obj.list, mdb_list);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void br_switchdev_mdb_notify(struct net_device *dev,
|
||||
struct net_bridge_mdb_entry *mp,
|
||||
struct net_bridge_port_group *pg,
|
||||
int type)
|
||||
{
|
||||
struct br_switchdev_mdb_complete_info *complete_info;
|
||||
struct switchdev_obj_port_mdb mdb = {
|
||||
.obj = {
|
||||
.id = SWITCHDEV_OBJ_ID_PORT_MDB,
|
||||
.flags = SWITCHDEV_F_DEFER,
|
||||
},
|
||||
};
|
||||
|
||||
if (!pg)
|
||||
return br_switchdev_host_mdb(dev, mp, type);
|
||||
|
||||
br_switchdev_mdb_populate(&mdb, mp);
|
||||
|
||||
mdb.obj.orig_dev = pg->key.port->dev;
|
||||
switch (type) {
|
||||
case RTM_NEWMDB:
|
||||
complete_info = kmalloc(sizeof(*complete_info), GFP_ATOMIC);
|
||||
if (!complete_info)
|
||||
break;
|
||||
complete_info->port = pg->key.port;
|
||||
complete_info->ip = mp->addr;
|
||||
mdb.obj.complete_priv = complete_info;
|
||||
mdb.obj.complete = br_switchdev_mdb_complete;
|
||||
if (switchdev_port_obj_add(pg->key.port->dev, &mdb.obj, NULL))
|
||||
kfree(complete_info);
|
||||
break;
|
||||
case RTM_DELMDB:
|
||||
switchdev_port_obj_del(pg->key.port->dev, &mdb.obj);
|
||||
break;
|
||||
}
|
||||
}
|
||||
#endif
|
||||
|
||||
static int
|
||||
br_switchdev_mdb_replay(struct net_device *br_dev, struct net_device *dev,
|
||||
const void *ctx, bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
#ifdef CONFIG_BRIDGE_IGMP_SNOOPING
|
||||
const struct net_bridge_mdb_entry *mp;
|
||||
struct switchdev_obj *obj, *tmp;
|
||||
struct net_bridge *br;
|
||||
unsigned long action;
|
||||
LIST_HEAD(mdb_list);
|
||||
int err = 0;
|
||||
|
||||
ASSERT_RTNL();
|
||||
|
||||
if (!nb)
|
||||
return 0;
|
||||
|
||||
if (!netif_is_bridge_master(br_dev) || !netif_is_bridge_port(dev))
|
||||
return -EINVAL;
|
||||
|
||||
br = netdev_priv(br_dev);
|
||||
|
||||
if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
|
||||
return 0;
|
||||
|
||||
/* We cannot walk over br->mdb_list protected just by the rtnl_mutex,
|
||||
* because the write-side protection is br->multicast_lock. But we
|
||||
* need to emulate the [ blocking ] calling context of a regular
|
||||
* switchdev event, so since both br->multicast_lock and RCU read side
|
||||
* critical sections are atomic, we have no choice but to pick the RCU
|
||||
* read side lock, queue up all our events, leave the critical section
|
||||
* and notify switchdev from blocking context.
|
||||
*/
|
||||
rcu_read_lock();
|
||||
|
||||
hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
|
||||
struct net_bridge_port_group __rcu * const *pp;
|
||||
const struct net_bridge_port_group *p;
|
||||
|
||||
if (mp->host_joined) {
|
||||
err = br_switchdev_mdb_queue_one(&mdb_list,
|
||||
SWITCHDEV_OBJ_ID_HOST_MDB,
|
||||
mp, br_dev);
|
||||
if (err) {
|
||||
rcu_read_unlock();
|
||||
goto out_free_mdb;
|
||||
}
|
||||
}
|
||||
|
||||
for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
|
||||
pp = &p->next) {
|
||||
if (p->key.port->dev != dev)
|
||||
continue;
|
||||
|
||||
err = br_switchdev_mdb_queue_one(&mdb_list,
|
||||
SWITCHDEV_OBJ_ID_PORT_MDB,
|
||||
mp, dev);
|
||||
if (err) {
|
||||
rcu_read_unlock();
|
||||
goto out_free_mdb;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
rcu_read_unlock();
|
||||
|
||||
if (adding)
|
||||
action = SWITCHDEV_PORT_OBJ_ADD;
|
||||
else
|
||||
action = SWITCHDEV_PORT_OBJ_DEL;
|
||||
|
||||
list_for_each_entry(obj, &mdb_list, list) {
|
||||
err = br_switchdev_mdb_replay_one(nb, dev,
|
||||
SWITCHDEV_OBJ_PORT_MDB(obj),
|
||||
action, ctx, extack);
|
||||
if (err)
|
||||
goto out_free_mdb;
|
||||
}
|
||||
|
||||
out_free_mdb:
|
||||
list_for_each_entry_safe(obj, tmp, &mdb_list, list) {
|
||||
list_del(&obj->list);
|
||||
kfree(SWITCHDEV_OBJ_PORT_MDB(obj));
|
||||
}
|
||||
|
||||
if (err)
|
||||
return err;
|
||||
#endif
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int nbp_switchdev_sync_objs(struct net_bridge_port *p, const void *ctx,
|
||||
struct notifier_block *atomic_nb,
|
||||
struct notifier_block *blocking_nb,
|
||||
@ -336,15 +681,17 @@ static int nbp_switchdev_sync_objs(struct net_bridge_port *p, const void *ctx,
|
||||
struct net_device *dev = p->dev;
|
||||
int err;
|
||||
|
||||
err = br_vlan_replay(br_dev, dev, ctx, true, blocking_nb, extack);
|
||||
err = br_switchdev_vlan_replay(br_dev, dev, ctx, true, blocking_nb,
|
||||
extack);
|
||||
if (err && err != -EOPNOTSUPP)
|
||||
return err;
|
||||
|
||||
err = br_mdb_replay(br_dev, dev, ctx, true, blocking_nb, extack);
|
||||
err = br_switchdev_mdb_replay(br_dev, dev, ctx, true, blocking_nb,
|
||||
extack);
|
||||
if (err && err != -EOPNOTSUPP)
|
||||
return err;
|
||||
|
||||
err = br_fdb_replay(br_dev, ctx, true, atomic_nb);
|
||||
err = br_switchdev_fdb_replay(br_dev, ctx, true, atomic_nb);
|
||||
if (err && err != -EOPNOTSUPP)
|
||||
return err;
|
||||
|
||||
@ -359,11 +706,11 @@ static void nbp_switchdev_unsync_objs(struct net_bridge_port *p,
|
||||
struct net_device *br_dev = p->br->dev;
|
||||
struct net_device *dev = p->dev;
|
||||
|
||||
br_vlan_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
|
||||
br_switchdev_vlan_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
|
||||
|
||||
br_mdb_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
|
||||
br_switchdev_mdb_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
|
||||
|
||||
br_fdb_replay(br_dev, ctx, false, atomic_nb);
|
||||
br_switchdev_fdb_replay(br_dev, ctx, false, atomic_nb);
|
||||
}
|
||||
|
||||
/* Let the bridge know that this port is offloaded, so that it can assign a
|
||||
|
@ -1860,90 +1860,6 @@ out_kfree:
|
||||
kfree_skb(skb);
|
||||
}
|
||||
|
||||
static int br_vlan_replay_one(struct notifier_block *nb,
|
||||
struct net_device *dev,
|
||||
struct switchdev_obj_port_vlan *vlan,
|
||||
const void *ctx, unsigned long action,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct switchdev_notifier_port_obj_info obj_info = {
|
||||
.info = {
|
||||
.dev = dev,
|
||||
.extack = extack,
|
||||
.ctx = ctx,
|
||||
},
|
||||
.obj = &vlan->obj,
|
||||
};
|
||||
int err;
|
||||
|
||||
err = nb->notifier_call(nb, action, &obj_info);
|
||||
return notifier_to_errno(err);
|
||||
}
|
||||
|
||||
int br_vlan_replay(struct net_device *br_dev, struct net_device *dev,
|
||||
const void *ctx, bool adding, struct notifier_block *nb,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
struct net_bridge_vlan_group *vg;
|
||||
struct net_bridge_vlan *v;
|
||||
struct net_bridge_port *p;
|
||||
struct net_bridge *br;
|
||||
unsigned long action;
|
||||
int err = 0;
|
||||
u16 pvid;
|
||||
|
||||
ASSERT_RTNL();
|
||||
|
||||
if (!nb)
|
||||
return 0;
|
||||
|
||||
if (!netif_is_bridge_master(br_dev))
|
||||
return -EINVAL;
|
||||
|
||||
if (!netif_is_bridge_master(dev) && !netif_is_bridge_port(dev))
|
||||
return -EINVAL;
|
||||
|
||||
if (netif_is_bridge_master(dev)) {
|
||||
br = netdev_priv(dev);
|
||||
vg = br_vlan_group(br);
|
||||
p = NULL;
|
||||
} else {
|
||||
p = br_port_get_rtnl(dev);
|
||||
if (WARN_ON(!p))
|
||||
return -EINVAL;
|
||||
vg = nbp_vlan_group(p);
|
||||
br = p->br;
|
||||
}
|
||||
|
||||
if (!vg)
|
||||
return 0;
|
||||
|
||||
if (adding)
|
||||
action = SWITCHDEV_PORT_OBJ_ADD;
|
||||
else
|
||||
action = SWITCHDEV_PORT_OBJ_DEL;
|
||||
|
||||
pvid = br_get_pvid(vg);
|
||||
|
||||
list_for_each_entry(v, &vg->vlan_list, vlist) {
|
||||
struct switchdev_obj_port_vlan vlan = {
|
||||
.obj.orig_dev = dev,
|
||||
.obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
|
||||
.flags = br_vlan_flags(v, pvid),
|
||||
.vid = v->vid,
|
||||
};
|
||||
|
||||
if (!br_vlan_should_use(v))
|
||||
continue;
|
||||
|
||||
err = br_vlan_replay_one(nb, dev, &vlan, ctx, action, extack);
|
||||
if (err)
|
||||
return err;
|
||||
}
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
/* check if v_curr can enter a range ending in range_end */
|
||||
bool br_vlan_can_enter_range(const struct net_bridge_vlan *v_curr,
|
||||
const struct net_bridge_vlan *range_end)
|
||||
|
Loading…
x
Reference in New Issue
Block a user