dc489f8625
Before this change, generation of the list of MDB events to replay
would race against the creation of new group memberships, either from
the IGMP/MLD snooping logic or from user configuration.
While new memberships are immediately visible to walkers of
br->mdb_list, the notification of their existence to switchdev event
subscribers is deferred until a later point in time. So if a replay
list was generated during a time that overlapped with such a window,
it would also contain a replay of the not-yet-delivered event.
The driver would thus receive two copies of what the bridge internally
considered to be one single event. On destruction of the bridge, only
a single membership deletion event was therefore sent. As a
consequence of this, drivers which reference count memberships (at
least DSA), would be left with orphan groups in their hardware
database when the bridge was destroyed.
This is only an issue when replaying additions. While deletion events
may still be pending on the deferred queue, they will already have
been removed from br->mdb_list, so no duplicates can be generated in
that scenario.
To a user this meant that old group memberships, from a bridge in
which a port was previously attached, could be reanimated (in
hardware) when the port joined a new bridge, without the new bridge's
knowledge.
For example, on an mv88e6xxx system, create a snooping bridge and
immediately add a port to it:
root@infix-06-0b-00:~$ ip link add dev br0 up type bridge mcast_snooping 1 && \
> ip link set dev x3 up master br0
And then destroy the bridge:
root@infix-06-0b-00:~$ ip link del dev br0
root@infix-06-0b-00:~$ mvls atu
ADDRESS FID STATE Q F 0 1 2 3 4 5 6 7 8 9 a
DEV:0 Marvell 88E6393X
33:33:00:00:00:6a 1 static - - 0 . . . . . . . . . .
33:33:ff:87:e4:3f 1 static - - 0 . . . . . . . . . .
ff:ff:ff:ff:ff:ff 1 static - - 0 1 2 3 4 5 6 7 8 9 a
root@infix-06-0b-00:~$
The two IPv6 groups remain in the hardware database because the
port (x3) is notified of the host's membership twice: once via the
original event and once via a replay. Since only a single delete
notification is sent, the count remains at 1 when the bridge is
destroyed.
Then add the same port (or another port belonging to the same hardware
domain) to a new bridge, this time with snooping disabled:
root@infix-06-0b-00:~$ ip link add dev br1 up type bridge mcast_snooping 0 && \
> ip link set dev x3 up master br1
All multicast, including the two IPv6 groups from br0, should now be
flooded, according to the policy of br1. But instead the old
memberships are still active in the hardware database, causing the
switch to only forward traffic to those groups towards the CPU (port
0).
Eliminate the race in two steps:
1. Grab the write-side lock of the MDB while generating the replay
list.
This prevents new memberships from showing up while we are generating
the replay list. But it leaves the scenario in which a deferred event
was already generated, but not delivered, before we grabbed the
lock. Therefore:
2. Make sure that no deferred version of a replay event is already
enqueued to the switchdev deferred queue, before adding it to the
replay list, when replaying additions.
Fixes: 4f2673b3a2
("net: bridge: add helper to replay port and host-joined mdb entries")
Signed-off-by: Tobias Waldekranz <tobias@waldekranz.com>
Reviewed-by: Vladimir Oltean <olteanv@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
529 lines
15 KiB
C
529 lines
15 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* include/net/switchdev.h - Switch device API
|
|
* Copyright (c) 2014-2015 Jiri Pirko <jiri@resnulli.us>
|
|
* Copyright (c) 2014-2015 Scott Feldman <sfeldma@gmail.com>
|
|
*/
|
|
#ifndef _LINUX_SWITCHDEV_H_
|
|
#define _LINUX_SWITCHDEV_H_
|
|
|
|
#include <linux/netdevice.h>
|
|
#include <linux/notifier.h>
|
|
#include <linux/list.h>
|
|
#include <net/ip_fib.h>
|
|
|
|
#define SWITCHDEV_F_NO_RECURSE BIT(0)
|
|
#define SWITCHDEV_F_SKIP_EOPNOTSUPP BIT(1)
|
|
#define SWITCHDEV_F_DEFER BIT(2)
|
|
|
|
enum switchdev_attr_id {
|
|
SWITCHDEV_ATTR_ID_UNDEFINED,
|
|
SWITCHDEV_ATTR_ID_PORT_STP_STATE,
|
|
SWITCHDEV_ATTR_ID_PORT_MST_STATE,
|
|
SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS,
|
|
SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS,
|
|
SWITCHDEV_ATTR_ID_PORT_MROUTER,
|
|
SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME,
|
|
SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING,
|
|
SWITCHDEV_ATTR_ID_BRIDGE_VLAN_PROTOCOL,
|
|
SWITCHDEV_ATTR_ID_BRIDGE_MC_DISABLED,
|
|
SWITCHDEV_ATTR_ID_BRIDGE_MROUTER,
|
|
SWITCHDEV_ATTR_ID_BRIDGE_MST,
|
|
SWITCHDEV_ATTR_ID_MRP_PORT_ROLE,
|
|
SWITCHDEV_ATTR_ID_VLAN_MSTI,
|
|
};
|
|
|
|
struct switchdev_mst_state {
|
|
u16 msti;
|
|
u8 state;
|
|
};
|
|
|
|
struct switchdev_brport_flags {
|
|
unsigned long val;
|
|
unsigned long mask;
|
|
};
|
|
|
|
struct switchdev_vlan_msti {
|
|
u16 vid;
|
|
u16 msti;
|
|
};
|
|
|
|
struct switchdev_attr {
|
|
struct net_device *orig_dev;
|
|
enum switchdev_attr_id id;
|
|
u32 flags;
|
|
void *complete_priv;
|
|
void (*complete)(struct net_device *dev, int err, void *priv);
|
|
union {
|
|
u8 stp_state; /* PORT_STP_STATE */
|
|
struct switchdev_mst_state mst_state; /* PORT_MST_STATE */
|
|
struct switchdev_brport_flags brport_flags; /* PORT_BRIDGE_FLAGS */
|
|
bool mrouter; /* PORT_MROUTER */
|
|
clock_t ageing_time; /* BRIDGE_AGEING_TIME */
|
|
bool vlan_filtering; /* BRIDGE_VLAN_FILTERING */
|
|
u16 vlan_protocol; /* BRIDGE_VLAN_PROTOCOL */
|
|
bool mst; /* BRIDGE_MST */
|
|
bool mc_disabled; /* MC_DISABLED */
|
|
u8 mrp_port_role; /* MRP_PORT_ROLE */
|
|
struct switchdev_vlan_msti vlan_msti; /* VLAN_MSTI */
|
|
} u;
|
|
};
|
|
|
|
enum switchdev_obj_id {
|
|
SWITCHDEV_OBJ_ID_UNDEFINED,
|
|
SWITCHDEV_OBJ_ID_PORT_VLAN,
|
|
SWITCHDEV_OBJ_ID_PORT_MDB,
|
|
SWITCHDEV_OBJ_ID_HOST_MDB,
|
|
SWITCHDEV_OBJ_ID_MRP,
|
|
SWITCHDEV_OBJ_ID_RING_TEST_MRP,
|
|
SWITCHDEV_OBJ_ID_RING_ROLE_MRP,
|
|
SWITCHDEV_OBJ_ID_RING_STATE_MRP,
|
|
SWITCHDEV_OBJ_ID_IN_TEST_MRP,
|
|
SWITCHDEV_OBJ_ID_IN_ROLE_MRP,
|
|
SWITCHDEV_OBJ_ID_IN_STATE_MRP,
|
|
};
|
|
|
|
struct switchdev_obj {
|
|
struct list_head list;
|
|
struct net_device *orig_dev;
|
|
enum switchdev_obj_id id;
|
|
u32 flags;
|
|
void *complete_priv;
|
|
void (*complete)(struct net_device *dev, int err, void *priv);
|
|
};
|
|
|
|
/* SWITCHDEV_OBJ_ID_PORT_VLAN */
|
|
struct switchdev_obj_port_vlan {
|
|
struct switchdev_obj obj;
|
|
u16 flags;
|
|
u16 vid;
|
|
/* If set, the notifier signifies a change of one of the following
|
|
* flags for a VLAN that already exists:
|
|
* - BRIDGE_VLAN_INFO_PVID
|
|
* - BRIDGE_VLAN_INFO_UNTAGGED
|
|
* Entries with BRIDGE_VLAN_INFO_BRENTRY unset are not notified at all.
|
|
*/
|
|
bool changed;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_PORT_VLAN(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_port_vlan, obj)
|
|
|
|
/* SWITCHDEV_OBJ_ID_PORT_MDB */
|
|
struct switchdev_obj_port_mdb {
|
|
struct switchdev_obj obj;
|
|
unsigned char addr[ETH_ALEN];
|
|
u16 vid;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_PORT_MDB(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_port_mdb, obj)
|
|
|
|
|
|
/* SWITCHDEV_OBJ_ID_MRP */
|
|
struct switchdev_obj_mrp {
|
|
struct switchdev_obj obj;
|
|
struct net_device *p_port;
|
|
struct net_device *s_port;
|
|
u32 ring_id;
|
|
u16 prio;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_mrp, obj)
|
|
|
|
/* SWITCHDEV_OBJ_ID_RING_TEST_MRP */
|
|
struct switchdev_obj_ring_test_mrp {
|
|
struct switchdev_obj obj;
|
|
/* The value is in us and a value of 0 represents to stop */
|
|
u32 interval;
|
|
u8 max_miss;
|
|
u32 ring_id;
|
|
u32 period;
|
|
bool monitor;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_RING_TEST_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_ring_test_mrp, obj)
|
|
|
|
/* SWICHDEV_OBJ_ID_RING_ROLE_MRP */
|
|
struct switchdev_obj_ring_role_mrp {
|
|
struct switchdev_obj obj;
|
|
u8 ring_role;
|
|
u32 ring_id;
|
|
u8 sw_backup;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_RING_ROLE_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_ring_role_mrp, obj)
|
|
|
|
struct switchdev_obj_ring_state_mrp {
|
|
struct switchdev_obj obj;
|
|
u8 ring_state;
|
|
u32 ring_id;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_RING_STATE_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_ring_state_mrp, obj)
|
|
|
|
/* SWITCHDEV_OBJ_ID_IN_TEST_MRP */
|
|
struct switchdev_obj_in_test_mrp {
|
|
struct switchdev_obj obj;
|
|
/* The value is in us and a value of 0 represents to stop */
|
|
u32 interval;
|
|
u32 in_id;
|
|
u32 period;
|
|
u8 max_miss;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_IN_TEST_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_in_test_mrp, obj)
|
|
|
|
/* SWICHDEV_OBJ_ID_IN_ROLE_MRP */
|
|
struct switchdev_obj_in_role_mrp {
|
|
struct switchdev_obj obj;
|
|
struct net_device *i_port;
|
|
u32 ring_id;
|
|
u16 in_id;
|
|
u8 in_role;
|
|
u8 sw_backup;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_IN_ROLE_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_in_role_mrp, obj)
|
|
|
|
struct switchdev_obj_in_state_mrp {
|
|
struct switchdev_obj obj;
|
|
u32 in_id;
|
|
u8 in_state;
|
|
};
|
|
|
|
#define SWITCHDEV_OBJ_IN_STATE_MRP(OBJ) \
|
|
container_of((OBJ), struct switchdev_obj_in_state_mrp, obj)
|
|
|
|
struct switchdev_brport {
|
|
struct net_device *dev;
|
|
const void *ctx;
|
|
struct notifier_block *atomic_nb;
|
|
struct notifier_block *blocking_nb;
|
|
bool tx_fwd_offload;
|
|
};
|
|
|
|
enum switchdev_notifier_type {
|
|
SWITCHDEV_FDB_ADD_TO_BRIDGE = 1,
|
|
SWITCHDEV_FDB_DEL_TO_BRIDGE,
|
|
SWITCHDEV_FDB_ADD_TO_DEVICE,
|
|
SWITCHDEV_FDB_DEL_TO_DEVICE,
|
|
SWITCHDEV_FDB_OFFLOADED,
|
|
SWITCHDEV_FDB_FLUSH_TO_BRIDGE,
|
|
|
|
SWITCHDEV_PORT_OBJ_ADD, /* Blocking. */
|
|
SWITCHDEV_PORT_OBJ_DEL, /* Blocking. */
|
|
SWITCHDEV_PORT_ATTR_SET, /* May be blocking . */
|
|
|
|
SWITCHDEV_VXLAN_FDB_ADD_TO_BRIDGE,
|
|
SWITCHDEV_VXLAN_FDB_DEL_TO_BRIDGE,
|
|
SWITCHDEV_VXLAN_FDB_ADD_TO_DEVICE,
|
|
SWITCHDEV_VXLAN_FDB_DEL_TO_DEVICE,
|
|
SWITCHDEV_VXLAN_FDB_OFFLOADED,
|
|
|
|
SWITCHDEV_BRPORT_OFFLOADED,
|
|
SWITCHDEV_BRPORT_UNOFFLOADED,
|
|
SWITCHDEV_BRPORT_REPLAY,
|
|
};
|
|
|
|
struct switchdev_notifier_info {
|
|
struct net_device *dev;
|
|
struct netlink_ext_ack *extack;
|
|
const void *ctx;
|
|
};
|
|
|
|
/* Remember to update br_switchdev_fdb_populate() when adding
|
|
* new members to this structure
|
|
*/
|
|
struct switchdev_notifier_fdb_info {
|
|
struct switchdev_notifier_info info; /* must be first */
|
|
const unsigned char *addr;
|
|
u16 vid;
|
|
u8 added_by_user:1,
|
|
is_local:1,
|
|
locked:1,
|
|
offloaded:1;
|
|
};
|
|
|
|
struct switchdev_notifier_port_obj_info {
|
|
struct switchdev_notifier_info info; /* must be first */
|
|
const struct switchdev_obj *obj;
|
|
bool handled;
|
|
};
|
|
|
|
struct switchdev_notifier_port_attr_info {
|
|
struct switchdev_notifier_info info; /* must be first */
|
|
const struct switchdev_attr *attr;
|
|
bool handled;
|
|
};
|
|
|
|
struct switchdev_notifier_brport_info {
|
|
struct switchdev_notifier_info info; /* must be first */
|
|
const struct switchdev_brport brport;
|
|
};
|
|
|
|
static inline struct net_device *
|
|
switchdev_notifier_info_to_dev(const struct switchdev_notifier_info *info)
|
|
{
|
|
return info->dev;
|
|
}
|
|
|
|
static inline struct netlink_ext_ack *
|
|
switchdev_notifier_info_to_extack(const struct switchdev_notifier_info *info)
|
|
{
|
|
return info->extack;
|
|
}
|
|
|
|
static inline bool
|
|
switchdev_fdb_is_dynamically_learned(const struct switchdev_notifier_fdb_info *fdb_info)
|
|
{
|
|
return !fdb_info->added_by_user && !fdb_info->is_local;
|
|
}
|
|
|
|
#ifdef CONFIG_NET_SWITCHDEV
|
|
|
|
int switchdev_bridge_port_offload(struct net_device *brport_dev,
|
|
struct net_device *dev, const void *ctx,
|
|
struct notifier_block *atomic_nb,
|
|
struct notifier_block *blocking_nb,
|
|
bool tx_fwd_offload,
|
|
struct netlink_ext_ack *extack);
|
|
void switchdev_bridge_port_unoffload(struct net_device *brport_dev,
|
|
const void *ctx,
|
|
struct notifier_block *atomic_nb,
|
|
struct notifier_block *blocking_nb);
|
|
int switchdev_bridge_port_replay(struct net_device *brport_dev,
|
|
struct net_device *dev, const void *ctx,
|
|
struct notifier_block *atomic_nb,
|
|
struct notifier_block *blocking_nb,
|
|
struct netlink_ext_ack *extack);
|
|
|
|
void switchdev_deferred_process(void);
|
|
int switchdev_port_attr_set(struct net_device *dev,
|
|
const struct switchdev_attr *attr,
|
|
struct netlink_ext_ack *extack);
|
|
bool switchdev_port_obj_act_is_deferred(struct net_device *dev,
|
|
enum switchdev_notifier_type nt,
|
|
const struct switchdev_obj *obj);
|
|
int switchdev_port_obj_add(struct net_device *dev,
|
|
const struct switchdev_obj *obj,
|
|
struct netlink_ext_ack *extack);
|
|
int switchdev_port_obj_del(struct net_device *dev,
|
|
const struct switchdev_obj *obj);
|
|
|
|
int register_switchdev_notifier(struct notifier_block *nb);
|
|
int unregister_switchdev_notifier(struct notifier_block *nb);
|
|
int call_switchdev_notifiers(unsigned long val, struct net_device *dev,
|
|
struct switchdev_notifier_info *info,
|
|
struct netlink_ext_ack *extack);
|
|
|
|
int register_switchdev_blocking_notifier(struct notifier_block *nb);
|
|
int unregister_switchdev_blocking_notifier(struct notifier_block *nb);
|
|
int call_switchdev_blocking_notifiers(unsigned long val, struct net_device *dev,
|
|
struct switchdev_notifier_info *info,
|
|
struct netlink_ext_ack *extack);
|
|
|
|
int switchdev_handle_fdb_event_to_device(struct net_device *dev, unsigned long event,
|
|
const struct switchdev_notifier_fdb_info *fdb_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
bool (*foreign_dev_check_cb)(const struct net_device *dev,
|
|
const struct net_device *foreign_dev),
|
|
int (*mod_cb)(struct net_device *dev, struct net_device *orig_dev,
|
|
unsigned long event, const void *ctx,
|
|
const struct switchdev_notifier_fdb_info *fdb_info));
|
|
|
|
int switchdev_handle_port_obj_add(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
int (*add_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj,
|
|
struct netlink_ext_ack *extack));
|
|
int switchdev_handle_port_obj_add_foreign(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
bool (*foreign_dev_check_cb)(const struct net_device *dev,
|
|
const struct net_device *foreign_dev),
|
|
int (*add_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj,
|
|
struct netlink_ext_ack *extack));
|
|
int switchdev_handle_port_obj_del(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
int (*del_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj));
|
|
int switchdev_handle_port_obj_del_foreign(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
bool (*foreign_dev_check_cb)(const struct net_device *dev,
|
|
const struct net_device *foreign_dev),
|
|
int (*del_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj));
|
|
|
|
int switchdev_handle_port_attr_set(struct net_device *dev,
|
|
struct switchdev_notifier_port_attr_info *port_attr_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
int (*set_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_attr *attr,
|
|
struct netlink_ext_ack *extack));
|
|
#else
|
|
|
|
static inline int
|
|
switchdev_bridge_port_offload(struct net_device *brport_dev,
|
|
struct net_device *dev, const void *ctx,
|
|
struct notifier_block *atomic_nb,
|
|
struct notifier_block *blocking_nb,
|
|
bool tx_fwd_offload,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline void
|
|
switchdev_bridge_port_unoffload(struct net_device *brport_dev,
|
|
const void *ctx,
|
|
struct notifier_block *atomic_nb,
|
|
struct notifier_block *blocking_nb)
|
|
{
|
|
}
|
|
|
|
static inline void switchdev_deferred_process(void)
|
|
{
|
|
}
|
|
|
|
static inline int switchdev_port_attr_set(struct net_device *dev,
|
|
const struct switchdev_attr *attr,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int switchdev_port_obj_add(struct net_device *dev,
|
|
const struct switchdev_obj *obj,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int switchdev_port_obj_del(struct net_device *dev,
|
|
const struct switchdev_obj *obj)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int register_switchdev_notifier(struct notifier_block *nb)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int unregister_switchdev_notifier(struct notifier_block *nb)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int call_switchdev_notifiers(unsigned long val,
|
|
struct net_device *dev,
|
|
struct switchdev_notifier_info *info,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
static inline int
|
|
register_switchdev_blocking_notifier(struct notifier_block *nb)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
unregister_switchdev_blocking_notifier(struct notifier_block *nb)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
call_switchdev_blocking_notifiers(unsigned long val,
|
|
struct net_device *dev,
|
|
struct switchdev_notifier_info *info,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
static inline int
|
|
switchdev_handle_fdb_event_to_device(struct net_device *dev, unsigned long event,
|
|
const struct switchdev_notifier_fdb_info *fdb_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
bool (*foreign_dev_check_cb)(const struct net_device *dev,
|
|
const struct net_device *foreign_dev),
|
|
int (*mod_cb)(struct net_device *dev, struct net_device *orig_dev,
|
|
unsigned long event, const void *ctx,
|
|
const struct switchdev_notifier_fdb_info *fdb_info))
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
switchdev_handle_port_obj_add(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
int (*add_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj,
|
|
struct netlink_ext_ack *extack))
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int switchdev_handle_port_obj_add_foreign(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
bool (*foreign_dev_check_cb)(const struct net_device *dev,
|
|
const struct net_device *foreign_dev),
|
|
int (*add_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj,
|
|
struct netlink_ext_ack *extack))
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
switchdev_handle_port_obj_del(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
int (*del_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj))
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
switchdev_handle_port_obj_del_foreign(struct net_device *dev,
|
|
struct switchdev_notifier_port_obj_info *port_obj_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
bool (*foreign_dev_check_cb)(const struct net_device *dev,
|
|
const struct net_device *foreign_dev),
|
|
int (*del_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_obj *obj))
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
switchdev_handle_port_attr_set(struct net_device *dev,
|
|
struct switchdev_notifier_port_attr_info *port_attr_info,
|
|
bool (*check_cb)(const struct net_device *dev),
|
|
int (*set_cb)(struct net_device *dev, const void *ctx,
|
|
const struct switchdev_attr *attr,
|
|
struct netlink_ext_ack *extack))
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#endif /* _LINUX_SWITCHDEV_H_ */
|