2021-07-29 05:20:41 +03:00
/* SPDX-License-Identifier: GPL-2.0 */
/*
* Management Component Transport Protocol ( MCTP )
*
* Copyright ( c ) 2021 Code Construct
* Copyright ( c ) 2021 Google
*/
# ifndef __NET_MCTP_H
# define __NET_MCTP_H
# include <linux/bits.h>
2021-07-29 05:20:44 +03:00
# include <linux/mctp.h>
2021-10-26 04:57:28 +03:00
# include <linux/netdevice.h>
2021-07-29 05:20:45 +03:00
# include <net/net_namespace.h>
2021-07-29 05:20:49 +03:00
# include <net/sock.h>
2021-07-29 05:20:41 +03:00
/* MCTP packet definitions */
struct mctp_hdr {
u8 ver ;
u8 dest ;
u8 src ;
u8 flags_seq_tag ;
} ;
# define MCTP_VER_MIN 1
# define MCTP_VER_MAX 1
/* Definitions for flags_seq_tag field */
# define MCTP_HDR_FLAG_SOM BIT(7)
# define MCTP_HDR_FLAG_EOM BIT(6)
# define MCTP_HDR_FLAG_TO BIT(3)
# define MCTP_HDR_FLAGS GENMASK(5, 3)
# define MCTP_HDR_SEQ_SHIFT 4
# define MCTP_HDR_SEQ_MASK GENMASK(1, 0)
# define MCTP_HDR_TAG_SHIFT 0
# define MCTP_HDR_TAG_MASK GENMASK(2, 0)
2021-07-29 05:20:52 +03:00
# define MCTP_INITIAL_DEFAULT_NET 1
2022-02-18 07:25:53 +03:00
static inline bool mctp_address_unicast ( mctp_eid_t eid )
2021-07-29 05:20:44 +03:00
{
return eid > = 8 & & eid < 255 ;
}
2022-02-18 07:25:53 +03:00
static inline bool mctp_address_broadcast ( mctp_eid_t eid )
{
return eid = = 255 ;
}
static inline bool mctp_address_null ( mctp_eid_t eid )
{
return eid = = 0 ;
}
2022-02-09 07:05:55 +03:00
static inline bool mctp_address_matches ( mctp_eid_t match , mctp_eid_t eid )
{
return match = = eid | | match = = MCTP_ADDR_ANY ;
}
2021-07-29 05:20:44 +03:00
static inline struct mctp_hdr * mctp_hdr ( struct sk_buff * skb )
{
return ( struct mctp_hdr * ) skb_network_header ( skb ) ;
}
2021-07-29 05:20:49 +03:00
/* socket implementation */
struct mctp_sock {
struct sock sk ;
/* bind() params */
2021-10-18 06:29:34 +03:00
unsigned int bind_net ;
2021-07-29 05:20:49 +03:00
mctp_eid_t bind_addr ;
__u8 bind_type ;
2021-10-26 04:57:28 +03:00
/* sendmsg()/recvmsg() uses struct sockaddr_mctp_ext */
bool addr_ext ;
2021-07-29 05:20:49 +03:00
/* list of mctp_sk_key, for incoming tag lookup. updates protected
* by sk - > net - > keys_lock
*/
struct hlist_head keys ;
2021-09-29 10:26:09 +03:00
/* mechanism for expiring allocated keys; will release an allocated
* tag , and any netdev state for a request / response pairing
*/
struct timer_list key_expiry ;
2021-07-29 05:20:49 +03:00
} ;
/* Key for matching incoming packets to sockets or reassembly contexts.
* Packets are matched on ( src , dest , tag ) .
*
2021-09-29 10:26:07 +03:00
* Lifetime / locking requirements :
2021-07-29 05:20:49 +03:00
*
2021-09-29 10:26:07 +03:00
* - individual key data ( ie , the struct itself ) is protected by key - > lock ;
* changes must be made with that lock held .
*
* - the lookup fields : peer_addr , local_addr and tag are set before the
* key is added to lookup lists , and never updated .
*
* - A ref to the key must be held ( throuh key - > refs ) if a pointer to the
* key is to be accessed after key - > lock is released .
2021-07-29 05:20:49 +03:00
*
* - a mctp_sk_key contains a reference to a struct sock ; this is valid
* for the life of the key . On sock destruction ( through unhash ) , the key is
2021-09-29 10:26:07 +03:00
* removed from lists ( see below ) , and marked invalid .
2021-07-29 05:20:49 +03:00
*
* - these mctp_sk_keys appear on two lists :
* 1 ) the struct mctp_sock - > keys list
* 2 ) the struct netns_mctp - > keys list
*
2021-09-29 10:26:07 +03:00
* presences on these lists requires a ( single ) refcount to be held ; both
* lists are updated as a single operation .
*
* Updates and lookups in either list are performed under the
* netns_mctp - > keys lock . Lookup functions will need to lock the key and
* take a reference before unlocking the keys_lock . Consequently , the list ' s
* keys_lock * cannot * be acquired with the individual key - > lock held .
2021-07-29 05:20:49 +03:00
*
mctp: Implement message fragmentation & reassembly
This change implements MCTP fragmentation (based on route & device MTU),
and corresponding reassembly.
The MCTP specification only allows for fragmentation on the originating
message endpoint, and reassembly on the destination endpoint -
intermediate nodes do not need to reassemble/refragment. Consequently,
we only fragment in the local transmit path, and reassemble
locally-bound packets. Messages are required to be in-order, so we
simply cancel reassembly on out-of-order or missing packets.
In the fragmentation path, we just break up the message into MTU-sized
fragments; the skb structure is a simple copy for now, which we can later
improve with a shared data implementation.
For reassembly, we keep track of incoming message fragments using the
existing tag infrastructure, allocating a key on the (src,dest,tag)
tuple, and reassembles matching fragments into a skb->frag_list.
Signed-off-by: Jeremy Kerr <jk@codeconstruct.com.au>
Signed-off-by: David S. Miller <davem@davemloft.net>
2021-07-29 05:20:50 +03:00
* - a key may have a sk_buff attached as part of an in - progress message
2021-09-29 10:26:07 +03:00
* reassembly ( - > reasm_head ) . The reasm data is protected by the individual
* key - > lock .
mctp: Implement message fragmentation & reassembly
This change implements MCTP fragmentation (based on route & device MTU),
and corresponding reassembly.
The MCTP specification only allows for fragmentation on the originating
message endpoint, and reassembly on the destination endpoint -
intermediate nodes do not need to reassemble/refragment. Consequently,
we only fragment in the local transmit path, and reassemble
locally-bound packets. Messages are required to be in-order, so we
simply cancel reassembly on out-of-order or missing packets.
In the fragmentation path, we just break up the message into MTU-sized
fragments; the skb structure is a simple copy for now, which we can later
improve with a shared data implementation.
For reassembly, we keep track of incoming message fragments using the
existing tag infrastructure, allocating a key on the (src,dest,tag)
tuple, and reassembles matching fragments into a skb->frag_list.
Signed-off-by: Jeremy Kerr <jk@codeconstruct.com.au>
Signed-off-by: David S. Miller <davem@davemloft.net>
2021-07-29 05:20:50 +03:00
*
* - there are two destruction paths for a mctp_sk_key :
*
* - through socket unhash ( see mctp_sk_unhash ) . This performs the list
* removal under keys_lock .
*
* - where a key is established to receive a reply message : after receiving
* the ( complete ) reply , or during reassembly errors . Here , we clean up
* the reassembly context ( marking reasm_dead , to prevent another from
* starting ) , and remove the socket from the netns & socket lists .
2021-09-29 10:26:09 +03:00
*
* - through an expiry timeout , on a per - socket timer
2021-07-29 05:20:49 +03:00
*/
struct mctp_sk_key {
mctp_eid_t peer_addr ;
2022-02-09 07:05:57 +03:00
mctp_eid_t local_addr ; /* MCTP_ADDR_ANY for local owned tags */
2021-07-29 05:20:49 +03:00
__u8 tag ; /* incoming tag match; invert TO for local */
/* we hold a ref to sk when set */
struct sock * sk ;
/* routing lookup list */
struct hlist_node hlist ;
/* per-socket list */
struct hlist_node sklist ;
2021-09-29 10:26:07 +03:00
/* lock protects against concurrent updates to the reassembly and
* expiry data below .
*/
spinlock_t lock ;
/* Keys are referenced during the output path, which may sleep */
refcount_t refs ;
mctp: Implement message fragmentation & reassembly
This change implements MCTP fragmentation (based on route & device MTU),
and corresponding reassembly.
The MCTP specification only allows for fragmentation on the originating
message endpoint, and reassembly on the destination endpoint -
intermediate nodes do not need to reassemble/refragment. Consequently,
we only fragment in the local transmit path, and reassemble
locally-bound packets. Messages are required to be in-order, so we
simply cancel reassembly on out-of-order or missing packets.
In the fragmentation path, we just break up the message into MTU-sized
fragments; the skb structure is a simple copy for now, which we can later
improve with a shared data implementation.
For reassembly, we keep track of incoming message fragments using the
existing tag infrastructure, allocating a key on the (src,dest,tag)
tuple, and reassembles matching fragments into a skb->frag_list.
Signed-off-by: Jeremy Kerr <jk@codeconstruct.com.au>
Signed-off-by: David S. Miller <davem@davemloft.net>
2021-07-29 05:20:50 +03:00
/* incoming fragment reassembly context */
struct sk_buff * reasm_head ;
struct sk_buff * * reasm_tailp ;
bool reasm_dead ;
u8 last_seq ;
2021-09-29 10:26:07 +03:00
/* key validity */
bool valid ;
2021-09-29 10:26:09 +03:00
/* expiry timeout; valid (above) cleared on expiry */
unsigned long expiry ;
2021-10-29 06:01:45 +03:00
/* free to use for device flow state tracking. Initialised to
* zero on initial key creation
*/
unsigned long dev_flow_state ;
struct mctp_dev * dev ;
2022-02-09 07:05:57 +03:00
/* a tag allocated with SIOCMCTPALLOCTAG ioctl will not expire
* automatically on timeout or response , instead SIOCMCTPDROPTAG
* is used .
*/
bool manual_alloc ;
2021-07-29 05:20:49 +03:00
} ;
2021-07-29 05:20:45 +03:00
struct mctp_skb_cb {
unsigned int magic ;
unsigned int net ;
2021-10-26 04:57:28 +03:00
int ifindex ; /* extended/direct addressing if set */
2021-07-29 05:20:45 +03:00
mctp_eid_t src ;
2021-10-26 04:57:28 +03:00
unsigned char halen ;
unsigned char haddr [ MAX_ADDR_LEN ] ;
2021-07-29 05:20:45 +03:00
} ;
/* skb control-block accessors with a little extra debugging for initial
* development .
*
* TODO : remove checks & mctp_skb_cb - > magic ; replace callers of __mctp_cb
* with mctp_cb ( ) .
*
* __mctp_cb ( ) is only for the initial ingress code ; we should see - > magic set
* at all times after this .
*/
static inline struct mctp_skb_cb * __mctp_cb ( struct sk_buff * skb )
{
struct mctp_skb_cb * cb = ( void * ) skb - > cb ;
cb - > magic = 0x4d435450 ;
return cb ;
}
static inline struct mctp_skb_cb * mctp_cb ( struct sk_buff * skb )
{
struct mctp_skb_cb * cb = ( void * ) skb - > cb ;
2021-10-26 04:57:28 +03:00
BUILD_BUG_ON ( sizeof ( struct mctp_skb_cb ) > sizeof ( skb - > cb ) ) ;
2021-07-29 05:20:45 +03:00
WARN_ON ( cb - > magic ! = 0x4d435450 ) ;
return ( void * ) ( skb - > cb ) ;
}
2021-10-29 06:01:44 +03:00
/* If CONFIG_MCTP_FLOWS, we may add one of these as a SKB extension,
* indicating the flow to the device driver .
*/
struct mctp_flow {
struct mctp_sk_key * key ;
} ;
2021-07-29 05:20:45 +03:00
/* Route definition.
*
* These are held in the pernet - > mctp . routes list , with RCU protection for
* removed routes . We hold a reference to the netdev ; routes need to be
* dropped on NETDEV_UNREGISTER events .
*
* Updates to the route table are performed under rtnl ; all reads under RCU ,
* so routes cannot be referenced over a RCU grace period . Specifically : A
2021-10-26 04:57:28 +03:00
* caller cannot block between mctp_route_lookup and mctp_route_release ( )
2021-07-29 05:20:45 +03:00
*/
struct mctp_route {
mctp_eid_t min , max ;
struct mctp_dev * dev ;
unsigned int mtu ;
2021-08-10 05:38:34 +03:00
unsigned char type ;
2021-07-29 05:20:45 +03:00
int ( * output ) ( struct mctp_route * route ,
struct sk_buff * skb ) ;
struct list_head list ;
refcount_t refs ;
struct rcu_head rcu ;
} ;
/* route interfaces */
struct mctp_route * mctp_route_lookup ( struct net * net , unsigned int dnet ,
mctp_eid_t daddr ) ;
int mctp_local_output ( struct sock * sk , struct mctp_route * rt ,
struct sk_buff * skb , mctp_eid_t daddr , u8 req_tag ) ;
2021-09-29 10:26:07 +03:00
void mctp_key_unref ( struct mctp_sk_key * key ) ;
2022-02-09 07:05:57 +03:00
struct mctp_sk_key * mctp_alloc_local_tag ( struct mctp_sock * msk ,
mctp_eid_t daddr , mctp_eid_t saddr ,
bool manual , u8 * tagp ) ;
2021-07-29 05:20:45 +03:00
/* routing <--> device interface */
2021-07-29 05:20:46 +03:00
unsigned int mctp_default_net ( struct net * net ) ;
int mctp_default_net_set ( struct net * net , unsigned int index ) ;
2021-07-29 05:20:45 +03:00
int mctp_route_add_local ( struct mctp_dev * mdev , mctp_eid_t addr ) ;
int mctp_route_remove_local ( struct mctp_dev * mdev , mctp_eid_t addr ) ;
void mctp_route_remove_dev ( struct mctp_dev * mdev ) ;
2021-07-29 05:20:47 +03:00
/* neighbour definitions */
enum mctp_neigh_source {
MCTP_NEIGH_STATIC ,
MCTP_NEIGH_DISCOVER ,
} ;
struct mctp_neigh {
struct mctp_dev * dev ;
mctp_eid_t eid ;
enum mctp_neigh_source source ;
unsigned char ha [ MAX_ADDR_LEN ] ;
struct list_head list ;
struct rcu_head rcu ;
} ;
int mctp_neigh_init ( void ) ;
void mctp_neigh_exit ( void ) ;
// ret_hwaddr may be NULL, otherwise must have space for MAX_ADDR_LEN
int mctp_neigh_lookup ( struct mctp_dev * dev , mctp_eid_t eid ,
void * ret_hwaddr ) ;
void mctp_neigh_remove_dev ( struct mctp_dev * mdev ) ;
2021-07-29 05:20:45 +03:00
int mctp_routes_init ( void ) ;
void mctp_routes_exit ( void ) ;
2021-07-29 05:20:44 +03:00
void mctp_device_init ( void ) ;
void mctp_device_exit ( void ) ;
2021-07-29 05:20:41 +03:00
# endif /* __NET_MCTP_H */