2020-05-20 22:20:53 +03:00
/* SPDX-License-Identifier: GPL-2.0 */
/* Copyright(c) 2020 Intel Corporation. */
# ifndef XSK_BUFF_POOL_H_
# define XSK_BUFF_POOL_H_
2020-05-20 22:21:02 +03:00
# include <linux/if_xdp.h>
2020-05-20 22:20:53 +03:00
# include <linux/types.h>
# include <linux/dma-mapping.h>
2021-09-22 10:56:06 +03:00
# include <linux/bpf.h>
2020-05-20 22:20:53 +03:00
# include <net/xdp.h>
struct xsk_buff_pool ;
struct xdp_rxq_info ;
struct xsk_queue ;
struct xdp_desc ;
2020-08-28 11:26:15 +03:00
struct xdp_umem ;
2020-08-28 11:26:17 +03:00
struct xdp_sock ;
2020-05-20 22:20:53 +03:00
struct device ;
struct page ;
struct xdp_buff_xsk {
struct xdp_buff xdp ;
dma_addr_t dma ;
dma_addr_t frame_dma ;
struct xsk_buff_pool * pool ;
u64 orig_addr ;
struct list_head free_list_node ;
} ;
2020-08-28 11:26:22 +03:00
struct xsk_dma_map {
dma_addr_t * dma_pages ;
struct device * dev ;
struct net_device * netdev ;
refcount_t users ;
struct list_head list ; /* Protected by the RTNL_LOCK */
u32 dma_pages_cnt ;
bool dma_need_sync ;
} ;
2020-05-20 22:21:02 +03:00
struct xsk_buff_pool {
2020-08-28 11:26:23 +03:00
/* Members only used in the control path first. */
struct device * dev ;
struct net_device * netdev ;
struct list_head xsk_tx_list ;
/* Protects modifications to the xsk_tx_list */
spinlock_t xsk_tx_list_lock ;
refcount_t users ;
struct xdp_umem * umem ;
struct work_struct work ;
2020-05-20 22:21:02 +03:00
struct list_head free_list ;
2020-08-28 11:26:23 +03:00
u32 heads_cnt ;
u16 queue_id ;
/* Data path members as close to free_heads at the end as possible. */
struct xsk_queue * fq ____cacheline_aligned_in_smp ;
struct xsk_queue * cq ;
2020-08-28 11:26:22 +03:00
/* For performance reasons, each buff pool has its own array of dma_pages
* even when they are identical .
*/
2020-05-20 22:21:02 +03:00
dma_addr_t * dma_pages ;
struct xdp_buff_xsk * heads ;
2022-01-25 19:04:43 +03:00
struct xdp_desc * tx_descs ;
2020-05-20 22:21:02 +03:00
u64 chunk_mask ;
u64 addrs_cnt ;
u32 free_list_cnt ;
u32 dma_pages_cnt ;
u32 free_heads_cnt ;
u32 headroom ;
u32 chunk_size ;
2021-09-22 10:56:06 +03:00
u32 chunk_shift ;
2020-05-20 22:21:02 +03:00
u32 frame_len ;
2020-08-28 11:26:19 +03:00
u8 cached_need_wakeup ;
bool uses_need_wakeup ;
2020-06-29 16:03:57 +03:00
bool dma_need_sync ;
2020-05-20 22:21:02 +03:00
bool unaligned ;
void * addrs ;
2020-12-18 16:45:24 +03:00
/* Mutual exclusion of the completion ring in the SKB mode. Two cases to protect:
* NAPI TX thread and sendmsg error paths in the SKB destructor callback and when
* sockets share a single cq when the same netdev and queue id is shared .
*/
spinlock_t cq_lock ;
2020-05-20 22:21:02 +03:00
struct xdp_buff_xsk * free_heads [ ] ;
} ;
2021-09-22 10:56:06 +03:00
/* Masks for xdp_umem_page flags.
* The low 12 - bits of the addr will be 0 since this is the page address , so we
* can use them for flags .
*/
# define XSK_NEXT_PG_CONTIG_SHIFT 0
# define XSK_NEXT_PG_CONTIG_MASK BIT_ULL(XSK_NEXT_PG_CONTIG_SHIFT)
2020-05-20 22:20:53 +03:00
/* AF_XDP core. */
2020-08-28 11:26:17 +03:00
struct xsk_buff_pool * xp_create_and_assign_umem ( struct xdp_sock * xs ,
struct xdp_umem * umem ) ;
int xp_assign_dev ( struct xsk_buff_pool * pool , struct net_device * dev ,
u16 queue_id , u16 flags ) ;
2022-09-21 16:57:01 +03:00
int xp_assign_dev_shared ( struct xsk_buff_pool * pool , struct xdp_sock * umem_xs ,
2020-08-28 11:26:25 +03:00
struct net_device * dev , u16 queue_id ) ;
2022-04-25 18:37:45 +03:00
int xp_alloc_tx_descs ( struct xsk_buff_pool * pool , struct xdp_sock * xs ) ;
2020-05-20 22:20:53 +03:00
void xp_destroy ( struct xsk_buff_pool * pool ) ;
2020-08-28 11:26:17 +03:00
void xp_get_pool ( struct xsk_buff_pool * pool ) ;
2020-10-27 15:32:01 +03:00
bool xp_put_pool ( struct xsk_buff_pool * pool ) ;
2020-08-28 11:26:17 +03:00
void xp_clear_dev ( struct xsk_buff_pool * pool ) ;
2020-08-28 11:26:20 +03:00
void xp_add_xsk ( struct xsk_buff_pool * pool , struct xdp_sock * xs ) ;
void xp_del_xsk ( struct xsk_buff_pool * pool , struct xdp_sock * xs ) ;
2020-05-20 22:20:53 +03:00
/* AF_XDP, and XDP core. */
void xp_free ( struct xdp_buff_xsk * xskb ) ;
2021-09-22 10:56:06 +03:00
static inline void xp_init_xskb_addr ( struct xdp_buff_xsk * xskb , struct xsk_buff_pool * pool ,
u64 addr )
{
xskb - > orig_addr = addr ;
xskb - > xdp . data_hard_start = pool - > addrs + addr + pool - > headroom ;
}
static inline void xp_init_xskb_dma ( struct xdp_buff_xsk * xskb , struct xsk_buff_pool * pool ,
dma_addr_t * dma_pages , u64 addr )
{
xskb - > frame_dma = ( dma_pages [ addr > > PAGE_SHIFT ] & ~ XSK_NEXT_PG_CONTIG_MASK ) +
( addr & ~ PAGE_MASK ) ;
xskb - > dma = xskb - > frame_dma + pool - > headroom + XDP_PACKET_HEADROOM ;
}
2020-05-20 22:20:53 +03:00
/* AF_XDP ZC drivers, via xdp_sock_buff.h */
void xp_set_rxq_info ( struct xsk_buff_pool * pool , struct xdp_rxq_info * rxq ) ;
int xp_dma_map ( struct xsk_buff_pool * pool , struct device * dev ,
unsigned long attrs , struct page * * pages , u32 nr_pages ) ;
void xp_dma_unmap ( struct xsk_buff_pool * pool , unsigned long attrs ) ;
struct xdp_buff * xp_alloc ( struct xsk_buff_pool * pool ) ;
xsk: Batched buffer allocation for the pool
Add a new driver interface xsk_buff_alloc_batch() offering batched
buffer allocations to improve performance. The new interface takes
three arguments: the buffer pool to allocated from, a pointer to an
array of struct xdp_buff pointers which will contain pointers to the
allocated xdp_buffs, and an unsigned integer specifying the max number
of buffers to allocate. The return value is the actual number of
buffers that the allocator managed to allocate and it will be in the
range 0 <= N <= max, where max is the third parameter to the function.
u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp,
u32 max);
A second driver interface is also introduced that need to be used in
conjunction with xsk_buff_alloc_batch(). It is a helper that sets the
size of struct xdp_buff and is used by the NIC Rx irq routine when
receiving a packet. This helper sets the three struct members data,
data_meta, and data_end. The two first ones is in the xsk_buff_alloc()
case set in the allocation routine and data_end is set when a packet
is received in the receive irq function. This unfortunately leads to
worse performance since the xdp_buff is touched twice with a long time
period in between leading to an extra cache miss. Instead, we fill out
the xdp_buff with all 3 fields at one single point in time in the
driver, when the size of the packet is known. Hence this helper. Note
that the driver has to use this helper (or set all three fields
itself) when using xsk_buff_alloc_batch(). xsk_buff_alloc() works as
before and does not require this.
void xsk_buff_set_size(struct xdp_buff *xdp, u32 size);
Signed-off-by: Magnus Karlsson <magnus.karlsson@intel.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Link: https://lore.kernel.org/bpf/20210922075613.12186-3-magnus.karlsson@gmail.com
2021-09-22 10:56:02 +03:00
u32 xp_alloc_batch ( struct xsk_buff_pool * pool , struct xdp_buff * * xdp , u32 max ) ;
2020-05-20 22:20:53 +03:00
bool xp_can_alloc ( struct xsk_buff_pool * pool , u32 count ) ;
void * xp_raw_get_data ( struct xsk_buff_pool * pool , u64 addr ) ;
dma_addr_t xp_raw_get_dma ( struct xsk_buff_pool * pool , u64 addr ) ;
2020-05-20 22:21:02 +03:00
static inline dma_addr_t xp_get_dma ( struct xdp_buff_xsk * xskb )
{
return xskb - > dma ;
}
static inline dma_addr_t xp_get_frame_dma ( struct xdp_buff_xsk * xskb )
{
return xskb - > frame_dma ;
}
void xp_dma_sync_for_cpu_slow ( struct xdp_buff_xsk * xskb ) ;
static inline void xp_dma_sync_for_cpu ( struct xdp_buff_xsk * xskb )
{
xp_dma_sync_for_cpu_slow ( xskb ) ;
}
void xp_dma_sync_for_device_slow ( struct xsk_buff_pool * pool , dma_addr_t dma ,
size_t size ) ;
static inline void xp_dma_sync_for_device ( struct xsk_buff_pool * pool ,
dma_addr_t dma , size_t size )
{
2020-06-29 16:03:57 +03:00
if ( ! pool - > dma_need_sync )
2020-05-20 22:21:02 +03:00
return ;
xp_dma_sync_for_device_slow ( pool , dma , size ) ;
}
/* Masks for xdp_umem_page flags.
* The low 12 - bits of the addr will be 0 since this is the page address , so we
* can use them for flags .
*/
# define XSK_NEXT_PG_CONTIG_SHIFT 0
# define XSK_NEXT_PG_CONTIG_MASK BIT_ULL(XSK_NEXT_PG_CONTIG_SHIFT)
static inline bool xp_desc_crosses_non_contig_pg ( struct xsk_buff_pool * pool ,
u64 addr , u32 len )
{
bool cross_pg = ( addr & ( PAGE_SIZE - 1 ) ) + len > PAGE_SIZE ;
2021-06-17 12:22:55 +03:00
if ( likely ( ! cross_pg ) )
return false ;
if ( pool - > dma_pages_cnt ) {
2020-05-20 22:21:02 +03:00
return ! ( pool - > dma_pages [ addr > > PAGE_SHIFT ] &
XSK_NEXT_PG_CONTIG_MASK ) ;
}
2021-06-17 12:22:55 +03:00
/* skb path */
return addr + len > pool - > addrs_cnt ;
2020-05-20 22:21:02 +03:00
}
static inline u64 xp_aligned_extract_addr ( struct xsk_buff_pool * pool , u64 addr )
{
return addr & pool - > chunk_mask ;
}
static inline u64 xp_unaligned_extract_addr ( u64 addr )
{
return addr & XSK_UNALIGNED_BUF_ADDR_MASK ;
}
static inline u64 xp_unaligned_extract_offset ( u64 addr )
{
return addr > > XSK_UNALIGNED_BUF_OFFSET_SHIFT ;
}
static inline u64 xp_unaligned_add_offset_to_addr ( u64 addr )
{
return xp_unaligned_extract_addr ( addr ) +
xp_unaligned_extract_offset ( addr ) ;
}
2020-05-20 22:20:53 +03:00
2021-09-22 10:56:06 +03:00
static inline u32 xp_aligned_extract_idx ( struct xsk_buff_pool * pool , u64 addr )
{
return xp_aligned_extract_addr ( pool , addr ) > > pool - > chunk_shift ;
}
static inline void xp_release ( struct xdp_buff_xsk * xskb )
{
if ( xskb - > pool - > unaligned )
xskb - > pool - > free_heads [ xskb - > pool - > free_heads_cnt + + ] = xskb ;
}
static inline u64 xp_get_handle ( struct xdp_buff_xsk * xskb )
{
u64 offset = xskb - > xdp . data - xskb - > xdp . data_hard_start ;
offset + = xskb - > pool - > headroom ;
if ( ! xskb - > pool - > unaligned )
return xskb - > orig_addr + offset ;
return xskb - > orig_addr + ( offset < < XSK_UNALIGNED_BUF_OFFSET_SHIFT ) ;
}
2020-05-20 22:20:53 +03:00
# endif /* XSK_BUFF_POOL_H_ */