eth: link netdev to page_pools in drivers
Link page pool instances to netdev for the drivers which already link to NAPI. Unless the driver is doing something very weird per-NAPI should imply per-netdev. Add netsec as well, Ilias indicates that it fits the mold. Reviewed-by: Eric Dumazet <edumazet@google.com> Acked-by: Jesper Dangaard Brouer <hawk@kernel.org> Signed-off-by: Jakub Kicinski <kuba@kernel.org> Signed-off-by: Paolo Abeni <pabeni@redhat.com>
This commit is contained in:
parent
02b3de80c5
commit
7cc9e6d77f
@ -3331,6 +3331,7 @@ static int bnxt_alloc_rx_page_pool(struct bnxt *bp,
|
||||
pp.pool_size += bp->rx_ring_size;
|
||||
pp.nid = dev_to_node(&bp->pdev->dev);
|
||||
pp.napi = &rxr->bnapi->napi;
|
||||
pp.netdev = bp->dev;
|
||||
pp.dev = &bp->pdev->dev;
|
||||
pp.dma_dir = bp->rx_dir;
|
||||
pp.max_len = PAGE_SIZE;
|
||||
|
@ -902,6 +902,7 @@ static int mlx5e_alloc_rq(struct mlx5e_params *params,
|
||||
pp_params.nid = node;
|
||||
pp_params.dev = rq->pdev;
|
||||
pp_params.napi = rq->cq.napi;
|
||||
pp_params.netdev = rq->netdev;
|
||||
pp_params.dma_dir = rq->buff.map_dir;
|
||||
pp_params.max_len = PAGE_SIZE;
|
||||
|
||||
|
@ -2137,6 +2137,7 @@ static int mana_create_page_pool(struct mana_rxq *rxq, struct gdma_context *gc)
|
||||
pprm.pool_size = RX_BUFFERS_PER_QUEUE;
|
||||
pprm.nid = gc->numa_node;
|
||||
pprm.napi = &rxq->rx_cq.napi;
|
||||
pprm.netdev = rxq->ndev;
|
||||
|
||||
rxq->page_pool = page_pool_create(&pprm);
|
||||
|
||||
|
@ -1302,6 +1302,8 @@ static int netsec_setup_rx_dring(struct netsec_priv *priv)
|
||||
.dma_dir = xdp_prog ? DMA_BIDIRECTIONAL : DMA_FROM_DEVICE,
|
||||
.offset = NETSEC_RXBUF_HEADROOM,
|
||||
.max_len = NETSEC_RX_BUF_SIZE,
|
||||
.napi = &priv->napi,
|
||||
.netdev = priv->ndev,
|
||||
};
|
||||
int i, err;
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user