mvpp2: use page_pool allocator
Use the page_pool API for memory management. This is a prerequisite for native XDP support. Tested-by: Sven Auhagen <sven.auhagen@voleatech.de> Signed-off-by: Matteo Croce <mcroce@microsoft.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
136bcd8425
commit
b27db2274b
@ -87,6 +87,7 @@ config MVPP2
|
|||||||
depends on ARCH_MVEBU || COMPILE_TEST
|
depends on ARCH_MVEBU || COMPILE_TEST
|
||||||
select MVMDIO
|
select MVMDIO
|
||||||
select PHYLINK
|
select PHYLINK
|
||||||
|
select PAGE_POOL
|
||||||
help
|
help
|
||||||
This driver supports the network interface units in the
|
This driver supports the network interface units in the
|
||||||
Marvell ARMADA 375, 7K and 8K SoCs.
|
Marvell ARMADA 375, 7K and 8K SoCs.
|
||||||
|
@ -15,6 +15,7 @@
|
|||||||
#include <linux/phy.h>
|
#include <linux/phy.h>
|
||||||
#include <linux/phylink.h>
|
#include <linux/phylink.h>
|
||||||
#include <net/flow_offload.h>
|
#include <net/flow_offload.h>
|
||||||
|
#include <net/page_pool.h>
|
||||||
|
|
||||||
/* Fifo Registers */
|
/* Fifo Registers */
|
||||||
#define MVPP2_RX_DATA_FIFO_SIZE_REG(port) (0x00 + 4 * (port))
|
#define MVPP2_RX_DATA_FIFO_SIZE_REG(port) (0x00 + 4 * (port))
|
||||||
@ -820,6 +821,9 @@ struct mvpp2 {
|
|||||||
|
|
||||||
/* RSS Indirection tables */
|
/* RSS Indirection tables */
|
||||||
struct mvpp2_rss_table *rss_tables[MVPP22_N_RSS_TABLES];
|
struct mvpp2_rss_table *rss_tables[MVPP22_N_RSS_TABLES];
|
||||||
|
|
||||||
|
/* page_pool allocator */
|
||||||
|
struct page_pool *page_pool[MVPP2_PORT_MAX_RXQ];
|
||||||
};
|
};
|
||||||
|
|
||||||
struct mvpp2_pcpu_stats {
|
struct mvpp2_pcpu_stats {
|
||||||
@ -1161,6 +1165,10 @@ struct mvpp2_rx_queue {
|
|||||||
|
|
||||||
/* Port's logic RXQ number to which physical RXQ is mapped */
|
/* Port's logic RXQ number to which physical RXQ is mapped */
|
||||||
int logic_rxq;
|
int logic_rxq;
|
||||||
|
|
||||||
|
/* XDP memory accounting */
|
||||||
|
struct xdp_rxq_info xdp_rxq_short;
|
||||||
|
struct xdp_rxq_info xdp_rxq_long;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct mvpp2_bm_pool {
|
struct mvpp2_bm_pool {
|
||||||
|
@ -95,6 +95,22 @@ static inline u32 mvpp2_cpu_to_thread(struct mvpp2 *priv, int cpu)
|
|||||||
return cpu % priv->nthreads;
|
return cpu % priv->nthreads;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static struct page_pool *
|
||||||
|
mvpp2_create_page_pool(struct device *dev, int num, int len)
|
||||||
|
{
|
||||||
|
struct page_pool_params pp_params = {
|
||||||
|
/* internal DMA mapping in page_pool */
|
||||||
|
.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV,
|
||||||
|
.pool_size = num,
|
||||||
|
.nid = NUMA_NO_NODE,
|
||||||
|
.dev = dev,
|
||||||
|
.dma_dir = DMA_FROM_DEVICE,
|
||||||
|
.max_len = len,
|
||||||
|
};
|
||||||
|
|
||||||
|
return page_pool_create(&pp_params);
|
||||||
|
}
|
||||||
|
|
||||||
/* These accessors should be used to access:
|
/* These accessors should be used to access:
|
||||||
*
|
*
|
||||||
* - per-thread registers, where each thread has its own copy of the
|
* - per-thread registers, where each thread has its own copy of the
|
||||||
@ -327,17 +343,25 @@ static inline int mvpp2_txq_phys(int port, int txq)
|
|||||||
return (MVPP2_MAX_TCONT + port) * MVPP2_MAX_TXQ + txq;
|
return (MVPP2_MAX_TCONT + port) * MVPP2_MAX_TXQ + txq;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void *mvpp2_frag_alloc(const struct mvpp2_bm_pool *pool)
|
/* Returns a struct page if page_pool is set, otherwise a buffer */
|
||||||
|
static void *mvpp2_frag_alloc(const struct mvpp2_bm_pool *pool,
|
||||||
|
struct page_pool *page_pool)
|
||||||
{
|
{
|
||||||
|
if (page_pool)
|
||||||
|
return page_pool_dev_alloc_pages(page_pool);
|
||||||
|
|
||||||
if (likely(pool->frag_size <= PAGE_SIZE))
|
if (likely(pool->frag_size <= PAGE_SIZE))
|
||||||
return netdev_alloc_frag(pool->frag_size);
|
return netdev_alloc_frag(pool->frag_size);
|
||||||
else
|
|
||||||
return kmalloc(pool->frag_size, GFP_ATOMIC);
|
return kmalloc(pool->frag_size, GFP_ATOMIC);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void mvpp2_frag_free(const struct mvpp2_bm_pool *pool, void *data)
|
static void mvpp2_frag_free(const struct mvpp2_bm_pool *pool,
|
||||||
|
struct page_pool *page_pool, void *data)
|
||||||
{
|
{
|
||||||
if (likely(pool->frag_size <= PAGE_SIZE))
|
if (page_pool)
|
||||||
|
page_pool_put_full_page(page_pool, virt_to_head_page(data), false);
|
||||||
|
else if (likely(pool->frag_size <= PAGE_SIZE))
|
||||||
skb_free_frag(data);
|
skb_free_frag(data);
|
||||||
else
|
else
|
||||||
kfree(data);
|
kfree(data);
|
||||||
@ -442,6 +466,7 @@ static void mvpp2_bm_bufs_get_addrs(struct device *dev, struct mvpp2 *priv,
|
|||||||
static void mvpp2_bm_bufs_free(struct device *dev, struct mvpp2 *priv,
|
static void mvpp2_bm_bufs_free(struct device *dev, struct mvpp2 *priv,
|
||||||
struct mvpp2_bm_pool *bm_pool, int buf_num)
|
struct mvpp2_bm_pool *bm_pool, int buf_num)
|
||||||
{
|
{
|
||||||
|
struct page_pool *pp = NULL;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
if (buf_num > bm_pool->buf_num) {
|
if (buf_num > bm_pool->buf_num) {
|
||||||
@ -450,6 +475,9 @@ static void mvpp2_bm_bufs_free(struct device *dev, struct mvpp2 *priv,
|
|||||||
buf_num = bm_pool->buf_num;
|
buf_num = bm_pool->buf_num;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (priv->percpu_pools)
|
||||||
|
pp = priv->page_pool[bm_pool->id];
|
||||||
|
|
||||||
for (i = 0; i < buf_num; i++) {
|
for (i = 0; i < buf_num; i++) {
|
||||||
dma_addr_t buf_dma_addr;
|
dma_addr_t buf_dma_addr;
|
||||||
phys_addr_t buf_phys_addr;
|
phys_addr_t buf_phys_addr;
|
||||||
@ -458,14 +486,15 @@ static void mvpp2_bm_bufs_free(struct device *dev, struct mvpp2 *priv,
|
|||||||
mvpp2_bm_bufs_get_addrs(dev, priv, bm_pool,
|
mvpp2_bm_bufs_get_addrs(dev, priv, bm_pool,
|
||||||
&buf_dma_addr, &buf_phys_addr);
|
&buf_dma_addr, &buf_phys_addr);
|
||||||
|
|
||||||
dma_unmap_single(dev, buf_dma_addr,
|
if (!pp)
|
||||||
bm_pool->buf_size, DMA_FROM_DEVICE);
|
dma_unmap_single(dev, buf_dma_addr,
|
||||||
|
bm_pool->buf_size, DMA_FROM_DEVICE);
|
||||||
|
|
||||||
data = (void *)phys_to_virt(buf_phys_addr);
|
data = (void *)phys_to_virt(buf_phys_addr);
|
||||||
if (!data)
|
if (!data)
|
||||||
break;
|
break;
|
||||||
|
|
||||||
mvpp2_frag_free(bm_pool, data);
|
mvpp2_frag_free(bm_pool, pp, data);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Update BM driver with number of buffers removed from pool */
|
/* Update BM driver with number of buffers removed from pool */
|
||||||
@ -496,6 +525,9 @@ static int mvpp2_bm_pool_destroy(struct device *dev, struct mvpp2 *priv,
|
|||||||
int buf_num;
|
int buf_num;
|
||||||
u32 val;
|
u32 val;
|
||||||
|
|
||||||
|
if (priv->percpu_pools)
|
||||||
|
page_pool_destroy(priv->page_pool[bm_pool->id]);
|
||||||
|
|
||||||
buf_num = mvpp2_check_hw_buf_num(priv, bm_pool);
|
buf_num = mvpp2_check_hw_buf_num(priv, bm_pool);
|
||||||
mvpp2_bm_bufs_free(dev, priv, bm_pool, buf_num);
|
mvpp2_bm_bufs_free(dev, priv, bm_pool, buf_num);
|
||||||
|
|
||||||
@ -548,8 +580,20 @@ static int mvpp2_bm_init(struct device *dev, struct mvpp2 *priv)
|
|||||||
{
|
{
|
||||||
int i, err, poolnum = MVPP2_BM_POOLS_NUM;
|
int i, err, poolnum = MVPP2_BM_POOLS_NUM;
|
||||||
|
|
||||||
if (priv->percpu_pools)
|
if (priv->percpu_pools) {
|
||||||
poolnum = mvpp2_get_nrxqs(priv) * 2;
|
poolnum = mvpp2_get_nrxqs(priv) * 2;
|
||||||
|
for (i = 0; i < poolnum; i++) {
|
||||||
|
/* the pool in use */
|
||||||
|
int pn = i / (poolnum / 2);
|
||||||
|
|
||||||
|
priv->page_pool[i] =
|
||||||
|
mvpp2_create_page_pool(dev,
|
||||||
|
mvpp2_pools[pn].buf_num,
|
||||||
|
mvpp2_pools[pn].pkt_size);
|
||||||
|
if (IS_ERR(priv->page_pool[i]))
|
||||||
|
return PTR_ERR(priv->page_pool[i]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
dev_info(dev, "using %d %s buffers\n", poolnum,
|
dev_info(dev, "using %d %s buffers\n", poolnum,
|
||||||
priv->percpu_pools ? "per-cpu" : "shared");
|
priv->percpu_pools ? "per-cpu" : "shared");
|
||||||
@ -632,23 +676,31 @@ static void mvpp2_rxq_short_pool_set(struct mvpp2_port *port,
|
|||||||
|
|
||||||
static void *mvpp2_buf_alloc(struct mvpp2_port *port,
|
static void *mvpp2_buf_alloc(struct mvpp2_port *port,
|
||||||
struct mvpp2_bm_pool *bm_pool,
|
struct mvpp2_bm_pool *bm_pool,
|
||||||
|
struct page_pool *page_pool,
|
||||||
dma_addr_t *buf_dma_addr,
|
dma_addr_t *buf_dma_addr,
|
||||||
phys_addr_t *buf_phys_addr,
|
phys_addr_t *buf_phys_addr,
|
||||||
gfp_t gfp_mask)
|
gfp_t gfp_mask)
|
||||||
{
|
{
|
||||||
dma_addr_t dma_addr;
|
dma_addr_t dma_addr;
|
||||||
|
struct page *page;
|
||||||
void *data;
|
void *data;
|
||||||
|
|
||||||
data = mvpp2_frag_alloc(bm_pool);
|
data = mvpp2_frag_alloc(bm_pool, page_pool);
|
||||||
if (!data)
|
if (!data)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
dma_addr = dma_map_single(port->dev->dev.parent, data,
|
if (page_pool) {
|
||||||
MVPP2_RX_BUF_SIZE(bm_pool->pkt_size),
|
page = (struct page *)data;
|
||||||
DMA_FROM_DEVICE);
|
dma_addr = page_pool_get_dma_addr(page);
|
||||||
if (unlikely(dma_mapping_error(port->dev->dev.parent, dma_addr))) {
|
data = page_to_virt(page);
|
||||||
mvpp2_frag_free(bm_pool, data);
|
} else {
|
||||||
return NULL;
|
dma_addr = dma_map_single(port->dev->dev.parent, data,
|
||||||
|
MVPP2_RX_BUF_SIZE(bm_pool->pkt_size),
|
||||||
|
DMA_FROM_DEVICE);
|
||||||
|
if (unlikely(dma_mapping_error(port->dev->dev.parent, dma_addr))) {
|
||||||
|
mvpp2_frag_free(bm_pool, NULL, data);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
*buf_dma_addr = dma_addr;
|
*buf_dma_addr = dma_addr;
|
||||||
*buf_phys_addr = virt_to_phys(data);
|
*buf_phys_addr = virt_to_phys(data);
|
||||||
@ -706,6 +758,7 @@ static int mvpp2_bm_bufs_add(struct mvpp2_port *port,
|
|||||||
int i, buf_size, total_size;
|
int i, buf_size, total_size;
|
||||||
dma_addr_t dma_addr;
|
dma_addr_t dma_addr;
|
||||||
phys_addr_t phys_addr;
|
phys_addr_t phys_addr;
|
||||||
|
struct page_pool *pp = NULL;
|
||||||
void *buf;
|
void *buf;
|
||||||
|
|
||||||
if (port->priv->percpu_pools &&
|
if (port->priv->percpu_pools &&
|
||||||
@ -726,8 +779,10 @@ static int mvpp2_bm_bufs_add(struct mvpp2_port *port,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (port->priv->percpu_pools)
|
||||||
|
pp = port->priv->page_pool[bm_pool->id];
|
||||||
for (i = 0; i < buf_num; i++) {
|
for (i = 0; i < buf_num; i++) {
|
||||||
buf = mvpp2_buf_alloc(port, bm_pool, &dma_addr,
|
buf = mvpp2_buf_alloc(port, bm_pool, pp, &dma_addr,
|
||||||
&phys_addr, GFP_KERNEL);
|
&phys_addr, GFP_KERNEL);
|
||||||
if (!buf)
|
if (!buf)
|
||||||
break;
|
break;
|
||||||
@ -2374,10 +2429,11 @@ static int mvpp2_aggr_txq_init(struct platform_device *pdev,
|
|||||||
/* Create a specified Rx queue */
|
/* Create a specified Rx queue */
|
||||||
static int mvpp2_rxq_init(struct mvpp2_port *port,
|
static int mvpp2_rxq_init(struct mvpp2_port *port,
|
||||||
struct mvpp2_rx_queue *rxq)
|
struct mvpp2_rx_queue *rxq)
|
||||||
|
|
||||||
{
|
{
|
||||||
|
struct mvpp2 *priv = port->priv;
|
||||||
unsigned int thread;
|
unsigned int thread;
|
||||||
u32 rxq_dma;
|
u32 rxq_dma;
|
||||||
|
int err;
|
||||||
|
|
||||||
rxq->size = port->rx_ring_size;
|
rxq->size = port->rx_ring_size;
|
||||||
|
|
||||||
@ -2415,7 +2471,43 @@ static int mvpp2_rxq_init(struct mvpp2_port *port,
|
|||||||
/* Add number of descriptors ready for receiving packets */
|
/* Add number of descriptors ready for receiving packets */
|
||||||
mvpp2_rxq_status_update(port, rxq->id, 0, rxq->size);
|
mvpp2_rxq_status_update(port, rxq->id, 0, rxq->size);
|
||||||
|
|
||||||
|
if (priv->percpu_pools) {
|
||||||
|
err = xdp_rxq_info_reg(&rxq->xdp_rxq_short, port->dev, rxq->id);
|
||||||
|
if (err < 0)
|
||||||
|
goto err_free_dma;
|
||||||
|
|
||||||
|
err = xdp_rxq_info_reg(&rxq->xdp_rxq_long, port->dev, rxq->id);
|
||||||
|
if (err < 0)
|
||||||
|
goto err_unregister_rxq_short;
|
||||||
|
|
||||||
|
/* Every RXQ has a pool for short and another for long packets */
|
||||||
|
err = xdp_rxq_info_reg_mem_model(&rxq->xdp_rxq_short,
|
||||||
|
MEM_TYPE_PAGE_POOL,
|
||||||
|
priv->page_pool[rxq->logic_rxq]);
|
||||||
|
if (err < 0)
|
||||||
|
goto err_unregister_rxq_long;
|
||||||
|
|
||||||
|
err = xdp_rxq_info_reg_mem_model(&rxq->xdp_rxq_long,
|
||||||
|
MEM_TYPE_PAGE_POOL,
|
||||||
|
priv->page_pool[rxq->logic_rxq +
|
||||||
|
port->nrxqs]);
|
||||||
|
if (err < 0)
|
||||||
|
goto err_unregister_mem_rxq_short;
|
||||||
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
err_unregister_mem_rxq_short:
|
||||||
|
xdp_rxq_info_unreg_mem_model(&rxq->xdp_rxq_short);
|
||||||
|
err_unregister_rxq_long:
|
||||||
|
xdp_rxq_info_unreg(&rxq->xdp_rxq_long);
|
||||||
|
err_unregister_rxq_short:
|
||||||
|
xdp_rxq_info_unreg(&rxq->xdp_rxq_short);
|
||||||
|
err_free_dma:
|
||||||
|
dma_free_coherent(port->dev->dev.parent,
|
||||||
|
rxq->size * MVPP2_DESC_ALIGNED_SIZE,
|
||||||
|
rxq->descs, rxq->descs_dma);
|
||||||
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Push packets received by the RXQ to BM pool */
|
/* Push packets received by the RXQ to BM pool */
|
||||||
@ -2449,6 +2541,12 @@ static void mvpp2_rxq_deinit(struct mvpp2_port *port,
|
|||||||
{
|
{
|
||||||
unsigned int thread;
|
unsigned int thread;
|
||||||
|
|
||||||
|
if (xdp_rxq_info_is_reg(&rxq->xdp_rxq_short))
|
||||||
|
xdp_rxq_info_unreg(&rxq->xdp_rxq_short);
|
||||||
|
|
||||||
|
if (xdp_rxq_info_is_reg(&rxq->xdp_rxq_long))
|
||||||
|
xdp_rxq_info_unreg(&rxq->xdp_rxq_long);
|
||||||
|
|
||||||
mvpp2_rxq_drop_pkts(port, rxq);
|
mvpp2_rxq_drop_pkts(port, rxq);
|
||||||
|
|
||||||
if (rxq->descs)
|
if (rxq->descs)
|
||||||
@ -2890,14 +2988,15 @@ static void mvpp2_rx_csum(struct mvpp2_port *port, u32 status,
|
|||||||
|
|
||||||
/* Allocate a new skb and add it to BM pool */
|
/* Allocate a new skb and add it to BM pool */
|
||||||
static int mvpp2_rx_refill(struct mvpp2_port *port,
|
static int mvpp2_rx_refill(struct mvpp2_port *port,
|
||||||
struct mvpp2_bm_pool *bm_pool, int pool)
|
struct mvpp2_bm_pool *bm_pool,
|
||||||
|
struct page_pool *page_pool, int pool)
|
||||||
{
|
{
|
||||||
dma_addr_t dma_addr;
|
dma_addr_t dma_addr;
|
||||||
phys_addr_t phys_addr;
|
phys_addr_t phys_addr;
|
||||||
void *buf;
|
void *buf;
|
||||||
|
|
||||||
buf = mvpp2_buf_alloc(port, bm_pool, &dma_addr, &phys_addr,
|
buf = mvpp2_buf_alloc(port, bm_pool, page_pool,
|
||||||
GFP_ATOMIC);
|
&dma_addr, &phys_addr, GFP_ATOMIC);
|
||||||
if (!buf)
|
if (!buf)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
@ -2956,6 +3055,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi,
|
|||||||
while (rx_done < rx_todo) {
|
while (rx_done < rx_todo) {
|
||||||
struct mvpp2_rx_desc *rx_desc = mvpp2_rxq_next_desc_get(rxq);
|
struct mvpp2_rx_desc *rx_desc = mvpp2_rxq_next_desc_get(rxq);
|
||||||
struct mvpp2_bm_pool *bm_pool;
|
struct mvpp2_bm_pool *bm_pool;
|
||||||
|
struct page_pool *pp = NULL;
|
||||||
struct sk_buff *skb;
|
struct sk_buff *skb;
|
||||||
unsigned int frag_size;
|
unsigned int frag_size;
|
||||||
dma_addr_t dma_addr;
|
dma_addr_t dma_addr;
|
||||||
@ -2989,6 +3089,9 @@ static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi,
|
|||||||
DMA_FROM_DEVICE);
|
DMA_FROM_DEVICE);
|
||||||
prefetch(data);
|
prefetch(data);
|
||||||
|
|
||||||
|
if (port->priv->percpu_pools)
|
||||||
|
pp = port->priv->page_pool[pool];
|
||||||
|
|
||||||
if (bm_pool->frag_size > PAGE_SIZE)
|
if (bm_pool->frag_size > PAGE_SIZE)
|
||||||
frag_size = 0;
|
frag_size = 0;
|
||||||
else
|
else
|
||||||
@ -3000,15 +3103,18 @@ static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi,
|
|||||||
goto err_drop_frame;
|
goto err_drop_frame;
|
||||||
}
|
}
|
||||||
|
|
||||||
err = mvpp2_rx_refill(port, bm_pool, pool);
|
err = mvpp2_rx_refill(port, bm_pool, pp, pool);
|
||||||
if (err) {
|
if (err) {
|
||||||
netdev_err(port->dev, "failed to refill BM pools\n");
|
netdev_err(port->dev, "failed to refill BM pools\n");
|
||||||
goto err_drop_frame;
|
goto err_drop_frame;
|
||||||
}
|
}
|
||||||
|
|
||||||
dma_unmap_single_attrs(dev->dev.parent, dma_addr,
|
if (pp)
|
||||||
bm_pool->buf_size, DMA_FROM_DEVICE,
|
page_pool_release_page(pp, virt_to_page(data));
|
||||||
DMA_ATTR_SKIP_CPU_SYNC);
|
else
|
||||||
|
dma_unmap_single_attrs(dev->dev.parent, dma_addr,
|
||||||
|
bm_pool->buf_size, DMA_FROM_DEVICE,
|
||||||
|
DMA_ATTR_SKIP_CPU_SYNC);
|
||||||
|
|
||||||
rcvd_pkts++;
|
rcvd_pkts++;
|
||||||
rcvd_bytes += rx_bytes;
|
rcvd_bytes += rx_bytes;
|
||||||
|
Loading…
x
Reference in New Issue
Block a user