Commit aa6be2b9 authored by David S. Miller's avatar David S. Miller

Merge branch 'cpsw-Add-XDP-support'

Ivan Khoronzhuk says:

====================
net: ethernet: ti: cpsw: Add XDP support

This patchset adds XDP support for TI cpsw driver and base it on
page_pool allocator. It was verified on af_xdp socket drop,
af_xdp l2f, ebpf XDP_DROP, XDP_REDIRECT, XDP_PASS, XDP_TX.

It was verified with following configs enabled:
CONFIG_JIT=y
CONFIG_BPFILTER=y
CONFIG_BPF_SYSCALL=y
CONFIG_XDP_SOCKETS=y
CONFIG_BPF_EVENTS=y
CONFIG_HAVE_EBPF_JIT=y
CONFIG_BPF_JIT=y
CONFIG_CGROUP_BPF=y

Link on previous v7:
https://lkml.org/lkml/2019/7/4/715

Also regular tests with iperf2 were done in order to verify impact on
regular netstack performance, compared with base commit:
https://pastebin.com/JSMT0iZ4

v8..v9:
- fix warnings on arm64 caused by typos in type casting

v7..v8:
- corrected dma calculation based on headroom instead of hard start
- minor comment changes

v6..v7:
- rolled back to v4 solution but with small modification
- picked up patch:
  https://www.spinics.net/lists/netdev/msg583145.html
- added changes related to netsec fix and cpsw

v5..v6:
- do changes that is rx_dev while redirect/flush cycle is kept the same
- dropped net: ethernet: ti: davinci_cpdma: return handler status
- other changes desc in patches

v4..v5:
- added two plreliminary patches:
  net: ethernet: ti: davinci_cpdma: allow desc split while down
  net: ethernet: ti: cpsw_ethtool: allow res split while down
- added xdp alocator refcnt on xdp level, avoiding page pool refcnt
- moved flush status as separate argument for cpdma_chan_process
- reworked cpsw code according to last changes to allocator
- added missed statistic counter

v3..v4:
- added page pool user counter
- use same pool for ndevs in dual mac
- restructured page pool create/destroy according to the last changes in API

v2..v3:
- each rxq and ndev has its own page pool

v1..v2:
- combined xdp_xmit functions
- used page allocation w/o refcnt juggle
- unmapped page for skb netstack
- moved rxq/page pool allocation to open/close pair
- added several preliminary patches:
  net: page_pool: add helper function to retrieve dma addresses
  net: page_pool: add helper function to unmap dma addresses
  net: ethernet: ti: cpsw: use cpsw as drv data
  net: ethernet: ti: cpsw_ethtool: simplify slave loops
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents 49db9228 9ed4050c
...@@ -577,8 +577,6 @@ static int mlx5e_alloc_rq(struct mlx5e_channel *c, ...@@ -577,8 +577,6 @@ static int mlx5e_alloc_rq(struct mlx5e_channel *c,
} }
err = xdp_rxq_info_reg_mem_model(&rq->xdp_rxq, err = xdp_rxq_info_reg_mem_model(&rq->xdp_rxq,
MEM_TYPE_PAGE_POOL, rq->page_pool); MEM_TYPE_PAGE_POOL, rq->page_pool);
if (err)
page_pool_free(rq->page_pool);
} }
if (err) if (err)
goto err_free; goto err_free;
...@@ -646,6 +644,7 @@ static int mlx5e_alloc_rq(struct mlx5e_channel *c, ...@@ -646,6 +644,7 @@ static int mlx5e_alloc_rq(struct mlx5e_channel *c,
if (rq->xdp_prog) if (rq->xdp_prog)
bpf_prog_put(rq->xdp_prog); bpf_prog_put(rq->xdp_prog);
xdp_rxq_info_unreg(&rq->xdp_rxq); xdp_rxq_info_unreg(&rq->xdp_rxq);
page_pool_destroy(rq->page_pool);
mlx5_wq_destroy(&rq->wq_ctrl); mlx5_wq_destroy(&rq->wq_ctrl);
return err; return err;
...@@ -680,6 +679,7 @@ static void mlx5e_free_rq(struct mlx5e_rq *rq) ...@@ -680,6 +679,7 @@ static void mlx5e_free_rq(struct mlx5e_rq *rq)
} }
xdp_rxq_info_unreg(&rq->xdp_rxq); xdp_rxq_info_unreg(&rq->xdp_rxq);
page_pool_destroy(rq->page_pool);
mlx5_wq_destroy(&rq->wq_ctrl); mlx5_wq_destroy(&rq->wq_ctrl);
} }
......
...@@ -1212,15 +1212,11 @@ static void netsec_uninit_pkt_dring(struct netsec_priv *priv, int id) ...@@ -1212,15 +1212,11 @@ static void netsec_uninit_pkt_dring(struct netsec_priv *priv, int id)
} }
} }
/* Rx is currently using page_pool /* Rx is currently using page_pool */
* since the pool is created during netsec_setup_rx_dring(), we need to
* free the pool manually if the registration failed
*/
if (id == NETSEC_RING_RX) { if (id == NETSEC_RING_RX) {
if (xdp_rxq_info_is_reg(&dring->xdp_rxq)) if (xdp_rxq_info_is_reg(&dring->xdp_rxq))
xdp_rxq_info_unreg(&dring->xdp_rxq); xdp_rxq_info_unreg(&dring->xdp_rxq);
else page_pool_destroy(dring->page_pool);
page_pool_free(dring->page_pool);
} }
memset(dring->desc, 0, sizeof(struct netsec_desc) * DESC_NUM); memset(dring->desc, 0, sizeof(struct netsec_desc) * DESC_NUM);
......
...@@ -50,6 +50,7 @@ config TI_CPSW ...@@ -50,6 +50,7 @@ config TI_CPSW
depends on ARCH_DAVINCI || ARCH_OMAP2PLUS || COMPILE_TEST depends on ARCH_DAVINCI || ARCH_OMAP2PLUS || COMPILE_TEST
select TI_DAVINCI_MDIO select TI_DAVINCI_MDIO
select MFD_SYSCON select MFD_SYSCON
select PAGE_POOL
select REGMAP select REGMAP
---help--- ---help---
This driver supports TI's CPSW Ethernet Switch. This driver supports TI's CPSW Ethernet Switch.
......
This diff is collapsed.
...@@ -578,6 +578,18 @@ static int cpsw_update_channels_res(struct cpsw_priv *priv, int ch_num, int rx, ...@@ -578,6 +578,18 @@ static int cpsw_update_channels_res(struct cpsw_priv *priv, int ch_num, int rx,
return 0; return 0;
} }
static void cpsw_fail(struct cpsw_common *cpsw)
{
struct net_device *ndev;
int i;
for (i = 0; i < cpsw->data.slaves; i++) {
ndev = cpsw->slaves[i].ndev;
if (ndev)
dev_close(ndev);
}
}
int cpsw_set_channels_common(struct net_device *ndev, int cpsw_set_channels_common(struct net_device *ndev,
struct ethtool_channels *chs, struct ethtool_channels *chs,
cpdma_handler_fn rx_handler) cpdma_handler_fn rx_handler)
...@@ -585,7 +597,7 @@ int cpsw_set_channels_common(struct net_device *ndev, ...@@ -585,7 +597,7 @@ int cpsw_set_channels_common(struct net_device *ndev,
struct cpsw_priv *priv = netdev_priv(ndev); struct cpsw_priv *priv = netdev_priv(ndev);
struct cpsw_common *cpsw = priv->cpsw; struct cpsw_common *cpsw = priv->cpsw;
struct net_device *sl_ndev; struct net_device *sl_ndev;
int i, ret; int i, new_pools, ret;
ret = cpsw_check_ch_settings(cpsw, chs); ret = cpsw_check_ch_settings(cpsw, chs);
if (ret < 0) if (ret < 0)
...@@ -593,6 +605,8 @@ int cpsw_set_channels_common(struct net_device *ndev, ...@@ -593,6 +605,8 @@ int cpsw_set_channels_common(struct net_device *ndev,
cpsw_suspend_data_pass(ndev); cpsw_suspend_data_pass(ndev);
new_pools = (chs->rx_count != cpsw->rx_ch_num) && cpsw->usage_count;
ret = cpsw_update_channels_res(priv, chs->rx_count, 1, rx_handler); ret = cpsw_update_channels_res(priv, chs->rx_count, 1, rx_handler);
if (ret) if (ret)
goto err; goto err;
...@@ -620,15 +634,21 @@ int cpsw_set_channels_common(struct net_device *ndev, ...@@ -620,15 +634,21 @@ int cpsw_set_channels_common(struct net_device *ndev,
} }
} }
if (cpsw->usage_count) cpsw_split_res(cpsw);
cpsw_split_res(cpsw);
if (new_pools) {
cpsw_destroy_xdp_rxqs(cpsw);
ret = cpsw_create_xdp_rxqs(cpsw);
if (ret)
goto err;
}
ret = cpsw_resume_data_pass(ndev); ret = cpsw_resume_data_pass(ndev);
if (!ret) if (!ret)
return 0; return 0;
err: err:
dev_err(priv->dev, "cannot update channels number, closing device\n"); dev_err(priv->dev, "cannot update channels number, closing device\n");
dev_close(ndev); cpsw_fail(cpsw);
return ret; return ret;
} }
...@@ -648,9 +668,8 @@ void cpsw_get_ringparam(struct net_device *ndev, ...@@ -648,9 +668,8 @@ void cpsw_get_ringparam(struct net_device *ndev,
int cpsw_set_ringparam(struct net_device *ndev, int cpsw_set_ringparam(struct net_device *ndev,
struct ethtool_ringparam *ering) struct ethtool_ringparam *ering)
{ {
struct cpsw_priv *priv = netdev_priv(ndev); struct cpsw_common *cpsw = ndev_to_cpsw(ndev);
struct cpsw_common *cpsw = priv->cpsw; int descs_num, ret;
int ret;
/* ignore ering->tx_pending - only rx_pending adjustment is supported */ /* ignore ering->tx_pending - only rx_pending adjustment is supported */
...@@ -659,22 +678,34 @@ int cpsw_set_ringparam(struct net_device *ndev, ...@@ -659,22 +678,34 @@ int cpsw_set_ringparam(struct net_device *ndev,
ering->rx_pending > (cpsw->descs_pool_size - CPSW_MAX_QUEUES)) ering->rx_pending > (cpsw->descs_pool_size - CPSW_MAX_QUEUES))
return -EINVAL; return -EINVAL;
if (ering->rx_pending == cpdma_get_num_rx_descs(cpsw->dma)) descs_num = cpdma_get_num_rx_descs(cpsw->dma);
if (ering->rx_pending == descs_num)
return 0; return 0;
cpsw_suspend_data_pass(ndev); cpsw_suspend_data_pass(ndev);
cpdma_set_num_rx_descs(cpsw->dma, ering->rx_pending); ret = cpdma_set_num_rx_descs(cpsw->dma, ering->rx_pending);
if (ret) {
if (cpsw_resume_data_pass(ndev))
goto err;
return ret;
}
if (cpsw->usage_count) if (cpsw->usage_count) {
cpdma_chan_split_pool(cpsw->dma); cpsw_destroy_xdp_rxqs(cpsw);
ret = cpsw_create_xdp_rxqs(cpsw);
if (ret)
goto err;
}
ret = cpsw_resume_data_pass(ndev); ret = cpsw_resume_data_pass(ndev);
if (!ret) if (!ret)
return 0; return 0;
err:
cpdma_set_num_rx_descs(cpsw->dma, descs_num);
dev_err(cpsw->dev, "cannot set ring params, closing device\n"); dev_err(cpsw->dev, "cannot set ring params, closing device\n");
dev_close(ndev); cpsw_fail(cpsw);
return ret; return ret;
} }
......
...@@ -346,6 +346,7 @@ struct cpsw_common { ...@@ -346,6 +346,7 @@ struct cpsw_common {
int rx_ch_num, tx_ch_num; int rx_ch_num, tx_ch_num;
int speed; int speed;
int usage_count; int usage_count;
struct page_pool *page_pool[CPSW_MAX_QUEUES];
}; };
struct cpsw_priv { struct cpsw_priv {
...@@ -360,6 +361,10 @@ struct cpsw_priv { ...@@ -360,6 +361,10 @@ struct cpsw_priv {
int shp_cfg_speed; int shp_cfg_speed;
int tx_ts_enabled; int tx_ts_enabled;
int rx_ts_enabled; int rx_ts_enabled;
struct bpf_prog *xdp_prog;
struct xdp_rxq_info xdp_rxq[CPSW_MAX_QUEUES];
struct xdp_attachment_info xdpi;
u32 emac_port; u32 emac_port;
struct cpsw_common *cpsw; struct cpsw_common *cpsw;
}; };
...@@ -391,6 +396,8 @@ int cpsw_fill_rx_channels(struct cpsw_priv *priv); ...@@ -391,6 +396,8 @@ int cpsw_fill_rx_channels(struct cpsw_priv *priv);
void cpsw_intr_enable(struct cpsw_common *cpsw); void cpsw_intr_enable(struct cpsw_common *cpsw);
void cpsw_intr_disable(struct cpsw_common *cpsw); void cpsw_intr_disable(struct cpsw_common *cpsw);
void cpsw_tx_handler(void *token, int len, int status); void cpsw_tx_handler(void *token, int len, int status);
int cpsw_create_xdp_rxqs(struct cpsw_common *cpsw);
void cpsw_destroy_xdp_rxqs(struct cpsw_common *cpsw);
/* ethtool */ /* ethtool */
u32 cpsw_get_msglevel(struct net_device *ndev); u32 cpsw_get_msglevel(struct net_device *ndev);
......
...@@ -139,6 +139,7 @@ struct submit_info { ...@@ -139,6 +139,7 @@ struct submit_info {
int directed; int directed;
void *token; void *token;
void *data; void *data;
int flags;
int len; int len;
}; };
...@@ -184,6 +185,8 @@ static struct cpdma_control_info controls[] = { ...@@ -184,6 +185,8 @@ static struct cpdma_control_info controls[] = {
(directed << CPDMA_TO_PORT_SHIFT)); \ (directed << CPDMA_TO_PORT_SHIFT)); \
} while (0) } while (0)
#define CPDMA_DMA_EXT_MAP BIT(16)
static void cpdma_desc_pool_destroy(struct cpdma_ctlr *ctlr) static void cpdma_desc_pool_destroy(struct cpdma_ctlr *ctlr)
{ {
struct cpdma_desc_pool *pool = ctlr->pool; struct cpdma_desc_pool *pool = ctlr->pool;
...@@ -1015,6 +1018,7 @@ static int cpdma_chan_submit_si(struct submit_info *si) ...@@ -1015,6 +1018,7 @@ static int cpdma_chan_submit_si(struct submit_info *si)
struct cpdma_chan *chan = si->chan; struct cpdma_chan *chan = si->chan;
struct cpdma_ctlr *ctlr = chan->ctlr; struct cpdma_ctlr *ctlr = chan->ctlr;
int len = si->len; int len = si->len;
int swlen = len;
struct cpdma_desc __iomem *desc; struct cpdma_desc __iomem *desc;
dma_addr_t buffer; dma_addr_t buffer;
u32 mode; u32 mode;
...@@ -1036,16 +1040,22 @@ static int cpdma_chan_submit_si(struct submit_info *si) ...@@ -1036,16 +1040,22 @@ static int cpdma_chan_submit_si(struct submit_info *si)
chan->stats.runt_transmit_buff++; chan->stats.runt_transmit_buff++;
} }
buffer = dma_map_single(ctlr->dev, si->data, len, chan->dir);
ret = dma_mapping_error(ctlr->dev, buffer);
if (ret) {
cpdma_desc_free(ctlr->pool, desc, 1);
return -EINVAL;
}
mode = CPDMA_DESC_OWNER | CPDMA_DESC_SOP | CPDMA_DESC_EOP; mode = CPDMA_DESC_OWNER | CPDMA_DESC_SOP | CPDMA_DESC_EOP;
cpdma_desc_to_port(chan, mode, si->directed); cpdma_desc_to_port(chan, mode, si->directed);
if (si->flags & CPDMA_DMA_EXT_MAP) {
buffer = (dma_addr_t)si->data;
dma_sync_single_for_device(ctlr->dev, buffer, len, chan->dir);
swlen |= CPDMA_DMA_EXT_MAP;
} else {
buffer = dma_map_single(ctlr->dev, si->data, len, chan->dir);
ret = dma_mapping_error(ctlr->dev, buffer);
if (ret) {
cpdma_desc_free(ctlr->pool, desc, 1);
return -EINVAL;
}
}
/* Relaxed IO accessors can be used here as there is read barrier /* Relaxed IO accessors can be used here as there is read barrier
* at the end of write sequence. * at the end of write sequence.
*/ */
...@@ -1055,7 +1065,7 @@ static int cpdma_chan_submit_si(struct submit_info *si) ...@@ -1055,7 +1065,7 @@ static int cpdma_chan_submit_si(struct submit_info *si)
writel_relaxed(mode | len, &desc->hw_mode); writel_relaxed(mode | len, &desc->hw_mode);
writel_relaxed((uintptr_t)si->token, &desc->sw_token); writel_relaxed((uintptr_t)si->token, &desc->sw_token);
writel_relaxed(buffer, &desc->sw_buffer); writel_relaxed(buffer, &desc->sw_buffer);
writel_relaxed(len, &desc->sw_len); writel_relaxed(swlen, &desc->sw_len);
desc_read(desc, sw_len); desc_read(desc, sw_len);
__cpdma_chan_submit(chan, desc); __cpdma_chan_submit(chan, desc);
...@@ -1079,6 +1089,32 @@ int cpdma_chan_idle_submit(struct cpdma_chan *chan, void *token, void *data, ...@@ -1079,6 +1089,32 @@ int cpdma_chan_idle_submit(struct cpdma_chan *chan, void *token, void *data,
si.data = data; si.data = data;
si.len = len; si.len = len;
si.directed = directed; si.directed = directed;
si.flags = 0;
spin_lock_irqsave(&chan->lock, flags);
if (chan->state == CPDMA_STATE_TEARDOWN) {
spin_unlock_irqrestore(&chan->lock, flags);
return -EINVAL;
}
ret = cpdma_chan_submit_si(&si);
spin_unlock_irqrestore(&chan->lock, flags);
return ret;
}
int cpdma_chan_idle_submit_mapped(struct cpdma_chan *chan, void *token,
dma_addr_t data, int len, int directed)
{
struct submit_info si;
unsigned long flags;
int ret;
si.chan = chan;
si.token = token;
si.data = (void *)data;
si.len = len;
si.directed = directed;
si.flags = CPDMA_DMA_EXT_MAP;
spin_lock_irqsave(&chan->lock, flags); spin_lock_irqsave(&chan->lock, flags);
if (chan->state == CPDMA_STATE_TEARDOWN) { if (chan->state == CPDMA_STATE_TEARDOWN) {
...@@ -1103,6 +1139,32 @@ int cpdma_chan_submit(struct cpdma_chan *chan, void *token, void *data, ...@@ -1103,6 +1139,32 @@ int cpdma_chan_submit(struct cpdma_chan *chan, void *token, void *data,
si.data = data; si.data = data;
si.len = len; si.len = len;
si.directed = directed; si.directed = directed;
si.flags = 0;
spin_lock_irqsave(&chan->lock, flags);
if (chan->state != CPDMA_STATE_ACTIVE) {
spin_unlock_irqrestore(&chan->lock, flags);
return -EINVAL;
}
ret = cpdma_chan_submit_si(&si);
spin_unlock_irqrestore(&chan->lock, flags);
return ret;
}
int cpdma_chan_submit_mapped(struct cpdma_chan *chan, void *token,
dma_addr_t data, int len, int directed)
{
struct submit_info si;
unsigned long flags;
int ret;
si.chan = chan;
si.token = token;
si.data = (void *)data;
si.len = len;
si.directed = directed;
si.flags = CPDMA_DMA_EXT_MAP;
spin_lock_irqsave(&chan->lock, flags); spin_lock_irqsave(&chan->lock, flags);
if (chan->state != CPDMA_STATE_ACTIVE) { if (chan->state != CPDMA_STATE_ACTIVE) {
...@@ -1140,10 +1202,17 @@ static void __cpdma_chan_free(struct cpdma_chan *chan, ...@@ -1140,10 +1202,17 @@ static void __cpdma_chan_free(struct cpdma_chan *chan,
uintptr_t token; uintptr_t token;
token = desc_read(desc, sw_token); token = desc_read(desc, sw_token);
buff_dma = desc_read(desc, sw_buffer);
origlen = desc_read(desc, sw_len); origlen = desc_read(desc, sw_len);
dma_unmap_single(ctlr->dev, buff_dma, origlen, chan->dir); buff_dma = desc_read(desc, sw_buffer);
if (origlen & CPDMA_DMA_EXT_MAP) {
origlen &= ~CPDMA_DMA_EXT_MAP;
dma_sync_single_for_cpu(ctlr->dev, buff_dma, origlen,
chan->dir);
} else {
dma_unmap_single(ctlr->dev, buff_dma, origlen, chan->dir);
}
cpdma_desc_free(pool, desc, 1); cpdma_desc_free(pool, desc, 1);
(*chan->handler)((void *)token, outlen, status); (*chan->handler)((void *)token, outlen, status);
} }
...@@ -1354,8 +1423,23 @@ int cpdma_get_num_tx_descs(struct cpdma_ctlr *ctlr) ...@@ -1354,8 +1423,23 @@ int cpdma_get_num_tx_descs(struct cpdma_ctlr *ctlr)
return ctlr->num_tx_desc; return ctlr->num_tx_desc;
} }
void cpdma_set_num_rx_descs(struct cpdma_ctlr *ctlr, int num_rx_desc) int cpdma_set_num_rx_descs(struct cpdma_ctlr *ctlr, int num_rx_desc)
{ {
unsigned long flags;
int temp, ret;
spin_lock_irqsave(&ctlr->lock, flags);
temp = ctlr->num_rx_desc;
ctlr->num_rx_desc = num_rx_desc; ctlr->num_rx_desc = num_rx_desc;
ctlr->num_tx_desc = ctlr->pool->num_desc - ctlr->num_rx_desc; ctlr->num_tx_desc = ctlr->pool->num_desc - ctlr->num_rx_desc;
ret = cpdma_chan_split_pool(ctlr);
if (ret) {
ctlr->num_rx_desc = temp;
ctlr->num_tx_desc = ctlr->pool->num_desc - ctlr->num_rx_desc;
}
spin_unlock_irqrestore(&ctlr->lock, flags);
return ret;
} }
...@@ -77,8 +77,12 @@ int cpdma_chan_stop(struct cpdma_chan *chan); ...@@ -77,8 +77,12 @@ int cpdma_chan_stop(struct cpdma_chan *chan);
int cpdma_chan_get_stats(struct cpdma_chan *chan, int cpdma_chan_get_stats(struct cpdma_chan *chan,
struct cpdma_chan_stats *stats); struct cpdma_chan_stats *stats);
int cpdma_chan_submit_mapped(struct cpdma_chan *chan, void *token,
dma_addr_t data, int len, int directed);
int cpdma_chan_submit(struct cpdma_chan *chan, void *token, void *data, int cpdma_chan_submit(struct cpdma_chan *chan, void *token, void *data,
int len, int directed); int len, int directed);
int cpdma_chan_idle_submit_mapped(struct cpdma_chan *chan, void *token,
dma_addr_t data, int len, int directed);
int cpdma_chan_idle_submit(struct cpdma_chan *chan, void *token, void *data, int cpdma_chan_idle_submit(struct cpdma_chan *chan, void *token, void *data,
int len, int directed); int len, int directed);
int cpdma_chan_process(struct cpdma_chan *chan, int quota); int cpdma_chan_process(struct cpdma_chan *chan, int quota);
...@@ -112,8 +116,7 @@ enum cpdma_control { ...@@ -112,8 +116,7 @@ enum cpdma_control {
int cpdma_control_get(struct cpdma_ctlr *ctlr, int control); int cpdma_control_get(struct cpdma_ctlr *ctlr, int control);
int cpdma_control_set(struct cpdma_ctlr *ctlr, int control, int value); int cpdma_control_set(struct cpdma_ctlr *ctlr, int control, int value);
int cpdma_get_num_rx_descs(struct cpdma_ctlr *ctlr); int cpdma_get_num_rx_descs(struct cpdma_ctlr *ctlr);
void cpdma_set_num_rx_descs(struct cpdma_ctlr *ctlr, int num_rx_desc); int cpdma_set_num_rx_descs(struct cpdma_ctlr *ctlr, int num_rx_desc);
int cpdma_get_num_tx_descs(struct cpdma_ctlr *ctlr); int cpdma_get_num_tx_descs(struct cpdma_ctlr *ctlr);
int cpdma_chan_split_pool(struct cpdma_ctlr *ctlr);
#endif #endif
...@@ -101,6 +101,12 @@ struct page_pool { ...@@ -101,6 +101,12 @@ struct page_pool {
struct ptr_ring ring; struct ptr_ring ring;
atomic_t pages_state_release_cnt; atomic_t pages_state_release_cnt;
/* A page_pool is strictly tied to a single RX-queue being
* protected by NAPI, due to above pp_alloc_cache. This
* refcnt serves purpose is to simplify drivers error handling.
*/
refcount_t user_cnt;
}; };
struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
...@@ -134,6 +140,15 @@ static inline void page_pool_free(struct page_pool *pool) ...@@ -134,6 +140,15 @@ static inline void page_pool_free(struct page_pool *pool)
#endif #endif
} }
/* Drivers use this instead of page_pool_free */
static inline void page_pool_destroy(struct page_pool *pool)
{
if (!pool)
return;
page_pool_free(pool);
}
/* Never call this directly, use helpers below */ /* Never call this directly, use helpers below */
void __page_pool_put_page(struct page_pool *pool, void __page_pool_put_page(struct page_pool *pool,
struct page *page, bool allow_direct); struct page *page, bool allow_direct);
...@@ -201,4 +216,14 @@ static inline bool is_page_pool_compiled_in(void) ...@@ -201,4 +216,14 @@ static inline bool is_page_pool_compiled_in(void)
#endif #endif
} }
static inline void page_pool_get(struct page_pool *pool)
{
refcount_inc(&pool->user_cnt);
}
static inline bool page_pool_put(struct page_pool *pool)
{
return refcount_dec_and_test(&pool->user_cnt);
}
#endif /* _NET_PAGE_POOL_H */ #endif /* _NET_PAGE_POOL_H */
...@@ -49,6 +49,9 @@ static int page_pool_init(struct page_pool *pool, ...@@ -49,6 +49,9 @@ static int page_pool_init(struct page_pool *pool,
atomic_set(&pool->pages_state_release_cnt, 0); atomic_set(&pool->pages_state_release_cnt, 0);
/* Driver calling page_pool_create() also call page_pool_destroy() */
refcount_set(&pool->user_cnt, 1);
if (pool->p.flags & PP_FLAG_DMA_MAP) if (pool->p.flags & PP_FLAG_DMA_MAP)
get_device(pool->p.dev); get_device(pool->p.dev);
...@@ -70,6 +73,7 @@ struct page_pool *page_pool_create(const struct page_pool_params *params) ...@@ -70,6 +73,7 @@ struct page_pool *page_pool_create(const struct page_pool_params *params)
kfree(pool); kfree(pool);
return ERR_PTR(err); return ERR_PTR(err);
} }
return pool; return pool;
} }
EXPORT_SYMBOL(page_pool_create); EXPORT_SYMBOL(page_pool_create);
...@@ -356,6 +360,10 @@ static void __warn_in_flight(struct page_pool *pool) ...@@ -356,6 +360,10 @@ static void __warn_in_flight(struct page_pool *pool)
void __page_pool_free(struct page_pool *pool) void __page_pool_free(struct page_pool *pool)
{ {
/* Only last user actually free/release resources */
if (!page_pool_put(pool))
return;
WARN(pool->alloc.count, "API usage violation"); WARN(pool->alloc.count, "API usage violation");
WARN(!ptr_ring_empty(&pool->ring), "ptr_ring is not empty"); WARN(!ptr_ring_empty(&pool->ring), "ptr_ring is not empty");
......
...@@ -370,6 +370,9 @@ int xdp_rxq_info_reg_mem_model(struct xdp_rxq_info *xdp_rxq, ...@@ -370,6 +370,9 @@ int xdp_rxq_info_reg_mem_model(struct xdp_rxq_info *xdp_rxq,
goto err; goto err;
} }
if (type == MEM_TYPE_PAGE_POOL)
page_pool_get(xdp_alloc->page_pool);
mutex_unlock(&mem_id_lock); mutex_unlock(&mem_id_lock);
trace_mem_connect(xdp_alloc, xdp_rxq); trace_mem_connect(xdp_alloc, xdp_rxq);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment