Commit fe2d55d2 authored by David S. Miller's avatar David S. Miller

Merge branch 'ibmvnic-Update-TX-pool-and-TX-routines'

Thomas Falcon says:

====================
ibmvnic: Update TX pool and TX routines

This patch restructures the TX pool data structure and provides a
separate TX pool array for TSO transmissions. This is already used
in some way due to our unique DMA situation, namely that we cannot
use single DMA mappings for packet data. Previously, both buffer
arrays used the same pool entry. This restructuring allows for
some additional cleanup in the driver code, especially in some
places in the device transmit routine.

In addition, it allows us to more easily track the consumer
and producer indexes of a particular pool. This has been
further improved by better tracking of in-use buffers to
prevent possible data corruption in case an invalid buffer
entry is used.

v5: Fix bisectability mistake in the first patch. Removed
    TSO-specific data in a later patch when it is no longer used.

v4: Fix error in 7th patch that causes an oops by using
    the older fixed value for number of buffers instead
    of the respective field in the tx pool data structure

v3: Forgot to update TX pool cleaning function to handle new data
    structures. Included 7th patch for that.

v2: Fix typo in 3/6 commit subject line
====================
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parents d47d08c8 76c15c91
...@@ -557,36 +557,41 @@ static int init_rx_pools(struct net_device *netdev) ...@@ -557,36 +557,41 @@ static int init_rx_pools(struct net_device *netdev)
return 0; return 0;
} }
static int reset_tx_pools(struct ibmvnic_adapter *adapter) static int reset_one_tx_pool(struct ibmvnic_adapter *adapter,
struct ibmvnic_tx_pool *tx_pool)
{ {
struct ibmvnic_tx_pool *tx_pool; int rc, i;
int tx_scrqs;
int i, j, rc;
tx_scrqs = be32_to_cpu(adapter->login_rsp_buf->num_txsubm_subcrqs);
for (i = 0; i < tx_scrqs; i++) {
netdev_dbg(adapter->netdev, "Re-setting tx_pool[%d]\n", i);
tx_pool = &adapter->tx_pool[i];
rc = reset_long_term_buff(adapter, &tx_pool->long_term_buff); rc = reset_long_term_buff(adapter, &tx_pool->long_term_buff);
if (rc) if (rc)
return rc; return rc;
rc = reset_long_term_buff(adapter, &tx_pool->tso_ltb);
if (rc)
return rc;
memset(tx_pool->tx_buff, 0, memset(tx_pool->tx_buff, 0,
adapter->req_tx_entries_per_subcrq * tx_pool->num_buffers *
sizeof(struct ibmvnic_tx_buff)); sizeof(struct ibmvnic_tx_buff));
for (j = 0; j < adapter->req_tx_entries_per_subcrq; j++) for (i = 0; i < tx_pool->num_buffers; i++)
tx_pool->free_map[j] = j; tx_pool->free_map[i] = i;
tx_pool->consumer_index = 0; tx_pool->consumer_index = 0;
tx_pool->producer_index = 0; tx_pool->producer_index = 0;
tx_pool->tso_index = 0;
return 0;
}
static int reset_tx_pools(struct ibmvnic_adapter *adapter)
{
int tx_scrqs;
int i, rc;
tx_scrqs = be32_to_cpu(adapter->login_rsp_buf->num_txsubm_subcrqs);
for (i = 0; i < tx_scrqs; i++) {
rc = reset_one_tx_pool(adapter, &adapter->tso_pool[i]);
if (rc)
return rc;
rc = reset_one_tx_pool(adapter, &adapter->tx_pool[i]);
if (rc)
return rc;
} }
return 0; return 0;
...@@ -603,35 +608,70 @@ static void release_vpd_data(struct ibmvnic_adapter *adapter) ...@@ -603,35 +608,70 @@ static void release_vpd_data(struct ibmvnic_adapter *adapter)
adapter->vpd = NULL; adapter->vpd = NULL;
} }
static void release_one_tx_pool(struct ibmvnic_adapter *adapter,
struct ibmvnic_tx_pool *tx_pool)
{
kfree(tx_pool->tx_buff);
kfree(tx_pool->free_map);
free_long_term_buff(adapter, &tx_pool->long_term_buff);
}
static void release_tx_pools(struct ibmvnic_adapter *adapter) static void release_tx_pools(struct ibmvnic_adapter *adapter)
{ {
struct ibmvnic_tx_pool *tx_pool;
int i; int i;
if (!adapter->tx_pool) if (!adapter->tx_pool)
return; return;
for (i = 0; i < adapter->num_active_tx_pools; i++) { for (i = 0; i < adapter->num_active_tx_pools; i++) {
netdev_dbg(adapter->netdev, "Releasing tx_pool[%d]\n", i); release_one_tx_pool(adapter, &adapter->tx_pool[i]);
tx_pool = &adapter->tx_pool[i]; release_one_tx_pool(adapter, &adapter->tso_pool[i]);
kfree(tx_pool->tx_buff);
free_long_term_buff(adapter, &tx_pool->long_term_buff);
free_long_term_buff(adapter, &tx_pool->tso_ltb);
kfree(tx_pool->free_map);
} }
kfree(adapter->tx_pool); kfree(adapter->tx_pool);
adapter->tx_pool = NULL; adapter->tx_pool = NULL;
kfree(adapter->tso_pool);
adapter->tso_pool = NULL;
adapter->num_active_tx_pools = 0; adapter->num_active_tx_pools = 0;
} }
static int init_one_tx_pool(struct net_device *netdev,
struct ibmvnic_tx_pool *tx_pool,
int num_entries, int buf_size)
{
struct ibmvnic_adapter *adapter = netdev_priv(netdev);
int i;
tx_pool->tx_buff = kcalloc(num_entries,
sizeof(struct ibmvnic_tx_buff),
GFP_KERNEL);
if (!tx_pool->tx_buff)
return -1;
if (alloc_long_term_buff(adapter, &tx_pool->long_term_buff,
num_entries * buf_size))
return -1;
tx_pool->free_map = kcalloc(num_entries, sizeof(int), GFP_KERNEL);
if (!tx_pool->free_map)
return -1;
for (i = 0; i < num_entries; i++)
tx_pool->free_map[i] = i;
tx_pool->consumer_index = 0;
tx_pool->producer_index = 0;
tx_pool->num_buffers = num_entries;
tx_pool->buf_size = buf_size;
return 0;
}
static int init_tx_pools(struct net_device *netdev) static int init_tx_pools(struct net_device *netdev)
{ {
struct ibmvnic_adapter *adapter = netdev_priv(netdev); struct ibmvnic_adapter *adapter = netdev_priv(netdev);
struct device *dev = &adapter->vdev->dev;
struct ibmvnic_tx_pool *tx_pool;
int tx_subcrqs; int tx_subcrqs;
int i, j; int i, rc;
tx_subcrqs = be32_to_cpu(adapter->login_rsp_buf->num_txsubm_subcrqs); tx_subcrqs = be32_to_cpu(adapter->login_rsp_buf->num_txsubm_subcrqs);
adapter->tx_pool = kcalloc(tx_subcrqs, adapter->tx_pool = kcalloc(tx_subcrqs,
...@@ -639,53 +679,29 @@ static int init_tx_pools(struct net_device *netdev) ...@@ -639,53 +679,29 @@ static int init_tx_pools(struct net_device *netdev)
if (!adapter->tx_pool) if (!adapter->tx_pool)
return -1; return -1;
adapter->num_active_tx_pools = tx_subcrqs; adapter->tso_pool = kcalloc(tx_subcrqs,
sizeof(struct ibmvnic_tx_pool), GFP_KERNEL);
for (i = 0; i < tx_subcrqs; i++) { if (!adapter->tso_pool)
tx_pool = &adapter->tx_pool[i];
netdev_dbg(adapter->netdev,
"Initializing tx_pool[%d], %lld buffs\n",
i, adapter->req_tx_entries_per_subcrq);
tx_pool->tx_buff = kcalloc(adapter->req_tx_entries_per_subcrq,
sizeof(struct ibmvnic_tx_buff),
GFP_KERNEL);
if (!tx_pool->tx_buff) {
dev_err(dev, "tx pool buffer allocation failed\n");
release_tx_pools(adapter);
return -1; return -1;
}
if (alloc_long_term_buff(adapter, &tx_pool->long_term_buff, adapter->num_active_tx_pools = tx_subcrqs;
adapter->req_tx_entries_per_subcrq *
(adapter->req_mtu + VLAN_HLEN))) {
release_tx_pools(adapter);
return -1;
}
/* alloc TSO ltb */ for (i = 0; i < tx_subcrqs; i++) {
if (alloc_long_term_buff(adapter, &tx_pool->tso_ltb, rc = init_one_tx_pool(netdev, &adapter->tx_pool[i],
IBMVNIC_TSO_BUFS * adapter->req_tx_entries_per_subcrq,
IBMVNIC_TSO_BUF_SZ)) { adapter->req_mtu + VLAN_HLEN);
if (rc) {
release_tx_pools(adapter); release_tx_pools(adapter);
return -1; return rc;
} }
tx_pool->tso_index = 0; init_one_tx_pool(netdev, &adapter->tso_pool[i],
IBMVNIC_TSO_BUFS,
tx_pool->free_map = kcalloc(adapter->req_tx_entries_per_subcrq, IBMVNIC_TSO_BUF_SZ);
sizeof(int), GFP_KERNEL); if (rc) {
if (!tx_pool->free_map) {
release_tx_pools(adapter); release_tx_pools(adapter);
return -1; return rc;
} }
for (j = 0; j < adapter->req_tx_entries_per_subcrq; j++)
tx_pool->free_map[j] = j;
tx_pool->consumer_index = 0;
tx_pool->producer_index = 0;
} }
return 0; return 0;
...@@ -1112,34 +1128,42 @@ static void clean_rx_pools(struct ibmvnic_adapter *adapter) ...@@ -1112,34 +1128,42 @@ static void clean_rx_pools(struct ibmvnic_adapter *adapter)
} }
} }
static void clean_tx_pools(struct ibmvnic_adapter *adapter) static void clean_one_tx_pool(struct ibmvnic_adapter *adapter,
struct ibmvnic_tx_pool *tx_pool)
{ {
struct ibmvnic_tx_pool *tx_pool;
struct ibmvnic_tx_buff *tx_buff; struct ibmvnic_tx_buff *tx_buff;
u64 tx_entries; u64 tx_entries;
int tx_scrqs; int i;
int i, j;
if (!adapter->tx_pool) if (!tx_pool && !tx_pool->tx_buff)
return; return;
tx_scrqs = be32_to_cpu(adapter->login_rsp_buf->num_txsubm_subcrqs); tx_entries = tx_pool->num_buffers;
tx_entries = adapter->req_tx_entries_per_subcrq;
/* Free any remaining skbs in the tx buffer pools */
for (i = 0; i < tx_scrqs; i++) {
tx_pool = &adapter->tx_pool[i];
if (!tx_pool && !tx_pool->tx_buff)
continue;
netdev_dbg(adapter->netdev, "Cleaning tx_pool[%d]\n", i); for (i = 0; i < tx_entries; i++) {
for (j = 0; j < tx_entries; j++) { tx_buff = &tx_pool->tx_buff[i];
tx_buff = &tx_pool->tx_buff[j];
if (tx_buff && tx_buff->skb) { if (tx_buff && tx_buff->skb) {
dev_kfree_skb_any(tx_buff->skb); dev_kfree_skb_any(tx_buff->skb);
tx_buff->skb = NULL; tx_buff->skb = NULL;
} }
} }
}
static void clean_tx_pools(struct ibmvnic_adapter *adapter)
{
int tx_scrqs;
int i;
if (!adapter->tx_pool || !adapter->tso_pool)
return;
tx_scrqs = be32_to_cpu(adapter->login_rsp_buf->num_txsubm_subcrqs);
/* Free any remaining skbs in the tx buffer pools */
for (i = 0; i < tx_scrqs; i++) {
netdev_dbg(adapter->netdev, "Cleaning tx_pool[%d]\n", i);
clean_one_tx_pool(adapter, &adapter->tx_pool[i]);
clean_one_tx_pool(adapter, &adapter->tso_pool[i]);
} }
} }
...@@ -1398,8 +1422,11 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1398,8 +1422,11 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
ret = NETDEV_TX_OK; ret = NETDEV_TX_OK;
goto out; goto out;
} }
if (skb_is_gso(skb))
tx_pool = &adapter->tso_pool[queue_num];
else
tx_pool = &adapter->tx_pool[queue_num]; tx_pool = &adapter->tx_pool[queue_num];
tx_scrq = adapter->tx_scrq[queue_num]; tx_scrq = adapter->tx_scrq[queue_num];
txq = netdev_get_tx_queue(netdev, skb_get_queue_mapping(skb)); txq = netdev_get_tx_queue(netdev, skb_get_queue_mapping(skb));
handle_array = (u64 *)((u8 *)(adapter->login_rsp_buf) + handle_array = (u64 *)((u8 *)(adapter->login_rsp_buf) +
...@@ -1407,20 +1434,20 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1407,20 +1434,20 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
index = tx_pool->free_map[tx_pool->consumer_index]; index = tx_pool->free_map[tx_pool->consumer_index];
if (skb_is_gso(skb)) { if (index == IBMVNIC_INVALID_MAP) {
offset = tx_pool->tso_index * IBMVNIC_TSO_BUF_SZ; dev_kfree_skb_any(skb);
dst = tx_pool->tso_ltb.buff + offset; tx_send_failed++;
memset(dst, 0, IBMVNIC_TSO_BUF_SZ); tx_dropped++;
data_dma_addr = tx_pool->tso_ltb.addr + offset; ret = NETDEV_TX_OK;
tx_pool->tso_index++; goto out;
if (tx_pool->tso_index == IBMVNIC_TSO_BUFS) }
tx_pool->tso_index = 0;
} else { tx_pool->free_map[tx_pool->consumer_index] = IBMVNIC_INVALID_MAP;
offset = index * (adapter->req_mtu + VLAN_HLEN);
offset = index * tx_pool->buf_size;
dst = tx_pool->long_term_buff.buff + offset; dst = tx_pool->long_term_buff.buff + offset;
memset(dst, 0, adapter->req_mtu + VLAN_HLEN); memset(dst, 0, tx_pool->buf_size);
data_dma_addr = tx_pool->long_term_buff.addr + offset; data_dma_addr = tx_pool->long_term_buff.addr + offset;
}
if (skb_shinfo(skb)->nr_frags) { if (skb_shinfo(skb)->nr_frags) {
int cur, i; int cur, i;
...@@ -1443,8 +1470,7 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1443,8 +1470,7 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
} }
tx_pool->consumer_index = tx_pool->consumer_index =
(tx_pool->consumer_index + 1) % (tx_pool->consumer_index + 1) % tx_pool->num_buffers;
adapter->req_tx_entries_per_subcrq;
tx_buff = &tx_pool->tx_buff[index]; tx_buff = &tx_pool->tx_buff[index];
tx_buff->skb = skb; tx_buff->skb = skb;
...@@ -1460,10 +1486,12 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1460,10 +1486,12 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
tx_crq.v1.n_crq_elem = 1; tx_crq.v1.n_crq_elem = 1;
tx_crq.v1.n_sge = 1; tx_crq.v1.n_sge = 1;
tx_crq.v1.flags1 = IBMVNIC_TX_COMP_NEEDED; tx_crq.v1.flags1 = IBMVNIC_TX_COMP_NEEDED;
tx_crq.v1.correlator = cpu_to_be32(index);
if (skb_is_gso(skb)) if (skb_is_gso(skb))
tx_crq.v1.dma_reg = cpu_to_be16(tx_pool->tso_ltb.map_id); tx_crq.v1.correlator =
cpu_to_be32(index | IBMVNIC_TSO_POOL_MASK);
else else
tx_crq.v1.correlator = cpu_to_be32(index);
tx_crq.v1.dma_reg = cpu_to_be16(tx_pool->long_term_buff.map_id); tx_crq.v1.dma_reg = cpu_to_be16(tx_pool->long_term_buff.map_id);
tx_crq.v1.sge_len = cpu_to_be32(skb->len); tx_crq.v1.sge_len = cpu_to_be32(skb->len);
tx_crq.v1.ioba = cpu_to_be64(data_dma_addr); tx_crq.v1.ioba = cpu_to_be64(data_dma_addr);
...@@ -1512,7 +1540,7 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1512,7 +1540,7 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
tx_map_failed++; tx_map_failed++;
tx_dropped++; tx_dropped++;
ret = NETDEV_TX_OK; ret = NETDEV_TX_OK;
goto out; goto tx_err_out;
} }
lpar_rc = send_subcrq_indirect(adapter, handle_array[queue_num], lpar_rc = send_subcrq_indirect(adapter, handle_array[queue_num],
(u64)tx_buff->indir_dma, (u64)tx_buff->indir_dma,
...@@ -1524,13 +1552,6 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1524,13 +1552,6 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
} }
if (lpar_rc != H_SUCCESS) { if (lpar_rc != H_SUCCESS) {
dev_err(dev, "tx failed with code %ld\n", lpar_rc); dev_err(dev, "tx failed with code %ld\n", lpar_rc);
if (tx_pool->consumer_index == 0)
tx_pool->consumer_index =
adapter->req_tx_entries_per_subcrq - 1;
else
tx_pool->consumer_index--;
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
tx_buff->skb = NULL; tx_buff->skb = NULL;
...@@ -1546,7 +1567,7 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1546,7 +1567,7 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
tx_send_failed++; tx_send_failed++;
tx_dropped++; tx_dropped++;
ret = NETDEV_TX_OK; ret = NETDEV_TX_OK;
goto out; goto tx_err_out;
} }
if (atomic_add_return(num_entries, &tx_scrq->used) if (atomic_add_return(num_entries, &tx_scrq->used)
...@@ -1559,7 +1580,16 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev) ...@@ -1559,7 +1580,16 @@ static int ibmvnic_xmit(struct sk_buff *skb, struct net_device *netdev)
tx_bytes += skb->len; tx_bytes += skb->len;
txq->trans_start = jiffies; txq->trans_start = jiffies;
ret = NETDEV_TX_OK; ret = NETDEV_TX_OK;
goto out;
tx_err_out:
/* roll back consumer index and map array*/
if (tx_pool->consumer_index == 0)
tx_pool->consumer_index =
tx_pool->num_buffers - 1;
else
tx_pool->consumer_index--;
tx_pool->free_map[tx_pool->consumer_index] = index;
out: out:
netdev->stats.tx_dropped += tx_dropped; netdev->stats.tx_dropped += tx_dropped;
netdev->stats.tx_bytes += tx_bytes; netdev->stats.tx_bytes += tx_bytes;
...@@ -2531,6 +2561,7 @@ static int ibmvnic_complete_tx(struct ibmvnic_adapter *adapter, ...@@ -2531,6 +2561,7 @@ static int ibmvnic_complete_tx(struct ibmvnic_adapter *adapter,
struct ibmvnic_sub_crq_queue *scrq) struct ibmvnic_sub_crq_queue *scrq)
{ {
struct device *dev = &adapter->vdev->dev; struct device *dev = &adapter->vdev->dev;
struct ibmvnic_tx_pool *tx_pool;
struct ibmvnic_tx_buff *txbuff; struct ibmvnic_tx_buff *txbuff;
union sub_crq *next; union sub_crq *next;
int index; int index;
...@@ -2550,7 +2581,14 @@ static int ibmvnic_complete_tx(struct ibmvnic_adapter *adapter, ...@@ -2550,7 +2581,14 @@ static int ibmvnic_complete_tx(struct ibmvnic_adapter *adapter,
continue; continue;
} }
index = be32_to_cpu(next->tx_comp.correlators[i]); index = be32_to_cpu(next->tx_comp.correlators[i]);
txbuff = &adapter->tx_pool[pool].tx_buff[index]; if (index & IBMVNIC_TSO_POOL_MASK) {
tx_pool = &adapter->tso_pool[pool];
index &= ~IBMVNIC_TSO_POOL_MASK;
} else {
tx_pool = &adapter->tx_pool[pool];
}
txbuff = &tx_pool->tx_buff[index];
for (j = 0; j < IBMVNIC_MAX_FRAGS_PER_CRQ; j++) { for (j = 0; j < IBMVNIC_MAX_FRAGS_PER_CRQ; j++) {
if (!txbuff->data_dma[j]) if (!txbuff->data_dma[j])
...@@ -2573,11 +2611,10 @@ static int ibmvnic_complete_tx(struct ibmvnic_adapter *adapter, ...@@ -2573,11 +2611,10 @@ static int ibmvnic_complete_tx(struct ibmvnic_adapter *adapter,
num_entries += txbuff->num_entries; num_entries += txbuff->num_entries;
adapter->tx_pool[pool].free_map[adapter->tx_pool[pool]. tx_pool->free_map[tx_pool->producer_index] = index;
producer_index] = index; tx_pool->producer_index =
adapter->tx_pool[pool].producer_index = (tx_pool->producer_index + 1) %
(adapter->tx_pool[pool].producer_index + 1) % tx_pool->num_buffers;
adapter->req_tx_entries_per_subcrq;
} }
/* remove tx_comp scrq*/ /* remove tx_comp scrq*/
next->tx_comp.first = 0; next->tx_comp.first = 0;
......
...@@ -43,6 +43,7 @@ ...@@ -43,6 +43,7 @@
#define IBMVNIC_TSO_BUF_SZ 65536 #define IBMVNIC_TSO_BUF_SZ 65536
#define IBMVNIC_TSO_BUFS 64 #define IBMVNIC_TSO_BUFS 64
#define IBMVNIC_TSO_POOL_MASK 0x80000000
#define IBMVNIC_MAX_LTB_SIZE ((1 << (MAX_ORDER - 1)) * PAGE_SIZE) #define IBMVNIC_MAX_LTB_SIZE ((1 << (MAX_ORDER - 1)) * PAGE_SIZE)
#define IBMVNIC_BUFFER_HLEN 500 #define IBMVNIC_BUFFER_HLEN 500
...@@ -917,11 +918,9 @@ struct ibmvnic_tx_pool { ...@@ -917,11 +918,9 @@ struct ibmvnic_tx_pool {
int *free_map; int *free_map;
int consumer_index; int consumer_index;
int producer_index; int producer_index;
wait_queue_head_t ibmvnic_tx_comp_q;
struct task_struct *work_thread;
struct ibmvnic_long_term_buff long_term_buff; struct ibmvnic_long_term_buff long_term_buff;
struct ibmvnic_long_term_buff tso_ltb; int num_buffers;
int tso_index; int buf_size;
}; };
struct ibmvnic_rx_buff { struct ibmvnic_rx_buff {
...@@ -1044,6 +1043,7 @@ struct ibmvnic_adapter { ...@@ -1044,6 +1043,7 @@ struct ibmvnic_adapter {
u64 promisc; u64 promisc;
struct ibmvnic_tx_pool *tx_pool; struct ibmvnic_tx_pool *tx_pool;
struct ibmvnic_tx_pool *tso_pool;
struct completion init_done; struct completion init_done;
int init_done_rc; int init_done_rc;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment