Commit c493ea45 authored by Alexander Duyck's avatar Alexander Duyck Committed by David S. Miller

igb: remove IGB_DESC_UNUSED since it is better handled by a function call

This patch removes IGB_DESC_UNUSED and replaces it with a function call
instead in order to cleanup some of the ugliness introduced by the macro.
Signed-off-by: default avatarAlexander Duyck <alexander.h.duyck@intel.com>
Signed-off-by: default avatarJeff Kirsher <jeffrey.t.kirsher@intel.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent fa4a7ef3
...@@ -182,10 +182,6 @@ struct igb_ring { ...@@ -182,10 +182,6 @@ struct igb_ring {
char name[IFNAMSIZ + 5]; char name[IFNAMSIZ + 5];
}; };
#define IGB_DESC_UNUSED(R) \
((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \
(R)->next_to_clean - (R)->next_to_use - 1)
#define E1000_RX_DESC_ADV(R, i) \ #define E1000_RX_DESC_ADV(R, i) \
(&(((union e1000_adv_rx_desc *)((R).desc))[i])) (&(((union e1000_adv_rx_desc *)((R).desc))[i]))
#define E1000_TX_DESC_ADV(R, i) \ #define E1000_TX_DESC_ADV(R, i) \
......
...@@ -277,6 +277,17 @@ static char *igb_get_time_str(struct igb_adapter *adapter, ...@@ -277,6 +277,17 @@ static char *igb_get_time_str(struct igb_adapter *adapter,
} }
#endif #endif
/**
* igb_desc_unused - calculate if we have unused descriptors
**/
static int igb_desc_unused(struct igb_ring *ring)
{
if (ring->next_to_clean > ring->next_to_use)
return ring->next_to_clean - ring->next_to_use - 1;
return ring->count + ring->next_to_clean - ring->next_to_use - 1;
}
/** /**
* igb_init_module - Driver Registration Routine * igb_init_module - Driver Registration Routine
* *
...@@ -873,12 +884,12 @@ static void igb_configure(struct igb_adapter *adapter) ...@@ -873,12 +884,12 @@ static void igb_configure(struct igb_adapter *adapter)
igb_rx_fifo_flush_82575(&adapter->hw); igb_rx_fifo_flush_82575(&adapter->hw);
/* call IGB_DESC_UNUSED which always leaves /* call igb_desc_unused which always leaves
* at least 1 descriptor unused to make sure * at least 1 descriptor unused to make sure
* next_to_use != next_to_clean */ * next_to_use != next_to_clean */
for (i = 0; i < adapter->num_rx_queues; i++) { for (i = 0; i < adapter->num_rx_queues; i++) {
struct igb_ring *ring = &adapter->rx_ring[i]; struct igb_ring *ring = &adapter->rx_ring[i];
igb_alloc_rx_buffers_adv(ring, IGB_DESC_UNUSED(ring)); igb_alloc_rx_buffers_adv(ring, igb_desc_unused(ring));
} }
...@@ -2661,7 +2672,7 @@ static void igb_watchdog_task(struct work_struct *work) ...@@ -2661,7 +2672,7 @@ static void igb_watchdog_task(struct work_struct *work)
igb_update_adaptive(&adapter->hw); igb_update_adaptive(&adapter->hw);
if (!netif_carrier_ok(netdev)) { if (!netif_carrier_ok(netdev)) {
if (IGB_DESC_UNUSED(tx_ring) + 1 < tx_ring->count) { if (igb_desc_unused(tx_ring) + 1 < tx_ring->count) {
/* We've lost link, so the controller stops DMA, /* We've lost link, so the controller stops DMA,
* but we've got queued Tx work that's never going * but we've got queued Tx work that's never going
* to get done, so reset controller to flush Tx. * to get done, so reset controller to flush Tx.
...@@ -3199,7 +3210,7 @@ static int __igb_maybe_stop_tx(struct net_device *netdev, ...@@ -3199,7 +3210,7 @@ static int __igb_maybe_stop_tx(struct net_device *netdev,
/* We need to check again in a case another CPU has just /* We need to check again in a case another CPU has just
* made room available. */ * made room available. */
if (IGB_DESC_UNUSED(tx_ring) < size) if (igb_desc_unused(tx_ring) < size)
return -EBUSY; return -EBUSY;
/* A reprieve! */ /* A reprieve! */
...@@ -3211,7 +3222,7 @@ static int __igb_maybe_stop_tx(struct net_device *netdev, ...@@ -3211,7 +3222,7 @@ static int __igb_maybe_stop_tx(struct net_device *netdev,
static int igb_maybe_stop_tx(struct net_device *netdev, static int igb_maybe_stop_tx(struct net_device *netdev,
struct igb_ring *tx_ring, int size) struct igb_ring *tx_ring, int size)
{ {
if (IGB_DESC_UNUSED(tx_ring) >= size) if (igb_desc_unused(tx_ring) >= size)
return 0; return 0;
return __igb_maybe_stop_tx(netdev, tx_ring, size); return __igb_maybe_stop_tx(netdev, tx_ring, size);
} }
...@@ -4310,7 +4321,7 @@ static bool igb_clean_tx_irq(struct igb_ring *tx_ring) ...@@ -4310,7 +4321,7 @@ static bool igb_clean_tx_irq(struct igb_ring *tx_ring)
if (unlikely(count && if (unlikely(count &&
netif_carrier_ok(netdev) && netif_carrier_ok(netdev) &&
IGB_DESC_UNUSED(tx_ring) >= IGB_TX_QUEUE_WAKE)) { igb_desc_unused(tx_ring) >= IGB_TX_QUEUE_WAKE)) {
/* Make sure that anybody stopping the queue after this /* Make sure that anybody stopping the queue after this
* sees the new next_to_clean. * sees the new next_to_clean.
*/ */
...@@ -4587,7 +4598,7 @@ static bool igb_clean_rx_irq_adv(struct igb_ring *rx_ring, ...@@ -4587,7 +4598,7 @@ static bool igb_clean_rx_irq_adv(struct igb_ring *rx_ring,
} }
rx_ring->next_to_clean = i; rx_ring->next_to_clean = i;
cleaned_count = IGB_DESC_UNUSED(rx_ring); cleaned_count = igb_desc_unused(rx_ring);
if (cleaned_count) if (cleaned_count)
igb_alloc_rx_buffers_adv(rx_ring, cleaned_count); igb_alloc_rx_buffers_adv(rx_ring, cleaned_count);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment