Commit 242722dd authored by Jacob Keller's avatar Jacob Keller Committed by Jeff Kirsher

fm10k: Update adaptive ITR algorithm

The existing adaptive ITR algorithm is overly restrictive. It throttles
incorrectly for various traffic rates, and does not produce good
performance. The algorithm now allows for more interrupts per second,
and does some calculation to help improve for smaller packet loads. In
addition, take into account the new itr_scale from the hardware which
indicates how much to scale due to PCIe link speed.
Reported-by: default avatarMatthew Vick <matthew.vick@intel.com>
Reported-by: default avatarAlex Duyck <alexander.duyck@gmail.com>
Signed-off-by: default avatarJacob Keller <jacob.e.keller@intel.com>
Tested-by: default avatarKrishneil Singh <krishneil.k.singh@intel.com>
Signed-off-by: default avatarJeff Kirsher <jeffrey.t.kirsher@intel.com>
parent 584373f5
...@@ -164,6 +164,7 @@ struct fm10k_ring_container { ...@@ -164,6 +164,7 @@ struct fm10k_ring_container {
unsigned int total_packets; /* total packets processed this int */ unsigned int total_packets; /* total packets processed this int */
u16 work_limit; /* total work allowed per interrupt */ u16 work_limit; /* total work allowed per interrupt */
u16 itr; /* interrupt throttle rate value */ u16 itr; /* interrupt throttle rate value */
u8 itr_scale; /* ITR adjustment scaler based on PCI speed */
u8 count; /* total number of rings in vector */ u8 count; /* total number of rings in vector */
}; };
......
...@@ -1363,7 +1363,7 @@ static bool fm10k_clean_tx_irq(struct fm10k_q_vector *q_vector, ...@@ -1363,7 +1363,7 @@ static bool fm10k_clean_tx_irq(struct fm10k_q_vector *q_vector,
**/ **/
static void fm10k_update_itr(struct fm10k_ring_container *ring_container) static void fm10k_update_itr(struct fm10k_ring_container *ring_container)
{ {
unsigned int avg_wire_size, packets; unsigned int avg_wire_size, packets, itr_round;
/* Only update ITR if we are using adaptive setting */ /* Only update ITR if we are using adaptive setting */
if (!ITR_IS_ADAPTIVE(ring_container->itr)) if (!ITR_IS_ADAPTIVE(ring_container->itr))
...@@ -1375,18 +1375,44 @@ static void fm10k_update_itr(struct fm10k_ring_container *ring_container) ...@@ -1375,18 +1375,44 @@ static void fm10k_update_itr(struct fm10k_ring_container *ring_container)
avg_wire_size = ring_container->total_bytes / packets; avg_wire_size = ring_container->total_bytes / packets;
/* Add 24 bytes to size to account for CRC, preamble, and gap */ /* The following is a crude approximation of:
avg_wire_size += 24; * wmem_default / (size + overhead) = desired_pkts_per_int
* rate / bits_per_byte / (size + ethernet overhead) = pkt_rate
/* Don't starve jumbo frames */ * (desired_pkt_rate / pkt_rate) * usecs_per_sec = ITR value
if (avg_wire_size > 3000) *
avg_wire_size = 3000; * Assuming wmem_default is 212992 and overhead is 640 bytes per
* packet, (256 skb, 64 headroom, 320 shared info), we can reduce the
* formula down to
*
* (34 * (size + 24)) / (size + 640) = ITR
*
* We first do some math on the packet size and then finally bitshift
* by 8 after rounding up. We also have to account for PCIe link speed
* difference as ITR scales based on this.
*/
if (avg_wire_size <= 360) {
/* Start at 250K ints/sec and gradually drop to 77K ints/sec */
avg_wire_size *= 8;
avg_wire_size += 376;
} else if (avg_wire_size <= 1152) {
/* 77K ints/sec to 45K ints/sec */
avg_wire_size *= 3;
avg_wire_size += 2176;
} else if (avg_wire_size <= 1920) {
/* 45K ints/sec to 38K ints/sec */
avg_wire_size += 4480;
} else {
/* plateau at a limit of 38K ints/sec */
avg_wire_size = 6656;
}
/* Give a little boost to mid-size frames */ /* Perform final bitshift for division after rounding up to ensure
if ((avg_wire_size > 300) && (avg_wire_size < 1200)) * that the calculation will never get below a 1. The bit shift
avg_wire_size /= 3; * accounts for changes in the ITR due to PCIe link speed.
else */
avg_wire_size /= 2; itr_round = ACCESS_ONCE(ring_container->itr_scale) + 8;
avg_wire_size += (1 << itr_round) - 1;
avg_wire_size >>= itr_round;
/* write back value and retain adaptive flag */ /* write back value and retain adaptive flag */
ring_container->itr = avg_wire_size | FM10K_ITR_ADAPTIVE; ring_container->itr = avg_wire_size | FM10K_ITR_ADAPTIVE;
...@@ -1604,6 +1630,7 @@ static int fm10k_alloc_q_vector(struct fm10k_intfc *interface, ...@@ -1604,6 +1630,7 @@ static int fm10k_alloc_q_vector(struct fm10k_intfc *interface,
q_vector->tx.ring = ring; q_vector->tx.ring = ring;
q_vector->tx.work_limit = FM10K_DEFAULT_TX_WORK; q_vector->tx.work_limit = FM10K_DEFAULT_TX_WORK;
q_vector->tx.itr = interface->tx_itr; q_vector->tx.itr = interface->tx_itr;
q_vector->tx.itr_scale = interface->hw.mac.itr_scale;
q_vector->tx.count = txr_count; q_vector->tx.count = txr_count;
while (txr_count) { while (txr_count) {
...@@ -1632,6 +1659,7 @@ static int fm10k_alloc_q_vector(struct fm10k_intfc *interface, ...@@ -1632,6 +1659,7 @@ static int fm10k_alloc_q_vector(struct fm10k_intfc *interface,
/* save Rx ring container info */ /* save Rx ring container info */
q_vector->rx.ring = ring; q_vector->rx.ring = ring;
q_vector->rx.itr = interface->rx_itr; q_vector->rx.itr = interface->rx_itr;
q_vector->rx.itr_scale = interface->hw.mac.itr_scale;
q_vector->rx.count = rxr_count; q_vector->rx.count = rxr_count;
while (rxr_count) { while (rxr_count) {
......
...@@ -880,7 +880,8 @@ static irqreturn_t fm10k_msix_mbx_vf(int __always_unused irq, void *data) ...@@ -880,7 +880,8 @@ static irqreturn_t fm10k_msix_mbx_vf(int __always_unused irq, void *data)
/* re-enable mailbox interrupt and indicate 20us delay */ /* re-enable mailbox interrupt and indicate 20us delay */
fm10k_write_reg(hw, FM10K_VFITR(FM10K_MBX_VECTOR), fm10k_write_reg(hw, FM10K_VFITR(FM10K_MBX_VECTOR),
FM10K_ITR_ENABLE | FM10K_MBX_INT_DELAY); FM10K_ITR_ENABLE | (FM10K_MBX_INT_DELAY >>
hw->mac.itr_scale));
/* service upstream mailbox */ /* service upstream mailbox */
if (fm10k_mbx_trylock(interface)) { if (fm10k_mbx_trylock(interface)) {
...@@ -1111,7 +1112,8 @@ static irqreturn_t fm10k_msix_mbx_pf(int __always_unused irq, void *data) ...@@ -1111,7 +1112,8 @@ static irqreturn_t fm10k_msix_mbx_pf(int __always_unused irq, void *data)
/* re-enable mailbox interrupt and indicate 20us delay */ /* re-enable mailbox interrupt and indicate 20us delay */
fm10k_write_reg(hw, FM10K_ITR(FM10K_MBX_VECTOR), fm10k_write_reg(hw, FM10K_ITR(FM10K_MBX_VECTOR),
FM10K_ITR_ENABLE | FM10K_MBX_INT_DELAY); FM10K_ITR_ENABLE | (FM10K_MBX_INT_DELAY >>
hw->mac.itr_scale));
return IRQ_HANDLED; return IRQ_HANDLED;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment