Commit 1900e30d authored by Robert Hancock's avatar Robert Hancock Committed by Jakub Kicinski

net: macb: simplify/cleanup NAPI reschedule checking

Previously the macb_poll method was checking the RSR register after
completing its RX receive work to see if additional packets had been
received since IRQs were disabled, since this controller does not
maintain the pending IRQ status across IRQ disable. It also had to
double-check the register after re-enabling IRQs to detect if packets
were received after the first check but before IRQs were enabled.

Using the RSR register for this purpose is problematic since it reflects
the global device state rather than the per-queue state, so if packets
are being received on multiple queues it may end up retriggering receive
on a queue where the packets did not actually arrive and not on the one
where they did arrive. This will also cause problems with an upcoming
change to use NAPI for the TX path where use of multiple queues is more
likely.

Add a macb_rx_pending function to check the RX ring to see if more
packets have arrived in the queue, and use that to check if NAPI should
be rescheduled rather than the RSR register. By doing this, we can just
ignore the global RSR register entirely, and thus save some extra device
register accesses at the same time.

This also makes the previous first check for pending packets rather
redundant, since it would be checking the RX ring state which was just
checked in the receive work function. Therefore we can get rid of it and
just check after enabling interrupts whether packets are already
pending.
Signed-off-by: default avatarRobert Hancock <robert.hancock@calian.com>
Signed-off-by: default avatarJakub Kicinski <kuba@kernel.org>
parent 11ecf341
...@@ -1554,54 +1554,51 @@ static int macb_rx(struct macb_queue *queue, struct napi_struct *napi, ...@@ -1554,54 +1554,51 @@ static int macb_rx(struct macb_queue *queue, struct napi_struct *napi,
return received; return received;
} }
static bool macb_rx_pending(struct macb_queue *queue)
{
struct macb *bp = queue->bp;
unsigned int entry;
struct macb_dma_desc *desc;
entry = macb_rx_ring_wrap(bp, queue->rx_tail);
desc = macb_rx_desc(queue, entry);
/* Make hw descriptor updates visible to CPU */
rmb();
return (desc->addr & MACB_BIT(RX_USED)) != 0;
}
static int macb_poll(struct napi_struct *napi, int budget) static int macb_poll(struct napi_struct *napi, int budget)
{ {
struct macb_queue *queue = container_of(napi, struct macb_queue, napi); struct macb_queue *queue = container_of(napi, struct macb_queue, napi);
struct macb *bp = queue->bp; struct macb *bp = queue->bp;
int work_done; int work_done;
u32 status;
status = macb_readl(bp, RSR); work_done = bp->macbgem_ops.mog_rx(queue, napi, budget);
macb_writel(bp, RSR, status);
netdev_vdbg(bp->dev, "poll: status = %08lx, budget = %d\n", netdev_vdbg(bp->dev, "poll: queue = %u, work_done = %d, budget = %d\n",
(unsigned long)status, budget); (unsigned int)(queue - bp->queues), work_done, budget);
work_done = bp->macbgem_ops.mog_rx(queue, napi, budget); if (work_done < budget && napi_complete_done(napi, work_done)) {
if (work_done < budget) { queue_writel(queue, IER, bp->rx_intr_mask);
napi_complete_done(napi, work_done);
/* RSR bits only seem to propagate to raise interrupts when /* Packet completions only seem to propagate to raise
* interrupts are enabled at the time, so if bits are already * interrupts when interrupts are enabled at the time, so if
* set due to packets received while interrupts were disabled, * packets were received while interrupts were disabled,
* they will not cause another interrupt to be generated when * they will not cause another interrupt to be generated when
* interrupts are re-enabled. * interrupts are re-enabled.
* Check for this case here. This has been seen to happen * Check for this case here to avoid losing a wakeup. This can
* around 30% of the time under heavy network load. * potentially race with the interrupt handler doing the same
* actions if an interrupt is raised just after enabling them,
* but this should be harmless.
*/ */
status = macb_readl(bp, RSR); if (macb_rx_pending(queue)) {
if (status) { queue_writel(queue, IDR, bp->rx_intr_mask);
if (bp->caps & MACB_CAPS_ISR_CLEAR_ON_WRITE) if (bp->caps & MACB_CAPS_ISR_CLEAR_ON_WRITE)
queue_writel(queue, ISR, MACB_BIT(RCOMP)); queue_writel(queue, ISR, MACB_BIT(RCOMP));
napi_reschedule(napi); netdev_vdbg(bp->dev, "poll: packets pending, reschedule\n");
} else { napi_schedule(napi);
queue_writel(queue, IER, bp->rx_intr_mask);
/* In rare cases, packets could have been received in
* the window between the check above and re-enabling
* interrupts. Therefore, a double-check is required
* to avoid losing a wakeup. This can potentially race
* with the interrupt handler doing the same actions
* if an interrupt is raised just after enabling them,
* but this should be harmless.
*/
status = macb_readl(bp, RSR);
if (unlikely(status)) {
queue_writel(queue, IDR, bp->rx_intr_mask);
if (bp->caps & MACB_CAPS_ISR_CLEAR_ON_WRITE)
queue_writel(queue, ISR, MACB_BIT(RCOMP));
napi_schedule(napi);
}
} }
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment