Commit 23bc06af authored by Bob Pearson's avatar Bob Pearson Committed by Jason Gunthorpe

RDMA/rxe: Don't call direct between tasks

Replace calls to rxe_run_task() with rxe_sched_task().  This prevents the
tasks from all running on the same cpu.

This change slightly reduces performance for single qp send and write
benchmarks in loopback mode but greatly improves the performance with
multiple qps because if run task is used all the work tends to be
performed on one cpu. For actual on the wire benchmarks there is no
noticeable performance change.

Link: https://lore.kernel.org/r/20240329145513.35381-11-rpearsonhpe@gmail.comSigned-off-by: default avatarBob Pearson <rpearsonhpe@gmail.com>
Signed-off-by: default avatarJason Gunthorpe <jgg@nvidia.com>
parent 3d807a3e
...@@ -129,18 +129,9 @@ void retransmit_timer(struct timer_list *t) ...@@ -129,18 +129,9 @@ void retransmit_timer(struct timer_list *t)
void rxe_comp_queue_pkt(struct rxe_qp *qp, struct sk_buff *skb) void rxe_comp_queue_pkt(struct rxe_qp *qp, struct sk_buff *skb)
{ {
int must_sched; rxe_counter_inc(SKB_TO_PKT(skb)->rxe, RXE_CNT_SENDER_SCHED);
must_sched = skb_queue_len(&qp->resp_pkts) > 0;
if (must_sched != 0)
rxe_counter_inc(SKB_TO_PKT(skb)->rxe, RXE_CNT_SENDER_SCHED);
skb_queue_tail(&qp->resp_pkts, skb); skb_queue_tail(&qp->resp_pkts, skb);
rxe_sched_task(&qp->send_task);
if (must_sched)
rxe_sched_task(&qp->send_task);
else
rxe_run_task(&qp->send_task);
} }
static inline enum comp_state get_wqe(struct rxe_qp *qp, static inline enum comp_state get_wqe(struct rxe_qp *qp,
......
...@@ -49,18 +49,8 @@ static char *resp_state_name[] = { ...@@ -49,18 +49,8 @@ static char *resp_state_name[] = {
/* rxe_recv calls here to add a request packet to the input queue */ /* rxe_recv calls here to add a request packet to the input queue */
void rxe_resp_queue_pkt(struct rxe_qp *qp, struct sk_buff *skb) void rxe_resp_queue_pkt(struct rxe_qp *qp, struct sk_buff *skb)
{ {
int must_sched;
struct rxe_pkt_info *pkt = SKB_TO_PKT(skb);
skb_queue_tail(&qp->req_pkts, skb); skb_queue_tail(&qp->req_pkts, skb);
rxe_sched_task(&qp->recv_task);
must_sched = (pkt->opcode == IB_OPCODE_RC_RDMA_READ_REQUEST) ||
(skb_queue_len(&qp->req_pkts) > 1);
if (must_sched)
rxe_sched_task(&qp->recv_task);
else
rxe_run_task(&qp->recv_task);
} }
static inline enum resp_states get_req(struct rxe_qp *qp, static inline enum resp_states get_req(struct rxe_qp *qp,
......
...@@ -935,7 +935,7 @@ static int rxe_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, ...@@ -935,7 +935,7 @@ static int rxe_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
if (qp->is_user) { if (qp->is_user) {
/* Utilize process context to do protocol processing */ /* Utilize process context to do protocol processing */
rxe_run_task(&qp->send_task); rxe_sched_task(&qp->send_task);
} else { } else {
err = rxe_post_send_kernel(qp, wr, bad_wr); err = rxe_post_send_kernel(qp, wr, bad_wr);
if (err) if (err)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment