Message ID | 20240722110325.195085-1-saravanan.vajravel@broadcom.com (mailing list archive) |
---|---|
State | Accepted |
Headers | show |
Series | [for-rc] rdma-core/mad: Improve handling of timed out WRs of mad agent | expand |
On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote: > Current timeout handler of mad agent aquires/releases mad_agent_priv > lock for every timed out WRs. This causes heavy locking contention > when higher no. of WRs are to be handled inside timeout handler. > > This leads to softlockup with below trace in some use cases where > rdma-cm path is used to establish connection between peer nodes > > Trace: > ----- > BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] > CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE > ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 > Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 > Workqueue: ib_mad1 timeout_sends [ib_core] > RIP: 0010:__do_softirq+0x78/0x2ac > RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 > RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f > RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b > RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 > R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 > FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > PKRU: 55555554 > Call Trace: > <IRQ> > ? show_trace_log_lvl+0x1c4/0x2df > ? show_trace_log_lvl+0x1c4/0x2df > ? __irq_exit_rcu+0xa1/0xc0 > ? watchdog_timer_fn+0x1b2/0x210 > ? __pfx_watchdog_timer_fn+0x10/0x10 > ? __hrtimer_run_queues+0x127/0x2c0 > ? hrtimer_interrupt+0xfc/0x210 > ? __sysvec_apic_timer_interrupt+0x5c/0x110 > ? sysvec_apic_timer_interrupt+0x37/0x90 > ? asm_sysvec_apic_timer_interrupt+0x16/0x20 > ? __do_softirq+0x78/0x2ac > ? __do_softirq+0x60/0x2ac > __irq_exit_rcu+0xa1/0xc0 > sysvec_call_function_single+0x72/0x90 > </IRQ> > <TASK> > asm_sysvec_call_function_single+0x16/0x20 > RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 > RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 > RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 > RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c > RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 > R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 > R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c > cm_process_send_error+0x122/0x1d0 [ib_cm] > timeout_sends+0x1dd/0x270 [ib_core] > process_one_work+0x1e2/0x3b0 > ? __pfx_worker_thread+0x10/0x10 > worker_thread+0x50/0x3a0 > ? __pfx_worker_thread+0x10/0x10 > kthread+0xdd/0x100 > ? __pfx_kthread+0x10/0x10 > ret_from_fork+0x29/0x50 > </TASK> > > Simplified timeout handler by creating local list of timed out WRs > and invoke send handler post creating the list. The new method acquires/ > releases lock once to fetch the list and hence helps to reduce locking > contetiong when processing higher no. of WRs > > Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com> > --- > drivers/infiniband/core/mad.c | 14 ++++++++------ > 1 file changed, 8 insertions(+), 6 deletions(-) > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c > index 674344eb8e2f..58befbaaf0ad 100644 > --- a/drivers/infiniband/core/mad.c > +++ b/drivers/infiniband/core/mad.c > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) > > static void timeout_sends(struct work_struct *work) > { > + struct ib_mad_send_wr_private *mad_send_wr, *n; > struct ib_mad_agent_private *mad_agent_priv; > - struct ib_mad_send_wr_private *mad_send_wr; > struct ib_mad_send_wc mad_send_wc; > + struct list_head local_list; > unsigned long flags, delay; > > mad_agent_priv = container_of(work, struct ib_mad_agent_private, > timed_work.work); > mad_send_wc.vendor_err = 0; > + INIT_LIST_HEAD(&local_list); > > spin_lock_irqsave(&mad_agent_priv->lock, flags); > while (!list_empty(&mad_agent_priv->wait_list)) { > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) > break; > } > > - list_del(&mad_send_wr->agent_list); > + list_del_init(&mad_send_wr->agent_list); > if (mad_send_wr->status == IB_WC_SUCCESS && > !retry_send(mad_send_wr)) > continue; > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > + list_add_tail(&mad_send_wr->agent_list, &local_list); > + } > + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { > if (mad_send_wr->status == IB_WC_SUCCESS) > mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > else > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) > mad_send_wc.send_buf = &mad_send_wr->send_buf; > mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, > &mad_send_wc); I understand the problem, but I'm not sure that this is safe to do. How can we be sure that this is safe to call the send_handler on entry in wait_list without the locking? Thanks > - > deref_mad_agent(mad_agent_priv); > - spin_lock_irqsave(&mad_agent_priv->lock, flags); > } > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > } > > /* > -- > 2.39.3 >
On Sun, Jul 28, 2024 at 1:08 PM Leon Romanovsky <leon@kernel.org> wrote: > > On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote: > > Current timeout handler of mad agent aquires/releases mad_agent_priv > > lock for every timed out WRs. This causes heavy locking contention > > when higher no. of WRs are to be handled inside timeout handler. > > > > This leads to softlockup with below trace in some use cases where > > rdma-cm path is used to establish connection between peer nodes > > > > Trace: > > ----- > > BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] > > CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE > > ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 > > Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 > > Workqueue: ib_mad1 timeout_sends [ib_core] > > RIP: 0010:__do_softirq+0x78/0x2ac > > RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 > > RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f > > RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b > > RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 > > R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 > > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 > > FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 > > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 > > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > PKRU: 55555554 > > Call Trace: > > <IRQ> > > ? show_trace_log_lvl+0x1c4/0x2df > > ? show_trace_log_lvl+0x1c4/0x2df > > ? __irq_exit_rcu+0xa1/0xc0 > > ? watchdog_timer_fn+0x1b2/0x210 > > ? __pfx_watchdog_timer_fn+0x10/0x10 > > ? __hrtimer_run_queues+0x127/0x2c0 > > ? hrtimer_interrupt+0xfc/0x210 > > ? __sysvec_apic_timer_interrupt+0x5c/0x110 > > ? sysvec_apic_timer_interrupt+0x37/0x90 > > ? asm_sysvec_apic_timer_interrupt+0x16/0x20 > > ? __do_softirq+0x78/0x2ac > > ? __do_softirq+0x60/0x2ac > > __irq_exit_rcu+0xa1/0xc0 > > sysvec_call_function_single+0x72/0x90 > > </IRQ> > > <TASK> > > asm_sysvec_call_function_single+0x16/0x20 > > RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 > > RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 > > RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 > > RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c > > RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 > > R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 > > R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c > > cm_process_send_error+0x122/0x1d0 [ib_cm] > > timeout_sends+0x1dd/0x270 [ib_core] > > process_one_work+0x1e2/0x3b0 > > ? __pfx_worker_thread+0x10/0x10 > > worker_thread+0x50/0x3a0 > > ? __pfx_worker_thread+0x10/0x10 > > kthread+0xdd/0x100 > > ? __pfx_kthread+0x10/0x10 > > ret_from_fork+0x29/0x50 > > </TASK> > > > > Simplified timeout handler by creating local list of timed out WRs > > and invoke send handler post creating the list. The new method acquires/ > > releases lock once to fetch the list and hence helps to reduce locking > > contetiong when processing higher no. of WRs > > > > Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com> > > --- > > drivers/infiniband/core/mad.c | 14 ++++++++------ > > 1 file changed, 8 insertions(+), 6 deletions(-) > > > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c > > index 674344eb8e2f..58befbaaf0ad 100644 > > --- a/drivers/infiniband/core/mad.c > > +++ b/drivers/infiniband/core/mad.c > > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) > > > > static void timeout_sends(struct work_struct *work) > > { > > + struct ib_mad_send_wr_private *mad_send_wr, *n; > > struct ib_mad_agent_private *mad_agent_priv; > > - struct ib_mad_send_wr_private *mad_send_wr; > > struct ib_mad_send_wc mad_send_wc; > > + struct list_head local_list; > > unsigned long flags, delay; > > > > mad_agent_priv = container_of(work, struct ib_mad_agent_private, > > timed_work.work); > > mad_send_wc.vendor_err = 0; > > + INIT_LIST_HEAD(&local_list); > > > > spin_lock_irqsave(&mad_agent_priv->lock, flags); > > while (!list_empty(&mad_agent_priv->wait_list)) { > > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) > > break; > > } > > > > - list_del(&mad_send_wr->agent_list); > > + list_del_init(&mad_send_wr->agent_list); > > if (mad_send_wr->status == IB_WC_SUCCESS && > > !retry_send(mad_send_wr)) > > continue; > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > + list_add_tail(&mad_send_wr->agent_list, &local_list); > > + } > > + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { > > if (mad_send_wr->status == IB_WC_SUCCESS) > > mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > > else > > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) > > mad_send_wc.send_buf = &mad_send_wr->send_buf; > > mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, > > &mad_send_wc); > > I understand the problem, but I'm not sure that this is safe to do. How > can we be sure that this is safe to call the send_handler on entry in > wait_list without the locking? > > Thanks Per existing implementation, the send_handler is invoked without locking. I didn't change that design. Existing implementation was acquiring and releasing the lock for every Work Request (WR) that it handles inside timeout_handler. I improved it in such a way that once lock is acquired to fetch list of WR to be handled and process all fetched WRs outside the scope of lock. In both implementations, send_handler is always called without lock. This patch aims to reduce locking contention Thanks, > > > - > > deref_mad_agent(mad_agent_priv); > > - spin_lock_irqsave(&mad_agent_priv->lock, flags); > > } > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > } > > > > /* > > -- > > 2.39.3 > >
On Mon, Jul 29, 2024 at 09:27:35AM +0530, Saravanan Vajravel wrote: > On Sun, Jul 28, 2024 at 1:08 PM Leon Romanovsky <leon@kernel.org> wrote: > > > > On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote: > > > Current timeout handler of mad agent aquires/releases mad_agent_priv > > > lock for every timed out WRs. This causes heavy locking contention > > > when higher no. of WRs are to be handled inside timeout handler. > > > > > > This leads to softlockup with below trace in some use cases where > > > rdma-cm path is used to establish connection between peer nodes > > > > > > Trace: > > > ----- > > > BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] > > > CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE > > > ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 > > > Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 > > > Workqueue: ib_mad1 timeout_sends [ib_core] > > > RIP: 0010:__do_softirq+0x78/0x2ac > > > RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 > > > RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f > > > RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b > > > RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 > > > R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 > > > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 > > > FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 > > > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > > CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 > > > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > > DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > > PKRU: 55555554 > > > Call Trace: > > > <IRQ> > > > ? show_trace_log_lvl+0x1c4/0x2df > > > ? show_trace_log_lvl+0x1c4/0x2df > > > ? __irq_exit_rcu+0xa1/0xc0 > > > ? watchdog_timer_fn+0x1b2/0x210 > > > ? __pfx_watchdog_timer_fn+0x10/0x10 > > > ? __hrtimer_run_queues+0x127/0x2c0 > > > ? hrtimer_interrupt+0xfc/0x210 > > > ? __sysvec_apic_timer_interrupt+0x5c/0x110 > > > ? sysvec_apic_timer_interrupt+0x37/0x90 > > > ? asm_sysvec_apic_timer_interrupt+0x16/0x20 > > > ? __do_softirq+0x78/0x2ac > > > ? __do_softirq+0x60/0x2ac > > > __irq_exit_rcu+0xa1/0xc0 > > > sysvec_call_function_single+0x72/0x90 > > > </IRQ> > > > <TASK> > > > asm_sysvec_call_function_single+0x16/0x20 > > > RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 > > > RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 > > > RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 > > > RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c > > > RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 > > > R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 > > > R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c > > > cm_process_send_error+0x122/0x1d0 [ib_cm] > > > timeout_sends+0x1dd/0x270 [ib_core] > > > process_one_work+0x1e2/0x3b0 > > > ? __pfx_worker_thread+0x10/0x10 > > > worker_thread+0x50/0x3a0 > > > ? __pfx_worker_thread+0x10/0x10 > > > kthread+0xdd/0x100 > > > ? __pfx_kthread+0x10/0x10 > > > ret_from_fork+0x29/0x50 > > > </TASK> > > > > > > Simplified timeout handler by creating local list of timed out WRs > > > and invoke send handler post creating the list. The new method acquires/ > > > releases lock once to fetch the list and hence helps to reduce locking > > > contetiong when processing higher no. of WRs > > > > > > Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com> > > > --- > > > drivers/infiniband/core/mad.c | 14 ++++++++------ > > > 1 file changed, 8 insertions(+), 6 deletions(-) > > > > > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c > > > index 674344eb8e2f..58befbaaf0ad 100644 > > > --- a/drivers/infiniband/core/mad.c > > > +++ b/drivers/infiniband/core/mad.c > > > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) > > > > > > static void timeout_sends(struct work_struct *work) > > > { > > > + struct ib_mad_send_wr_private *mad_send_wr, *n; > > > struct ib_mad_agent_private *mad_agent_priv; > > > - struct ib_mad_send_wr_private *mad_send_wr; > > > struct ib_mad_send_wc mad_send_wc; > > > + struct list_head local_list; > > > unsigned long flags, delay; > > > > > > mad_agent_priv = container_of(work, struct ib_mad_agent_private, > > > timed_work.work); > > > mad_send_wc.vendor_err = 0; > > > + INIT_LIST_HEAD(&local_list); > > > > > > spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > while (!list_empty(&mad_agent_priv->wait_list)) { > > > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) > > > break; > > > } > > > > > > - list_del(&mad_send_wr->agent_list); > > > + list_del_init(&mad_send_wr->agent_list); > > > if (mad_send_wr->status == IB_WC_SUCCESS && > > > !retry_send(mad_send_wr)) > > > continue; > > > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > + list_add_tail(&mad_send_wr->agent_list, &local_list); > > > + } > > > + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > > + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { > > > if (mad_send_wr->status == IB_WC_SUCCESS) > > > mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > > > else > > > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) > > > mad_send_wc.send_buf = &mad_send_wr->send_buf; > > > mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, > > > &mad_send_wc); > > > > I understand the problem, but I'm not sure that this is safe to do. How > > can we be sure that this is safe to call the send_handler on entry in > > wait_list without the locking? > > > > Thanks > > Per existing implementation, the send_handler is invoked without locking. > I didn't change that design. Existing implementation was acquiring and > releasing the lock for every Work Request (WR) that it handles inside > timeout_handler. > I improved it in such a way that once lock is acquired to fetch list > of WR to be handled > and process all fetched WRs outside the scope of lock. In both implementations, > send_handler is always called without lock. This patch aims to reduce locking > contention Indeed send_handler is called without lock, but not on the wait_list. I've asked explicitly about operations on the wait_list. Thanks > > Thanks, > > > > > - > > > deref_mad_agent(mad_agent_priv); > > > - spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > } > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > } > > > > > > /* > > > -- > > > 2.39.3 > > >
On Mon, Jul 29, 2024 at 11:40 AM Leon Romanovsky <leon@kernel.org> wrote: > > On Mon, Jul 29, 2024 at 09:27:35AM +0530, Saravanan Vajravel wrote: > > On Sun, Jul 28, 2024 at 1:08 PM Leon Romanovsky <leon@kernel.org> wrote: > > > > > > On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote: > > > > Current timeout handler of mad agent aquires/releases mad_agent_priv > > > > lock for every timed out WRs. This causes heavy locking contention > > > > when higher no. of WRs are to be handled inside timeout handler. > > > > > > > > This leads to softlockup with below trace in some use cases where > > > > rdma-cm path is used to establish connection between peer nodes > > > > > > > > Trace: > > > > ----- > > > > BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] > > > > CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE > > > > ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 > > > > Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 > > > > Workqueue: ib_mad1 timeout_sends [ib_core] > > > > RIP: 0010:__do_softirq+0x78/0x2ac > > > > RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 > > > > RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f > > > > RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b > > > > RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 > > > > R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 > > > > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 > > > > FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 > > > > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > > > CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 > > > > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > > > DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > > > PKRU: 55555554 > > > > Call Trace: > > > > <IRQ> > > > > ? show_trace_log_lvl+0x1c4/0x2df > > > > ? show_trace_log_lvl+0x1c4/0x2df > > > > ? __irq_exit_rcu+0xa1/0xc0 > > > > ? watchdog_timer_fn+0x1b2/0x210 > > > > ? __pfx_watchdog_timer_fn+0x10/0x10 > > > > ? __hrtimer_run_queues+0x127/0x2c0 > > > > ? hrtimer_interrupt+0xfc/0x210 > > > > ? __sysvec_apic_timer_interrupt+0x5c/0x110 > > > > ? sysvec_apic_timer_interrupt+0x37/0x90 > > > > ? asm_sysvec_apic_timer_interrupt+0x16/0x20 > > > > ? __do_softirq+0x78/0x2ac > > > > ? __do_softirq+0x60/0x2ac > > > > __irq_exit_rcu+0xa1/0xc0 > > > > sysvec_call_function_single+0x72/0x90 > > > > </IRQ> > > > > <TASK> > > > > asm_sysvec_call_function_single+0x16/0x20 > > > > RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 > > > > RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 > > > > RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 > > > > RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c > > > > RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 > > > > R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 > > > > R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c > > > > cm_process_send_error+0x122/0x1d0 [ib_cm] > > > > timeout_sends+0x1dd/0x270 [ib_core] > > > > process_one_work+0x1e2/0x3b0 > > > > ? __pfx_worker_thread+0x10/0x10 > > > > worker_thread+0x50/0x3a0 > > > > ? __pfx_worker_thread+0x10/0x10 > > > > kthread+0xdd/0x100 > > > > ? __pfx_kthread+0x10/0x10 > > > > ret_from_fork+0x29/0x50 > > > > </TASK> > > > > > > > > Simplified timeout handler by creating local list of timed out WRs > > > > and invoke send handler post creating the list. The new method acquires/ > > > > releases lock once to fetch the list and hence helps to reduce locking > > > > contetiong when processing higher no. of WRs > > > > > > > > Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com> > > > > --- > > > > drivers/infiniband/core/mad.c | 14 ++++++++------ > > > > 1 file changed, 8 insertions(+), 6 deletions(-) > > > > > > > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c > > > > index 674344eb8e2f..58befbaaf0ad 100644 > > > > --- a/drivers/infiniband/core/mad.c > > > > +++ b/drivers/infiniband/core/mad.c > > > > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) > > > > > > > > static void timeout_sends(struct work_struct *work) > > > > { > > > > + struct ib_mad_send_wr_private *mad_send_wr, *n; > > > > struct ib_mad_agent_private *mad_agent_priv; > > > > - struct ib_mad_send_wr_private *mad_send_wr; > > > > struct ib_mad_send_wc mad_send_wc; > > > > + struct list_head local_list; > > > > unsigned long flags, delay; > > > > > > > > mad_agent_priv = container_of(work, struct ib_mad_agent_private, > > > > timed_work.work); > > > > mad_send_wc.vendor_err = 0; > > > > + INIT_LIST_HEAD(&local_list); > > > > > > > > spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > > while (!list_empty(&mad_agent_priv->wait_list)) { > > > > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) > > > > break; > > > > } > > > > > > > > - list_del(&mad_send_wr->agent_list); > > > > + list_del_init(&mad_send_wr->agent_list); > > > > if (mad_send_wr->status == IB_WC_SUCCESS && > > > > !retry_send(mad_send_wr)) > > > > continue; > > > > > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > + list_add_tail(&mad_send_wr->agent_list, &local_list); > > > > + } > > > > + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > > > > + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { > > > > if (mad_send_wr->status == IB_WC_SUCCESS) > > > > mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > > > > else > > > > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) > > > > mad_send_wc.send_buf = &mad_send_wr->send_buf; > > > > mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, > > > > &mad_send_wc); > > > > > > I understand the problem, but I'm not sure that this is safe to do. How > > > can we be sure that this is safe to call the send_handler on entry in > > > wait_list without the locking? > > > > > > Thanks > > > > Per existing implementation, the send_handler is invoked without locking. > > I didn't change that design. Existing implementation was acquiring and > > releasing the lock for every Work Request (WR) that it handles inside > > timeout_handler. > > I improved it in such a way that once lock is acquired to fetch list > > of WR to be handled > > and process all fetched WRs outside the scope of lock. In both implementations, > > send_handler is always called without lock. This patch aims to reduce locking > > contention > > Indeed send_handler is called without lock, but not on the wait_list. > I've asked explicitly about operations on the wait_list. > > Thanks The entry in wait_list is removed under the scope of mad_agent_priv->lock. It is then added to local_list declared inside timeout_sends() handler under the same lock. Once all such entries are removed and added to the local_list, this timeout_handler processes each entry from the local_list and it invokes send_handler. This local_list doesn't need any locking as there is no possibility of race condition. Existing implementation also removed entry from wait_list and invoked send_handler. So operation on the wait_list should not change in the new patch Thanks, > > > > > > > - > > > > deref_mad_agent(mad_agent_priv); > > > > - spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > > } > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > } > > > > > > > > /* > > > > -- > > > > 2.39.3 > > > > > >
On Mon, Jul 29, 2024 at 12:08:52PM +0530, Saravanan Vajravel wrote: > On Mon, Jul 29, 2024 at 11:40 AM Leon Romanovsky <leon@kernel.org> wrote: > > > > On Mon, Jul 29, 2024 at 09:27:35AM +0530, Saravanan Vajravel wrote: > > > On Sun, Jul 28, 2024 at 1:08 PM Leon Romanovsky <leon@kernel.org> wrote: > > > > > > > > On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote: > > > > > Current timeout handler of mad agent aquires/releases mad_agent_priv > > > > > lock for every timed out WRs. This causes heavy locking contention > > > > > when higher no. of WRs are to be handled inside timeout handler. > > > > > > > > > > This leads to softlockup with below trace in some use cases where > > > > > rdma-cm path is used to establish connection between peer nodes > > > > > > > > > > Trace: > > > > > ----- > > > > > BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] > > > > > CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE > > > > > ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 > > > > > Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 > > > > > Workqueue: ib_mad1 timeout_sends [ib_core] > > > > > RIP: 0010:__do_softirq+0x78/0x2ac > > > > > RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 > > > > > RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f > > > > > RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b > > > > > RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 > > > > > R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 > > > > > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 > > > > > FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 > > > > > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > > > > CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 > > > > > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > > > > DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > > > > PKRU: 55555554 > > > > > Call Trace: > > > > > <IRQ> > > > > > ? show_trace_log_lvl+0x1c4/0x2df > > > > > ? show_trace_log_lvl+0x1c4/0x2df > > > > > ? __irq_exit_rcu+0xa1/0xc0 > > > > > ? watchdog_timer_fn+0x1b2/0x210 > > > > > ? __pfx_watchdog_timer_fn+0x10/0x10 > > > > > ? __hrtimer_run_queues+0x127/0x2c0 > > > > > ? hrtimer_interrupt+0xfc/0x210 > > > > > ? __sysvec_apic_timer_interrupt+0x5c/0x110 > > > > > ? sysvec_apic_timer_interrupt+0x37/0x90 > > > > > ? asm_sysvec_apic_timer_interrupt+0x16/0x20 > > > > > ? __do_softirq+0x78/0x2ac > > > > > ? __do_softirq+0x60/0x2ac > > > > > __irq_exit_rcu+0xa1/0xc0 > > > > > sysvec_call_function_single+0x72/0x90 > > > > > </IRQ> > > > > > <TASK> > > > > > asm_sysvec_call_function_single+0x16/0x20 > > > > > RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 > > > > > RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 > > > > > RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 > > > > > RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c > > > > > RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 > > > > > R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 > > > > > R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c > > > > > cm_process_send_error+0x122/0x1d0 [ib_cm] > > > > > timeout_sends+0x1dd/0x270 [ib_core] > > > > > process_one_work+0x1e2/0x3b0 > > > > > ? __pfx_worker_thread+0x10/0x10 > > > > > worker_thread+0x50/0x3a0 > > > > > ? __pfx_worker_thread+0x10/0x10 > > > > > kthread+0xdd/0x100 > > > > > ? __pfx_kthread+0x10/0x10 > > > > > ret_from_fork+0x29/0x50 > > > > > </TASK> > > > > > > > > > > Simplified timeout handler by creating local list of timed out WRs > > > > > and invoke send handler post creating the list. The new method acquires/ > > > > > releases lock once to fetch the list and hence helps to reduce locking > > > > > contetiong when processing higher no. of WRs > > > > > > > > > > Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com> > > > > > --- > > > > > drivers/infiniband/core/mad.c | 14 ++++++++------ > > > > > 1 file changed, 8 insertions(+), 6 deletions(-) > > > > > > > > > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c > > > > > index 674344eb8e2f..58befbaaf0ad 100644 > > > > > --- a/drivers/infiniband/core/mad.c > > > > > +++ b/drivers/infiniband/core/mad.c > > > > > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) > > > > > > > > > > static void timeout_sends(struct work_struct *work) > > > > > { > > > > > + struct ib_mad_send_wr_private *mad_send_wr, *n; > > > > > struct ib_mad_agent_private *mad_agent_priv; > > > > > - struct ib_mad_send_wr_private *mad_send_wr; > > > > > struct ib_mad_send_wc mad_send_wc; > > > > > + struct list_head local_list; > > > > > unsigned long flags, delay; > > > > > > > > > > mad_agent_priv = container_of(work, struct ib_mad_agent_private, > > > > > timed_work.work); > > > > > mad_send_wc.vendor_err = 0; > > > > > + INIT_LIST_HEAD(&local_list); > > > > > > > > > > spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > > > while (!list_empty(&mad_agent_priv->wait_list)) { > > > > > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) > > > > > break; > > > > > } > > > > > > > > > > - list_del(&mad_send_wr->agent_list); > > > > > + list_del_init(&mad_send_wr->agent_list); > > > > > if (mad_send_wr->status == IB_WC_SUCCESS && > > > > > !retry_send(mad_send_wr)) > > > > > continue; > > > > > > > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > + list_add_tail(&mad_send_wr->agent_list, &local_list); > > > > > + } > > > > > + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > > > > > > + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { > > > > > if (mad_send_wr->status == IB_WC_SUCCESS) > > > > > mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > > > > > else > > > > > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) > > > > > mad_send_wc.send_buf = &mad_send_wr->send_buf; > > > > > mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, > > > > > &mad_send_wc); > > > > > > > > I understand the problem, but I'm not sure that this is safe to do. How > > > > can we be sure that this is safe to call the send_handler on entry in > > > > wait_list without the locking? > > > > > > > > Thanks > > > > > > Per existing implementation, the send_handler is invoked without locking. > > > I didn't change that design. Existing implementation was acquiring and > > > releasing the lock for every Work Request (WR) that it handles inside > > > timeout_handler. > > > I improved it in such a way that once lock is acquired to fetch list > > > of WR to be handled > > > and process all fetched WRs outside the scope of lock. In both implementations, > > > send_handler is always called without lock. This patch aims to reduce locking > > > contention > > > > Indeed send_handler is called without lock, but not on the wait_list. > > I've asked explicitly about operations on the wait_list. > > > > Thanks > > The entry in wait_list is removed under the scope of mad_agent_priv->lock. > It is then added to local_list declared inside timeout_sends() handler > under the same > lock. Once all such entries are removed and added to the local_list, > this timeout_handler > processes each entry from the local_list and it invokes send_handler. > This local_list > doesn't need any locking as there is no possibility of race condition. > > Existing implementation also removed entry from wait_list and invoked > send_handler. Yes, and what is the difference between the current and proposed code? 2649 spin_unlock_irqrestore(&mad_agent_priv->lock, flags); <----- unlock 2650 2651 if (mad_send_wr->status == IB_WC_SUCCESS) 2652 mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; 2653 else 2654 mad_send_wc.status = mad_send_wr->status; 2655 mad_send_wc.send_buf = &mad_send_wr->send_buf; 2656 mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, <-- performed without lock 2657 &mad_send_wc); 2658 2659 deref_mad_agent(mad_agent_priv); 2660 spin_lock_irqsave(&mad_agent_priv->lock, flags); 2661 } 2662 spin_unlock_irqrestore(&mad_agent_priv->lock, flags); <-- lock again. > So operation on the wait_list should not change in the new patch > > Thanks, > > > > > > > > > > - > > > > > deref_mad_agent(mad_agent_priv); > > > > > - spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > > > } > > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > } > > > > > > > > > > /* > > > > > -- > > > > > 2.39.3 > > > > > > > > >
On Mon, Jul 29, 2024 at 12:27 PM Leon Romanovsky <leon@kernel.org> wrote: > > On Mon, Jul 29, 2024 at 12:08:52PM +0530, Saravanan Vajravel wrote: > > On Mon, Jul 29, 2024 at 11:40 AM Leon Romanovsky <leon@kernel.org> wrote: > > > > > > On Mon, Jul 29, 2024 at 09:27:35AM +0530, Saravanan Vajravel wrote: > > > > On Sun, Jul 28, 2024 at 1:08 PM Leon Romanovsky <leon@kernel.org> wrote: > > > > > > > > > > On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote: > > > > > > Current timeout handler of mad agent aquires/releases mad_agent_priv > > > > > > lock for every timed out WRs. This causes heavy locking contention > > > > > > when higher no. of WRs are to be handled inside timeout handler. > > > > > > > > > > > > This leads to softlockup with below trace in some use cases where > > > > > > rdma-cm path is used to establish connection between peer nodes > > > > > > > > > > > > Trace: > > > > > > ----- > > > > > > BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] > > > > > > CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE > > > > > > ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 > > > > > > Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 > > > > > > Workqueue: ib_mad1 timeout_sends [ib_core] > > > > > > RIP: 0010:__do_softirq+0x78/0x2ac > > > > > > RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 > > > > > > RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f > > > > > > RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b > > > > > > RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 > > > > > > R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 > > > > > > R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 > > > > > > FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 > > > > > > CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > > > > > CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 > > > > > > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > > > > > DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > > > > > PKRU: 55555554 > > > > > > Call Trace: > > > > > > <IRQ> > > > > > > ? show_trace_log_lvl+0x1c4/0x2df > > > > > > ? show_trace_log_lvl+0x1c4/0x2df > > > > > > ? __irq_exit_rcu+0xa1/0xc0 > > > > > > ? watchdog_timer_fn+0x1b2/0x210 > > > > > > ? __pfx_watchdog_timer_fn+0x10/0x10 > > > > > > ? __hrtimer_run_queues+0x127/0x2c0 > > > > > > ? hrtimer_interrupt+0xfc/0x210 > > > > > > ? __sysvec_apic_timer_interrupt+0x5c/0x110 > > > > > > ? sysvec_apic_timer_interrupt+0x37/0x90 > > > > > > ? asm_sysvec_apic_timer_interrupt+0x16/0x20 > > > > > > ? __do_softirq+0x78/0x2ac > > > > > > ? __do_softirq+0x60/0x2ac > > > > > > __irq_exit_rcu+0xa1/0xc0 > > > > > > sysvec_call_function_single+0x72/0x90 > > > > > > </IRQ> > > > > > > <TASK> > > > > > > asm_sysvec_call_function_single+0x16/0x20 > > > > > > RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 > > > > > > RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 > > > > > > RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 > > > > > > RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c > > > > > > RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 > > > > > > R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 > > > > > > R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c > > > > > > cm_process_send_error+0x122/0x1d0 [ib_cm] > > > > > > timeout_sends+0x1dd/0x270 [ib_core] > > > > > > process_one_work+0x1e2/0x3b0 > > > > > > ? __pfx_worker_thread+0x10/0x10 > > > > > > worker_thread+0x50/0x3a0 > > > > > > ? __pfx_worker_thread+0x10/0x10 > > > > > > kthread+0xdd/0x100 > > > > > > ? __pfx_kthread+0x10/0x10 > > > > > > ret_from_fork+0x29/0x50 > > > > > > </TASK> > > > > > > > > > > > > Simplified timeout handler by creating local list of timed out WRs > > > > > > and invoke send handler post creating the list. The new method acquires/ > > > > > > releases lock once to fetch the list and hence helps to reduce locking > > > > > > contetiong when processing higher no. of WRs > > > > > > > > > > > > Signed-off-by: Saravanan Vajravel < saravanan.vajravel@broadcom.com> > > > > > > --- > > > > > > drivers/infiniband/core/mad.c | 14 ++++++++------ > > > > > > 1 file changed, 8 insertions(+), 6 deletions(-) > > > > > > > > > > > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c > > > > > > index 674344eb8e2f..58befbaaf0ad 100644 > > > > > > --- a/drivers/infiniband/core/mad.c > > > > > > +++ b/drivers/infiniband/core/mad.c > > > > > > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) > > > > > > > > > > > > static void timeout_sends(struct work_struct *work) > > > > > > { > > > > > > + struct ib_mad_send_wr_private *mad_send_wr, *n; > > > > > > struct ib_mad_agent_private *mad_agent_priv; > > > > > > - struct ib_mad_send_wr_private *mad_send_wr; > > > > > > struct ib_mad_send_wc mad_send_wc; > > > > > > + struct list_head local_list; > > > > > > unsigned long flags, delay; > > > > > > > > > > > > mad_agent_priv = container_of(work, struct ib_mad_agent_private, > > > > > > timed_work.work); > > > > > > mad_send_wc.vendor_err = 0; > > > > > > + INIT_LIST_HEAD(&local_list); > > > > > > > > > > > > spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > > > > while (!list_empty(&mad_agent_priv->wait_list)) { > > > > > > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) > > > > > > break; > > > > > > } > > > > > > > > > > > > - list_del(&mad_send_wr->agent_list); > > > > > > + list_del_init(&mad_send_wr->agent_list); > > > > > > if (mad_send_wr->status == IB_WC_SUCCESS && > > > > > > !retry_send(mad_send_wr)) > > > > > > continue; > > > > > > > > > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > > + list_add_tail(&mad_send_wr->agent_list, &local_list); > > > > > > + } > > > > > > + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > > > > > > > > + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { > > > > > > if (mad_send_wr->status == IB_WC_SUCCESS) > > > > > > mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > > > > > > else > > > > > > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) > > > > > > mad_send_wc.send_buf = &mad_send_wr->send_buf; > > > > > > mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, > > > > > > &mad_send_wc); > > > > > > > > > > I understand the problem, but I'm not sure that this is safe to do. How > > > > > can we be sure that this is safe to call the send_handler on entry in > > > > > wait_list without the locking? > > > > > > > > > > Thanks > > > > > > > > Per existing implementation, the send_handler is invoked without locking. > > > > I didn't change that design. Existing implementation was acquiring and > > > > releasing the lock for every Work Request (WR) that it handles inside > > > > timeout_handler. > > > > I improved it in such a way that once lock is acquired to fetch list > > > > of WR to be handled > > > > and process all fetched WRs outside the scope of lock. In both implementations, > > > > send_handler is always called without lock. This patch aims to reduce locking > > > > contention > > > > > > Indeed send_handler is called without lock, but not on the wait_list. > > > I've asked explicitly about operations on the wait_list. > > > > > > Thanks > > > > The entry in wait_list is removed under the scope of mad_agent_priv->lock. > > It is then added to local_list declared inside timeout_sends() handler > > under the same > > lock. Once all such entries are removed and added to the local_list, > > this timeout_handler > > processes each entry from the local_list and it invokes send_handler. > > This local_list > > doesn't need any locking as there is no possibility of race condition. > > > > Existing implementation also removed entry from wait_list and invoked > > send_handler. > > Yes, and what is the difference between the current and proposed code? > 2649 spin_unlock_irqrestore(&mad_agent_priv->lock, flags); <----- unlock > 2650 > 2651 if (mad_send_wr->status == IB_WC_SUCCESS) > 2652 mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; > 2653 else > 2654 mad_send_wc.status = mad_send_wr->status; > 2655 mad_send_wc.send_buf = &mad_send_wr->send_buf; > 2656 mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, <-- performed without lock > 2657 &mad_send_wc); > 2658 > 2659 deref_mad_agent(mad_agent_priv); > 2660 spin_lock_irqsave(&mad_agent_priv->lock, flags); <--- ## lock() again > 2661 } > 2662 spin_unlock_irqrestore(&mad_agent_priv->lock, flags); <-- ## unlock. > > > So operation on the wait_list should not change in the new patch > > > > Thanks, Please check your snippet of code above. I corrected it with ## tag. It is acquiring / releasing / again acquiring / again releasing the lock for handling one WR. So the existing implementation has heavy locking contention. It will be surfacing as an issue only when higher no. of WRs are handled inside time_out handler(). Thanks. > > > > > > > > > > > > > - > > > > > > deref_mad_agent(mad_agent_priv); > > > > > > - spin_lock_irqsave(&mad_agent_priv->lock, flags); > > > > > > } > > > > > > - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); > > > > > > } > > > > > > > > > > > > /* > > > > > > -- > > > > > > 2.39.3 > > > > > > > > > > > > > >
On Mon, 22 Jul 2024 16:33:25 +0530, Saravanan Vajravel wrote: > Current timeout handler of mad agent aquires/releases mad_agent_priv > lock for every timed out WRs. This causes heavy locking contention > when higher no. of WRs are to be handled inside timeout handler. > > This leads to softlockup with below trace in some use cases where > rdma-cm path is used to establish connection between peer nodes > > [...] Applied, thanks! [1/1] rdma-core/mad: Improve handling of timed out WRs of mad agent https://git.kernel.org/rdma/rdma/c/b1a72f369973f5 Best regards,
diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c index 674344eb8e2f..58befbaaf0ad 100644 --- a/drivers/infiniband/core/mad.c +++ b/drivers/infiniband/core/mad.c @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr) static void timeout_sends(struct work_struct *work) { + struct ib_mad_send_wr_private *mad_send_wr, *n; struct ib_mad_agent_private *mad_agent_priv; - struct ib_mad_send_wr_private *mad_send_wr; struct ib_mad_send_wc mad_send_wc; + struct list_head local_list; unsigned long flags, delay; mad_agent_priv = container_of(work, struct ib_mad_agent_private, timed_work.work); mad_send_wc.vendor_err = 0; + INIT_LIST_HEAD(&local_list); spin_lock_irqsave(&mad_agent_priv->lock, flags); while (!list_empty(&mad_agent_priv->wait_list)) { @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work) break; } - list_del(&mad_send_wr->agent_list); + list_del_init(&mad_send_wr->agent_list); if (mad_send_wr->status == IB_WC_SUCCESS && !retry_send(mad_send_wr)) continue; - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); + list_add_tail(&mad_send_wr->agent_list, &local_list); + } + spin_unlock_irqrestore(&mad_agent_priv->lock, flags); + list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) { if (mad_send_wr->status == IB_WC_SUCCESS) mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR; else @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work) mad_send_wc.send_buf = &mad_send_wr->send_buf; mad_agent_priv->agent.send_handler(&mad_agent_priv->agent, &mad_send_wc); - deref_mad_agent(mad_agent_priv); - spin_lock_irqsave(&mad_agent_priv->lock, flags); } - spin_unlock_irqrestore(&mad_agent_priv->lock, flags); } /*
Current timeout handler of mad agent aquires/releases mad_agent_priv lock for every timed out WRs. This causes heavy locking contention when higher no. of WRs are to be handled inside timeout handler. This leads to softlockup with below trace in some use cases where rdma-cm path is used to establish connection between peer nodes Trace: ----- BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767] CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE ------- --- 5.14.0-427.13.1.el9_4.x86_64 #1 Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019 Workqueue: ib_mad1 timeout_sends [ib_core] RIP: 0010:__do_softirq+0x78/0x2ac RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246 RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000 R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000 R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040 FS: 0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0 DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 PKRU: 55555554 Call Trace: <IRQ> ? show_trace_log_lvl+0x1c4/0x2df ? show_trace_log_lvl+0x1c4/0x2df ? __irq_exit_rcu+0xa1/0xc0 ? watchdog_timer_fn+0x1b2/0x210 ? __pfx_watchdog_timer_fn+0x10/0x10 ? __hrtimer_run_queues+0x127/0x2c0 ? hrtimer_interrupt+0xfc/0x210 ? __sysvec_apic_timer_interrupt+0x5c/0x110 ? sysvec_apic_timer_interrupt+0x37/0x90 ? asm_sysvec_apic_timer_interrupt+0x16/0x20 ? __do_softirq+0x78/0x2ac ? __do_softirq+0x60/0x2ac __irq_exit_rcu+0xa1/0xc0 sysvec_call_function_single+0x72/0x90 </IRQ> <TASK> asm_sysvec_call_function_single+0x16/0x20 RIP: 0010:_raw_spin_unlock_irq+0x14/0x30 RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247 RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800 RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000 R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538 R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c cm_process_send_error+0x122/0x1d0 [ib_cm] timeout_sends+0x1dd/0x270 [ib_core] process_one_work+0x1e2/0x3b0 ? __pfx_worker_thread+0x10/0x10 worker_thread+0x50/0x3a0 ? __pfx_worker_thread+0x10/0x10 kthread+0xdd/0x100 ? __pfx_kthread+0x10/0x10 ret_from_fork+0x29/0x50 </TASK> Simplified timeout handler by creating local list of timed out WRs and invoke send handler post creating the list. The new method acquires/ releases lock once to fetch the list and hence helps to reduce locking contetiong when processing higher no. of WRs Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com> --- drivers/infiniband/core/mad.c | 14 ++++++++------ 1 file changed, 8 insertions(+), 6 deletions(-)