diff mbox series

[v4,1/2] block: disable iopoll for split bio

Message ID 20201117075625.46118-2-jefflexu@linux.alibaba.com (mailing list archive)
State New, archived
Headers show
Series block, iomap: disable iopoll for split bio | expand

Commit Message

Jingbo Xu Nov. 17, 2020, 7:56 a.m. UTC
iopoll is initially for small size, latency sensitive IO. It doesn't
work well for big IO, especially when it needs to be split to multiple
bios. In this case, the returned cookie of __submit_bio_noacct_mq() is
indeed the cookie of the last split bio. The completion of *this* last
split bio done by iopoll doesn't mean the whole original bio has
completed. Callers of iopoll still need to wait for completion of other
split bios.

Besides bio splitting may cause more trouble for iopoll which isn't
supposed to be used in case of big IO.

iopoll for split bio may cause potential race if CPU migration happens
during bio submission. Since the returned cookie is that of the last
split bio, polling on the corresponding hardware queue doesn't help
complete other split bios, if these split bios are enqueued into
different hardware queues. Since interrupts are disabled for polling
queues, the completion of these other split bios depends on timeout
mechanism, thus causing a potential hang.

iopoll for split bio may also cause hang for sync polling. Currently
both the blkdev and iomap-based fs (ext4/xfs, etc) support sync polling
in direct IO routine. These routines will submit bio without REQ_NOWAIT
flag set, and then start sync polling in current process context. The
process may hang in blk_mq_get_tag() if the submitted bio has to be
split into multiple bios and can rapidly exhaust the queue depth. The
process are waiting for the completion of the previously allocated
requests, which should be reaped by the following polling, and thus
causing a deadlock.

To avoid these subtle trouble described above, just disable iopoll for
split bio.

Suggested-by: Ming Lei <ming.lei@redhat.com>
Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com>
---
 block/blk-merge.c | 7 +++++++
 block/blk-mq.c    | 6 ++++--
 2 files changed, 11 insertions(+), 2 deletions(-)

Comments

Jingbo Xu Nov. 19, 2020, 3:06 a.m. UTC | #1
Hi Jens,

Would you mind giving a glance at this, at least patch 1?


Patch 1 could fix the potential deadlock may be triggered by sync io polling

in direct IO routine.


Actually this patch could alleviate another potential hang in io_uring. The

returned cookie of submit_bio() is actually the cookie of the last split bio

if splitting happened. If the returned cookie is BLK_QC_T_NONE ( the

last split bio get merged into another request, or queue depth exhausted

and REQ_NOWAIT set by io_uring), then iocb->ki_cookie is BLK_QC_T_NONE.


If there's only this one IO (though split to several bios) in the 
polling hw queue,

then io_do_iopoll() will get stuck in indefinite calling of blk_poll(), 
while blk_poll()

will return 0 immediately because of BLK_QC_T_NONE, without reaping any

former completed requests.


This hang rarly happened, maybe because as long as there's other IO, polling

of these other IOs will help reap completed requests of the stuck IO 
described

above.


Of course we could refactor the problematic logic of returning cookie in 
submit_bio(),

But this patch 1 could also fix this issue. Besides this patch 1 is also 
needed to fix

the potential deadlock triggered by sync io polling described above.


Thanks,

Jeffle



On 11/17/20 3:56 PM, Jeffle Xu wrote:
> iopoll is initially for small size, latency sensitive IO. It doesn't
> work well for big IO, especially when it needs to be split to multiple
> bios. In this case, the returned cookie of __submit_bio_noacct_mq() is
> indeed the cookie of the last split bio. The completion of *this* last
> split bio done by iopoll doesn't mean the whole original bio has
> completed. Callers of iopoll still need to wait for completion of other
> split bios.
>
> Besides bio splitting may cause more trouble for iopoll which isn't
> supposed to be used in case of big IO.
>
> iopoll for split bio may cause potential race if CPU migration happens
> during bio submission. Since the returned cookie is that of the last
> split bio, polling on the corresponding hardware queue doesn't help
> complete other split bios, if these split bios are enqueued into
> different hardware queues. Since interrupts are disabled for polling
> queues, the completion of these other split bios depends on timeout
> mechanism, thus causing a potential hang.
>
> iopoll for split bio may also cause hang for sync polling. Currently
> both the blkdev and iomap-based fs (ext4/xfs, etc) support sync polling
> in direct IO routine. These routines will submit bio without REQ_NOWAIT
> flag set, and then start sync polling in current process context. The
> process may hang in blk_mq_get_tag() if the submitted bio has to be
> split into multiple bios and can rapidly exhaust the queue depth. The
> process are waiting for the completion of the previously allocated
> requests, which should be reaped by the following polling, and thus
> causing a deadlock.
>
> To avoid these subtle trouble described above, just disable iopoll for
> split bio.
>
> Suggested-by: Ming Lei <ming.lei@redhat.com>
> Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com>
> ---
>   block/blk-merge.c | 7 +++++++
>   block/blk-mq.c    | 6 ++++--
>   2 files changed, 11 insertions(+), 2 deletions(-)
>
> diff --git a/block/blk-merge.c b/block/blk-merge.c
> index bcf5e4580603..53ad781917a2 100644
> --- a/block/blk-merge.c
> +++ b/block/blk-merge.c
> @@ -279,6 +279,13 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>   	return NULL;
>   split:
>   	*segs = nsegs;
> +
> +	/*
> +	 * bio splitting may cause subtle trouble such as hang when doing iopoll,
> +	 * not to mention iopoll isn't supposed to be used in case of big IO.
> +	 */
> +	bio->bi_opf &= ~REQ_HIPRI;
> +
>   	return bio_split(bio, sectors, GFP_NOIO, bs);
>   }
>   
> diff --git a/block/blk-mq.c b/block/blk-mq.c
> index 55bcee5dc032..6d10652a7ed0 100644
> --- a/block/blk-mq.c
> +++ b/block/blk-mq.c
> @@ -3853,11 +3853,13 @@ int blk_poll(struct request_queue *q, blk_qc_t cookie, bool spin)
>   	    !test_bit(QUEUE_FLAG_POLL, &q->queue_flags))
>   		return 0;
>   
> +	hctx = q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)];
> +	if (hctx->type != HCTX_TYPE_POLL)
> +		return 0;
> +
>   	if (current->plug)
>   		blk_flush_plug_list(current->plug, false);
>   
> -	hctx = q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)];
> -
>   	/*
>   	 * If we sleep, have the caller restart the poll loop to reset
>   	 * the state. Like for the other success return cases, the
Christoph Hellwig Nov. 19, 2020, 5:52 p.m. UTC | #2
On Tue, Nov 17, 2020 at 03:56:24PM +0800, Jeffle Xu wrote:
> iopoll is initially for small size, latency sensitive IO. It doesn't
> work well for big IO, especially when it needs to be split to multiple
> bios. In this case, the returned cookie of __submit_bio_noacct_mq() is
> indeed the cookie of the last split bio. The completion of *this* last
> split bio done by iopoll doesn't mean the whole original bio has
> completed. Callers of iopoll still need to wait for completion of other
> split bios.
> 
> Besides bio splitting may cause more trouble for iopoll which isn't
> supposed to be used in case of big IO.
> 
> iopoll for split bio may cause potential race if CPU migration happens
> during bio submission. Since the returned cookie is that of the last
> split bio, polling on the corresponding hardware queue doesn't help
> complete other split bios, if these split bios are enqueued into
> different hardware queues. Since interrupts are disabled for polling
> queues, the completion of these other split bios depends on timeout
> mechanism, thus causing a potential hang.
> 
> iopoll for split bio may also cause hang for sync polling. Currently
> both the blkdev and iomap-based fs (ext4/xfs, etc) support sync polling
> in direct IO routine. These routines will submit bio without REQ_NOWAIT
> flag set, and then start sync polling in current process context. The
> process may hang in blk_mq_get_tag() if the submitted bio has to be
> split into multiple bios and can rapidly exhaust the queue depth. The
> process are waiting for the completion of the previously allocated
> requests, which should be reaped by the following polling, and thus
> causing a deadlock.
> 
> To avoid these subtle trouble described above, just disable iopoll for
> split bio.
> 
> Suggested-by: Ming Lei <ming.lei@redhat.com>
> Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com>
> ---
>  block/blk-merge.c | 7 +++++++
>  block/blk-mq.c    | 6 ++++--
>  2 files changed, 11 insertions(+), 2 deletions(-)
> 
> diff --git a/block/blk-merge.c b/block/blk-merge.c
> index bcf5e4580603..53ad781917a2 100644
> --- a/block/blk-merge.c
> +++ b/block/blk-merge.c
> @@ -279,6 +279,13 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>  	return NULL;
>  split:
>  	*segs = nsegs;
> +
> +	/*
> +	 * bio splitting may cause subtle trouble such as hang when doing iopoll,

Please capitalize the first character of a multi-line comments.  Also
this adds an overly long line.

> +	hctx = q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)];
> +	if (hctx->type != HCTX_TYPE_POLL)
> +		return 0;

I think this is good as a sanity check, but shouldn't we be able to
avoid even hitting this patch if we ensure that BLK_QC_T_NONE is
returned after a bio is split?
Jingbo Xu Nov. 20, 2020, 9:22 a.m. UTC | #3
On 11/20/20 1:52 AM, Christoph Hellwig wrote:
> On Tue, Nov 17, 2020 at 03:56:24PM +0800, Jeffle Xu wrote:
>> iopoll is initially for small size, latency sensitive IO. It doesn't
>> work well for big IO, especially when it needs to be split to multiple
>> bios. In this case, the returned cookie of __submit_bio_noacct_mq() is
>> indeed the cookie of the last split bio. The completion of *this* last
>> split bio done by iopoll doesn't mean the whole original bio has
>> completed. Callers of iopoll still need to wait for completion of other
>> split bios.
>>
>> Besides bio splitting may cause more trouble for iopoll which isn't
>> supposed to be used in case of big IO.
>>
>> iopoll for split bio may cause potential race if CPU migration happens
>> during bio submission. Since the returned cookie is that of the last
>> split bio, polling on the corresponding hardware queue doesn't help
>> complete other split bios, if these split bios are enqueued into
>> different hardware queues. Since interrupts are disabled for polling
>> queues, the completion of these other split bios depends on timeout
>> mechanism, thus causing a potential hang.
>>
>> iopoll for split bio may also cause hang for sync polling. Currently
>> both the blkdev and iomap-based fs (ext4/xfs, etc) support sync polling
>> in direct IO routine. These routines will submit bio without REQ_NOWAIT
>> flag set, and then start sync polling in current process context. The
>> process may hang in blk_mq_get_tag() if the submitted bio has to be
>> split into multiple bios and can rapidly exhaust the queue depth. The
>> process are waiting for the completion of the previously allocated
>> requests, which should be reaped by the following polling, and thus
>> causing a deadlock.
>>
>> To avoid these subtle trouble described above, just disable iopoll for
>> split bio.
>>
>> Suggested-by: Ming Lei <ming.lei@redhat.com>
>> Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com>
>> ---
>>   block/blk-merge.c | 7 +++++++
>>   block/blk-mq.c    | 6 ++++--
>>   2 files changed, 11 insertions(+), 2 deletions(-)
>>
>> diff --git a/block/blk-merge.c b/block/blk-merge.c
>> index bcf5e4580603..53ad781917a2 100644
>> --- a/block/blk-merge.c
>> +++ b/block/blk-merge.c
>> @@ -279,6 +279,13 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>   	return NULL;
>>   split:
>>   	*segs = nsegs;
>> +
>> +	/*
>> +	 * bio splitting may cause subtle trouble such as hang when doing iopoll,
> Please capitalize the first character of a multi-line comments.  Also
> this adds an overly long line.

Regards.


>
>> +	hctx = q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)];
>> +	if (hctx->type != HCTX_TYPE_POLL)
>> +		return 0;
> I think this is good as a sanity check, but shouldn't we be able to
> avoid even hitting this patch if we ensure that BLK_QC_T_NONE is
> returned after a bio is split?

Actually I had thought about returning  BLK_QC_T_NONE for split bio, but 
got blocked.


At the beginning, I want to identify split bio by checking if @split is 
NULL in __blk_queue_split().

```

                 split = blk_bio_segment_split(q, *bio, &q->bio_split, 
nr_segs);
                 break;
         }

         if (split) {

             /* bio got split */

```

But it's not the case. Even if @split is NULL, the input @bio may be the 
*last* split bio.


Then I want to identify split bio by checking loop times in 
__submit_bio_noacct_mq().

--- a/block/blk-core.c
+++ b/block/blk-core.c
@@ -1008,12 +1008,15 @@ static blk_qc_t __submit_bio_noacct_mq(struct 
bio *bio)
  {
         struct bio_list bio_list[2] = { };
         blk_qc_t ret = BLK_QC_T_NONE;
+       int split = -1;

         current->bio_list = bio_list;

         do {
                 struct gendisk *disk = bio->bi_disk;

+               split = min(split + 1, 1)
+
                 if (unlikely(bio_queue_enter(bio) != 0))
                         continue;

@@ -1027,7 +1030,7 @@ static blk_qc_t __submit_bio_noacct_mq(struct bio 
*bio)
         } while ((bio = bio_list_pop(&bio_list[0])));

         current->bio_list = NULL;
-       return ret;
+       return split ? BLK_QC_T_NONE : ret;
  }

But the bio-based routine will call blk_mq_submit_bio() directly, bypassing

__submit_bio_noacct_mq().


It seems that we have to add one specific flag to identify split bio.


Or we could use BIO_CHAIN to identify the *last* split bio from normal 
bio, since the

last split bio is always marked with BIO_CHAIN. Then we can identify the 
last split

bio by BIO_CHAIN, and the others by checking if @split is NULL in 
__blk_queue_split().
diff mbox series

Patch

diff --git a/block/blk-merge.c b/block/blk-merge.c
index bcf5e4580603..53ad781917a2 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -279,6 +279,13 @@  static struct bio *blk_bio_segment_split(struct request_queue *q,
 	return NULL;
 split:
 	*segs = nsegs;
+
+	/*
+	 * bio splitting may cause subtle trouble such as hang when doing iopoll,
+	 * not to mention iopoll isn't supposed to be used in case of big IO.
+	 */
+	bio->bi_opf &= ~REQ_HIPRI;
+
 	return bio_split(bio, sectors, GFP_NOIO, bs);
 }
 
diff --git a/block/blk-mq.c b/block/blk-mq.c
index 55bcee5dc032..6d10652a7ed0 100644
--- a/block/blk-mq.c
+++ b/block/blk-mq.c
@@ -3853,11 +3853,13 @@  int blk_poll(struct request_queue *q, blk_qc_t cookie, bool spin)
 	    !test_bit(QUEUE_FLAG_POLL, &q->queue_flags))
 		return 0;
 
+	hctx = q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)];
+	if (hctx->type != HCTX_TYPE_POLL)
+		return 0;
+
 	if (current->plug)
 		blk_flush_plug_list(current->plug, false);
 
-	hctx = q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)];
-
 	/*
 	 * If we sleep, have the caller restart the poll loop to reset
 	 * the state. Like for the other success return cases, the