diff mbox series

cfq: clear queue pointers from cfqg after unpinning them in cfq_pd_offline

Message ID ba086112-65b1-c20f-0694-35d1ca0415d9@maciej.szmigiero.name (mailing list archive)
State New, archived
Headers show
Series cfq: clear queue pointers from cfqg after unpinning them in cfq_pd_offline | expand

Commit Message

Maciej S. Szmigiero Aug. 17, 2018, 5:28 p.m. UTC
The current linux-block, 4.18 and 4.17 can reliably be crashed within few
minutes by running the following bash snippet:

mkfs.ext4 -v /dev/sda3 && mount /dev/sda3 /mnt/test/ -t ext4;
while true; do
	mkdir /sys/fs/cgroup/unified/test/;
	echo $$ >/sys/fs/cgroup/unified/test/cgroup.procs;
	dd if=/dev/zero of=/mnt/test/test-$(( RANDOM * 10 / 32768 )) bs=1M count=1024 &
	echo $$ >/sys/fs/cgroup/unified/cgroup.procs;
	sleep 1;
	kill -KILL $!; wait $!;
	rmdir /sys/fs/cgroup/unified/test;
done

# cat /sys/block/sda/queue/scheduler
noop [cfq]
# cat /sys/block/sda/queue/rotational
1
# cat /sys/fs/cgroup/unified/cgroup.subtree_control
cpu io memory pids

The backtraces vary but often they are NULL pointer dereferences due to
various cfqq fields being NULL.
Or BUG_ON(cfqq->ref <= 0) in cfq_put_queue() triggers due to cfqq reference
count being zero.

Bisection points at
commit 4c6994806f70 ("blk-throttle: fix race between blkcg_bio_issue_check() and cgroup_rmdir()").
The prime suspect looked like .pd_offline_fn() method being called multiple
times, but from analyzing the mentioned commit this didn't seem possible
and runtime trials have confirmed that.

However, CFQ's cfq_pd_offline() implementation of the above method were
leaving queue pointers intact in cfqg after unpinning them.
After making sure that they are cleared to NULL in this function I can no
longer reproduce the crash.

Signed-off-by: Maciej S. Szmigiero <mail@maciej.szmigiero.name>
Fixes: 4c6994806f70 ("blk-throttle: fix race between blkcg_bio_issue_check() and cgroup_rmdir()").
Cc: stable@vger.kernel.org
---
 block/cfq-iosched.c | 12 +++++++++---
 1 file changed, 9 insertions(+), 3 deletions(-)

Comments

Paolo Valente Aug. 17, 2018, 5:30 p.m. UTC | #1
> Il giorno 17 ago 2018, alle ore 19:28, Maciej S. Szmigiero <mail@maciej.szmigiero.name> ha scritto:
> 
> The current linux-block, 4.18 and 4.17 can reliably be crashed within few
> minutes by running the following bash snippet:
> 
> mkfs.ext4 -v /dev/sda3 && mount /dev/sda3 /mnt/test/ -t ext4;
> while true; do
> 	mkdir /sys/fs/cgroup/unified/test/;
> 	echo $$ >/sys/fs/cgroup/unified/test/cgroup.procs;
> 	dd if=/dev/zero of=/mnt/test/test-$(( RANDOM * 10 / 32768 )) bs=1M count=1024 &
> 	echo $$ >/sys/fs/cgroup/unified/cgroup.procs;
> 	sleep 1;
> 	kill -KILL $!; wait $!;
> 	rmdir /sys/fs/cgroup/unified/test;
> done
> 
> # cat /sys/block/sda/queue/scheduler
> noop [cfq]
> # cat /sys/block/sda/queue/rotational
> 1
> # cat /sys/fs/cgroup/unified/cgroup.subtree_control
> cpu io memory pids
> 
> The backtraces vary but often they are NULL pointer dereferences due to
> various cfqq fields being NULL.
> Or BUG_ON(cfqq->ref <= 0) in cfq_put_queue() triggers due to cfqq reference
> count being zero.
> 
> Bisection points at
> commit 4c6994806f70 ("blk-throttle: fix race between blkcg_bio_issue_check() and cgroup_rmdir()").
> The prime suspect looked like .pd_offline_fn() method being called multiple
> times, but from analyzing the mentioned commit this didn't seem possible
> and runtime trials have confirmed that.
> 
> However, CFQ's cfq_pd_offline() implementation of the above method were
> leaving queue pointers intact in cfqg after unpinning them.
> After making sure that they are cleared to NULL in this function I can no
> longer reproduce the crash.
> 

By chance, did you check whether is BFQ is ok in this respect?

Thanks,
Paolo

> Signed-off-by: Maciej S. Szmigiero <mail@maciej.szmigiero.name>
> Fixes: 4c6994806f70 ("blk-throttle: fix race between blkcg_bio_issue_check() and cgroup_rmdir()").
> Cc: stable@vger.kernel.org
> ---
> block/cfq-iosched.c | 12 +++++++++---
> 1 file changed, 9 insertions(+), 3 deletions(-)
> 
> diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
> index 2eb87444b157..ed41aa978c4a 100644
> --- a/block/cfq-iosched.c
> +++ b/block/cfq-iosched.c
> @@ -1644,14 +1644,20 @@ static void cfq_pd_offline(struct blkg_policy_data *pd)
> 	int i;
> 
> 	for (i = 0; i < IOPRIO_BE_NR; i++) {
> -		if (cfqg->async_cfqq[0][i])
> +		if (cfqg->async_cfqq[0][i]) {
> 			cfq_put_queue(cfqg->async_cfqq[0][i]);
> -		if (cfqg->async_cfqq[1][i])
> +			cfqg->async_cfqq[0][i] = NULL;
> +		}
> +		if (cfqg->async_cfqq[1][i]) {
> 			cfq_put_queue(cfqg->async_cfqq[1][i]);
> +			cfqg->async_cfqq[1][i] = NULL;
> +		}
> 	}
> 
> -	if (cfqg->async_idle_cfqq)
> +	if (cfqg->async_idle_cfqq) {
> 		cfq_put_queue(cfqg->async_idle_cfqq);
> +		cfqg->async_idle_cfqq = NULL;
> +	}
> 
> 	/*
> 	 * @blkg is going offline and will be ignored by
Maciej S. Szmigiero Aug. 17, 2018, 5:38 p.m. UTC | #2
On 17.08.2018 19:30, Paolo Valente wrote:
> 
> 
>> Il giorno 17 ago 2018, alle ore 19:28, Maciej S. Szmigiero <mail@maciej.szmigiero.name> ha scritto:
>>
>> The current linux-block, 4.18 and 4.17 can reliably be crashed within few
>> minutes by running the following bash snippet:
>>
>> mkfs.ext4 -v /dev/sda3 && mount /dev/sda3 /mnt/test/ -t ext4;
>> while true; do
>> 	mkdir /sys/fs/cgroup/unified/test/;
>> 	echo $$ >/sys/fs/cgroup/unified/test/cgroup.procs;
>> 	dd if=/dev/zero of=/mnt/test/test-$(( RANDOM * 10 / 32768 )) bs=1M count=1024 &
>> 	echo $$ >/sys/fs/cgroup/unified/cgroup.procs;
>> 	sleep 1;
>> 	kill -KILL $!; wait $!;
>> 	rmdir /sys/fs/cgroup/unified/test;
>> done
>>
>> # cat /sys/block/sda/queue/scheduler
>> noop [cfq]
>> # cat /sys/block/sda/queue/rotational
>> 1
>> # cat /sys/fs/cgroup/unified/cgroup.subtree_control
>> cpu io memory pids
>>
>> The backtraces vary but often they are NULL pointer dereferences due to
>> various cfqq fields being NULL.
>> Or BUG_ON(cfqq->ref <= 0) in cfq_put_queue() triggers due to cfqq reference
>> count being zero.
>>
>> Bisection points at
>> commit 4c6994806f70 ("blk-throttle: fix race between blkcg_bio_issue_check() and cgroup_rmdir()").
>> The prime suspect looked like .pd_offline_fn() method being called multiple
>> times, but from analyzing the mentioned commit this didn't seem possible
>> and runtime trials have confirmed that.
>>
>> However, CFQ's cfq_pd_offline() implementation of the above method were
>> leaving queue pointers intact in cfqg after unpinning them.
>> After making sure that they are cleared to NULL in this function I can no
>> longer reproduce the crash.
>>
> 
> By chance, did you check whether is BFQ is ok in this respect?

I wasn't able to crash BFQ with the above test and in fact had run my
machines on BFQ until I was able to find a fix for this in CFQ.

Also, BFQ has a bit similar code in bfq_put_async_queues() that is
called from bfq_pd_offline() that is already NULL-ing the passed
pointer.

> Thanks,
> Paolo

Regards,
Maciej
diff mbox series

Patch

diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index 2eb87444b157..ed41aa978c4a 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -1644,14 +1644,20 @@  static void cfq_pd_offline(struct blkg_policy_data *pd)
 	int i;
 
 	for (i = 0; i < IOPRIO_BE_NR; i++) {
-		if (cfqg->async_cfqq[0][i])
+		if (cfqg->async_cfqq[0][i]) {
 			cfq_put_queue(cfqg->async_cfqq[0][i]);
-		if (cfqg->async_cfqq[1][i])
+			cfqg->async_cfqq[0][i] = NULL;
+		}
+		if (cfqg->async_cfqq[1][i]) {
 			cfq_put_queue(cfqg->async_cfqq[1][i]);
+			cfqg->async_cfqq[1][i] = NULL;
+		}
 	}
 
-	if (cfqg->async_idle_cfqq)
+	if (cfqg->async_idle_cfqq) {
 		cfq_put_queue(cfqg->async_idle_cfqq);
+		cfqg->async_idle_cfqq = NULL;
+	}
 
 	/*
 	 * @blkg is going offline and will be ignored by