@@ -131,7 +131,8 @@ static inline bool preempt_tag(struct blk_mq_alloc_data *data,
struct sbitmap_queue *bt)
{
return data->preempt ||
- atomic_read(&bt->ws_active) <= SBQ_WAIT_QUEUES;
+ atomic_read(&bt->ws_active) <= SBQ_WAIT_QUEUES ||
+ bt->force_tag_preemption;
}
unsigned int blk_mq_get_tag(struct blk_mq_alloc_data *data)
@@ -143,6 +143,8 @@ struct sbitmap_queue {
* sbitmap_queue_get_shallow()
*/
unsigned int min_shallow_depth;
+
+ bool force_tag_preemption;
};
/**
@@ -434,6 +434,7 @@ int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth,
sbq->wake_batch = sbq_calc_wake_batch(sbq, depth);
atomic_set(&sbq->wake_index, 0);
atomic_set(&sbq->ws_active, 0);
+ sbq->force_tag_preemption = true;
sbq->ws = kzalloc_node(SBQ_WAIT_QUEUES * sizeof(*sbq->ws), flags, node);
if (!sbq->ws) {
@@ -604,6 +605,34 @@ static void sbq_update_wake_index(struct sbitmap_queue *sbq,
atomic_cmpxchg(&sbq->wake_index, old_wake_index, index);
}
+static inline void sbq_update_preemption(struct sbitmap_queue *sbq,
+ unsigned int wake_batch)
+{
+ unsigned int free;
+
+ if (wake_batch == 1) {
+ /*
+ * Waiters will be woken up one by one, no risk of declining
+ * io concurrency.
+ */
+ sbq->force_tag_preemption = false;
+ return;
+ }
+
+ free = sbq->sb.depth - sbitmap_weight(&sbq->sb);
+ if (sbq->force_tag_preemption) {
+ if (free <= wake_batch)
+ sbq->force_tag_preemption = false;
+ } else {
+ if (free > wake_batch << 1)
+ sbq->force_tag_preemption = true;
+
+ }
+ sbq->force_tag_preemption =
+ (sbq->sb.depth - sbitmap_weight(&sbq->sb)) >= wake_batch << 1 ?
+ true : false;
+}
+
static bool __sbq_wake_up(struct sbitmap_queue *sbq)
{
struct sbq_wait_state *ws;
@@ -642,6 +671,7 @@ static bool __sbq_wake_up(struct sbitmap_queue *sbq)
*/
smp_mb__before_atomic();
atomic_set(&ws->wait_cnt, wake_batch);
+ sbq_update_preemption(sbq, wake_batch);
wake_up_nr(&ws->wait, wake_batch);
return false;
Now that tag preemption is disabled under heavy load, if wakers doesn't use up 'wake_batch' tags while preemption is still disabled, io concurrency will be declined. To fix the problem, add a detection before wake up, and force tag preemption is free tags are sufficient, so that the extra tags can be used by new io. And tag preemption will be disabled again if the extra tags are used up. Signed-off-by: Yu Kuai <yukuai3@huawei.com> --- block/blk-mq-tag.c | 3 ++- include/linux/sbitmap.h | 2 ++ lib/sbitmap.c | 30 ++++++++++++++++++++++++++++++ 3 files changed, 34 insertions(+), 1 deletion(-)