@@ -112,8 +112,11 @@ background_gc=%s Turn on/off cleaning operations, namely garbage
collection and if background_gc=off, garbage collection
will be turned off. If background_gc=sync, it will turn
on synchronous garbage collection running in background.
- Default value for this option is on. So garbage
- collection is on by default.
+ If background_gc=adjust, gc will adjust its policy depends
+ on conditions: speed up if there no free segments, and slow
+ down if there is no free space.
+ Default value for this option is on. So garbage collection
+ is on by default.
gc_merge When background_gc is on, this option can be enabled to
let background GC thread to handle foreground GC requests,
it can eliminate the sluggish issue caused by slow foreground
@@ -1333,6 +1333,10 @@ enum {
* background gc is on, migrating blocks
* like foreground gc
*/
+ BGGC_MODE_ADJUST, /*
+ * background gc is on, and tune its speed
+ * dependso n conditions
+ */
};
enum {
@@ -28,6 +28,67 @@ static struct kmem_cache *victim_entry_slab;
static unsigned int count_bits(const unsigned long *addr,
unsigned int offset, unsigned int len);
+static inline int free_user_block_ratio(struct f2fs_sb_info *sbi)
+{
+ block_t unused_user_blocks = sbi->user_block_count -
+ written_block_count(sbi);
+ return unused_user_blocks == 0 ? 100 :
+ (100 * free_user_blocks(sbi) / unused_user_blocks);
+}
+
+static bool has_few_free_segments(struct f2fs_sb_info *sbi)
+{
+ unsigned int free_segs = free_segments(sbi);
+ unsigned int ovp_segs = overprovision_segments(sbi);
+
+ return free_segs <= DEF_FEW_FREE_SEGMENT_MULTIPLE * ovp_segs;
+}
+
+static bool has_few_free_space(struct f2fs_sb_info *sbi)
+{
+ block_t total_user_block = sbi->user_block_count;
+ block_t free_user_blocks = total_user_block - written_block_count(sbi);
+
+ return 100 * free_user_blocks / total_user_block <=
+ DEF_FEW_FREE_SPACE_RATIO;
+}
+
+static bool has_enough_reclaimable_blocks(struct f2fs_sb_info *sbi)
+{
+ return 100 - free_user_block_ratio(sbi) >=
+ DEF_ENOUGH_RECLAIMABLE_BLOCK_RATIO;
+}
+
+static void adjust_gc_perference(struct f2fs_sb_info *sbi,
+ unsigned int *wait_ms)
+{
+ struct f2fs_gc_kthread *gc_th = sbi->gc_thread;
+
+ if (has_few_free_space(sbi))
+ gc_th->gc_preference = GC_LIFETIME_MODE;
+ else if (has_few_free_segments(sbi) &&
+ has_enough_reclaimable_blocks(sbi))
+ gc_th->gc_preference = GC_PERFORMANCE_MODE;
+ else
+ gc_th->gc_preference = GC_BALANCE_MODE;
+
+ switch (gc_th->gc_preference) {
+ case GC_PERFORMANCE_MODE:
+ *wait_ms = max(DEF_GC_BALANCE_MIN_SLEEP_TIME *
+ free_user_block_ratio(sbi) / 100,
+ DEF_GC_PERFORMANCE_MIN_SLEEP_TIME);
+ break;
+ case GC_LIFETIME_MODE:
+ gc_th->min_sleep_time = DEF_GC_THREAD_MIN_SLEEP_TIME;
+ break;
+ case GC_BALANCE_MODE:
+ gc_th->min_sleep_time = DEF_GC_BALANCE_MIN_SLEEP_TIME;
+ break;
+ default:
+ f2fs_bug_on(sbi, 1);
+ }
+}
+
static int gc_thread_func(void *data)
{
struct f2fs_sb_info *sbi = data;
@@ -46,6 +107,9 @@ static int gc_thread_func(void *data)
do {
bool sync_mode, foreground = false;
+ if (F2FS_OPTION(sbi).bggc_mode == BGGC_MODE_ADJUST)
+ adjust_gc_perference(sbi, &wait_ms);
+
wait_event_interruptible_timeout(*wq,
kthread_should_stop() || freezing(current) ||
waitqueue_active(fggc_wq) ||
@@ -109,7 +173,8 @@ static int gc_thread_func(void *data)
goto next;
}
- if (!is_idle(sbi, GC_TIME)) {
+ if (!is_idle(sbi, GC_TIME) &&
+ F2FS_OPTION(sbi).bggc_mode != BGGC_MODE_ADJUST) {
increase_sleep_time(gc_th, &wait_ms);
f2fs_up_write(&sbi->gc_lock);
stat_io_skip_bggc_count(sbi);
@@ -183,6 +248,8 @@ int f2fs_start_gc_thread(struct f2fs_sb_info *sbi)
gc_th->max_sleep_time = DEF_GC_THREAD_MAX_SLEEP_TIME;
gc_th->no_gc_sleep_time = DEF_GC_THREAD_NOGC_SLEEP_TIME;
+ gc_th->gc_preference = GC_BALANCE_MODE;
+
gc_th->gc_wake = false;
sbi->gc_thread = gc_th;
@@ -329,6 +396,23 @@ static unsigned int check_bg_victims(struct f2fs_sb_info *sbi)
return NULL_SEGNO;
}
+static unsigned char get_max_age(struct f2fs_sb_info *sbi)
+{
+ struct f2fs_gc_kthread *gc_th = sbi->gc_thread;
+ unsigned char max_age = 100;
+ unsigned char ratio;
+
+ if (!gc_th || gc_th->gc_preference != GC_LIFETIME_MODE)
+ goto out;
+
+ /* if free block count is less than 10%, reduce cost weight of age */
+ ratio = free_user_block_ratio(sbi);
+ if (ratio <= DEF_FEW_FREE_SEGMENT_RATIO)
+ max_age = max(10 * ratio, 1);
+out:
+ return max_age;
+}
+
static unsigned int get_cb_cost(struct f2fs_sb_info *sbi, unsigned int segno)
{
struct sit_info *sit_i = SIT_I(sbi);
@@ -336,6 +420,7 @@ static unsigned int get_cb_cost(struct f2fs_sb_info *sbi, unsigned int segno)
unsigned int start = GET_SEG_FROM_SEC(sbi, secno);
unsigned long long mtime = 0;
unsigned int vblocks;
+ unsigned char max_age;
unsigned char age = 0;
unsigned char u;
unsigned int i;
@@ -355,8 +440,11 @@ static unsigned int get_cb_cost(struct f2fs_sb_info *sbi, unsigned int segno)
sit_i->min_mtime = mtime;
if (mtime > sit_i->max_mtime)
sit_i->max_mtime = mtime;
+
+ max_age = get_max_age(sbi);
+
if (sit_i->max_mtime != sit_i->min_mtime)
- age = 100 - div64_u64(100 * (mtime - sit_i->min_mtime),
+ age = max_age - div64_u64(max_age * (mtime - sit_i->min_mtime),
sit_i->max_mtime - sit_i->min_mtime);
return UINT_MAX - ((100 * (100 - u) * age) / (100 + u));
@@ -15,6 +15,14 @@
#define DEF_GC_THREAD_MAX_SLEEP_TIME 60000
#define DEF_GC_THREAD_NOGC_SLEEP_TIME 300000 /* wait 5 min */
+/* for BGGC_MODE_ADJUST */
+#define DEF_GC_PERFORMANCE_MIN_SLEEP_TIME 100 /* 100 ms */
+#define DEF_GC_BALANCE_MIN_SLEEP_TIME 10000 /* 10 sec */
+#define DEF_FEW_FREE_SPACE_RATIO 10 /* few free space ratio */
+#define DEF_FEW_FREE_SEGMENT_MULTIPLE 10 /* few free segments multiple */
+#define DEF_ENOUGH_RECLAIMABLE_BLOCK_RATIO 20 /* enough reclaimable block ratio */
+#define DEF_FEW_FREE_SEGMENT_RATIO 10 /* few free segment ratio */
+
/* choose candidates from sections which has age of more than 7 days */
#define DEF_GC_THREAD_AGE_THRESHOLD (60 * 60 * 24 * 7)
#define DEF_GC_THREAD_CANDIDATE_RATIO 20 /* select 20% oldest sections as candidates */
@@ -32,6 +40,19 @@
#define NR_GC_CHECKPOINT_SECS (3) /* data/node/dentry sections */
+/* GC preference */
+enum {
+ GC_PERFORMANCE_MODE, /*
+ * speed up background gc to recycle
+ * slack space for better performance
+ */
+ GC_LIFETIME_MODE, /*
+ * slow down background gc to avoid high
+ * WAF if there is less free space.
+ */
+ GC_BALANCE_MODE, /* tradeoff in between perf and lifetime */
+};
+
struct f2fs_gc_kthread {
struct task_struct *f2fs_gc_task;
wait_queue_head_t gc_wait_queue_head;
@@ -42,6 +63,8 @@ struct f2fs_gc_kthread {
unsigned int max_sleep_time;
unsigned int no_gc_sleep_time;
+ unsigned char gc_preference; /* gc perference */
+
/* for changing gc mode */
bool gc_wake;
@@ -693,6 +693,8 @@ static int parse_options(struct super_block *sb, char *options, bool is_remount)
F2FS_OPTION(sbi).bggc_mode = BGGC_MODE_OFF;
} else if (!strcmp(name, "sync")) {
F2FS_OPTION(sbi).bggc_mode = BGGC_MODE_SYNC;
+ } else if (!strcmp(name, "adjust")) {
+ F2FS_OPTION(sbi).bggc_mode = BGGC_MODE_ADJUST;
} else {
kfree(name);
return -EINVAL;
@@ -1927,6 +1929,8 @@ static int f2fs_show_options(struct seq_file *seq, struct dentry *root)
seq_printf(seq, ",background_gc=%s", "on");
else if (F2FS_OPTION(sbi).bggc_mode == BGGC_MODE_OFF)
seq_printf(seq, ",background_gc=%s", "off");
+ else if (F2FS_OPTION(sbi).bggc_mode == BGGC_MODE_ADJUST)
+ seq_printf(seq, ",background_gc=%s", "adjust");
if (test_opt(sbi, GC_MERGE))
seq_puts(seq, ",gc_merge");