@@ -449,6 +449,7 @@ static noinline void compress_file_range(struct inode *inode,
u64 num_bytes;
u64 blocksize = fs_info->sectorsize;
u64 actual_end;
+ u64 compress_range_end;
u64 isize = i_size_read(inode);
int ret = 0;
struct page **pages = NULL;
@@ -500,6 +501,8 @@ static noinline void compress_file_range(struct inode *inode,
num_bytes = max(blocksize, num_bytes);
total_in = 0;
ret = 0;
+ compress_range_end = start + min_t(u64, BTRFS_MAX_UNCOMPRESSED,
+ end - start);
/*
* we do compression for mount -o compress and when the
@@ -528,7 +531,8 @@ static noinline void compress_file_range(struct inode *inode,
* If the compression fails for any reason, we set the pages
* dirty again later on.
*/
- extent_range_clear_dirty_for_io(inode, start, end);
+ extent_range_clear_dirty_for_io(inode, start,
+ compress_range_end);
redirty = 1;
ret = btrfs_compress_pages(compress_type,
inode->i_mapping, start,
@@ -667,7 +671,7 @@ static noinline void compress_file_range(struct inode *inode,
/* unlocked later on in the async handlers */
if (redirty)
- extent_range_redirty_for_io(inode, start, end);
+ extent_range_redirty_for_io(inode, start, compress_range_end);
add_async_extent(async_cow, start, end - start + 1, 0, NULL, 0,
BTRFS_COMPRESS_NONE);
*num_added += 1;
At now while compressing data range code touch dirty page status on whole range on each 128kb iteration, that's costs time As we care only about page status in range that will be compressed in current iteration, let's touch dirty status only for actual compression range Signed-off-by: Timofey Titovets <nefelim4ag@gmail.com> --- fs/btrfs/inode.c | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) -- 2.14.1 -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html