@@ -83,6 +83,7 @@ struct send_ctx {
u32 send_size;
u32 send_max_size;
bool put_data;
+ struct page **send_buf_pages;
u64 flags; /* 'flags' member of btrfs_ioctl_send_args is u64 */
/* Protocol version compatibility requested */
u32 proto;
@@ -7497,6 +7498,7 @@ long btrfs_ioctl_send(struct inode *inode, struct btrfs_ioctl_send_args *arg)
struct btrfs_root *clone_root;
struct send_ctx *sctx = NULL;
u32 i;
+ u32 send_buf_num_pages = 0;
u64 *clone_sources_tmp = NULL;
int clone_sources_to_rollback = 0;
size_t alloc_size;
@@ -7588,10 +7590,28 @@ long btrfs_ioctl_send(struct inode *inode, struct btrfs_ioctl_send_args *arg)
if (sctx->proto >= 2) {
sctx->send_max_size = ALIGN(SZ_16K + BTRFS_MAX_COMPRESSED,
PAGE_SIZE);
+ send_buf_num_pages = sctx->send_max_size >> PAGE_SHIFT;
+ sctx->send_buf_pages = kcalloc(send_buf_num_pages,
+ sizeof(*sctx->send_buf_pages),
+ GFP_KERNEL);
+ if (!sctx->send_buf_pages) {
+ send_buf_num_pages = 0;
+ ret = -ENOMEM;
+ goto out;
+ }
+ for (i = 0; i < send_buf_num_pages; i++) {
+ sctx->send_buf_pages[i] = alloc_page(GFP_KERNEL);
+ if (!sctx->send_buf_pages[i]) {
+ ret = -ENOMEM;
+ goto out;
+ }
+ }
+ sctx->send_buf = vmap(sctx->send_buf_pages, send_buf_num_pages,
+ VM_MAP, PAGE_KERNEL);
} else {
sctx->send_max_size = BTRFS_SEND_BUF_SIZE_V1;
+ sctx->send_buf = kvmalloc(sctx->send_max_size, GFP_KERNEL);
}
- sctx->send_buf = kvmalloc(sctx->send_max_size, GFP_KERNEL);
if (!sctx->send_buf) {
ret = -ENOMEM;
goto out;
@@ -7784,7 +7804,16 @@ long btrfs_ioctl_send(struct inode *inode, struct btrfs_ioctl_send_args *arg)
fput(sctx->send_filp);
kvfree(sctx->clone_roots);
- kvfree(sctx->send_buf);
+ if (sctx->proto >= 2) {
+ vunmap(sctx->send_buf);
+ for (i = 0; i < send_buf_num_pages; i++) {
+ if (sctx->send_buf_pages[i])
+ __free_page(sctx->send_buf_pages[i]);
+ }
+ kfree(sctx->send_buf_pages);
+ } else {
+ kvfree(sctx->send_buf);
+ }
name_cache_free(sctx);