Message ID | 20230829081142.3619-9-urezki@gmail.com (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | Mitigate a vmap lock contention v2 | expand |
On 08/29/23 at 10:11am, Uladzislau Rezki (Sony) wrote: > Allocated areas are spread among nodes, it implies that > the scanning has to be performed individually of each node > in order to dump all existing VAs. > > Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com> > --- > mm/vmalloc.c | 120 ++++++++++++++++++++------------------------------- > 1 file changed, 47 insertions(+), 73 deletions(-) LGTM, Reviewed-by: Baoquan He <bhe@redhat.com> > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index 968144c16237..9cce012aecdb 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -4636,30 +4636,6 @@ bool vmalloc_dump_obj(void *object) > #endif > > #ifdef CONFIG_PROC_FS > -static void *s_start(struct seq_file *m, loff_t *pos) > -{ > - struct vmap_node *vn = addr_to_node(0); > - > - mutex_lock(&vmap_purge_lock); > - spin_lock(&vn->busy.lock); > - > - return seq_list_start(&vn->busy.head, *pos); > -} > - > -static void *s_next(struct seq_file *m, void *p, loff_t *pos) > -{ > - struct vmap_node *vn = addr_to_node(0); > - return seq_list_next(p, &vn->busy.head, pos); > -} > - > -static void s_stop(struct seq_file *m, void *p) > -{ > - struct vmap_node *vn = addr_to_node(0); > - > - spin_unlock(&vn->busy.lock); > - mutex_unlock(&vmap_purge_lock); > -} > - > static void show_numa_info(struct seq_file *m, struct vm_struct *v) > { > if (IS_ENABLED(CONFIG_NUMA)) { > @@ -4703,84 +4679,82 @@ static void show_purge_info(struct seq_file *m) > } > } > > -static int s_show(struct seq_file *m, void *p) > +static int vmalloc_info_show(struct seq_file *m, void *p) > { > struct vmap_node *vn; > struct vmap_area *va; > struct vm_struct *v; > + int i; > > - vn = addr_to_node(0); > - va = list_entry(p, struct vmap_area, list); > + for (i = 0; i < nr_nodes; i++) { > + vn = &nodes[i]; > > - if (!va->vm) { > - if (va->flags & VMAP_RAM) > - seq_printf(m, "0x%pK-0x%pK %7ld vm_map_ram\n", > - (void *)va->va_start, (void *)va->va_end, > - va->va_end - va->va_start); > + spin_lock(&vn->busy.lock); > + list_for_each_entry(va, &vn->busy.head, list) { > + if (!va->vm) { > + if (va->flags & VMAP_RAM) > + seq_printf(m, "0x%pK-0x%pK %7ld vm_map_ram\n", > + (void *)va->va_start, (void *)va->va_end, > + va->va_end - va->va_start); > > - goto final; > - } > + continue; > + } > > - v = va->vm; > + v = va->vm; > > - seq_printf(m, "0x%pK-0x%pK %7ld", > - v->addr, v->addr + v->size, v->size); > + seq_printf(m, "0x%pK-0x%pK %7ld", > + v->addr, v->addr + v->size, v->size); > > - if (v->caller) > - seq_printf(m, " %pS", v->caller); > + if (v->caller) > + seq_printf(m, " %pS", v->caller); > > - if (v->nr_pages) > - seq_printf(m, " pages=%d", v->nr_pages); > + if (v->nr_pages) > + seq_printf(m, " pages=%d", v->nr_pages); > > - if (v->phys_addr) > - seq_printf(m, " phys=%pa", &v->phys_addr); > + if (v->phys_addr) > + seq_printf(m, " phys=%pa", &v->phys_addr); > > - if (v->flags & VM_IOREMAP) > - seq_puts(m, " ioremap"); > + if (v->flags & VM_IOREMAP) > + seq_puts(m, " ioremap"); > > - if (v->flags & VM_ALLOC) > - seq_puts(m, " vmalloc"); > + if (v->flags & VM_ALLOC) > + seq_puts(m, " vmalloc"); > > - if (v->flags & VM_MAP) > - seq_puts(m, " vmap"); > + if (v->flags & VM_MAP) > + seq_puts(m, " vmap"); > > - if (v->flags & VM_USERMAP) > - seq_puts(m, " user"); > + if (v->flags & VM_USERMAP) > + seq_puts(m, " user"); > > - if (v->flags & VM_DMA_COHERENT) > - seq_puts(m, " dma-coherent"); > + if (v->flags & VM_DMA_COHERENT) > + seq_puts(m, " dma-coherent"); > > - if (is_vmalloc_addr(v->pages)) > - seq_puts(m, " vpages"); > + if (is_vmalloc_addr(v->pages)) > + seq_puts(m, " vpages"); > > - show_numa_info(m, v); > - seq_putc(m, '\n'); > + show_numa_info(m, v); > + seq_putc(m, '\n'); > + } > + spin_unlock(&vn->busy.lock); > + } > > /* > * As a final step, dump "unpurged" areas. > */ > -final: > - if (list_is_last(&va->list, &vn->busy.head)) > - show_purge_info(m); > - > + show_purge_info(m); > return 0; > } > > -static const struct seq_operations vmalloc_op = { > - .start = s_start, > - .next = s_next, > - .stop = s_stop, > - .show = s_show, > -}; > - > static int __init proc_vmalloc_init(void) > { > + void *priv_data = NULL; > + > if (IS_ENABLED(CONFIG_NUMA)) > - proc_create_seq_private("vmallocinfo", 0400, NULL, > - &vmalloc_op, > - nr_node_ids * sizeof(unsigned int), NULL); > - else > - proc_create_seq("vmallocinfo", 0400, NULL, &vmalloc_op); > + priv_data = kmalloc(nr_node_ids * sizeof(unsigned int), GFP_KERNEL); > + > + proc_create_single_data("vmallocinfo", > + 0400, NULL, vmalloc_info_show, priv_data); > + > return 0; > } > module_init(proc_vmalloc_init); > -- > 2.30.2 >
On Fri, Sep 15, 2023 at 09:02:37PM +0800, Baoquan He wrote: > On 08/29/23 at 10:11am, Uladzislau Rezki (Sony) wrote: > > Allocated areas are spread among nodes, it implies that > > the scanning has to be performed individually of each node > > in order to dump all existing VAs. > > > > Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com> > > --- > > mm/vmalloc.c | 120 ++++++++++++++++++++------------------------------- > > 1 file changed, 47 insertions(+), 73 deletions(-) > > LGTM, > > Reviewed-by: Baoquan He <bhe@redhat.com> > Thank you for review, applied for v3. -- Uladzislau Rezki
diff --git a/mm/vmalloc.c b/mm/vmalloc.c index 968144c16237..9cce012aecdb 100644 --- a/mm/vmalloc.c +++ b/mm/vmalloc.c @@ -4636,30 +4636,6 @@ bool vmalloc_dump_obj(void *object) #endif #ifdef CONFIG_PROC_FS -static void *s_start(struct seq_file *m, loff_t *pos) -{ - struct vmap_node *vn = addr_to_node(0); - - mutex_lock(&vmap_purge_lock); - spin_lock(&vn->busy.lock); - - return seq_list_start(&vn->busy.head, *pos); -} - -static void *s_next(struct seq_file *m, void *p, loff_t *pos) -{ - struct vmap_node *vn = addr_to_node(0); - return seq_list_next(p, &vn->busy.head, pos); -} - -static void s_stop(struct seq_file *m, void *p) -{ - struct vmap_node *vn = addr_to_node(0); - - spin_unlock(&vn->busy.lock); - mutex_unlock(&vmap_purge_lock); -} - static void show_numa_info(struct seq_file *m, struct vm_struct *v) { if (IS_ENABLED(CONFIG_NUMA)) { @@ -4703,84 +4679,82 @@ static void show_purge_info(struct seq_file *m) } } -static int s_show(struct seq_file *m, void *p) +static int vmalloc_info_show(struct seq_file *m, void *p) { struct vmap_node *vn; struct vmap_area *va; struct vm_struct *v; + int i; - vn = addr_to_node(0); - va = list_entry(p, struct vmap_area, list); + for (i = 0; i < nr_nodes; i++) { + vn = &nodes[i]; - if (!va->vm) { - if (va->flags & VMAP_RAM) - seq_printf(m, "0x%pK-0x%pK %7ld vm_map_ram\n", - (void *)va->va_start, (void *)va->va_end, - va->va_end - va->va_start); + spin_lock(&vn->busy.lock); + list_for_each_entry(va, &vn->busy.head, list) { + if (!va->vm) { + if (va->flags & VMAP_RAM) + seq_printf(m, "0x%pK-0x%pK %7ld vm_map_ram\n", + (void *)va->va_start, (void *)va->va_end, + va->va_end - va->va_start); - goto final; - } + continue; + } - v = va->vm; + v = va->vm; - seq_printf(m, "0x%pK-0x%pK %7ld", - v->addr, v->addr + v->size, v->size); + seq_printf(m, "0x%pK-0x%pK %7ld", + v->addr, v->addr + v->size, v->size); - if (v->caller) - seq_printf(m, " %pS", v->caller); + if (v->caller) + seq_printf(m, " %pS", v->caller); - if (v->nr_pages) - seq_printf(m, " pages=%d", v->nr_pages); + if (v->nr_pages) + seq_printf(m, " pages=%d", v->nr_pages); - if (v->phys_addr) - seq_printf(m, " phys=%pa", &v->phys_addr); + if (v->phys_addr) + seq_printf(m, " phys=%pa", &v->phys_addr); - if (v->flags & VM_IOREMAP) - seq_puts(m, " ioremap"); + if (v->flags & VM_IOREMAP) + seq_puts(m, " ioremap"); - if (v->flags & VM_ALLOC) - seq_puts(m, " vmalloc"); + if (v->flags & VM_ALLOC) + seq_puts(m, " vmalloc"); - if (v->flags & VM_MAP) - seq_puts(m, " vmap"); + if (v->flags & VM_MAP) + seq_puts(m, " vmap"); - if (v->flags & VM_USERMAP) - seq_puts(m, " user"); + if (v->flags & VM_USERMAP) + seq_puts(m, " user"); - if (v->flags & VM_DMA_COHERENT) - seq_puts(m, " dma-coherent"); + if (v->flags & VM_DMA_COHERENT) + seq_puts(m, " dma-coherent"); - if (is_vmalloc_addr(v->pages)) - seq_puts(m, " vpages"); + if (is_vmalloc_addr(v->pages)) + seq_puts(m, " vpages"); - show_numa_info(m, v); - seq_putc(m, '\n'); + show_numa_info(m, v); + seq_putc(m, '\n'); + } + spin_unlock(&vn->busy.lock); + } /* * As a final step, dump "unpurged" areas. */ -final: - if (list_is_last(&va->list, &vn->busy.head)) - show_purge_info(m); - + show_purge_info(m); return 0; } -static const struct seq_operations vmalloc_op = { - .start = s_start, - .next = s_next, - .stop = s_stop, - .show = s_show, -}; - static int __init proc_vmalloc_init(void) { + void *priv_data = NULL; + if (IS_ENABLED(CONFIG_NUMA)) - proc_create_seq_private("vmallocinfo", 0400, NULL, - &vmalloc_op, - nr_node_ids * sizeof(unsigned int), NULL); - else - proc_create_seq("vmallocinfo", 0400, NULL, &vmalloc_op); + priv_data = kmalloc(nr_node_ids * sizeof(unsigned int), GFP_KERNEL); + + proc_create_single_data("vmallocinfo", + 0400, NULL, vmalloc_info_show, priv_data); + return 0; } module_init(proc_vmalloc_init);
Allocated areas are spread among nodes, it implies that the scanning has to be performed individually of each node in order to dump all existing VAs. Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com> --- mm/vmalloc.c | 120 ++++++++++++++++++++------------------------------- 1 file changed, 47 insertions(+), 73 deletions(-)