Message ID | 20240202215332.118728-5-david@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | libvhost-user: support more memslots and cleanup memslot handling code | expand |
On Fri, Feb 2, 2024 at 4:54 PM David Hildenbrand <david@redhat.com> wrote: > > Let's factor it out. Note that the check for MAP_FAILED was wrong as > we never set mmap_addr if mmap() failed. We'll remove the NULL check > separately. > > Signed-off-by: David Hildenbrand <david@redhat.com> Reviewed-by: Raphael Norwitz <raphael@enfabrica.net> > --- > subprojects/libvhost-user/libvhost-user.c | 34 ++++++++++++----------- > 1 file changed, 18 insertions(+), 16 deletions(-) > > diff --git a/subprojects/libvhost-user/libvhost-user.c b/subprojects/libvhost-user/libvhost-user.c > index 8a5a7a2295..d5b3468e43 100644 > --- a/subprojects/libvhost-user/libvhost-user.c > +++ b/subprojects/libvhost-user/libvhost-user.c > @@ -240,6 +240,22 @@ qva_to_va(VuDev *dev, uint64_t qemu_addr) > return NULL; > } > > +static void > +vu_remove_all_mem_regs(VuDev *dev) > +{ > + unsigned int i; > + > + for (i = 0; i < dev->nregions; i++) { > + VuDevRegion *r = &dev->regions[i]; > + void *ma = (void *)(uintptr_t)r->mmap_addr; > + > + if (ma) { > + munmap(ma, r->size + r->mmap_offset); > + } > + } > + dev->nregions = 0; > +} > + > static void > vmsg_close_fds(VhostUserMsg *vmsg) > { > @@ -1003,14 +1019,7 @@ vu_set_mem_table_exec(VuDev *dev, VhostUserMsg *vmsg) > unsigned int i; > VhostUserMemory m = vmsg->payload.memory, *memory = &m; > > - for (i = 0; i < dev->nregions; i++) { > - VuDevRegion *r = &dev->regions[i]; > - void *ma = (void *) (uintptr_t) r->mmap_addr; > - > - if (ma) { > - munmap(ma, r->size + r->mmap_offset); > - } > - } > + vu_remove_all_mem_regs(dev); > dev->nregions = memory->nregions; > > if (dev->postcopy_listening) { > @@ -2112,14 +2121,7 @@ vu_deinit(VuDev *dev) > { > unsigned int i; > > - for (i = 0; i < dev->nregions; i++) { > - VuDevRegion *r = &dev->regions[i]; > - void *m = (void *) (uintptr_t) r->mmap_addr; > - if (m != MAP_FAILED) { > - munmap(m, r->size + r->mmap_offset); > - } > - } > - dev->nregions = 0; > + vu_remove_all_mem_regs(dev); > > for (i = 0; i < dev->max_queues; i++) { > VuVirtq *vq = &dev->vq[i]; > -- > 2.43.0 > >
diff --git a/subprojects/libvhost-user/libvhost-user.c b/subprojects/libvhost-user/libvhost-user.c index 8a5a7a2295..d5b3468e43 100644 --- a/subprojects/libvhost-user/libvhost-user.c +++ b/subprojects/libvhost-user/libvhost-user.c @@ -240,6 +240,22 @@ qva_to_va(VuDev *dev, uint64_t qemu_addr) return NULL; } +static void +vu_remove_all_mem_regs(VuDev *dev) +{ + unsigned int i; + + for (i = 0; i < dev->nregions; i++) { + VuDevRegion *r = &dev->regions[i]; + void *ma = (void *)(uintptr_t)r->mmap_addr; + + if (ma) { + munmap(ma, r->size + r->mmap_offset); + } + } + dev->nregions = 0; +} + static void vmsg_close_fds(VhostUserMsg *vmsg) { @@ -1003,14 +1019,7 @@ vu_set_mem_table_exec(VuDev *dev, VhostUserMsg *vmsg) unsigned int i; VhostUserMemory m = vmsg->payload.memory, *memory = &m; - for (i = 0; i < dev->nregions; i++) { - VuDevRegion *r = &dev->regions[i]; - void *ma = (void *) (uintptr_t) r->mmap_addr; - - if (ma) { - munmap(ma, r->size + r->mmap_offset); - } - } + vu_remove_all_mem_regs(dev); dev->nregions = memory->nregions; if (dev->postcopy_listening) { @@ -2112,14 +2121,7 @@ vu_deinit(VuDev *dev) { unsigned int i; - for (i = 0; i < dev->nregions; i++) { - VuDevRegion *r = &dev->regions[i]; - void *m = (void *) (uintptr_t) r->mmap_addr; - if (m != MAP_FAILED) { - munmap(m, r->size + r->mmap_offset); - } - } - dev->nregions = 0; + vu_remove_all_mem_regs(dev); for (i = 0; i < dev->max_queues; i++) { VuVirtq *vq = &dev->vq[i];
Let's factor it out. Note that the check for MAP_FAILED was wrong as we never set mmap_addr if mmap() failed. We'll remove the NULL check separately. Signed-off-by: David Hildenbrand <david@redhat.com> --- subprojects/libvhost-user/libvhost-user.c | 34 ++++++++++++----------- 1 file changed, 18 insertions(+), 16 deletions(-)