Message ID | 20190131130850.6850-7-yuval.shaia@oracle.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Misc fixes to pvrdma device | expand |
Yuval Shaia <yuval.shaia@oracle.com> writes: > Signed-off-by: Yuval Shaia <yuval.shaia@oracle.com> > --- > hw/rdma/vmw/pvrdma.h | 1 + > hw/rdma/vmw/pvrdma_main.c | 72 +++++++++++++++++++++++++++++++++++++++ > 2 files changed, 73 insertions(+) > > diff --git a/hw/rdma/vmw/pvrdma.h b/hw/rdma/vmw/pvrdma.h > index 167706ec2c..dc10f21ca0 100644 > --- a/hw/rdma/vmw/pvrdma.h > +++ b/hw/rdma/vmw/pvrdma.h > @@ -133,5 +133,6 @@ static inline void post_interrupt(PVRDMADev *dev, unsigned vector) > } > > int pvrdma_exec_cmd(PVRDMADev *dev); > +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func); > > #endif The only user appears in the next patch. I'd squash the two patches. Matter of taste. > diff --git a/hw/rdma/vmw/pvrdma_main.c b/hw/rdma/vmw/pvrdma_main.c > index cf82e78f08..79900076ec 100644 > --- a/hw/rdma/vmw/pvrdma_main.c > +++ b/hw/rdma/vmw/pvrdma_main.c > @@ -14,6 +14,7 @@ > */ > > #include "qemu/osdep.h" > +#include "qemu/units.h" > #include "qapi/error.h" > #include "hw/hw.h" > #include "hw/pci/pci.h" > @@ -36,6 +37,8 @@ > #include "standard-headers/drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h" > #include "pvrdma_qp_ops.h" > > +GSList *devices; > + > static Property pvrdma_dev_properties[] = { > DEFINE_PROP_STRING("netdev", PVRDMADev, backend_eth_device_name), > DEFINE_PROP_STRING("ibdev", PVRDMADev, backend_device_name), > @@ -55,6 +58,72 @@ static Property pvrdma_dev_properties[] = { > DEFINE_PROP_END_OF_LIST(), > }; > > +static void pvrdma_dump_device_statistics(gpointer data, gpointer user_data) > +{ > + CPUListState *s = user_data; > + PCIDevice *pdev = data; > + PVRDMADev *dev = PVRDMA_DEV(pdev); > + > + (*s->cpu_fprintf)(s->file, "%s_%x.%x\n", pdev->name, > + PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn)); Why the indirection through CPUListState? What's wrong with straight monitor_printf()? > + (*s->cpu_fprintf)(s->file, "\tcommands : %" PRId64 "\n", > + dev->stats.commands); > + (*s->cpu_fprintf)(s->file, "\ttx : %" PRId64 "\n", > + dev->rdma_dev_res.stats.tx); > + (*s->cpu_fprintf)(s->file, "\ttx_len : %" PRId64 "\n", > + dev->rdma_dev_res.stats.tx_len); > + (*s->cpu_fprintf)(s->file, "\ttx_err : %" PRId64 "\n", > + dev->rdma_dev_res.stats.tx_err); > + (*s->cpu_fprintf)(s->file, "\trx_bufs : %" PRId64 "\n", > + dev->rdma_dev_res.stats.rx_bufs); > + (*s->cpu_fprintf)(s->file, "\trx_bufs_len : %" PRId64 "\n", > + dev->rdma_dev_res.stats.rx_bufs_len); > + (*s->cpu_fprintf)(s->file, "\trx_bufs_err : %" PRId64 "\n", > + dev->rdma_dev_res.stats.rx_bufs_err); > + (*s->cpu_fprintf)(s->file, "\tcompletions : %" PRId64 "\n", > + dev->rdma_dev_res.stats.completions); > + (*s->cpu_fprintf)(s->file, "\tpoll_cq (bk) : %" PRId64 "\n", > + dev->rdma_dev_res.stats.poll_cq_from_bk); > + (*s->cpu_fprintf)(s->file, "\tpoll_cq_ppoll_to : %" PRId64 "\n", > + dev->rdma_dev_res.stats.poll_cq_ppoll_to); > + (*s->cpu_fprintf)(s->file, "\tpoll_cq (fe) : %" PRId64 "\n", > + dev->rdma_dev_res.stats.poll_cq_from_guest); > + (*s->cpu_fprintf)(s->file, "\tmad_tx : %" PRId64 "\n", > + dev->rdma_dev_res.stats.mad_tx); > + (*s->cpu_fprintf)(s->file, "\tmad_tx_err : %" PRId64 "\n", > + dev->rdma_dev_res.stats.mad_tx_err); > + (*s->cpu_fprintf)(s->file, "\tmad_rx : %" PRId64 "\n", > + dev->rdma_dev_res.stats.mad_rx); > + (*s->cpu_fprintf)(s->file, "\tmad_rx_err : %" PRId64 "\n", > + dev->rdma_dev_res.stats.mad_rx_err); > + (*s->cpu_fprintf)(s->file, "\tmad_rx_bufs : %" PRId64 "\n", > + dev->rdma_dev_res.stats.mad_rx_bufs); > + (*s->cpu_fprintf)(s->file, "\tmad_rx_bufs_err : %" PRId64 "\n", > + dev->rdma_dev_res.stats.mad_rx_bufs_err); > + (*s->cpu_fprintf)(s->file, "\tPDs : %" PRId32 "\n", > + dev->rdma_dev_res.pd_tbl.used); > + (*s->cpu_fprintf)(s->file, "\tMRs : %" PRId32 "\n", > + dev->rdma_dev_res.mr_tbl.used); > + (*s->cpu_fprintf)(s->file, "\tUCs : %" PRId32 "\n", > + dev->rdma_dev_res.uc_tbl.used); > + (*s->cpu_fprintf)(s->file, "\tQPs : %" PRId32 "\n", > + dev->rdma_dev_res.qp_tbl.used); > + (*s->cpu_fprintf)(s->file, "\tCQs : %" PRId32 "\n", > + dev->rdma_dev_res.cq_tbl.used); > + (*s->cpu_fprintf)(s->file, "\tCEQ_CTXs : %" PRId32 "\n", > + dev->rdma_dev_res.cqe_ctx_tbl.used); > +} > + > +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func) > +{ > + CPUListState s = { > + .file = f, > + .cpu_fprintf = fprintf_func, > + }; > + > + g_slist_foreach(devices, pvrdma_dump_device_statistics, &s); > +} > + > static void free_dev_ring(PCIDevice *pci_dev, PvrdmaRing *ring, > void *ring_state) > { > @@ -618,6 +687,8 @@ static void pvrdma_realize(PCIDevice *pdev, Error **errp) > dev->shutdown_notifier.notify = pvrdma_shutdown_notifier; > qemu_register_shutdown_notifier(&dev->shutdown_notifier); > > + devices = g_slist_append(devices, pdev); > + > out: > if (rc) { > pvrdma_fini(pdev); > @@ -627,6 +698,7 @@ out: > > static void pvrdma_exit(PCIDevice *pdev) > { > + devices = g_slist_remove(devices, pdev); > pvrdma_fini(pdev); > }
On Mon, Feb 04, 2019 at 02:03:58PM +0100, Markus Armbruster wrote: > Yuval Shaia <yuval.shaia@oracle.com> writes: > > > Signed-off-by: Yuval Shaia <yuval.shaia@oracle.com> > > --- > > hw/rdma/vmw/pvrdma.h | 1 + > > hw/rdma/vmw/pvrdma_main.c | 72 +++++++++++++++++++++++++++++++++++++++ > > 2 files changed, 73 insertions(+) > > > > diff --git a/hw/rdma/vmw/pvrdma.h b/hw/rdma/vmw/pvrdma.h > > index 167706ec2c..dc10f21ca0 100644 > > --- a/hw/rdma/vmw/pvrdma.h > > +++ b/hw/rdma/vmw/pvrdma.h > > @@ -133,5 +133,6 @@ static inline void post_interrupt(PVRDMADev *dev, unsigned vector) > > } > > > > int pvrdma_exec_cmd(PVRDMADev *dev); > > +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func); > > > > #endif > > The only user appears in the next patch. I'd squash the two patches. > Matter of taste. Agree with you. I just did to help reviewers so ones that familiar with 'monitor' can review the other patch while rdma folks can review this one. Will probably squash them as soon as the conversion on the other patch will be over. > > > diff --git a/hw/rdma/vmw/pvrdma_main.c b/hw/rdma/vmw/pvrdma_main.c > > index cf82e78f08..79900076ec 100644 > > --- a/hw/rdma/vmw/pvrdma_main.c > > +++ b/hw/rdma/vmw/pvrdma_main.c > > @@ -14,6 +14,7 @@ > > */ > > > > #include "qemu/osdep.h" > > +#include "qemu/units.h" > > #include "qapi/error.h" > > #include "hw/hw.h" > > #include "hw/pci/pci.h" > > @@ -36,6 +37,8 @@ > > #include "standard-headers/drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h" > > #include "pvrdma_qp_ops.h" > > > > +GSList *devices; > > + > > static Property pvrdma_dev_properties[] = { > > DEFINE_PROP_STRING("netdev", PVRDMADev, backend_eth_device_name), > > DEFINE_PROP_STRING("ibdev", PVRDMADev, backend_device_name), > > @@ -55,6 +58,72 @@ static Property pvrdma_dev_properties[] = { > > DEFINE_PROP_END_OF_LIST(), > > }; > > > > +static void pvrdma_dump_device_statistics(gpointer data, gpointer user_data) > > +{ > > + CPUListState *s = user_data; > > + PCIDevice *pdev = data; > > + PVRDMADev *dev = PVRDMA_DEV(pdev); > > + > > + (*s->cpu_fprintf)(s->file, "%s_%x.%x\n", pdev->name, > > + PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn)); > > Why the indirection through CPUListState? What's wrong with straight > monitor_printf()? No special reasoning, just wanted to utilize an existing mechanism and design. > > > + (*s->cpu_fprintf)(s->file, "\tcommands : %" PRId64 "\n", > > + dev->stats.commands); > > + (*s->cpu_fprintf)(s->file, "\ttx : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.tx); > > + (*s->cpu_fprintf)(s->file, "\ttx_len : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.tx_len); > > + (*s->cpu_fprintf)(s->file, "\ttx_err : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.tx_err); > > + (*s->cpu_fprintf)(s->file, "\trx_bufs : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.rx_bufs); > > + (*s->cpu_fprintf)(s->file, "\trx_bufs_len : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.rx_bufs_len); > > + (*s->cpu_fprintf)(s->file, "\trx_bufs_err : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.rx_bufs_err); > > + (*s->cpu_fprintf)(s->file, "\tcompletions : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.completions); > > + (*s->cpu_fprintf)(s->file, "\tpoll_cq (bk) : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.poll_cq_from_bk); > > + (*s->cpu_fprintf)(s->file, "\tpoll_cq_ppoll_to : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.poll_cq_ppoll_to); > > + (*s->cpu_fprintf)(s->file, "\tpoll_cq (fe) : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.poll_cq_from_guest); > > + (*s->cpu_fprintf)(s->file, "\tmad_tx : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.mad_tx); > > + (*s->cpu_fprintf)(s->file, "\tmad_tx_err : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.mad_tx_err); > > + (*s->cpu_fprintf)(s->file, "\tmad_rx : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.mad_rx); > > + (*s->cpu_fprintf)(s->file, "\tmad_rx_err : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.mad_rx_err); > > + (*s->cpu_fprintf)(s->file, "\tmad_rx_bufs : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.mad_rx_bufs); > > + (*s->cpu_fprintf)(s->file, "\tmad_rx_bufs_err : %" PRId64 "\n", > > + dev->rdma_dev_res.stats.mad_rx_bufs_err); > > + (*s->cpu_fprintf)(s->file, "\tPDs : %" PRId32 "\n", > > + dev->rdma_dev_res.pd_tbl.used); > > + (*s->cpu_fprintf)(s->file, "\tMRs : %" PRId32 "\n", > > + dev->rdma_dev_res.mr_tbl.used); > > + (*s->cpu_fprintf)(s->file, "\tUCs : %" PRId32 "\n", > > + dev->rdma_dev_res.uc_tbl.used); > > + (*s->cpu_fprintf)(s->file, "\tQPs : %" PRId32 "\n", > > + dev->rdma_dev_res.qp_tbl.used); > > + (*s->cpu_fprintf)(s->file, "\tCQs : %" PRId32 "\n", > > + dev->rdma_dev_res.cq_tbl.used); > > + (*s->cpu_fprintf)(s->file, "\tCEQ_CTXs : %" PRId32 "\n", > > + dev->rdma_dev_res.cqe_ctx_tbl.used); > > +} > > + > > +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func) > > +{ > > + CPUListState s = { > > + .file = f, > > + .cpu_fprintf = fprintf_func, > > + }; > > + > > + g_slist_foreach(devices, pvrdma_dump_device_statistics, &s); > > +} > > + > > static void free_dev_ring(PCIDevice *pci_dev, PvrdmaRing *ring, > > void *ring_state) > > { > > @@ -618,6 +687,8 @@ static void pvrdma_realize(PCIDevice *pdev, Error **errp) > > dev->shutdown_notifier.notify = pvrdma_shutdown_notifier; > > qemu_register_shutdown_notifier(&dev->shutdown_notifier); > > > > + devices = g_slist_append(devices, pdev); > > + > > out: > > if (rc) { > > pvrdma_fini(pdev); > > @@ -627,6 +698,7 @@ out: > > > > static void pvrdma_exit(PCIDevice *pdev) > > { > > + devices = g_slist_remove(devices, pdev); > > pvrdma_fini(pdev); > > }
Yuval Shaia <yuval.shaia@oracle.com> writes: > On Mon, Feb 04, 2019 at 02:03:58PM +0100, Markus Armbruster wrote: >> Yuval Shaia <yuval.shaia@oracle.com> writes: >> >> > Signed-off-by: Yuval Shaia <yuval.shaia@oracle.com> >> > --- >> > hw/rdma/vmw/pvrdma.h | 1 + >> > hw/rdma/vmw/pvrdma_main.c | 72 +++++++++++++++++++++++++++++++++++++++ >> > 2 files changed, 73 insertions(+) >> > >> > diff --git a/hw/rdma/vmw/pvrdma.h b/hw/rdma/vmw/pvrdma.h >> > index 167706ec2c..dc10f21ca0 100644 >> > --- a/hw/rdma/vmw/pvrdma.h >> > +++ b/hw/rdma/vmw/pvrdma.h >> > @@ -133,5 +133,6 @@ static inline void post_interrupt(PVRDMADev *dev, unsigned vector) >> > } >> > >> > int pvrdma_exec_cmd(PVRDMADev *dev); >> > +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func); >> > >> > #endif >> >> The only user appears in the next patch. I'd squash the two patches. >> Matter of taste. > > Agree with you. > I just did to help reviewers so ones that familiar with 'monitor' can > review the other patch while rdma folks can review this one. > > Will probably squash them as soon as the conversion on the other patch will > be over. > >> >> > diff --git a/hw/rdma/vmw/pvrdma_main.c b/hw/rdma/vmw/pvrdma_main.c >> > index cf82e78f08..79900076ec 100644 >> > --- a/hw/rdma/vmw/pvrdma_main.c >> > +++ b/hw/rdma/vmw/pvrdma_main.c >> > @@ -14,6 +14,7 @@ >> > */ >> > >> > #include "qemu/osdep.h" >> > +#include "qemu/units.h" >> > #include "qapi/error.h" >> > #include "hw/hw.h" >> > #include "hw/pci/pci.h" >> > @@ -36,6 +37,8 @@ >> > #include "standard-headers/drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h" >> > #include "pvrdma_qp_ops.h" >> > >> > +GSList *devices; >> > + >> > static Property pvrdma_dev_properties[] = { >> > DEFINE_PROP_STRING("netdev", PVRDMADev, backend_eth_device_name), >> > DEFINE_PROP_STRING("ibdev", PVRDMADev, backend_device_name), >> > @@ -55,6 +58,72 @@ static Property pvrdma_dev_properties[] = { >> > DEFINE_PROP_END_OF_LIST(), >> > }; >> > >> > +static void pvrdma_dump_device_statistics(gpointer data, gpointer user_data) >> > +{ >> > + CPUListState *s = user_data; >> > + PCIDevice *pdev = data; >> > + PVRDMADev *dev = PVRDMA_DEV(pdev); >> > + >> > + (*s->cpu_fprintf)(s->file, "%s_%x.%x\n", pdev->name, >> > + PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn)); >> >> Why the indirection through CPUListState? What's wrong with straight >> monitor_printf()? > > No special reasoning, just wanted to utilize an existing mechanism and > design. Please use the simplest available mechanism unless you have an actual need for a more complex one. Complicating things just to prepare for future needs rarely pays off, because YAGNI (you ain't gonna need it).
diff --git a/hw/rdma/vmw/pvrdma.h b/hw/rdma/vmw/pvrdma.h index 167706ec2c..dc10f21ca0 100644 --- a/hw/rdma/vmw/pvrdma.h +++ b/hw/rdma/vmw/pvrdma.h @@ -133,5 +133,6 @@ static inline void post_interrupt(PVRDMADev *dev, unsigned vector) } int pvrdma_exec_cmd(PVRDMADev *dev); +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func); #endif diff --git a/hw/rdma/vmw/pvrdma_main.c b/hw/rdma/vmw/pvrdma_main.c index cf82e78f08..79900076ec 100644 --- a/hw/rdma/vmw/pvrdma_main.c +++ b/hw/rdma/vmw/pvrdma_main.c @@ -14,6 +14,7 @@ */ #include "qemu/osdep.h" +#include "qemu/units.h" #include "qapi/error.h" #include "hw/hw.h" #include "hw/pci/pci.h" @@ -36,6 +37,8 @@ #include "standard-headers/drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h" #include "pvrdma_qp_ops.h" +GSList *devices; + static Property pvrdma_dev_properties[] = { DEFINE_PROP_STRING("netdev", PVRDMADev, backend_eth_device_name), DEFINE_PROP_STRING("ibdev", PVRDMADev, backend_device_name), @@ -55,6 +58,72 @@ static Property pvrdma_dev_properties[] = { DEFINE_PROP_END_OF_LIST(), }; +static void pvrdma_dump_device_statistics(gpointer data, gpointer user_data) +{ + CPUListState *s = user_data; + PCIDevice *pdev = data; + PVRDMADev *dev = PVRDMA_DEV(pdev); + + (*s->cpu_fprintf)(s->file, "%s_%x.%x\n", pdev->name, + PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn)); + (*s->cpu_fprintf)(s->file, "\tcommands : %" PRId64 "\n", + dev->stats.commands); + (*s->cpu_fprintf)(s->file, "\ttx : %" PRId64 "\n", + dev->rdma_dev_res.stats.tx); + (*s->cpu_fprintf)(s->file, "\ttx_len : %" PRId64 "\n", + dev->rdma_dev_res.stats.tx_len); + (*s->cpu_fprintf)(s->file, "\ttx_err : %" PRId64 "\n", + dev->rdma_dev_res.stats.tx_err); + (*s->cpu_fprintf)(s->file, "\trx_bufs : %" PRId64 "\n", + dev->rdma_dev_res.stats.rx_bufs); + (*s->cpu_fprintf)(s->file, "\trx_bufs_len : %" PRId64 "\n", + dev->rdma_dev_res.stats.rx_bufs_len); + (*s->cpu_fprintf)(s->file, "\trx_bufs_err : %" PRId64 "\n", + dev->rdma_dev_res.stats.rx_bufs_err); + (*s->cpu_fprintf)(s->file, "\tcompletions : %" PRId64 "\n", + dev->rdma_dev_res.stats.completions); + (*s->cpu_fprintf)(s->file, "\tpoll_cq (bk) : %" PRId64 "\n", + dev->rdma_dev_res.stats.poll_cq_from_bk); + (*s->cpu_fprintf)(s->file, "\tpoll_cq_ppoll_to : %" PRId64 "\n", + dev->rdma_dev_res.stats.poll_cq_ppoll_to); + (*s->cpu_fprintf)(s->file, "\tpoll_cq (fe) : %" PRId64 "\n", + dev->rdma_dev_res.stats.poll_cq_from_guest); + (*s->cpu_fprintf)(s->file, "\tmad_tx : %" PRId64 "\n", + dev->rdma_dev_res.stats.mad_tx); + (*s->cpu_fprintf)(s->file, "\tmad_tx_err : %" PRId64 "\n", + dev->rdma_dev_res.stats.mad_tx_err); + (*s->cpu_fprintf)(s->file, "\tmad_rx : %" PRId64 "\n", + dev->rdma_dev_res.stats.mad_rx); + (*s->cpu_fprintf)(s->file, "\tmad_rx_err : %" PRId64 "\n", + dev->rdma_dev_res.stats.mad_rx_err); + (*s->cpu_fprintf)(s->file, "\tmad_rx_bufs : %" PRId64 "\n", + dev->rdma_dev_res.stats.mad_rx_bufs); + (*s->cpu_fprintf)(s->file, "\tmad_rx_bufs_err : %" PRId64 "\n", + dev->rdma_dev_res.stats.mad_rx_bufs_err); + (*s->cpu_fprintf)(s->file, "\tPDs : %" PRId32 "\n", + dev->rdma_dev_res.pd_tbl.used); + (*s->cpu_fprintf)(s->file, "\tMRs : %" PRId32 "\n", + dev->rdma_dev_res.mr_tbl.used); + (*s->cpu_fprintf)(s->file, "\tUCs : %" PRId32 "\n", + dev->rdma_dev_res.uc_tbl.used); + (*s->cpu_fprintf)(s->file, "\tQPs : %" PRId32 "\n", + dev->rdma_dev_res.qp_tbl.used); + (*s->cpu_fprintf)(s->file, "\tCQs : %" PRId32 "\n", + dev->rdma_dev_res.cq_tbl.used); + (*s->cpu_fprintf)(s->file, "\tCEQ_CTXs : %" PRId32 "\n", + dev->rdma_dev_res.cqe_ctx_tbl.used); +} + +void pvrdma_dump_statistics(FILE *f, fprintf_function fprintf_func) +{ + CPUListState s = { + .file = f, + .cpu_fprintf = fprintf_func, + }; + + g_slist_foreach(devices, pvrdma_dump_device_statistics, &s); +} + static void free_dev_ring(PCIDevice *pci_dev, PvrdmaRing *ring, void *ring_state) { @@ -618,6 +687,8 @@ static void pvrdma_realize(PCIDevice *pdev, Error **errp) dev->shutdown_notifier.notify = pvrdma_shutdown_notifier; qemu_register_shutdown_notifier(&dev->shutdown_notifier); + devices = g_slist_append(devices, pdev); + out: if (rc) { pvrdma_fini(pdev); @@ -627,6 +698,7 @@ out: static void pvrdma_exit(PCIDevice *pdev) { + devices = g_slist_remove(devices, pdev); pvrdma_fini(pdev); }
Signed-off-by: Yuval Shaia <yuval.shaia@oracle.com> --- hw/rdma/vmw/pvrdma.h | 1 + hw/rdma/vmw/pvrdma_main.c | 72 +++++++++++++++++++++++++++++++++++++++ 2 files changed, 73 insertions(+)