Message ID | 1652132902-27109-1-git-send-email-quic_jhugo@quicinc.com (mailing list archive) |
---|---|
Headers | show |
Series | hyperv compose_msi_msg fixups | expand |
> From: Jeffrey Hugo <quic_jhugo@quicinc.com> > Sent: Monday, May 9, 2022 2:48 PM Thank you Jeff for working out the patches! Thanks, -- Dexuan
On 5/9/2022 5:23 PM, Dexuan Cui wrote: >> From: Jeffrey Hugo <quic_jhugo@quicinc.com> >> Sent: Monday, May 9, 2022 2:48 PM > > Thank you Jeff for working out the patches! Thanks for the feedback and reviews. This certainly would have been more challenging without your assistance. -Jeff
From: Jeffrey Hugo <quic_jhugo@quicinc.com> Sent: Monday, May 9, 2022 2:48 PM > > While multi-MSI appears to work with pci-hyperv.c, there was a concern about > how linux was doing the ITRE allocations. Patch 2 addresses the concern. > > However, patch 2 exposed an issue with how compose_msi_msg() was freeing a > previous allocation when called for the Nth time. Imagine a driver using > pci_alloc_irq_vectors() to request 32 MSIs. This would cause compose_msi_msg() > to be called 32 times, once for each MSI. With patch 2, MSI0 would allocate > the ITREs needed, and MSI1-31 would use the cached information. Then the driver > uses request_irq() on MSI1-17. This would call compose_msi_msg() again on those > MSIs, which would again use the cached information. Then unmask() would be > called to retarget the MSIs to the right VCPU vectors. Finally, the driver > calls request_irq() on MSI0. This would call conpose_msi_msg(), which would > free the block of 32 MSIs, and allocate a new block. This would undo the > retarget of MSI1-17, and likely leave those MSIs targeting invalid VCPU vectors. > This is addressed by patch 1, which is introduced first to prevent a regression. > > Jeffrey Hugo (2): > PCI: hv: Reuse existing ITRE allocation in compose_msi_msg() > PCI: hv: Fix interrupt mapping for multi-MSI > > drivers/pci/controller/pci-hyperv.c | 76 ++++++++++++++++++++++++++++--------- > 1 file changed, 59 insertions(+), 17 deletions(-) > > -- > 2.7.4 I tested these two patches in combination with the earlier two on 5.18-rc6 in an ARM64 VM in Azure. This was a smoke-test to ensure everything compiled and that the changes aren't fundamentally broken on ARM64. The PCI device in this case is the Mellanox Virtual Function offered to VMs in Azure, which is MSI-X. As such, the new MSI "batch" handling is not tested. Tested-by: Michael Kelley <mikelley@microsoft.com>
On Mon, May 09, 2022 at 03:48:20PM -0600, Jeffrey Hugo wrote: > While multi-MSI appears to work with pci-hyperv.c, there was a concern about > how linux was doing the ITRE allocations. Patch 2 addresses the concern. > > However, patch 2 exposed an issue with how compose_msi_msg() was freeing a > previous allocation when called for the Nth time. Imagine a driver using > pci_alloc_irq_vectors() to request 32 MSIs. This would cause compose_msi_msg() > to be called 32 times, once for each MSI. With patch 2, MSI0 would allocate > the ITREs needed, and MSI1-31 would use the cached information. Then the driver > uses request_irq() on MSI1-17. This would call compose_msi_msg() again on those > MSIs, which would again use the cached information. Then unmask() would be > called to retarget the MSIs to the right VCPU vectors. Finally, the driver > calls request_irq() on MSI0. This would call conpose_msi_msg(), which would > free the block of 32 MSIs, and allocate a new block. This would undo the > retarget of MSI1-17, and likely leave those MSIs targeting invalid VCPU vectors. > This is addressed by patch 1, which is introduced first to prevent a regression. > > Jeffrey Hugo (2): > PCI: hv: Reuse existing ITRE allocation in compose_msi_msg() > PCI: hv: Fix interrupt mapping for multi-MSI > Applied to hyperv-next. Thanks.
On 5/11/2022 8:41 AM, Wei Liu wrote: > On Mon, May 09, 2022 at 03:48:20PM -0600, Jeffrey Hugo wrote: >> While multi-MSI appears to work with pci-hyperv.c, there was a concern about >> how linux was doing the ITRE allocations. Patch 2 addresses the concern. >> >> However, patch 2 exposed an issue with how compose_msi_msg() was freeing a >> previous allocation when called for the Nth time. Imagine a driver using >> pci_alloc_irq_vectors() to request 32 MSIs. This would cause compose_msi_msg() >> to be called 32 times, once for each MSI. With patch 2, MSI0 would allocate >> the ITREs needed, and MSI1-31 would use the cached information. Then the driver >> uses request_irq() on MSI1-17. This would call compose_msi_msg() again on those >> MSIs, which would again use the cached information. Then unmask() would be >> called to retarget the MSIs to the right VCPU vectors. Finally, the driver >> calls request_irq() on MSI0. This would call conpose_msi_msg(), which would >> free the block of 32 MSIs, and allocate a new block. This would undo the >> retarget of MSI1-17, and likely leave those MSIs targeting invalid VCPU vectors. >> This is addressed by patch 1, which is introduced first to prevent a regression. >> >> Jeffrey Hugo (2): >> PCI: hv: Reuse existing ITRE allocation in compose_msi_msg() >> PCI: hv: Fix interrupt mapping for multi-MSI >> > > Applied to hyperv-next. Thanks. Huh? I thought you wanted a V2. I was intending on sending that out today. -Jeff
On Wed, May 11, 2022 at 08:47:23AM -0600, Jeffrey Hugo wrote: > On 5/11/2022 8:41 AM, Wei Liu wrote: > > On Mon, May 09, 2022 at 03:48:20PM -0600, Jeffrey Hugo wrote: > > > While multi-MSI appears to work with pci-hyperv.c, there was a concern about > > > how linux was doing the ITRE allocations. Patch 2 addresses the concern. > > > > > > However, patch 2 exposed an issue with how compose_msi_msg() was freeing a > > > previous allocation when called for the Nth time. Imagine a driver using > > > pci_alloc_irq_vectors() to request 32 MSIs. This would cause compose_msi_msg() > > > to be called 32 times, once for each MSI. With patch 2, MSI0 would allocate > > > the ITREs needed, and MSI1-31 would use the cached information. Then the driver > > > uses request_irq() on MSI1-17. This would call compose_msi_msg() again on those > > > MSIs, which would again use the cached information. Then unmask() would be > > > called to retarget the MSIs to the right VCPU vectors. Finally, the driver > > > calls request_irq() on MSI0. This would call conpose_msi_msg(), which would > > > free the block of 32 MSIs, and allocate a new block. This would undo the > > > retarget of MSI1-17, and likely leave those MSIs targeting invalid VCPU vectors. > > > This is addressed by patch 1, which is introduced first to prevent a regression. > > > > > > Jeffrey Hugo (2): > > > PCI: hv: Reuse existing ITRE allocation in compose_msi_msg() > > > PCI: hv: Fix interrupt mapping for multi-MSI > > > > > > > Applied to hyperv-next. Thanks. > > Huh? I thought you wanted a V2. I was intending on sending that out today. > Please send them out. I will apply the new version. Thanks, Wei. > -Jeff
On 5/11/2022 9:19 AM, Wei Liu wrote: > On Wed, May 11, 2022 at 08:47:23AM -0600, Jeffrey Hugo wrote: >> On 5/11/2022 8:41 AM, Wei Liu wrote: >>> On Mon, May 09, 2022 at 03:48:20PM -0600, Jeffrey Hugo wrote: >>>> While multi-MSI appears to work with pci-hyperv.c, there was a concern about >>>> how linux was doing the ITRE allocations. Patch 2 addresses the concern. >>>> >>>> However, patch 2 exposed an issue with how compose_msi_msg() was freeing a >>>> previous allocation when called for the Nth time. Imagine a driver using >>>> pci_alloc_irq_vectors() to request 32 MSIs. This would cause compose_msi_msg() >>>> to be called 32 times, once for each MSI. With patch 2, MSI0 would allocate >>>> the ITREs needed, and MSI1-31 would use the cached information. Then the driver >>>> uses request_irq() on MSI1-17. This would call compose_msi_msg() again on those >>>> MSIs, which would again use the cached information. Then unmask() would be >>>> called to retarget the MSIs to the right VCPU vectors. Finally, the driver >>>> calls request_irq() on MSI0. This would call conpose_msi_msg(), which would >>>> free the block of 32 MSIs, and allocate a new block. This would undo the >>>> retarget of MSI1-17, and likely leave those MSIs targeting invalid VCPU vectors. >>>> This is addressed by patch 1, which is introduced first to prevent a regression. >>>> >>>> Jeffrey Hugo (2): >>>> PCI: hv: Reuse existing ITRE allocation in compose_msi_msg() >>>> PCI: hv: Fix interrupt mapping for multi-MSI >>>> >>> >>> Applied to hyperv-next. Thanks. >> >> Huh? I thought you wanted a V2. I was intending on sending that out today. >> > > Please send them out. I will apply the new version. Sure, sending shortly. -Jeff