Message ID | 20240604050940.859909-1-mhklinux@outlook.com (mailing list archive) |
---|---|
Headers | show |
Series | Hyper-V guests use Linux IRQs for channel interrupts | expand |
From: mhkelley58@gmail.com <mhkelley58@gmail.com> Sent: Monday, June 3, 2024 10:09 PM > > This patch set makes significant changes in how Linux guests running on Hyper-V > handle interrupts from VMBus devices. It's "RFC" because of the level of change, > and because I'm interested in macro-level feedback on the whole idea. The usual > detailed-level code review feedback is also welcome. I want to get back to this RFC patch set and incorporate the changes suggested by Thomas Gleixner. But before I do that, is there any high-level feedback on whether the whole idea should be pursued at all? I think there's value in having the VMBus devices appear in a standard way in /proc/irq, so that we can eventually get away from the custom stuff in /sys/bus/vmbus. It also makes VMBus devices show up in a way similar to virtio devices. But on the flip side, it's churn to code that's already working well. And having VMBus devices appear in /proc/irq means they are visible to user space utilities like irqbalance, giving them the ability to affect performance (particularly for networking). As such, this patch set has implications beyond just the Linux kernel, and probably means disabling irqbalance by default in Azure cloud images. I'm especially interested in input from Microsoft folks, since the changes are specific to Hyper-V, with implications for the Azure cloud. Michael > > Background > ---------- > Linux guests running on Hyper-V are presented with a range of synthetic > devices, including storage, networking, mouse, keyboard, etc. Guests are also > presented with management-related pseudo-devices, including time > synchronization, heartbeat, and host-initiated shutdown. These devices use the > VMBus connection between the host and the guest, and each device has one or > more VMBus channels for passing messages between the host and guest. > Separately, a control path exists for creating and tearing down VMBus channels > when needed. Hyper-V VMBus and its synthetic devices play a role similar to > virtio and virtio devices. > > The host interrupts the guest when it sends a new message to the guest in an > otherwise empty channel, and when it sends a new control message to the guest. > All interrupts are multiplexed onto a single per-CPU architectural interrupt as > defined by the processor architecture. On arm64, this architectural interrupt > is a PPI, and is represented in Linux as a Linux per-CPU IRQ. The x86/x64 > architecture does not provide per-CPU interrupts, so Linux reserves a fixed x86 > interrupt vector (HYPERVISOR_CALLBACK_VECTOR) on all CPUs, with a hard-coded > interrupt handler. No Linux IRQ is assigned on x86/x64. > > The interrupt handler for the architectural interrupt is the VMBus interrupt > handler, and it runs whenever the host generates an interrupt for a channel or > a control message. The VMBus interrupt handler consults data structures > provided by the per-CPU Hyper-V synthetic interrupt controller (synic) to > determine which channel or channels have a pending interrupt, or if there is a > pending control message. When doing this demultiplexing, the VMBus interrupt > handler directly invokes the per-device handlers, and processes any control > messages. > > In this scheme, the individual VMBus devices and their channels are not > modelled as Linux IRQs. /proc/interrupts shows counts for the top-level > architectural interrupt, but not for the individual VMBus devices. The VMBus > driver has implemented separate reporting of per-device data under /sys. > Similarly, /proc/irq/<nn> does not show affinity information for individual > VMBus devices, and the affinity cannot be changed via /proc/irq. Again, a > separate interface under /sys is provided for changing the vCPU that a VMBus > channel should interrupt. > > What's New > ---------- > Linux kernel community members have commented that it would be better for the > Hyper-V synic and the handling of VMBus channel interrupts to be modelled as > Linux IRQs. Then they would participate in existing IRQ handling mechanisms, > and not need something separate under /sys. This patch set provides such an > implementation. > > With this patch set, the output of /proc/interrupts looks like this in a Linux > guest on a Hyper-V x86/x64 Generation 2 VM with 8 vCPUs (the columns for > CPUs 2 thru 5 are omitted due to text width considerations): > > CPU0 CPU1 CPU6 CPU7 > 8: 0 0 0 0 IO-APIC 8-edge rtc0 > 9: 2 0 0 0 IO-APIC 9-fasteoi acpi > 24: 16841 0 0 0 VMBus 7 heartbeat > 25: 1 0 0 0 VMBus 9 shutdown > 26: 6752 0 0 0 VMBus 10 timesync > 27: 1 0 0 0 VMBus 11 vss > 28: 22095 0 0 0 VMBus 14 pri@storvsc1 > 29: 0 85 0 0 VMBus 15 pri@storvsc0 > 30: 3 0 0 0 VMBus 1 balloon > 31: 106 0 0 0 VMBus 3 mouse > 32: 73 0 0 0 VMBus 4 keyboard > 33: 6 0 0 0 VMBus 5 framebuffer > 34: 0 0 0 0 VMBus 13 netvsc > 35: 1 0 0 0 VMBus 8 kvp > 36: 0 0 0 0 VMBus 16 netvsc > 37: 0 0 0 0 VMBus 17 netvsc > 38: 0 0 0 0 VMBus 18 netvsc > 39: 0 0 2375 0 VMBus 19 netvsc > 40: 0 0 0 1178 VMBus 20 netvsc > 41: 1003 0 0 0 VMBus 21 netvsc > 42: 0 4337 0 0 VMBus 22 netvsc > 43: 0 0 0 0 VMBus 23 sub@storvsc1 > 44: 0 0 0 0 VMBus 24 sub@storvsc0 > NMI: 0 0 0 0 Non-maskable interrupts > LOC: 0 0 0 0 Local timer interrupts > SPU: 0 0 0 0 Spurious interrupts > PMI: 0 0 0 0 Performance monitoring interrupts > IWI: 1 0 0 2 IRQ work interrupts > RTR: 0 0 0 0 APIC ICR read retries > RES: 411 233 349 318 Rescheduling interrupts > CAL: 135236 29211 99526 36424 Function call interrupts > TLB: 0 0 0 0 TLB shootdowns > TRM: 0 0 0 0 Thermal event interrupts > THR: 0 0 0 0 Threshold APIC interrupts > DFR: 0 0 0 0 Deferred Error APIC interrupts > MCE: 0 0 0 0 Machine check exceptions > MCP: 109 109 109 109 Machine check polls > HYP: 71 0 0 0 Hypervisor callback interrupts > HRE: 0 0 0 0 Hyper-V reenlightenment interrupts > HVS: 183391 109695 181539 339239 Hyper-V stimer0 interrupts > ERR: 0 > MIS: 0 > PIN: 0 0 0 0 Posted-interrupt notification event > NPI: 0 0 0 0 Nested posted-interrupt event > PIW: 0 0 0 0 Posted-interrupt wakeup event > > IRQs 24 thru 44 are the VMBus channel IRQs that didn't previously exist. Some > devices, such as storvsc and netvsc, have multiple channels, each of which has > a separate IRQ. The HYP line now shows counts only for control messages instead > of for all VMBus interrupts. The HVS line continues to show Hyper-V synthetic > timer (stimer0) interrupts. (In older versions of Hyper-V where stimer0 > interrupts are delivered as control messages, those interrupts are accounted > for on the HYP line. Since this is legacy behavior, it seemed unnecessary to > create a separate Linux IRQ to model these messages.) > > The basic approach is to update the VMBus channel create/open/close/delete > infrastructure, and the VMBus interrupt handler, to create and process the > additional IRQs. The goal is that individual VMBus drivers require no code > changes to work with the new IRQs. However, a driver may optionally annotate > the IRQ using knowledge that only the driver has. For example, in the above > /proc/interrupts output, the storvsc driver for synthetic disks has been > updated to distinguish the primary channel from the subchannels, and to > distinguish controller 0 from controller 1. Since storvsc models a SCSI > controller, the numeric designations are set to match the "host0" and "host1" > SCSI controllers that could be seen with "lsscsi -H -v". Similarly annotating > the "netvsc" entries is a "to do" item. > > In furtherance of the "no changes to existing VMBus drivers" goal, all VMBus > IRQs use the same interrupt handler. Each channel has a callback function > associated with it, and the interrupt handler invokes this callback in the same > way as before VMBus IRQs were introduced. > > VMBus code creates a standalone linear IRQ domain for the VMBus IRQs. The hwirq > #'s are the VMBus channel relid's, which are integers in the range 1 to 2047. > Each relid uniquely identifies a VMBus channel, and a channel interrupt from > the host provides the relid. A mapping from IRQ to hwirq (relid) is created > when a new channel is created, and the mapping is deleted when the channel is > deleted, so adding and removing VMBus devices in a running VM is fully > functional. > > Getting the IRQ counting correct requires a new IRQ flow handler in > /kernel/irq/chip.c. See Patch 6. The existing flow handlers don't handle this > case, and creating a custom flow handler outside of chip.c isn't feasible > because the needed components aren't available to code outside of /kernel/irq. > > When a guest CPU is taken offline, Linux automatically changes the affinity of > IRQs assigned to that CPU so that the IRQ is handled on another CPU that is > still online. Unfortunately, VMBus channel IRQs are not able to take advantage > of that mechanism. At the point the mechanism runs, it isn't possible to know > when Hyper-V started interrupting the new CPU; hence pending interrupts may be > stranded on the old offline CPU. Existing VMBus code in Linux prevents a CPU > from going offline when a VMBus channel interrupt is affined to it, and that > code must be retained. Of course, VMBus channel IRQs can manually be affined to > a different CPU, which could then allow a CPU to be taken offline. I have an > idea for a somewhat convoluted way to allow the automatic Linux mechanism to > work, but that will be another patch set. > > Having VMBus channel interrupts appear in /proc/irq means that user-space > irqbalance can change the affinity of the interrupts, where previously it could > not. For example, this gives irqbalance the ability to change (and perhaps > degrade) the default affinity configuration that was designed to maximize > network throughput. This is yet another reason to disable irqbalance in VM > images. > > The custom VMBus entries under /sys are retained for backwards compatibility, > but they are integrated. For example, changing the affinity via /proc/irq is > reflected in /sys/bus/vmbus/devices/<guid>/channels/<nn>/cpu, and vice versa. > > Patches > ------- > The patches have been tested on x86/x64 Generation 1 and Generation 2 VMs, and > on arm64 VMs. > > * Patches 1 and 2 fixes some error handling that is needed by subsequent > patches. They could be applied independent of this patch set. > > * Patches 3,4, and 5 add support for IRQ names, and add annotations in the > storvsc and the Hyper-V vPCI driver so that descriptions in /proc/interrupts > convey useful information. > > * Patch 6 adds a new IRQ flow handler needed to properly account for IRQs as > VMBus IRQs or as the top-level architectural interrupt. > > * Patch 7 creates the new IRQ domain for VMBus channel IRQs. > > * Patch 8 allocates the VMBus channel IRQs, and creates the mapping between > VMBus relid (used as the hwirq) and the Linux IRQ number. That mapping is then > used for lookups. > > * Patch 9 does request_irq() for the new VMBus channel IRQs, and converts the > top-level VMBus architectural interrupt handler to use the new VMBus channel > IRQs for handling channel interrupts. With this patch, /proc/interrupts shows > counts for the VMBus channel IRQs. > > * Patch 10 implements the irq_set_affinity() function for VMBus channel IRQs, > and integrates it with the existing Hyper-V-specific sysfs mechanism for > changing the CPU that a VMBus channel will interrupt. > > * Patch 11 updates hv_synic_cleanup() during CPU offlining to handle the lack > of waiting for the MODIFYCHANNEL message in vmbus_irq_set_affinity() in Patch > 10. > > * Patch 12 fixes a race in VMBus channel interrupt assignments that existed > prior to this patch set when taking a CPU offline. > > Open Topics > ----------- > * The performance impact of the additional IRQ processing in the interrupt path > appears to be minimal, based on looking at the code. Measurements are still to > be taken to confirm this. > > * Does the netvsc driver have sufficient information to annotate what appears > in /proc/interrupts, particularly when multiple synthetic NICs are present? At > first glance, it appears that the identifying info isn't generated until after > vmbus_open() is called, and by then it's too late to do the IRQ annotation. > Need some input from a netvsc expert. > > * The VMBus irq domain and irq chip data structures are placed in the > vmbus_connection structure. Starting with VMBus protocol version 5.0, the VMBus > host side supports multiple connections from a single guest instance (see > commit ae20b254306a6). While there's no upstream kernel code that creates > multiple VMBus connections, it's unclear whether the IRQ domain should be > global across all VMBus connection, or per connection. Are relid's global (to > the VM) or per connection? > > * I have not tried or tested any hv_sock channels. Is there a handy test > program that would be convenient for opening multiple hv_sock connections to > the guest, have them persist long enough to see what /proc/interrupts shows, > and try changing the IRQ affinity? > > * I have not tried or tested Linux guest hibernation paths. > > The patches build and run against 6.10-rc2 and next-20240603. > > Michael Kelley (12): > Drivers: hv: vmbus: Drop unsupported VMBus devices earlier > Drivers: hv: vmbus: Fix error path that deletes non-existent sysfs > group > Drivers: hv: vmbus: Add an IRQ name to VMBus channels > PCI: hv: Annotate the VMBus channel IRQ name > scsi: storvsc: Annotate the VMBus channel IRQ name > genirq: Add per-cpu flow handler with conditional IRQ stats > Drivers: hv: vmbus: Set up irqdomain and irqchip for the VMBus > connection > Drivers: hv: vmbus: Allocate an IRQ per channel and use for relid > mapping > Drivers: hv: vmbus: Use Linux IRQs to handle VMBus channel interrupts > Drivers: hv: vmbus: Implement vmbus_irq_set_affinity > Drivers: hv: vmbus: Wait for MODIFYCHANNEL to finish when offlining > CPUs > Drivers: hv: vmbus: Ensure IRQ affinity isn't set to a CPU going > offline > > arch/x86/kernel/cpu/mshyperv.c | 9 +- > drivers/hv/channel.c | 125 ++++------ > drivers/hv/channel_mgmt.c | 139 ++++++++---- > drivers/hv/connection.c | 48 ++-- > drivers/hv/hv.c | 90 +++++++- > drivers/hv/hv_common.c | 2 +- > drivers/hv/hyperv_vmbus.h | 22 +- > drivers/hv/vmbus_drv.c | 338 +++++++++++++++++++--------- > drivers/pci/controller/pci-hyperv.c | 5 + > drivers/scsi/storvsc_drv.c | 9 + > include/asm-generic/mshyperv.h | 9 +- > include/linux/hyperv.h | 9 + > include/linux/irq.h | 1 + > kernel/irq/chip.c | 29 +++ > 14 files changed, 567 insertions(+), 268 deletions(-) > > -- > 2.25.1 >
From: Michael Kelley <mhklinux@outlook.com> This patch set makes significant changes in how Linux guests running on Hyper-V handle interrupts from VMBus devices. It's "RFC" because of the level of change, and because I'm interested in macro-level feedback on the whole idea. The usual detailed-level code review feedback is also welcome. Background ---------- Linux guests running on Hyper-V are presented with a range of synthetic devices, including storage, networking, mouse, keyboard, etc. Guests are also presented with management-related pseudo-devices, including time synchronization, heartbeat, and host-initiated shutdown. These devices use the VMBus connection between the host and the guest, and each device has one or more VMBus channels for passing messages between the host and guest. Separately, a control path exists for creating and tearing down VMBus channels when needed. Hyper-V VMBus and its synthetic devices play a role similar to virtio and virtio devices. The host interrupts the guest when it sends a new message to the guest in an otherwise empty channel, and when it sends a new control message to the guest. All interrupts are multiplexed onto a single per-CPU architectural interrupt as defined by the processor architecture. On arm64, this architectural interrupt is a PPI, and is represented in Linux as a Linux per-CPU IRQ. The x86/x64 architecture does not provide per-CPU interrupts, so Linux reserves a fixed x86 interrupt vector (HYPERVISOR_CALLBACK_VECTOR) on all CPUs, with a hard-coded interrupt handler. No Linux IRQ is assigned on x86/x64. The interrupt handler for the architectural interrupt is the VMBus interrupt handler, and it runs whenever the host generates an interrupt for a channel or a control message. The VMBus interrupt handler consults data structures provided by the per-CPU Hyper-V synthetic interrupt controller (synic) to determine which channel or channels have a pending interrupt, or if there is a pending control message. When doing this demultiplexing, the VMBus interrupt handler directly invokes the per-device handlers, and processes any control messages. In this scheme, the individual VMBus devices and their channels are not modelled as Linux IRQs. /proc/interrupts shows counts for the top-level architectural interrupt, but not for the individual VMBus devices. The VMBus driver has implemented separate reporting of per-device data under /sys. Similarly, /proc/irq/<nn> does not show affinity information for individual VMBus devices, and the affinity cannot be changed via /proc/irq. Again, a separate interface under /sys is provided for changing the vCPU that a VMBus channel should interrupt. What's New ---------- Linux kernel community members have commented that it would be better for the Hyper-V synic and the handling of VMBus channel interrupts to be modelled as Linux IRQs. Then they would participate in existing IRQ handling mechanisms, and not need something separate under /sys. This patch set provides such an implementation. With this patch set, the output of /proc/interrupts looks like this in a Linux guest on a Hyper-V x86/x64 Generation 2 VM with 8 vCPUs (the columns for CPUs 2 thru 5 are omitted due to text width considerations): CPU0 CPU1 CPU6 CPU7 8: 0 0 0 0 IO-APIC 8-edge rtc0 9: 2 0 0 0 IO-APIC 9-fasteoi acpi 24: 16841 0 0 0 VMBus 7 heartbeat 25: 1 0 0 0 VMBus 9 shutdown 26: 6752 0 0 0 VMBus 10 timesync 27: 1 0 0 0 VMBus 11 vss 28: 22095 0 0 0 VMBus 14 pri@storvsc1 29: 0 85 0 0 VMBus 15 pri@storvsc0 30: 3 0 0 0 VMBus 1 balloon 31: 106 0 0 0 VMBus 3 mouse 32: 73 0 0 0 VMBus 4 keyboard 33: 6 0 0 0 VMBus 5 framebuffer 34: 0 0 0 0 VMBus 13 netvsc 35: 1 0 0 0 VMBus 8 kvp 36: 0 0 0 0 VMBus 16 netvsc 37: 0 0 0 0 VMBus 17 netvsc 38: 0 0 0 0 VMBus 18 netvsc 39: 0 0 2375 0 VMBus 19 netvsc 40: 0 0 0 1178 VMBus 20 netvsc 41: 1003 0 0 0 VMBus 21 netvsc 42: 0 4337 0 0 VMBus 22 netvsc 43: 0 0 0 0 VMBus 23 sub@storvsc1 44: 0 0 0 0 VMBus 24 sub@storvsc0 NMI: 0 0 0 0 Non-maskable interrupts LOC: 0 0 0 0 Local timer interrupts SPU: 0 0 0 0 Spurious interrupts PMI: 0 0 0 0 Performance monitoring interrupts IWI: 1 0 0 2 IRQ work interrupts RTR: 0 0 0 0 APIC ICR read retries RES: 411 233 349 318 Rescheduling interrupts CAL: 135236 29211 99526 36424 Function call interrupts TLB: 0 0 0 0 TLB shootdowns TRM: 0 0 0 0 Thermal event interrupts THR: 0 0 0 0 Threshold APIC interrupts DFR: 0 0 0 0 Deferred Error APIC interrupts MCE: 0 0 0 0 Machine check exceptions MCP: 109 109 109 109 Machine check polls HYP: 71 0 0 0 Hypervisor callback interrupts HRE: 0 0 0 0 Hyper-V reenlightenment interrupts HVS: 183391 109695 181539 339239 Hyper-V stimer0 interrupts ERR: 0 MIS: 0 PIN: 0 0 0 0 Posted-interrupt notification event NPI: 0 0 0 0 Nested posted-interrupt event PIW: 0 0 0 0 Posted-interrupt wakeup event IRQs 24 thru 44 are the VMBus channel IRQs that didn't previously exist. Some devices, such as storvsc and netvsc, have multiple channels, each of which has a separate IRQ. The HYP line now shows counts only for control messages instead of for all VMBus interrupts. The HVS line continues to show Hyper-V synthetic timer (stimer0) interrupts. (In older versions of Hyper-V where stimer0 interrupts are delivered as control messages, those interrupts are accounted for on the HYP line. Since this is legacy behavior, it seemed unnecessary to create a separate Linux IRQ to model these messages.) The basic approach is to update the VMBus channel create/open/close/delete infrastructure, and the VMBus interrupt handler, to create and process the additional IRQs. The goal is that individual VMBus drivers require no code changes to work with the new IRQs. However, a driver may optionally annotate the IRQ using knowledge that only the driver has. For example, in the above /proc/interrupts output, the storvsc driver for synthetic disks has been updated to distinguish the primary channel from the subchannels, and to distinguish controller 0 from controller 1. Since storvsc models a SCSI controller, the numeric designations are set to match the "host0" and "host1" SCSI controllers that could be seen with "lsscsi -H -v". Similarly annotating the "netvsc" entries is a "to do" item. In furtherance of the "no changes to existing VMBus drivers" goal, all VMBus IRQs use the same interrupt handler. Each channel has a callback function associated with it, and the interrupt handler invokes this callback in the same way as before VMBus IRQs were introduced. VMBus code creates a standalone linear IRQ domain for the VMBus IRQs. The hwirq #'s are the VMBus channel relid's, which are integers in the range 1 to 2047. Each relid uniquely identifies a VMBus channel, and a channel interrupt from the host provides the relid. A mapping from IRQ to hwirq (relid) is created when a new channel is created, and the mapping is deleted when the channel is deleted, so adding and removing VMBus devices in a running VM is fully functional. Getting the IRQ counting correct requires a new IRQ flow handler in /kernel/irq/chip.c. See Patch 6. The existing flow handlers don't handle this case, and creating a custom flow handler outside of chip.c isn't feasible because the needed components aren't available to code outside of /kernel/irq. When a guest CPU is taken offline, Linux automatically changes the affinity of IRQs assigned to that CPU so that the IRQ is handled on another CPU that is still online. Unfortunately, VMBus channel IRQs are not able to take advantage of that mechanism. At the point the mechanism runs, it isn't possible to know when Hyper-V started interrupting the new CPU; hence pending interrupts may be stranded on the old offline CPU. Existing VMBus code in Linux prevents a CPU from going offline when a VMBus channel interrupt is affined to it, and that code must be retained. Of course, VMBus channel IRQs can manually be affined to a different CPU, which could then allow a CPU to be taken offline. I have an idea for a somewhat convoluted way to allow the automatic Linux mechanism to work, but that will be another patch set. Having VMBus channel interrupts appear in /proc/irq means that user-space irqbalance can change the affinity of the interrupts, where previously it could not. For example, this gives irqbalance the ability to change (and perhaps degrade) the default affinity configuration that was designed to maximize network throughput. This is yet another reason to disable irqbalance in VM images. The custom VMBus entries under /sys are retained for backwards compatibility, but they are integrated. For example, changing the affinity via /proc/irq is reflected in /sys/bus/vmbus/devices/<guid>/channels/<nn>/cpu, and vice versa. Patches ------- The patches have been tested on x86/x64 Generation 1 and Generation 2 VMs, and on arm64 VMs. * Patches 1 and 2 fixes some error handling that is needed by subsequent patches. They could be applied independent of this patch set. * Patches 3,4, and 5 add support for IRQ names, and add annotations in the storvsc and the Hyper-V vPCI driver so that descriptions in /proc/interrupts convey useful information. * Patch 6 adds a new IRQ flow handler needed to properly account for IRQs as VMBus IRQs or as the top-level architectural interrupt. * Patch 7 creates the new IRQ domain for VMBus channel IRQs. * Patch 8 allocates the VMBus channel IRQs, and creates the mapping between VMBus relid (used as the hwirq) and the Linux IRQ number. That mapping is then used for lookups. * Patch 9 does request_irq() for the new VMBus channel IRQs, and converts the top-level VMBus architectural interrupt handler to use the new VMBus channel IRQs for handling channel interrupts. With this patch, /proc/interrupts shows counts for the VMBus channel IRQs. * Patch 10 implements the irq_set_affinity() function for VMBus channel IRQs, and integrates it with the existing Hyper-V-specific sysfs mechanism for changing the CPU that a VMBus channel will interrupt. * Patch 11 updates hv_synic_cleanup() during CPU offlining to handle the lack of waiting for the MODIFYCHANNEL message in vmbus_irq_set_affinity() in Patch 10. * Patch 12 fixes a race in VMBus channel interrupt assignments that existed prior to this patch set when taking a CPU offline. Open Topics ----------- * The performance impact of the additional IRQ processing in the interrupt path appears to be minimal, based on looking at the code. Measurements are still to be taken to confirm this. * Does the netvsc driver have sufficient information to annotate what appears in /proc/interrupts, particularly when multiple synthetic NICs are present? At first glance, it appears that the identifying info isn't generated until after vmbus_open() is called, and by then it's too late to do the IRQ annotation. Need some input from a netvsc expert. * The VMBus irq domain and irq chip data structures are placed in the vmbus_connection structure. Starting with VMBus protocol version 5.0, the VMBus host side supports multiple connections from a single guest instance (see commit ae20b254306a6). While there's no upstream kernel code that creates multiple VMBus connections, it's unclear whether the IRQ domain should be global across all VMBus connection, or per connection. Are relid's global (to the VM) or per connection? * I have not tried or tested any hv_sock channels. Is there a handy test program that would be convenient for opening multiple hv_sock connections to the guest, have them persist long enough to see what /proc/interrupts shows, and try changing the IRQ affinity? * I have not tried or tested Linux guest hibernation paths. The patches build and run against 6.10-rc2 and next-20240603. Michael Kelley (12): Drivers: hv: vmbus: Drop unsupported VMBus devices earlier Drivers: hv: vmbus: Fix error path that deletes non-existent sysfs group Drivers: hv: vmbus: Add an IRQ name to VMBus channels PCI: hv: Annotate the VMBus channel IRQ name scsi: storvsc: Annotate the VMBus channel IRQ name genirq: Add per-cpu flow handler with conditional IRQ stats Drivers: hv: vmbus: Set up irqdomain and irqchip for the VMBus connection Drivers: hv: vmbus: Allocate an IRQ per channel and use for relid mapping Drivers: hv: vmbus: Use Linux IRQs to handle VMBus channel interrupts Drivers: hv: vmbus: Implement vmbus_irq_set_affinity Drivers: hv: vmbus: Wait for MODIFYCHANNEL to finish when offlining CPUs Drivers: hv: vmbus: Ensure IRQ affinity isn't set to a CPU going offline arch/x86/kernel/cpu/mshyperv.c | 9 +- drivers/hv/channel.c | 125 ++++------ drivers/hv/channel_mgmt.c | 139 ++++++++---- drivers/hv/connection.c | 48 ++-- drivers/hv/hv.c | 90 +++++++- drivers/hv/hv_common.c | 2 +- drivers/hv/hyperv_vmbus.h | 22 +- drivers/hv/vmbus_drv.c | 338 +++++++++++++++++++--------- drivers/pci/controller/pci-hyperv.c | 5 + drivers/scsi/storvsc_drv.c | 9 + include/asm-generic/mshyperv.h | 9 +- include/linux/hyperv.h | 9 + include/linux/irq.h | 1 + kernel/irq/chip.c | 29 +++ 14 files changed, 567 insertions(+), 268 deletions(-)