Message ID | 1564479235-25074-1-git-send-email-wanpengli@tencent.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | KVM: Disable wake-affine vCPU process to mitigate lock holder preemption | expand |
On 30/07/19 11:33, Wanpeng Li wrote: > When qemu/other vCPU inject virtual interrupt to guest through waking up one > sleeping vCPU, it increases the probability to stack vCPUs/qemu by scheduler > wake-affine. vCPU stacking issue can greately inceases the lock synchronization > latency in a virtualized environment. This patch disables wake-affine vCPU > process to mitigtate lock holder preemption. There is no guarantee that the vCPU remains on the thread where it's created, so the patch is not enough. If many vCPUs are stacked on the same pCPU, why doesn't the wake_cap kick in sooner or later? Paolo > Cc: Peter Zijlstra <peterz@infradead.org> > Cc: Thomas Gleixner <tglx@linutronix.de> > Cc: Paolo Bonzini <pbonzini@redhat.com> > Cc: Radim Krčmář <rkrcmar@redhat.com> > Signed-off-by: Wanpeng Li <wanpengli@tencent.com> > --- > include/linux/sched.h | 1 + > kernel/sched/fair.c | 3 +++ > virt/kvm/kvm_main.c | 1 + > 3 files changed, 5 insertions(+) > > diff --git a/include/linux/sched.h b/include/linux/sched.h > index 8dc1811..3dd33d8 100644 > --- a/include/linux/sched.h > +++ b/include/linux/sched.h > @@ -1468,6 +1468,7 @@ extern struct pid *cad_pid; > #define PF_NO_SETAFFINITY 0x04000000 /* Userland is not allowed to meddle with cpus_mask */ > #define PF_MCE_EARLY 0x08000000 /* Early kill for mce process policy */ > #define PF_MEMALLOC_NOCMA 0x10000000 /* All allocation request will have _GFP_MOVABLE cleared */ > +#define PF_NO_WAKE_AFFINE 0x20000000 /* This thread should not be wake affine */ > #define PF_FREEZER_SKIP 0x40000000 /* Freezer should not count it as freezable */ > #define PF_SUSPEND_TASK 0x80000000 /* This thread called freeze_processes() and should not be frozen */ > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > index 036be95..18eb1fa 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -5428,6 +5428,9 @@ static int wake_wide(struct task_struct *p) > unsigned int slave = p->wakee_flips; > int factor = this_cpu_read(sd_llc_size); > > + if (unlikely(p->flags & PF_NO_WAKE_AFFINE)) > + return 1; > + > if (master < slave) > swap(master, slave); > if (slave < factor || master < slave * factor) > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index 887f3b0..b9f75c3 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -2680,6 +2680,7 @@ static int kvm_vm_ioctl_create_vcpu(struct kvm *kvm, u32 id) > > mutex_unlock(&kvm->lock); > kvm_arch_vcpu_postcreate(vcpu); > + current->flags |= PF_NO_WAKE_AFFINE; > return r;
On Tue, Jul 30, 2019 at 05:33:55PM +0800, Wanpeng Li wrote: > From: Wanpeng Li <wanpengli@tencent.com> > > Wake-affine is a feature inside scheduler which we attempt to make processes > running closely, it gains benefit mostly from cache-hit. When waker tries > to wakup wakee, it needs to select cpu to run wakee, wake affine heuristic > mays select the cpu which waker is running on currently instead of the prev > cpu which wakee was last time running. > > However, in multiple VMs over-subscribe virtualization scenario, it increases > the probability to incur vCPU stacking which means that the sibling vCPUs from > the same VM will be stacked on one pCPU. I test three 80 vCPUs VMs running on > one 80 pCPUs Skylake server(PLE is supported), the ebizzy score can increase 17% > after disabling wake-affine for vCPU process. > > When qemu/other vCPU inject virtual interrupt to guest through waking up one > sleeping vCPU, it increases the probability to stack vCPUs/qemu by scheduler > wake-affine. vCPU stacking issue can greately inceases the lock synchronization > latency in a virtualized environment. This patch disables wake-affine vCPU > process to mitigtate lock holder preemption. > > Cc: Peter Zijlstra <peterz@infradead.org> > Cc: Thomas Gleixner <tglx@linutronix.de> > Cc: Paolo Bonzini <pbonzini@redhat.com> > Cc: Radim Krčmář <rkrcmar@redhat.com> > Signed-off-by: Wanpeng Li <wanpengli@tencent.com> > --- > include/linux/sched.h | 1 + > kernel/sched/fair.c | 3 +++ > virt/kvm/kvm_main.c | 1 + > 3 files changed, 5 insertions(+) > index 036be95..18eb1fa 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -5428,6 +5428,9 @@ static int wake_wide(struct task_struct *p) > unsigned int slave = p->wakee_flips; > int factor = this_cpu_read(sd_llc_size); > > + if (unlikely(p->flags & PF_NO_WAKE_AFFINE)) > + return 1; > + > if (master < slave) > swap(master, slave); > if (slave < factor || master < slave * factor) I intensely dislike how you misrepresent this patch as a KVM patch. Also the above is very much not the right place, even if this PF_flag were to live.
On Tue, 2019-07-30 at 13:46 +0200, Paolo Bonzini wrote: > On 30/07/19 11:33, Wanpeng Li wrote: > > When qemu/other vCPU inject virtual interrupt to guest through > > waking up one > > sleeping vCPU, it increases the probability to stack vCPUs/qemu by > > scheduler > > wake-affine. vCPU stacking issue can greately inceases the lock > > synchronization > > latency in a virtualized environment. This patch disables wake- > > affine vCPU > > process to mitigtate lock holder preemption. > > There is no guarantee that the vCPU remains on the thread where it's > created, so the patch is not enough. > > If many vCPUs are stacked on the same pCPU, why doesn't the wake_cap > kick in sooner or later? > Assuming it actually is the case that vcpus *do* get stacked *and* that wake_cap() *doesn't* kick in, maybe it could be because of this check? /* Minimum capacity is close to max, no need to abort wake_affine */ if (max_cap - min_cap < max_cap >> 3) return 0; Regards
On Tue, 2019-07-30 at 17:33 +0800, Wanpeng Li wrote: > However, in multiple VMs over-subscribe virtualization scenario, it > increases > the probability to incur vCPU stacking which means that the sibling > vCPUs from > the same VM will be stacked on one pCPU. I test three 80 vCPUs VMs > running on > one 80 pCPUs Skylake server(PLE is supported), the ebizzy score can > increase 17% > after disabling wake-affine for vCPU process. > Can't we achieve this by removing SD_WAKE_AFFINE from the relevant scheduling domains? By acting on /proc/sys/kernel/sched_domain/cpuX/domainY/flags, I mean? Of course this will impact all tasks, not only KVM vcpus. But if the host does KVM only anyway... Regards
On Thu, 1 Aug 2019 at 20:57, Dario Faggioli <dfaggioli@suse.com> wrote: > > On Tue, 2019-07-30 at 17:33 +0800, Wanpeng Li wrote: > > However, in multiple VMs over-subscribe virtualization scenario, it > > increases > > the probability to incur vCPU stacking which means that the sibling > > vCPUs from > > the same VM will be stacked on one pCPU. I test three 80 vCPUs VMs > > running on > > one 80 pCPUs Skylake server(PLE is supported), the ebizzy score can > > increase 17% > > after disabling wake-affine for vCPU process. > > > Can't we achieve this by removing SD_WAKE_AFFINE from the relevant > scheduling domains? By acting on > /proc/sys/kernel/sched_domain/cpuX/domainY/flags, I mean? > > Of course this will impact all tasks, not only KVM vcpus. But if the > host does KVM only anyway... Yes, not only kvm host and dedicated kvm host, unless introduce per-process flags, otherwise can't appeal to both. Regards, Wanpeng Li
Dario Faggioli writes: > On Tue, 2019-07-30 at 17:33 +0800, Wanpeng Li wrote: >> However, in multiple VMs over-subscribe virtualization scenario, it >> increases >> the probability to incur vCPU stacking which means that the sibling >> vCPUs from >> the same VM will be stacked on one pCPU. I test three 80 vCPUs VMs >> running on >> one 80 pCPUs Skylake server(PLE is supported), the ebizzy score can >> increase 17% >> after disabling wake-affine for vCPU process. >> > Can't we achieve this by removing SD_WAKE_AFFINE from the relevant > scheduling domains? By acting on > /proc/sys/kernel/sched_domain/cpuX/domainY/flags, I mean? > > Of course this will impact all tasks, not only KVM vcpus. But if the > host does KVM only anyway... Even a host dedicated to KVM has many non-KVM processes. I suspect an increasing number of hosts will be split between VMs and containers. > > Regards
On 01/08/19 14:57, Dario Faggioli wrote: > Can't we achieve this by removing SD_WAKE_AFFINE from the relevant > scheduling domains? By acting on > /proc/sys/kernel/sched_domain/cpuX/domainY/flags, I mean? > > Of course this will impact all tasks, not only KVM vcpus. But if the > host does KVM only anyway... Perhaps add flags to the unified cgroups hierarchy instead. But if the "min_cap close to max_cap" heuristics are wrong they should indeed be fixed. Paolo
diff --git a/include/linux/sched.h b/include/linux/sched.h index 8dc1811..3dd33d8 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -1468,6 +1468,7 @@ extern struct pid *cad_pid; #define PF_NO_SETAFFINITY 0x04000000 /* Userland is not allowed to meddle with cpus_mask */ #define PF_MCE_EARLY 0x08000000 /* Early kill for mce process policy */ #define PF_MEMALLOC_NOCMA 0x10000000 /* All allocation request will have _GFP_MOVABLE cleared */ +#define PF_NO_WAKE_AFFINE 0x20000000 /* This thread should not be wake affine */ #define PF_FREEZER_SKIP 0x40000000 /* Freezer should not count it as freezable */ #define PF_SUSPEND_TASK 0x80000000 /* This thread called freeze_processes() and should not be frozen */ diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 036be95..18eb1fa 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -5428,6 +5428,9 @@ static int wake_wide(struct task_struct *p) unsigned int slave = p->wakee_flips; int factor = this_cpu_read(sd_llc_size); + if (unlikely(p->flags & PF_NO_WAKE_AFFINE)) + return 1; + if (master < slave) swap(master, slave); if (slave < factor || master < slave * factor) diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 887f3b0..b9f75c3 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -2680,6 +2680,7 @@ static int kvm_vm_ioctl_create_vcpu(struct kvm *kvm, u32 id) mutex_unlock(&kvm->lock); kvm_arch_vcpu_postcreate(vcpu); + current->flags |= PF_NO_WAKE_AFFINE; return r; unlock_vcpu_destroy: