Message ID | a34cef71-d60f-43ae-f61d-13d6c846eaa4@suse.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | runstate/time area registration by (guest) physical address | expand |
On Thu, Sep 28, 2023 at 09:16:48AM +0200, Jan Beulich wrote: > The registration by virtual/linear address has downsides: At least on > x86 the access is expensive for HVM/PVH domains. Furthermore for 64-bit > PV domains the areas are inaccessible (and hence cannot be updated by > Xen) when in guest-user mode, and for HVM guests they may be > inaccessible when Meltdown mitigations are in place. (There are yet > more issues.) > > In preparation of the introduction of new vCPU operations allowing to > register the respective areas (one of the two is x86-specific) by > guest-physical address, flesh out the map/unmap functions. > > Noteworthy differences from map_vcpu_info(): > - areas can be registered more than once (and de-registered), > - remote vCPU-s are paused rather than checked for being down (which in > principle can change right after the check), > - the domain lock is taken for a much smaller region. > > Signed-off-by: Jan Beulich <jbeulich@suse.com> Acked-by: Roger Pau Monné <roger.pau@citrix.com> Thanks, Roger.
On 28.09.2023 12:04, Roger Pau Monné wrote: > On Thu, Sep 28, 2023 at 09:16:48AM +0200, Jan Beulich wrote: >> The registration by virtual/linear address has downsides: At least on >> x86 the access is expensive for HVM/PVH domains. Furthermore for 64-bit >> PV domains the areas are inaccessible (and hence cannot be updated by >> Xen) when in guest-user mode, and for HVM guests they may be >> inaccessible when Meltdown mitigations are in place. (There are yet >> more issues.) >> >> In preparation of the introduction of new vCPU operations allowing to >> register the respective areas (one of the two is x86-specific) by >> guest-physical address, flesh out the map/unmap functions. >> >> Noteworthy differences from map_vcpu_info(): >> - areas can be registered more than once (and de-registered), >> - remote vCPU-s are paused rather than checked for being down (which in >> principle can change right after the check), >> - the domain lock is taken for a much smaller region. >> >> Signed-off-by: Jan Beulich <jbeulich@suse.com> > > Acked-by: Roger Pau Monné <roger.pau@citrix.com> Thanks, but to be clear: Formally this doesn't help this patch make any progress, aiui. I'll still need an A-b by a REST maintainer then. An R-b from you would be different in this regard. Jan
On Thu, Sep 28, 2023 at 12:14:17PM +0200, Jan Beulich wrote: > On 28.09.2023 12:04, Roger Pau Monné wrote: > > On Thu, Sep 28, 2023 at 09:16:48AM +0200, Jan Beulich wrote: > >> The registration by virtual/linear address has downsides: At least on > >> x86 the access is expensive for HVM/PVH domains. Furthermore for 64-bit > >> PV domains the areas are inaccessible (and hence cannot be updated by > >> Xen) when in guest-user mode, and for HVM guests they may be > >> inaccessible when Meltdown mitigations are in place. (There are yet > >> more issues.) > >> > >> In preparation of the introduction of new vCPU operations allowing to > >> register the respective areas (one of the two is x86-specific) by > >> guest-physical address, flesh out the map/unmap functions. > >> > >> Noteworthy differences from map_vcpu_info(): > >> - areas can be registered more than once (and de-registered), > >> - remote vCPU-s are paused rather than checked for being down (which in > >> principle can change right after the check), > >> - the domain lock is taken for a much smaller region. > >> > >> Signed-off-by: Jan Beulich <jbeulich@suse.com> > > > > Acked-by: Roger Pau Monné <roger.pau@citrix.com> > > Thanks, but to be clear: Formally this doesn't help this patch make any > progress, aiui. I'll still need an A-b by a REST maintainer then. An R-b > from you would be different in this regard. I see. Reviewed-by: Roger Pau Monné <roger.pau@citrix.com> Thanks, Roger.
--- a/xen/common/domain.c +++ b/xen/common/domain.c @@ -1605,7 +1605,82 @@ int map_guest_area(struct vcpu *v, paddr struct guest_area *area, void (*populate)(void *dst, struct vcpu *v)) { - return -EOPNOTSUPP; + struct domain *d = v->domain; + void *map = NULL; + struct page_info *pg = NULL; + int rc = 0; + + if ( ~gaddr ) /* Map (i.e. not just unmap)? */ + { + unsigned long gfn = PFN_DOWN(gaddr); + unsigned int align; + p2m_type_t p2mt; + + if ( gfn != PFN_DOWN(gaddr + size - 1) ) + return -ENXIO; + +#ifdef CONFIG_COMPAT + if ( has_32bit_shinfo(d) ) + align = alignof(compat_ulong_t); + else +#endif + align = alignof(xen_ulong_t); + if ( !IS_ALIGNED(gaddr, align) ) + return -ENXIO; + + rc = check_get_page_from_gfn(d, _gfn(gfn), false, &p2mt, &pg); + if ( rc ) + return rc; + + if ( !get_page_type(pg, PGT_writable_page) ) + { + put_page(pg); + return -EACCES; + } + + map = __map_domain_page_global(pg); + if ( !map ) + { + put_page_and_type(pg); + return -ENOMEM; + } + map += PAGE_OFFSET(gaddr); + } + + if ( v != current ) + { + if ( !spin_trylock(&d->hypercall_deadlock_mutex) ) + { + rc = -ERESTART; + goto unmap; + } + + vcpu_pause(v); + + spin_unlock(&d->hypercall_deadlock_mutex); + } + + domain_lock(d); + + if ( map && populate ) + populate(map, v); + + SWAP(area->pg, pg); + SWAP(area->map, map); + + domain_unlock(d); + + if ( v != current ) + vcpu_unpause(v); + + unmap: + if ( pg ) + { + unmap_domain_page_global(map); + put_page_and_type(pg); + } + + return rc; } /* @@ -1616,9 +1691,24 @@ int map_guest_area(struct vcpu *v, paddr void unmap_guest_area(struct vcpu *v, struct guest_area *area) { struct domain *d = v->domain; + void *map; + struct page_info *pg; if ( v != current ) ASSERT(atomic_read(&v->pause_count) | atomic_read(&d->pause_count)); + + domain_lock(d); + map = area->map; + area->map = NULL; + pg = area->pg; + area->pg = NULL; + domain_unlock(d); + + if ( pg ) + { + unmap_domain_page_global(map); + put_page_and_type(pg); + } } int default_initialise_vcpu(struct vcpu *v, XEN_GUEST_HANDLE_PARAM(void) arg)
The registration by virtual/linear address has downsides: At least on x86 the access is expensive for HVM/PVH domains. Furthermore for 64-bit PV domains the areas are inaccessible (and hence cannot be updated by Xen) when in guest-user mode, and for HVM guests they may be inaccessible when Meltdown mitigations are in place. (There are yet more issues.) In preparation of the introduction of new vCPU operations allowing to register the respective areas (one of the two is x86-specific) by guest-physical address, flesh out the map/unmap functions. Noteworthy differences from map_vcpu_info(): - areas can be registered more than once (and de-registered), - remote vCPU-s are paused rather than checked for being down (which in principle can change right after the check), - the domain lock is taken for a much smaller region. Signed-off-by: Jan Beulich <jbeulich@suse.com> --- RFC: By using global domain page mappings the demand on the underlying VA range may increase significantly. I did consider to use per- domain mappings instead, but they exist for x86 only. Of course we could have arch_{,un}map_guest_area() aliasing global domain page mapping functions on Arm and using per-domain mappings on x86. Yet then again map_vcpu_info() doesn't (and can't) do so. RFC: In map_guest_area() I'm not checking the P2M type, instead - just like map_vcpu_info() - solely relying on the type ref acquisition. Checking for p2m_ram_rw alone would be wrong, as at least p2m_ram_logdirty ought to also be okay to use here (and in similar cases, e.g. in Argo's find_ring_mfn()). p2m_is_pageable() could be used here (like altp2m_vcpu_enable_ve() does) as well as in map_vcpu_info(), yet then again the P2M type is stale by the time it is being looked at anyway without the P2M lock held. --- v4: Add another comment. Use IS_ALIGNED(). Add another NULL check. v2: currd -> d, to cover mem-sharing's copy_guest_area(). Re-base over change(s) earlier in the series. Use ~0 as "unmap" request indicator.