Message ID | 20240928085727.56883-2-phil@philjordan.eu (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | macOS PV Graphics and new vmapple machine type | expand |
On 2024/09/28 17:57, Phil Dennis-Jordan wrote: > MacOS provides a framework (library) that allows any vmm to implement a > paravirtualized 3d graphics passthrough to the host metal stack called > ParavirtualizedGraphics.Framework (PVG). The library abstracts away > almost every aspect of the paravirtualized device model and only provides > and receives callbacks on MMIO access as well as to share memory address > space between the VM and PVG. > > This patch implements a QEMU device that drives PVG for the VMApple > variant of it. I think it is better to name it MMIO variant instead of VMApple. There is nothing specific to VMApple in: hw/display/apple-gfx-vmapple.m > > Signed-off-by: Alexander Graf <graf@amazon.com> > Co-authored-by: Alexander Graf <graf@amazon.com> > > Subsequent changes: > > * Cherry-pick/rebase conflict fixes > * BQL function renaming > * Moved from hw/vmapple/ (useful outside that machine type) > * Code review comments: Switched to DEFINE_TYPES macro & little endian > MMIO. > * Removed some dead/superfluous code > * Mad set_mode thread & memory safe > * Added migration blocker due to lack of (de-)serialisation. > * Fixes to ObjC refcounting and autorelease pool usage. > * Fixed ObjC new/init misuse > * Switched to ObjC category extension for private property. > * Simplified task memory mapping and made it thread safe. > * Refactoring to split generic and vmapple MMIO variant specific > code. > * Switched to asynchronous MMIO writes on x86-64 > * Rendering and graphics update are now done asynchronously > * Fixed cursor handling > * Coding convention fixes > * Removed software cursor compositing > > Signed-off-by: Phil Dennis-Jordan <phil@philjordan.eu> > > --- > > v3: > > * Rebased on latest upstream, fixed breakages including switching to Resettable methods. > * Squashed patches dealing with dGPUs, MMIO area size, and GPU picking. > * Allow re-entrant MMIO; this simplifies the code and solves the divergence > between x86-64 and arm64 variants. > > hw/display/Kconfig | 9 + > hw/display/apple-gfx-vmapple.m | 215 +++++++++++++ > hw/display/apple-gfx.h | 57 ++++ > hw/display/apple-gfx.m | 536 +++++++++++++++++++++++++++++++++ > hw/display/meson.build | 2 + > hw/display/trace-events | 26 ++ > meson.build | 4 + > 7 files changed, 849 insertions(+) > create mode 100644 hw/display/apple-gfx-vmapple.m > create mode 100644 hw/display/apple-gfx.h > create mode 100644 hw/display/apple-gfx.m > > diff --git a/hw/display/Kconfig b/hw/display/Kconfig > index a4552c8ed78..179a479d220 100644 > --- a/hw/display/Kconfig > +++ b/hw/display/Kconfig > @@ -143,3 +143,12 @@ config XLNX_DISPLAYPORT > > config DM163 > bool > + > +config MAC_PVG > + bool > + default y > + > +config MAC_PVG_VMAPPLE > + bool > + depends on MAC_PVG > + depends on ARM Use AARCH64 instead. > diff --git a/hw/display/apple-gfx-vmapple.m b/hw/display/apple-gfx-vmapple.m > new file mode 100644 > index 00000000000..d8fc7651dde > --- /dev/null > +++ b/hw/display/apple-gfx-vmapple.m > @@ -0,0 +1,215 @@ > +/* > + * QEMU Apple ParavirtualizedGraphics.framework device, vmapple (arm64) variant > + * > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All Rights Reserved. > + * > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > + * See the COPYING file in the top-level directory. > + * > + * ParavirtualizedGraphics.framework is a set of libraries that macOS provides > + * which implements 3d graphics passthrough to the host as well as a > + * proprietary guest communication channel to drive it. This device model > + * implements support to drive that library from within QEMU as an MMIO-based > + * system device for macOS on arm64 VMs. > + */ > + > +#include "apple-gfx.h" > +#include "monitor/monitor.h" > +#include "hw/sysbus.h" > +#include "hw/irq.h" > +#include "trace.h" > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > + > +_Static_assert(__aarch64__, ""); I don't think this assertion is worthwhile. This assertion will trigger if you accidentally remove depends on AARCH64 from Kconfig, but I don't think such code change happens by accident, and there is no reason to believe that this assertion won't be removed in such a case. > + > +/* > + * ParavirtualizedGraphics.Framework only ships header files for the PCI > + * variant which does not include IOSFC descriptors and host devices. We add > + * their definitions here so that we can also work with the ARM version. > + */ > +typedef bool(^IOSFCRaiseInterrupt)(uint32_t vector); > +typedef bool(^IOSFCUnmapMemory)( > + void *a, void *b, void *c, void *d, void *e, void *f); Omit dummy parameter names. > +typedef bool(^IOSFCMapMemory)( > + uint64_t phys, uint64_t len, bool ro, void **va, void *e, void *f); > + > +@interface PGDeviceDescriptor (IOSurfaceMapper) > +@property (readwrite, nonatomic) bool usingIOSurfaceMapper; > +@end > + > +@interface PGIOSurfaceHostDeviceDescriptor : NSObject > +-(PGIOSurfaceHostDeviceDescriptor *)init; > +@property (readwrite, nonatomic, copy, nullable) IOSFCMapMemory mapMemory; > +@property (readwrite, nonatomic, copy, nullable) IOSFCUnmapMemory unmapMemory; > +@property (readwrite, nonatomic, copy, nullable) IOSFCRaiseInterrupt raiseInterrupt; > +@end > + > +@interface PGIOSurfaceHostDevice : NSObject > +-(instancetype)initWithDescriptor:(PGIOSurfaceHostDeviceDescriptor *) desc; > +-(uint32_t)mmioReadAtOffset:(size_t) offset; > +-(void)mmioWriteAtOffset:(size_t) offset value:(uint32_t)value; It has inconsistent spacing between types and parameter names. It is conventional to omit spaces between them as far as I remember. > +@end > + > +typedef struct AppleGFXVmappleState { > + SysBusDevice parent_obj; > + > + AppleGFXState common; > + > + qemu_irq irq_gfx; > + qemu_irq irq_iosfc; > + MemoryRegion iomem_iosfc; > + PGIOSurfaceHostDevice *pgiosfc; > +} AppleGFXVmappleState; > + > +OBJECT_DECLARE_SIMPLE_TYPE(AppleGFXVmappleState, APPLE_GFX_VMAPPLE) > + > + > +static uint64_t apple_iosfc_read(void *opaque, hwaddr offset, unsigned size) > +{ > + AppleGFXVmappleState *s = opaque; > + uint64_t res = 0; > + > + bql_unlock(); It is dangerous to unlock BQL at an arbitrary place. Instead of unlocking, I suggest: - running [s->pgiosfc mmioReadAtOffset:offset] on another thread - using a bottom half to request operations that require BQL from the thread running [s->pgiosfc mmioReadAtOffset:offset] - calling AIO_WAIT_WHILE() to process the bottom half and to wait for the completion of [s->pgiosfc mmioReadAtOffset:offset] > + res = [s->pgiosfc mmioReadAtOffset:offset]; > + bql_lock(); > + > + trace_apple_iosfc_read(offset, res); > + > + return res; > +} > + > +static void apple_iosfc_write( > + void *opaque, hwaddr offset, uint64_t val, unsigned size) > +{ > + AppleGFXVmappleState *s = opaque; > + > + trace_apple_iosfc_write(offset, val); > + > + [s->pgiosfc mmioWriteAtOffset:offset value:val]; > +} > + > +static const MemoryRegionOps apple_iosfc_ops = { > + .read = apple_iosfc_read, > + .write = apple_iosfc_write, > + .endianness = DEVICE_LITTLE_ENDIAN, > + .valid = { > + .min_access_size = 4, > + .max_access_size = 8, > + }, > + .impl = { > + .min_access_size = 4, > + .max_access_size = 8, > + }, > +}; > + > +static PGIOSurfaceHostDevice *apple_gfx_prepare_iosurface_host_device( > + AppleGFXVmappleState *s) > +{ > + PGIOSurfaceHostDeviceDescriptor *iosfc_desc = > + [PGIOSurfaceHostDeviceDescriptor new]; > + PGIOSurfaceHostDevice *iosfc_host_dev = nil; > + > + iosfc_desc.mapMemory = > + ^(uint64_t phys, uint64_t len, bool ro, void **va, void *e, void *f) { > + trace_apple_iosfc_map_memory(phys, len, ro, va, e, f); > + MemoryRegion *tmp_mr; > + *va = gpa2hva(&tmp_mr, phys, len, NULL); Use: dma_memory_map() > + return (bool)true; Why cast? > + }; > + > + iosfc_desc.unmapMemory = > + ^(void *a, void *b, void *c, void *d, void *e, void *f) { > + trace_apple_iosfc_unmap_memory(a, b, c, d, e, f); > + return (bool)true; > + }; > + > + iosfc_desc.raiseInterrupt = ^(uint32_t vector) { > + trace_apple_iosfc_raise_irq(vector); > + bool locked = bql_locked(); > + if (!locked) { > + bql_lock(); > + } > + qemu_irq_pulse(s->irq_iosfc);> + if (!locked) { > + bql_unlock(); > + } > + return (bool)true; > + }; > + > + iosfc_host_dev = > + [[PGIOSurfaceHostDevice alloc] initWithDescriptor:iosfc_desc]; > + [iosfc_desc release]; > + return iosfc_host_dev; > +} > + > +static void apple_gfx_vmapple_realize(DeviceState *dev, Error **errp) > +{ > + @autoreleasepool { This autoreleasepool is not used. > + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(dev); > + > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > + desc.usingIOSurfaceMapper = true; > + desc.raiseInterrupt = ^(uint32_t vector) { > + bool locked; > + > + trace_apple_gfx_raise_irq(vector); > + locked = bql_locked(); > + if (!locked) { > + bql_lock(); > + } > + qemu_irq_pulse(s->irq_gfx); > + if (!locked) { > + bql_unlock(); > + } > + }; > + > + s->pgiosfc = apple_gfx_prepare_iosurface_host_device(s); > + > + apple_gfx_common_realize(&s->common, desc); > + [desc release]; > + desc = nil; > + } > +} > + > +static void apple_gfx_vmapple_init(Object *obj) > +{ > + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(obj); > + > + apple_gfx_common_init(obj, &s->common, TYPE_APPLE_GFX_VMAPPLE); > + > + memory_region_init_io(&s->iomem_iosfc, obj, &apple_iosfc_ops, s, > + TYPE_APPLE_GFX_VMAPPLE, 0x10000); > + sysbus_init_mmio(SYS_BUS_DEVICE(s), &s->common.iomem_gfx); > + sysbus_init_mmio(SYS_BUS_DEVICE(s), &s->iomem_iosfc); > + sysbus_init_irq(SYS_BUS_DEVICE(s), &s->irq_gfx); > + sysbus_init_irq(SYS_BUS_DEVICE(s), &s->irq_iosfc); > +} > + > +static void apple_gfx_vmapple_reset(Object *obj, ResetType type) > +{ > + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(obj); > + [s->common.pgdev reset]; > +} > + > + > +static void apple_gfx_vmapple_class_init(ObjectClass *klass, void *data) > +{ > + DeviceClass *dc = DEVICE_CLASS(klass); > + ResettableClass *rc = RESETTABLE_CLASS(klass); > + > + assert(rc->phases.hold == NULL); Please remove this assertion. Having such an assertion for all types that inherit TYPE_SYS_BUS_DEVICE will make the code base too cluttered, and there is nothing special with TYPE_APPLE_GFX_VMAPPLE. > + rc->phases.hold = apple_gfx_vmapple_reset; > + > + dc->realize = apple_gfx_vmapple_realize; > +} > + > +static TypeInfo apple_gfx_vmapple_types[] = { > + { > + .name = TYPE_APPLE_GFX_VMAPPLE, > + .parent = TYPE_SYS_BUS_DEVICE, > + .instance_size = sizeof(AppleGFXVmappleState), > + .class_init = apple_gfx_vmapple_class_init, > + .instance_init = apple_gfx_vmapple_init, > + } > +}; > +DEFINE_TYPES(apple_gfx_vmapple_types) > diff --git a/hw/display/apple-gfx.h b/hw/display/apple-gfx.h > new file mode 100644 > index 00000000000..995ecf7f4a7 > --- /dev/null > +++ b/hw/display/apple-gfx.h > @@ -0,0 +1,57 @@ > +#ifndef QEMU_APPLE_GFX_H > +#define QEMU_APPLE_GFX_H > + > +#define TYPE_APPLE_GFX_VMAPPLE "apple-gfx-vmapple" > +#define TYPE_APPLE_GFX_PCI "apple-gfx-pci" > + > +#include "qemu/typedefs.h" > + > +typedef struct AppleGFXState AppleGFXState; > + > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const char* obj_name); > + > +#ifdef __OBJC__ This ifdef is unnecessary. > + > +#include "qemu/osdep.h" > +#include "exec/memory.h" > +#include "ui/surface.h" > +#include <dispatch/dispatch.h> > + > +@class PGDeviceDescriptor; > +@protocol PGDevice; > +@protocol PGDisplay; > +@protocol MTLDevice; > +@protocol MTLTexture; > +@protocol MTLCommandQueue; > + > +typedef QTAILQ_HEAD(, PGTask_s) AppleGFXTaskList; > + > +struct AppleGFXState { > + MemoryRegion iomem_gfx; > + id<PGDevice> pgdev; > + id<PGDisplay> pgdisp; > + AppleGFXTaskList tasks; > + QemuConsole *con; > + id<MTLDevice> mtl; > + id<MTLCommandQueue> mtl_queue; > + bool handles_frames; > + bool new_frame; > + bool cursor_show; > + QEMUCursor *cursor; > + > + dispatch_queue_t render_queue; > + /* The following fields should only be accessed from render_queue: */ > + bool gfx_update_requested; > + bool new_frame_ready; > + bool using_managed_texture_storage; > + int32_t pending_frames; > + void *vram; > + DisplaySurface *surface; > + id<MTLTexture> texture; > +}; > + > +void apple_gfx_common_realize(AppleGFXState *s, PGDeviceDescriptor *desc); > + > +#endif /* __OBJC__ */ > + > +#endif > diff --git a/hw/display/apple-gfx.m b/hw/display/apple-gfx.m > new file mode 100644 > index 00000000000..837300f9cd4 > --- /dev/null > +++ b/hw/display/apple-gfx.m > @@ -0,0 +1,536 @@ > +/* > + * QEMU Apple ParavirtualizedGraphics.framework device > + * > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All Rights Reserved. > + * > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > + * See the COPYING file in the top-level directory. > + * > + * ParavirtualizedGraphics.framework is a set of libraries that macOS provides > + * which implements 3d graphics passthrough to the host as well as a > + * proprietary guest communication channel to drive it. This device model > + * implements support to drive that library from within QEMU. > + */ > + > +#include "apple-gfx.h" > +#include "trace.h" > +#include "qemu/main-loop.h" > +#include "ui/console.h" > +#include "monitor/monitor.h" > +#include "qapi/error.h" > +#include "migration/blocker.h" > +#include <mach/mach_vm.h> > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > + > +static const PGDisplayCoord_t apple_gfx_modes[] = { > + { .x = 1440, .y = 1080 }, > + { .x = 1280, .y = 1024 }, > +}; > + > +typedef struct PGTask_s { // Name matches forward declaration in PG header Let's name it AppleGFXTask. It is a common practice to have the same tag name and typedef in QEMU. Please also avoid using // comment as suggested in: docs/devel/style.rst You can just remove the comment in this case as the usage of struct tag names in QTAILQ_HEAD() is very common. > + QTAILQ_ENTRY(PGTask_s) node; > + mach_vm_address_t address; > + uint64_t len; > +} AppleGFXTask; > + > +static Error *apple_gfx_mig_blocker; > + > +static void apple_gfx_render_frame_completed(AppleGFXState *s, void *vram, > + id<MTLTexture> texture); > + > +static AppleGFXTask *apple_gfx_new_task(AppleGFXState *s, uint64_t len) > +{ > + mach_vm_address_t task_mem; > + AppleGFXTask *task; > + kern_return_t r; > + > + r = mach_vm_allocate(mach_task_self(), &task_mem, len, VM_FLAGS_ANYWHERE); > + if (r != KERN_SUCCESS || task_mem == 0) { I think you can just check r. There is no reason to reject if the kernel insists it successfully allocated memory at 0 though that should never happen. > + return NULL; > + } > + > + task = g_new0(AppleGFXTask, 1); > + > + task->address = task_mem; > + task->len = len; > + QTAILQ_INSERT_TAIL(&s->tasks, task, node); > + > + return task; > +} > + > +static uint64_t apple_gfx_read(void *opaque, hwaddr offset, unsigned size) > +{ > + AppleGFXState *s = opaque; > + uint64_t res = 0; Don't initialize it with an arbitrary value as such an initialization suppresses an uninitialized variable warning. Please also keep naming consistent; you have used "r" for a similar variable previously. > + > + bql_unlock(); > + res = [s->pgdev mmioReadAtOffset:offset]; > + bql_lock(); > + > + trace_apple_gfx_read(offset, res); > + > + return res; > +} > + > +static void apple_gfx_write(void *opaque, hwaddr offset, uint64_t val, > + unsigned size) > +{ > + AppleGFXState *s = opaque; > + > + trace_apple_gfx_write(offset, val); > + > + bql_unlock(); > + [s->pgdev mmioWriteAtOffset:offset value:val]; > + bql_lock(); > +} > + > +static const MemoryRegionOps apple_gfx_ops = { > + .read = apple_gfx_read, > + .write = apple_gfx_write, > + .endianness = DEVICE_LITTLE_ENDIAN, > + .valid = { > + .min_access_size = 4, > + .max_access_size = 8, > + }, > + .impl = { > + .min_access_size = 4, > + .max_access_size = 4, > + }, > +}; > + > +static void apple_gfx_render_new_frame(AppleGFXState *s) > +{ > + BOOL r; > + void *vram = s->vram; > + uint32_t width = surface_width(s->surface); > + uint32_t height = surface_height(s->surface); > + MTLRegion region = MTLRegionMake2D(0, 0, width, height); > + id<MTLCommandBuffer> command_buffer = [s->mtl_queue commandBuffer]; > + id<MTLTexture> texture = s->texture; > + r = [s->pgdisp encodeCurrentFrameToCommandBuffer:command_buffer > + texture:texture > + region:region]; > + if (!r) { > + return; > + } > + [texture retain]; > + if (s->using_managed_texture_storage) { > + /* "Managed" textures exist in both VRAM and RAM and must be synced. */ > + id<MTLBlitCommandEncoder> blit = [command_buffer blitCommandEncoder]; > + [blit synchronizeResource:texture]; > + [blit endEncoding]; > + } > + [command_buffer retain]; I don't think this call of retain is necessary. The completion handler gets the command buffer via parameter, which implies the command buffer is automatically retained until the completion handler finishes. > + [command_buffer addCompletedHandler: > + ^(id<MTLCommandBuffer> cb) > + { > + dispatch_async(s->render_queue, ^{ > + apple_gfx_render_frame_completed(s, vram, texture); > + [texture release]; > + }); > + [command_buffer release]; > + }]; > + [command_buffer commit]; > +} > + > +static void copy_mtl_texture_to_surface_mem(id<MTLTexture> texture, void *vram) > +{ > + /* TODO: Skip this entirely on a pure Metal or headless/guest-only > + * rendering path, else use a blit command encoder? Needs careful > + * (double?) buffering design. */ > + size_t width = texture.width, height = texture.height; > + MTLRegion region = MTLRegionMake2D(0, 0, width, height); > + [texture getBytes:vram > + bytesPerRow:(width * 4) > + bytesPerImage:(width * height * 4) > + fromRegion:region > + mipmapLevel:0 > + slice:0]; > +} > + > +static void apple_gfx_render_frame_completed(AppleGFXState *s, void *vram, > + id<MTLTexture> texture) > +{ > + --s->pending_frames; > + assert(s->pending_frames >= 0); > + > + if (vram != s->vram) { > + /* Display mode has changed, drop this old frame. */ > + assert(texture != s->texture); > + g_free(vram); This management of buffers looks a bit convoluted. I suggest remembering the width and height instead of pointers and comparing them. This way you can free resources in set_mode(). > + } else { > + copy_mtl_texture_to_surface_mem(texture, vram); Writing vram outside BQL may result in tearing. > + if (s->gfx_update_requested) { > + s->gfx_update_requested = false; > + dpy_gfx_update_full(s->con); > + graphic_hw_update_done(s->con); > + s->new_frame_ready = false; This assignment is unnecessary as s->new_frame_ready is always false when s->gfx_update_requested. If you want to make sure s->gfx_update_requested and s->new_frame_ready are mutually exclusive, use one enum value instead of having two bools. > + } else { > + s->new_frame_ready = true; > + } > + } > + if (s->pending_frames > 0) { > + apple_gfx_render_new_frame(s); > + } > +} > + > +static void apple_gfx_fb_update_display(void *opaque) > +{ > + AppleGFXState *s = opaque; > + > + dispatch_async(s->render_queue, ^{ > + if (s->pending_frames > 0) { It should check for s->new_frame_ready as apple_gfx_render_frame_completed() doesn't check if s->pending_frames > 0 before calling graphic_hw_update_done(), which is inconsistent. Checking if s->pending_frames > 0 both in apple_gfx_fb_update_display() and apple_gfx_render_frame_completed() is also problematic as that can defer graphic_hw_update_done() indefinitely if we are getting new frames too fast. > + s->gfx_update_requested = true; > + } else { > + if (s->new_frame_ready) { > + dpy_gfx_update_full(s->con); > + s->new_frame_ready = false; > + } > + graphic_hw_update_done(s->con); > + }> + }); > +} > + > +static const GraphicHwOps apple_gfx_fb_ops = { > + .gfx_update = apple_gfx_fb_update_display, > + .gfx_update_async = true, > +}; > + > +static void update_cursor(AppleGFXState *s) > +{ > + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, > + s->pgdisp.cursorPosition.y, s->cursor_show); > +} > + > +static void set_mode(AppleGFXState *s, uint32_t width, uint32_t height) > +{ > + void *vram = NULL; > + DisplaySurface *surface; > + MTLTextureDescriptor *textureDescriptor; > + id<MTLTexture> texture = nil; > + __block bool no_change = false; > + > + dispatch_sync(s->render_queue, Calling dispatch_sync() while holding BQL may result in deadlock. > + ^{ > + if (s->surface && > + width == surface_width(s->surface) && > + height == surface_height(s->surface)) { > + no_change = true; > + } > + }); > + > + if (no_change) { > + return; > + } > + > + vram = g_malloc0(width * height * 4); > + surface = qemu_create_displaysurface_from(width, height, PIXMAN_LE_a8r8g8b8, > + width * 4, vram); > + > + @autoreleasepool { > + textureDescriptor = > + [MTLTextureDescriptor > + texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm > + width:width > + height:height > + mipmapped:NO]; > + textureDescriptor.usage = s->pgdisp.minimumTextureUsage; > + texture = [s->mtl newTextureWithDescriptor:textureDescriptor]; > + } > + > + s->using_managed_texture_storage = > + (texture.storageMode == MTLStorageModeManaged); > + > + dispatch_sync(s->render_queue, > + ^{ > + id<MTLTexture> old_texture = nil; > + void *old_vram = s->vram; > + s->vram = vram; > + s->surface = surface; > + > + dpy_gfx_replace_surface(s->con, surface); > + > + old_texture = s->texture; > + s->texture = texture; > + [old_texture release]; You can just do: [s->texture release]; s->texture = texture; This will make s->texture dangling between the two statements, but that don't matter since s->texture is not an atomic variable that can be safely observed from another thread anyway. > + > + if (s->pending_frames == 0) { > + g_free(old_vram); > + } > + }); > +} > + > +static void create_fb(AppleGFXState *s) > +{ > + s->con = graphic_console_init(NULL, 0, &apple_gfx_fb_ops, s); > + set_mode(s, 1440, 1080); > + > + s->cursor_show = true; > +} > + > +static size_t apple_gfx_get_default_mmio_range_size(void) > +{ > + size_t mmio_range_size; > + @autoreleasepool { > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > + mmio_range_size = desc.mmioLength; > + [desc release]; > + } > + return mmio_range_size; > +} > + > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const char* obj_name) This function can be merged into apple_gfx_common_realize(). > +{ > + Error *local_err = NULL; > + int r; > + size_t mmio_range_size = apple_gfx_get_default_mmio_range_size(); > + > + trace_apple_gfx_common_init(obj_name, mmio_range_size); > + memory_region_init_io(&s->iomem_gfx, obj, &apple_gfx_ops, s, obj_name, > + mmio_range_size); > + s->iomem_gfx.disable_reentrancy_guard = true; Why do you disable reentrancy guard? > + > + /* TODO: PVG framework supports serialising device state: integrate it! */ > + if (apple_gfx_mig_blocker == NULL) { > + error_setg(&apple_gfx_mig_blocker, > + "Migration state blocked by apple-gfx display device"); > + r = migrate_add_blocker(&apple_gfx_mig_blocker, &local_err); > + if (r < 0) { > + error_report_err(local_err); Please report the error to the caller of apple_gfx_common_realize() instead. > + } > + } > +}> + > +static void apple_gfx_register_task_mapping_handlers(AppleGFXState *s, > + PGDeviceDescriptor *desc) > +{ > + desc.createTask = ^(uint64_t vmSize, void * _Nullable * _Nonnull baseAddress) { > + AppleGFXTask *task = apple_gfx_new_task(s, vmSize); > + *baseAddress = (void*)task->address; nit: please write as (void *) instead of (void*). > + trace_apple_gfx_create_task(vmSize, *baseAddress); > + return task; > + }; > + > + desc.destroyTask = ^(AppleGFXTask * _Nonnull task) { > + trace_apple_gfx_destroy_task(task); > + QTAILQ_REMOVE(&s->tasks, task, node); > + mach_vm_deallocate(mach_task_self(), task->address, task->len); > + g_free(task); > + }; > + > + desc.mapMemory = ^(AppleGFXTask * _Nonnull task, uint32_t rangeCount, > + uint64_t virtualOffset, bool readOnly, > + PGPhysicalMemoryRange_t * _Nonnull ranges) { > + kern_return_t r; > + mach_vm_address_t target, source; > + trace_apple_gfx_map_memory(task, rangeCount, virtualOffset, readOnly); > + for (int i = 0; i < rangeCount; i++) { > + PGPhysicalMemoryRange_t *range = &ranges[i]; > + MemoryRegion *tmp_mr; > + /* TODO: Bounds checks? r/o? */ > + bql_lock(); > + > + trace_apple_gfx_map_memory_range(i, range->physicalAddress, > + range->physicalLength); > + > + target = task->address + virtualOffset; > + source = (mach_vm_address_t)gpa2hva(&tmp_mr, > + range->physicalAddress, > + range->physicalLength, NULL); > + vm_prot_t cur_protection = 0; > + vm_prot_t max_protection = 0; > + // Map guest RAM at range->physicalAddress into PG task memory range > + r = mach_vm_remap(mach_task_self(), > + &target, range->physicalLength, vm_page_size - 1, > + VM_FLAGS_FIXED | VM_FLAGS_OVERWRITE, > + mach_task_self(), > + source, false /* shared mapping, no copy */, > + &cur_protection, &max_protection, > + VM_INHERIT_COPY); > + trace_apple_gfx_remap(r, source, target); > + g_assert(r == KERN_SUCCESS); > + > + bql_unlock(); > + > + virtualOffset += range->physicalLength; > + } > + return (bool)true; > + }; > + > + desc.unmapMemory = ^(AppleGFXTask * _Nonnull task, uint64_t virtualOffset, > + uint64_t length) { > + kern_return_t r; > + mach_vm_address_t range_address; > + > + trace_apple_gfx_unmap_memory(task, virtualOffset, length); > + > + /* Replace task memory range with fresh pages, undoing the mapping > + * from guest RAM. */ > + range_address = task->address + virtualOffset; > + r = mach_vm_allocate(mach_task_self(), &range_address, length, > + VM_FLAGS_FIXED | VM_FLAGS_OVERWRITE); > + g_assert(r == KERN_SUCCESS); > + > + return (bool)true; > + }; > + > + desc.readMemory = ^(uint64_t physicalAddress, uint64_t length, > + void * _Nonnull dst) { > + trace_apple_gfx_read_memory(physicalAddress, length, dst); > + cpu_physical_memory_read(physicalAddress, dst, length); > + return (bool)true; > + }; > + > +} > + > +static PGDisplayDescriptor *apple_gfx_prepare_display_handlers(AppleGFXState *s) This name is misleading as it does a bit more than preparing handlers but sets properties like name and sizeInMillimeters. > +{ > + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor new]; > + > + disp_desc.name = @"QEMU display"; > + disp_desc.sizeInMillimeters = NSMakeSize(400., 300.); /* A 20" display */ > + disp_desc.queue = dispatch_get_main_queue(); > + disp_desc.newFrameEventHandler = ^(void) { > + trace_apple_gfx_new_frame(); > + dispatch_async(s->render_queue, ^{ > + /* Drop frames if we get too far ahead. */ > + if (s->pending_frames >= 2) > + return; > + ++s->pending_frames; > + if (s->pending_frames > 1) { > + return; > + } > + @autoreleasepool { > + apple_gfx_render_new_frame(s); > + } > + }); > + }; > + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t sizeInPixels, > + OSType pixelFormat) { > + trace_apple_gfx_mode_change(sizeInPixels.x, sizeInPixels.y); > + set_mode(s, sizeInPixels.x, sizeInPixels.y); > + }; > + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep *glyph, > + PGDisplayCoord_t hotSpot) { > + uint32_t bpp = glyph.bitsPerPixel; > + size_t width = glyph.pixelsWide; > + size_t height = glyph.pixelsHigh; > + size_t padding_bytes_per_row = glyph.bytesPerRow - width * 4; > + const uint8_t* px_data = glyph.bitmapData; > + > + trace_apple_gfx_cursor_set(bpp, width, height); > + > + if (s->cursor) { > + cursor_unref(s->cursor); > + s->cursor = NULL; > + } > + > + if (bpp == 32) { /* Shouldn't be anything else, but just to be safe...*/ > + s->cursor = cursor_alloc(width, height); > + s->cursor->hot_x = hotSpot.x; > + s->cursor->hot_y = hotSpot.y; > + > + uint32_t *dest_px = s->cursor->data; > + > + for (size_t y = 0; y < height; ++y) { > + for (size_t x = 0; x < width; ++x) { > + /* NSBitmapImageRep's red & blue channels are swapped > + * compared to QEMUCursor's. */ > + *dest_px = > + (px_data[0] << 16u) | > + (px_data[1] << 8u) | > + (px_data[2] << 0u) | > + (px_data[3] << 24u); > + ++dest_px; > + px_data += 4; > + } > + px_data += padding_bytes_per_row; > + } > + dpy_cursor_define(s->con, s->cursor); > + update_cursor(s); > + } > + }; > + disp_desc.cursorShowHandler = ^(BOOL show) { > + trace_apple_gfx_cursor_show(show); > + s->cursor_show = show; > + update_cursor(s); > + }; > + disp_desc.cursorMoveHandler = ^(void) { > + trace_apple_gfx_cursor_move(); > + update_cursor(s); > + }; > + > + return disp_desc; > +} > + > +static NSArray<PGDisplayMode*>* apple_gfx_prepare_display_mode_array(void) > +{ > + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; > + NSArray<PGDisplayMode*>* mode_array = nil; > + int i; > + > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > + modes[i] = > + [[PGDisplayMode alloc] initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; > + } > + > + mode_array = [NSArray arrayWithObjects:modes count:ARRAY_SIZE(apple_gfx_modes)]; > + > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > + [modes[i] release]; > + modes[i] = nil; > + } > + > + return mode_array; > +} > + > +static id<MTLDevice> copy_suitable_metal_device(void) > +{ > + id<MTLDevice> dev = nil; > + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); > + > + /* Prefer a unified memory GPU. Failing that, pick a non-removable GPU. */ > + for (size_t i = 0; i < devs.count; ++i) { > + if (devs[i].hasUnifiedMemory) { > + dev = devs[i]; > + break; > + } > + if (!devs[i].removable) { > + dev = devs[i]; > + } > + } > + > + if (dev != nil) { > + [dev retain]; > + } else { > + dev = MTLCreateSystemDefaultDevice(); > + } > + [devs release]; > + > + return dev; > +} > + > +void apple_gfx_common_realize(AppleGFXState *s, PGDeviceDescriptor *desc) > +{ > + PGDisplayDescriptor *disp_desc = nil; > + > + QTAILQ_INIT(&s->tasks); > + s->render_queue = dispatch_queue_create("apple-gfx.render", > + DISPATCH_QUEUE_SERIAL); > + s->mtl = copy_suitable_metal_device(); > + s->mtl_queue = [s->mtl newCommandQueue]; > + > + desc.device = s->mtl; > + > + apple_gfx_register_task_mapping_handlers(s, desc); > + > + s->pgdev = PGNewDeviceWithDescriptor(desc); > + > + disp_desc = apple_gfx_prepare_display_handlers(s); > + s->pgdisp = [s->pgdev newDisplayWithDescriptor:disp_desc > + port:0 serialNum:1234]; > + [disp_desc release]; > + s->pgdisp.modeList = apple_gfx_prepare_display_mode_array(); > + > + create_fb(s); > +} > diff --git a/hw/display/meson.build b/hw/display/meson.build > index 7db05eace97..70d855749c0 100644 > --- a/hw/display/meson.build > +++ b/hw/display/meson.build > @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', if_true: files('artist.c')) > > system_ss.add(when: 'CONFIG_ATI_VGA', if_true: [files('ati.c', 'ati_2d.c', 'ati_dbg.c'), pixman]) > > +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: [files('apple-gfx.m'), pvg, metal]) > +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: [files('apple-gfx-vmapple.m'), pvg, metal]) > > if config_all_devices.has_key('CONFIG_VIRTIO_GPU') > virtio_gpu_ss = ss.source_set() > diff --git a/hw/display/trace-events b/hw/display/trace-events > index 781f8a33203..1809a358e36 100644 > --- a/hw/display/trace-events > +++ b/hw/display/trace-events > @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) "dest_width : %u" > dm163_leds(int led, uint32_t value) "led %d: 0x%x" > dm163_channels(int channel, uint8_t value) "channel %d: 0x%x" > dm163_refresh_rate(uint32_t rr) "refresh rate %d" > + > +# apple-gfx.m > +apple_gfx_read(uint64_t offset, uint64_t res) "offset=0x%"PRIx64" res=0x%"PRIx64 > +apple_gfx_write(uint64_t offset, uint64_t val) "offset=0x%"PRIx64" val=0x%"PRIx64 > +apple_gfx_create_task(uint32_t vm_size, void *va) "vm_size=0x%x base_addr=%p" > +apple_gfx_destroy_task(void *task) "task=%p" > +apple_gfx_map_memory(void *task, uint32_t range_count, uint64_t virtual_offset, uint32_t read_only) "task=%p range_count=0x%x virtual_offset=0x%"PRIx64" read_only=%d" > +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, uint64_t phys_len) "[%d] phys_addr=0x%"PRIx64" phys_len=0x%"PRIx64 > +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t target) "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 > +apple_gfx_unmap_memory(void *task, uint64_t virtual_offset, uint64_t length) "task=%p virtual_offset=0x%"PRIx64" length=0x%"PRIx64 > +apple_gfx_read_memory(uint64_t phys_address, uint64_t length, void *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" > +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" > +apple_gfx_new_frame(void) "" > +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" y=%"PRId64 > +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, uint64_t height) "bpp=%d width=%"PRId64" height=0x%"PRId64 > +apple_gfx_cursor_show(uint32_t show) "show=%d" > +apple_gfx_cursor_move(void) "" > +apple_gfx_common_init(const char *device_name, size_t mmio_size) "device: %s; MMIO size: %zu bytes" > + > +# apple-gfx-vmapple.m > +apple_iosfc_read(uint64_t offset, uint64_t res) "offset=0x%"PRIx64" res=0x%"PRIx64 > +apple_iosfc_write(uint64_t offset, uint64_t val) "offset=0x%"PRIx64" val=0x%"PRIx64 > +apple_iosfc_map_memory(uint64_t phys, uint64_t len, uint32_t ro, void *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" ro=%d va=%p e=%p f=%p" > +apple_iosfc_unmap_memory(void *a, void *b, void *c, void *d, void *e, void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" > +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" > + > diff --git a/meson.build b/meson.build > index 10464466ff3..f09df3f09d5 100644 > --- a/meson.build > +++ b/meson.build > @@ -741,6 +741,8 @@ socket = [] > version_res = [] > coref = [] > iokit = [] > +pvg = [] > +metal = [] > emulator_link_args = [] > midl = not_found > widl = not_found > @@ -762,6 +764,8 @@ elif host_os == 'darwin' > coref = dependency('appleframeworks', modules: 'CoreFoundation') > iokit = dependency('appleframeworks', modules: 'IOKit', required: false) > host_dsosuf = '.dylib' > + pvg = dependency('appleframeworks', modules: 'ParavirtualizedGraphics') > + metal = dependency('appleframeworks', modules: 'Metal') > elif host_os == 'sunos' > socket = [cc.find_library('socket'), > cc.find_library('nsl'),
Hi, Thanks for taking a close look at this. There are some further comments, explanations, and also a few questions inline below. Where I've not commented, I'll just go ahead and make the suggested change for v4. On Tue, 1 Oct 2024 at 11:40, Akihiko Odaki <akihiko.odaki@daynix.com> wrote: > > > This patch implements a QEMU device that drives PVG for the VMApple > > variant of it. > > I think it is better to name it MMIO variant instead of VMApple. There > is nothing specific to VMApple in: hw/display/apple-gfx-vmapple.m > I mean, I don't see it being useful for anything whatsoever outside the vmapple machine type… But I guess there's little harm in renaming it. > > +#include "apple-gfx.h" > > +#include "monitor/monitor.h" > > +#include "hw/sysbus.h" > > +#include "hw/irq.h" > > +#include "trace.h" > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > + > > +_Static_assert(__aarch64__, ""); > > I don't think this assertion is worthwhile. This assertion will trigger > if you accidentally remove depends on AARCH64 from Kconfig, but I don't > think such code change happens by accident, and there is no reason to > believe that this assertion won't be removed in such a case. > As far as I'm aware the Kconfig AARCH64 dependency is for the *target* architecture, not the *host* architecture? The static assert checks for the latter. The PGIOSurfaceHostDeviceDescriptor type isn't available at all on non-aarch64 macOS hosts. I've not had any luck with using this variant of the device on x86-64 hosts simply by disabling any surface mapper code. Incidentally, if you know of a way to depend on a specific *host* architecture in the Kconfig, that would be even better. I couldn't spot a way of doing that though. > + > > +/* > > + * ParavirtualizedGraphics.Framework only ships header files for the PCI > > + * variant which does not include IOSFC descriptors and host devices. > We add > > + * their definitions here so that we can also work with the ARM version. > > + */ > > +typedef bool(^IOSFCRaiseInterrupt)(uint32_t vector); > > +typedef bool(^IOSFCUnmapMemory)( > > + void *a, void *b, void *c, void *d, void *e, void *f); > > Omit dummy parameter names. > > > +@end > > + > > +typedef struct AppleGFXVmappleState { > > + SysBusDevice parent_obj; > > + > > + AppleGFXState common; > > + > > + qemu_irq irq_gfx; > > + qemu_irq irq_iosfc; > > + MemoryRegion iomem_iosfc; > > + PGIOSurfaceHostDevice *pgiosfc; > > +} AppleGFXVmappleState; > > + > > +OBJECT_DECLARE_SIMPLE_TYPE(AppleGFXVmappleState, APPLE_GFX_VMAPPLE) > > + > > + > > +static uint64_t apple_iosfc_read(void *opaque, hwaddr offset, unsigned > size) > > +{ > > + AppleGFXVmappleState *s = opaque; > > + uint64_t res = 0; > > + > > + bql_unlock(); > > It is dangerous to unlock BQL at an arbitrary place. Instead of > unlocking, I suggest: > - running [s->pgiosfc mmioReadAtOffset:offset] on another thread > - using a bottom half to request operations that require BQL from the > thread running [s->pgiosfc mmioReadAtOffset:offset] > - calling AIO_WAIT_WHILE() to process the bottom half and to wait for > the completion of [s->pgiosfc mmioReadAtOffset:offset] > OK, I think I see what you mean, I'll try to rework things around that pattern. Any preference on how I kick off the job on the other thread? As we necessarily need to use libdispatch in a bunch of places in this code anyway, I guess dispatch_async() would probably be the simplest? > > + res = [s->pgiosfc mmioReadAtOffset:offset]; > > + bql_lock(); > > + > > + trace_apple_iosfc_read(offset, res); > > + > > + return res; > > +} > > + > > +static void apple_iosfc_write( > > + void *opaque, hwaddr offset, uint64_t val, unsigned size) > > +{ > > + AppleGFXVmappleState *s = opaque; > > + > > + trace_apple_iosfc_write(offset, val); > > + > > + [s->pgiosfc mmioWriteAtOffset:offset value:val]; > > +} > > + > > +static const MemoryRegionOps apple_iosfc_ops = { > > + .read = apple_iosfc_read, > > + .write = apple_iosfc_write, > > + .endianness = DEVICE_LITTLE_ENDIAN, > > + .valid = { > > + .min_access_size = 4, > > + .max_access_size = 8, > > + }, > > + .impl = { > > + .min_access_size = 4, > > + .max_access_size = 8, > > + }, > > +}; > > + > > +static PGIOSurfaceHostDevice *apple_gfx_prepare_iosurface_host_device( > > + AppleGFXVmappleState *s) > > +{ > > + PGIOSurfaceHostDeviceDescriptor *iosfc_desc = > > + [PGIOSurfaceHostDeviceDescriptor new]; > > + PGIOSurfaceHostDevice *iosfc_host_dev = nil; > > + > > + iosfc_desc.mapMemory = > > + ^(uint64_t phys, uint64_t len, bool ro, void **va, void *e, > void *f) { > > + trace_apple_iosfc_map_memory(phys, len, ro, va, e, f); > > + MemoryRegion *tmp_mr; > > + *va = gpa2hva(&tmp_mr, phys, len, NULL); > > Use: dma_memory_map() > > That doesn't seem to be a precisely equivalent operation. It also says in its headerdoc, > Use only for reads OR writes - not for read-modify-write operations. which I don't think we can guarantee here at all. I guess I can call it twice, once for writing and once for reading, but given that the dma_memory_unmap operation marks the area dirty, I'm not it's intended for what I understand the use case here to be: As far as I can tell, the PV graphics device uses (some) of this memory to exchange data in a cache-coherent way between host and guest, e.g. as a lock-free ring buffer, using atomic operations as necessary. (This works because it's a PV device: it "knows" the other end just another CPU core (or even the same one) executing in a different hypervisor context.) This doesn't really match "traditional" DMA patterns where there's either a read or a write happening. Hunting around some more for alternative APIs, there's also memory_region_get_ram_ptr(), but I'm not sure its restrictions apply here either. > + return (bool)true; > > Why cast? > Good question. Not originally my code, so I've fixed all the instances I could find now. > > + }; > > + > > + iosfc_desc.unmapMemory = > > + ^(void *a, void *b, void *c, void *d, void *e, void *f) { > > + trace_apple_iosfc_unmap_memory(a, b, c, d, e, f); > > + return (bool)true; > > + }; > > + > > + iosfc_desc.raiseInterrupt = ^(uint32_t vector) { > > + trace_apple_iosfc_raise_irq(vector); > > + bool locked = bql_locked(); > > + if (!locked) { > > + bql_lock(); > > + } > > + qemu_irq_pulse(s->irq_iosfc);> + if (!locked) { > > + bql_unlock(); > > + } > > + return (bool)true; > > + }; > > + > > + iosfc_host_dev = > > + [[PGIOSurfaceHostDevice alloc] initWithDescriptor:iosfc_desc]; > > + [iosfc_desc release]; > > + return iosfc_host_dev; > > +} > > + > > +static void apple_gfx_vmapple_realize(DeviceState *dev, Error **errp) > > +{ > > + @autoreleasepool { > > This autoreleasepool is not used. > It is definitely used inside the apple_gfx_common_realize() call. It's also impossible to say whether [PGDeviceDescriptor new] uses autorelease semantics internally, so it seemed safer to wrap the whole thing in an outer pool. > + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(dev); > > + > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > > + desc.usingIOSurfaceMapper = true; > > + desc.raiseInterrupt = ^(uint32_t vector) { > > + bool locked; > > + > > + trace_apple_gfx_raise_irq(vector); > > + locked = bql_locked(); > > + if (!locked) { > > + bql_lock(); > > + } > > + qemu_irq_pulse(s->irq_gfx); > > + if (!locked) { > > + bql_unlock(); > > + } > > + }; > > + > > + s->pgiosfc = apple_gfx_prepare_iosurface_host_device(s); > > + > > + apple_gfx_common_realize(&s->common, desc); > > + [desc release]; > > + desc = nil; > > + } > > +} > > + > > > > +++ b/hw/display/apple-gfx.h > > @@ -0,0 +1,57 @@ > > +#ifndef QEMU_APPLE_GFX_H > > +#define QEMU_APPLE_GFX_H > > + > > +#define TYPE_APPLE_GFX_VMAPPLE "apple-gfx-vmapple" > > +#define TYPE_APPLE_GFX_PCI "apple-gfx-pci" > > + > > +#include "qemu/typedefs.h" > > + > > +typedef struct AppleGFXState AppleGFXState; > > + > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const char* > obj_name); > > + > > +#ifdef __OBJC__ > > This ifdef is unnecessary. > Ah indeed; at one point, vmapple.c was #including this file, but that's no longer necessary. > > > diff --git a/hw/display/apple-gfx.m b/hw/display/apple-gfx.m > > new file mode 100644 > > index 00000000000..837300f9cd4 > > --- /dev/null > > +++ b/hw/display/apple-gfx.m > > @@ -0,0 +1,536 @@ > > +/* > > + * QEMU Apple ParavirtualizedGraphics.framework device > > + * > > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All Rights > Reserved. > > + * > > + * This work is licensed under the terms of the GNU GPL, version 2 or > later. > > + * See the COPYING file in the top-level directory. > > + * > > + * ParavirtualizedGraphics.framework is a set of libraries that macOS > provides > > + * which implements 3d graphics passthrough to the host as well as a > > + * proprietary guest communication channel to drive it. This device > model > > + * implements support to drive that library from within QEMU. > > + */ > > + > > +#include "apple-gfx.h" > > +#include "trace.h" > > +#include "qemu/main-loop.h" > > +#include "ui/console.h" > > +#include "monitor/monitor.h" > > +#include "qapi/error.h" > > +#include "migration/blocker.h" > > +#include <mach/mach_vm.h> > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > + > > +static const PGDisplayCoord_t apple_gfx_modes[] = { > > + { .x = 1440, .y = 1080 }, > > + { .x = 1280, .y = 1024 }, > > +}; > > + > > +typedef struct PGTask_s { // Name matches forward declaration in PG > header > > Let's name it AppleGFXTask. It is a common practice to have the same tag > name and typedef in QEMU. > This is defining a forward-declared type from framework headers which is opaque from the framework's point of view. We do not get to choose its struct name. The alternative is having casts wherever these objects are being passed between our code and the framework. (See the original v1/v2 vmapple patch series for how messy this is.) > > +static void apple_gfx_render_new_frame(AppleGFXState *s) > > +{ > > + BOOL r; > > + void *vram = s->vram; > > + uint32_t width = surface_width(s->surface); > > + uint32_t height = surface_height(s->surface); > > + MTLRegion region = MTLRegionMake2D(0, 0, width, height); > > + id<MTLCommandBuffer> command_buffer = [s->mtl_queue commandBuffer]; > > + id<MTLTexture> texture = s->texture; > > + r = [s->pgdisp encodeCurrentFrameToCommandBuffer:command_buffer > > + texture:texture > > + region:region]; > > + if (!r) { > > + return; > > + } > > + [texture retain]; > > + if (s->using_managed_texture_storage) { > > + /* "Managed" textures exist in both VRAM and RAM and must be > synced. */ > > + id<MTLBlitCommandEncoder> blit = [command_buffer > blitCommandEncoder]; > > + [blit synchronizeResource:texture]; > > + [blit endEncoding]; > > + } > > + [command_buffer retain]; > > I don't think this call of retain is necessary. The completion handler > gets the command buffer via parameter, which implies the command buffer > is automatically retained until the completion handler finishes. > I couldn't find any hard guarantees for this documented anywhere, hence erring on the side of caution. I think the harm of an additional retain/release pair here is quite minor. > > +static void apple_gfx_render_frame_completed(AppleGFXState *s, void > *vram, > > + id<MTLTexture> texture) > > +{ > > + --s->pending_frames; > > + assert(s->pending_frames >= 0); > > + > > + if (vram != s->vram) { > > + /* Display mode has changed, drop this old frame. */ > > + assert(texture != s->texture); > > + g_free(vram); > > This management of buffers looks a bit convoluted. I suggest remembering > the width and height instead of pointers and comparing them. This way > you can free resources in set_mode(). > Yeah, I suppose that works, I can change that around. > + } else { > > + copy_mtl_texture_to_surface_mem(texture, vram); > > Writing vram outside BQL may result in tearing. > As far as I can tell(*), QXL does the same. I couldn't find any examples of double-buffering in any of the existing display devices, which would be the only way to do async updates efficiently and without tearing. In any case, this solution is still vastly better than a regular VGA device, which suffers from very visible tearing with macOS on the guest side anyway. And in an ideal world, we'd pass through the rendered texture directly to the Cocoa UI code rather than copying out only for the CPU to draw it back into a window surface which is then passed to the GPU for host side rendering. But I felt this patch is already very, very large, and if anyone cares, we can fix imperfections in subsequent updates. (*)The rendering code in that device is also fairly complex, so I may be misreading it. > > + if (s->gfx_update_requested) { > > + s->gfx_update_requested = false; > > + dpy_gfx_update_full(s->con); > > + graphic_hw_update_done(s->con); > > + s->new_frame_ready = false; > > This assignment is unnecessary as s->new_frame_ready is always false > when s->gfx_update_requested. If you want to make sure > s->gfx_update_requested and s->new_frame_ready are mutually exclusive, > use one enum value instead of having two bools. > I'll need to refresh my memory and get back to you on this one, it's been so many months since I actively worked on this code. > > + } else { > > + s->new_frame_ready = true; > > + } > > + } > > + if (s->pending_frames > 0) { > > + apple_gfx_render_new_frame(s); > > + } > > +} > > + > > +static void apple_gfx_fb_update_display(void *opaque) > > +{ > > + AppleGFXState *s = opaque; > > + > > + dispatch_async(s->render_queue, ^{ > > + if (s->pending_frames > 0) { > > It should check for s->new_frame_ready as > apple_gfx_render_frame_completed() doesn't check if > s->pending_frames > 0 before calling graphic_hw_update_done(), which is > inconsistent. > pending_frames is about guest-side frames that are queued to be rendered by the host GPU. new_frame_ready being true indicates that the contents of the Qemu console surface has been updated with new frame data since the last gfx_update. gfx_update_requested indicates that gfx_update is currently awaiting an async completion (graphic_hw_update_done) but the surface has not received a new frame content, but the GPU is stily busy drawing one. apple_gfx_render_frame_completed is scheduled exactly once per pending frame, so pending_frames > 0 is an invariant there. (Hence the assert.) I don't think there is any inconsistency here, but I'll double check. It's possible that there's an easier way to express the state machine, and I'll take a look at that. > Checking if s->pending_frames > 0 both in apple_gfx_fb_update_display() > and apple_gfx_render_frame_completed() is also problematic as that can > defer graphic_hw_update_done() indefinitely if we are getting new frames > too fast. > I see what you mean about this part. I'll have to test it, but I guess we should reverse the priority, like this: if (s->new_frame_ready) { dpy_gfx_update_full(s->con); s->new_frame_ready = false; graphic_hw_update_done(s->con); } else if (s->pending_frames > 0) { s->gfx_update_requested = true; } else { graphic_hw_update_done(s->con); } 1. If we already have a frame, ready to be displayed, return it immediately. 2. If the guest has reported that it's completed a frame and the GPU is currently busy rendering it, defer graphic_hw_update_done until it's done. 3. If the guest reports no changes to its display, indicate this back to Qemu as a no-op display update graphic_hw_update_done() with no dpy_gfx_update* call. > > + s->gfx_update_requested = true; > > + } else { > > + if (s->new_frame_ready) { > > + dpy_gfx_update_full(s->con); > > + s->new_frame_ready = false; > > + } > > + graphic_hw_update_done(s->con); > > + }> + }); > > +} > > + > > +static const GraphicHwOps apple_gfx_fb_ops = { > > + .gfx_update = apple_gfx_fb_update_display, > > + .gfx_update_async = true, > > +}; > > + > > +static void update_cursor(AppleGFXState *s) > > +{ > > + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, > > + s->pgdisp.cursorPosition.y, s->cursor_show); > > +} > > + > > +static void set_mode(AppleGFXState *s, uint32_t width, uint32_t height) > > +{ > > + void *vram = NULL; > > + DisplaySurface *surface; > > + MTLTextureDescriptor *textureDescriptor; > > + id<MTLTexture> texture = nil; > > + __block bool no_change = false; > > + > > + dispatch_sync(s->render_queue, > > Calling dispatch_sync() while holding BQL may result in deadlock. > Only if any code executed on the same dispatch queue acquires the lock either directly or transitively. I believe I have ensure this is not done on the reqnder_queue, have you found anywhere this is the case? > > + ^{ > > + if (s->surface && > > + width == surface_width(s->surface) && > > + height == surface_height(s->surface)) { > > + no_change = true; > > + } > > + }); > > + > > + if (no_change) { > > + return; > > + } > > + > > + vram = g_malloc0(width * height * 4); > > + surface = qemu_create_displaysurface_from(width, height, > PIXMAN_LE_a8r8g8b8, > > + width * 4, vram); > > + > > + @autoreleasepool { > > + textureDescriptor = > > + [MTLTextureDescriptor > > + > texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm > > + width:width > > + height:height > > + mipmapped:NO]; > > + textureDescriptor.usage = s->pgdisp.minimumTextureUsage; > > + texture = [s->mtl newTextureWithDescriptor:textureDescriptor]; > > + } > > + > > + s->using_managed_texture_storage = > > + (texture.storageMode == MTLStorageModeManaged); > > + > > + dispatch_sync(s->render_queue, > > + ^{ > > + id<MTLTexture> old_texture = nil; > > + void *old_vram = s->vram; > > + s->vram = vram; > > + s->surface = surface; > > + > > + dpy_gfx_replace_surface(s->con, surface); > > + > > + old_texture = s->texture; > > + s->texture = texture; > > + [old_texture release]; > > You can just do: > [s->texture release]; > s->texture = texture; > > This will make s->texture dangling between the two statements, but that > don't matter since s->texture is not an atomic variable that can be > safely observed from another thread anyway. > > > + > > + if (s->pending_frames == 0) { > > + g_free(old_vram); > > + } > > + }); > > +} > > + > > +static void create_fb(AppleGFXState *s) > > +{ > > + s->con = graphic_console_init(NULL, 0, &apple_gfx_fb_ops, s); > > + set_mode(s, 1440, 1080); > > + > > + s->cursor_show = true; > > +} > > + > > +static size_t apple_gfx_get_default_mmio_range_size(void) > > +{ > > + size_t mmio_range_size; > > + @autoreleasepool { > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > > + mmio_range_size = desc.mmioLength; > > + [desc release]; > > + } > > + return mmio_range_size; > > +} > > + > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const char* > obj_name) > > This function can be merged into apple_gfx_common_realize(). > Probably. I'll try it. > > +{ > > + Error *local_err = NULL; > > + int r; > > + size_t mmio_range_size = apple_gfx_get_default_mmio_range_size(); > > + > > + trace_apple_gfx_common_init(obj_name, mmio_range_size); > > + memory_region_init_io(&s->iomem_gfx, obj, &apple_gfx_ops, s, > obj_name, > > + mmio_range_size); > > + s->iomem_gfx.disable_reentrancy_guard = true; > > Why do you disable reentrancy guard? > Perhaps with the proposed AIO_WAIT_WHILE based I/O scheme, this won't be an issue anymore, but the guard would otherwise keep dropping MMIOs which immediately caused the PV graphics device to stop making progress. The MMIO APIs in the PVG framework are thread- and reentrancy-safe, so we certainly don't need to serialise them on our side. > > + > > + /* TODO: PVG framework supports serialising device state: integrate > it! */ > > + if (apple_gfx_mig_blocker == NULL) { > > + error_setg(&apple_gfx_mig_blocker, > > + "Migration state blocked by apple-gfx display > device"); > > + r = migrate_add_blocker(&apple_gfx_mig_blocker, &local_err); > > + if (r < 0) { > > + error_report_err(local_err); > > Please report the error to the caller of apple_gfx_common_realize() > instead. > > > + } > > + } > > +}> + > > +static void apple_gfx_register_task_mapping_handlers(AppleGFXState *s, > > + PGDeviceDescriptor > *desc) > > +{ > > + desc.createTask = ^(uint64_t vmSize, void * _Nullable * _Nonnull > baseAddress) { > > + AppleGFXTask *task = apple_gfx_new_task(s, vmSize); > > + *baseAddress = (void*)task->address; > > nit: please write as (void *) instead of (void*). > > > + trace_apple_gfx_create_task(vmSize, *baseAddress); > > + return task; > > + }; > > + > > > +{ > > + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor new]; > > + > > + disp_desc.name = @"QEMU display"; > > + disp_desc.sizeInMillimeters = NSMakeSize(400., 300.); /* A 20" > display */ > > + disp_desc.queue = dispatch_get_main_queue(); > > + disp_desc.newFrameEventHandler = ^(void) { > > + trace_apple_gfx_new_frame(); > > + dispatch_async(s->render_queue, ^{ > > + /* Drop frames if we get too far ahead. */ > > + if (s->pending_frames >= 2) > > + return; > > + ++s->pending_frames; > > + if (s->pending_frames > 1) { > > + return; > > + } > > + @autoreleasepool { > > + apple_gfx_render_new_frame(s); > > + } > > + }); > > + }; > > + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t sizeInPixels, > > + OSType pixelFormat) { > > + trace_apple_gfx_mode_change(sizeInPixels.x, sizeInPixels.y); > > + set_mode(s, sizeInPixels.x, sizeInPixels.y); > > + }; > > + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep *glyph, > > + PGDisplayCoord_t hotSpot) { > > + uint32_t bpp = glyph.bitsPerPixel; > > + size_t width = glyph.pixelsWide; > > + size_t height = glyph.pixelsHigh; > > + size_t padding_bytes_per_row = glyph.bytesPerRow - width * 4; > > + const uint8_t* px_data = glyph.bitmapData; > > + > > + trace_apple_gfx_cursor_set(bpp, width, height); > > + > > + if (s->cursor) { > > + cursor_unref(s->cursor); > > + s->cursor = NULL; > > + } > > + > > + if (bpp == 32) { /* Shouldn't be anything else, but just to be > safe...*/ > > + s->cursor = cursor_alloc(width, height); > > + s->cursor->hot_x = hotSpot.x; > > + s->cursor->hot_y = hotSpot.y; > > + > > + uint32_t *dest_px = s->cursor->data; > > + > > + for (size_t y = 0; y < height; ++y) { > > + for (size_t x = 0; x < width; ++x) { > > + /* NSBitmapImageRep's red & blue channels are > swapped > > + * compared to QEMUCursor's. */ > > + *dest_px = > > + (px_data[0] << 16u) | > > + (px_data[1] << 8u) | > > + (px_data[2] << 0u) | > > + (px_data[3] << 24u); > > + ++dest_px; > > + px_data += 4; > > + } > > + px_data += padding_bytes_per_row; > > + } > > + dpy_cursor_define(s->con, s->cursor); > > + update_cursor(s); > > + } > > + }; > > + disp_desc.cursorShowHandler = ^(BOOL show) { > > + trace_apple_gfx_cursor_show(show); > > + s->cursor_show = show; > > + update_cursor(s); > > + }; > > + disp_desc.cursorMoveHandler = ^(void) { > > + trace_apple_gfx_cursor_move(); > > + update_cursor(s); > > + }; > > + > > + return disp_desc; > > +} > > + > > +static NSArray<PGDisplayMode*>* > apple_gfx_prepare_display_mode_array(void) > > +{ > > + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; > > + NSArray<PGDisplayMode*>* mode_array = nil; > > + int i; > > + > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > + modes[i] = > > + [[PGDisplayMode alloc] > initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; > > + } > > + > > + mode_array = [NSArray arrayWithObjects:modes > count:ARRAY_SIZE(apple_gfx_modes)]; > > + > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > + [modes[i] release]; > > + modes[i] = nil; > > + } > > + > > + return mode_array; > > +} > > + > > +static id<MTLDevice> copy_suitable_metal_device(void) > > +{ > > + id<MTLDevice> dev = nil; > > + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); > > + > > + /* Prefer a unified memory GPU. Failing that, pick a non-removable > GPU. */ > > + for (size_t i = 0; i < devs.count; ++i) { > > + if (devs[i].hasUnifiedMemory) { > > + dev = devs[i]; > > + break; > > + } > > + if (!devs[i].removable) { > > + dev = devs[i]; > > + } > > + } > > + > > + if (dev != nil) { > > + [dev retain]; > > + } else { > > + dev = MTLCreateSystemDefaultDevice(); > > + } > > + [devs release]; > > + > > + return dev; > > +} > > + > > +void apple_gfx_common_realize(AppleGFXState *s, PGDeviceDescriptor > *desc) > > +{ > > + PGDisplayDescriptor *disp_desc = nil; > > + > > + QTAILQ_INIT(&s->tasks); > > + s->render_queue = dispatch_queue_create("apple-gfx.render", > > + DISPATCH_QUEUE_SERIAL); > > + s->mtl = copy_suitable_metal_device(); > > + s->mtl_queue = [s->mtl newCommandQueue]; > > + > > + desc.device = s->mtl; > > + > > + apple_gfx_register_task_mapping_handlers(s, desc); > > + > > + s->pgdev = PGNewDeviceWithDescriptor(desc); > > + > > + disp_desc = apple_gfx_prepare_display_handlers(s); > > + s->pgdisp = [s->pgdev newDisplayWithDescriptor:disp_desc > > + port:0 serialNum:1234]; > > + [disp_desc release]; > > + s->pgdisp.modeList = apple_gfx_prepare_display_mode_array(); > > + > > + create_fb(s); > > +} > > diff --git a/hw/display/meson.build b/hw/display/meson.build > > index 7db05eace97..70d855749c0 100644 > > --- a/hw/display/meson.build > > +++ b/hw/display/meson.build > > @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', if_true: > files('artist.c')) > > > > system_ss.add(when: 'CONFIG_ATI_VGA', if_true: [files('ati.c', > 'ati_2d.c', 'ati_dbg.c'), pixman]) > > > > +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: > [files('apple-gfx.m'), pvg, metal]) > > +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: > [files('apple-gfx-vmapple.m'), pvg, metal]) > > > > if config_all_devices.has_key('CONFIG_VIRTIO_GPU') > > virtio_gpu_ss = ss.source_set() > > diff --git a/hw/display/trace-events b/hw/display/trace-events > > index 781f8a33203..1809a358e36 100644 > > --- a/hw/display/trace-events > > +++ b/hw/display/trace-events > > @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) "dest_width : > %u" > > dm163_leds(int led, uint32_t value) "led %d: 0x%x" > > dm163_channels(int channel, uint8_t value) "channel %d: 0x%x" > > dm163_refresh_rate(uint32_t rr) "refresh rate %d" > > + > > +# apple-gfx.m > > +apple_gfx_read(uint64_t offset, uint64_t res) "offset=0x%"PRIx64" > res=0x%"PRIx64 > > +apple_gfx_write(uint64_t offset, uint64_t val) "offset=0x%"PRIx64" > val=0x%"PRIx64 > > +apple_gfx_create_task(uint32_t vm_size, void *va) "vm_size=0x%x > base_addr=%p" > > +apple_gfx_destroy_task(void *task) "task=%p" > > +apple_gfx_map_memory(void *task, uint32_t range_count, uint64_t > virtual_offset, uint32_t read_only) "task=%p range_count=0x%x > virtual_offset=0x%"PRIx64" read_only=%d" > > +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, uint64_t > phys_len) "[%d] phys_addr=0x%"PRIx64" phys_len=0x%"PRIx64 > > +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t target) > "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 > > +apple_gfx_unmap_memory(void *task, uint64_t virtual_offset, uint64_t > length) "task=%p virtual_offset=0x%"PRIx64" length=0x%"PRIx64 > > +apple_gfx_read_memory(uint64_t phys_address, uint64_t length, void > *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" > > +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" > > +apple_gfx_new_frame(void) "" > > +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" y=%"PRId64 > > +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, uint64_t height) > "bpp=%d width=%"PRId64" height=0x%"PRId64 > > +apple_gfx_cursor_show(uint32_t show) "show=%d" > > +apple_gfx_cursor_move(void) "" > > +apple_gfx_common_init(const char *device_name, size_t mmio_size) > "device: %s; MMIO size: %zu bytes" > > + > > +# apple-gfx-vmapple.m > > +apple_iosfc_read(uint64_t offset, uint64_t res) "offset=0x%"PRIx64" > res=0x%"PRIx64 > > +apple_iosfc_write(uint64_t offset, uint64_t val) "offset=0x%"PRIx64" > val=0x%"PRIx64 > > +apple_iosfc_map_memory(uint64_t phys, uint64_t len, uint32_t ro, void > *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" ro=%d va=%p e=%p > f=%p" > > +apple_iosfc_unmap_memory(void *a, void *b, void *c, void *d, void *e, > void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" > > +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" > > + > > diff --git a/meson.build b/meson.build > > index 10464466ff3..f09df3f09d5 100644 > > --- a/meson.build > > +++ b/meson.build > > @@ -741,6 +741,8 @@ socket = [] > > version_res = [] > > coref = [] > > iokit = [] > > +pvg = [] > > +metal = [] > > emulator_link_args = [] > > midl = not_found > > widl = not_found > > @@ -762,6 +764,8 @@ elif host_os == 'darwin' > > coref = dependency('appleframeworks', modules: 'CoreFoundation') > > iokit = dependency('appleframeworks', modules: 'IOKit', required: > false) > > host_dsosuf = '.dylib' > > + pvg = dependency('appleframeworks', modules: > 'ParavirtualizedGraphics') > > + metal = dependency('appleframeworks', modules: 'Metal') > > elif host_os == 'sunos' > > socket = [cc.find_library('socket'), > > cc.find_library('nsl'), > >
On 2024/10/02 22:33, Phil Dennis-Jordan wrote: > Hi, > > Thanks for taking a close look at this. > > There are some further comments, explanations, and also a few questions > inline below. Where I've not commented, I'll just go ahead and make the > suggested change for v4. > > On Tue, 1 Oct 2024 at 11:40, Akihiko Odaki <akihiko.odaki@daynix.com > <mailto:akihiko.odaki@daynix.com>> wrote: > > > > This patch implements a QEMU device that drives PVG for the VMApple > > variant of it. > > I think it is better to name it MMIO variant instead of VMApple. There > is nothing specific to VMApple in: hw/display/apple-gfx-vmapple.m > > > I mean, I don't see it being useful for anything whatsoever outside the > vmapple machine type… But I guess there's little harm in renaming it. I don't know but perhaps we may get vmapple2 or something in the future. > > > > +#include "apple-gfx.h" > > +#include "monitor/monitor.h" > > +#include "hw/sysbus.h" > > +#include "hw/irq.h" > > +#include "trace.h" > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > + > > +_Static_assert(__aarch64__, ""); > > I don't think this assertion is worthwhile. This assertion will trigger > if you accidentally remove depends on AARCH64 from Kconfig, but I don't > think such code change happens by accident, and there is no reason to > believe that this assertion won't be removed in such a case. > > > As far as I'm aware the Kconfig AARCH64 dependency is for the /target/ > architecture, not the /host/ architecture? The static assert checks for > the latter. The PGIOSurfaceHostDeviceDescriptor type isn't available at > all on non-aarch64 macOS hosts. I've not had any luck with using this > variant of the device on x86-64 hosts simply by disabling any surface > mapper code. > > Incidentally, if you know of a way to depend on a specific /host/ > architecture in the Kconfig, that would be even better. I couldn't spot > a way of doing that though. I got your intention now. The correct way to do that is to check for cpu == 'aarch64'. Having assertion will break qemu-system-aarch64 on Intel Macs. > > > + > > +/* > > + * ParavirtualizedGraphics.Framework only ships header files for > the PCI > > + * variant which does not include IOSFC descriptors and host > devices. We add > > + * their definitions here so that we can also work with the ARM > version. > > + */ > > +typedef bool(^IOSFCRaiseInterrupt)(uint32_t vector); > > +typedef bool(^IOSFCUnmapMemory)( > > + void *a, void *b, void *c, void *d, void *e, void *f); > > Omit dummy parameter names. > > > > +@end > > + > > +typedef struct AppleGFXVmappleState { > > + SysBusDevice parent_obj; > > + > > + AppleGFXState common; > > + > > + qemu_irq irq_gfx; > > + qemu_irq irq_iosfc; > > + MemoryRegion iomem_iosfc; > > + PGIOSurfaceHostDevice *pgiosfc; > > +} AppleGFXVmappleState; > > + > > +OBJECT_DECLARE_SIMPLE_TYPE(AppleGFXVmappleState, APPLE_GFX_VMAPPLE) > > + > > + > > +static uint64_t apple_iosfc_read(void *opaque, hwaddr offset, > unsigned size) > > +{ > > + AppleGFXVmappleState *s = opaque; > > + uint64_t res = 0; > > + > > + bql_unlock(); > > It is dangerous to unlock BQL at an arbitrary place. Instead of > unlocking, I suggest: > - running [s->pgiosfc mmioReadAtOffset:offset] on another thread > - using a bottom half to request operations that require BQL from the > thread running [s->pgiosfc mmioReadAtOffset:offset] > - calling AIO_WAIT_WHILE() to process the bottom half and to wait for > the completion of [s->pgiosfc mmioReadAtOffset:offset] > > > OK, I think I see what you mean, I'll try to rework things around that > pattern. Any preference on how I kick off the job on the other thread? > As we necessarily need to use libdispatch in a bunch of places in this > code anyway, I guess dispatch_async() would probably be the simplest? Perhaps so. The QEMU way is to use a bottom half with AioContext, but you can't simultaneously run a dispatch queue and AioContext in one thread so you have to use the dispatch queue if you need one. > > > + res = [s->pgiosfc mmioReadAtOffset:offset]; > > + bql_lock(); > > + > > + trace_apple_iosfc_read(offset, res); > > + > > + return res; > > +} > > + > > +static void apple_iosfc_write( > > + void *opaque, hwaddr offset, uint64_t val, unsigned size) > > +{ > > + AppleGFXVmappleState *s = opaque; > > + > > + trace_apple_iosfc_write(offset, val); > > + > > + [s->pgiosfc mmioWriteAtOffset:offset value:val]; > > +} > > + > > +static const MemoryRegionOps apple_iosfc_ops = { > > + .read = apple_iosfc_read, > > + .write = apple_iosfc_write, > > + .endianness = DEVICE_LITTLE_ENDIAN, > > + .valid = { > > + .min_access_size = 4, > > + .max_access_size = 8, > > + }, > > + .impl = { > > + .min_access_size = 4, > > + .max_access_size = 8, > > + }, > > +}; > > + > > +static PGIOSurfaceHostDevice > *apple_gfx_prepare_iosurface_host_device( > > + AppleGFXVmappleState *s) > > +{ > > + PGIOSurfaceHostDeviceDescriptor *iosfc_desc = > > + [PGIOSurfaceHostDeviceDescriptor new]; > > + PGIOSurfaceHostDevice *iosfc_host_dev = nil; > > + > > + iosfc_desc.mapMemory = > > + ^(uint64_t phys, uint64_t len, bool ro, void **va, void > *e, void *f) { > > + trace_apple_iosfc_map_memory(phys, len, ro, va, e, f); > > + MemoryRegion *tmp_mr; > > + *va = gpa2hva(&tmp_mr, phys, len, NULL); > > Use: dma_memory_map() > > > That doesn't seem to be a precisely equivalent operation. It also says > in its headerdoc, > > Use only for reads OR writes - not for read-modify-write operations. > > which I don't think we can guarantee here at all. > > I guess I can call it twice, once for writing and once for reading, but > given that the dma_memory_unmap operation marks the area dirty, I'm not > it's intended for what I understand the use case here to be: As far as I > can tell, the PV graphics device uses (some) of this memory to exchange > data in a cache-coherent way between host and guest, e.g. as a lock-free > ring buffer, using atomic operations as necessary. (This works because > it's a PV device: it "knows" the other end just another CPU core (or > even the same one) executing in a different hypervisor context.) This > doesn't really match "traditional" DMA patterns where there's either a > read or a write happening. I think the story is a bit different for this VMApple variant. Probably the CPU and GPU in Apple Silicon is cache-coherent so you can map normal memory for GPU without any kind of cache maintenance. Cache conherency of CPU and GPU in Apple Silicon is implied with Apple's documentation; it says you don't need to synchronize resources for MTLStorageModeShared, which is the default for Apple Silicon. https://developer.apple.com/documentation/metal/resource_fundamentals/synchronizing_a_managed_resource_in_macos The name "IOSurface" also implies it is used not only for e.g., ring buffer but also for real data. > > Hunting around some more for alternative APIs, there's also > memory_region_get_ram_ptr(), but I'm not sure its restrictions apply > here either. I think you can call memory_access_is_direct() to check if the requirement is satisfied. It will still break dirty page tracking implemented by dma_memory_unmap() and others, but it's broken for hvf, which does not implement dirty page tracking either. > > > + return (bool)true; > > Why cast? > > > Good question. Not originally my code, so I've fixed all the instances I > could find now. > > > + }; > > + > > + iosfc_desc.unmapMemory = > > + ^(void *a, void *b, void *c, void *d, void *e, void *f) { > > + trace_apple_iosfc_unmap_memory(a, b, c, d, e, f); > > + return (bool)true; > > + }; > > + > > + iosfc_desc.raiseInterrupt = ^(uint32_t vector) { > > + trace_apple_iosfc_raise_irq(vector); > > + bool locked = bql_locked(); > > + if (!locked) { > > + bql_lock(); > > + } > > + qemu_irq_pulse(s->irq_iosfc);> + if (!locked) { > > + bql_unlock(); > > + } > > + return (bool)true; > > + }; > > + > > + iosfc_host_dev = > > + [[PGIOSurfaceHostDevice alloc] > initWithDescriptor:iosfc_desc]; > > + [iosfc_desc release]; > > + return iosfc_host_dev; > > +} > > + > > +static void apple_gfx_vmapple_realize(DeviceState *dev, Error > **errp) > > +{ > > + @autoreleasepool { > > This autoreleasepool is not used. > > > It is definitely used inside the apple_gfx_common_realize() call. It's > also impossible to say whether [PGDeviceDescriptor new] uses autorelease > semantics internally, so it seemed safer to wrap the whole thing in an > outer pool. Theoretically, It should be safe to assume the callee creates autoreleasepool by themselves as needed in general. We have bunch of code to call Objective-C APIs without creating autoreleasepool in the caller. Practically, [PGDeviceDescriptor new] is likely to be implemented with ARC, which wraps methods in autoreleasepool as necessary. Functions that uses a method that returns autorelease resources should be wrapped with autoreleasepool instead of assuming the caller creates autoreleasepool for them. > > > + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(dev); > > + > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > > + desc.usingIOSurfaceMapper = true; > > + desc.raiseInterrupt = ^(uint32_t vector) { > > + bool locked; > > + > > + trace_apple_gfx_raise_irq(vector); > > + locked = bql_locked(); > > + if (!locked) { > > + bql_lock(); > > + } > > + qemu_irq_pulse(s->irq_gfx); > > + if (!locked) { > > + bql_unlock(); > > + } > > + }; > > + > > + s->pgiosfc = apple_gfx_prepare_iosurface_host_device(s); > > + > > + apple_gfx_common_realize(&s->common, desc); > > + [desc release]; > > + desc = nil; > > + } > > +} > > + > > > +++ b/hw/display/apple-gfx.h > > @@ -0,0 +1,57 @@ > > +#ifndef QEMU_APPLE_GFX_H > > +#define QEMU_APPLE_GFX_H > > + > > +#define TYPE_APPLE_GFX_VMAPPLE "apple-gfx-vmapple" > > +#define TYPE_APPLE_GFX_PCI "apple-gfx-pci" > > + > > +#include "qemu/typedefs.h" > > + > > +typedef struct AppleGFXState AppleGFXState; > > + > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const > char* obj_name); > > + > > +#ifdef __OBJC__ > > This ifdef is unnecessary. > > Ah indeed; at one point, vmapple.c was #including this file, but that's > no longer necessary. > > > > diff --git a/hw/display/apple-gfx.m b/hw/display/apple-gfx.m > > new file mode 100644 > > index 00000000000..837300f9cd4 > > --- /dev/null > > +++ b/hw/display/apple-gfx.m > > @@ -0,0 +1,536 @@ > > +/* > > + * QEMU Apple ParavirtualizedGraphics.framework device > > + * > > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All > Rights Reserved. > > + * > > + * This work is licensed under the terms of the GNU GPL, version > 2 or later. > > + * See the COPYING file in the top-level directory. > > + * > > + * ParavirtualizedGraphics.framework is a set of libraries that > macOS provides > > + * which implements 3d graphics passthrough to the host as well as a > > + * proprietary guest communication channel to drive it. This > device model > > + * implements support to drive that library from within QEMU. > > + */ > > + > > +#include "apple-gfx.h" > > +#include "trace.h" > > +#include "qemu/main-loop.h" > > +#include "ui/console.h" > > +#include "monitor/monitor.h" > > +#include "qapi/error.h" > > +#include "migration/blocker.h" > > +#include <mach/mach_vm.h> > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > + > > +static const PGDisplayCoord_t apple_gfx_modes[] = { > > + { .x = 1440, .y = 1080 }, > > + { .x = 1280, .y = 1024 }, > > +}; > > + > > +typedef struct PGTask_s { // Name matches forward declaration in > PG header > > Let's name it AppleGFXTask. It is a common practice to have the same > tag > name and typedef in QEMU. > > > This is defining a forward-declared type from framework headers which is > opaque from the framework's point of view. We do not get to choose its > struct name. The alternative is having casts wherever these objects are > being passed between our code and the framework. (See the original v1/v2 > vmapple patch series for how messy this is.) I got the idea. Let's not avoid the typedef then to clarify the naming is not under our control. > > > +static void apple_gfx_render_new_frame(AppleGFXState *s) > > +{ > > + BOOL r; > > + void *vram = s->vram; > > + uint32_t width = surface_width(s->surface); > > + uint32_t height = surface_height(s->surface); > > + MTLRegion region = MTLRegionMake2D(0, 0, width, height); > > + id<MTLCommandBuffer> command_buffer = [s->mtl_queue > commandBuffer]; > > + id<MTLTexture> texture = s->texture; > > + r = [s->pgdisp encodeCurrentFrameToCommandBuffer:command_buffer > > + texture:texture > > + region:region]; > > + if (!r) { > > + return; > > + } > > + [texture retain]; > > + if (s->using_managed_texture_storage) { > > + /* "Managed" textures exist in both VRAM and RAM and > must be synced. */ > > + id<MTLBlitCommandEncoder> blit = [command_buffer > blitCommandEncoder]; > > + [blit synchronizeResource:texture]; > > + [blit endEncoding]; > > + } > > + [command_buffer retain]; > > I don't think this call of retain is necessary. The completion handler > gets the command buffer via parameter, which implies the command buffer > is automatically retained until the completion handler finishes. > > > I couldn't find any hard guarantees for this documented anywhere, hence > erring on the side of caution. I think the harm of an additional retain/ > release pair here is quite minor. Apple's sample code doesn't retain for such a case so we can certainly omit it: https://developer.apple.com/documentation/metal/indirect_command_encoding/encoding_indirect_command_buffers_on_the_cpu It's better to avoid additional resource management when the underlying API automatically does things correct. There is a hazard of forgetting release anytime adding one call of retain. > > > > +static void apple_gfx_render_frame_completed(AppleGFXState *s, > void *vram, > > + id<MTLTexture> texture) > > +{ > > + --s->pending_frames; > > + assert(s->pending_frames >= 0); > > + > > + if (vram != s->vram) { > > + /* Display mode has changed, drop this old frame. */ > > + assert(texture != s->texture); > > + g_free(vram); > > This management of buffers looks a bit convoluted. I suggest > remembering > the width and height instead of pointers and comparing them. This way > you can free resources in set_mode(). > > > Yeah, I suppose that works, I can change that around. > > > + } else { > > + copy_mtl_texture_to_surface_mem(texture, vram); > > Writing vram outside BQL may result in tearing. > > > As far as I can tell(*), QXL does the same. I couldn't find any examples > of double-buffering in any of the existing display devices, which would > be the only way to do async updates efficiently and without tearing. In > any case, this solution is still vastly better than a regular VGA > device, which suffers from very visible tearing with macOS on the guest > side anyway. And in an ideal world, we'd pass through the rendered > texture directly to the Cocoa UI code rather than copying out only for > the CPU to draw it back into a window surface which is then passed to > the GPU for host side rendering. But I felt this patch is already very, > very large, and if anyone cares, we can fix imperfections in subsequent > updates. > > (*)The rendering code in that device is also fairly complex, so I may be > misreading it. QXL always modifies the surface with BQL. The surface is modified with qxl_blit(), which is called by qxl_render_update_area_unlocked(). qxl_render_update_area_unlocked() is called by either of qxl_render_update() and qxl_render_update_area_bh(). Both of them are called with BQL. The name includes "unlocked", but it means it is called without holding QXL-internal lock. Most devices works entirely with BQL so they don't perform double buffering. apple-gfx can do the same. > > > + if (s->gfx_update_requested) { > > + s->gfx_update_requested = false; > > + dpy_gfx_update_full(s->con); > > + graphic_hw_update_done(s->con); > > + s->new_frame_ready = false; > > This assignment is unnecessary as s->new_frame_ready is always false > when s->gfx_update_requested. If you want to make sure > s->gfx_update_requested and s->new_frame_ready are mutually exclusive, > use one enum value instead of having two bools. > > > I'll need to refresh my memory and get back to you on this one, it's > been so many months since I actively worked on this code. > > > + } else { > > + s->new_frame_ready = true; > > + } > > + } > > + if (s->pending_frames > 0) { > > + apple_gfx_render_new_frame(s); > > + } > > +} > > + > > +static void apple_gfx_fb_update_display(void *opaque) > > +{ > > + AppleGFXState *s = opaque; > > + > > + dispatch_async(s->render_queue, ^{ > > + if (s->pending_frames > 0) { > > It should check for s->new_frame_ready as > apple_gfx_render_frame_completed() doesn't check if > s->pending_frames > 0 before calling graphic_hw_update_done(), which is > inconsistent. > > > pending_frames is about guest-side frames that are queued to be rendered > by the host GPU. > new_frame_ready being true indicates that the contents of the Qemu > console surface has been updated with new frame data since the last > gfx_update. > gfx_update_requested indicates that gfx_update is currently awaiting an > async completion (graphic_hw_update_done) but the surface has not > received a new frame content, but the GPU is stily busy drawing one. > > apple_gfx_render_frame_completed is scheduled exactly once per pending > frame, so pending_frames > 0 is an invariant there. (Hence the assert.)> > I don't think there is any inconsistency here, but I'll double check. > It's possible that there's an easier way to express the state machine, > and I'll take a look at that. I meant that apple_gfx_render_frame_completed() does not check if the frame is the last one currently pending. apple_gfx_fb_update_display() ignores a new ready frame when there is a more pending frame, but apple_gfx_render_frame_completed() unconditionally fires graphic_hw_update_done() even if there is a more pending frame. And I think apple_gfx_render_frame_completed() is right and apple_gfx_fb_update_display() is wrong in such a situation. > > Checking if s->pending_frames > 0 both in apple_gfx_fb_update_display() > and apple_gfx_render_frame_completed() is also problematic as that can > defer graphic_hw_update_done() indefinitely if we are getting new > frames > too fast. > > > I see what you mean about this part. I'll have to test it, but I guess > we should reverse the priority, like this: > > if (s->new_frame_ready) { > dpy_gfx_update_full(s->con); > s->new_frame_ready = false; > graphic_hw_update_done(s->con); > } else if (s->pending_frames > 0) { > s->gfx_update_requested = true; > } else { > graphic_hw_update_done(s->con); > } > > 1. If we already have a frame, ready to be displayed, return it immediately. > 2. If the guest has reported that it's completed a frame and the GPU is > currently busy rendering it, defer graphic_hw_update_done until it's done. > 3. If the guest reports no changes to its display, indicate this back to > Qemu as a no-op display update graphic_hw_update_done() with no > dpy_gfx_update* call. Yes, that looks correct. > > > + s->gfx_update_requested = true; > > + } else { > > + if (s->new_frame_ready) { > > + dpy_gfx_update_full(s->con); > > + s->new_frame_ready = false; > > + } > > + graphic_hw_update_done(s->con); > > + }> + }); > > +} > > + > > +static const GraphicHwOps apple_gfx_fb_ops = { > > + .gfx_update = apple_gfx_fb_update_display, > > + .gfx_update_async = true, > > +}; > > + > > +static void update_cursor(AppleGFXState *s) > > +{ > > + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, > > + s->pgdisp.cursorPosition.y, s->cursor_show); > > +} > > + > > +static void set_mode(AppleGFXState *s, uint32_t width, uint32_t > height) > > +{ > > + void *vram = NULL; > > + DisplaySurface *surface; > > + MTLTextureDescriptor *textureDescriptor; > > + id<MTLTexture> texture = nil; > > + __block bool no_change = false; > > + > > + dispatch_sync(s->render_queue, > > Calling dispatch_sync() while holding BQL may result in deadlock. > > Only if any code executed on the same dispatch queue acquires the lock > either directly or transitively. I believe I have ensure this is not > done on the reqnder_queue, have you found anywhere this is the case? The documentation is not clear what threads a dispatch queue runs on. We can have a deadlock if they lock the BQL. > > > + ^{ > > + if (s->surface && > > + width == surface_width(s->surface) && > > + height == surface_height(s->surface)) { > > + no_change = true; > > + } > > + }); > > + > > + if (no_change) { > > + return; > > + } > > + > > + vram = g_malloc0(width * height * 4); > > + surface = qemu_create_displaysurface_from(width, height, > PIXMAN_LE_a8r8g8b8, > > + width * 4, vram); > > + > > + @autoreleasepool { > > + textureDescriptor = > > + [MTLTextureDescriptor > > + > texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm > > + width:width > > + height:height > > + mipmapped:NO]; > > + textureDescriptor.usage = s->pgdisp.minimumTextureUsage; > > + texture = [s->mtl > newTextureWithDescriptor:textureDescriptor]; > > + } > > + > > + s->using_managed_texture_storage = > > + (texture.storageMode == MTLStorageModeManaged); > > + > > + dispatch_sync(s->render_queue, > > + ^{ > > + id<MTLTexture> old_texture = nil; > > + void *old_vram = s->vram; > > + s->vram = vram; > > + s->surface = surface; > > + > > + dpy_gfx_replace_surface(s->con, surface); > > + > > + old_texture = s->texture; > > + s->texture = texture; > > + [old_texture release]; > > You can just do: > [s->texture release]; > s->texture = texture; > > This will make s->texture dangling between the two statements, but that > don't matter since s->texture is not an atomic variable that can be > safely observed from another thread anyway. > > > + > > + if (s->pending_frames == 0) { > > + g_free(old_vram); > > + } > > + }); > > +} > > + > > +static void create_fb(AppleGFXState *s) > > +{ > > + s->con = graphic_console_init(NULL, 0, &apple_gfx_fb_ops, s); > > + set_mode(s, 1440, 1080); > > + > > + s->cursor_show = true; > > +} > > + > > +static size_t apple_gfx_get_default_mmio_range_size(void) > > +{ > > + size_t mmio_range_size; > > + @autoreleasepool { > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > > + mmio_range_size = desc.mmioLength; > > + [desc release]; > > + } > > + return mmio_range_size; > > +} > > + > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const > char* obj_name) > > This function can be merged into apple_gfx_common_realize(). > > > Probably. I'll try it. > > > +{ > > + Error *local_err = NULL; > > + int r; > > + size_t mmio_range_size = > apple_gfx_get_default_mmio_range_size(); > > + > > + trace_apple_gfx_common_init(obj_name, mmio_range_size); > > + memory_region_init_io(&s->iomem_gfx, obj, &apple_gfx_ops, s, > obj_name, > > + mmio_range_size); > > + s->iomem_gfx.disable_reentrancy_guard = true; > > Why do you disable reentrancy guard? > > > Perhaps with the proposed AIO_WAIT_WHILE based I/O scheme, this won't be > an issue anymore, but the guard would otherwise keep dropping MMIOs > which immediately caused the PV graphics device to stop making progress. > The MMIO APIs in the PVG framework are thread- and reentrancy-safe, so > we certainly don't need to serialise them on our side. It's better to understand why such reentrancy happens. Reentrancy itself is often a sign of bug. > > > + > > + /* TODO: PVG framework supports serialising device state: > integrate it! */ > > + if (apple_gfx_mig_blocker == NULL) { > > + error_setg(&apple_gfx_mig_blocker, > > + "Migration state blocked by apple-gfx display > device"); > > + r = migrate_add_blocker(&apple_gfx_mig_blocker, &local_err); > > + if (r < 0) { > > + error_report_err(local_err); > > Please report the error to the caller of apple_gfx_common_realize() > instead. > > > + } > > + } > > +}> + > > +static void > apple_gfx_register_task_mapping_handlers(AppleGFXState *s, > > + > PGDeviceDescriptor *desc) > > +{ > > + desc.createTask = ^(uint64_t vmSize, void * _Nullable * > _Nonnull baseAddress) { > > + AppleGFXTask *task = apple_gfx_new_task(s, vmSize); > > + *baseAddress = (void*)task->address; > > nit: please write as (void *) instead of (void*). > > > + trace_apple_gfx_create_task(vmSize, *baseAddress); > > + return task; > > + }; > > + > > > +{ > > + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor new]; > > + > > + disp_desc.name <http://disp_desc.name> = @"QEMU display"; > > + disp_desc.sizeInMillimeters = NSMakeSize(400., 300.); /* A > 20" display */ > > + disp_desc.queue = dispatch_get_main_queue(); > > + disp_desc.newFrameEventHandler = ^(void) { > > + trace_apple_gfx_new_frame(); > > + dispatch_async(s->render_queue, ^{ > > + /* Drop frames if we get too far ahead. */ > > + if (s->pending_frames >= 2) > > + return; > > + ++s->pending_frames; > > + if (s->pending_frames > 1) { > > + return; > > + } > > + @autoreleasepool { > > + apple_gfx_render_new_frame(s); > > + } > > + }); > > + }; > > + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t sizeInPixels, > > + OSType pixelFormat) { > > + trace_apple_gfx_mode_change(sizeInPixels.x, sizeInPixels.y); > > + set_mode(s, sizeInPixels.x, sizeInPixels.y); > > + }; > > + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep *glyph, > > + PGDisplayCoord_t hotSpot) { > > + uint32_t bpp = glyph.bitsPerPixel; > > + size_t width = glyph.pixelsWide; > > + size_t height = glyph.pixelsHigh; > > + size_t padding_bytes_per_row = glyph.bytesPerRow - width > * 4; > > + const uint8_t* px_data = glyph.bitmapData; > > + > > + trace_apple_gfx_cursor_set(bpp, width, height); > > + > > + if (s->cursor) { > > + cursor_unref(s->cursor); > > + s->cursor = NULL; > > + } > > + > > + if (bpp == 32) { /* Shouldn't be anything else, but just > to be safe...*/ > > + s->cursor = cursor_alloc(width, height); > > + s->cursor->hot_x = hotSpot.x; > > + s->cursor->hot_y = hotSpot.y; > > + > > + uint32_t *dest_px = s->cursor->data; > > + > > + for (size_t y = 0; y < height; ++y) { > > + for (size_t x = 0; x < width; ++x) { > > + /* NSBitmapImageRep's red & blue channels > are swapped > > + * compared to QEMUCursor's. */ > > + *dest_px = > > + (px_data[0] << 16u) | > > + (px_data[1] << 8u) | > > + (px_data[2] << 0u) | > > + (px_data[3] << 24u); > > + ++dest_px; > > + px_data += 4; > > + } > > + px_data += padding_bytes_per_row; > > + } > > + dpy_cursor_define(s->con, s->cursor); > > + update_cursor(s); > > + } > > + }; > > + disp_desc.cursorShowHandler = ^(BOOL show) { > > + trace_apple_gfx_cursor_show(show); > > + s->cursor_show = show; > > + update_cursor(s); > > + }; > > + disp_desc.cursorMoveHandler = ^(void) { > > + trace_apple_gfx_cursor_move(); > > + update_cursor(s); > > + }; > > + > > + return disp_desc; > > +} > > + > > +static NSArray<PGDisplayMode*>* > apple_gfx_prepare_display_mode_array(void) > > +{ > > + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; > > + NSArray<PGDisplayMode*>* mode_array = nil; > > + int i; > > + > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > + modes[i] = > > + [[PGDisplayMode alloc] > initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; > > + } > > + > > + mode_array = [NSArray arrayWithObjects:modes > count:ARRAY_SIZE(apple_gfx_modes)]; > > + > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > + [modes[i] release]; > > + modes[i] = nil; > > + } > > + > > + return mode_array; > > +} > > + > > +static id<MTLDevice> copy_suitable_metal_device(void) > > +{ > > + id<MTLDevice> dev = nil; > > + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); > > + > > + /* Prefer a unified memory GPU. Failing that, pick a non- > removable GPU. */ > > + for (size_t i = 0; i < devs.count; ++i) { > > + if (devs[i].hasUnifiedMemory) { > > + dev = devs[i]; > > + break; > > + } > > + if (!devs[i].removable) { > > + dev = devs[i]; > > + } > > + } > > + > > + if (dev != nil) { > > + [dev retain]; > > + } else { > > + dev = MTLCreateSystemDefaultDevice(); > > + } > > + [devs release]; > > + > > + return dev; > > +} > > + > > +void apple_gfx_common_realize(AppleGFXState *s, > PGDeviceDescriptor *desc) > > +{ > > + PGDisplayDescriptor *disp_desc = nil; > > + > > + QTAILQ_INIT(&s->tasks); > > + s->render_queue = dispatch_queue_create("apple-gfx.render", > > + DISPATCH_QUEUE_SERIAL); > > + s->mtl = copy_suitable_metal_device(); > > + s->mtl_queue = [s->mtl newCommandQueue]; > > + > > + desc.device = s->mtl; > > + > > + apple_gfx_register_task_mapping_handlers(s, desc); > > + > > + s->pgdev = PGNewDeviceWithDescriptor(desc); > > + > > + disp_desc = apple_gfx_prepare_display_handlers(s); > > + s->pgdisp = [s->pgdev newDisplayWithDescriptor:disp_desc > > + port:0 > serialNum:1234]; > > + [disp_desc release]; > > + s->pgdisp.modeList = apple_gfx_prepare_display_mode_array(); > > + > > + create_fb(s); > > +} > > diff --git a/hw/display/meson.build b/hw/display/meson.build > > index 7db05eace97..70d855749c0 100644 > > --- a/hw/display/meson.build > > +++ b/hw/display/meson.build > > @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', if_true: > files('artist.c')) > > > > system_ss.add(when: 'CONFIG_ATI_VGA', if_true: [files('ati.c', > 'ati_2d.c', 'ati_dbg.c'), pixman]) > > > > +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: > [files('apple-gfx.m'), pvg, metal]) > > +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: > [files('apple-gfx-vmapple.m'), pvg, metal]) > > > > if config_all_devices.has_key('CONFIG_VIRTIO_GPU') > > virtio_gpu_ss = ss.source_set() > > diff --git a/hw/display/trace-events b/hw/display/trace-events > > index 781f8a33203..1809a358e36 100644 > > --- a/hw/display/trace-events > > +++ b/hw/display/trace-events > > @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) > "dest_width : %u" > > dm163_leds(int led, uint32_t value) "led %d: 0x%x" > > dm163_channels(int channel, uint8_t value) "channel %d: 0x%x" > > dm163_refresh_rate(uint32_t rr) "refresh rate %d" > > + > > +# apple-gfx.m > > +apple_gfx_read(uint64_t offset, uint64_t res) > "offset=0x%"PRIx64" res=0x%"PRIx64 > > +apple_gfx_write(uint64_t offset, uint64_t val) > "offset=0x%"PRIx64" val=0x%"PRIx64 > > +apple_gfx_create_task(uint32_t vm_size, void *va) "vm_size=0x%x > base_addr=%p" > > +apple_gfx_destroy_task(void *task) "task=%p" > > +apple_gfx_map_memory(void *task, uint32_t range_count, uint64_t > virtual_offset, uint32_t read_only) "task=%p range_count=0x%x > virtual_offset=0x%"PRIx64" read_only=%d" > > +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, > uint64_t phys_len) "[%d] phys_addr=0x%"PRIx64" phys_len=0x%"PRIx64 > > +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t > target) "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 > > +apple_gfx_unmap_memory(void *task, uint64_t virtual_offset, > uint64_t length) "task=%p virtual_offset=0x%"PRIx64" length=0x%"PRIx64 > > +apple_gfx_read_memory(uint64_t phys_address, uint64_t length, > void *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" > > +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" > > +apple_gfx_new_frame(void) "" > > +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" > y=%"PRId64 > > +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, uint64_t > height) "bpp=%d width=%"PRId64" height=0x%"PRId64 > > +apple_gfx_cursor_show(uint32_t show) "show=%d" > > +apple_gfx_cursor_move(void) "" > > +apple_gfx_common_init(const char *device_name, size_t mmio_size) > "device: %s; MMIO size: %zu bytes" > > + > > +# apple-gfx-vmapple.m > > +apple_iosfc_read(uint64_t offset, uint64_t res) > "offset=0x%"PRIx64" res=0x%"PRIx64 > > +apple_iosfc_write(uint64_t offset, uint64_t val) > "offset=0x%"PRIx64" val=0x%"PRIx64 > > +apple_iosfc_map_memory(uint64_t phys, uint64_t len, uint32_t ro, > void *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" ro=%d > va=%p e=%p f=%p" > > +apple_iosfc_unmap_memory(void *a, void *b, void *c, void *d, > void *e, void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" > > +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" > > + > > diff --git a/meson.build b/meson.build > > index 10464466ff3..f09df3f09d5 100644 > > --- a/meson.build > > +++ b/meson.build > > @@ -741,6 +741,8 @@ socket = [] > > version_res = [] > > coref = [] > > iokit = [] > > +pvg = [] > > +metal = [] > > emulator_link_args = [] > > midl = not_found > > widl = not_found > > @@ -762,6 +764,8 @@ elif host_os == 'darwin' > > coref = dependency('appleframeworks', modules: 'CoreFoundation') > > iokit = dependency('appleframeworks', modules: 'IOKit', > required: false) > > host_dsosuf = '.dylib' > > + pvg = dependency('appleframeworks', modules: > 'ParavirtualizedGraphics') > > + metal = dependency('appleframeworks', modules: 'Metal') > > elif host_os == 'sunos' > > socket = [cc.find_library('socket'), > > cc.find_library('nsl'), >
On Thu, 3 Oct 2024 at 09:09, Akihiko Odaki <akihiko.odaki@daynix.com> wrote: > On 2024/10/02 22:33, Phil Dennis-Jordan wrote: > > > > > > > +#include "apple-gfx.h" > > > +#include "monitor/monitor.h" > > > +#include "hw/sysbus.h" > > > +#include "hw/irq.h" > > > +#include "trace.h" > > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > > + > > > +_Static_assert(__aarch64__, ""); > > > > I don't think this assertion is worthwhile. This assertion will > trigger > > if you accidentally remove depends on AARCH64 from Kconfig, but I > don't > > think such code change happens by accident, and there is no reason to > > believe that this assertion won't be removed in such a case. > > > > > > As far as I'm aware the Kconfig AARCH64 dependency is for the /target/ > > architecture, not the /host/ architecture? The static assert checks for > > the latter. The PGIOSurfaceHostDeviceDescriptor type isn't available at > > all on non-aarch64 macOS hosts. I've not had any luck with using this > > variant of the device on x86-64 hosts simply by disabling any surface > > mapper code. > > > > Incidentally, if you know of a way to depend on a specific /host/ > > architecture in the Kconfig, that would be even better. I couldn't spot > > a way of doing that though. > > I got your intention now. The correct way to do that is to check for cpu > == 'aarch64'. Having assertion will break qemu-system-aarch64 on Intel > Macs. > OK, looks Iike that needs to be done at the meson.build level not Kconfig, but this seems to work: if cpu == 'aarch64' system_ss.add(when: 'CONFIG_MAC_PVG_MMIO', if_true: [files('apple-gfx-mmio.m'), pvg, metal]) endif > > > > It is dangerous to unlock BQL at an arbitrary place. Instead of > > unlocking, I suggest: > > - running [s->pgiosfc mmioReadAtOffset:offset] on another thread > > - using a bottom half to request operations that require BQL from the > > thread running [s->pgiosfc mmioReadAtOffset:offset] > > - calling AIO_WAIT_WHILE() to process the bottom half and to wait for > > the completion of [s->pgiosfc mmioReadAtOffset:offset] > > > > > > OK, I think I see what you mean, I'll try to rework things around that > > pattern. Any preference on how I kick off the job on the other thread? > > As we necessarily need to use libdispatch in a bunch of places in this > > code anyway, I guess dispatch_async() would probably be the simplest? > > Perhaps so. The QEMU way is to use a bottom half with AioContext, but > you can't simultaneously run a dispatch queue and AioContext in one > thread so you have to use the dispatch queue if you need one. > > > > > > + res = [s->pgiosfc mmioReadAtOffset:offset]; > > > + bql_lock(); > > > + > > > + trace_apple_iosfc_read(offset, res); > > > + > > > + return res; > > > +} > > > + > > > +static void apple_iosfc_write( > > > + void *opaque, hwaddr offset, uint64_t val, unsigned size) > > > +{ > > > + AppleGFXVmappleState *s = opaque; > > > + > > > + trace_apple_iosfc_write(offset, val); > > > + > > > + [s->pgiosfc mmioWriteAtOffset:offset value:val]; > > > +} > > > + > > > +static const MemoryRegionOps apple_iosfc_ops = { > > > + .read = apple_iosfc_read, > > > + .write = apple_iosfc_write, > > > + .endianness = DEVICE_LITTLE_ENDIAN, > > > + .valid = { > > > + .min_access_size = 4, > > > + .max_access_size = 8, > > > + }, > > > + .impl = { > > > + .min_access_size = 4, > > > + .max_access_size = 8, > > > + }, > > > +}; > > > + > > > +static PGIOSurfaceHostDevice > > *apple_gfx_prepare_iosurface_host_device( > > > + AppleGFXVmappleState *s) > > > +{ > > > + PGIOSurfaceHostDeviceDescriptor *iosfc_desc = > > > + [PGIOSurfaceHostDeviceDescriptor new]; > > > + PGIOSurfaceHostDevice *iosfc_host_dev = nil; > > > + > > > + iosfc_desc.mapMemory = > > > + ^(uint64_t phys, uint64_t len, bool ro, void **va, void > > *e, void *f) { > > > + trace_apple_iosfc_map_memory(phys, len, ro, va, e, > f); > > > + MemoryRegion *tmp_mr; > > > + *va = gpa2hva(&tmp_mr, phys, len, NULL); > > > > Use: dma_memory_map() > > > > > > That doesn't seem to be a precisely equivalent operation. It also says > > in its headerdoc, > > > > Use only for reads OR writes - not for read-modify-write operations. > > > > which I don't think we can guarantee here at all. > > > > I guess I can call it twice, once for writing and once for reading, but > > given that the dma_memory_unmap operation marks the area dirty, I'm not > > it's intended for what I understand the use case here to be: As far as I > > can tell, the PV graphics device uses (some) of this memory to exchange > > data in a cache-coherent way between host and guest, e.g. as a lock-free > > ring buffer, using atomic operations as necessary. (This works because > > it's a PV device: it "knows" the other end just another CPU core (or > > even the same one) executing in a different hypervisor context.) This > > doesn't really match "traditional" DMA patterns where there's either a > > read or a write happening. > > I think the story is a bit different for this VMApple variant. Probably > the CPU and GPU in Apple Silicon is cache-coherent so you can map normal > memory for GPU without any kind of cache maintenance. > > Cache conherency of CPU and GPU in Apple Silicon is implied with Apple's > documentation; it says you don't need to synchronize resources for > MTLStorageModeShared, which is the default for Apple Silicon. > > https://developer.apple.com/documentation/metal/resource_fundamentals/synchronizing_a_managed_resource_in_macos > > The name "IOSurface" also implies it is used not only for e.g., ring > buffer but also for real data. > Note that the PGTask map/unmap callbacks appear to have equivalent semantics, so it's not just the surface mapping. > > > > Hunting around some more for alternative APIs, there's also > > memory_region_get_ram_ptr(), but I'm not sure its restrictions apply > > here either. > > I think you can call memory_access_is_direct() to check if the > requirement is satisfied. > > It will still break dirty page tracking implemented by > dma_memory_unmap() and others, but it's broken for hvf, which does not > implement dirty page tracking either. > > > > > + return (bool)true; > > > > Why cast? > > > > > > Good question. Not originally my code, so I've fixed all the instances I > > could find now. > OK, it turns out the reason for this is that C treats 'true' as an int, which then becomes the block's inferred return type - and the callbacks are expecting bool-returning blocks. I've fixed it by explicitly specifying the block return type and removing the cast in the return statement: iosfc_desc.unmapMemory = ^bool(…) { … return true; }; > > > + > > > + iosfc_desc.unmapMemory = > > > + ^(void *a, void *b, void *c, void *d, void *e, void *f) { > > > + trace_apple_iosfc_unmap_memory(a, b, c, d, e, f); > > > + return (bool)true; > > > + }; > > > + > > > + iosfc_desc.raiseInterrupt = ^(uint32_t vector) { > > > + trace_apple_iosfc_raise_irq(vector); > > > + bool locked = bql_locked(); > > > + if (!locked) { > > > + bql_lock(); > > > + } > > > + qemu_irq_pulse(s->irq_iosfc);> + if (!locked) { > > > + bql_unlock(); > > > + } > > > + return (bool)true; > > > + }; > > > + > > > + iosfc_host_dev = > > > + [[PGIOSurfaceHostDevice alloc] > > initWithDescriptor:iosfc_desc]; > > > + [iosfc_desc release]; > > > + return iosfc_host_dev; > > > +} > > > + > > > +static void apple_gfx_vmapple_realize(DeviceState *dev, Error > > **errp) > > > +{ > > > + @autoreleasepool { > > > > This autoreleasepool is not used. > > > > > > It is definitely used inside the apple_gfx_common_realize() call. It's > > also impossible to say whether [PGDeviceDescriptor new] uses autorelease > > semantics internally, so it seemed safer to wrap the whole thing in an > > outer pool. > > Theoretically, It should be safe to assume the callee creates > autoreleasepool by themselves as needed in general. We have bunch of > code to call Objective-C APIs without creating autoreleasepool in the > caller. Practically, [PGDeviceDescriptor new] is likely to be > implemented with ARC, which wraps methods in autoreleasepool as necessary. > As far as I'm aware, ARC does NOT automatically insert autorelease pool blocks. The reason you rarely need to create autoreleasepool blocks in "plain" Objective-C programming is that Cocoa/CFRunloop/libdispatch event handlers run each event in an autoreleasepool. So you don't need to create them explicitly when using dispatch_async and similar, or when running code on the main thread (which runs inside NSApplicationMain/CFRunloopRun/dispatch_main). As far as I'm aware, if you don't explicitly define autoreleasepools in raw threads created with the pthreads API, any autoreleased objects will leak. At least I've not found any specification/documentation contradicting this. And most code in Qemu runs on such raw threads, so we need to play it safe with regard to autorelease semantics. Whether the existing Qemu Objective-C code is safe in this regard I don't know for certain, but I've certainly paid attention to this aspect when modifying ui/cocoa.m in the past, and indeed most of that code runs on the main thread. Note also how I wrap the apple_gfx_render_new_frame call in a pool when it can't be guaranteed it's running on a dispatch queue because the command buffer inside that uses autorelease semantics. Functions that uses a method that returns autorelease resources should > be wrapped with autoreleasepool instead of assuming the caller creates > autoreleasepool for them. > I'm treating apple_gfx_common_realize as an internal API, and I don't think expecting its callers to wrap it in an autoreleasepool block is unreasonable. I can certainly explicitly document this in a comment. > > > diff --git a/hw/display/apple-gfx.m b/hw/display/apple-gfx.m > > > new file mode 100644 > > > index 00000000000..837300f9cd4 > > > --- /dev/null > > > +++ b/hw/display/apple-gfx.m > > > @@ -0,0 +1,536 @@ > > > +/* > > > + * QEMU Apple ParavirtualizedGraphics.framework device > > > + * > > > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All > > Rights Reserved. > > > + * > > > + * This work is licensed under the terms of the GNU GPL, version > > 2 or later. > > > + * See the COPYING file in the top-level directory. > > > + * > > > + * ParavirtualizedGraphics.framework is a set of libraries that > > macOS provides > > > + * which implements 3d graphics passthrough to the host as well > as a > > > + * proprietary guest communication channel to drive it. This > > device model > > > + * implements support to drive that library from within QEMU. > > > + */ > > > + > > > +#include "apple-gfx.h" > > > +#include "trace.h" > > > +#include "qemu/main-loop.h" > > > +#include "ui/console.h" > > > +#include "monitor/monitor.h" > > > +#include "qapi/error.h" > > > +#include "migration/blocker.h" > > > +#include <mach/mach_vm.h> > > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > > + > > > +static const PGDisplayCoord_t apple_gfx_modes[] = { > > > + { .x = 1440, .y = 1080 }, > > > + { .x = 1280, .y = 1024 }, > > > +}; > > > + > > > +typedef struct PGTask_s { // Name matches forward declaration in > > PG header > > > > Let's name it AppleGFXTask. It is a common practice to have the same > > tag > > name and typedef in QEMU. > > > > > > This is defining a forward-declared type from framework headers which is > > opaque from the framework's point of view. We do not get to choose its > > struct name. The alternative is having casts wherever these objects are > > being passed between our code and the framework. (See the original v1/v2 > > vmapple patch series for how messy this is.) > > I got the idea. Let's not avoid the typedef then to clarify the naming > is not under our control. > I'm not sure what you mean by this double negative. Are you saying, don't add our own typedef for struct PGTask_s at all, just use the framework-supplied PGTask_t where appropriate? > > > > > > +static void apple_gfx_render_frame_completed(AppleGFXState *s, > > void *vram, > > > + id<MTLTexture> > texture) > > > +{ > > > + --s->pending_frames; > > > + assert(s->pending_frames >= 0); > > > + > > > + if (vram != s->vram) { > > > + /* Display mode has changed, drop this old frame. */ > > > + assert(texture != s->texture); > > > + g_free(vram); > > > > This management of buffers looks a bit convoluted. I suggest > > remembering > > the width and height instead of pointers and comparing them. This way > > you can free resources in set_mode(). > > > > > > Yeah, I suppose that works, I can change that around. > > > > > + } else { > > > + copy_mtl_texture_to_surface_mem(texture, vram); > > > > Writing vram outside BQL may result in tearing. > > > > > > As far as I can tell(*), QXL does the same. I couldn't find any examples > > of double-buffering in any of the existing display devices, which would > > be the only way to do async updates efficiently and without tearing. In > > any case, this solution is still vastly better than a regular VGA > > device, which suffers from very visible tearing with macOS on the guest > > side anyway. And in an ideal world, we'd pass through the rendered > > texture directly to the Cocoa UI code rather than copying out only for > > the CPU to draw it back into a window surface which is then passed to > > the GPU for host side rendering. But I felt this patch is already very, > > very large, and if anyone cares, we can fix imperfections in subsequent > > updates. > > > > (*)The rendering code in that device is also fairly complex, so I may be > > misreading it. > > QXL always modifies the surface with BQL. The surface is modified with > qxl_blit(), which is called by qxl_render_update_area_unlocked(). > qxl_render_update_area_unlocked() is called by either of > qxl_render_update() and qxl_render_update_area_bh(). Both of them are > called with BQL. The name includes "unlocked", but it means it is called > without holding QXL-internal lock. > > Most devices works entirely with BQL so they don't perform double > buffering. apple-gfx can do the same. > I think we can safely move apple-gfx's framebuffer state management back inside the BQL, yes. I just figured that copying dozens of megabytes of framebuffer data on every frame while holding the lock was not going to help BQL contention. Especially as PVG does not have a concept of dirty areas, so we must copy the whole framebuffer every time. (Unless we were to implement dirty area detection ourselves.) Unfortunately, implementing double-buffering would require a major rework of Qemu's whole surface management, console code, and probably most of the UI implementations. I'm guessing the OpenGL fast-path sidesteps all of this, so replicating that with Metal would probably be the easier way forward. (Although doing all this graphics stuff inside the BQL generally seems like a major architectural flaw; I suppose most enterprise use of Qemu does not involve the framebuffer, so it's not shown up in BQL contention profiling there. It certainly does in desktop use, although at least on macOS hosts there are far worse culprits in that regard.) > > > > + if (s->gfx_update_requested) { > > > + s->gfx_update_requested = false; > > > + dpy_gfx_update_full(s->con); > > > + graphic_hw_update_done(s->con); > > > + s->new_frame_ready = false; > > > > This assignment is unnecessary as s->new_frame_ready is always false > > when s->gfx_update_requested. If you want to make sure > > s->gfx_update_requested and s->new_frame_ready are mutually > exclusive, > > use one enum value instead of having two bools. > > > > > > I'll need to refresh my memory and get back to you on this one, it's > > been so many months since I actively worked on this code. > > > > > + } else { > > > + s->new_frame_ready = true; > > > + } > > > + } > > > + if (s->pending_frames > 0) { > > > + apple_gfx_render_new_frame(s); > > > + } > > > +} > > > + > > > +static void apple_gfx_fb_update_display(void *opaque) > > > +{ > > > + AppleGFXState *s = opaque; > > > + > > > + dispatch_async(s->render_queue, ^{ > > > + if (s->pending_frames > 0) { > > > > It should check for s->new_frame_ready as > > apple_gfx_render_frame_completed() doesn't check if > > s->pending_frames > 0 before calling graphic_hw_update_done(), which > is > > inconsistent. > > > > > > pending_frames is about guest-side frames that are queued to be rendered > > by the host GPU. > > new_frame_ready being true indicates that the contents of the Qemu > > console surface has been updated with new frame data since the last > > gfx_update. > > gfx_update_requested indicates that gfx_update is currently awaiting an > > async completion (graphic_hw_update_done) but the surface has not > > received a new frame content, but the GPU is stily busy drawing one. > > > > apple_gfx_render_frame_completed is scheduled exactly once per pending > > frame, so pending_frames > 0 is an invariant there. (Hence the assert.)> > > I don't think there is any inconsistency here, but I'll double check. > > It's possible that there's an easier way to express the state machine, > > and I'll take a look at that. > > I meant that apple_gfx_render_frame_completed() does not check if the > frame is the last one currently pending. apple_gfx_fb_update_display() > ignores a new ready frame when there is a more pending frame, but > apple_gfx_render_frame_completed() unconditionally fires > graphic_hw_update_done() even if there is a more pending frame. And I > think apple_gfx_render_frame_completed() is right and > apple_gfx_fb_update_display() is wrong in such a situation. > > OK, got it. And yes, I agree. > > > > Checking if s->pending_frames > 0 both in > apple_gfx_fb_update_display() > > and apple_gfx_render_frame_completed() is also problematic as that > can > > defer graphic_hw_update_done() indefinitely if we are getting new > > frames > > too fast. > > > > > > I see what you mean about this part. I'll have to test it, but I guess > > we should reverse the priority, like this: > > > > if (s->new_frame_ready) { > > dpy_gfx_update_full(s->con); > > s->new_frame_ready = false; > > graphic_hw_update_done(s->con); > > } else if (s->pending_frames > 0) { > > s->gfx_update_requested = true; > > } else { > > graphic_hw_update_done(s->con); > > } > > > > 1. If we already have a frame, ready to be displayed, return it > immediately. > > 2. If the guest has reported that it's completed a frame and the GPU is > > currently busy rendering it, defer graphic_hw_update_done until it's > done. > > 3. If the guest reports no changes to its display, indicate this back to > > Qemu as a no-op display update graphic_hw_update_done() with no > > dpy_gfx_update* call. > > Yes, that looks correct. > > > > > > + s->gfx_update_requested = true; > > > + } else { > > > + if (s->new_frame_ready) { > > > + dpy_gfx_update_full(s->con); > > > + s->new_frame_ready = false; > > > + } > > > + graphic_hw_update_done(s->con); > > > + }> + }); > > > +} > > > + > > > +static const GraphicHwOps apple_gfx_fb_ops = { > > > + .gfx_update = apple_gfx_fb_update_display, > > > + .gfx_update_async = true, > > > +}; > > > + > > > +static void update_cursor(AppleGFXState *s) > > > +{ > > > + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, > > > + s->pgdisp.cursorPosition.y, s->cursor_show); > > > +} > > > + > > > +static void set_mode(AppleGFXState *s, uint32_t width, uint32_t > > height) > > > +{ > > > + void *vram = NULL; > > > + DisplaySurface *surface; > > > + MTLTextureDescriptor *textureDescriptor; > > > + id<MTLTexture> texture = nil; > > > + __block bool no_change = false; > > > + > > > + dispatch_sync(s->render_queue, > > > > Calling dispatch_sync() while holding BQL may result in deadlock. > > > > Only if any code executed on the same dispatch queue acquires the lock > > either directly or transitively. I believe I have ensure this is not > > done on the reqnder_queue, have you found anywhere this is the case? > > The documentation is not clear what threads a dispatch queue runs on. We > can have a deadlock if they lock the BQL. > dispatch_sync is a synchronisation primitive (it waits for and asserts exclusive access to the given queue), it doesn't actually do any thread scheduling. Work scheduled asynchronously to non-main dispatch queues will otherwise run on libdispatch pool threads. Running blocks on dispatch queues will not preempt and schedule it on other threads which may or may not be holding some locks. So the only way this code will deadlock is if any code scheduled to render_queue directly or transitively acquires the BQL. None of it does, although updating the console while holding the BQL rather complicates this. > > > > > + ^{ > > > + if (s->surface && > > > + width == surface_width(s->surface) && > > > + height == surface_height(s->surface)) { > > > + no_change = true; > > > + } > > > + }); > > > + > > > + if (no_change) { > > > + return; > > > + } > > > + > > > + vram = g_malloc0(width * height * 4); > > > + surface = qemu_create_displaysurface_from(width, height, > > PIXMAN_LE_a8r8g8b8, > > > + width * 4, vram); > > > + > > > + @autoreleasepool { > > > + textureDescriptor = > > > + [MTLTextureDescriptor > > > + > > texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm > > > + width:width > > > + height:height > > > + mipmapped:NO]; > > > + textureDescriptor.usage = s->pgdisp.minimumTextureUsage; > > > + texture = [s->mtl > > newTextureWithDescriptor:textureDescriptor]; > > > + } > > > + > > > + s->using_managed_texture_storage = > > > + (texture.storageMode == MTLStorageModeManaged); > > > + > > > + dispatch_sync(s->render_queue, > > > + ^{ > > > + id<MTLTexture> old_texture = nil; > > > + void *old_vram = s->vram; > > > + s->vram = vram; > > > + s->surface = surface; > > > + > > > + dpy_gfx_replace_surface(s->con, surface); > > > + > > > + old_texture = s->texture; > > > + s->texture = texture; > > > + [old_texture release]; > > > > You can just do: > > [s->texture release]; > > s->texture = texture; > > > > This will make s->texture dangling between the two statements, but > that > > don't matter since s->texture is not an atomic variable that can be > > safely observed from another thread anyway. > > > > > + > > > + if (s->pending_frames == 0) { > > > + g_free(old_vram); > > > + } > > > + }); > > > +} > > > + > > > +static void create_fb(AppleGFXState *s) > > > +{ > > > + s->con = graphic_console_init(NULL, 0, &apple_gfx_fb_ops, s); > > > + set_mode(s, 1440, 1080); > > > + > > > + s->cursor_show = true; > > > +} > > > + > > > +static size_t apple_gfx_get_default_mmio_range_size(void) > > > +{ > > > + size_t mmio_range_size; > > > + @autoreleasepool { > > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > > > + mmio_range_size = desc.mmioLength; > > > + [desc release]; > > > + } > > > + return mmio_range_size; > > > +} > > > + > > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const > > char* obj_name) > > > > This function can be merged into apple_gfx_common_realize(). > > > > > > Probably. I'll try it. > Upon further inspection, we need to call cocoa_enable_runloop_on_main_thread() during the init phase, not realize(). So we can't get rid of this entirely. Is there any value in moving the other init code into _realize()? > > > +{ > > > + Error *local_err = NULL; > > > + int r; > > > + size_t mmio_range_size = > > apple_gfx_get_default_mmio_range_size(); > > > + > > > + trace_apple_gfx_common_init(obj_name, mmio_range_size); > > > + memory_region_init_io(&s->iomem_gfx, obj, &apple_gfx_ops, s, > > obj_name, > > > + mmio_range_size); > > > + s->iomem_gfx.disable_reentrancy_guard = true; > > > > Why do you disable reentrancy guard? > > > > > > Perhaps with the proposed AIO_WAIT_WHILE based I/O scheme, this won't be > > an issue anymore, but the guard would otherwise keep dropping MMIOs > > which immediately caused the PV graphics device to stop making progress. > > The MMIO APIs in the PVG framework are thread- and reentrancy-safe, so > > we certainly don't need to serialise them on our side. > > It's better to understand why such reentrancy happens. Reentrancy itself > is often a sign of bug. > > > > > > + > > > + /* TODO: PVG framework supports serialising device state: > > integrate it! */ > > > + if (apple_gfx_mig_blocker == NULL) { > > > + error_setg(&apple_gfx_mig_blocker, > > > + "Migration state blocked by apple-gfx display > > device"); > > > + r = migrate_add_blocker(&apple_gfx_mig_blocker, > &local_err); > > > + if (r < 0) { > > > + error_report_err(local_err); > > > > Please report the error to the caller of apple_gfx_common_realize() > > instead. > > > > > + } > > > + } > > > +}> + > > > +static void > > apple_gfx_register_task_mapping_handlers(AppleGFXState *s, > > > + > > PGDeviceDescriptor *desc) > > > +{ > > > + desc.createTask = ^(uint64_t vmSize, void * _Nullable * > > _Nonnull baseAddress) { > > > + AppleGFXTask *task = apple_gfx_new_task(s, vmSize); > > > + *baseAddress = (void*)task->address; > > > > nit: please write as (void *) instead of (void*). > > > > > + trace_apple_gfx_create_task(vmSize, *baseAddress); > > > + return task; > > > + }; > > > + > > > > > +{ > > > + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor new]; > > > + > > > + disp_desc.name <http://disp_desc.name> = @"QEMU display"; > > > + disp_desc.sizeInMillimeters = NSMakeSize(400., 300.); /* A > > 20" display */ > > > + disp_desc.queue = dispatch_get_main_queue(); > > > + disp_desc.newFrameEventHandler = ^(void) { > > > + trace_apple_gfx_new_frame(); > > > + dispatch_async(s->render_queue, ^{ > > > + /* Drop frames if we get too far ahead. */ > > > + if (s->pending_frames >= 2) > > > + return; > > > + ++s->pending_frames; > > > + if (s->pending_frames > 1) { > > > + return; > > > + } > > > + @autoreleasepool { > > > + apple_gfx_render_new_frame(s); > > > + } > > > + }); > > > + }; > > > + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t > sizeInPixels, > > > + OSType pixelFormat) { > > > + trace_apple_gfx_mode_change(sizeInPixels.x, > sizeInPixels.y); > > > + set_mode(s, sizeInPixels.x, sizeInPixels.y); > > > + }; > > > + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep *glyph, > > > + PGDisplayCoord_t hotSpot) { > > > + uint32_t bpp = glyph.bitsPerPixel; > > > + size_t width = glyph.pixelsWide; > > > + size_t height = glyph.pixelsHigh; > > > + size_t padding_bytes_per_row = glyph.bytesPerRow - width > > * 4; > > > + const uint8_t* px_data = glyph.bitmapData; > > > + > > > + trace_apple_gfx_cursor_set(bpp, width, height); > > > + > > > + if (s->cursor) { > > > + cursor_unref(s->cursor); > > > + s->cursor = NULL; > > > + } > > > + > > > + if (bpp == 32) { /* Shouldn't be anything else, but just > > to be safe...*/ > > > + s->cursor = cursor_alloc(width, height); > > > + s->cursor->hot_x = hotSpot.x; > > > + s->cursor->hot_y = hotSpot.y; > > > + > > > + uint32_t *dest_px = s->cursor->data; > > > + > > > + for (size_t y = 0; y < height; ++y) { > > > + for (size_t x = 0; x < width; ++x) { > > > + /* NSBitmapImageRep's red & blue channels > > are swapped > > > + * compared to QEMUCursor's. */ > > > + *dest_px = > > > + (px_data[0] << 16u) | > > > + (px_data[1] << 8u) | > > > + (px_data[2] << 0u) | > > > + (px_data[3] << 24u); > > > + ++dest_px; > > > + px_data += 4; > > > + } > > > + px_data += padding_bytes_per_row; > > > + } > > > + dpy_cursor_define(s->con, s->cursor); > > > + update_cursor(s); > > > + } > > > + }; > > > + disp_desc.cursorShowHandler = ^(BOOL show) { > > > + trace_apple_gfx_cursor_show(show); > > > + s->cursor_show = show; > > > + update_cursor(s); > > > + }; > > > + disp_desc.cursorMoveHandler = ^(void) { > > > + trace_apple_gfx_cursor_move(); > > > + update_cursor(s); > > > + }; > > > + > > > + return disp_desc; > > > +} > > > + > > > +static NSArray<PGDisplayMode*>* > > apple_gfx_prepare_display_mode_array(void) > > > +{ > > > + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; > > > + NSArray<PGDisplayMode*>* mode_array = nil; > > > + int i; > > > + > > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > > + modes[i] = > > > + [[PGDisplayMode alloc] > > initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; > > > + } > > > + > > > + mode_array = [NSArray arrayWithObjects:modes > > count:ARRAY_SIZE(apple_gfx_modes)]; > > > + > > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > > + [modes[i] release]; > > > + modes[i] = nil; > > > + } > > > + > > > + return mode_array; > > > +} > > > + > > > +static id<MTLDevice> copy_suitable_metal_device(void) > > > +{ > > > + id<MTLDevice> dev = nil; > > > + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); > > > + > > > + /* Prefer a unified memory GPU. Failing that, pick a non- > > removable GPU. */ > > > + for (size_t i = 0; i < devs.count; ++i) { > > > + if (devs[i].hasUnifiedMemory) { > > > + dev = devs[i]; > > > + break; > > > + } > > > + if (!devs[i].removable) { > > > + dev = devs[i]; > > > + } > > > + } > > > + > > > + if (dev != nil) { > > > + [dev retain]; > > > + } else { > > > + dev = MTLCreateSystemDefaultDevice(); > > > + } > > > + [devs release]; > > > + > > > + return dev; > > > +} > > > + > > > +void apple_gfx_common_realize(AppleGFXState *s, > > PGDeviceDescriptor *desc) > > > +{ > > > + PGDisplayDescriptor *disp_desc = nil; > > > + > > > + QTAILQ_INIT(&s->tasks); > > > + s->render_queue = dispatch_queue_create("apple-gfx.render", > > > + > DISPATCH_QUEUE_SERIAL); > > > + s->mtl = copy_suitable_metal_device(); > > > + s->mtl_queue = [s->mtl newCommandQueue]; > > > + > > > + desc.device = s->mtl; > > > + > > > + apple_gfx_register_task_mapping_handlers(s, desc); > > > + > > > + s->pgdev = PGNewDeviceWithDescriptor(desc); > > > + > > > + disp_desc = apple_gfx_prepare_display_handlers(s); > > > + s->pgdisp = [s->pgdev newDisplayWithDescriptor:disp_desc > > > + port:0 > > serialNum:1234]; > > > + [disp_desc release]; > > > + s->pgdisp.modeList = apple_gfx_prepare_display_mode_array(); > > > + > > > + create_fb(s); > > > +} > > > diff --git a/hw/display/meson.build b/hw/display/meson.build > > > index 7db05eace97..70d855749c0 100644 > > > --- a/hw/display/meson.build > > > +++ b/hw/display/meson.build > > > @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', if_true: > > files('artist.c')) > > > > > > system_ss.add(when: 'CONFIG_ATI_VGA', if_true: [files('ati.c', > > 'ati_2d.c', 'ati_dbg.c'), pixman]) > > > > > > +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: > > [files('apple-gfx.m'), pvg, metal]) > > > +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: > > [files('apple-gfx-vmapple.m'), pvg, metal]) > > > > > > if config_all_devices.has_key('CONFIG_VIRTIO_GPU') > > > virtio_gpu_ss = ss.source_set() > > > diff --git a/hw/display/trace-events b/hw/display/trace-events > > > index 781f8a33203..1809a358e36 100644 > > > --- a/hw/display/trace-events > > > +++ b/hw/display/trace-events > > > @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) > > "dest_width : %u" > > > dm163_leds(int led, uint32_t value) "led %d: 0x%x" > > > dm163_channels(int channel, uint8_t value) "channel %d: 0x%x" > > > dm163_refresh_rate(uint32_t rr) "refresh rate %d" > > > + > > > +# apple-gfx.m > > > +apple_gfx_read(uint64_t offset, uint64_t res) > > "offset=0x%"PRIx64" res=0x%"PRIx64 > > > +apple_gfx_write(uint64_t offset, uint64_t val) > > "offset=0x%"PRIx64" val=0x%"PRIx64 > > > +apple_gfx_create_task(uint32_t vm_size, void *va) "vm_size=0x%x > > base_addr=%p" > > > +apple_gfx_destroy_task(void *task) "task=%p" > > > +apple_gfx_map_memory(void *task, uint32_t range_count, uint64_t > > virtual_offset, uint32_t read_only) "task=%p range_count=0x%x > > virtual_offset=0x%"PRIx64" read_only=%d" > > > +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, > > uint64_t phys_len) "[%d] phys_addr=0x%"PRIx64" phys_len=0x%"PRIx64 > > > +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t > > target) "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 > > > +apple_gfx_unmap_memory(void *task, uint64_t virtual_offset, > > uint64_t length) "task=%p virtual_offset=0x%"PRIx64" > length=0x%"PRIx64 > > > +apple_gfx_read_memory(uint64_t phys_address, uint64_t length, > > void *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" > > > +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" > > > +apple_gfx_new_frame(void) "" > > > +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" > > y=%"PRId64 > > > +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, uint64_t > > height) "bpp=%d width=%"PRId64" height=0x%"PRId64 > > > +apple_gfx_cursor_show(uint32_t show) "show=%d" > > > +apple_gfx_cursor_move(void) "" > > > +apple_gfx_common_init(const char *device_name, size_t mmio_size) > > "device: %s; MMIO size: %zu bytes" > > > + > > > +# apple-gfx-vmapple.m > > > +apple_iosfc_read(uint64_t offset, uint64_t res) > > "offset=0x%"PRIx64" res=0x%"PRIx64 > > > +apple_iosfc_write(uint64_t offset, uint64_t val) > > "offset=0x%"PRIx64" val=0x%"PRIx64 > > > +apple_iosfc_map_memory(uint64_t phys, uint64_t len, uint32_t ro, > > void *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" ro=%d > > va=%p e=%p f=%p" > > > +apple_iosfc_unmap_memory(void *a, void *b, void *c, void *d, > > void *e, void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" > > > +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" > > > + > > > diff --git a/meson.build b/meson.build > > > index 10464466ff3..f09df3f09d5 100644 > > > --- a/meson.build > > > +++ b/meson.build > > > @@ -741,6 +741,8 @@ socket = [] > > > version_res = [] > > > coref = [] > > > iokit = [] > > > +pvg = [] > > > +metal = [] > > > emulator_link_args = [] > > > midl = not_found > > > widl = not_found > > > @@ -762,6 +764,8 @@ elif host_os == 'darwin' > > > coref = dependency('appleframeworks', modules: > 'CoreFoundation') > > > iokit = dependency('appleframeworks', modules: 'IOKit', > > required: false) > > > host_dsosuf = '.dylib' > > > + pvg = dependency('appleframeworks', modules: > > 'ParavirtualizedGraphics') > > > + metal = dependency('appleframeworks', modules: 'Metal') > > > elif host_os == 'sunos' > > > socket = [cc.find_library('socket'), > > > cc.find_library('nsl'), > > > >
On 2024/10/06 19:39, Phil Dennis-Jordan wrote: > > > On Thu, 3 Oct 2024 at 09:09, Akihiko Odaki <akihiko.odaki@daynix.com > <mailto:akihiko.odaki@daynix.com>> wrote: > > On 2024/10/02 22:33, Phil Dennis-Jordan wrote: > > > > > > > +#include "apple-gfx.h" > > > +#include "monitor/monitor.h" > > > +#include "hw/sysbus.h" > > > +#include "hw/irq.h" > > > +#include "trace.h" > > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > > + > > > +_Static_assert(__aarch64__, ""); > > > > I don't think this assertion is worthwhile. This assertion > will trigger > > if you accidentally remove depends on AARCH64 from Kconfig, > but I don't > > think such code change happens by accident, and there is no > reason to > > believe that this assertion won't be removed in such a case. > > > > > > As far as I'm aware the Kconfig AARCH64 dependency is for the / > target/ > > architecture, not the /host/ architecture? The static assert > checks for > > the latter. The PGIOSurfaceHostDeviceDescriptor type isn't > available at > > all on non-aarch64 macOS hosts. I've not had any luck with using > this > > variant of the device on x86-64 hosts simply by disabling any > surface > > mapper code. > > > > Incidentally, if you know of a way to depend on a specific /host/ > > architecture in the Kconfig, that would be even better. I > couldn't spot > > a way of doing that though. > > I got your intention now. The correct way to do that is to check for > cpu > == 'aarch64'. Having assertion will break qemu-system-aarch64 on > Intel Macs. > > > OK, looks Iike that needs to be done at the meson.build level not > Kconfig, but this seems to work: > > if cpu == 'aarch64' > system_ss.add(when: 'CONFIG_MAC_PVG_MMIO', if_true: [files('apple- > gfx-mmio.m'), pvg, metal]) > endif > > > > > It is dangerous to unlock BQL at an arbitrary place. Instead of > > unlocking, I suggest: > > - running [s->pgiosfc mmioReadAtOffset:offset] on another thread > > - using a bottom half to request operations that require BQL > from the > > thread running [s->pgiosfc mmioReadAtOffset:offset] > > - calling AIO_WAIT_WHILE() to process the bottom half and to > wait for > > the completion of [s->pgiosfc mmioReadAtOffset:offset] > > > > > > OK, I think I see what you mean, I'll try to rework things around > that > > pattern. Any preference on how I kick off the job on the other > thread? > > As we necessarily need to use libdispatch in a bunch of places in > this > > code anyway, I guess dispatch_async() would probably be the simplest? > > Perhaps so. The QEMU way is to use a bottom half with AioContext, but > you can't simultaneously run a dispatch queue and AioContext in one > thread so you have to use the dispatch queue if you need one. > > > > > > + res = [s->pgiosfc mmioReadAtOffset:offset]; > > > + bql_lock(); > > > + > > > + trace_apple_iosfc_read(offset, res); > > > + > > > + return res; > > > +} > > > + > > > +static void apple_iosfc_write( > > > + void *opaque, hwaddr offset, uint64_t val, unsigned size) > > > +{ > > > + AppleGFXVmappleState *s = opaque; > > > + > > > + trace_apple_iosfc_write(offset, val); > > > + > > > + [s->pgiosfc mmioWriteAtOffset:offset value:val]; > > > +} > > > + > > > +static const MemoryRegionOps apple_iosfc_ops = { > > > + .read = apple_iosfc_read, > > > + .write = apple_iosfc_write, > > > + .endianness = DEVICE_LITTLE_ENDIAN, > > > + .valid = { > > > + .min_access_size = 4, > > > + .max_access_size = 8, > > > + }, > > > + .impl = { > > > + .min_access_size = 4, > > > + .max_access_size = 8, > > > + }, > > > +}; > > > + > > > +static PGIOSurfaceHostDevice > > *apple_gfx_prepare_iosurface_host_device( > > > + AppleGFXVmappleState *s) > > > +{ > > > + PGIOSurfaceHostDeviceDescriptor *iosfc_desc = > > > + [PGIOSurfaceHostDeviceDescriptor new]; > > > + PGIOSurfaceHostDevice *iosfc_host_dev = nil; > > > + > > > + iosfc_desc.mapMemory = > > > + ^(uint64_t phys, uint64_t len, bool ro, void > **va, void > > *e, void *f) { > > > + trace_apple_iosfc_map_memory(phys, len, ro, > va, e, f); > > > + MemoryRegion *tmp_mr; > > > + *va = gpa2hva(&tmp_mr, phys, len, NULL); > > > > Use: dma_memory_map() > > > > > > That doesn't seem to be a precisely equivalent operation. It also > says > > in its headerdoc, > > > > Use only for reads OR writes - not for read-modify-write > operations. > > > > which I don't think we can guarantee here at all. > > > > I guess I can call it twice, once for writing and once for > reading, but > > given that the dma_memory_unmap operation marks the area dirty, > I'm not > > it's intended for what I understand the use case here to be: As > far as I > > can tell, the PV graphics device uses (some) of this memory to > exchange > > data in a cache-coherent way between host and guest, e.g. as a > lock-free > > ring buffer, using atomic operations as necessary. (This works > because > > it's a PV device: it "knows" the other end just another CPU core (or > > even the same one) executing in a different hypervisor context.) > This > > doesn't really match "traditional" DMA patterns where there's > either a > > read or a write happening. > > I think the story is a bit different for this VMApple variant. Probably > the CPU and GPU in Apple Silicon is cache-coherent so you can map > normal > memory for GPU without any kind of cache maintenance. > > Cache conherency of CPU and GPU in Apple Silicon is implied with > Apple's > documentation; it says you don't need to synchronize resources for > MTLStorageModeShared, which is the default for Apple Silicon. > https://developer.apple.com/documentation/metal/ > resource_fundamentals/synchronizing_a_managed_resource_in_macos > <https://developer.apple.com/documentation/metal/ > resource_fundamentals/synchronizing_a_managed_resource_in_macos> > > The name "IOSurface" also implies it is used not only for e.g., ring > buffer but also for real data. > > Note that the PGTask map/unmap callbacks appear to have equivalent > semantics, so it's not just the surface mapping. > > > > > Hunting around some more for alternative APIs, there's also > > memory_region_get_ram_ptr(), but I'm not sure its restrictions apply > > here either. > > I think you can call memory_access_is_direct() to check if the > requirement is satisfied. > > It will still break dirty page tracking implemented by > dma_memory_unmap() and others, but it's broken for hvf, which does not > implement dirty page tracking either. > > > > > > > > > + return (bool)true; > > > > Why cast? > > > > > > Good question. Not originally my code, so I've fixed all the > instances I > > could find now. > > > OK, it turns out the reason for this is that C treats 'true' as an int, > which then becomes the block's inferred return type - and the callbacks > are expecting bool-returning blocks. > > I've fixed it by explicitly specifying the block return type and > removing the cast in the return statement: > > iosfc_desc.unmapMemory = > ^bool(…) { > … > return true; > }; > > > > + > > > + iosfc_desc.unmapMemory = > > > + ^(void *a, void *b, void *c, void *d, void *e, > void *f) { > > > + trace_apple_iosfc_unmap_memory(a, b, c, d, e, f); > > > + return (bool)true; > > > + }; > > > + > > > + iosfc_desc.raiseInterrupt = ^(uint32_t vector) { > > > + trace_apple_iosfc_raise_irq(vector); > > > + bool locked = bql_locked(); > > > + if (!locked) { > > > + bql_lock(); > > > + } > > > + qemu_irq_pulse(s->irq_iosfc);> + if (! > locked) { > > > + bql_unlock(); > > > + } > > > + return (bool)true; > > > + }; > > > + > > > + iosfc_host_dev = > > > + [[PGIOSurfaceHostDevice alloc] > > initWithDescriptor:iosfc_desc]; > > > + [iosfc_desc release]; > > > + return iosfc_host_dev; > > > +} > > > + > > > +static void apple_gfx_vmapple_realize(DeviceState *dev, Error > > **errp) > > > +{ > > > + @autoreleasepool { > > > > This autoreleasepool is not used. > > > > > > It is definitely used inside the apple_gfx_common_realize() call. > It's > > also impossible to say whether [PGDeviceDescriptor new] uses > autorelease > > semantics internally, so it seemed safer to wrap the whole thing > in an > > outer pool. > > Theoretically, It should be safe to assume the callee creates > autoreleasepool by themselves as needed in general. We have bunch of > code to call Objective-C APIs without creating autoreleasepool in the > caller. Practically, [PGDeviceDescriptor new] is likely to be > implemented with ARC, which wraps methods in autoreleasepool as > necessary. > > > As far as I'm aware, ARC does NOT automatically insert autorelease pool > blocks. The reason you rarely need to create autoreleasepool blocks in > "plain" Objective-C programming is that Cocoa/CFRunloop/libdispatch > event handlers run each event in an autoreleasepool. So you don't need > to create them explicitly when using dispatch_async and similar, or when > running code on the main thread (which runs inside NSApplicationMain/ > CFRunloopRun/dispatch_main). My statement regarding ARC was wrong; It seems ARC just retains a value annoted as autoreleased. > > As far as I'm aware, if you don't explicitly define autoreleasepools in > raw threads created with the pthreads API, any autoreleased objects will > leak. At least I've not found any specification/documentation > contradicting this. And most code in Qemu runs on such raw threads, so > we need to play it safe with regard to autorelease semantics. I decided to dig deeper and found this documentation: https://developer.apple.com/library/archive/documentation/Cocoa/Conceptual/MemoryMgmt/Articles/mmAutoreleasePools.html It says: > Cocoa always expects code to be executed within an autorelease pool > block, otherwise autoreleased objects do not get released and your > application leaks memory. So yes, we must wrap everything in @autoreleasepool at least for Cocoa. It is probably a good practice to wrap all Objective-C code in @autoreleasepool. > > Whether the existing Qemu Objective-C code is safe in this regard I > don't know for certain, but I've certainly paid attention to this aspect > when modifying ui/cocoa.m in the past, and indeed most of that code runs > on the main thread. Note also how I wrap the apple_gfx_render_new_frame > call in a pool when it can't be guaranteed it's running on a dispatch > queue because the command buffer inside that uses autorelease semantics. It is more about event loop rather than thread. Resources allocated before [NSApp run] will leak even if they are autoreleased in the main thread. apple_gfx_common_realize() is one of such functions that run in the main thread before [NSApp run]. In ui/cocoa, cocoa_display_init() runs before [NSApp run]. Fortunately we already have NSAutoreleasePool for this function. > > Functions that uses a method that returns autorelease resources should > be wrapped with autoreleasepool instead of assuming the caller creates > autoreleasepool for them. > > > I'm treating apple_gfx_common_realize as an internal API, and I don't > think expecting its callers to wrap it in an autoreleasepool block is > unreasonable. I can certainly explicitly document this in a comment. We don't have a comment for cocoa_display_init() and it's more about generic macOS programming so it's not necessary. > > > > diff --git a/hw/display/apple-gfx.m b/hw/display/apple-gfx.m > > > new file mode 100644 > > > index 00000000000..837300f9cd4 > > > --- /dev/null > > > +++ b/hw/display/apple-gfx.m > > > @@ -0,0 +1,536 @@ > > > +/* > > > + * QEMU Apple ParavirtualizedGraphics.framework device > > > + * > > > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All > > Rights Reserved. > > > + * > > > + * This work is licensed under the terms of the GNU GPL, > version > > 2 or later. > > > + * See the COPYING file in the top-level directory. > > > + * > > > + * ParavirtualizedGraphics.framework is a set of > libraries that > > macOS provides > > > + * which implements 3d graphics passthrough to the host > as well as a > > > + * proprietary guest communication channel to drive it. This > > device model > > > + * implements support to drive that library from within QEMU. > > > + */ > > > + > > > +#include "apple-gfx.h" > > > +#include "trace.h" > > > +#include "qemu/main-loop.h" > > > +#include "ui/console.h" > > > +#include "monitor/monitor.h" > > > +#include "qapi/error.h" > > > +#include "migration/blocker.h" > > > +#include <mach/mach_vm.h> > > > +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > > + > > > +static const PGDisplayCoord_t apple_gfx_modes[] = { > > > + { .x = 1440, .y = 1080 }, > > > + { .x = 1280, .y = 1024 }, > > > +}; > > > + > > > +typedef struct PGTask_s { // Name matches forward > declaration in > > PG header > > > > Let's name it AppleGFXTask. It is a common practice to have > the same > > tag > > name and typedef in QEMU. > > > > > > This is defining a forward-declared type from framework headers > which is > > opaque from the framework's point of view. We do not get to > choose its > > struct name. The alternative is having casts wherever these > objects are > > being passed between our code and the framework. (See the > original v1/v2 > > vmapple patch series for how messy this is.) > > I got the idea. Let's not avoid the typedef then to clarify the naming > is not under our control. > > > I'm not sure what you mean by this double negative. Are you saying, > don't add our own typedef for struct PGTask_s at all, just use the > framework-supplied PGTask_t where appropriate? Yes. > > > > > > > +static void > apple_gfx_render_frame_completed(AppleGFXState *s, > > void *vram, > > > + > id<MTLTexture> texture) > > > +{ > > > + --s->pending_frames; > > > + assert(s->pending_frames >= 0); > > > + > > > + if (vram != s->vram) { > > > + /* Display mode has changed, drop this old frame. */ > > > + assert(texture != s->texture); > > > + g_free(vram); > > > > This management of buffers looks a bit convoluted. I suggest > > remembering > > the width and height instead of pointers and comparing them. > This way > > you can free resources in set_mode(). > > > > > > Yeah, I suppose that works, I can change that around. > > > > > + } else { > > > + copy_mtl_texture_to_surface_mem(texture, vram); > > > > Writing vram outside BQL may result in tearing. > > > > > > As far as I can tell(*), QXL does the same. I couldn't find any > examples > > of double-buffering in any of the existing display devices, which > would > > be the only way to do async updates efficiently and without > tearing. In > > any case, this solution is still vastly better than a regular VGA > > device, which suffers from very visible tearing with macOS on the > guest > > side anyway. And in an ideal world, we'd pass through the rendered > > texture directly to the Cocoa UI code rather than copying out > only for > > the CPU to draw it back into a window surface which is then > passed to > > the GPU for host side rendering. But I felt this patch is already > very, > > very large, and if anyone cares, we can fix imperfections in > subsequent > > updates. > > > > (*)The rendering code in that device is also fairly complex, so I > may be > > misreading it. > > QXL always modifies the surface with BQL. The surface is modified with > qxl_blit(), which is called by qxl_render_update_area_unlocked(). > qxl_render_update_area_unlocked() is called by either of > qxl_render_update() and qxl_render_update_area_bh(). Both of them are > called with BQL. The name includes "unlocked", but it means it is > called > without holding QXL-internal lock. > > Most devices works entirely with BQL so they don't perform double > buffering. apple-gfx can do the same. > > > I think we can safely move apple-gfx's framebuffer state management back > inside the BQL, yes. I just figured that copying dozens of megabytes of > framebuffer data on every frame while holding the lock was not going to > help BQL contention. Especially as PVG does not have a concept of dirty > areas, so we must copy the whole framebuffer every time. (Unless we were > to implement dirty area detection ourselves.) > > Unfortunately, implementing double-buffering would require a major > rework of Qemu's whole surface management, console code, and probably > most of the UI implementations. I'm guessing the OpenGL fast-path > sidesteps all of this, so replicating that with Metal would probably be > the easier way forward. (Although doing all this graphics stuff inside > the BQL generally seems like a major architectural flaw; I suppose most > enterprise use of Qemu does not involve the framebuffer, so it's not > shown up in BQL contention profiling there. It certainly does in desktop > use, although at least on macOS hosts there are far worse culprits in > that regard.) We don't need double-buffering; instead we can have one thread that renders the UI and also [PGDisplay -encodeCurrentFrameToCommandBuffer:texture:region:]. Rendering the UI in machine event loop in one thread is a bad idea in the first place so it will make sense to extract all UI work into a separate thread. This still requires a lot of work because the UI code assumes BQL everywhere. In my understanding BQL is less problematic for KVM users because QEMU can (sometimes) avoid locking BQL in the vCPU threads. > > > > > > + if (s->gfx_update_requested) { > > > + s->gfx_update_requested = false; > > > + dpy_gfx_update_full(s->con); > > > + graphic_hw_update_done(s->con); > > > + s->new_frame_ready = false; > > > > This assignment is unnecessary as s->new_frame_ready is > always false > > when s->gfx_update_requested. If you want to make sure > > s->gfx_update_requested and s->new_frame_ready are mutually > exclusive, > > use one enum value instead of having two bools. > > > > > > I'll need to refresh my memory and get back to you on this one, it's > > been so many months since I actively worked on this code. > > > > > + } else { > > > + s->new_frame_ready = true; > > > + } > > > + } > > > + if (s->pending_frames > 0) { > > > + apple_gfx_render_new_frame(s); > > > + } > > > +} > > > + > > > +static void apple_gfx_fb_update_display(void *opaque) > > > +{ > > > + AppleGFXState *s = opaque; > > > + > > > + dispatch_async(s->render_queue, ^{ > > > + if (s->pending_frames > 0) { > > > > It should check for s->new_frame_ready as > > apple_gfx_render_frame_completed() doesn't check if > > s->pending_frames > 0 before calling > graphic_hw_update_done(), which is > > inconsistent. > > > > > > pending_frames is about guest-side frames that are queued to be > rendered > > by the host GPU. > > new_frame_ready being true indicates that the contents of the Qemu > > console surface has been updated with new frame data since the last > > gfx_update. > > gfx_update_requested indicates that gfx_update is currently > awaiting an > > async completion (graphic_hw_update_done) but the surface has not > > received a new frame content, but the GPU is stily busy drawing one. > > > > apple_gfx_render_frame_completed is scheduled exactly once per > pending > > frame, so pending_frames > 0 is an invariant there. (Hence the > assert.)> > I don't think there is any inconsistency here, but I'll > double check. > > It's possible that there's an easier way to express the state > machine, > > and I'll take a look at that. > > I meant that apple_gfx_render_frame_completed() does not check if the > frame is the last one currently pending. apple_gfx_fb_update_display() > ignores a new ready frame when there is a more pending frame, but > apple_gfx_render_frame_completed() unconditionally fires > graphic_hw_update_done() even if there is a more pending frame. And I > think apple_gfx_render_frame_completed() is right and > apple_gfx_fb_update_display() is wrong in such a situation. > > > OK, got it. And yes, I agree. > > > > > Checking if s->pending_frames > 0 both in > apple_gfx_fb_update_display() > > and apple_gfx_render_frame_completed() is also problematic as > that can > > defer graphic_hw_update_done() indefinitely if we are getting new > > frames > > too fast. > > > > > > I see what you mean about this part. I'll have to test it, but I > guess > > we should reverse the priority, like this: > > > > if (s->new_frame_ready) { > > dpy_gfx_update_full(s->con); > > s->new_frame_ready = false; > > graphic_hw_update_done(s->con); > > } else if (s->pending_frames > 0) { > > s->gfx_update_requested = true; > > } else { > > graphic_hw_update_done(s->con); > > } > > > > 1. If we already have a frame, ready to be displayed, return it > immediately. > > 2. If the guest has reported that it's completed a frame and the > GPU is > > currently busy rendering it, defer graphic_hw_update_done until > it's done. > > 3. If the guest reports no changes to its display, indicate this > back to > > Qemu as a no-op display update graphic_hw_update_done() with no > > dpy_gfx_update* call. > > Yes, that looks correct. > > > > > > + s->gfx_update_requested = true; > > > + } else { > > > + if (s->new_frame_ready) { > > > + dpy_gfx_update_full(s->con); > > > + s->new_frame_ready = false; > > > + } > > > + graphic_hw_update_done(s->con); > > > + }> + }); > > > +} > > > + > > > +static const GraphicHwOps apple_gfx_fb_ops = { > > > + .gfx_update = apple_gfx_fb_update_display, > > > + .gfx_update_async = true, > > > +}; > > > + > > > +static void update_cursor(AppleGFXState *s) > > > +{ > > > + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, > > > + s->pgdisp.cursorPosition.y, s- > >cursor_show); > > > +} > > > + > > > +static void set_mode(AppleGFXState *s, uint32_t width, > uint32_t > > height) > > > +{ > > > + void *vram = NULL; > > > + DisplaySurface *surface; > > > + MTLTextureDescriptor *textureDescriptor; > > > + id<MTLTexture> texture = nil; > > > + __block bool no_change = false; > > > + > > > + dispatch_sync(s->render_queue, > > > > Calling dispatch_sync() while holding BQL may result in deadlock. > > > > Only if any code executed on the same dispatch queue acquires the > lock > > either directly or transitively. I believe I have ensure this is not > > done on the reqnder_queue, have you found anywhere this is the case? > > The documentation is not clear what threads a dispatch queue runs > on. We > can have a deadlock if they lock the BQL. > > > dispatch_sync is a synchronisation primitive (it waits for and asserts > exclusive access to the given queue), it doesn't actually do any thread > scheduling. Work scheduled asynchronously to non-main dispatch queues > will otherwise run on libdispatch pool threads. Running blocks on > dispatch queues will not preempt and schedule it on other threads which > may or may not be holding some locks. What if all pool threads are waiting for BQL? > > So the only way this code will deadlock is if any code scheduled to > render_queue directly or transitively acquires the BQL. None of it does, > although updating the console while holding the BQL rather complicates this. > > > > > > + ^{ > > > + if (s->surface && > > > + width == surface_width(s->surface) && > > > + height == surface_height(s->surface)) { > > > + no_change = true; > > > + } > > > + }); > > > + > > > + if (no_change) { > > > + return; > > > + } > > > + > > > + vram = g_malloc0(width * height * 4); > > > + surface = qemu_create_displaysurface_from(width, height, > > PIXMAN_LE_a8r8g8b8, > > > + width * 4, > vram); > > > + > > > + @autoreleasepool { > > > + textureDescriptor = > > > + [MTLTextureDescriptor > > > + > > texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm > > > + width:width > > > + height:height > > > + mipmapped:NO]; > > > + textureDescriptor.usage = s- > >pgdisp.minimumTextureUsage; > > > + texture = [s->mtl > > newTextureWithDescriptor:textureDescriptor]; > > > + } > > > + > > > + s->using_managed_texture_storage = > > > + (texture.storageMode == MTLStorageModeManaged); > > > + > > > + dispatch_sync(s->render_queue, > > > + ^{ > > > + id<MTLTexture> old_texture = nil; > > > + void *old_vram = s->vram; > > > + s->vram = vram; > > > + s->surface = surface; > > > + > > > + dpy_gfx_replace_surface(s->con, surface); > > > + > > > + old_texture = s->texture; > > > + s->texture = texture; > > > + [old_texture release]; > > > > You can just do: > > [s->texture release]; > > s->texture = texture; > > > > This will make s->texture dangling between the two > statements, but that > > don't matter since s->texture is not an atomic variable that > can be > > safely observed from another thread anyway. > > > > > + > > > + if (s->pending_frames == 0) { > > > + g_free(old_vram); > > > + } > > > + }); > > > +} > > > + > > > +static void create_fb(AppleGFXState *s) > > > +{ > > > + s->con = graphic_console_init(NULL, 0, > &apple_gfx_fb_ops, s); > > > + set_mode(s, 1440, 1080); > > > + > > > + s->cursor_show = true; > > > +} > > > + > > > +static size_t apple_gfx_get_default_mmio_range_size(void) > > > +{ > > > + size_t mmio_range_size; > > > + @autoreleasepool { > > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; > > > + mmio_range_size = desc.mmioLength; > > > + [desc release]; > > > + } > > > + return mmio_range_size; > > > +} > > > + > > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, > const > > char* obj_name) > > > > This function can be merged into apple_gfx_common_realize(). > > > > > > Probably. I'll try it. > > > Upon further inspection, we need to call > cocoa_enable_runloop_on_main_thread() during the init phase, not > realize(). So we can't get rid of this entirely. Is there any value in > moving the other init code into _realize()? Calling cocoa_enable_runloop_on_main_thread() should be avoided even in apple_gfx_common_init(). QEMU can plug a device at runtime instead of initialization time, and in such a case, apple_gfx_common_init() will run after calling qemu_main. I had a closer look and found it has a memory_region_init_io() call, which should remain in apple_gfx_common_init(). This leads to the same conclusion that we cannot remove this function so let's only move migrate_add_blocker() to apple_gfx_common_realize() to report its error. > > > > +{ > > > + Error *local_err = NULL; > > > + int r; > > > + size_t mmio_range_size = > > apple_gfx_get_default_mmio_range_size(); > > > + > > > + trace_apple_gfx_common_init(obj_name, mmio_range_size); > > > + memory_region_init_io(&s->iomem_gfx, obj, > &apple_gfx_ops, s, > > obj_name, > > > + mmio_range_size); > > > + s->iomem_gfx.disable_reentrancy_guard = true; > > > > Why do you disable reentrancy guard? > > > > > > Perhaps with the proposed AIO_WAIT_WHILE based I/O scheme, this > won't be > > an issue anymore, but the guard would otherwise keep dropping MMIOs > > which immediately caused the PV graphics device to stop making > progress. > > The MMIO APIs in the PVG framework are thread- and reentrancy- > safe, so > > we certainly don't need to serialise them on our side. > > It's better to understand why such reentrancy happens. Reentrancy > itself > is often a sign of bug. > > > > > > + > > > + /* TODO: PVG framework supports serialising device state: > > integrate it! */ > > > + if (apple_gfx_mig_blocker == NULL) { > > > + error_setg(&apple_gfx_mig_blocker, > > > + "Migration state blocked by apple-gfx > display > > device"); > > > + r = migrate_add_blocker(&apple_gfx_mig_blocker, > &local_err); > > > + if (r < 0) { > > > + error_report_err(local_err); > > > > Please report the error to the caller of > apple_gfx_common_realize() > > instead. > > > > > + } > > > + } > > > +}> + > > > +static void > > apple_gfx_register_task_mapping_handlers(AppleGFXState *s, > > > + > > PGDeviceDescriptor *desc) > > > +{ > > > + desc.createTask = ^(uint64_t vmSize, void * _Nullable * > > _Nonnull baseAddress) { > > > + AppleGFXTask *task = apple_gfx_new_task(s, vmSize); > > > + *baseAddress = (void*)task->address; > > > > nit: please write as (void *) instead of (void*). > > > > > + trace_apple_gfx_create_task(vmSize, *baseAddress); > > > + return task; > > > + }; > > > + > > > > > +{ > > > + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor > new]; > > > + > > > + disp_desc.name <http://disp_desc.name> <http:// > disp_desc.name <http://disp_desc.name>> = @"QEMU display"; > > > + disp_desc.sizeInMillimeters = NSMakeSize(400., > 300.); /* A > > 20" display */ > > > + disp_desc.queue = dispatch_get_main_queue(); > > > + disp_desc.newFrameEventHandler = ^(void) { > > > + trace_apple_gfx_new_frame(); > > > + dispatch_async(s->render_queue, ^{ > > > + /* Drop frames if we get too far ahead. */ > > > + if (s->pending_frames >= 2) > > > + return; > > > + ++s->pending_frames; > > > + if (s->pending_frames > 1) { > > > + return; > > > + } > > > + @autoreleasepool { > > > + apple_gfx_render_new_frame(s); > > > + } > > > + }); > > > + }; > > > + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t > sizeInPixels, > > > + OSType pixelFormat) { > > > + trace_apple_gfx_mode_change(sizeInPixels.x, > sizeInPixels.y); > > > + set_mode(s, sizeInPixels.x, sizeInPixels.y); > > > + }; > > > + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep *glyph, > > > + PGDisplayCoord_t > hotSpot) { > > > + uint32_t bpp = glyph.bitsPerPixel; > > > + size_t width = glyph.pixelsWide; > > > + size_t height = glyph.pixelsHigh; > > > + size_t padding_bytes_per_row = glyph.bytesPerRow > - width > > * 4; > > > + const uint8_t* px_data = glyph.bitmapData; > > > + > > > + trace_apple_gfx_cursor_set(bpp, width, height); > > > + > > > + if (s->cursor) { > > > + cursor_unref(s->cursor); > > > + s->cursor = NULL; > > > + } > > > + > > > + if (bpp == 32) { /* Shouldn't be anything else, > but just > > to be safe...*/ > > > + s->cursor = cursor_alloc(width, height); > > > + s->cursor->hot_x = hotSpot.x; > > > + s->cursor->hot_y = hotSpot.y; > > > + > > > + uint32_t *dest_px = s->cursor->data; > > > + > > > + for (size_t y = 0; y < height; ++y) { > > > + for (size_t x = 0; x < width; ++x) { > > > + /* NSBitmapImageRep's red & blue channels > > are swapped > > > + * compared to QEMUCursor's. */ > > > + *dest_px = > > > + (px_data[0] << 16u) | > > > + (px_data[1] << 8u) | > > > + (px_data[2] << 0u) | > > > + (px_data[3] << 24u); > > > + ++dest_px; > > > + px_data += 4; > > > + } > > > + px_data += padding_bytes_per_row; > > > + } > > > + dpy_cursor_define(s->con, s->cursor); > > > + update_cursor(s); > > > + } > > > + }; > > > + disp_desc.cursorShowHandler = ^(BOOL show) { > > > + trace_apple_gfx_cursor_show(show); > > > + s->cursor_show = show; > > > + update_cursor(s); > > > + }; > > > + disp_desc.cursorMoveHandler = ^(void) { > > > + trace_apple_gfx_cursor_move(); > > > + update_cursor(s); > > > + }; > > > + > > > + return disp_desc; > > > +} > > > + > > > +static NSArray<PGDisplayMode*>* > > apple_gfx_prepare_display_mode_array(void) > > > +{ > > > + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; > > > + NSArray<PGDisplayMode*>* mode_array = nil; > > > + int i; > > > + > > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > > + modes[i] = > > > + [[PGDisplayMode alloc] > > initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; > > > + } > > > + > > > + mode_array = [NSArray arrayWithObjects:modes > > count:ARRAY_SIZE(apple_gfx_modes)]; > > > + > > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > > + [modes[i] release]; > > > + modes[i] = nil; > > > + } > > > + > > > + return mode_array; > > > +} > > > + > > > +static id<MTLDevice> copy_suitable_metal_device(void) > > > +{ > > > + id<MTLDevice> dev = nil; > > > + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); > > > + > > > + /* Prefer a unified memory GPU. Failing that, pick a non- > > removable GPU. */ > > > + for (size_t i = 0; i < devs.count; ++i) { > > > + if (devs[i].hasUnifiedMemory) { > > > + dev = devs[i]; > > > + break; > > > + } > > > + if (!devs[i].removable) { > > > + dev = devs[i]; > > > + } > > > + } > > > + > > > + if (dev != nil) { > > > + [dev retain]; > > > + } else { > > > + dev = MTLCreateSystemDefaultDevice(); > > > + } > > > + [devs release]; > > > + > > > + return dev; > > > +} > > > + > > > +void apple_gfx_common_realize(AppleGFXState *s, > > PGDeviceDescriptor *desc) > > > +{ > > > + PGDisplayDescriptor *disp_desc = nil; > > > + > > > + QTAILQ_INIT(&s->tasks); > > > + s->render_queue = dispatch_queue_create("apple- > gfx.render", > > > + > DISPATCH_QUEUE_SERIAL); > > > + s->mtl = copy_suitable_metal_device(); > > > + s->mtl_queue = [s->mtl newCommandQueue]; > > > + > > > + desc.device = s->mtl; > > > + > > > + apple_gfx_register_task_mapping_handlers(s, desc); > > > + > > > + s->pgdev = PGNewDeviceWithDescriptor(desc); > > > + > > > + disp_desc = apple_gfx_prepare_display_handlers(s); > > > + s->pgdisp = [s->pgdev newDisplayWithDescriptor:disp_desc > > > + port:0 > > serialNum:1234]; > > > + [disp_desc release]; > > > + s->pgdisp.modeList = > apple_gfx_prepare_display_mode_array(); > > > + > > > + create_fb(s); > > > +} > > > diff --git a/hw/display/meson.build b/hw/display/meson.build > > > index 7db05eace97..70d855749c0 100644 > > > --- a/hw/display/meson.build > > > +++ b/hw/display/meson.build > > > @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', > if_true: > > files('artist.c')) > > > > > > system_ss.add(when: 'CONFIG_ATI_VGA', if_true: > [files('ati.c', > > 'ati_2d.c', 'ati_dbg.c'), pixman]) > > > > > > +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: > > [files('apple-gfx.m'), pvg, metal]) > > > +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: > > [files('apple-gfx-vmapple.m'), pvg, metal]) > > > > > > if config_all_devices.has_key('CONFIG_VIRTIO_GPU') > > > virtio_gpu_ss = ss.source_set() > > > diff --git a/hw/display/trace-events b/hw/display/trace-events > > > index 781f8a33203..1809a358e36 100644 > > > --- a/hw/display/trace-events > > > +++ b/hw/display/trace-events > > > @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) > > "dest_width : %u" > > > dm163_leds(int led, uint32_t value) "led %d: 0x%x" > > > dm163_channels(int channel, uint8_t value) "channel %d: > 0x%x" > > > dm163_refresh_rate(uint32_t rr) "refresh rate %d" > > > + > > > +# apple-gfx.m > > > +apple_gfx_read(uint64_t offset, uint64_t res) > > "offset=0x%"PRIx64" res=0x%"PRIx64 > > > +apple_gfx_write(uint64_t offset, uint64_t val) > > "offset=0x%"PRIx64" val=0x%"PRIx64 > > > +apple_gfx_create_task(uint32_t vm_size, void *va) > "vm_size=0x%x > > base_addr=%p" > > > +apple_gfx_destroy_task(void *task) "task=%p" > > > +apple_gfx_map_memory(void *task, uint32_t range_count, > uint64_t > > virtual_offset, uint32_t read_only) "task=%p range_count=0x%x > > virtual_offset=0x%"PRIx64" read_only=%d" > > > +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, > > uint64_t phys_len) "[%d] phys_addr=0x%"PRIx64" > phys_len=0x%"PRIx64 > > > +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t > > target) "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 > > > +apple_gfx_unmap_memory(void *task, uint64_t virtual_offset, > > uint64_t length) "task=%p virtual_offset=0x%"PRIx64" > length=0x%"PRIx64 > > > +apple_gfx_read_memory(uint64_t phys_address, uint64_t length, > > void *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" > > > +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" > > > +apple_gfx_new_frame(void) "" > > > +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" > > y=%"PRId64 > > > +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, uint64_t > > height) "bpp=%d width=%"PRId64" height=0x%"PRId64 > > > +apple_gfx_cursor_show(uint32_t show) "show=%d" > > > +apple_gfx_cursor_move(void) "" > > > +apple_gfx_common_init(const char *device_name, size_t > mmio_size) > > "device: %s; MMIO size: %zu bytes" > > > + > > > +# apple-gfx-vmapple.m > > > +apple_iosfc_read(uint64_t offset, uint64_t res) > > "offset=0x%"PRIx64" res=0x%"PRIx64 > > > +apple_iosfc_write(uint64_t offset, uint64_t val) > > "offset=0x%"PRIx64" val=0x%"PRIx64 > > > +apple_iosfc_map_memory(uint64_t phys, uint64_t len, > uint32_t ro, > > void *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" > ro=%d > > va=%p e=%p f=%p" > > > +apple_iosfc_unmap_memory(void *a, void *b, void *c, void *d, > > void *e, void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" > > > +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" > > > + > > > diff --git a/meson.build b/meson.build > > > index 10464466ff3..f09df3f09d5 100644 > > > --- a/meson.build > > > +++ b/meson.build > > > @@ -741,6 +741,8 @@ socket = [] > > > version_res = [] > > > coref = [] > > > iokit = [] > > > +pvg = [] > > > +metal = [] > > > emulator_link_args = [] > > > midl = not_found > > > widl = not_found > > > @@ -762,6 +764,8 @@ elif host_os == 'darwin' > > > coref = dependency('appleframeworks', modules: > 'CoreFoundation') > > > iokit = dependency('appleframeworks', modules: 'IOKit', > > required: false) > > > host_dsosuf = '.dylib' > > > + pvg = dependency('appleframeworks', modules: > > 'ParavirtualizedGraphics') > > > + metal = dependency('appleframeworks', modules: 'Metal') > > > elif host_os == 'sunos' > > > socket = [cc.find_library('socket'), > > > cc.find_library('nsl'), > > >
On Mon, 7 Oct 2024 at 10:25, Akihiko Odaki <akihiko.odaki@daynix.com> wrote: > > > Theoretically, It should be safe to assume the callee creates > > autoreleasepool by themselves as needed in general. We have bunch of > > code to call Objective-C APIs without creating autoreleasepool in the > > caller. Practically, [PGDeviceDescriptor new] is likely to be > > implemented with ARC, which wraps methods in autoreleasepool as > > necessary. > > > > > > As far as I'm aware, ARC does NOT automatically insert autorelease pool > > blocks. The reason you rarely need to create autoreleasepool blocks in > > "plain" Objective-C programming is that Cocoa/CFRunloop/libdispatch > > event handlers run each event in an autoreleasepool. So you don't need > > to create them explicitly when using dispatch_async and similar, or when > > running code on the main thread (which runs inside NSApplicationMain/ > > CFRunloopRun/dispatch_main). > > My statement regarding ARC was wrong; It seems ARC just retains a value > annoted as autoreleased. > Yes, the runtime maintains a sort of shadow stack of pointers to autoreleased objects for each thread. When you enter a pool block, it remembers the position in the shadow stack. On exiting the pool, all objects added since entering the pool are popped from the shadow stack and released. Hence, if objects are added to the stack outside of ANY autorelease pool, those are never popped and released. > > > > As far as I'm aware, if you don't explicitly define autoreleasepools in > > raw threads created with the pthreads API, any autoreleased objects will > > leak. At least I've not found any specification/documentation > > contradicting this. And most code in Qemu runs on such raw threads, so > > we need to play it safe with regard to autorelease semantics. > > I decided to dig deeper and found this documentation: > > https://developer.apple.com/library/archive/documentation/Cocoa/Conceptual/MemoryMgmt/Articles/mmAutoreleasePools.html > > It says: > > Cocoa always expects code to be executed within an autorelease pool > > block, otherwise autoreleased objects do not get released and your > > application leaks memory. > > So yes, we must wrap everything in @autoreleasepool at least for Cocoa. > It is probably a good practice to wrap all Objective-C code in > @autoreleasepool. > Yes; any method or function returning an Objective-C object pointer using autorelease semantics - either annotated or by naming convention - will use the pool. But unless it's all your own code all the way down, you don't know if anything inside a method whose own return value is not autoreleased might use autorelease semantics. (The naming convention says most methods return autoreleased, except those named create*, new*, or similar.) > Whether the existing Qemu Objective-C code is safe in this regard I > > don't know for certain, but I've certainly paid attention to this aspect > > when modifying ui/cocoa.m in the past, and indeed most of that code runs > > on the main thread. Note also how I wrap the apple_gfx_render_new_frame > > call in a pool when it can't be guaranteed it's running on a dispatch > > queue because the command buffer inside that uses autorelease semantics. > > It is more about event loop rather than thread. Resources allocated > before [NSApp run] will leak even if they are autoreleased in the main > thread. apple_gfx_common_realize() is one of such functions that run in > the main thread before [NSApp run]. In ui/cocoa, cocoa_display_init() > runs before [NSApp run]. Fortunately we already have NSAutoreleasePool > for this function. > Yes. Unless it's very clear that the code will run only in a Cocoa event handler or dispatch queue context, you need to use an autoreleasepool. > > > > Functions that uses a method that returns autorelease resources > should > > be wrapped with autoreleasepool instead of assuming the caller > creates > > autoreleasepool for them. > > > > > > I'm treating apple_gfx_common_realize as an internal API, and I don't > > think expecting its callers to wrap it in an autoreleasepool block is > > unreasonable. I can certainly explicitly document this in a comment. > > We don't have a comment for cocoa_display_init() and it's more about > generic macOS programming so it's not necessary. > > > > > > diff --git a/hw/display/apple-gfx.m > b/hw/display/apple-gfx.m > > > > new file mode 100644 > > > > index 00000000000..837300f9cd4 > > > > --- /dev/null > > > > +++ b/hw/display/apple-gfx.m > > > > @@ -0,0 +1,536 @@ > > > > +/* > > > > + * QEMU Apple ParavirtualizedGraphics.framework device > > > > + * > > > > + * Copyright © 2023 Amazon.com, Inc. or its affiliates. > All > > > Rights Reserved. > > > > + * > > > > + * This work is licensed under the terms of the GNU GPL, > > version > > > 2 or later. > > > > + * See the COPYING file in the top-level directory. > > > > + * > > > > + * ParavirtualizedGraphics.framework is a set of > > libraries that > > > macOS provides > > > > + * which implements 3d graphics passthrough to the host > > as well as a > > > > + * proprietary guest communication channel to drive it. > This > > > device model > > > > + * implements support to drive that library from within > QEMU. > > > > + */ > > > > + > > > > +#include "apple-gfx.h" > > > > +#include "trace.h" > > > > +#include "qemu/main-loop.h" > > > > +#include "ui/console.h" > > > > +#include "monitor/monitor.h" > > > > +#include "qapi/error.h" > > > > +#include "migration/blocker.h" > > > > +#include <mach/mach_vm.h> > > > > +#import > <ParavirtualizedGraphics/ParavirtualizedGraphics.h> > > > > + > > > > +static const PGDisplayCoord_t apple_gfx_modes[] = { > > > > + { .x = 1440, .y = 1080 }, > > > > + { .x = 1280, .y = 1024 }, > > > > +}; > > > > + > > > > +typedef struct PGTask_s { // Name matches forward > > declaration in > > > PG header > > > > > > Let's name it AppleGFXTask. It is a common practice to have > > the same > > > tag > > > name and typedef in QEMU. > > > > > > > > > This is defining a forward-declared type from framework headers > > which is > > > opaque from the framework's point of view. We do not get to > > choose its > > > struct name. The alternative is having casts wherever these > > objects are > > > being passed between our code and the framework. (See the > > original v1/v2 > > > vmapple patch series for how messy this is.) > > > > I got the idea. Let's not avoid the typedef then to clarify the > naming > > is not under our control. > > > > > > I'm not sure what you mean by this double negative. Are you saying, > > don't add our own typedef for struct PGTask_s at all, just use the > > framework-supplied PGTask_t where appropriate? > > Yes. > > > > > > > > > > > > +static void > > apple_gfx_render_frame_completed(AppleGFXState *s, > > > void *vram, > > > > + > > id<MTLTexture> texture) > > > > +{ > > > > + --s->pending_frames; > > > > + assert(s->pending_frames >= 0); > > > > + > > > > + if (vram != s->vram) { > > > > + /* Display mode has changed, drop this old frame. > */ > > > > + assert(texture != s->texture); > > > > + g_free(vram); > > > > > > This management of buffers looks a bit convoluted. I suggest > > > remembering > > > the width and height instead of pointers and comparing them. > > This way > > > you can free resources in set_mode(). > > > > > > > > > Yeah, I suppose that works, I can change that around. > > > > > > > + } else { > > > > + copy_mtl_texture_to_surface_mem(texture, vram); > > > > > > Writing vram outside BQL may result in tearing. > > > > > > > > > As far as I can tell(*), QXL does the same. I couldn't find any > > examples > > > of double-buffering in any of the existing display devices, which > > would > > > be the only way to do async updates efficiently and without > > tearing. In > > > any case, this solution is still vastly better than a regular VGA > > > device, which suffers from very visible tearing with macOS on the > > guest > > > side anyway. And in an ideal world, we'd pass through the rendered > > > texture directly to the Cocoa UI code rather than copying out > > only for > > > the CPU to draw it back into a window surface which is then > > passed to > > > the GPU for host side rendering. But I felt this patch is already > > very, > > > very large, and if anyone cares, we can fix imperfections in > > subsequent > > > updates. > > > > > > (*)The rendering code in that device is also fairly complex, so I > > may be > > > misreading it. > > > > QXL always modifies the surface with BQL. The surface is modified > with > > qxl_blit(), which is called by qxl_render_update_area_unlocked(). > > qxl_render_update_area_unlocked() is called by either of > > qxl_render_update() and qxl_render_update_area_bh(). Both of them are > > called with BQL. The name includes "unlocked", but it means it is > > called > > without holding QXL-internal lock. > > > > Most devices works entirely with BQL so they don't perform double > > buffering. apple-gfx can do the same. > > > > > > I think we can safely move apple-gfx's framebuffer state management back > > inside the BQL, yes. I just figured that copying dozens of megabytes of > > framebuffer data on every frame while holding the lock was not going to > > help BQL contention. Especially as PVG does not have a concept of dirty > > areas, so we must copy the whole framebuffer every time. (Unless we were > > to implement dirty area detection ourselves.) > > > > Unfortunately, implementing double-buffering would require a major > > rework of Qemu's whole surface management, console code, and probably > > most of the UI implementations. I'm guessing the OpenGL fast-path > > sidesteps all of this, so replicating that with Metal would probably be > > the easier way forward. (Although doing all this graphics stuff inside > > the BQL generally seems like a major architectural flaw; I suppose most > > enterprise use of Qemu does not involve the framebuffer, so it's not > > shown up in BQL contention profiling there. It certainly does in desktop > > use, although at least on macOS hosts there are far worse culprits in > > that regard.) > > We don't need double-buffering; instead we can have one thread that > renders the UI and also > [PGDisplay -encodeCurrentFrameToCommandBuffer:texture:region:]. > Rendering the UI in machine event loop in one thread is a bad idea in > the first place so it will make sense to extract all UI work into a > separate thread. This still requires a lot of work because the UI code > assumes BQL everywhere. > I'm not reworking the UI layer in this patch set, so it sounds like the best we can do is to acquire the BQL for updating the console framebuffer and just take the lock contention hit. > In my understanding BQL is less problematic for KVM users because QEMU > can (sometimes) avoid locking BQL in the vCPU threads. > Yeah, the current hvf implementation is unfortunately still very BQL heavy. Part of that is down to the interrupt controller: with the KVM accel, you typically use the interrupt controller provided by the kernel. With the hvf accel, you get a lot of VM exits for servicing interrupt controller I/O; at least on x86, the APIC implementation only works inside the BQL. HVF on x86-64 does technically have an in-kernel APIC implementation; unfortunately, it has some severe bugs which prevent using it in the general case. (EOI signalling to the IOAPIC does not work/FB14425590; the behaviour for coalescing and retriggering of level-triggered interrupts is also flawed/FB14425412.) On the aarch64 side, an in-kernel GIC implementation was only just added to hvf with macOS 15. I have submitted a patch that at least adds x2apic support to the hvf accel and thus significantly reduces BQL pressure a few times, but the maintainers of the x86 hvf backend don't seem to be active at the moment. > > > > > > > > > + if (s->gfx_update_requested) { > > > > + s->gfx_update_requested = false; > > > > + dpy_gfx_update_full(s->con); > > > > + graphic_hw_update_done(s->con); > > > > + s->new_frame_ready = false; > > > > > > This assignment is unnecessary as s->new_frame_ready is > > always false > > > when s->gfx_update_requested. If you want to make sure > > > s->gfx_update_requested and s->new_frame_ready are mutually > > exclusive, > > > use one enum value instead of having two bools. > > > > > > > > > I'll need to refresh my memory and get back to you on this one, > it's > > > been so many months since I actively worked on this code. > > > > > > > + } else { > > > > + s->new_frame_ready = true; > > > > + } > > > > + } > > > > + if (s->pending_frames > 0) { > > > > + apple_gfx_render_new_frame(s); > > > > + } > > > > +} > > > > + > > > > +static void apple_gfx_fb_update_display(void *opaque) > > > > +{ > > > > + AppleGFXState *s = opaque; > > > > + > > > > + dispatch_async(s->render_queue, ^{ > > > > + if (s->pending_frames > 0) { > > > > > > It should check for s->new_frame_ready as > > > apple_gfx_render_frame_completed() doesn't check if > > > s->pending_frames > 0 before calling > > graphic_hw_update_done(), which is > > > inconsistent. > > > > > > > > > pending_frames is about guest-side frames that are queued to be > > rendered > > > by the host GPU. > > > new_frame_ready being true indicates that the contents of the Qemu > > > console surface has been updated with new frame data since the > last > > > gfx_update. > > > gfx_update_requested indicates that gfx_update is currently > > awaiting an > > > async completion (graphic_hw_update_done) but the surface has not > > > received a new frame content, but the GPU is stily busy drawing > one. > > > > > > apple_gfx_render_frame_completed is scheduled exactly once per > > pending > > > frame, so pending_frames > 0 is an invariant there. (Hence the > > assert.)> > I don't think there is any inconsistency here, but I'll > > double check. > > > It's possible that there's an easier way to express the state > > machine, > > > and I'll take a look at that. > > > > I meant that apple_gfx_render_frame_completed() does not check if the > > frame is the last one currently pending. > apple_gfx_fb_update_display() > > ignores a new ready frame when there is a more pending frame, but > > apple_gfx_render_frame_completed() unconditionally fires > > graphic_hw_update_done() even if there is a more pending frame. And I > > think apple_gfx_render_frame_completed() is right and > > apple_gfx_fb_update_display() is wrong in such a situation. > > > > > > OK, got it. And yes, I agree. > > > > > > > > Checking if s->pending_frames > 0 both in > > apple_gfx_fb_update_display() > > > and apple_gfx_render_frame_completed() is also problematic as > > that can > > > defer graphic_hw_update_done() indefinitely if we are getting > new > > > frames > > > too fast. > > > > > > > > > I see what you mean about this part. I'll have to test it, but I > > guess > > > we should reverse the priority, like this: > > > > > > if (s->new_frame_ready) { > > > dpy_gfx_update_full(s->con); > > > s->new_frame_ready = false; > > > graphic_hw_update_done(s->con); > > > } else if (s->pending_frames > 0) { > > > s->gfx_update_requested = true; > > > } else { > > > graphic_hw_update_done(s->con); > > > } > > > > > > 1. If we already have a frame, ready to be displayed, return it > > immediately. > > > 2. If the guest has reported that it's completed a frame and the > > GPU is > > > currently busy rendering it, defer graphic_hw_update_done until > > it's done. > > > 3. If the guest reports no changes to its display, indicate this > > back to > > > Qemu as a no-op display update graphic_hw_update_done() with no > > > dpy_gfx_update* call. > > > > Yes, that looks correct. > > > > > > > > > + s->gfx_update_requested = true; > > > > + } else { > > > > + if (s->new_frame_ready) { > > > > + dpy_gfx_update_full(s->con); > > > > + s->new_frame_ready = false; > > > > + } > > > > + graphic_hw_update_done(s->con); > > > > + }> + }); > > > > +} > > > > + > > > > +static const GraphicHwOps apple_gfx_fb_ops = { > > > > + .gfx_update = apple_gfx_fb_update_display, > > > > + .gfx_update_async = true, > > > > +}; > > > > + > > > > +static void update_cursor(AppleGFXState *s) > > > > +{ > > > > + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, > > > > + s->pgdisp.cursorPosition.y, s- > > >cursor_show); > > > > +} > > > > + > > > > +static void set_mode(AppleGFXState *s, uint32_t width, > > uint32_t > > > height) > > > > +{ > > > > + void *vram = NULL; > > > > + DisplaySurface *surface; > > > > + MTLTextureDescriptor *textureDescriptor; > > > > + id<MTLTexture> texture = nil; > > > > + __block bool no_change = false; > > > > + > > > > + dispatch_sync(s->render_queue, > > > > > > Calling dispatch_sync() while holding BQL may result in > deadlock. > > > > > > Only if any code executed on the same dispatch queue acquires the > > lock > > > either directly or transitively. I believe I have ensure this is > not > > > done on the reqnder_queue, have you found anywhere this is the > case? > > > > The documentation is not clear what threads a dispatch queue runs > > on. We > > can have a deadlock if they lock the BQL. > > > > > > dispatch_sync is a synchronisation primitive (it waits for and asserts > > exclusive access to the given queue), it doesn't actually do any thread > > scheduling. Work scheduled asynchronously to non-main dispatch queues > > will otherwise run on libdispatch pool threads. Running blocks on > > dispatch queues will not preempt and schedule it on other threads which > > may or may not be holding some locks. > > What if all pool threads are waiting for BQL? > The pool is elastic, so if threads are blocked (in syscalls) and there are queued tasks ready to run, the runtime will create more threads as necessary. > > > > So the only way this code will deadlock is if any code scheduled to > > render_queue directly or transitively acquires the BQL. None of it does, > > although updating the console while holding the BQL rather complicates > this. > > > > > > > > > + ^{ > > > > + if (s->surface && > > > > + width == surface_width(s->surface) && > > > > + height == surface_height(s->surface)) { > > > > + no_change = true; > > > > + } > > > > + }); > > > > + > > > > + if (no_change) { > > > > + return; > > > > + } > > > > + > > > > + vram = g_malloc0(width * height * 4); > > > > + surface = qemu_create_displaysurface_from(width, > height, > > > PIXMAN_LE_a8r8g8b8, > > > > + width * 4, > > vram); > > > > + > > > > + @autoreleasepool { > > > > + textureDescriptor = > > > > + [MTLTextureDescriptor > > > > + > > > texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm > > > > + width:width > > > > + height:height > > > > + mipmapped:NO]; > > > > + textureDescriptor.usage = s- > > >pgdisp.minimumTextureUsage; > > > > + texture = [s->mtl > > > newTextureWithDescriptor:textureDescriptor]; > > > > + } > > > > + > > > > + s->using_managed_texture_storage = > > > > + (texture.storageMode == MTLStorageModeManaged); > > > > + > > > > + dispatch_sync(s->render_queue, > > > > + ^{ > > > > + id<MTLTexture> old_texture = nil; > > > > + void *old_vram = s->vram; > > > > + s->vram = vram; > > > > + s->surface = surface; > > > > + > > > > + dpy_gfx_replace_surface(s->con, surface); > > > > + > > > > + old_texture = s->texture; > > > > + s->texture = texture; > > > > + [old_texture release]; > > > > > > You can just do: > > > [s->texture release]; > > > s->texture = texture; > > > > > > This will make s->texture dangling between the two > > statements, but that > > > don't matter since s->texture is not an atomic variable that > > can be > > > safely observed from another thread anyway. > > > > > > > + > > > > + if (s->pending_frames == 0) { > > > > + g_free(old_vram); > > > > + } > > > > + }); > > > > +} > > > > + > > > > +static void create_fb(AppleGFXState *s) > > > > +{ > > > > + s->con = graphic_console_init(NULL, 0, > > &apple_gfx_fb_ops, s); > > > > + set_mode(s, 1440, 1080); > > > > + > > > > + s->cursor_show = true; > > > > +} > > > > + > > > > +static size_t apple_gfx_get_default_mmio_range_size(void) > > > > +{ > > > > + size_t mmio_range_size; > > > > + @autoreleasepool { > > > > + PGDeviceDescriptor *desc = [PGDeviceDescriptor > new]; > > > > + mmio_range_size = desc.mmioLength; > > > > + [desc release]; > > > > + } > > > > + return mmio_range_size; > > > > +} > > > > + > > > > +void apple_gfx_common_init(Object *obj, AppleGFXState *s, > > const > > > char* obj_name) > > > > > > This function can be merged into apple_gfx_common_realize(). > > > > > > > > > Probably. I'll try it. > > > > > > Upon further inspection, we need to call > > cocoa_enable_runloop_on_main_thread() during the init phase, not > > realize(). So we can't get rid of this entirely. Is there any value in > > moving the other init code into _realize()? > > Calling cocoa_enable_runloop_on_main_thread() should be avoided even in > apple_gfx_common_init(). QEMU can plug a device at runtime instead of > initialization time, and in such a case, apple_gfx_common_init() will > run after calling qemu_main. > Good point, it looks like I forgot to mark the mmio variant of apple-gfx devices as non-hotpluggable; I've done that now. I strongly suspect the guest driver doesn't support hotplugging anyway. > I had a closer look and found it has a memory_region_init_io() call, > which should remain in apple_gfx_common_init(). This leads to the same > conclusion that we cannot remove this function so let's only move > migrate_add_blocker() to apple_gfx_common_realize() to report its error. > Sounds good. > > > > > > +{ > > > > + Error *local_err = NULL; > > > > + int r; > > > > + size_t mmio_range_size = > > > apple_gfx_get_default_mmio_range_size(); > > > > + > > > > + trace_apple_gfx_common_init(obj_name, > mmio_range_size); > > > > + memory_region_init_io(&s->iomem_gfx, obj, > > &apple_gfx_ops, s, > > > obj_name, > > > > + mmio_range_size); > > > > + s->iomem_gfx.disable_reentrancy_guard = true; > > > > > > Why do you disable reentrancy guard? > > > > > > > > > Perhaps with the proposed AIO_WAIT_WHILE based I/O scheme, this > > won't be > > > an issue anymore, but the guard would otherwise keep dropping > MMIOs > > > which immediately caused the PV graphics device to stop making > > progress. > > > The MMIO APIs in the PVG framework are thread- and reentrancy- > > safe, so > > > we certainly don't need to serialise them on our side. > > > > It's better to understand why such reentrancy happens. Reentrancy > > itself > > is often a sign of bug. > > > > > > > > > + > > > > + /* TODO: PVG framework supports serialising device > state: > > > integrate it! */ > > > > + if (apple_gfx_mig_blocker == NULL) { > > > > + error_setg(&apple_gfx_mig_blocker, > > > > + "Migration state blocked by apple-gfx > > display > > > device"); > > > > + r = migrate_add_blocker(&apple_gfx_mig_blocker, > > &local_err); > > > > + if (r < 0) { > > > > + error_report_err(local_err); > > > > > > Please report the error to the caller of > > apple_gfx_common_realize() > > > instead. > > > > > > > + } > > > > + } > > > > +}> + > > > > +static void > > > apple_gfx_register_task_mapping_handlers(AppleGFXState *s, > > > > + > > > PGDeviceDescriptor *desc) > > > > +{ > > > > + desc.createTask = ^(uint64_t vmSize, void * _Nullable > * > > > _Nonnull baseAddress) { > > > > + AppleGFXTask *task = apple_gfx_new_task(s, > vmSize); > > > > + *baseAddress = (void*)task->address; > > > > > > nit: please write as (void *) instead of (void*). > > > > > > > + trace_apple_gfx_create_task(vmSize, *baseAddress); > > > > + return task; > > > > + }; > > > > + > > > > > > > +{ > > > > + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor > > new]; > > > > + > > > > + disp_desc.name <http://disp_desc.name> <http:// > > disp_desc.name <http://disp_desc.name>> = @"QEMU display"; > > > > + disp_desc.sizeInMillimeters = NSMakeSize(400., > > 300.); /* A > > > 20" display */ > > > > + disp_desc.queue = dispatch_get_main_queue(); > > > > + disp_desc.newFrameEventHandler = ^(void) { > > > > + trace_apple_gfx_new_frame(); > > > > + dispatch_async(s->render_queue, ^{ > > > > + /* Drop frames if we get too far ahead. */ > > > > + if (s->pending_frames >= 2) > > > > + return; > > > > + ++s->pending_frames; > > > > + if (s->pending_frames > 1) { > > > > + return; > > > > + } > > > > + @autoreleasepool { > > > > + apple_gfx_render_new_frame(s); > > > > + } > > > > + }); > > > > + }; > > > > + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t > > sizeInPixels, > > > > + OSType pixelFormat) { > > > > + trace_apple_gfx_mode_change(sizeInPixels.x, > > sizeInPixels.y); > > > > + set_mode(s, sizeInPixels.x, sizeInPixels.y); > > > > + }; > > > > + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep > *glyph, > > > > + PGDisplayCoord_t > > hotSpot) { > > > > + uint32_t bpp = glyph.bitsPerPixel; > > > > + size_t width = glyph.pixelsWide; > > > > + size_t height = glyph.pixelsHigh; > > > > + size_t padding_bytes_per_row = glyph.bytesPerRow > > - width > > > * 4; > > > > + const uint8_t* px_data = glyph.bitmapData; > > > > + > > > > + trace_apple_gfx_cursor_set(bpp, width, height); > > > > + > > > > + if (s->cursor) { > > > > + cursor_unref(s->cursor); > > > > + s->cursor = NULL; > > > > + } > > > > + > > > > + if (bpp == 32) { /* Shouldn't be anything else, > > but just > > > to be safe...*/ > > > > + s->cursor = cursor_alloc(width, height); > > > > + s->cursor->hot_x = hotSpot.x; > > > > + s->cursor->hot_y = hotSpot.y; > > > > + > > > > + uint32_t *dest_px = s->cursor->data; > > > > + > > > > + for (size_t y = 0; y < height; ++y) { > > > > + for (size_t x = 0; x < width; ++x) { > > > > + /* NSBitmapImageRep's red & blue > channels > > > are swapped > > > > + * compared to QEMUCursor's. */ > > > > + *dest_px = > > > > + (px_data[0] << 16u) | > > > > + (px_data[1] << 8u) | > > > > + (px_data[2] << 0u) | > > > > + (px_data[3] << 24u); > > > > + ++dest_px; > > > > + px_data += 4; > > > > + } > > > > + px_data += padding_bytes_per_row; > > > > + } > > > > + dpy_cursor_define(s->con, s->cursor); > > > > + update_cursor(s); > > > > + } > > > > + }; > > > > + disp_desc.cursorShowHandler = ^(BOOL show) { > > > > + trace_apple_gfx_cursor_show(show); > > > > + s->cursor_show = show; > > > > + update_cursor(s); > > > > + }; > > > > + disp_desc.cursorMoveHandler = ^(void) { > > > > + trace_apple_gfx_cursor_move(); > > > > + update_cursor(s); > > > > + }; > > > > + > > > > + return disp_desc; > > > > +} > > > > + > > > > +static NSArray<PGDisplayMode*>* > > > apple_gfx_prepare_display_mode_array(void) > > > > +{ > > > > + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; > > > > + NSArray<PGDisplayMode*>* mode_array = nil; > > > > + int i; > > > > + > > > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > > > + modes[i] = > > > > + [[PGDisplayMode alloc] > > > initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; > > > > + } > > > > + > > > > + mode_array = [NSArray arrayWithObjects:modes > > > count:ARRAY_SIZE(apple_gfx_modes)]; > > > > + > > > > + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { > > > > + [modes[i] release]; > > > > + modes[i] = nil; > > > > + } > > > > + > > > > + return mode_array; > > > > +} > > > > + > > > > +static id<MTLDevice> copy_suitable_metal_device(void) > > > > +{ > > > > + id<MTLDevice> dev = nil; > > > > + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); > > > > + > > > > + /* Prefer a unified memory GPU. Failing that, pick a > non- > > > removable GPU. */ > > > > + for (size_t i = 0; i < devs.count; ++i) { > > > > + if (devs[i].hasUnifiedMemory) { > > > > + dev = devs[i]; > > > > + break; > > > > + } > > > > + if (!devs[i].removable) { > > > > + dev = devs[i]; > > > > + } > > > > + } > > > > + > > > > + if (dev != nil) { > > > > + [dev retain]; > > > > + } else { > > > > + dev = MTLCreateSystemDefaultDevice(); > > > > + } > > > > + [devs release]; > > > > + > > > > + return dev; > > > > +} > > > > + > > > > +void apple_gfx_common_realize(AppleGFXState *s, > > > PGDeviceDescriptor *desc) > > > > +{ > > > > + PGDisplayDescriptor *disp_desc = nil; > > > > + > > > > + QTAILQ_INIT(&s->tasks); > > > > + s->render_queue = dispatch_queue_create("apple- > > gfx.render", > > > > + > > DISPATCH_QUEUE_SERIAL); > > > > + s->mtl = copy_suitable_metal_device(); > > > > + s->mtl_queue = [s->mtl newCommandQueue]; > > > > + > > > > + desc.device = s->mtl; > > > > + > > > > + apple_gfx_register_task_mapping_handlers(s, desc); > > > > + > > > > + s->pgdev = PGNewDeviceWithDescriptor(desc); > > > > + > > > > + disp_desc = apple_gfx_prepare_display_handlers(s); > > > > + s->pgdisp = [s->pgdev > newDisplayWithDescriptor:disp_desc > > > > + port:0 > > > serialNum:1234]; > > > > + [disp_desc release]; > > > > + s->pgdisp.modeList = > > apple_gfx_prepare_display_mode_array(); > > > > + > > > > + create_fb(s); > > > > +} > > > > diff --git a/hw/display/meson.build > b/hw/display/meson.build > > > > index 7db05eace97..70d855749c0 100644 > > > > --- a/hw/display/meson.build > > > > +++ b/hw/display/meson.build > > > > @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', > > if_true: > > > files('artist.c')) > > > > > > > > system_ss.add(when: 'CONFIG_ATI_VGA', if_true: > > [files('ati.c', > > > 'ati_2d.c', 'ati_dbg.c'), pixman]) > > > > > > > > +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: > > > [files('apple-gfx.m'), pvg, metal]) > > > > +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: > > > [files('apple-gfx-vmapple.m'), pvg, metal]) > > > > > > > > if config_all_devices.has_key('CONFIG_VIRTIO_GPU') > > > > virtio_gpu_ss = ss.source_set() > > > > diff --git a/hw/display/trace-events > b/hw/display/trace-events > > > > index 781f8a33203..1809a358e36 100644 > > > > --- a/hw/display/trace-events > > > > +++ b/hw/display/trace-events > > > > @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) > > > "dest_width : %u" > > > > dm163_leds(int led, uint32_t value) "led %d: 0x%x" > > > > dm163_channels(int channel, uint8_t value) "channel %d: > > 0x%x" > > > > dm163_refresh_rate(uint32_t rr) "refresh rate %d" > > > > + > > > > +# apple-gfx.m > > > > +apple_gfx_read(uint64_t offset, uint64_t res) > > > "offset=0x%"PRIx64" res=0x%"PRIx64 > > > > +apple_gfx_write(uint64_t offset, uint64_t val) > > > "offset=0x%"PRIx64" val=0x%"PRIx64 > > > > +apple_gfx_create_task(uint32_t vm_size, void *va) > > "vm_size=0x%x > > > base_addr=%p" > > > > +apple_gfx_destroy_task(void *task) "task=%p" > > > > +apple_gfx_map_memory(void *task, uint32_t range_count, > > uint64_t > > > virtual_offset, uint32_t read_only) "task=%p range_count=0x%x > > > virtual_offset=0x%"PRIx64" read_only=%d" > > > > +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, > > > uint64_t phys_len) "[%d] phys_addr=0x%"PRIx64" > > phys_len=0x%"PRIx64 > > > > +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t > > > target) "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 > > > > +apple_gfx_unmap_memory(void *task, uint64_t > virtual_offset, > > > uint64_t length) "task=%p virtual_offset=0x%"PRIx64" > > length=0x%"PRIx64 > > > > +apple_gfx_read_memory(uint64_t phys_address, uint64_t > length, > > > void *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" > > > > +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" > > > > +apple_gfx_new_frame(void) "" > > > > +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" > > > y=%"PRId64 > > > > +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, > uint64_t > > > height) "bpp=%d width=%"PRId64" height=0x%"PRId64 > > > > +apple_gfx_cursor_show(uint32_t show) "show=%d" > > > > +apple_gfx_cursor_move(void) "" > > > > +apple_gfx_common_init(const char *device_name, size_t > > mmio_size) > > > "device: %s; MMIO size: %zu bytes" > > > > + > > > > +# apple-gfx-vmapple.m > > > > +apple_iosfc_read(uint64_t offset, uint64_t res) > > > "offset=0x%"PRIx64" res=0x%"PRIx64 > > > > +apple_iosfc_write(uint64_t offset, uint64_t val) > > > "offset=0x%"PRIx64" val=0x%"PRIx64 > > > > +apple_iosfc_map_memory(uint64_t phys, uint64_t len, > > uint32_t ro, > > > void *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" > > ro=%d > > > va=%p e=%p f=%p" > > > > +apple_iosfc_unmap_memory(void *a, void *b, void *c, void > *d, > > > void *e, void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" > > > > +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" > > > > + > > > > diff --git a/meson.build b/meson.build > > > > index 10464466ff3..f09df3f09d5 100644 > > > > --- a/meson.build > > > > +++ b/meson.build > > > > @@ -741,6 +741,8 @@ socket = [] > > > > version_res = [] > > > > coref = [] > > > > iokit = [] > > > > +pvg = [] > > > > +metal = [] > > > > emulator_link_args = [] > > > > midl = not_found > > > > widl = not_found > > > > @@ -762,6 +764,8 @@ elif host_os == 'darwin' > > > > coref = dependency('appleframeworks', modules: > > 'CoreFoundation') > > > > iokit = dependency('appleframeworks', modules: 'IOKit', > > > required: false) > > > > host_dsosuf = '.dylib' > > > > + pvg = dependency('appleframeworks', modules: > > > 'ParavirtualizedGraphics') > > > > + metal = dependency('appleframeworks', modules: 'Metal') > > > > elif host_os == 'sunos' > > > > socket = [cc.find_library('socket'), > > > > cc.find_library('nsl'), > > > > > > >
diff --git a/hw/display/Kconfig b/hw/display/Kconfig index a4552c8ed78..179a479d220 100644 --- a/hw/display/Kconfig +++ b/hw/display/Kconfig @@ -143,3 +143,12 @@ config XLNX_DISPLAYPORT config DM163 bool + +config MAC_PVG + bool + default y + +config MAC_PVG_VMAPPLE + bool + depends on MAC_PVG + depends on ARM diff --git a/hw/display/apple-gfx-vmapple.m b/hw/display/apple-gfx-vmapple.m new file mode 100644 index 00000000000..d8fc7651dde --- /dev/null +++ b/hw/display/apple-gfx-vmapple.m @@ -0,0 +1,215 @@ +/* + * QEMU Apple ParavirtualizedGraphics.framework device, vmapple (arm64) variant + * + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All Rights Reserved. + * + * This work is licensed under the terms of the GNU GPL, version 2 or later. + * See the COPYING file in the top-level directory. + * + * ParavirtualizedGraphics.framework is a set of libraries that macOS provides + * which implements 3d graphics passthrough to the host as well as a + * proprietary guest communication channel to drive it. This device model + * implements support to drive that library from within QEMU as an MMIO-based + * system device for macOS on arm64 VMs. + */ + +#include "apple-gfx.h" +#include "monitor/monitor.h" +#include "hw/sysbus.h" +#include "hw/irq.h" +#include "trace.h" +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> + +_Static_assert(__aarch64__, ""); + +/* + * ParavirtualizedGraphics.Framework only ships header files for the PCI + * variant which does not include IOSFC descriptors and host devices. We add + * their definitions here so that we can also work with the ARM version. + */ +typedef bool(^IOSFCRaiseInterrupt)(uint32_t vector); +typedef bool(^IOSFCUnmapMemory)( + void *a, void *b, void *c, void *d, void *e, void *f); +typedef bool(^IOSFCMapMemory)( + uint64_t phys, uint64_t len, bool ro, void **va, void *e, void *f); + +@interface PGDeviceDescriptor (IOSurfaceMapper) +@property (readwrite, nonatomic) bool usingIOSurfaceMapper; +@end + +@interface PGIOSurfaceHostDeviceDescriptor : NSObject +-(PGIOSurfaceHostDeviceDescriptor *)init; +@property (readwrite, nonatomic, copy, nullable) IOSFCMapMemory mapMemory; +@property (readwrite, nonatomic, copy, nullable) IOSFCUnmapMemory unmapMemory; +@property (readwrite, nonatomic, copy, nullable) IOSFCRaiseInterrupt raiseInterrupt; +@end + +@interface PGIOSurfaceHostDevice : NSObject +-(instancetype)initWithDescriptor:(PGIOSurfaceHostDeviceDescriptor *) desc; +-(uint32_t)mmioReadAtOffset:(size_t) offset; +-(void)mmioWriteAtOffset:(size_t) offset value:(uint32_t)value; +@end + +typedef struct AppleGFXVmappleState { + SysBusDevice parent_obj; + + AppleGFXState common; + + qemu_irq irq_gfx; + qemu_irq irq_iosfc; + MemoryRegion iomem_iosfc; + PGIOSurfaceHostDevice *pgiosfc; +} AppleGFXVmappleState; + +OBJECT_DECLARE_SIMPLE_TYPE(AppleGFXVmappleState, APPLE_GFX_VMAPPLE) + + +static uint64_t apple_iosfc_read(void *opaque, hwaddr offset, unsigned size) +{ + AppleGFXVmappleState *s = opaque; + uint64_t res = 0; + + bql_unlock(); + res = [s->pgiosfc mmioReadAtOffset:offset]; + bql_lock(); + + trace_apple_iosfc_read(offset, res); + + return res; +} + +static void apple_iosfc_write( + void *opaque, hwaddr offset, uint64_t val, unsigned size) +{ + AppleGFXVmappleState *s = opaque; + + trace_apple_iosfc_write(offset, val); + + [s->pgiosfc mmioWriteAtOffset:offset value:val]; +} + +static const MemoryRegionOps apple_iosfc_ops = { + .read = apple_iosfc_read, + .write = apple_iosfc_write, + .endianness = DEVICE_LITTLE_ENDIAN, + .valid = { + .min_access_size = 4, + .max_access_size = 8, + }, + .impl = { + .min_access_size = 4, + .max_access_size = 8, + }, +}; + +static PGIOSurfaceHostDevice *apple_gfx_prepare_iosurface_host_device( + AppleGFXVmappleState *s) +{ + PGIOSurfaceHostDeviceDescriptor *iosfc_desc = + [PGIOSurfaceHostDeviceDescriptor new]; + PGIOSurfaceHostDevice *iosfc_host_dev = nil; + + iosfc_desc.mapMemory = + ^(uint64_t phys, uint64_t len, bool ro, void **va, void *e, void *f) { + trace_apple_iosfc_map_memory(phys, len, ro, va, e, f); + MemoryRegion *tmp_mr; + *va = gpa2hva(&tmp_mr, phys, len, NULL); + return (bool)true; + }; + + iosfc_desc.unmapMemory = + ^(void *a, void *b, void *c, void *d, void *e, void *f) { + trace_apple_iosfc_unmap_memory(a, b, c, d, e, f); + return (bool)true; + }; + + iosfc_desc.raiseInterrupt = ^(uint32_t vector) { + trace_apple_iosfc_raise_irq(vector); + bool locked = bql_locked(); + if (!locked) { + bql_lock(); + } + qemu_irq_pulse(s->irq_iosfc); + if (!locked) { + bql_unlock(); + } + return (bool)true; + }; + + iosfc_host_dev = + [[PGIOSurfaceHostDevice alloc] initWithDescriptor:iosfc_desc]; + [iosfc_desc release]; + return iosfc_host_dev; +} + +static void apple_gfx_vmapple_realize(DeviceState *dev, Error **errp) +{ + @autoreleasepool { + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(dev); + + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; + desc.usingIOSurfaceMapper = true; + desc.raiseInterrupt = ^(uint32_t vector) { + bool locked; + + trace_apple_gfx_raise_irq(vector); + locked = bql_locked(); + if (!locked) { + bql_lock(); + } + qemu_irq_pulse(s->irq_gfx); + if (!locked) { + bql_unlock(); + } + }; + + s->pgiosfc = apple_gfx_prepare_iosurface_host_device(s); + + apple_gfx_common_realize(&s->common, desc); + [desc release]; + desc = nil; + } +} + +static void apple_gfx_vmapple_init(Object *obj) +{ + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(obj); + + apple_gfx_common_init(obj, &s->common, TYPE_APPLE_GFX_VMAPPLE); + + memory_region_init_io(&s->iomem_iosfc, obj, &apple_iosfc_ops, s, + TYPE_APPLE_GFX_VMAPPLE, 0x10000); + sysbus_init_mmio(SYS_BUS_DEVICE(s), &s->common.iomem_gfx); + sysbus_init_mmio(SYS_BUS_DEVICE(s), &s->iomem_iosfc); + sysbus_init_irq(SYS_BUS_DEVICE(s), &s->irq_gfx); + sysbus_init_irq(SYS_BUS_DEVICE(s), &s->irq_iosfc); +} + +static void apple_gfx_vmapple_reset(Object *obj, ResetType type) +{ + AppleGFXVmappleState *s = APPLE_GFX_VMAPPLE(obj); + [s->common.pgdev reset]; +} + + +static void apple_gfx_vmapple_class_init(ObjectClass *klass, void *data) +{ + DeviceClass *dc = DEVICE_CLASS(klass); + ResettableClass *rc = RESETTABLE_CLASS(klass); + + assert(rc->phases.hold == NULL); + rc->phases.hold = apple_gfx_vmapple_reset; + + dc->realize = apple_gfx_vmapple_realize; +} + +static TypeInfo apple_gfx_vmapple_types[] = { + { + .name = TYPE_APPLE_GFX_VMAPPLE, + .parent = TYPE_SYS_BUS_DEVICE, + .instance_size = sizeof(AppleGFXVmappleState), + .class_init = apple_gfx_vmapple_class_init, + .instance_init = apple_gfx_vmapple_init, + } +}; +DEFINE_TYPES(apple_gfx_vmapple_types) diff --git a/hw/display/apple-gfx.h b/hw/display/apple-gfx.h new file mode 100644 index 00000000000..995ecf7f4a7 --- /dev/null +++ b/hw/display/apple-gfx.h @@ -0,0 +1,57 @@ +#ifndef QEMU_APPLE_GFX_H +#define QEMU_APPLE_GFX_H + +#define TYPE_APPLE_GFX_VMAPPLE "apple-gfx-vmapple" +#define TYPE_APPLE_GFX_PCI "apple-gfx-pci" + +#include "qemu/typedefs.h" + +typedef struct AppleGFXState AppleGFXState; + +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const char* obj_name); + +#ifdef __OBJC__ + +#include "qemu/osdep.h" +#include "exec/memory.h" +#include "ui/surface.h" +#include <dispatch/dispatch.h> + +@class PGDeviceDescriptor; +@protocol PGDevice; +@protocol PGDisplay; +@protocol MTLDevice; +@protocol MTLTexture; +@protocol MTLCommandQueue; + +typedef QTAILQ_HEAD(, PGTask_s) AppleGFXTaskList; + +struct AppleGFXState { + MemoryRegion iomem_gfx; + id<PGDevice> pgdev; + id<PGDisplay> pgdisp; + AppleGFXTaskList tasks; + QemuConsole *con; + id<MTLDevice> mtl; + id<MTLCommandQueue> mtl_queue; + bool handles_frames; + bool new_frame; + bool cursor_show; + QEMUCursor *cursor; + + dispatch_queue_t render_queue; + /* The following fields should only be accessed from render_queue: */ + bool gfx_update_requested; + bool new_frame_ready; + bool using_managed_texture_storage; + int32_t pending_frames; + void *vram; + DisplaySurface *surface; + id<MTLTexture> texture; +}; + +void apple_gfx_common_realize(AppleGFXState *s, PGDeviceDescriptor *desc); + +#endif /* __OBJC__ */ + +#endif diff --git a/hw/display/apple-gfx.m b/hw/display/apple-gfx.m new file mode 100644 index 00000000000..837300f9cd4 --- /dev/null +++ b/hw/display/apple-gfx.m @@ -0,0 +1,536 @@ +/* + * QEMU Apple ParavirtualizedGraphics.framework device + * + * Copyright © 2023 Amazon.com, Inc. or its affiliates. All Rights Reserved. + * + * This work is licensed under the terms of the GNU GPL, version 2 or later. + * See the COPYING file in the top-level directory. + * + * ParavirtualizedGraphics.framework is a set of libraries that macOS provides + * which implements 3d graphics passthrough to the host as well as a + * proprietary guest communication channel to drive it. This device model + * implements support to drive that library from within QEMU. + */ + +#include "apple-gfx.h" +#include "trace.h" +#include "qemu/main-loop.h" +#include "ui/console.h" +#include "monitor/monitor.h" +#include "qapi/error.h" +#include "migration/blocker.h" +#include <mach/mach_vm.h> +#import <ParavirtualizedGraphics/ParavirtualizedGraphics.h> + +static const PGDisplayCoord_t apple_gfx_modes[] = { + { .x = 1440, .y = 1080 }, + { .x = 1280, .y = 1024 }, +}; + +typedef struct PGTask_s { // Name matches forward declaration in PG header + QTAILQ_ENTRY(PGTask_s) node; + mach_vm_address_t address; + uint64_t len; +} AppleGFXTask; + +static Error *apple_gfx_mig_blocker; + +static void apple_gfx_render_frame_completed(AppleGFXState *s, void *vram, + id<MTLTexture> texture); + +static AppleGFXTask *apple_gfx_new_task(AppleGFXState *s, uint64_t len) +{ + mach_vm_address_t task_mem; + AppleGFXTask *task; + kern_return_t r; + + r = mach_vm_allocate(mach_task_self(), &task_mem, len, VM_FLAGS_ANYWHERE); + if (r != KERN_SUCCESS || task_mem == 0) { + return NULL; + } + + task = g_new0(AppleGFXTask, 1); + + task->address = task_mem; + task->len = len; + QTAILQ_INSERT_TAIL(&s->tasks, task, node); + + return task; +} + +static uint64_t apple_gfx_read(void *opaque, hwaddr offset, unsigned size) +{ + AppleGFXState *s = opaque; + uint64_t res = 0; + + bql_unlock(); + res = [s->pgdev mmioReadAtOffset:offset]; + bql_lock(); + + trace_apple_gfx_read(offset, res); + + return res; +} + +static void apple_gfx_write(void *opaque, hwaddr offset, uint64_t val, + unsigned size) +{ + AppleGFXState *s = opaque; + + trace_apple_gfx_write(offset, val); + + bql_unlock(); + [s->pgdev mmioWriteAtOffset:offset value:val]; + bql_lock(); +} + +static const MemoryRegionOps apple_gfx_ops = { + .read = apple_gfx_read, + .write = apple_gfx_write, + .endianness = DEVICE_LITTLE_ENDIAN, + .valid = { + .min_access_size = 4, + .max_access_size = 8, + }, + .impl = { + .min_access_size = 4, + .max_access_size = 4, + }, +}; + +static void apple_gfx_render_new_frame(AppleGFXState *s) +{ + BOOL r; + void *vram = s->vram; + uint32_t width = surface_width(s->surface); + uint32_t height = surface_height(s->surface); + MTLRegion region = MTLRegionMake2D(0, 0, width, height); + id<MTLCommandBuffer> command_buffer = [s->mtl_queue commandBuffer]; + id<MTLTexture> texture = s->texture; + r = [s->pgdisp encodeCurrentFrameToCommandBuffer:command_buffer + texture:texture + region:region]; + if (!r) { + return; + } + [texture retain]; + if (s->using_managed_texture_storage) { + /* "Managed" textures exist in both VRAM and RAM and must be synced. */ + id<MTLBlitCommandEncoder> blit = [command_buffer blitCommandEncoder]; + [blit synchronizeResource:texture]; + [blit endEncoding]; + } + [command_buffer retain]; + [command_buffer addCompletedHandler: + ^(id<MTLCommandBuffer> cb) + { + dispatch_async(s->render_queue, ^{ + apple_gfx_render_frame_completed(s, vram, texture); + [texture release]; + }); + [command_buffer release]; + }]; + [command_buffer commit]; +} + +static void copy_mtl_texture_to_surface_mem(id<MTLTexture> texture, void *vram) +{ + /* TODO: Skip this entirely on a pure Metal or headless/guest-only + * rendering path, else use a blit command encoder? Needs careful + * (double?) buffering design. */ + size_t width = texture.width, height = texture.height; + MTLRegion region = MTLRegionMake2D(0, 0, width, height); + [texture getBytes:vram + bytesPerRow:(width * 4) + bytesPerImage:(width * height * 4) + fromRegion:region + mipmapLevel:0 + slice:0]; +} + +static void apple_gfx_render_frame_completed(AppleGFXState *s, void *vram, + id<MTLTexture> texture) +{ + --s->pending_frames; + assert(s->pending_frames >= 0); + + if (vram != s->vram) { + /* Display mode has changed, drop this old frame. */ + assert(texture != s->texture); + g_free(vram); + } else { + copy_mtl_texture_to_surface_mem(texture, vram); + if (s->gfx_update_requested) { + s->gfx_update_requested = false; + dpy_gfx_update_full(s->con); + graphic_hw_update_done(s->con); + s->new_frame_ready = false; + } else { + s->new_frame_ready = true; + } + } + if (s->pending_frames > 0) { + apple_gfx_render_new_frame(s); + } +} + +static void apple_gfx_fb_update_display(void *opaque) +{ + AppleGFXState *s = opaque; + + dispatch_async(s->render_queue, ^{ + if (s->pending_frames > 0) { + s->gfx_update_requested = true; + } else { + if (s->new_frame_ready) { + dpy_gfx_update_full(s->con); + s->new_frame_ready = false; + } + graphic_hw_update_done(s->con); + } + }); +} + +static const GraphicHwOps apple_gfx_fb_ops = { + .gfx_update = apple_gfx_fb_update_display, + .gfx_update_async = true, +}; + +static void update_cursor(AppleGFXState *s) +{ + dpy_mouse_set(s->con, s->pgdisp.cursorPosition.x, + s->pgdisp.cursorPosition.y, s->cursor_show); +} + +static void set_mode(AppleGFXState *s, uint32_t width, uint32_t height) +{ + void *vram = NULL; + DisplaySurface *surface; + MTLTextureDescriptor *textureDescriptor; + id<MTLTexture> texture = nil; + __block bool no_change = false; + + dispatch_sync(s->render_queue, + ^{ + if (s->surface && + width == surface_width(s->surface) && + height == surface_height(s->surface)) { + no_change = true; + } + }); + + if (no_change) { + return; + } + + vram = g_malloc0(width * height * 4); + surface = qemu_create_displaysurface_from(width, height, PIXMAN_LE_a8r8g8b8, + width * 4, vram); + + @autoreleasepool { + textureDescriptor = + [MTLTextureDescriptor + texture2DDescriptorWithPixelFormat:MTLPixelFormatBGRA8Unorm + width:width + height:height + mipmapped:NO]; + textureDescriptor.usage = s->pgdisp.minimumTextureUsage; + texture = [s->mtl newTextureWithDescriptor:textureDescriptor]; + } + + s->using_managed_texture_storage = + (texture.storageMode == MTLStorageModeManaged); + + dispatch_sync(s->render_queue, + ^{ + id<MTLTexture> old_texture = nil; + void *old_vram = s->vram; + s->vram = vram; + s->surface = surface; + + dpy_gfx_replace_surface(s->con, surface); + + old_texture = s->texture; + s->texture = texture; + [old_texture release]; + + if (s->pending_frames == 0) { + g_free(old_vram); + } + }); +} + +static void create_fb(AppleGFXState *s) +{ + s->con = graphic_console_init(NULL, 0, &apple_gfx_fb_ops, s); + set_mode(s, 1440, 1080); + + s->cursor_show = true; +} + +static size_t apple_gfx_get_default_mmio_range_size(void) +{ + size_t mmio_range_size; + @autoreleasepool { + PGDeviceDescriptor *desc = [PGDeviceDescriptor new]; + mmio_range_size = desc.mmioLength; + [desc release]; + } + return mmio_range_size; +} + +void apple_gfx_common_init(Object *obj, AppleGFXState *s, const char* obj_name) +{ + Error *local_err = NULL; + int r; + size_t mmio_range_size = apple_gfx_get_default_mmio_range_size(); + + trace_apple_gfx_common_init(obj_name, mmio_range_size); + memory_region_init_io(&s->iomem_gfx, obj, &apple_gfx_ops, s, obj_name, + mmio_range_size); + s->iomem_gfx.disable_reentrancy_guard = true; + + /* TODO: PVG framework supports serialising device state: integrate it! */ + if (apple_gfx_mig_blocker == NULL) { + error_setg(&apple_gfx_mig_blocker, + "Migration state blocked by apple-gfx display device"); + r = migrate_add_blocker(&apple_gfx_mig_blocker, &local_err); + if (r < 0) { + error_report_err(local_err); + } + } +} + +static void apple_gfx_register_task_mapping_handlers(AppleGFXState *s, + PGDeviceDescriptor *desc) +{ + desc.createTask = ^(uint64_t vmSize, void * _Nullable * _Nonnull baseAddress) { + AppleGFXTask *task = apple_gfx_new_task(s, vmSize); + *baseAddress = (void*)task->address; + trace_apple_gfx_create_task(vmSize, *baseAddress); + return task; + }; + + desc.destroyTask = ^(AppleGFXTask * _Nonnull task) { + trace_apple_gfx_destroy_task(task); + QTAILQ_REMOVE(&s->tasks, task, node); + mach_vm_deallocate(mach_task_self(), task->address, task->len); + g_free(task); + }; + + desc.mapMemory = ^(AppleGFXTask * _Nonnull task, uint32_t rangeCount, + uint64_t virtualOffset, bool readOnly, + PGPhysicalMemoryRange_t * _Nonnull ranges) { + kern_return_t r; + mach_vm_address_t target, source; + trace_apple_gfx_map_memory(task, rangeCount, virtualOffset, readOnly); + for (int i = 0; i < rangeCount; i++) { + PGPhysicalMemoryRange_t *range = &ranges[i]; + MemoryRegion *tmp_mr; + /* TODO: Bounds checks? r/o? */ + bql_lock(); + + trace_apple_gfx_map_memory_range(i, range->physicalAddress, + range->physicalLength); + + target = task->address + virtualOffset; + source = (mach_vm_address_t)gpa2hva(&tmp_mr, + range->physicalAddress, + range->physicalLength, NULL); + vm_prot_t cur_protection = 0; + vm_prot_t max_protection = 0; + // Map guest RAM at range->physicalAddress into PG task memory range + r = mach_vm_remap(mach_task_self(), + &target, range->physicalLength, vm_page_size - 1, + VM_FLAGS_FIXED | VM_FLAGS_OVERWRITE, + mach_task_self(), + source, false /* shared mapping, no copy */, + &cur_protection, &max_protection, + VM_INHERIT_COPY); + trace_apple_gfx_remap(r, source, target); + g_assert(r == KERN_SUCCESS); + + bql_unlock(); + + virtualOffset += range->physicalLength; + } + return (bool)true; + }; + + desc.unmapMemory = ^(AppleGFXTask * _Nonnull task, uint64_t virtualOffset, + uint64_t length) { + kern_return_t r; + mach_vm_address_t range_address; + + trace_apple_gfx_unmap_memory(task, virtualOffset, length); + + /* Replace task memory range with fresh pages, undoing the mapping + * from guest RAM. */ + range_address = task->address + virtualOffset; + r = mach_vm_allocate(mach_task_self(), &range_address, length, + VM_FLAGS_FIXED | VM_FLAGS_OVERWRITE); + g_assert(r == KERN_SUCCESS); + + return (bool)true; + }; + + desc.readMemory = ^(uint64_t physicalAddress, uint64_t length, + void * _Nonnull dst) { + trace_apple_gfx_read_memory(physicalAddress, length, dst); + cpu_physical_memory_read(physicalAddress, dst, length); + return (bool)true; + }; + +} + +static PGDisplayDescriptor *apple_gfx_prepare_display_handlers(AppleGFXState *s) +{ + PGDisplayDescriptor *disp_desc = [PGDisplayDescriptor new]; + + disp_desc.name = @"QEMU display"; + disp_desc.sizeInMillimeters = NSMakeSize(400., 300.); /* A 20" display */ + disp_desc.queue = dispatch_get_main_queue(); + disp_desc.newFrameEventHandler = ^(void) { + trace_apple_gfx_new_frame(); + dispatch_async(s->render_queue, ^{ + /* Drop frames if we get too far ahead. */ + if (s->pending_frames >= 2) + return; + ++s->pending_frames; + if (s->pending_frames > 1) { + return; + } + @autoreleasepool { + apple_gfx_render_new_frame(s); + } + }); + }; + disp_desc.modeChangeHandler = ^(PGDisplayCoord_t sizeInPixels, + OSType pixelFormat) { + trace_apple_gfx_mode_change(sizeInPixels.x, sizeInPixels.y); + set_mode(s, sizeInPixels.x, sizeInPixels.y); + }; + disp_desc.cursorGlyphHandler = ^(NSBitmapImageRep *glyph, + PGDisplayCoord_t hotSpot) { + uint32_t bpp = glyph.bitsPerPixel; + size_t width = glyph.pixelsWide; + size_t height = glyph.pixelsHigh; + size_t padding_bytes_per_row = glyph.bytesPerRow - width * 4; + const uint8_t* px_data = glyph.bitmapData; + + trace_apple_gfx_cursor_set(bpp, width, height); + + if (s->cursor) { + cursor_unref(s->cursor); + s->cursor = NULL; + } + + if (bpp == 32) { /* Shouldn't be anything else, but just to be safe...*/ + s->cursor = cursor_alloc(width, height); + s->cursor->hot_x = hotSpot.x; + s->cursor->hot_y = hotSpot.y; + + uint32_t *dest_px = s->cursor->data; + + for (size_t y = 0; y < height; ++y) { + for (size_t x = 0; x < width; ++x) { + /* NSBitmapImageRep's red & blue channels are swapped + * compared to QEMUCursor's. */ + *dest_px = + (px_data[0] << 16u) | + (px_data[1] << 8u) | + (px_data[2] << 0u) | + (px_data[3] << 24u); + ++dest_px; + px_data += 4; + } + px_data += padding_bytes_per_row; + } + dpy_cursor_define(s->con, s->cursor); + update_cursor(s); + } + }; + disp_desc.cursorShowHandler = ^(BOOL show) { + trace_apple_gfx_cursor_show(show); + s->cursor_show = show; + update_cursor(s); + }; + disp_desc.cursorMoveHandler = ^(void) { + trace_apple_gfx_cursor_move(); + update_cursor(s); + }; + + return disp_desc; +} + +static NSArray<PGDisplayMode*>* apple_gfx_prepare_display_mode_array(void) +{ + PGDisplayMode *modes[ARRAY_SIZE(apple_gfx_modes)]; + NSArray<PGDisplayMode*>* mode_array = nil; + int i; + + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { + modes[i] = + [[PGDisplayMode alloc] initWithSizeInPixels:apple_gfx_modes[i] refreshRateInHz:60.]; + } + + mode_array = [NSArray arrayWithObjects:modes count:ARRAY_SIZE(apple_gfx_modes)]; + + for (i = 0; i < ARRAY_SIZE(apple_gfx_modes); i++) { + [modes[i] release]; + modes[i] = nil; + } + + return mode_array; +} + +static id<MTLDevice> copy_suitable_metal_device(void) +{ + id<MTLDevice> dev = nil; + NSArray<id<MTLDevice>> *devs = MTLCopyAllDevices(); + + /* Prefer a unified memory GPU. Failing that, pick a non-removable GPU. */ + for (size_t i = 0; i < devs.count; ++i) { + if (devs[i].hasUnifiedMemory) { + dev = devs[i]; + break; + } + if (!devs[i].removable) { + dev = devs[i]; + } + } + + if (dev != nil) { + [dev retain]; + } else { + dev = MTLCreateSystemDefaultDevice(); + } + [devs release]; + + return dev; +} + +void apple_gfx_common_realize(AppleGFXState *s, PGDeviceDescriptor *desc) +{ + PGDisplayDescriptor *disp_desc = nil; + + QTAILQ_INIT(&s->tasks); + s->render_queue = dispatch_queue_create("apple-gfx.render", + DISPATCH_QUEUE_SERIAL); + s->mtl = copy_suitable_metal_device(); + s->mtl_queue = [s->mtl newCommandQueue]; + + desc.device = s->mtl; + + apple_gfx_register_task_mapping_handlers(s, desc); + + s->pgdev = PGNewDeviceWithDescriptor(desc); + + disp_desc = apple_gfx_prepare_display_handlers(s); + s->pgdisp = [s->pgdev newDisplayWithDescriptor:disp_desc + port:0 serialNum:1234]; + [disp_desc release]; + s->pgdisp.modeList = apple_gfx_prepare_display_mode_array(); + + create_fb(s); +} diff --git a/hw/display/meson.build b/hw/display/meson.build index 7db05eace97..70d855749c0 100644 --- a/hw/display/meson.build +++ b/hw/display/meson.build @@ -65,6 +65,8 @@ system_ss.add(when: 'CONFIG_ARTIST', if_true: files('artist.c')) system_ss.add(when: 'CONFIG_ATI_VGA', if_true: [files('ati.c', 'ati_2d.c', 'ati_dbg.c'), pixman]) +system_ss.add(when: 'CONFIG_MAC_PVG', if_true: [files('apple-gfx.m'), pvg, metal]) +system_ss.add(when: 'CONFIG_MAC_PVG_VMAPPLE', if_true: [files('apple-gfx-vmapple.m'), pvg, metal]) if config_all_devices.has_key('CONFIG_VIRTIO_GPU') virtio_gpu_ss = ss.source_set() diff --git a/hw/display/trace-events b/hw/display/trace-events index 781f8a33203..1809a358e36 100644 --- a/hw/display/trace-events +++ b/hw/display/trace-events @@ -191,3 +191,29 @@ dm163_bits_ppi(unsigned dest_width) "dest_width : %u" dm163_leds(int led, uint32_t value) "led %d: 0x%x" dm163_channels(int channel, uint8_t value) "channel %d: 0x%x" dm163_refresh_rate(uint32_t rr) "refresh rate %d" + +# apple-gfx.m +apple_gfx_read(uint64_t offset, uint64_t res) "offset=0x%"PRIx64" res=0x%"PRIx64 +apple_gfx_write(uint64_t offset, uint64_t val) "offset=0x%"PRIx64" val=0x%"PRIx64 +apple_gfx_create_task(uint32_t vm_size, void *va) "vm_size=0x%x base_addr=%p" +apple_gfx_destroy_task(void *task) "task=%p" +apple_gfx_map_memory(void *task, uint32_t range_count, uint64_t virtual_offset, uint32_t read_only) "task=%p range_count=0x%x virtual_offset=0x%"PRIx64" read_only=%d" +apple_gfx_map_memory_range(uint32_t i, uint64_t phys_addr, uint64_t phys_len) "[%d] phys_addr=0x%"PRIx64" phys_len=0x%"PRIx64 +apple_gfx_remap(uint64_t retval, uint64_t source, uint64_t target) "retval=%"PRId64" source=0x%"PRIx64" target=0x%"PRIx64 +apple_gfx_unmap_memory(void *task, uint64_t virtual_offset, uint64_t length) "task=%p virtual_offset=0x%"PRIx64" length=0x%"PRIx64 +apple_gfx_read_memory(uint64_t phys_address, uint64_t length, void *dst) "phys_addr=0x%"PRIx64" length=0x%"PRIx64" dest=%p" +apple_gfx_raise_irq(uint32_t vector) "vector=0x%x" +apple_gfx_new_frame(void) "" +apple_gfx_mode_change(uint64_t x, uint64_t y) "x=%"PRId64" y=%"PRId64 +apple_gfx_cursor_set(uint32_t bpp, uint64_t width, uint64_t height) "bpp=%d width=%"PRId64" height=0x%"PRId64 +apple_gfx_cursor_show(uint32_t show) "show=%d" +apple_gfx_cursor_move(void) "" +apple_gfx_common_init(const char *device_name, size_t mmio_size) "device: %s; MMIO size: %zu bytes" + +# apple-gfx-vmapple.m +apple_iosfc_read(uint64_t offset, uint64_t res) "offset=0x%"PRIx64" res=0x%"PRIx64 +apple_iosfc_write(uint64_t offset, uint64_t val) "offset=0x%"PRIx64" val=0x%"PRIx64 +apple_iosfc_map_memory(uint64_t phys, uint64_t len, uint32_t ro, void *va, void *e, void *f) "phys=0x%"PRIx64" len=0x%"PRIx64" ro=%d va=%p e=%p f=%p" +apple_iosfc_unmap_memory(void *a, void *b, void *c, void *d, void *e, void *f) "a=%p b=%p c=%p d=%p e=%p f=%p" +apple_iosfc_raise_irq(uint32_t vector) "vector=0x%x" + diff --git a/meson.build b/meson.build index 10464466ff3..f09df3f09d5 100644 --- a/meson.build +++ b/meson.build @@ -741,6 +741,8 @@ socket = [] version_res = [] coref = [] iokit = [] +pvg = [] +metal = [] emulator_link_args = [] midl = not_found widl = not_found @@ -762,6 +764,8 @@ elif host_os == 'darwin' coref = dependency('appleframeworks', modules: 'CoreFoundation') iokit = dependency('appleframeworks', modules: 'IOKit', required: false) host_dsosuf = '.dylib' + pvg = dependency('appleframeworks', modules: 'ParavirtualizedGraphics') + metal = dependency('appleframeworks', modules: 'Metal') elif host_os == 'sunos' socket = [cc.find_library('socket'), cc.find_library('nsl'),