Message ID | 20240613233639.202896-5-salil.mehta@huawei.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Support of Virtual CPU Hotplug for ARMv8 Arch | expand |
On 6/14/24 9:36 AM, Salil Mehta wrote: > Factor out CPU properties code common for {hot,cold}-plugged CPUs. This allows > code reuse. > > Signed-off-by: Salil Mehta <salil.mehta@huawei.com> > --- > hw/arm/virt.c | 261 ++++++++++++++++++++++++++++-------------- > include/hw/arm/virt.h | 4 + > 2 files changed, 182 insertions(+), 83 deletions(-) > If this series are going to be split for easier review, this is an candidate for a prepatory patch. > diff --git a/hw/arm/virt.c b/hw/arm/virt.c > index 3e1c4d2d2f..2e0ec7d869 100644 > --- a/hw/arm/virt.c > +++ b/hw/arm/virt.c > @@ -1753,6 +1753,46 @@ static uint64_t virt_cpu_mp_affinity(VirtMachineState *vms, int idx) > return arm_build_mp_affinity(idx, clustersz); > } > > +static CPUArchId *virt_find_cpu_slot(MachineState *ms, int vcpuid) > +{ > + VirtMachineState *vms = VIRT_MACHINE(ms); > + CPUArchId *found_cpu; > + uint64_t mp_affinity; > + > + assert(vcpuid >= 0 && vcpuid < ms->possible_cpus->len); > + > + mp_affinity = virt_cpu_mp_affinity(vms, vcpuid); > + found_cpu = &ms->possible_cpus->cpus[vcpuid]; > + > + assert(found_cpu->arch_id == mp_affinity); > + > + /* > + * RFC: Question: > + * Slot-id is the index where vCPU with certain arch-id(=mpidr/ap-affinity) > + * is plugged. For Host KVM, MPIDR for vCPU is derived using vcpu-id. > + * As I understand, MPIDR and vcpu-id are property of vCPU but slot-id is > + * more related to machine? Current code assumes slot-id and vcpu-id are > + * same i.e. meaning of slot is bit vague. > + * > + * Q1: Is there any requirement to clearly represent slot and dissociate it > + * from vcpu-id? > + * Q2: Should we make MPIDR within host KVM user configurable? > + * > + * +----+----+----+----+----+----+----+----+ > + * MPIDR ||| Res | Aff2 | Aff1 | Aff0 | > + * +----+----+----+----+----+----+----+----+ > + * \ \ \ | | > + * \ 8bit \ 8bit \ |4bit| > + * \<------->\<------->\ |<-->| > + * \ \ \| | > + * +----+----+----+----+----+----+----+----+ > + * VCPU-ID | Byte4 | Byte2 | Byte1 | Byte0 | > + * +----+----+----+----+----+----+----+----+ > + */ > + > + return found_cpu; > +} > + I don't see why virt_find_cpu_slot() is needed. Apart from the sanity check, it basically returns ms->possible_cpus->cpus[]. The caller can dereference ms->possible_cpus->cpus[] directly. About the sanity check, the mp_affinity has been properly populated in virt_possible_cpu_arch_ids(). I don't see why we it needs to be checked again. > static inline bool *virt_get_high_memmap_enabled(VirtMachineState *vms, > int index) > { > @@ -2065,16 +2105,129 @@ static void virt_cpu_post_init(VirtMachineState *vms, MemoryRegion *sysmem) > } > } > > +static void virt_cpu_set_properties(Object *cpuobj, const CPUArchId *cpu_slot, > + Error **errp) > +{ > + MachineState *ms = MACHINE(qdev_get_machine()); > + VirtMachineState *vms = VIRT_MACHINE(ms); > + Error *local_err = NULL; > + VirtMachineClass *vmc; > + > + vmc = VIRT_MACHINE_GET_CLASS(ms); > + > + /* now, set the cpu object property values */ > + numa_cpu_pre_plug(cpu_slot, DEVICE(cpuobj), &local_err); > + if (local_err) { > + goto out; > + } > + > + object_property_set_int(cpuobj, "mp-affinity", cpu_slot->arch_id, NULL); > + > + if (!vms->secure) { > + object_property_set_bool(cpuobj, "has_el3", false, NULL); > + } > + > + if (!vms->virt && object_property_find(cpuobj, "has_el2")) { > + object_property_set_bool(cpuobj, "has_el2", false, NULL); > + } > + > + if (vmc->kvm_no_adjvtime && > + object_property_find(cpuobj, "kvm-no-adjvtime")) { > + object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL); > + } > + > + if (vmc->no_kvm_steal_time && > + object_property_find(cpuobj, "kvm-steal-time")) { > + object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL); > + } > + > + if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) { > + object_property_set_bool(cpuobj, "pmu", false, NULL); > + } > + > + if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) { > + object_property_set_bool(cpuobj, "lpa2", false, NULL); > + } > + > + if (object_property_find(cpuobj, "reset-cbar")) { > + object_property_set_int(cpuobj, "reset-cbar", > + vms->memmap[VIRT_CPUPERIPHS].base, > + &local_err); > + if (local_err) { > + goto out; > + } > + } > + > + /* link already initialized {secure,tag}-memory regions to this cpu */ > + object_property_set_link(cpuobj, "memory", OBJECT(vms->sysmem), &local_err); > + if (local_err) { > + goto out; > + } > + > + if (vms->secure) { > + object_property_set_link(cpuobj, "secure-memory", > + OBJECT(vms->secure_sysmem), &local_err); > + if (local_err) { > + goto out; > + } > + } > + > + if (vms->mte) { > + if (!object_property_find(cpuobj, "tag-memory")) { > + error_setg(&local_err, "MTE requested, but not supported " > + "by the guest CPU"); > + if (local_err) { > + goto out; > + } > + } > + > + object_property_set_link(cpuobj, "tag-memory", OBJECT(vms->tag_sysmem), > + &local_err); > + if (local_err) { > + goto out; > + } > + > + if (vms->secure) { > + object_property_set_link(cpuobj, "secure-tag-memory", > + OBJECT(vms->secure_tag_sysmem), > + &local_err); > + if (local_err) { > + goto out; > + } > + } > + } > + > + /* > + * RFC: Question: this must only be called for the hotplugged cpus. For the > + * cold booted secondary cpus this is being taken care in arm_load_kernel() > + * in boot.c. Perhaps we should remove that code now? > + */ > + if (vms->psci_conduit != QEMU_PSCI_CONDUIT_DISABLED) { > + object_property_set_int(cpuobj, "psci-conduit", vms->psci_conduit, > + NULL); > + > + /* Secondary CPUs start in PSCI powered-down state */ > + if (CPU(cpuobj)->cpu_index > 0) { > + object_property_set_bool(cpuobj, "start-powered-off", true, NULL); > + } > + } > + > +out: > + if (local_err) { > + error_propagate(errp, local_err); > + } > +} > + > static void machvirt_init(MachineState *machine) > { > VirtMachineState *vms = VIRT_MACHINE(machine); > VirtMachineClass *vmc = VIRT_MACHINE_GET_CLASS(machine); > MachineClass *mc = MACHINE_GET_CLASS(machine); > const CPUArchIdList *possible_cpus; > - MemoryRegion *sysmem = get_system_memory(); > + MemoryRegion *secure_tag_sysmem = NULL; > MemoryRegion *secure_sysmem = NULL; > MemoryRegion *tag_sysmem = NULL; > - MemoryRegion *secure_tag_sysmem = NULL; > + MemoryRegion *sysmem; > int n, virt_max_cpus; > bool firmware_loaded; > bool aarch64 = true; > @@ -2148,6 +2301,8 @@ static void machvirt_init(MachineState *machine) > /* uses smp.max_cpus to initialize all possible vCPUs */ > possible_cpus = mc->possible_cpu_arch_ids(machine); > > + sysmem = vms->sysmem = get_system_memory(); > + > if (vms->secure) { > /* > * The Secure view of the world is the same as the NonSecure, > @@ -2155,7 +2310,7 @@ static void machvirt_init(MachineState *machine) > * containing the system memory at low priority; any secure-only > * devices go in at higher priority and take precedence. > */ > - secure_sysmem = g_new(MemoryRegion, 1); > + secure_sysmem = vms->secure_sysmem = g_new(MemoryRegion, 1); > memory_region_init(secure_sysmem, OBJECT(machine), "secure-memory", > UINT64_MAX); > memory_region_add_subregion_overlap(secure_sysmem, 0, sysmem, -1); > @@ -2203,10 +2358,28 @@ static void machvirt_init(MachineState *machine) > exit(1); > } > > + if (vms->mte) { > + /* Create the memory region only once, but link to all cpus later */ > + tag_sysmem = vms->tag_sysmem = g_new(MemoryRegion, 1); > + memory_region_init(tag_sysmem, OBJECT(machine), > + "tag-memory", UINT64_MAX / 32); > + > + if (vms->secure) { > + secure_tag_sysmem = vms->secure_tag_sysmem = g_new(MemoryRegion, 1); > + memory_region_init(secure_tag_sysmem, OBJECT(machine), > + "secure-tag-memory", UINT64_MAX / 32); > + > + /* As with ram, secure-tag takes precedence over tag. */ > + memory_region_add_subregion_overlap(secure_tag_sysmem, 0, > + tag_sysmem, -1); > + } > + } > + > create_fdt(vms); > > assert(possible_cpus->len == max_cpus); > for (n = 0; n < possible_cpus->len; n++) { > + CPUArchId *cpu_slot; > Object *cpuobj; > CPUState *cs; > > @@ -2215,15 +2388,10 @@ static void machvirt_init(MachineState *machine) > } > > cpuobj = object_new(possible_cpus->cpus[n].type); > - object_property_set_int(cpuobj, "mp-affinity", > - possible_cpus->cpus[n].arch_id, NULL); > > cs = CPU(cpuobj); > cs->cpu_index = n; > > - numa_cpu_pre_plug(&possible_cpus->cpus[cs->cpu_index], DEVICE(cpuobj), > - &error_fatal); > - > aarch64 &= object_property_get_bool(cpuobj, "aarch64", NULL); > object_property_set_int(cpuobj, "socket-id", > virt_get_socket_id(machine, n), NULL); > @@ -2234,81 +2402,8 @@ static void machvirt_init(MachineState *machine) > object_property_set_int(cpuobj, "thread-id", > virt_get_thread_id(machine, n), NULL); > > - if (!vms->secure) { > - object_property_set_bool(cpuobj, "has_el3", false, NULL); > - } > - > - if (!vms->virt && object_property_find(cpuobj, "has_el2")) { > - object_property_set_bool(cpuobj, "has_el2", false, NULL); > - } > - > - if (vmc->kvm_no_adjvtime && > - object_property_find(cpuobj, "kvm-no-adjvtime")) { > - object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL); > - } > - > - if (vmc->no_kvm_steal_time && > - object_property_find(cpuobj, "kvm-steal-time")) { > - object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL); > - } > - > - if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) { > - object_property_set_bool(cpuobj, "pmu", false, NULL); > - } > - > - if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) { > - object_property_set_bool(cpuobj, "lpa2", false, NULL); > - } > - > - if (object_property_find(cpuobj, "reset-cbar")) { > - object_property_set_int(cpuobj, "reset-cbar", > - vms->memmap[VIRT_CPUPERIPHS].base, > - &error_abort); > - } > - > - object_property_set_link(cpuobj, "memory", OBJECT(sysmem), > - &error_abort); > - if (vms->secure) { > - object_property_set_link(cpuobj, "secure-memory", > - OBJECT(secure_sysmem), &error_abort); > - } > - > - if (vms->mte) { > - /* Create the memory region only once, but link to all cpus. */ > - if (!tag_sysmem) { > - /* > - * The property exists only if MemTag is supported. > - * If it is, we must allocate the ram to back that up. > - */ > - if (!object_property_find(cpuobj, "tag-memory")) { > - error_report("MTE requested, but not supported " > - "by the guest CPU"); > - exit(1); > - } > - > - tag_sysmem = g_new(MemoryRegion, 1); > - memory_region_init(tag_sysmem, OBJECT(machine), > - "tag-memory", UINT64_MAX / 32); > - > - if (vms->secure) { > - secure_tag_sysmem = g_new(MemoryRegion, 1); > - memory_region_init(secure_tag_sysmem, OBJECT(machine), > - "secure-tag-memory", UINT64_MAX / 32); > - > - /* As with ram, secure-tag takes precedence over tag. */ > - memory_region_add_subregion_overlap(secure_tag_sysmem, 0, > - tag_sysmem, -1); > - } > - } > - > - object_property_set_link(cpuobj, "tag-memory", OBJECT(tag_sysmem), > - &error_abort); > - if (vms->secure) { > - object_property_set_link(cpuobj, "secure-tag-memory", > - OBJECT(secure_tag_sysmem), > - &error_abort); > - } > - } > + cpu_slot = virt_find_cpu_slot(machine, cs->cpu_index); > + virt_cpu_set_properties(cpuobj, cpu_slot, &error_fatal); > > qdev_realize(DEVICE(cpuobj), NULL, &error_fatal); > object_unref(cpuobj); > diff --git a/include/hw/arm/virt.h b/include/hw/arm/virt.h > index 6f9a7bb60b..780bd53ceb 100644 > --- a/include/hw/arm/virt.h > +++ b/include/hw/arm/virt.h > @@ -139,6 +139,10 @@ struct VirtMachineState { > DeviceState *platform_bus_dev; > FWCfgState *fw_cfg; > PFlashCFI01 *flash[2]; > + MemoryRegion *sysmem; > + MemoryRegion *secure_sysmem; > + MemoryRegion *tag_sysmem; > + MemoryRegion *secure_tag_sysmem; > bool secure; > bool highmem; > bool highmem_compact; Thanks, Gavin
diff --git a/hw/arm/virt.c b/hw/arm/virt.c index 3e1c4d2d2f..2e0ec7d869 100644 --- a/hw/arm/virt.c +++ b/hw/arm/virt.c @@ -1753,6 +1753,46 @@ static uint64_t virt_cpu_mp_affinity(VirtMachineState *vms, int idx) return arm_build_mp_affinity(idx, clustersz); } +static CPUArchId *virt_find_cpu_slot(MachineState *ms, int vcpuid) +{ + VirtMachineState *vms = VIRT_MACHINE(ms); + CPUArchId *found_cpu; + uint64_t mp_affinity; + + assert(vcpuid >= 0 && vcpuid < ms->possible_cpus->len); + + mp_affinity = virt_cpu_mp_affinity(vms, vcpuid); + found_cpu = &ms->possible_cpus->cpus[vcpuid]; + + assert(found_cpu->arch_id == mp_affinity); + + /* + * RFC: Question: + * Slot-id is the index where vCPU with certain arch-id(=mpidr/ap-affinity) + * is plugged. For Host KVM, MPIDR for vCPU is derived using vcpu-id. + * As I understand, MPIDR and vcpu-id are property of vCPU but slot-id is + * more related to machine? Current code assumes slot-id and vcpu-id are + * same i.e. meaning of slot is bit vague. + * + * Q1: Is there any requirement to clearly represent slot and dissociate it + * from vcpu-id? + * Q2: Should we make MPIDR within host KVM user configurable? + * + * +----+----+----+----+----+----+----+----+ + * MPIDR ||| Res | Aff2 | Aff1 | Aff0 | + * +----+----+----+----+----+----+----+----+ + * \ \ \ | | + * \ 8bit \ 8bit \ |4bit| + * \<------->\<------->\ |<-->| + * \ \ \| | + * +----+----+----+----+----+----+----+----+ + * VCPU-ID | Byte4 | Byte2 | Byte1 | Byte0 | + * +----+----+----+----+----+----+----+----+ + */ + + return found_cpu; +} + static inline bool *virt_get_high_memmap_enabled(VirtMachineState *vms, int index) { @@ -2065,16 +2105,129 @@ static void virt_cpu_post_init(VirtMachineState *vms, MemoryRegion *sysmem) } } +static void virt_cpu_set_properties(Object *cpuobj, const CPUArchId *cpu_slot, + Error **errp) +{ + MachineState *ms = MACHINE(qdev_get_machine()); + VirtMachineState *vms = VIRT_MACHINE(ms); + Error *local_err = NULL; + VirtMachineClass *vmc; + + vmc = VIRT_MACHINE_GET_CLASS(ms); + + /* now, set the cpu object property values */ + numa_cpu_pre_plug(cpu_slot, DEVICE(cpuobj), &local_err); + if (local_err) { + goto out; + } + + object_property_set_int(cpuobj, "mp-affinity", cpu_slot->arch_id, NULL); + + if (!vms->secure) { + object_property_set_bool(cpuobj, "has_el3", false, NULL); + } + + if (!vms->virt && object_property_find(cpuobj, "has_el2")) { + object_property_set_bool(cpuobj, "has_el2", false, NULL); + } + + if (vmc->kvm_no_adjvtime && + object_property_find(cpuobj, "kvm-no-adjvtime")) { + object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL); + } + + if (vmc->no_kvm_steal_time && + object_property_find(cpuobj, "kvm-steal-time")) { + object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL); + } + + if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) { + object_property_set_bool(cpuobj, "pmu", false, NULL); + } + + if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) { + object_property_set_bool(cpuobj, "lpa2", false, NULL); + } + + if (object_property_find(cpuobj, "reset-cbar")) { + object_property_set_int(cpuobj, "reset-cbar", + vms->memmap[VIRT_CPUPERIPHS].base, + &local_err); + if (local_err) { + goto out; + } + } + + /* link already initialized {secure,tag}-memory regions to this cpu */ + object_property_set_link(cpuobj, "memory", OBJECT(vms->sysmem), &local_err); + if (local_err) { + goto out; + } + + if (vms->secure) { + object_property_set_link(cpuobj, "secure-memory", + OBJECT(vms->secure_sysmem), &local_err); + if (local_err) { + goto out; + } + } + + if (vms->mte) { + if (!object_property_find(cpuobj, "tag-memory")) { + error_setg(&local_err, "MTE requested, but not supported " + "by the guest CPU"); + if (local_err) { + goto out; + } + } + + object_property_set_link(cpuobj, "tag-memory", OBJECT(vms->tag_sysmem), + &local_err); + if (local_err) { + goto out; + } + + if (vms->secure) { + object_property_set_link(cpuobj, "secure-tag-memory", + OBJECT(vms->secure_tag_sysmem), + &local_err); + if (local_err) { + goto out; + } + } + } + + /* + * RFC: Question: this must only be called for the hotplugged cpus. For the + * cold booted secondary cpus this is being taken care in arm_load_kernel() + * in boot.c. Perhaps we should remove that code now? + */ + if (vms->psci_conduit != QEMU_PSCI_CONDUIT_DISABLED) { + object_property_set_int(cpuobj, "psci-conduit", vms->psci_conduit, + NULL); + + /* Secondary CPUs start in PSCI powered-down state */ + if (CPU(cpuobj)->cpu_index > 0) { + object_property_set_bool(cpuobj, "start-powered-off", true, NULL); + } + } + +out: + if (local_err) { + error_propagate(errp, local_err); + } +} + static void machvirt_init(MachineState *machine) { VirtMachineState *vms = VIRT_MACHINE(machine); VirtMachineClass *vmc = VIRT_MACHINE_GET_CLASS(machine); MachineClass *mc = MACHINE_GET_CLASS(machine); const CPUArchIdList *possible_cpus; - MemoryRegion *sysmem = get_system_memory(); + MemoryRegion *secure_tag_sysmem = NULL; MemoryRegion *secure_sysmem = NULL; MemoryRegion *tag_sysmem = NULL; - MemoryRegion *secure_tag_sysmem = NULL; + MemoryRegion *sysmem; int n, virt_max_cpus; bool firmware_loaded; bool aarch64 = true; @@ -2148,6 +2301,8 @@ static void machvirt_init(MachineState *machine) /* uses smp.max_cpus to initialize all possible vCPUs */ possible_cpus = mc->possible_cpu_arch_ids(machine); + sysmem = vms->sysmem = get_system_memory(); + if (vms->secure) { /* * The Secure view of the world is the same as the NonSecure, @@ -2155,7 +2310,7 @@ static void machvirt_init(MachineState *machine) * containing the system memory at low priority; any secure-only * devices go in at higher priority and take precedence. */ - secure_sysmem = g_new(MemoryRegion, 1); + secure_sysmem = vms->secure_sysmem = g_new(MemoryRegion, 1); memory_region_init(secure_sysmem, OBJECT(machine), "secure-memory", UINT64_MAX); memory_region_add_subregion_overlap(secure_sysmem, 0, sysmem, -1); @@ -2203,10 +2358,28 @@ static void machvirt_init(MachineState *machine) exit(1); } + if (vms->mte) { + /* Create the memory region only once, but link to all cpus later */ + tag_sysmem = vms->tag_sysmem = g_new(MemoryRegion, 1); + memory_region_init(tag_sysmem, OBJECT(machine), + "tag-memory", UINT64_MAX / 32); + + if (vms->secure) { + secure_tag_sysmem = vms->secure_tag_sysmem = g_new(MemoryRegion, 1); + memory_region_init(secure_tag_sysmem, OBJECT(machine), + "secure-tag-memory", UINT64_MAX / 32); + + /* As with ram, secure-tag takes precedence over tag. */ + memory_region_add_subregion_overlap(secure_tag_sysmem, 0, + tag_sysmem, -1); + } + } + create_fdt(vms); assert(possible_cpus->len == max_cpus); for (n = 0; n < possible_cpus->len; n++) { + CPUArchId *cpu_slot; Object *cpuobj; CPUState *cs; @@ -2215,15 +2388,10 @@ static void machvirt_init(MachineState *machine) } cpuobj = object_new(possible_cpus->cpus[n].type); - object_property_set_int(cpuobj, "mp-affinity", - possible_cpus->cpus[n].arch_id, NULL); cs = CPU(cpuobj); cs->cpu_index = n; - numa_cpu_pre_plug(&possible_cpus->cpus[cs->cpu_index], DEVICE(cpuobj), - &error_fatal); - aarch64 &= object_property_get_bool(cpuobj, "aarch64", NULL); object_property_set_int(cpuobj, "socket-id", virt_get_socket_id(machine, n), NULL); @@ -2234,81 +2402,8 @@ static void machvirt_init(MachineState *machine) object_property_set_int(cpuobj, "thread-id", virt_get_thread_id(machine, n), NULL); - if (!vms->secure) { - object_property_set_bool(cpuobj, "has_el3", false, NULL); - } - - if (!vms->virt && object_property_find(cpuobj, "has_el2")) { - object_property_set_bool(cpuobj, "has_el2", false, NULL); - } - - if (vmc->kvm_no_adjvtime && - object_property_find(cpuobj, "kvm-no-adjvtime")) { - object_property_set_bool(cpuobj, "kvm-no-adjvtime", true, NULL); - } - - if (vmc->no_kvm_steal_time && - object_property_find(cpuobj, "kvm-steal-time")) { - object_property_set_bool(cpuobj, "kvm-steal-time", false, NULL); - } - - if (vmc->no_pmu && object_property_find(cpuobj, "pmu")) { - object_property_set_bool(cpuobj, "pmu", false, NULL); - } - - if (vmc->no_tcg_lpa2 && object_property_find(cpuobj, "lpa2")) { - object_property_set_bool(cpuobj, "lpa2", false, NULL); - } - - if (object_property_find(cpuobj, "reset-cbar")) { - object_property_set_int(cpuobj, "reset-cbar", - vms->memmap[VIRT_CPUPERIPHS].base, - &error_abort); - } - - object_property_set_link(cpuobj, "memory", OBJECT(sysmem), - &error_abort); - if (vms->secure) { - object_property_set_link(cpuobj, "secure-memory", - OBJECT(secure_sysmem), &error_abort); - } - - if (vms->mte) { - /* Create the memory region only once, but link to all cpus. */ - if (!tag_sysmem) { - /* - * The property exists only if MemTag is supported. - * If it is, we must allocate the ram to back that up. - */ - if (!object_property_find(cpuobj, "tag-memory")) { - error_report("MTE requested, but not supported " - "by the guest CPU"); - exit(1); - } - - tag_sysmem = g_new(MemoryRegion, 1); - memory_region_init(tag_sysmem, OBJECT(machine), - "tag-memory", UINT64_MAX / 32); - - if (vms->secure) { - secure_tag_sysmem = g_new(MemoryRegion, 1); - memory_region_init(secure_tag_sysmem, OBJECT(machine), - "secure-tag-memory", UINT64_MAX / 32); - - /* As with ram, secure-tag takes precedence over tag. */ - memory_region_add_subregion_overlap(secure_tag_sysmem, 0, - tag_sysmem, -1); - } - } - - object_property_set_link(cpuobj, "tag-memory", OBJECT(tag_sysmem), - &error_abort); - if (vms->secure) { - object_property_set_link(cpuobj, "secure-tag-memory", - OBJECT(secure_tag_sysmem), - &error_abort); - } - } + cpu_slot = virt_find_cpu_slot(machine, cs->cpu_index); + virt_cpu_set_properties(cpuobj, cpu_slot, &error_fatal); qdev_realize(DEVICE(cpuobj), NULL, &error_fatal); object_unref(cpuobj); diff --git a/include/hw/arm/virt.h b/include/hw/arm/virt.h index 6f9a7bb60b..780bd53ceb 100644 --- a/include/hw/arm/virt.h +++ b/include/hw/arm/virt.h @@ -139,6 +139,10 @@ struct VirtMachineState { DeviceState *platform_bus_dev; FWCfgState *fw_cfg; PFlashCFI01 *flash[2]; + MemoryRegion *sysmem; + MemoryRegion *secure_sysmem; + MemoryRegion *tag_sysmem; + MemoryRegion *secure_tag_sysmem; bool secure; bool highmem; bool highmem_compact;
Factor out CPU properties code common for {hot,cold}-plugged CPUs. This allows code reuse. Signed-off-by: Salil Mehta <salil.mehta@huawei.com> --- hw/arm/virt.c | 261 ++++++++++++++++++++++++++++-------------- include/hw/arm/virt.h | 4 + 2 files changed, 182 insertions(+), 83 deletions(-)