Message ID | 20191115075352.17734-9-tao3.xu@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Build ACPI Heterogeneous Memory Attribute Table (HMAT) | expand |
On Fri, 15 Nov 2019 15:53:46 +0800 Tao Xu <tao3.xu@intel.com> wrote: > From: Liu Jingqi <jingqi.liu@intel.com> > > Add -numa hmat-lb option to provide System Locality Latency and > Bandwidth Information. These memory attributes help to build > System Locality Latency and Bandwidth Information Structure(s) > in ACPI Heterogeneous Memory Attribute Table (HMAT). > > Signed-off-by: Liu Jingqi <jingqi.liu@intel.com> > Signed-off-by: Tao Xu <tao3.xu@intel.com> looks good to me, so Reviewed-by: Igor Mammedov <imammedo@redhat.com> PS: also see question below > --- > > Changes in v16: > - Initialize HMAT_LB_Data lb_data (Igor) > - Remove punctuation from error_setg (Igor) > - Correct some description (Igor) > - Drop statement about max value (Igor) > - Simplify struct HMAT_LB_Info and related code, unify latency > and bandwidth (Igor) > > Changes in v15: > - Change the QAPI version tag to 5.0 (Eric) > > Changes in v14: > - Use qemu ctz64 and clz64 instead of builtin function > - Improve help message in qemu-options.hx > --- > hw/core/numa.c | 174 ++++++++++++++++++++++++++++++++++++++++++ > include/sysemu/numa.h | 53 +++++++++++++ > qapi/machine.json | 94 ++++++++++++++++++++++- > qemu-options.hx | 49 +++++++++++- > 4 files changed, 367 insertions(+), 3 deletions(-) > > diff --git a/hw/core/numa.c b/hw/core/numa.c > index e60da99293..47c7a96863 100644 > --- a/hw/core/numa.c > +++ b/hw/core/numa.c > @@ -23,6 +23,7 @@ > */ > > #include "qemu/osdep.h" > +#include "qemu/units.h" > #include "sysemu/hostmem.h" > #include "sysemu/numa.h" > #include "sysemu/sysemu.h" > @@ -198,6 +199,166 @@ void parse_numa_distance(MachineState *ms, NumaDistOptions *dist, Error **errp) > ms->numa_state->have_numa_distance = true; > } > > +void parse_numa_hmat_lb(NumaState *numa_state, NumaHmatLBOptions *node, > + Error **errp) > +{ > + int i, first_bit, last_bit; > + uint64_t max_entry, temp_base_la; > + NodeInfo *numa_info = numa_state->nodes; > + HMAT_LB_Info *hmat_lb = > + numa_state->hmat_lb[node->hierarchy][node->data_type]; > + HMAT_LB_Data lb_data = {}; > + HMAT_LB_Data *lb_temp; > + > + /* Error checking */ > + if (node->initiator > numa_state->num_nodes) { > + error_setg(errp, "Invalid initiator=%d, it should be less than %d", > + node->initiator, numa_state->num_nodes); > + return; > + } > + if (node->target > numa_state->num_nodes) { > + error_setg(errp, "Invalid target=%d, it should be less than %d", > + node->target, numa_state->num_nodes); > + return; > + } > + if (!numa_info[node->initiator].has_cpu) { > + error_setg(errp, "Invalid initiator=%d, it isn't an " > + "initiator proximity domain", node->initiator); > + return; > + } > + if (!numa_info[node->target].present) { > + error_setg(errp, "The target=%d should point to an existing node", > + node->target); > + return; > + } > + > + if (!hmat_lb) { > + hmat_lb = g_malloc0(sizeof(*hmat_lb)); > + numa_state->hmat_lb[node->hierarchy][node->data_type] = hmat_lb; > + hmat_lb->list = g_array_new(false, true, sizeof(HMAT_LB_Data)); > + } > + hmat_lb->hierarchy = node->hierarchy; > + hmat_lb->data_type = node->data_type; > + lb_data.initiator = node->initiator; > + lb_data.target = node->target; > + > + if (node->data_type <= HMATLB_DATA_TYPE_WRITE_LATENCY) { > + /* Input latency data */ > + > + if (!node->has_latency) { > + error_setg(errp, "Missing 'latency' option"); > + return; > + } > + if (node->has_bandwidth) { > + error_setg(errp, "Invalid option 'bandwidth' since " > + "the data type is latency"); > + return; > + } > + > + /* Detect duplicate configuration */ > + for (i = 0; i < hmat_lb->list->len; i++) { > + lb_temp = &g_array_index(hmat_lb->list, HMAT_LB_Data, i); > + > + if (node->initiator == lb_temp->initiator && > + node->target == lb_temp->target) { > + error_setg(errp, "Duplicate configuration of the latency for " > + "initiator=%d and target=%d", node->initiator, > + node->target); > + return; > + } > + } > + > + hmat_lb->base = hmat_lb->base ? hmat_lb->base : UINT64_MAX; > + > + /* Calculate the temporary base and compressed latency */ > + max_entry = node->latency; > + temp_base_la = 1; > + while (QEMU_IS_ALIGNED(max_entry, 10)) { > + max_entry /= 10; > + temp_base_la *= 10; > + } > + > + /* Calculate the max compressed latency */ > + hmat_lb->base = MIN(hmat_lb->base, temp_base_la); > + max_entry = node->latency / hmat_lb->base; > + hmat_lb->range_bitmap = MAX(hmat_lb->range_bitmap, max_entry); > + > + /* > + * For latency hmat_lb->range_bitmap record the max compressed latency > + * which should be less than 0xFFFF (UINT16_MAX) > + */ > + if (hmat_lb->range_bitmap >= UINT16_MAX) { > + error_setg(errp, "Latency %" PRIu64 " between initiator=%d and " > + "target=%d should not differ from previously entered " > + "min or max values on more than %d", node->latency, > + node->initiator, node->target, UINT16_MAX - 1); > + return; > + } > + > + /* Set lb_info_provided bit 0 as 1, latency information is provided */ > + numa_info[node->target].lb_info_provided |= BIT(0); > + lb_data.data = node->latency; > + } else if (node->data_type >= HMATLB_DATA_TYPE_ACCESS_BANDWIDTH) { > + /* Input bandwidth data */ > + > + if (!node->has_bandwidth) { > + error_setg(errp, "Missing 'bandwidth' option"); > + return; > + } > + if (node->has_latency) { > + error_setg(errp, "Invalid option 'latency' since " > + "the data type is bandwidth"); > + return; > + } > + if (!QEMU_IS_ALIGNED(node->bandwidth, MiB)) { > + error_setg(errp, "Bandwidth %" PRIu64 " between initiator=%d and " > + "target=%d should be 1MB aligned", node->bandwidth, > + node->initiator, node->target); > + return; > + } > + > + /* Detect duplicate configuration */ > + for (i = 0; i < hmat_lb->list->len; i++) { > + lb_temp = &g_array_index(hmat_lb->list, HMAT_LB_Data, i); > + > + if (node->initiator == lb_temp->initiator && > + node->target == lb_temp->target) { > + error_setg(errp, "Duplicate configuration of the bandwidth for " > + "initiator=%d and target=%d", node->initiator, > + node->target); > + return; > + } > + } > + > + hmat_lb->range_bitmap |= node->bandwidth; > + first_bit = ctz64(hmat_lb->range_bitmap); > + hmat_lb->base = UINT64_C(1) << first_bit; > + max_entry = node->bandwidth / hmat_lb->base; > + last_bit = 64 - clz64(hmat_lb->range_bitmap); > + > + /* > + * For bandwidth, first_bit record the base unit of bandwidth bits, > + * last_bit record the last bit of the max bandwidth. The max compressed > + * bandwidth should be less than 0xFFFF (UINT16_MAX) > + */ > + if ((last_bit - first_bit) > UINT16_BITS || max_entry >= UINT16_MAX) { ^^^^^^^^^^^^^^^^^^^ what bandwidth combination is going to trigger above condition? > + error_setg(errp, "Bandwidth %" PRIu64 " between initiator=%d and " > + "target=%d should not differ from previously entered " > + "values on more than %d", node->bandwidth, > + node->initiator, node->target, UINT16_MAX - 1); > + return; > + } > + > + /* Set lb_info_provided bit 1 as 1, bandwidth information is provided */ > + numa_info[node->target].lb_info_provided |= BIT(1); > + lb_data.data = node->bandwidth; > + } else { > + assert(0); > + } > + > + g_array_append_val(hmat_lb->list, lb_data); > +} > + > void set_numa_options(MachineState *ms, NumaOptions *object, Error **errp) > { > Error *err = NULL; > @@ -236,6 +397,19 @@ void set_numa_options(MachineState *ms, NumaOptions *object, Error **errp) > machine_set_cpu_numa_node(ms, qapi_NumaCpuOptions_base(&object->u.cpu), > &err); > break; > + case NUMA_OPTIONS_TYPE_HMAT_LB: > + if (!ms->numa_state->hmat_enabled) { > + error_setg(errp, "ACPI Heterogeneous Memory Attribute Table " > + "(HMAT) is disabled, enable it with -machine hmat=on " > + "before using any of hmat specific options"); > + return; > + } > + > + parse_numa_hmat_lb(ms->numa_state, &object->u.hmat_lb, &err); > + if (err) { > + goto end; > + } > + break; > default: > abort(); > } > diff --git a/include/sysemu/numa.h b/include/sysemu/numa.h > index 788cbec7a2..70f93c83d7 100644 > --- a/include/sysemu/numa.h > +++ b/include/sysemu/numa.h > @@ -14,11 +14,34 @@ struct CPUArchId; > #define NUMA_DISTANCE_MAX 254 > #define NUMA_DISTANCE_UNREACHABLE 255 > > +/* the value of AcpiHmatLBInfo flags */ > +enum { > + HMAT_LB_MEM_MEMORY = 0, > + HMAT_LB_MEM_CACHE_1ST_LEVEL = 1, > + HMAT_LB_MEM_CACHE_2ND_LEVEL = 2, > + HMAT_LB_MEM_CACHE_3RD_LEVEL = 3, > + HMAT_LB_LEVELS /* must be the last entry */ > +}; > + > +/* the value of AcpiHmatLBInfo data type */ > +enum { > + HMAT_LB_DATA_ACCESS_LATENCY = 0, > + HMAT_LB_DATA_READ_LATENCY = 1, > + HMAT_LB_DATA_WRITE_LATENCY = 2, > + HMAT_LB_DATA_ACCESS_BANDWIDTH = 3, > + HMAT_LB_DATA_READ_BANDWIDTH = 4, > + HMAT_LB_DATA_WRITE_BANDWIDTH = 5, > + HMAT_LB_TYPES /* must be the last entry */ > +}; > + > +#define UINT16_BITS 16 > + > struct NodeInfo { > uint64_t node_mem; > struct HostMemoryBackend *node_memdev; > bool present; > bool has_cpu; > + uint8_t lb_info_provided; > uint16_t initiator; > uint8_t distance[MAX_NODES]; > }; > @@ -28,6 +51,31 @@ struct NumaNodeMem { > uint64_t node_plugged_mem; > }; > > +struct HMAT_LB_Data { > + uint8_t initiator; > + uint8_t target; > + uint64_t data; > +}; > +typedef struct HMAT_LB_Data HMAT_LB_Data; > + > +struct HMAT_LB_Info { > + /* Indicates it's memory or the specified level memory side cache. */ > + uint8_t hierarchy; > + > + /* Present the type of data, access/read/write latency or bandwidth. */ > + uint8_t data_type; > + > + /* The range bitmap of bandwidth for calculating common base */ > + uint64_t range_bitmap; > + > + /* The common base unit for latencies or bandwidths */ > + uint64_t base; > + > + /* Array to store the latencies or bandwidths */ > + GArray *list; > +}; > +typedef struct HMAT_LB_Info HMAT_LB_Info; > + > struct NumaState { > /* Number of NUMA nodes */ > int num_nodes; > @@ -40,11 +88,16 @@ struct NumaState { > > /* NUMA nodes information */ > NodeInfo nodes[MAX_NODES]; > + > + /* NUMA nodes HMAT Locality Latency and Bandwidth Information */ > + HMAT_LB_Info *hmat_lb[HMAT_LB_LEVELS][HMAT_LB_TYPES]; > }; > typedef struct NumaState NumaState; > > void set_numa_options(MachineState *ms, NumaOptions *object, Error **errp); > void parse_numa_opts(MachineState *ms); > +void parse_numa_hmat_lb(NumaState *numa_state, NumaHmatLBOptions *node, > + Error **errp); > void numa_complete_configuration(MachineState *ms); > void query_numa_node_mem(NumaNodeMem node_mem[], MachineState *ms); > extern QemuOptsList qemu_numa_opts; > diff --git a/qapi/machine.json b/qapi/machine.json > index 27d0e37534..67f5910400 100644 > --- a/qapi/machine.json > +++ b/qapi/machine.json > @@ -426,10 +426,12 @@ > # > # @cpu: property based CPU(s) to node mapping (Since: 2.10) > # > +# @hmat-lb: memory latency and bandwidth information (Since: 5.0) > +# > # Since: 2.1 > ## > { 'enum': 'NumaOptionsType', > - 'data': [ 'node', 'dist', 'cpu' ] } > + 'data': [ 'node', 'dist', 'cpu', 'hmat-lb' ] } > > ## > # @NumaOptions: > @@ -444,7 +446,8 @@ > 'data': { > 'node': 'NumaNodeOptions', > 'dist': 'NumaDistOptions', > - 'cpu': 'NumaCpuOptions' }} > + 'cpu': 'NumaCpuOptions', > + 'hmat-lb': 'NumaHmatLBOptions' }} > > ## > # @NumaNodeOptions: > @@ -557,6 +560,93 @@ > 'base': 'CpuInstanceProperties', > 'data' : {} } > > +## > +# @HmatLBMemoryHierarchy: > +# > +# The memory hierarchy in the System Locality Latency > +# and Bandwidth Information Structure of HMAT (Heterogeneous > +# Memory Attribute Table) > +# > +# For more information about @HmatLBMemoryHierarchy see > +# the chapter 5.2.27.4: Table 5-146: Field "Flags" of ACPI 6.3 spec. > +# > +# @memory: the structure represents the memory performance > +# > +# @first-level: first level of memory side cache > +# > +# @second-level: second level of memory side cache > +# > +# @third-level: third level of memory side cache > +# > +# Since: 5.0 > +## > +{ 'enum': 'HmatLBMemoryHierarchy', > + 'data': [ 'memory', 'first-level', 'second-level', 'third-level' ] } > + > +## > +# @HmatLBDataType: > +# > +# Data type in the System Locality Latency > +# and Bandwidth Information Structure of HMAT (Heterogeneous > +# Memory Attribute Table) > +# > +# For more information about @HmatLBDataType see > +# the chapter 5.2.27.4: Table 5-146: Field "Data Type" of ACPI 6.3 spec. > +# > +# @access-latency: access latency (nanoseconds) > +# > +# @read-latency: read latency (nanoseconds) > +# > +# @write-latency: write latency (nanoseconds) > +# > +# @access-bandwidth: access bandwidth (B/s) > +# > +# @read-bandwidth: read bandwidth (B/s) > +# > +# @write-bandwidth: write bandwidth (B/s) > +# > +# Since: 5.0 > +## > +{ 'enum': 'HmatLBDataType', > + 'data': [ 'access-latency', 'read-latency', 'write-latency', > + 'access-bandwidth', 'read-bandwidth', 'write-bandwidth' ] } > + > +## > +# @NumaHmatLBOptions: > +# > +# Set the system locality latency and bandwidth information > +# between Initiator and Target proximity Domains. > +# > +# For more information about @NumaHmatLBOptions see > +# the chapter 5.2.27.4: Table 5-146 of ACPI 6.3 spec. > +# > +# @initiator: the Initiator Proximity Domain. > +# > +# @target: the Target Proximity Domain. > +# > +# @hierarchy: the Memory Hierarchy. Indicates the performance > +# of memory or side cache. > +# > +# @data-type: presents the type of data, access/read/write > +# latency or hit latency. > +# > +# @latency: the value of latency from @initiator to @target proximity domain, > +# the latency unit is "ns(nanosecond)". > +# > +# @bandwidth: the value of bandwidth between @initiator and @target proximity > +# domain, the bandwidth unit is "B(/s)". > +# > +# Since: 5.0 > +## > +{ 'struct': 'NumaHmatLBOptions', > + 'data': { > + 'initiator': 'uint16', > + 'target': 'uint16', > + 'hierarchy': 'HmatLBMemoryHierarchy', > + 'data-type': 'HmatLBDataType', > + '*latency': 'time', > + '*bandwidth': 'size' }} > + > ## > # @HostMemPolicy: > # > diff --git a/qemu-options.hx b/qemu-options.hx > index 63f6b33322..929d275450 100644 > --- a/qemu-options.hx > +++ b/qemu-options.hx > @@ -168,16 +168,19 @@ DEF("numa", HAS_ARG, QEMU_OPTION_numa, > "-numa node[,mem=size][,cpus=firstcpu[-lastcpu]][,nodeid=node][,initiator=node]\n" > "-numa node[,memdev=id][,cpus=firstcpu[-lastcpu]][,nodeid=node][,initiator=node]\n" > "-numa dist,src=source,dst=destination,val=distance\n" > - "-numa cpu,node-id=node[,socket-id=x][,core-id=y][,thread-id=z]\n", > + "-numa cpu,node-id=node[,socket-id=x][,core-id=y][,thread-id=z]\n" > + "-numa hmat-lb,initiator=node,target=node,hierarchy=memory|first-level|second-level|third-level,data-type=access-latency|read-latency|write-latency[,latency=lat][,bandwidth=bw]\n", > QEMU_ARCH_ALL) > STEXI > @item -numa node[,mem=@var{size}][,cpus=@var{firstcpu}[-@var{lastcpu}]][,nodeid=@var{node}][,initiator=@var{initiator}] > @itemx -numa node[,memdev=@var{id}][,cpus=@var{firstcpu}[-@var{lastcpu}]][,nodeid=@var{node}][,initiator=@var{initiator}] > @itemx -numa dist,src=@var{source},dst=@var{destination},val=@var{distance} > @itemx -numa cpu,node-id=@var{node}[,socket-id=@var{x}][,core-id=@var{y}][,thread-id=@var{z}] > +@itemx -numa hmat-lb,initiator=@var{node},target=@var{node},hierarchy=@var{hierarchy},data-type=@var{tpye}[,latency=@var{lat}][,bandwidth=@var{bw}] > @findex -numa > Define a NUMA node and assign RAM and VCPUs to it. > Set the NUMA distance from a source node to a destination node. > +Set the ACPI Heterogeneous Memory Attributes for the given nodes. > > Legacy VCPU assignment uses @samp{cpus} option where > @var{firstcpu} and @var{lastcpu} are CPU indexes. Each > @@ -256,6 +259,50 @@ specified resources, it just assigns existing resources to NUMA > nodes. This means that one still has to use the @option{-m}, > @option{-smp} options to allocate RAM and VCPUs respectively. > > +Use @samp{hmat-lb} to set System Locality Latency and Bandwidth Information > +between initiator and target NUMA nodes in ACPI Heterogeneous Attribute Memory Table (HMAT). > +Initiator NUMA node can create memory requests, usually it has one or more processors. > +Target NUMA node contains addressable memory. > + > +In @samp{hmat-lb} option, @var{node} are NUMA node IDs. @var{hierarchy} is the memory > +hierarchy of the target NUMA node: if @var{hierarchy} is 'memory', the structure > +represents the memory performance; if @var{hierarchy} is 'first-level|second-level|third-level', > +this structure represents aggregated performance of memory side caches for each domain. > +@var{type} of 'data-type' is type of data represented by this structure instance: > +if 'hierarchy' is 'memory', 'data-type' is 'access|read|write' latency or 'access|read|write' > +bandwidth of the target memory; if 'hierarchy' is 'first-level|second-level|third-level', > +'data-type' is 'access|read|write' hit latency or 'access|read|write' hit bandwidth of the > +target memory side cache. > + > +@var{lat} is latency value, the possible value and units are NUM[ns|us|ms] > +(nanosecond|microsecond|millisecond), the recommended unit is 'ns'. @var{bw} is > +bandwidth value, the possible value and units are NUM[M|G|T], mean that the > +bandwidth value are NUM byte (or MB/s, GB/s or TB/s depending on used suffix). > +Note that if NUM is 0, means the corresponding latency or bandwidth information > +is not provided. And if input numbers without any unit, the latency unit will be > +'ns' and the bandwidth will be B/s. > + > +For example, the following options describe 2 NUMA nodes. Node 0 has 2 cpus and > +a ram, node 1 has only a ram. The processors in node 0 access memory in node > +0 with access-latency 5 nanoseconds, access-bandwidth is 200 MB/s; > +The processors in NUMA node 0 access memory in NUMA node 1 with access-latency 10 > +nanoseconds, access-bandwidth is 100 MB/s. > +@example > +-machine hmat=on \ > +-m 2G \ > +-object memory-backend-ram,size=1G,id=m0 \ > +-object memory-backend-ram,size=1G,id=m1 \ > +-smp 2 \ > +-numa node,nodeid=0,memdev=m0 \ > +-numa node,nodeid=1,memdev=m1,initiator=0 \ > +-numa cpu,node-id=0,socket-id=0 \ > +-numa cpu,node-id=0,socket-id=1 \ > +-numa hmat-lb,initiator=0,target=0,hierarchy=memory,data-type=access-latency,latency=5ns \ > +-numa hmat-lb,initiator=0,target=0,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M \ > +-numa hmat-lb,initiator=0,target=1,hierarchy=memory,data-type=access-latency,latency=10ns \ > +-numa hmat-lb,initiator=0,target=1,hierarchy=memory,data-type=access-bandwidth,bandwidth=100M > +@end example > + > ETEXI > > DEF("add-fd", HAS_ARG, QEMU_OPTION_add_fd,
On 11/19/2019 7:03 PM, Igor Mammedov wrote: > On Fri, 15 Nov 2019 15:53:46 +0800 > Tao Xu <tao3.xu@intel.com> wrote: > >> From: Liu Jingqi <jingqi.liu@intel.com> >> >> Add -numa hmat-lb option to provide System Locality Latency and >> Bandwidth Information. These memory attributes help to build >> System Locality Latency and Bandwidth Information Structure(s) >> in ACPI Heterogeneous Memory Attribute Table (HMAT). >> >> Signed-off-by: Liu Jingqi <jingqi.liu@intel.com> >> Signed-off-by: Tao Xu <tao3.xu@intel.com> > > looks good to me, so > > Reviewed-by: Igor Mammedov <imammedo@redhat.com> > > > PS: > also see question below > [...] >> + >> + hmat_lb->range_bitmap |= node->bandwidth; >> + first_bit = ctz64(hmat_lb->range_bitmap); >> + hmat_lb->base = UINT64_C(1) << first_bit; >> + max_entry = node->bandwidth / hmat_lb->base; >> + last_bit = 64 - clz64(hmat_lb->range_bitmap); >> + >> + /* >> + * For bandwidth, first_bit record the base unit of bandwidth bits, >> + * last_bit record the last bit of the max bandwidth. The max compressed >> + * bandwidth should be less than 0xFFFF (UINT16_MAX) >> + */ >> + if ((last_bit - first_bit) > UINT16_BITS || max_entry >= UINT16_MAX) { > ^^^^^^^^^^^^^^^^^^^ > what bandwidth combination is going to trigger above condition? > Only use (last_bit - first_bit) > UINT16_BITS, we can't trigger error if the max compressed bandwidth is 0xFFFF. Because in that condition, "last_bit - first_bit == UINT16_BITS". So I add "max_entry >= UINT16_MAX" to catch 0xFFFF. For example: Combination 1 (Error): bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) bandwidth2 = ...0000 1111 1111 1111 1111 0000... (max_entry 65535) range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) Combination 2 (Error): bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) bandwidth2 = ...0001 1111 1111 1111 1110 0000... (max_entry 65535) range_bitmap = ...0001 1111 1111 1111 1110 0000... (range is 16 bits) Combination 3 (OK, because bandwidth1 will be compressed to 65534): bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) bandwidth2 = ...0000 0111 1111 1111 1111 0000... (max_entry 32767) range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) Combination 4 (Error): bandwidth1 = ...0000 1111 1111 1111 1111 0000... (max_entry 65535) range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits)
On Wed, 20 Nov 2019 15:55:04 +0800 Tao Xu <tao3.xu@intel.com> wrote: > On 11/19/2019 7:03 PM, Igor Mammedov wrote: > > On Fri, 15 Nov 2019 15:53:46 +0800 > > Tao Xu <tao3.xu@intel.com> wrote: > > > >> From: Liu Jingqi <jingqi.liu@intel.com> > >> > >> Add -numa hmat-lb option to provide System Locality Latency and > >> Bandwidth Information. These memory attributes help to build > >> System Locality Latency and Bandwidth Information Structure(s) > >> in ACPI Heterogeneous Memory Attribute Table (HMAT). > >> > >> Signed-off-by: Liu Jingqi <jingqi.liu@intel.com> > >> Signed-off-by: Tao Xu <tao3.xu@intel.com> > > > > looks good to me, so > > > > Reviewed-by: Igor Mammedov <imammedo@redhat.com> > > > > > > PS: > > also see question below > > > [...] > >> + > >> + hmat_lb->range_bitmap |= node->bandwidth; > >> + first_bit = ctz64(hmat_lb->range_bitmap); > >> + hmat_lb->base = UINT64_C(1) << first_bit; > >> + max_entry = node->bandwidth / hmat_lb->base; > >> + last_bit = 64 - clz64(hmat_lb->range_bitmap); > >> + > >> + /* > >> + * For bandwidth, first_bit record the base unit of bandwidth bits, > >> + * last_bit record the last bit of the max bandwidth. The max compressed > >> + * bandwidth should be less than 0xFFFF (UINT16_MAX) > >> + */ > >> + if ((last_bit - first_bit) > UINT16_BITS || max_entry >= UINT16_MAX) { > > ^^^^^^^^^^^^^^^^^^^ > > what bandwidth combination is going to trigger above condition? > > > Only use (last_bit - first_bit) > UINT16_BITS, we can't trigger error if > the max compressed bandwidth is 0xFFFF. Because in that condition, > "last_bit - first_bit == UINT16_BITS". So I add "max_entry >= > UINT16_MAX" to catch 0xFFFF. For example: > > Combination 1 (Error): > bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) > range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) > bandwidth2 = ...0000 1111 1111 1111 1111 0000... (max_entry 65535) > range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) > > Combination 2 (Error): > bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) > range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) > bandwidth2 = ...0001 1111 1111 1111 1110 0000... (max_entry 65535) > range_bitmap = ...0001 1111 1111 1111 1110 0000... (range is 16 bits) > > Combination 3 (OK, because bandwidth1 will be compressed to 65534): > bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) > range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) > bandwidth2 = ...0000 0111 1111 1111 1111 0000... (max_entry 32767) > range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) > > Combination 4 (Error): > bandwidth1 = ...0000 1111 1111 1111 1111 0000... (max_entry 65535) > range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) ok, I'd use in max/min possible values in bios-tables-test, to make sure that we are testing whole range and would be able to detect a error in case the valid ranges regressed (shrink) and x-fail tests I've asked for in QMP test should detect error other way around.
On 11/20/2019 8:56 PM, Igor Mammedov wrote: > On Wed, 20 Nov 2019 15:55:04 +0800 > Tao Xu <tao3.xu@intel.com> wrote: > >> On 11/19/2019 7:03 PM, Igor Mammedov wrote: >>> On Fri, 15 Nov 2019 15:53:46 +0800 >>> Tao Xu <tao3.xu@intel.com> wrote: >>> >>>> From: Liu Jingqi <jingqi.liu@intel.com> >>>> >>>> Add -numa hmat-lb option to provide System Locality Latency and >>>> Bandwidth Information. These memory attributes help to build >>>> System Locality Latency and Bandwidth Information Structure(s) >>>> in ACPI Heterogeneous Memory Attribute Table (HMAT). >>>> >>>> Signed-off-by: Liu Jingqi <jingqi.liu@intel.com> >>>> Signed-off-by: Tao Xu <tao3.xu@intel.com> >>> >>> looks good to me, so >>> >>> Reviewed-by: Igor Mammedov <imammedo@redhat.com> >>> >>> >>> PS: >>> also see question below >>> >> [...] >>>> + >>>> + hmat_lb->range_bitmap |= node->bandwidth; >>>> + first_bit = ctz64(hmat_lb->range_bitmap); >>>> + hmat_lb->base = UINT64_C(1) << first_bit; >>>> + max_entry = node->bandwidth / hmat_lb->base; >>>> + last_bit = 64 - clz64(hmat_lb->range_bitmap); >>>> + >>>> + /* >>>> + * For bandwidth, first_bit record the base unit of bandwidth bits, >>>> + * last_bit record the last bit of the max bandwidth. The max compressed >>>> + * bandwidth should be less than 0xFFFF (UINT16_MAX) >>>> + */ >>>> + if ((last_bit - first_bit) > UINT16_BITS || max_entry >= UINT16_MAX) { >>> ^^^^^^^^^^^^^^^^^^^ >>> what bandwidth combination is going to trigger above condition? >>> >> Only use (last_bit - first_bit) > UINT16_BITS, we can't trigger error if >> the max compressed bandwidth is 0xFFFF. Because in that condition, >> "last_bit - first_bit == UINT16_BITS". So I add "max_entry >= >> UINT16_MAX" to catch 0xFFFF. For example: >> >> Combination 1 (Error): >> bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) >> range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) >> bandwidth2 = ...0000 1111 1111 1111 1111 0000... (max_entry 65535) >> range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) >> >> Combination 2 (Error): >> bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) >> range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) >> bandwidth2 = ...0001 1111 1111 1111 1110 0000... (max_entry 65535) >> range_bitmap = ...0001 1111 1111 1111 1110 0000... (range is 16 bits) >> >> Combination 3 (OK, because bandwidth1 will be compressed to 65534): >> bandwidth1 = ...0000 1111 1111 1111 1110 0000... (max_entry 32767) >> range_bitmap = ...0000 1111 1111 1111 1110 0000... (range is 15 bits) >> bandwidth2 = ...0000 0111 1111 1111 1111 0000... (max_entry 32767) >> range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) >> >> Combination 4 (Error): >> bandwidth1 = ...0000 1111 1111 1111 1111 0000... (max_entry 65535) >> range_bitmap = ...0000 1111 1111 1111 1111 0000... (range is 16 bits) > > ok, I'd use in max/min possible values in bios-tables-test, > to make sure that we are testing whole range and would be able > to detect a error in case the valid ranges regressed (shrink) > and x-fail tests I've asked for in QMP test should detect > error other way around. > OK I will add these tests.
diff --git a/hw/core/numa.c b/hw/core/numa.c index e60da99293..47c7a96863 100644 --- a/hw/core/numa.c +++ b/hw/core/numa.c @@ -23,6 +23,7 @@ */ #include "qemu/osdep.h" +#include "qemu/units.h" #include "sysemu/hostmem.h" #include "sysemu/numa.h" #include "sysemu/sysemu.h" @@ -198,6 +199,166 @@ void parse_numa_distance(MachineState *ms, NumaDistOptions *dist, Error **errp) ms->numa_state->have_numa_distance = true; } +void parse_numa_hmat_lb(NumaState *numa_state, NumaHmatLBOptions *node, + Error **errp) +{ + int i, first_bit, last_bit; + uint64_t max_entry, temp_base_la; + NodeInfo *numa_info = numa_state->nodes; + HMAT_LB_Info *hmat_lb = + numa_state->hmat_lb[node->hierarchy][node->data_type]; + HMAT_LB_Data lb_data = {}; + HMAT_LB_Data *lb_temp; + + /* Error checking */ + if (node->initiator > numa_state->num_nodes) { + error_setg(errp, "Invalid initiator=%d, it should be less than %d", + node->initiator, numa_state->num_nodes); + return; + } + if (node->target > numa_state->num_nodes) { + error_setg(errp, "Invalid target=%d, it should be less than %d", + node->target, numa_state->num_nodes); + return; + } + if (!numa_info[node->initiator].has_cpu) { + error_setg(errp, "Invalid initiator=%d, it isn't an " + "initiator proximity domain", node->initiator); + return; + } + if (!numa_info[node->target].present) { + error_setg(errp, "The target=%d should point to an existing node", + node->target); + return; + } + + if (!hmat_lb) { + hmat_lb = g_malloc0(sizeof(*hmat_lb)); + numa_state->hmat_lb[node->hierarchy][node->data_type] = hmat_lb; + hmat_lb->list = g_array_new(false, true, sizeof(HMAT_LB_Data)); + } + hmat_lb->hierarchy = node->hierarchy; + hmat_lb->data_type = node->data_type; + lb_data.initiator = node->initiator; + lb_data.target = node->target; + + if (node->data_type <= HMATLB_DATA_TYPE_WRITE_LATENCY) { + /* Input latency data */ + + if (!node->has_latency) { + error_setg(errp, "Missing 'latency' option"); + return; + } + if (node->has_bandwidth) { + error_setg(errp, "Invalid option 'bandwidth' since " + "the data type is latency"); + return; + } + + /* Detect duplicate configuration */ + for (i = 0; i < hmat_lb->list->len; i++) { + lb_temp = &g_array_index(hmat_lb->list, HMAT_LB_Data, i); + + if (node->initiator == lb_temp->initiator && + node->target == lb_temp->target) { + error_setg(errp, "Duplicate configuration of the latency for " + "initiator=%d and target=%d", node->initiator, + node->target); + return; + } + } + + hmat_lb->base = hmat_lb->base ? hmat_lb->base : UINT64_MAX; + + /* Calculate the temporary base and compressed latency */ + max_entry = node->latency; + temp_base_la = 1; + while (QEMU_IS_ALIGNED(max_entry, 10)) { + max_entry /= 10; + temp_base_la *= 10; + } + + /* Calculate the max compressed latency */ + hmat_lb->base = MIN(hmat_lb->base, temp_base_la); + max_entry = node->latency / hmat_lb->base; + hmat_lb->range_bitmap = MAX(hmat_lb->range_bitmap, max_entry); + + /* + * For latency hmat_lb->range_bitmap record the max compressed latency + * which should be less than 0xFFFF (UINT16_MAX) + */ + if (hmat_lb->range_bitmap >= UINT16_MAX) { + error_setg(errp, "Latency %" PRIu64 " between initiator=%d and " + "target=%d should not differ from previously entered " + "min or max values on more than %d", node->latency, + node->initiator, node->target, UINT16_MAX - 1); + return; + } + + /* Set lb_info_provided bit 0 as 1, latency information is provided */ + numa_info[node->target].lb_info_provided |= BIT(0); + lb_data.data = node->latency; + } else if (node->data_type >= HMATLB_DATA_TYPE_ACCESS_BANDWIDTH) { + /* Input bandwidth data */ + + if (!node->has_bandwidth) { + error_setg(errp, "Missing 'bandwidth' option"); + return; + } + if (node->has_latency) { + error_setg(errp, "Invalid option 'latency' since " + "the data type is bandwidth"); + return; + } + if (!QEMU_IS_ALIGNED(node->bandwidth, MiB)) { + error_setg(errp, "Bandwidth %" PRIu64 " between initiator=%d and " + "target=%d should be 1MB aligned", node->bandwidth, + node->initiator, node->target); + return; + } + + /* Detect duplicate configuration */ + for (i = 0; i < hmat_lb->list->len; i++) { + lb_temp = &g_array_index(hmat_lb->list, HMAT_LB_Data, i); + + if (node->initiator == lb_temp->initiator && + node->target == lb_temp->target) { + error_setg(errp, "Duplicate configuration of the bandwidth for " + "initiator=%d and target=%d", node->initiator, + node->target); + return; + } + } + + hmat_lb->range_bitmap |= node->bandwidth; + first_bit = ctz64(hmat_lb->range_bitmap); + hmat_lb->base = UINT64_C(1) << first_bit; + max_entry = node->bandwidth / hmat_lb->base; + last_bit = 64 - clz64(hmat_lb->range_bitmap); + + /* + * For bandwidth, first_bit record the base unit of bandwidth bits, + * last_bit record the last bit of the max bandwidth. The max compressed + * bandwidth should be less than 0xFFFF (UINT16_MAX) + */ + if ((last_bit - first_bit) > UINT16_BITS || max_entry >= UINT16_MAX) { + error_setg(errp, "Bandwidth %" PRIu64 " between initiator=%d and " + "target=%d should not differ from previously entered " + "values on more than %d", node->bandwidth, + node->initiator, node->target, UINT16_MAX - 1); + return; + } + + /* Set lb_info_provided bit 1 as 1, bandwidth information is provided */ + numa_info[node->target].lb_info_provided |= BIT(1); + lb_data.data = node->bandwidth; + } else { + assert(0); + } + + g_array_append_val(hmat_lb->list, lb_data); +} + void set_numa_options(MachineState *ms, NumaOptions *object, Error **errp) { Error *err = NULL; @@ -236,6 +397,19 @@ void set_numa_options(MachineState *ms, NumaOptions *object, Error **errp) machine_set_cpu_numa_node(ms, qapi_NumaCpuOptions_base(&object->u.cpu), &err); break; + case NUMA_OPTIONS_TYPE_HMAT_LB: + if (!ms->numa_state->hmat_enabled) { + error_setg(errp, "ACPI Heterogeneous Memory Attribute Table " + "(HMAT) is disabled, enable it with -machine hmat=on " + "before using any of hmat specific options"); + return; + } + + parse_numa_hmat_lb(ms->numa_state, &object->u.hmat_lb, &err); + if (err) { + goto end; + } + break; default: abort(); } diff --git a/include/sysemu/numa.h b/include/sysemu/numa.h index 788cbec7a2..70f93c83d7 100644 --- a/include/sysemu/numa.h +++ b/include/sysemu/numa.h @@ -14,11 +14,34 @@ struct CPUArchId; #define NUMA_DISTANCE_MAX 254 #define NUMA_DISTANCE_UNREACHABLE 255 +/* the value of AcpiHmatLBInfo flags */ +enum { + HMAT_LB_MEM_MEMORY = 0, + HMAT_LB_MEM_CACHE_1ST_LEVEL = 1, + HMAT_LB_MEM_CACHE_2ND_LEVEL = 2, + HMAT_LB_MEM_CACHE_3RD_LEVEL = 3, + HMAT_LB_LEVELS /* must be the last entry */ +}; + +/* the value of AcpiHmatLBInfo data type */ +enum { + HMAT_LB_DATA_ACCESS_LATENCY = 0, + HMAT_LB_DATA_READ_LATENCY = 1, + HMAT_LB_DATA_WRITE_LATENCY = 2, + HMAT_LB_DATA_ACCESS_BANDWIDTH = 3, + HMAT_LB_DATA_READ_BANDWIDTH = 4, + HMAT_LB_DATA_WRITE_BANDWIDTH = 5, + HMAT_LB_TYPES /* must be the last entry */ +}; + +#define UINT16_BITS 16 + struct NodeInfo { uint64_t node_mem; struct HostMemoryBackend *node_memdev; bool present; bool has_cpu; + uint8_t lb_info_provided; uint16_t initiator; uint8_t distance[MAX_NODES]; }; @@ -28,6 +51,31 @@ struct NumaNodeMem { uint64_t node_plugged_mem; }; +struct HMAT_LB_Data { + uint8_t initiator; + uint8_t target; + uint64_t data; +}; +typedef struct HMAT_LB_Data HMAT_LB_Data; + +struct HMAT_LB_Info { + /* Indicates it's memory or the specified level memory side cache. */ + uint8_t hierarchy; + + /* Present the type of data, access/read/write latency or bandwidth. */ + uint8_t data_type; + + /* The range bitmap of bandwidth for calculating common base */ + uint64_t range_bitmap; + + /* The common base unit for latencies or bandwidths */ + uint64_t base; + + /* Array to store the latencies or bandwidths */ + GArray *list; +}; +typedef struct HMAT_LB_Info HMAT_LB_Info; + struct NumaState { /* Number of NUMA nodes */ int num_nodes; @@ -40,11 +88,16 @@ struct NumaState { /* NUMA nodes information */ NodeInfo nodes[MAX_NODES]; + + /* NUMA nodes HMAT Locality Latency and Bandwidth Information */ + HMAT_LB_Info *hmat_lb[HMAT_LB_LEVELS][HMAT_LB_TYPES]; }; typedef struct NumaState NumaState; void set_numa_options(MachineState *ms, NumaOptions *object, Error **errp); void parse_numa_opts(MachineState *ms); +void parse_numa_hmat_lb(NumaState *numa_state, NumaHmatLBOptions *node, + Error **errp); void numa_complete_configuration(MachineState *ms); void query_numa_node_mem(NumaNodeMem node_mem[], MachineState *ms); extern QemuOptsList qemu_numa_opts; diff --git a/qapi/machine.json b/qapi/machine.json index 27d0e37534..67f5910400 100644 --- a/qapi/machine.json +++ b/qapi/machine.json @@ -426,10 +426,12 @@ # # @cpu: property based CPU(s) to node mapping (Since: 2.10) # +# @hmat-lb: memory latency and bandwidth information (Since: 5.0) +# # Since: 2.1 ## { 'enum': 'NumaOptionsType', - 'data': [ 'node', 'dist', 'cpu' ] } + 'data': [ 'node', 'dist', 'cpu', 'hmat-lb' ] } ## # @NumaOptions: @@ -444,7 +446,8 @@ 'data': { 'node': 'NumaNodeOptions', 'dist': 'NumaDistOptions', - 'cpu': 'NumaCpuOptions' }} + 'cpu': 'NumaCpuOptions', + 'hmat-lb': 'NumaHmatLBOptions' }} ## # @NumaNodeOptions: @@ -557,6 +560,93 @@ 'base': 'CpuInstanceProperties', 'data' : {} } +## +# @HmatLBMemoryHierarchy: +# +# The memory hierarchy in the System Locality Latency +# and Bandwidth Information Structure of HMAT (Heterogeneous +# Memory Attribute Table) +# +# For more information about @HmatLBMemoryHierarchy see +# the chapter 5.2.27.4: Table 5-146: Field "Flags" of ACPI 6.3 spec. +# +# @memory: the structure represents the memory performance +# +# @first-level: first level of memory side cache +# +# @second-level: second level of memory side cache +# +# @third-level: third level of memory side cache +# +# Since: 5.0 +## +{ 'enum': 'HmatLBMemoryHierarchy', + 'data': [ 'memory', 'first-level', 'second-level', 'third-level' ] } + +## +# @HmatLBDataType: +# +# Data type in the System Locality Latency +# and Bandwidth Information Structure of HMAT (Heterogeneous +# Memory Attribute Table) +# +# For more information about @HmatLBDataType see +# the chapter 5.2.27.4: Table 5-146: Field "Data Type" of ACPI 6.3 spec. +# +# @access-latency: access latency (nanoseconds) +# +# @read-latency: read latency (nanoseconds) +# +# @write-latency: write latency (nanoseconds) +# +# @access-bandwidth: access bandwidth (B/s) +# +# @read-bandwidth: read bandwidth (B/s) +# +# @write-bandwidth: write bandwidth (B/s) +# +# Since: 5.0 +## +{ 'enum': 'HmatLBDataType', + 'data': [ 'access-latency', 'read-latency', 'write-latency', + 'access-bandwidth', 'read-bandwidth', 'write-bandwidth' ] } + +## +# @NumaHmatLBOptions: +# +# Set the system locality latency and bandwidth information +# between Initiator and Target proximity Domains. +# +# For more information about @NumaHmatLBOptions see +# the chapter 5.2.27.4: Table 5-146 of ACPI 6.3 spec. +# +# @initiator: the Initiator Proximity Domain. +# +# @target: the Target Proximity Domain. +# +# @hierarchy: the Memory Hierarchy. Indicates the performance +# of memory or side cache. +# +# @data-type: presents the type of data, access/read/write +# latency or hit latency. +# +# @latency: the value of latency from @initiator to @target proximity domain, +# the latency unit is "ns(nanosecond)". +# +# @bandwidth: the value of bandwidth between @initiator and @target proximity +# domain, the bandwidth unit is "B(/s)". +# +# Since: 5.0 +## +{ 'struct': 'NumaHmatLBOptions', + 'data': { + 'initiator': 'uint16', + 'target': 'uint16', + 'hierarchy': 'HmatLBMemoryHierarchy', + 'data-type': 'HmatLBDataType', + '*latency': 'time', + '*bandwidth': 'size' }} + ## # @HostMemPolicy: # diff --git a/qemu-options.hx b/qemu-options.hx index 63f6b33322..929d275450 100644 --- a/qemu-options.hx +++ b/qemu-options.hx @@ -168,16 +168,19 @@ DEF("numa", HAS_ARG, QEMU_OPTION_numa, "-numa node[,mem=size][,cpus=firstcpu[-lastcpu]][,nodeid=node][,initiator=node]\n" "-numa node[,memdev=id][,cpus=firstcpu[-lastcpu]][,nodeid=node][,initiator=node]\n" "-numa dist,src=source,dst=destination,val=distance\n" - "-numa cpu,node-id=node[,socket-id=x][,core-id=y][,thread-id=z]\n", + "-numa cpu,node-id=node[,socket-id=x][,core-id=y][,thread-id=z]\n" + "-numa hmat-lb,initiator=node,target=node,hierarchy=memory|first-level|second-level|third-level,data-type=access-latency|read-latency|write-latency[,latency=lat][,bandwidth=bw]\n", QEMU_ARCH_ALL) STEXI @item -numa node[,mem=@var{size}][,cpus=@var{firstcpu}[-@var{lastcpu}]][,nodeid=@var{node}][,initiator=@var{initiator}] @itemx -numa node[,memdev=@var{id}][,cpus=@var{firstcpu}[-@var{lastcpu}]][,nodeid=@var{node}][,initiator=@var{initiator}] @itemx -numa dist,src=@var{source},dst=@var{destination},val=@var{distance} @itemx -numa cpu,node-id=@var{node}[,socket-id=@var{x}][,core-id=@var{y}][,thread-id=@var{z}] +@itemx -numa hmat-lb,initiator=@var{node},target=@var{node},hierarchy=@var{hierarchy},data-type=@var{tpye}[,latency=@var{lat}][,bandwidth=@var{bw}] @findex -numa Define a NUMA node and assign RAM and VCPUs to it. Set the NUMA distance from a source node to a destination node. +Set the ACPI Heterogeneous Memory Attributes for the given nodes. Legacy VCPU assignment uses @samp{cpus} option where @var{firstcpu} and @var{lastcpu} are CPU indexes. Each @@ -256,6 +259,50 @@ specified resources, it just assigns existing resources to NUMA nodes. This means that one still has to use the @option{-m}, @option{-smp} options to allocate RAM and VCPUs respectively. +Use @samp{hmat-lb} to set System Locality Latency and Bandwidth Information +between initiator and target NUMA nodes in ACPI Heterogeneous Attribute Memory Table (HMAT). +Initiator NUMA node can create memory requests, usually it has one or more processors. +Target NUMA node contains addressable memory. + +In @samp{hmat-lb} option, @var{node} are NUMA node IDs. @var{hierarchy} is the memory +hierarchy of the target NUMA node: if @var{hierarchy} is 'memory', the structure +represents the memory performance; if @var{hierarchy} is 'first-level|second-level|third-level', +this structure represents aggregated performance of memory side caches for each domain. +@var{type} of 'data-type' is type of data represented by this structure instance: +if 'hierarchy' is 'memory', 'data-type' is 'access|read|write' latency or 'access|read|write' +bandwidth of the target memory; if 'hierarchy' is 'first-level|second-level|third-level', +'data-type' is 'access|read|write' hit latency or 'access|read|write' hit bandwidth of the +target memory side cache. + +@var{lat} is latency value, the possible value and units are NUM[ns|us|ms] +(nanosecond|microsecond|millisecond), the recommended unit is 'ns'. @var{bw} is +bandwidth value, the possible value and units are NUM[M|G|T], mean that the +bandwidth value are NUM byte (or MB/s, GB/s or TB/s depending on used suffix). +Note that if NUM is 0, means the corresponding latency or bandwidth information +is not provided. And if input numbers without any unit, the latency unit will be +'ns' and the bandwidth will be B/s. + +For example, the following options describe 2 NUMA nodes. Node 0 has 2 cpus and +a ram, node 1 has only a ram. The processors in node 0 access memory in node +0 with access-latency 5 nanoseconds, access-bandwidth is 200 MB/s; +The processors in NUMA node 0 access memory in NUMA node 1 with access-latency 10 +nanoseconds, access-bandwidth is 100 MB/s. +@example +-machine hmat=on \ +-m 2G \ +-object memory-backend-ram,size=1G,id=m0 \ +-object memory-backend-ram,size=1G,id=m1 \ +-smp 2 \ +-numa node,nodeid=0,memdev=m0 \ +-numa node,nodeid=1,memdev=m1,initiator=0 \ +-numa cpu,node-id=0,socket-id=0 \ +-numa cpu,node-id=0,socket-id=1 \ +-numa hmat-lb,initiator=0,target=0,hierarchy=memory,data-type=access-latency,latency=5ns \ +-numa hmat-lb,initiator=0,target=0,hierarchy=memory,data-type=access-bandwidth,bandwidth=200M \ +-numa hmat-lb,initiator=0,target=1,hierarchy=memory,data-type=access-latency,latency=10ns \ +-numa hmat-lb,initiator=0,target=1,hierarchy=memory,data-type=access-bandwidth,bandwidth=100M +@end example + ETEXI DEF("add-fd", HAS_ARG, QEMU_OPTION_add_fd,