qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Gavin Shan <gshan@redhat.com>
To: Igor Mammedov <imammedo@redhat.com>,
	"wangyanan (Y)" <wangyanan55@huawei.com>
Cc: peter.maydell@linaro.org, drjones@redhat.com,
	richard.henderson@linaro.org, qemu-devel@nongnu.org,
	zhenyzha@redhat.com, qemu-arm@nongnu.org, shan.gavin@gmail.com
Subject: Re: [PATCH v2 2/3] hw/acpi/aml-build: Use existing CPU topology to build PPTT table
Date: Wed, 23 Mar 2022 11:31:48 +0800	[thread overview]
Message-ID: <3f55457e-bf2e-03f4-1eef-17d277c4d094@redhat.com> (raw)
In-Reply-To: <20220318142811.1d8eb072@redhat.com>

Hi Igor and Yanan,

On 3/18/22 9:28 PM, Igor Mammedov wrote:
> On Fri, 18 Mar 2022 14:34:12 +0800
> "wangyanan (Y)" <wangyanan55@huawei.com> wrote:
>> On 2022/3/3 11:11, Gavin Shan wrote:
>>> When the PPTT table is built, the CPU topology is re-calculated, but
>>> it's unecessary because the CPU topology, except the cluster IDs,
>>> has been populated in virt_possible_cpu_arch_ids() on arm/virt machine.
>>>
>>> This avoids to re-calculate the CPU topology by reusing the existing
>>> one in ms->possible_cpus. However, the cluster ID for the CPU instance
>>> has to be calculated dynamically because there is no corresponding
>>> field in struct CpuInstanceProperties. Currently, the only user of
>>> build_pptt() is arm/virt machine.
>>>
>>> Signed-off-by: Gavin Shan <gshan@redhat.com>
>>> ---
>>>    hw/acpi/aml-build.c | 106 ++++++++++++++++++++++++++++++++++----------
>>>    1 file changed, 82 insertions(+), 24 deletions(-)
>>>
>>> diff --git a/hw/acpi/aml-build.c b/hw/acpi/aml-build.c
>>> index 8966e16320..572cf5fc00 100644
>>> --- a/hw/acpi/aml-build.c
>>> +++ b/hw/acpi/aml-build.c
>>> @@ -2002,18 +2002,27 @@ void build_pptt(GArray *table_data, BIOSLinker *linker, MachineState *ms,
>>>                    const char *oem_id, const char *oem_table_id)
>>>    {
>>>        MachineClass *mc = MACHINE_GET_CLASS(ms);
>>> +    CPUArchIdList *cpus = ms->possible_cpus;
>>> +    GQueue *socket_list = g_queue_new();
>>> +    GQueue *cluster_list = g_queue_new();
>>> +    GQueue *core_list = g_queue_new();
>>>        GQueue *list = g_queue_new();
>>>        guint pptt_start = table_data->len;
>>>        guint parent_offset;
>>>        guint length, i;
>>> -    int uid = 0;
>>> -    int socket;
>>> +    int n, id, socket_id, cluster_id, core_id, thread_id;
>>>        AcpiTable table = { .sig = "PPTT", .rev = 2,
>>>                            .oem_id = oem_id, .oem_table_id = oem_table_id };
>>>    
>>>        acpi_table_begin(&table, table_data);
>>>    
>>> -    for (socket = 0; socket < ms->smp.sockets; socket++) {
>>> +    for (n = 0; n < cpus->len; n++) {
>>> +        socket_id = cpus->cpus[n].props.socket_id;
>>> +        if (g_queue_find(socket_list, GUINT_TO_POINTER(socket_id))) {
>>> +            continue;
>>> +        }
>>> +
>>> +        g_queue_push_tail(socket_list, GUINT_TO_POINTER(socket_id));
>>>            g_queue_push_tail(list,
>>>                GUINT_TO_POINTER(table_data->len - pptt_start));
>>>            build_processor_hierarchy_node(
>>> @@ -2023,65 +2032,114 @@ void build_pptt(GArray *table_data, BIOSLinker *linker, MachineState *ms,
>>>                 * of a physical package
>>>                 */
>>>                (1 << 0),
>>> -            0, socket, NULL, 0);
>>> +            0, socket_id, NULL, 0);
>>>        }
>>>    
>>>        if (mc->smp_props.clusters_supported) {
>>>            length = g_queue_get_length(list);
>>>            for (i = 0; i < length; i++) {
>>> -            int cluster;
>>> -
>>>                parent_offset = GPOINTER_TO_UINT(g_queue_pop_head(list));
>>> -            for (cluster = 0; cluster < ms->smp.clusters; cluster++) {
>>> +            socket_id = GPOINTER_TO_UINT(g_queue_pop_head(socket_list));
>>> +
>>> +            for (n = 0; n < cpus->len; n++) {
>>> +                if (cpus->cpus[n].props.socket_id != socket_id) {
>>> +                    continue;
>>> +                }
>>> +
>>> +                /*
>>> +                 * We have to calculate the cluster ID because it isn't
>>> +                 * available in the CPU instance properties.
>>> +                 */
>> Since we need cluster ID now, maybe we can simply make it supported
>> in the CPU instance properties.
> 
> agreed
> 

Thanks for your review. I will add it in v3. FYI, the addition
needs to be done in the PATCH[v3 01/04] where the CPU topology
is populated :)

Thanks,
Gavin

>>> +                cluster_id = cpus->cpus[n].props.thread_id /
>>> +                             (ms->smp.cores * ms->smp.threads);
>>> +                if (g_queue_find(cluster_list, GUINT_TO_POINTER(cluster_id))) {
>>> +                    continue;
>>> +                }
>>> +
>>> +                g_queue_push_tail(cluster_list, GUINT_TO_POINTER(cluster_id));
>>>                    g_queue_push_tail(list,
>>>                        GUINT_TO_POINTER(table_data->len - pptt_start));
>>>                    build_processor_hierarchy_node(
>>>                        table_data,
>>>                        (0 << 0), /* not a physical package */
>>> -                    parent_offset, cluster, NULL, 0);
>>> +                    parent_offset, cluster_id, NULL, 0);
>>>                }
>>>            }
>>>        }
>>>    
>>>        length = g_queue_get_length(list);
>>>        for (i = 0; i < length; i++) {
>>> -        int core;
>>> -
>>>            parent_offset = GPOINTER_TO_UINT(g_queue_pop_head(list));
>>> -        for (core = 0; core < ms->smp.cores; core++) {
>>> -            if (ms->smp.threads > 1) {
>>> -                g_queue_push_tail(list,
>>> -                    GUINT_TO_POINTER(table_data->len - pptt_start));
>>> -                build_processor_hierarchy_node(
>>> -                    table_data,
>>> -                    (0 << 0), /* not a physical package */
>>> -                    parent_offset, core, NULL, 0);
>>> -            } else {
>>> +        if (!mc->smp_props.clusters_supported) {
>>> +            socket_id = GPOINTER_TO_UINT(g_queue_pop_head(socket_list));
>>> +        } else {
>>> +            cluster_id = GPOINTER_TO_UINT(g_queue_pop_head(cluster_list));
>>> +        }
>>> +
>>> +        for (n = 0; n < cpus->len; n++) {
>>> +            if (!mc->smp_props.clusters_supported &&
>>> +                cpus->cpus[n].props.socket_id != socket_id) {
>>> +                continue;
>>> +            }
>>> +
>>> +            /*
>>> +             * We have to calculate the cluster ID because it isn't
>>> +             * available in the CPU instance properties.
>>> +             */
>>> +            id = cpus->cpus[n].props.thread_id /
>>> +                (ms->smp.cores * ms->smp.threads);
>>> +            if (mc->smp_props.clusters_supported && id != cluster_id) {
>>> +                continue;
>>> +            }
>>> +
>>> +            core_id = cpus->cpus[n].props.core_id;
>>> +            if (ms->smp.threads <= 1) {
>>>                    build_processor_hierarchy_node(
>>>                        table_data,
>>>                        (1 << 1) | /* ACPI Processor ID valid */
>>>                        (1 << 3),  /* Node is a Leaf */
>>> -                    parent_offset, uid++, NULL, 0);
>>> +                    parent_offset, core_id, NULL, 0);
>>> +                continue;
>>>                }
>>> +
>>> +            if (g_queue_find(core_list, GUINT_TO_POINTER(core_id))) {
>>> +                continue;
>>> +            }
>>> +
>>> +            g_queue_push_tail(core_list, GUINT_TO_POINTER(core_id));
>>> +            g_queue_push_tail(list,
>>> +                GUINT_TO_POINTER(table_data->len - pptt_start));
>>> +            build_processor_hierarchy_node(
>>> +                table_data,
>>> +                (0 << 0), /* not a physical package */
>>> +                parent_offset, core_id, NULL, 0);
>>>            }
>>>        }
>>>    
>>>        length = g_queue_get_length(list);
>>>        for (i = 0; i < length; i++) {
>>> -        int thread;
>>> -
>>>            parent_offset = GPOINTER_TO_UINT(g_queue_pop_head(list));
>>> -        for (thread = 0; thread < ms->smp.threads; thread++) {
>>> +        core_id = GPOINTER_TO_UINT(g_queue_pop_head(core_list));
>>> +
>>> +        for (n = 0; n < cpus->len; n++) {
>>> +            if (cpus->cpus[n].props.core_id != core_id) {
>>> +                continue;
>>> +            }
>>> +
>>> +            thread_id = cpus->cpus[n].props.thread_id;
>>>                build_processor_hierarchy_node(
>>>                    table_data,
>>>                    (1 << 1) | /* ACPI Processor ID valid */
>>>                    (1 << 2) | /* Processor is a Thread */
>>>                    (1 << 3),  /* Node is a Leaf */
>>> -                parent_offset, uid++, NULL, 0);
>>> +                parent_offset, thread_id, NULL, 0);
>>>            }
>>>        }
>>>    
>>>        g_queue_free(list);
>>> +    g_queue_free(core_list);
>>> +    g_queue_free(cluster_list);
>>> +    g_queue_free(socket_list);
>>>        acpi_table_end(linker, &table);
>>>    }
>>>      
>>
> 



  reply	other threads:[~2022-03-23  3:33 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-03-03  3:11 [PATCH v2 0/3] hw/arm/virt: Fix CPU's default NUMA node ID Gavin Shan
2022-03-03  3:11 ` [PATCH v2 1/3] " Gavin Shan
2022-03-18  6:23   ` wangyanan (Y) via
2022-03-18  9:56     ` Igor Mammedov
2022-03-18 13:00       ` wangyanan (Y) via
2022-03-18 13:27         ` Igor Mammedov
2022-03-21  2:28           ` wangyanan (Y) via
2022-03-23  3:26             ` Gavin Shan
2022-03-23  3:29             ` Gavin Shan
2022-03-03  3:11 ` [PATCH v2 2/3] hw/acpi/aml-build: Use existing CPU topology to build PPTT table Gavin Shan
2022-03-18  6:34   ` wangyanan (Y) via
2022-03-18 13:28     ` Igor Mammedov
2022-03-23  3:31       ` Gavin Shan [this message]
2022-03-03  3:11 ` [PATCH v2 3/3] hw/arm/virt: Unify ACPI processor ID in MADT and SRAT table Gavin Shan
2022-03-14  6:24 ` [PATCH v2 0/3] hw/arm/virt: Fix CPU's default NUMA node ID Gavin Shan

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=3f55457e-bf2e-03f4-1eef-17d277c4d094@redhat.com \
    --to=gshan@redhat.com \
    --cc=drjones@redhat.com \
    --cc=imammedo@redhat.com \
    --cc=peter.maydell@linaro.org \
    --cc=qemu-arm@nongnu.org \
    --cc=qemu-devel@nongnu.org \
    --cc=richard.henderson@linaro.org \
    --cc=shan.gavin@gmail.com \
    --cc=wangyanan55@huawei.com \
    --cc=zhenyzha@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).