From patchwork Wed Apr 25 23:31:21 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jeremy Linton X-Patchwork-Id: 134390 Delivered-To: patch@linaro.org Received: by 10.46.151.6 with SMTP id r6csp1532933lji; Wed, 25 Apr 2018 17:26:44 -0700 (PDT) X-Google-Smtp-Source: AIpwx4+0rwPQ23QmrwyN/1dICScV/JiCEic34b7jHstdy5k1grhDI/wRyr9JwFINN9jRgpOYYpxG X-Received: by 10.98.2.72 with SMTP id 69mr29900805pfc.12.1524702404316; Wed, 25 Apr 2018 17:26:44 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1524702404; cv=none; d=google.com; s=arc-20160816; b=yh4so1Imm4BqhdMoJwBtsE6muRD5bRH5ri8/OUtLdhGFa/upY4Uc9Ej/dAjdgbRw22 N/EMt/QapIGHhrvnthaapfXjExNsKGvmRkiD3atMQkgG1Xvr+fH7HIJrulikeViXQK9N 32yFfY7ZQTJP7b0HtxqK6RPhzPWo1maJe0RcoDeBP699VnDSLPzY9Qyy2GSwmMNe0Coi S6UvKuvq+HnJgWi/tYVUea0s7FlKmmEoiVnF53k+BIn7g6UcwLDQJQ8kTbvtFJ6Oi/kT oLheZgGwnd1BYoAowyL50xt2Y7tzLKDjhbt23r7wwGTH7n7WrxV1JtDthfHWXrU9LV1d Ejbg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:arc-authentication-results; bh=QwhnOt04PddViD3urVl6ntTl9TkwS6jhYwGF9Q9Ck30=; b=JnGphXfSM0EkfjWjdfNPZfu0KSxdNdIB4o2Mf0VKo/k1V0I5zo2dPuY2FlcoNSPKQn 7kx1uCSRPYbKC8PBXsoU/ms6qbQ6v68mYyS+RCSSR9Usizb/d6RmoOUKMAzMpgLoNwpu //6RljhmtucWYucVEaBAGntFvPe2gLhfsEQM9wmmdpNs/kvrDphjb8v5ia9F7g1Kx6ol 0Js5Kpnt3wLga6rBPj/XdBu7WUqoj/t4sEmhNPqJ9xza06vPz3MigIRQZjHSbH1cCfiH uuCvZmQUZdxhcDMgRNhyBhAQ2h4hwudsdaQL2aK8vA7GPe2750pi1yX7qo3njUS5YZoW 6bdA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id bb4-v6si12911680plb.169.2018.04.25.17.26.43; Wed, 25 Apr 2018 17:26:44 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752972AbeDZA0l (ORCPT + 29 others); Wed, 25 Apr 2018 20:26:41 -0400 Received: from foss.arm.com ([217.140.101.70]:45886 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752129AbeDZA02 (ORCPT ); Wed, 25 Apr 2018 20:26:28 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id B7AF7165D; Wed, 25 Apr 2018 17:26:27 -0700 (PDT) Received: from beelzebub.austin.arm.com (beelzebub.austin.arm.com [10.118.12.119]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPA id A5BE93F590; Wed, 25 Apr 2018 17:26:26 -0700 (PDT) From: Jeremy Linton To: linux-acpi@vger.kernel.org Cc: Sudeep.Holla@arm.com, linux-arm-kernel@lists.infradead.org, Lorenzo.Pieralisi@arm.com, hanjun.guo@linaro.org, rjw@rjwysocki.net, Will.Deacon@arm.com, Catalin.Marinas@arm.com, gregkh@linuxfoundation.org, Mark.Rutland@arm.com, linux-kernel@vger.kernel.org, linux-riscv@lists.infradead.org, wangxiongfeng2@huawei.com, vkilari@codeaurora.org, ahs3@redhat.com, Dietmar.Eggemann@arm.com, Morten.Rasmussen@arm.com, palmer@sifive.com, lenb@kernel.org, john.garry@huawei.com, austinwc@codeaurora.org, tnowicki@caviumnetworks.com, jhugo@qti.qualcomm.com, timur@qti.qualcomm.com, ard.biesheuvel@linaro.org, Jeremy Linton Subject: [PATCH v8 13/13] arm64: topology: divorce MC scheduling domain from core_siblings Date: Wed, 25 Apr 2018 18:31:21 -0500 Message-Id: <20180425233121.13270-14-jeremy.linton@arm.com> X-Mailer: git-send-email 2.13.6 In-Reply-To: <20180425233121.13270-1-jeremy.linton@arm.com> References: <20180425233121.13270-1-jeremy.linton@arm.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Now that we have an accurate view of the physical topology we need to represent it correctly to the scheduler. Generally MC should equal the LLC in the system, but there are a number of special cases that need to be dealt with. In the case of NUMA in socket, we need to assure that the sched domain we build for the MC layer isn't larger than the DIE above it. Similarly for LLC's that might exist in cross socket interconnect or directory hardware we need to assure that MC is shrunk to the socket or NUMA node. This patch builds a sibling mask for the LLC, and then picks the smallest of LLC, socket siblings, or NUMA node siblings, which gives us the behavior described above. This is ever so slightly different than the similar alternative where we look for a cache layer less than or equal to the socket/NUMA siblings. The logic to pick the MC layer affects all arm64 machines, but only changes the behavior for DT/MPIDR systems if the NUMA domain is smaller than the core siblings (generally set to the cluster). Potentially this fixes a possible bug in DT systems, but really it only affects ACPI systems where the core siblings is correctly set to the socket siblings. Thus all currently available ACPI systems should have MC equal to LLC, including the NUMA in socket machines where the LLC is partitioned between the NUMA nodes. Signed-off-by: Jeremy Linton --- arch/arm64/include/asm/topology.h | 2 ++ arch/arm64/kernel/topology.c | 32 +++++++++++++++++++++++++++++++- 2 files changed, 33 insertions(+), 1 deletion(-) -- 2.13.6 diff --git a/arch/arm64/include/asm/topology.h b/arch/arm64/include/asm/topology.h index 6b10459e6905..df48212f767b 100644 --- a/arch/arm64/include/asm/topology.h +++ b/arch/arm64/include/asm/topology.h @@ -8,8 +8,10 @@ struct cpu_topology { int thread_id; int core_id; int package_id; + int llc_id; cpumask_t thread_sibling; cpumask_t core_sibling; + cpumask_t llc_siblings; }; extern struct cpu_topology cpu_topology[NR_CPUS]; diff --git a/arch/arm64/kernel/topology.c b/arch/arm64/kernel/topology.c index bd1aae438a31..20b4341dc527 100644 --- a/arch/arm64/kernel/topology.c +++ b/arch/arm64/kernel/topology.c @@ -13,6 +13,7 @@ #include #include +#include #include #include #include @@ -214,7 +215,19 @@ EXPORT_SYMBOL_GPL(cpu_topology); const struct cpumask *cpu_coregroup_mask(int cpu) { - return &cpu_topology[cpu].core_sibling; + const cpumask_t *core_mask = cpumask_of_node(cpu_to_node(cpu)); + + /* Find the smaller of NUMA, core or LLC siblings */ + if (cpumask_subset(&cpu_topology[cpu].core_sibling, core_mask)) { + /* not numa in package, lets use the package siblings */ + core_mask = &cpu_topology[cpu].core_sibling; + } + if (cpu_topology[cpu].llc_id != -1) { + if (cpumask_subset(&cpu_topology[cpu].llc_siblings, core_mask)) + core_mask = &cpu_topology[cpu].llc_siblings; + } + + return core_mask; } static void update_siblings_masks(unsigned int cpuid) @@ -226,6 +239,9 @@ static void update_siblings_masks(unsigned int cpuid) for_each_possible_cpu(cpu) { cpu_topo = &cpu_topology[cpu]; + if (cpuid_topo->llc_id == cpu_topo->llc_id) + cpumask_set_cpu(cpu, &cpuid_topo->llc_siblings); + if (cpuid_topo->package_id != cpu_topo->package_id) continue; @@ -291,6 +307,10 @@ static void __init reset_cpu_topology(void) cpu_topo->core_id = 0; cpu_topo->package_id = -1; + cpu_topo->llc_id = -1; + cpumask_clear(&cpu_topo->llc_siblings); + cpumask_set_cpu(cpu, &cpu_topo->llc_siblings); + cpumask_clear(&cpu_topo->core_sibling); cpumask_set_cpu(cpu, &cpu_topo->core_sibling); cpumask_clear(&cpu_topo->thread_sibling); @@ -311,6 +331,8 @@ static int __init parse_acpi_topology(void) is_threaded = read_cpuid_mpidr() & MPIDR_MT_BITMASK; for_each_possible_cpu(cpu) { + int i; + topology_id = find_acpi_cpu_topology(cpu, 0); if (topology_id < 0) return topology_id; @@ -325,6 +347,14 @@ static int __init parse_acpi_topology(void) } topology_id = find_acpi_cpu_topology_package(cpu); cpu_topology[cpu].package_id = topology_id; + + i = acpi_find_last_cache_level(cpu); + + if (i > 0) { + topology_id = find_acpi_cpu_cache_topology(cpu, i); + if (topology_id > 0) + cpu_topology[cpu].llc_id = topology_id; + } } return 0;