From patchwork Wed Mar 31 11:00:48 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Daniel Lezcano X-Patchwork-Id: 412873 Delivered-To: patch@linaro.org Received: by 2002:a02:8562:0:0:0:0:0 with SMTP id g89csp386285jai; Wed, 31 Mar 2021 04:02:17 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwnSKzpnbLdJckEJS3BNj3relmmrTQhtCRKsMl0hvYI+/aHBFU9cUuWZCoksaGdFzDhsAFi X-Received: by 2002:a05:6402:c1:: with SMTP id i1mr2880769edu.315.1617188537482; Wed, 31 Mar 2021 04:02:17 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1617188537; cv=none; d=google.com; s=arc-20160816; b=k27sJOyV18Hk6Fh+epQC2Edem7dmA6luS5aV5XCXfwChfWu32TalUWHMQwXX0kYBHY uYc1K8y6aHcS3HsrlYwNnVSclTN6TScJERgyR4a2QWXC13Qjnvqbp21a6l8zTWI7ZCGM Gb7POWcKHnyUNT/F/inXPRpmQvdrArjH3j96Y78VSirm5LXm0kQC6Qivj7HBmnx4+tIA g77tlqJBLeJYPtOb8cEloA32jU8yKzGKLs6S/XnqjBMTzGaKh1fPZJz90KHbVFEkmudG g0UMjDxDxMkvNCe4QvA4MWFBuuQFhYGbJyLUfPKr8hl3bvwNJJD7SRrH53rJ5fDiV5DA +MQA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:references:in-reply-to:message-id:date:subject :cc:to:from:dkim-signature; bh=IN1SjCUiJXttEzEwKJpOwO8v8lxJB5yWOIgxTzXkjf4=; b=k9d/Y8Y8TZBm+adt+ulA7RE4M3uL4L8yo3/yVFAyoXNVpE7QdOgbZp2gYmBhHBAIhW 4fWfr5BFxRStgiM5XlK4/WmFhnb/W0cg+pYhJeKbs61rT8N9rZL8lmCETP/n0kfJgIjT tdKQLfAUBCNDF148rV20xkE5vEK5MQcWKCth4rgL/XAX5LZK2JQPMtFNOv9SLMF0Np9H 4vXiaUqViBZ7d4pjn4T3CfKuaq0TdX2wUUqME7k5kVp8YtBq4SpeAzJqMEI5CMftkUPX MuvpUbOgcPVVgET8AkNRn/GJraZDWsmWLhChgsKgrWlcU/5j2h3NarUEiT3AZTsz7zE3 tYtA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b="mES/btcE"; spf=pass (google.com: domain of linux-pm-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-pm-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id r16si1445467edw.458.2021.03.31.04.02.17; Wed, 31 Mar 2021 04:02:17 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-pm-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b="mES/btcE"; spf=pass (google.com: domain of linux-pm-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-pm-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235466AbhCaLBq (ORCPT + 7 others); Wed, 31 Mar 2021 07:01:46 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:57376 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235524AbhCaLBJ (ORCPT ); Wed, 31 Mar 2021 07:01:09 -0400 Received: from mail-wr1-x433.google.com (mail-wr1-x433.google.com [IPv6:2a00:1450:4864:20::433]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id AB59DC0613D7 for ; Wed, 31 Mar 2021 04:01:08 -0700 (PDT) Received: by mail-wr1-x433.google.com with SMTP id z2so19216019wrl.5 for ; Wed, 31 Mar 2021 04:01:08 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=IN1SjCUiJXttEzEwKJpOwO8v8lxJB5yWOIgxTzXkjf4=; b=mES/btcERywUzdWQs1lJxM93buJiHXH0MILQ1g6KKVjqG1sh10JFCAQI2AA2xTNX46 fYc9nHcE8G81eSRcDO9fesgjtZcNXyE+ZCZ07MYPvDxkig2XST3Sv1VBAkV4ZAAWHcMW P1E8dzFXRygT10p1+Z8dhDxqY/icRWDlEcKKCrRpW/GJPrIEVQQQwUoa9mphcogE9lwB Ye65w2zXC7BjafTuGEhrNfvG+5KIy3OVbmgePRstcMZTq9bZAoZccoa9Rzb/h16fW7Cf jzPH422YDX0/p8XLb/Vpt/n8VlCuaNRqgpESHWRMx4BiknJaRQkqylA3nORJ6GGoFIxL 2nTw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=IN1SjCUiJXttEzEwKJpOwO8v8lxJB5yWOIgxTzXkjf4=; b=FjGmMki8Z1ASxILvc09riPqq5yl6H2zmsMoS5mQR5nLE0bTKT4aWoyi55n3dcmnpvK /DOBP3ekuZqAwCSDHyIJBbaqLJI7tZ7bEFi1RqjSoB3eok8onXXieLP6CqboCnDLufSF 6IEe1lC+GJvDVshtlQKkdNymqOiekHOfU8btYQZXI3GoGZ0jChT29VHni5RlX9c3z5PK fpZEWPpR1O0uta1uxB7IjIU45ezSWUzMBwXW0qhuD6gPGM7BDlVXpOJGjc96EynWPPZO w6NUqR3pb6fPfnRML1g9uTke4EERKvuT1bQiBqqZ4xWIVVJByg6RI6171gXsqacySD48 IVBg== X-Gm-Message-State: AOAM533gL3LyioHlj7pJ12mxl3JL5A/HBJSVP1SLM5T/S2f3OHo2E4/2 ufwEfImLXn2Z7/OaQmHFntkYUQ== X-Received: by 2002:a5d:6684:: with SMTP id l4mr2973647wru.381.1617188467320; Wed, 31 Mar 2021 04:01:07 -0700 (PDT) Received: from localhost.localdomain ([82.142.26.252]) by smtp.gmail.com with ESMTPSA id u23sm3376275wmn.26.2021.03.31.04.01.06 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 31 Mar 2021 04:01:06 -0700 (PDT) From: Daniel Lezcano To: daniel.lezcano@linaro.org Cc: linux-pm@vger.kernel.org, linux-kernel@vger.kernel.org, lukasz.luba@arm.com, rafael@kernel.org, gregkh@linuxfoundation.org Subject: [PATCH v5 5/5] powercap/drivers/dtpm: Scale the power with the load Date: Wed, 31 Mar 2021 13:00:48 +0200 Message-Id: <20210331110048.24956-5-daniel.lezcano@linaro.org> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20210331110048.24956-1-daniel.lezcano@linaro.org> References: <20210331110048.24956-1-daniel.lezcano@linaro.org> Precedence: bulk List-ID: X-Mailing-List: linux-pm@vger.kernel.org Currently the power consumption is based on the current OPP power assuming the entire performance domain is fully loaded. That gives very gross power estimation and we can do much better by using the load to scale the power consumption. Use the utilization to normalize and scale the power usage over the max possible power. Tested on a rock960 with 2 big CPUS, the power consumption estimation conforms with the expected one. Before this change: ~$ ~/dhrystone -t 1 -l 10000& ~$ cat /sys/devices/virtual/powercap/dtpm/dtpm:0/dtpm:0:1/constraint_0_max_power_uw 2260000 After this change: ~$ ~/dhrystone -t 1 -l 10000& ~$ cat /sys/devices/virtual/powercap/dtpm/dtpm:0/dtpm:0:1/constraint_0_max_power_uw 1130000 ~$ ~/dhrystone -t 2 -l 10000& ~$ cat /sys/devices/virtual/powercap/dtpm/dtpm:0/dtpm:0:1/constraint_0_max_power_uw 2260000 Signed-off-by: Daniel Lezcano Reviewed-by: Lukasz Luba --- V3: - Fixed uninitialized 'cpu' in scaled_power_uw() V2: - Replaced cpumask by em_span_cpus - Changed 'util' metrics variable types - Optimized utilization scaling power computation - Renamed parameter name for scale_pd_power_uw() --- drivers/powercap/dtpm_cpu.c | 46 +++++++++++++++++++++++++++++++------ 1 file changed, 39 insertions(+), 7 deletions(-) -- 2.17.1 diff --git a/drivers/powercap/dtpm_cpu.c b/drivers/powercap/dtpm_cpu.c index f4092d1b01d7..eae35ae3c42e 100644 --- a/drivers/powercap/dtpm_cpu.c +++ b/drivers/powercap/dtpm_cpu.c @@ -68,27 +68,59 @@ static u64 set_pd_power_limit(struct dtpm *dtpm, u64 power_limit) return power_limit; } +static u64 scale_pd_power_uw(struct cpumask *pd_mask, u64 power) +{ + unsigned long max = 0, sum_util = 0; + int cpu; + + for_each_cpu_and(cpu, pd_mask, cpu_online_mask) { + + /* + * The capacity is the same for all CPUs belonging to + * the same perf domain, so a single call to + * arch_scale_cpu_capacity() is enough. However, we + * need the CPU parameter to be initialized by the + * loop, so the call ends up in this block. + * + * We can initialize 'max' with a cpumask_first() call + * before the loop but the bits computation is not + * worth given the arch_scale_cpu_capacity() just + * returns a value where the resulting assembly code + * will be optimized by the compiler. + */ + max = arch_scale_cpu_capacity(cpu); + sum_util += sched_cpu_util(cpu, max); + } + + /* + * In the improbable case where all the CPUs of the perf + * domain are offline, 'max' will be zero and will lead to an + * illegal operation with a zero division. + */ + return max ? (power * ((sum_util << 10) / max)) >> 10 : 0; +} + static u64 get_pd_power_uw(struct dtpm *dtpm) { struct dtpm_cpu *dtpm_cpu = to_dtpm_cpu(dtpm); struct em_perf_domain *pd; - struct cpumask cpus; + struct cpumask *pd_mask; unsigned long freq; - int i, nr_cpus; + int i; pd = em_cpu_get(dtpm_cpu->cpu); - freq = cpufreq_quick_get(dtpm_cpu->cpu); - cpumask_and(&cpus, cpu_online_mask, to_cpumask(pd->cpus)); - nr_cpus = cpumask_weight(&cpus); + pd_mask = em_span_cpus(pd); + + freq = cpufreq_quick_get(dtpm_cpu->cpu); for (i = 0; i < pd->nr_perf_states; i++) { if (pd->table[i].frequency < freq) continue; - return pd->table[i].power * - MICROWATT_PER_MILLIWATT * nr_cpus; + return scale_pd_power_uw(pd_mask, pd->table[i].power * + MICROWATT_PER_MILLIWATT); } return 0;