From patchwork Fri Aug 26 20:17:53 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Lina Iyer X-Patchwork-Id: 9301969 X-Patchwork-Delegate: agross@codeaurora.org Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork.web.codeaurora.org (Postfix) with ESMTP id 2D18660757 for ; Fri, 26 Aug 2016 20:18:53 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 1E83829653 for ; Fri, 26 Aug 2016 20:18:53 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 122A629654; Fri, 26 Aug 2016 20:18:53 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-6.8 required=2.0 tests=BAYES_00,DKIM_SIGNED, RCVD_IN_DNSWL_HI,T_DKIM_INVALID autolearn=ham version=3.3.1 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 5E94E2963A for ; Fri, 26 Aug 2016 20:18:52 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754184AbcHZUSo (ORCPT ); Fri, 26 Aug 2016 16:18:44 -0400 Received: from mail-pa0-f53.google.com ([209.85.220.53]:36812 "EHLO mail-pa0-f53.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754134AbcHZUSm (ORCPT ); Fri, 26 Aug 2016 16:18:42 -0400 Received: by mail-pa0-f53.google.com with SMTP id di2so30184971pad.3 for ; Fri, 26 Aug 2016 13:18:42 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=gCuKDtDmHYDLbWXeioHIifudz6JbjkpkXDksVfcJIGk=; b=bpJ58GG/eS6v7TpTzxBJI/lBJ4NjY/X4lrf4t+p9sqXaagYy12ZYO25KgejRJm5g2T RyctHNJSrq1zEqNaRJd9zJFzrdyMIo8egN67D0lhdlDd/aszdBycQD/nzJ8Hz7xvFKwp jE0NG+5c2TZYuR1xllRL8V67wIzVNdNuf43Bg= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=gCuKDtDmHYDLbWXeioHIifudz6JbjkpkXDksVfcJIGk=; b=TXw9kd4oMKfj+2YSnVjuhaL01XoL2CrZv2+4W152mc1QN1IGWMf8wT/aBltYJX5QtN dTuUVBC3NaId6fpBPeK2nA+RMg8nygdIt4LAllJtx3eEjqxZJPBCKNhlXEKQPV24/OYw VGqrPXscgrJ5AQYDAJ6zrQuMqywK4ciop56/GqYwkuiZaOBM7eWj8sK8GVw4mn7JC4vy 4nKaD73OPqTo7ZZxUbxwXoIIi0dgqFEalo2QZklPXJqjn9Inw/d4O9pZv4dGI7fhXlob kFqJQWUUW1VI61y1qNFNYuMUNL+qopbnDbpEMYp3nS9dlwHIfM8zdPpm39ThbK1N7jcz M73w== X-Gm-Message-State: AE9vXwOpdoNiKL5i6PFtT6hSNlVF5bdojkyh6PMFSHg3AQQSJ7mZXJhe1LRxZuJvUOUWT2qK X-Received: by 10.66.86.1 with SMTP id l1mr9211153paz.106.1472242721631; Fri, 26 Aug 2016 13:18:41 -0700 (PDT) Received: from ubuntu.localdomain (i-global254.qualcomm.com. [199.106.103.254]) by smtp.gmail.com with ESMTPSA id m128sm30761463pfm.42.2016.08.26.13.18.39 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Fri, 26 Aug 2016 13:18:40 -0700 (PDT) From: Lina Iyer To: ulf.hansson@linaro.org, khilman@kernel.org, rjw@rjwysocki.net, linux-pm@vger.kernel.org, linux-arm-kernel@lists.infradead.org Cc: andy.gross@linaro.org, sboyd@codeaurora.org, linux-arm-msm@vger.kernel.org, brendan.jackman@arm.com, lorenzo.pieralisi@arm.com, sudeep.holla@arm.com, Juri.Lelli@arm.com, Lina Iyer Subject: [PATCH v5 11/16] PM / cpu_domains: Add PM Domain governor for CPUs Date: Fri, 26 Aug 2016 14:17:53 -0600 Message-Id: <1472242678-33700-12-git-send-email-lina.iyer@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1472242678-33700-1-git-send-email-lina.iyer@linaro.org> References: <1472242678-33700-1-git-send-email-lina.iyer@linaro.org> Sender: linux-arm-msm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-arm-msm@vger.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP A PM domain comprising of CPUs may be powered off when all the CPUs in the domain are powered down. Powering down a CPU domain is generally a expensive operation and therefore the power performance trade offs should be considered. The time between the last CPU powering down and the first CPU powering up in a domain, is the time available for the domain to sleep. Ideally, the sleep time of the domain should fulfill the residency requirement of the domains' idle state. To do this effectively, read the time before the wakeup of the cluster's CPUs and ensure that the domain's idle state sleep time guarantees the QoS requirements of each of the CPU, the PM QoS CPU_DMA_LATENCY and the state's residency. Signed-off-by: Lina Iyer --- drivers/base/power/cpu_domains.c | 80 +++++++++++++++++++++++++++++++++++++++- 1 file changed, 79 insertions(+), 1 deletion(-) diff --git a/drivers/base/power/cpu_domains.c b/drivers/base/power/cpu_domains.c index 8bf61e2..79fa4ae 100644 --- a/drivers/base/power/cpu_domains.c +++ b/drivers/base/power/cpu_domains.c @@ -17,9 +17,12 @@ #include #include #include +#include +#include #include #include #include +#include #define CPU_PD_NAME_MAX 36 @@ -51,6 +54,81 @@ static inline struct cpu_pm_domain *to_cpu_pd(struct generic_pm_domain *d) return res; } +static bool cpu_pd_down_ok(struct dev_pm_domain *pd) +{ + struct generic_pm_domain *genpd = pd_to_genpd(pd); + struct cpu_pm_domain *cpu_pd = to_cpu_pd(genpd); + int qos_ns = pm_qos_request(PM_QOS_CPU_DMA_LATENCY); + u64 sleep_ns; + ktime_t earliest, next_wakeup; + int cpu; + int i; + + /* Reset the last set genpd state, default to index 0 */ + genpd->state_idx = 0; + + /* We don't want to power down, if QoS is 0 */ + if (!qos_ns) + return false; + + /* + * Find the sleep time for the cluster. + * The time between now and the first wake up of any CPU that + * are in this domain hierarchy is the time available for the + * domain to be idle. + * + * We only care about the next wakeup for any online CPU in that + * cluster. Hotplug off any of the CPUs that we care about will + * wait on the genpd lock, until we are done. Any other CPU hotplug + * is not of consequence to our sleep time. + */ + earliest = ktime_set(KTIME_SEC_MAX, 0); + for_each_cpu_and(cpu, cpu_pd->cpus, cpu_online_mask) { + next_wakeup = tick_nohz_get_next_wakeup(cpu); + if (earliest.tv64 > next_wakeup.tv64) + earliest = next_wakeup; + } + + sleep_ns = ktime_to_ns(ktime_sub(earliest, ktime_get())); + if (sleep_ns <= 0) + return false; + + /* + * Find the deepest sleep state that satisfies the residency + * requirement and the QoS constraint + */ + for (i = genpd->state_count - 1; i >= 0; i--) { + u64 state_sleep_ns; + + state_sleep_ns = genpd->states[i].power_off_latency_ns + + genpd->states[i].power_on_latency_ns + + genpd->states[i].residency_ns; + + /* + * If we can't sleep to save power in the state, move on + * to the next lower idle state. + */ + if (state_sleep_ns > sleep_ns) + continue; + + /* + * We also don't want to sleep more than we should to + * gaurantee QoS. + */ + if (state_sleep_ns < (qos_ns * NSEC_PER_USEC)) + break; + } + + if (i >= 0) + genpd->state_idx = i; + + return (i >= 0); +} + +static struct dev_power_governor cpu_pd_gov = { + .power_down_ok = cpu_pd_down_ok, +}; + static int cpu_pd_power_on(struct generic_pm_domain *genpd) { struct cpu_pm_domain *pd = to_cpu_pd(genpd); @@ -172,7 +250,7 @@ struct generic_pm_domain *cpu_pd_init(struct generic_pm_domain *genpd, list_add_rcu(&pd->link, &of_cpu_pd_list); mutex_unlock(&cpu_pd_list_lock); - ret = pm_genpd_init(genpd, &simple_qos_governor, false); + ret = pm_genpd_init(genpd, &cpu_pd_gov, false); if (ret) { pr_err("Unable to initialize domain %s\n", genpd->name); goto fail;