From patchwork Tue Mar 1 21:15:26 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Lina Iyer X-Patchwork-Id: 8470361 Return-Path: X-Original-To: patchwork-linux-pm@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork1.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.136]) by patchwork1.web.kernel.org (Postfix) with ESMTP id A4FC09F8A8 for ; Tue, 1 Mar 2016 21:16:30 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id AB5A82034E for ; Tue, 1 Mar 2016 21:16:29 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 8D8E620351 for ; Tue, 1 Mar 2016 21:16:28 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753563AbcCAVQU (ORCPT ); Tue, 1 Mar 2016 16:16:20 -0500 Received: from mail-pf0-f178.google.com ([209.85.192.178]:35502 "EHLO mail-pf0-f178.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753676AbcCAVQT (ORCPT ); Tue, 1 Mar 2016 16:16:19 -0500 Received: by mail-pf0-f178.google.com with SMTP id w128so75060931pfb.2 for ; Tue, 01 Mar 2016 13:16:19 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=/YSX3Bj2IjidvnWJSVzA5NdO8B/M8CQUnUvN8q7IQZw=; b=eDp0udlnGWa7qzbz6SO5iS9DuT03qxTXyvzK3YISAfeVSDuKVi/nqtNAcJZEsDdtNi Y33zYsW4Jic5U9ZcuZ8xHz1oCD7uWig5/jG0Gag6DQW931m6HZ2mU39rypJEawC3EjrG lfzYUl8QmokkS09sbGcmqw4SAPIXZkFY8Fv70= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=/YSX3Bj2IjidvnWJSVzA5NdO8B/M8CQUnUvN8q7IQZw=; b=S3HyAGyiMXQtlOEgCXrZY6vTGYUvydK0T8QBabKUK4LahL3giCpdRMOUXHTzacoWFw w8VcNkxeBkE25aip2wRiVeoM70zjg4j8SRZBw2I2RQhzIaAbMnhm4KlQxWgMXdobf8GM 0/ta/V/7uQjg5QPbqxaFyRF0aeFjh1HlG1p8yBpvxCtY3LLjBspGfHir0BwcV+TpBAb/ +nlce8v8mdoA6sfLdKhVZwqn555p3PedGKAt1MbrvGV5wdN/sfYbPJca9wiK3Mf6dRL0 eJYj6sRxhamhBolYVw3lkzWXJ+ye04bx7Miwe7Bls+5c9eCm6wNzG8dUm9aIfjygpRGK 9QMA== X-Gm-Message-State: AD7BkJIpYXPVCjcJD36Uxx5pDZCRWVIjPxggnKCCvU52iTHui2XjTMq3Pw+36xVHl6NkLUhW X-Received: by 10.98.87.90 with SMTP id l87mr33478217pfb.70.1456866979076; Tue, 01 Mar 2016 13:16:19 -0800 (PST) Received: from ubuntu.localdomain (i-global254.qualcomm.com. [199.106.103.254]) by smtp.gmail.com with ESMTPSA id kw10sm47749328pab.0.2016.03.01.13.16.16 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Tue, 01 Mar 2016 13:16:17 -0800 (PST) From: Lina Iyer To: ulf.hansson@linaro.org, khilman@kernel.org, rjw@rjwysocki.net, linux-pm@vger.kernel.org, linux-arm-kernel@lists.infradead.org Cc: geert@linux-m68k.org, k.kozlowski@samsung.com, msivasub@codeaurora.org, agross@codeaurora.org, sboyd@codeaurora.org, linux-arm-msm@vger.kernel.org, lorenzo.pieralisi@arm.com, ahaslam@baylibre.com, mtitinger@baylibre.com, Lina Iyer Subject: [RFC v3 07/12] PM / cpu_domains: Add PM Domain governor for CPUs Date: Tue, 1 Mar 2016 14:15:26 -0700 Message-Id: <1456866931-37851-9-git-send-email-lina.iyer@linaro.org> X-Mailer: git-send-email 2.1.4 In-Reply-To: <1456866931-37851-1-git-send-email-lina.iyer@linaro.org> References: <1456866931-37851-1-git-send-email-lina.iyer@linaro.org> Sender: linux-pm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-pm@vger.kernel.org X-Spam-Status: No, score=-6.8 required=5.0 tests=BAYES_00,DKIM_SIGNED, RCVD_IN_DNSWL_HI,RP_MATCHES_RCVD,T_DKIM_INVALID,UNPARSEABLE_RELAY autolearn=unavailable version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP A PM domain comprising of CPUs may be powered off when all the CPUs in the domain are powered down. Powering down a CPU domain is generally a expensive operation and therefore the power performance trade offs should be considered. The time between the last CPU powering down and the first CPU powering up in a domain, is the time available for the domain to sleep. Ideally, the sleep time of the domain should fulfill the residency requirement of the domains' idle state. To do this effectively, read the time before the wakeup of the cluster's CPUs and ensure that the domain's idle state sleep time guarantees the QoS requirements of each of the CPU, the PM QoS CPU_DMA_LATENCY and the state's residency. Signed-off-by: Lina Iyer --- drivers/base/power/cpu_domains.c | 80 +++++++++++++++++++++++++++++++++++++++- 1 file changed, 79 insertions(+), 1 deletion(-) diff --git a/drivers/base/power/cpu_domains.c b/drivers/base/power/cpu_domains.c index 62c1ab9..6ba05f8 100644 --- a/drivers/base/power/cpu_domains.c +++ b/drivers/base/power/cpu_domains.c @@ -17,9 +17,12 @@ #include #include #include +#include +#include #include #include #include +#include #define CPU_PD_NAME_MAX 36 @@ -52,6 +55,81 @@ struct cpu_pm_domain *to_cpu_pd(struct generic_pm_domain *d) return res; } +static bool cpu_pd_down_ok(struct dev_pm_domain *pd) +{ + struct generic_pm_domain *genpd = pd_to_genpd(pd); + struct cpu_pm_domain *cpu_pd = to_cpu_pd(genpd); + int qos_ns = pm_qos_request(PM_QOS_CPU_DMA_LATENCY); + u64 sleep_ns; + ktime_t earliest, next_wakeup; + int cpu; + int i; + + /* Reset the last set genpd state, default to index 0 */ + genpd->state_idx = 0; + + /* We don't want to power down, if QoS is 0 */ + if (!qos_ns) + return false; + + /* + * Find the sleep time for the cluster. + * The time between now and the first wake up of any CPU that + * are in this domain hierarchy is the time available for the + * domain to be idle. + * + * We only care about the next wakeup for any online CPU in that + * cluster. Hotplug off any of the CPUs that we care about will + * wait on the genpd lock, until we are done. Any other CPU hotplug + * is not of consequence to our sleep time. + */ + earliest = ktime_set(KTIME_SEC_MAX, 0); + for_each_cpu_and(cpu, cpu_pd->cpus, cpu_online_mask) { + next_wakeup = tick_nohz_get_next_wakeup(cpu); + if (earliest.tv64 > next_wakeup.tv64) + earliest = next_wakeup; + } + + sleep_ns = ktime_to_ns(ktime_sub(earliest, ktime_get())); + if (sleep_ns <= 0) + return false; + + /* + * Find the deepest sleep state that satisfies the residency + * requirement and the QoS constraint + */ + for (i = genpd->state_count - 1; i >= 0; i--) { + u64 state_sleep_ns; + + state_sleep_ns = genpd->states[i].power_off_latency_ns + + genpd->states[i].power_on_latency_ns + + genpd->states[i].residency_ns; + + /* + * If we can't sleep to save power in the state, move on + * to the next lower idle state. + */ + if (state_sleep_ns > sleep_ns) + continue; + + /* + * We also don't want to sleep more than we should to + * gaurantee QoS. + */ + if (state_sleep_ns < (qos_ns * NSEC_PER_USEC)) + break; + } + + if (i >= 0) + genpd->state_idx = i; + + return (i >= 0); +} + +static struct dev_power_governor cpu_pd_gov = { + .power_down_ok = cpu_pd_down_ok, +}; + static int cpu_pd_attach_cpu(struct cpu_pm_domain *cpu_pd, int cpu) { int ret; @@ -143,7 +221,7 @@ static struct generic_pm_domain *of_init_cpu_pm_domain(struct device_node *dn, /* Register the CPU genpd */ pr_debug("adding %s as CPU PM domain\n", pd->genpd->name); - ret = of_pm_genpd_init(dn, pd->genpd, &simple_qos_governor, false); + ret = of_pm_genpd_init(dn, pd->genpd, &cpu_pd_gov, false); if (ret) { pr_err("Unable to initialize domain %s\n", dn->full_name); goto fail;