Received: by 2002:a25:1506:0:0:0:0:0 with SMTP id 6csp3658850ybv; Mon, 10 Feb 2020 04:11:03 -0800 (PST) X-Google-Smtp-Source: APXvYqwxe1IPHFqBm/K2LZ05Bi21ObbtRDCwST5qxepxfimLxEldupyHGKDAtLd4XR/Nb20Jhtd7 X-Received: by 2002:aca:220c:: with SMTP id b12mr590908oic.55.1581336663366; Mon, 10 Feb 2020 04:11:03 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1581336663; cv=none; d=google.com; s=arc-20160816; b=PPvJA3zyT1Xk7fZ4vS6Rb1Q4x9fuTVfJrCKJByQ6DPGpaS4ArCLKuQ4GOc/45BoAgR o6IxBKmX7FX2GIfh3E7z66MmmVb2Y/wywHmLMEe8MnubjlP9NHaoRuD+CqK6ydJ/UrO1 DB8Jn9baptbv93ugqr5T1+9u9RnaJFu5PC6xQ2XBMEqF8myyjdY+BFuXGTIVKmkM8wxO H3GPhS+mrYggg6HCSARzLYg1AsLWm7kXQpmRbqwbawe30gGnrT7Nc88LE5Z9XLBX9dhQ QX3Vw9M/iiX2mDT9H69xof2y2Y/SCs4H4czlZXKar7kiepvVhH+W+3Kg5sPLSMDvBplF sDLQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=Jh1T7CTtPbuRS7dbZY8cQIGHEoh8axLIX2AWEU99/Ig=; b=petTatjjvwhvjpXWCxKZR0K3T4anN4SINyF/MqwSOVSdhe7raN3GKOfM2SV+Uoz8t2 BHSImXW6nyhq2VQgW/ZiOMLEkauQbBgGBdehfJsmoFus5/fCajg0tXv0QdqZ5sYwhFOG iBYFgWCGnWk2qO7N8j+v3vt/OfdoLaHDv/Pdmpm69emYN30OBNukejT4I2NUggJUiaxz TWNT3+luK5TDMnvhJap3I4Afce2B3weg2S7kOK5MHQqqMUx4Dl9ahCRo3/GJrGYsFmTT nBHk9dIM8Pzj6MKU5a/8Lgvw+tMV5mwqJyyN6yDgS0fJDA6DgnWySp3/p0aty975Ohbp fXTg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j2si97672otr.255.2020.02.10.04.10.51; Mon, 10 Feb 2020 04:11:03 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727969AbgBJMJH (ORCPT + 99 others); Mon, 10 Feb 2020 07:09:07 -0500 Received: from foss.arm.com ([217.140.110.172]:59386 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727570AbgBJMHY (ORCPT ); Mon, 10 Feb 2020 07:07:24 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id DD9A81FB; Mon, 10 Feb 2020 04:07:23 -0800 (PST) Received: from [192.168.0.7] (unknown [172.31.20.19]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id C0B883F6CF; Mon, 10 Feb 2020 04:07:20 -0800 (PST) Subject: Re: [Patch v9 0/8] Introduce Thermal Pressure To: Thara Gopinath , mingo@redhat.com, peterz@infradead.org, ionela.voinescu@arm.com, vincent.guittot@linaro.org, rui.zhang@intel.com, qperret@google.com, daniel.lezcano@linaro.org, viresh.kumar@linaro.org, rostedt@goodmis.org, will@kernel.org, catalin.marinas@arm.com, sudeep.holla@arm.com, juri.lelli@redhat.com, corbet@lwn.net Cc: linux-kernel@vger.kernel.org, amit.kachhap@gmail.com, javi.merino@kernel.org, amit.kucheria@verdurent.com References: <1580250967-4386-1-git-send-email-thara.gopinath@linaro.org> From: Dietmar Eggemann Message-ID: <95164e1d-12e4-b155-f0d6-f869ee982aae@arm.com> Date: Mon, 10 Feb 2020 13:07:19 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.4.1 MIME-Version: 1.0 In-Reply-To: <1580250967-4386-1-git-send-email-thara.gopinath@linaro.org> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 28/01/2020 23:35, Thara Gopinath wrote: > Thermal governors can respond to an overheat event of a cpu by > capping the cpu's maximum possible frequency. This in turn > means that the maximum available compute capacity of the > cpu is restricted. But today in the kernel, task scheduler is > not notified of capping of maximum frequency of a cpu. > In other words, scheduler is unaware of maximum capacity > restrictions placed on a cpu due to thermal activity. > This patch series attempts to address this issue. > The benefits identified are better task placement among available > cpus in event of overheating which in turn leads to better > performance numbers. > > The reduction in the maximum possible capacity of a cpu due to a > thermal event can be considered as thermal pressure. Instantaneous > thermal pressure is hard to record and can sometime be erroneous > as there can be mismatch between the actual capping of capacity > and scheduler recording it. Thus solution is to have a weighted > average per cpu value for thermal pressure over time. > The weight reflects the amount of time the cpu has spent at a > capped maximum frequency. Since thermal pressure is recorded as > an average, it must be decayed periodically. Exisiting algorithm > in the kernel scheduler pelt framework is re-used to calculate > the weighted average. This patch series also defines a sysctl > inerface to allow for a configurable decay period. > > Regarding testing, basic build, boot and sanity testing have been > performed on db845c platform with debian file system. > Further, dhrystone and hackbench tests have been > run with the thermal pressure algorithm. During testing, due to > constraints of step wise governor in dealing with big little systems, > trip point 0 temperature was made assymetric between cpus in little > cluster and big cluster; the idea being that > big core will heat up and cpu cooling device will throttle the > frequency of the big cores faster, there by limiting the maximum available > capacity and the scheduler will spread out tasks to little cores as well. > > Test Results > > Hackbench: 1 group , 30000 loops, 10 runs > Result SD > (Secs) (% of mean) > No Thermal Pressure 14.03 2.69% > Thermal Pressure PELT Algo. Decay : 32 ms 13.29 0.56% > Thermal Pressure PELT Algo. Decay : 64 ms 12.57 1.56% > Thermal Pressure PELT Algo. Decay : 128 ms 12.71 1.04% > Thermal Pressure PELT Algo. Decay : 256 ms 12.29 1.42% > Thermal Pressure PELT Algo. Decay : 512 ms 12.42 1.15% > > Dhrystone Run Time : 20 threads, 3000 MLOOPS > Result SD > (Secs) (% of mean) > No Thermal Pressure 9.452 4.49% > Thermal Pressure PELT Algo. Decay : 32 ms 8.793 5.30% > Thermal Pressure PELT Algo. Decay : 64 ms 8.981 5.29% > Thermal Pressure PELT Algo. Decay : 128 ms 8.647 6.62% > Thermal Pressure PELT Algo. Decay : 256 ms 8.774 6.45% > Thermal Pressure PELT Algo. Decay : 512 ms 8.603 5.41% What do we do on systems on which one Frequency domain spawns all the CPUs (e.g. Hikey620)? perf stat --null --repeat 10 -- perf bench sched messaging -g 10 -l 1000 # Running 'sched/messaging' benchmark: # 20 sender and receiver processes per group # 10 groups == 400 processes run Total time: 4.697 [sec] # Running 'sched/messaging' benchmark: [ 8082.882751] hisi_thermal f7030700.tsensor: sensor <2> THERMAL ALARM: 66385 > 65000 # 20 sender and receiver processes per group # 10 groups == 400 processes run Total time: 4.910 [sec] # Running 'sched/messaging' benchmark: [ 8091.070386] CPU3 cpus=0-7 th_pressure=205 [ 8091.178390] CPU3 cpus=0-7 th_pressure=0 [ 8091.286389] CPU3 cpus=0-7 th_pressure=205 [ 8091.398397] CPU3 cpus=0-7 th_pressure=0