Received: by 2002:a25:d7c1:0:0:0:0:0 with SMTP id o184csp1733833ybg; Sat, 19 Oct 2019 01:20:07 -0700 (PDT) X-Google-Smtp-Source: APXvYqz6mfD+ZmBj/WMioORdLgX0kfCq2LIh2CEybbhicCoDVrvqOyjMVAsne3j/6tGHQw9yawLp X-Received: by 2002:a17:906:3615:: with SMTP id q21mr12249330ejb.152.1571473207543; Sat, 19 Oct 2019 01:20:07 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1571473207; cv=none; d=google.com; s=arc-20160816; b=n3dLc2elZtqd6PZ4IoBbXWjCqnHVbU0JhLcsnxvgpYAWqNzmNuYJmfPA4gDmwXNzfc pECI55TvEKEoH5xpwp9CyqT44TbktQQQaql4KOWUlrG9CcsiMcAS4FtAgcGVhArakG2g 023e5FcbnYLbNuJWMn8U7ZMAJjjXFDge8goilxtsVbjhZZEuqYS9U+sXpK83XXx1yLbV tHv0bXIUOB7VyptM9d2o5TqW4DjmiDE8cb6tbNmrnSMkTmiY4embHYJwzvbsUiogNQhC 7EVD7elP59gkLWAl+Z4/8hK2FmXygF9MxddB59XKi3Dctvp8QgePnTtNN/jc3VUX8xOJ /IhA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=//0iKN6X/Y+78TaiyI0/7LBWHob2KiqGQms0QVTjCQo=; b=nVAwcWIwPgT+P794m04N9gtX/ozBuHJM0OoeW/3MGUs1UlJ9DCaKXRh9VQSPv251u+ Cw27ogR8JgO1ZUxBj4mjpA7q8TAM0lqUZJWS/pyXwz9dmu2U+Yse0qIAxMZIKqypD24K j+lrVezCnovBmPt67tzWlnWGoc28QV19/fBz4oNni4b2OY5C0BkOhPO7vMJX7WRsVYWj gmbMtdolVIzqgtqXYlR69oCB+nTqApS/L2dSaBMtumqB1FyzbTYTqn1EDPp6H/XZPx/T p6QBvr4cLT9kqe2yMlprkpAY/ElMQ8U90bjlWK4C5SxhiU/HF8cTdfOIg8by35FeqxrM AnYw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id t18si4995416eju.329.2019.10.19.01.19.44; Sat, 19 Oct 2019 01:20:07 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2442746AbfJRMXO (ORCPT + 99 others); Fri, 18 Oct 2019 08:23:14 -0400 Received: from szxga05-in.huawei.com ([45.249.212.191]:4733 "EHLO huawei.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S2442702AbfJRMXO (ORCPT ); Fri, 18 Oct 2019 08:23:14 -0400 Received: from DGGEMS402-HUB.china.huawei.com (unknown [172.30.72.59]) by Forcepoint Email with ESMTP id 1802D6A250DE9ACE03E4; Fri, 18 Oct 2019 20:23:10 +0800 (CST) Received: from [127.0.0.1] (10.177.251.225) by DGGEMS402-HUB.china.huawei.com (10.3.19.202) with Microsoft SMTP Server id 14.3.439.0; Fri, 18 Oct 2019 20:23:03 +0800 Subject: Re: [PATCH V3] arm64: psci: Reduce waiting time for cpu_psci_cpu_kill() To: Mark Rutland CC: , , , , , , , , , "hushiyuan@huawei.com" , "linfeilong@huawei.com" , , , References: <433980c7-f246-f741-f00c-fce103a60af7@huawei.com> <20191018114143.GE27759@lakrids.cambridge.arm.com> From: Yunfeng Ye Message-ID: <4ce8e384-598c-cf93-96a4-5a4fca82aff8@huawei.com> Date: Fri, 18 Oct 2019 20:22:50 +0800 User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: <20191018114143.GE27759@lakrids.cambridge.arm.com> Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 7bit X-Originating-IP: [10.177.251.225] X-CFilter-Loop: Reflected Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2019/10/18 19:41, Mark Rutland wrote: > On Fri, Oct 18, 2019 at 07:24:14PM +0800, Yunfeng Ye wrote: >> In a case like suspend-to-disk, a large number of CPU cores need to be >> shut down. At present, the CPU hotplug operation is serialised, and the >> CPU cores can only be shut down one by one. In this process, if PSCI >> affinity_info() does not return LEVEL_OFF quickly, cpu_psci_cpu_kill() >> needs to wait for 10ms. If hundreds of CPU cores need to be shut down, >> it will take a long time. > > Do we have an idea of roughly how long a CPU _usually_ takes to > transition state? > > i.e. are we _just_ missing the transition the first time we call > AFFINITY_INFO? > we have test that in most case is less than 1ms, 50us-500us. the time not only include hardware state transition, but also include flush caches in BIOS. and flush caches operation is time-consuming. >> Normally, it is no need to wait 10ms in cpu_psci_cpu_kill(). So change >> the wait interval from 10 ms to max 1 ms and use usleep_range() instead >> of msleep() for more accurate schedule. >> >> In addition, reduce the time interval will increase the messages output, >> so remove the "Retry ..." message, instead, put the number of waiting >> times to the sucessful message. >> >> Signed-off-by: Yunfeng Ye >> --- >> v2 -> v3: >> - update the comment >> - remove the busy-wait logic, modify the loop logic and output message >> >> v1 -> v2: >> - use usleep_range() instead of udelay() after waiting for a while >> >> arch/arm64/kernel/psci.c | 7 +++---- >> 1 file changed, 3 insertions(+), 4 deletions(-) >> >> diff --git a/arch/arm64/kernel/psci.c b/arch/arm64/kernel/psci.c >> index c9f72b2665f1..00b8c0825a08 100644 >> --- a/arch/arm64/kernel/psci.c >> +++ b/arch/arm64/kernel/psci.c >> @@ -91,15 +91,14 @@ static int cpu_psci_cpu_kill(unsigned int cpu) >> * while it is dying. So, try again a few times. >> */ >> >> - for (i = 0; i < 10; i++) { >> + for (i = 0; i < 100; i++) { >> err = psci_ops.affinity_info(cpu_logical_map(cpu), 0); >> if (err == PSCI_0_2_AFFINITY_LEVEL_OFF) { >> - pr_info("CPU%d killed.\n", cpu); >> + pr_info("CPU%d killed by waiting %d loops.\n", cpu, i); > > Could we please make that: > > pr_info("CPU%d killed (polled %d times)\n", cpu, i + 1); > ok, thanks. > > >> return 0; >> } >> >> - msleep(10); >> - pr_info("Retrying again to check for CPU kill\n"); >> + usleep_range(100, 1000); > > Hmm, so now we'll wait somewhere between 10ms and 100ms before giving up > on a CPU depending on how long we actually sleep for each iteration of > the loop. That should be called out in the commit message. > > That could matter for kdump when you have a large number of CPUs, as in > the worst case for 256 CPUs we've gone from ~2.6s to ~26s. But tbh in > that case I'm not sure I care that much... > > In the majority of cases I'd hope AFFINITY_INFO would return OFF after > an iteration or two. > Normally it will no need so much time. > Thanks, > Mark. > > . >