Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp585777pxb; Wed, 27 Jan 2021 15:55:28 -0800 (PST) X-Google-Smtp-Source: ABdhPJwSHmf7zVY4w58ZundBV4KinF4FbdcEeNSPOoHD+glxi4cgD/xcg5vvW9zasbOB5sM9yM4K X-Received: by 2002:a17:906:4050:: with SMTP id y16mr3185124ejj.43.1611791728743; Wed, 27 Jan 2021 15:55:28 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1611791728; cv=none; d=google.com; s=arc-20160816; b=przB7eyr57+CZb8ZvS/JFu2hb0Be0VW8AXX0p1VDUgah/V9XAeTAXMVuTGoJ+eSt2M EpTZOONKkT6BmkbqhylCqH+tk33CVLjEQHxpxAq6eQJqdUWGWkAKuHRTLld3xl9FafYh zCWX+dzDrANoSNLYncLlpkbL6bDDHZ3NCkKHLgx/ejPRoorBZWhSUkED2jiZIYa3Q3Vr 4cDPFtqNcBJmfM1bgwXgDVvC+bYZHB0oM1hDXm9g2xH4eGIbKmwAmnC8t7eTS9xwgYJJ RyxVBeLAereTSFzl9z+gpi3a2puo/xMFL80pvQoRfFabHGHGS0Ysm8i2BSBrQJUo/9Sh b7Xw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :to:subject; bh=0B+lmRgqcAqf+3eE+57Bhg2VOoWWqQ2DxATfMl/XFss=; b=hOceAcNDrMqP8VDMnF3hzRSi0MvTMaAalbJivL9xViLxw5PIXEjfscv1WflbEtbFys JSK/GykprzKXAhhG3P9zsCkKn4AnXdxFiXHEw8wjGhK+97ofugabOfYuSUFOuXD+Z0Tx jWFYcBUZSTGXDDFlO9GArRM0SRZkNpLvQPFK5z/TalDSqXVdl9RiMT8ABqFYdiAJlISr 9l+K8nSTxJAN4noXmD5r/IhPBmyewmsw9Pm0DqnrRlW1pMTMmdrta1/oqSpJRTl2f2W9 QddAH+roQmSRK2nCHQF8T44JX8l73Daz0d4AV8BcgUmUFgNS5Eln1r08eaxUeFDQXkWL YaEQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id dg28si1247102edb.53.2021.01.27.15.55.04; Wed, 27 Jan 2021 15:55:28 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S237294AbhA0MAo (ORCPT + 99 others); Wed, 27 Jan 2021 07:00:44 -0500 Received: from foss.arm.com ([217.140.110.172]:42136 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S237453AbhA0L6G (ORCPT ); Wed, 27 Jan 2021 06:58:06 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id AF2361FB; Wed, 27 Jan 2021 03:57:20 -0800 (PST) Received: from [10.57.47.135] (unknown [10.57.47.135]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 8BDAE3F68F; Wed, 27 Jan 2021 03:57:17 -0800 (PST) Subject: Re: [Patch v4 1/3] lib: Restrict cpumask_local_spread to houskeeping CPUs To: Nitesh Narayan Lal , linux-kernel@vger.kernel.org, linux-api@vger.kernel.org, frederic@kernel.org, mtosatti@redhat.com, juri.lelli@redhat.com, abelits@marvell.com, bhelgaas@google.com, linux-pci@vger.kernel.org, rostedt@goodmis.org, mingo@kernel.org, peterz@infradead.org, tglx@linutronix.de, davem@davemloft.net, akpm@linux-foundation.org, sfr@canb.auug.org.au, stephen@networkplumber.org, rppt@linux.vnet.ibm.com, jinyuqi@huawei.com, zhangshaokun@hisilicon.com References: <20200625223443.2684-1-nitesh@redhat.com> <20200625223443.2684-2-nitesh@redhat.com> From: Robin Murphy Message-ID: <3e9ce666-c9cd-391b-52b6-3471fe2be2e6@arm.com> Date: Wed, 27 Jan 2021 11:57:16 +0000 User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:78.0) Gecko/20100101 Thunderbird/78.6.1 MIME-Version: 1.0 In-Reply-To: <20200625223443.2684-2-nitesh@redhat.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-GB Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, On 2020-06-25 23:34, Nitesh Narayan Lal wrote: > From: Alex Belits > > The current implementation of cpumask_local_spread() does not respect the > isolated CPUs, i.e., even if a CPU has been isolated for Real-Time task, > it will return it to the caller for pinning of its IRQ threads. Having > these unwanted IRQ threads on an isolated CPU adds up to a latency > overhead. > > Restrict the CPUs that are returned for spreading IRQs only to the > available housekeeping CPUs. > > Signed-off-by: Alex Belits > Signed-off-by: Nitesh Narayan Lal > --- > lib/cpumask.c | 16 +++++++++++----- > 1 file changed, 11 insertions(+), 5 deletions(-) > > diff --git a/lib/cpumask.c b/lib/cpumask.c > index fb22fb266f93..85da6ab4fbb5 100644 > --- a/lib/cpumask.c > +++ b/lib/cpumask.c > @@ -6,6 +6,7 @@ > #include > #include > #include > +#include > > /** > * cpumask_next - get the next cpu in a cpumask > @@ -205,22 +206,27 @@ void __init free_bootmem_cpumask_var(cpumask_var_t mask) > */ > unsigned int cpumask_local_spread(unsigned int i, int node) > { > - int cpu; > + int cpu, hk_flags; > + const struct cpumask *mask; > > + hk_flags = HK_FLAG_DOMAIN | HK_FLAG_MANAGED_IRQ; > + mask = housekeeping_cpumask(hk_flags); AFAICS, this generally resolves to something based on cpu_possible_mask rather than cpu_online_mask as before, so could now potentially return an offline CPU. Was that an intentional change? I was just looking at the current code since I had the rare presence of mind to check if something suitable already existed before I start open-coding "any online CPU, but local node preferred" logic for handling IRQ affinity in a driver - cpumask_local_spread() appears to be almost what I want (if a bit more heavyweight), if only it would actually guarantee an online CPU as the kerneldoc claims :( Robin. > /* Wrap: we always want a cpu. */ > - i %= num_online_cpus(); > + i %= cpumask_weight(mask); > > if (node == NUMA_NO_NODE) { > - for_each_cpu(cpu, cpu_online_mask) > + for_each_cpu(cpu, mask) { > if (i-- == 0) > return cpu; > + } > } else { > /* NUMA first. */ > - for_each_cpu_and(cpu, cpumask_of_node(node), cpu_online_mask) > + for_each_cpu_and(cpu, cpumask_of_node(node), mask) { > if (i-- == 0) > return cpu; > + } > > - for_each_cpu(cpu, cpu_online_mask) { > + for_each_cpu(cpu, mask) { > /* Skip NUMA nodes, done above. */ > if (cpumask_test_cpu(cpu, cpumask_of_node(node))) > continue; >