Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp3620830pxj; Mon, 7 Jun 2021 15:55:35 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzS5+1f8PRBCHjbNEhfV3PKe6fMWhYf4QVGE0Y9jvAXk7VPTpx2WmwrOgaFqnxZku6FT2Yu X-Received: by 2002:a05:6402:342:: with SMTP id r2mr22679353edw.69.1623106535058; Mon, 07 Jun 2021 15:55:35 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1623106535; cv=none; d=google.com; s=arc-20160816; b=pFjG8sSV6ppkg93U0QxtW212nxnfH/1+YKz94fSnl8s0NgnZ4dkTXHtKv5C7N+9Xda JDeishL4yrM9HNs21Lp4v6JLQvPnCPO095xoHZblcNLXZOxSm6XV1mrbE1FCZgvrvPrW P36uTlLfOxV58w8TsnxNv8tfs2TwcRE+41RL8//haNp1NY+tsxXCZc9n99EqCnR56e4j m4AUmHF3VOiTFSw8BGn+LWKnzQb6EaEtfuHu0ma6E7LI6I6mWL9IpiizkkiLu5ilmx9s NGAJP1dca3CdQR6soH5fxBCHWbfV9AkfRmaaQksfMYDZS78R7DL+gu2db2PjIWk+2Eqz MQvQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature; bh=FfdTWYNkKWrBSn9e1mH+6zdZXbgVR7+FaDZWEdEvJvE=; b=Kby4szCTijBeLuvN+cif2+ZEhsMilTIlyWo65pguRXmCmlg+iMeVBRW73pjr00dTro 1aZeD9HXMN4Rzz2PXtGy0PcTtGLpqBVWBv7XJ5ijBBx21h5hnW0VneOesUSYVU1wlmpz RBXaycmbpb77hvBlc1oDbbCVr7FiQR2gWIAnos0GutWvVFwmKXM+G1bcEylPVdsVs9Y5 n2zjqwgipgkHRXRE8+BWe9ZOv2vWAhUgfZZleWzZfWdtlaGUp8nUpSC5a4yMJj7yaxvU 5toh7UJp7NE5R/lw6m7Q5fwTNYt2DXMlphfPNBPl0FdfWGB+w5c8CurQWiUcM3NcwS0M XPOA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=KgBgiXXf; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id m12si13452857ejq.256.2021.06.07.15.55.11; Mon, 07 Jun 2021 15:55:35 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=KgBgiXXf; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230355AbhFGWyB (ORCPT + 99 others); Mon, 7 Jun 2021 18:54:01 -0400 Received: from mail.kernel.org ([198.145.29.99]:60770 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230183AbhFGWyB (ORCPT ); Mon, 7 Jun 2021 18:54:01 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id 82632610C7; Mon, 7 Jun 2021 22:52:06 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1623106329; bh=WaNp1TylVxIDpajJun/RGMsjtK/LLdNZo+wOLZ/qvqo=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=KgBgiXXfTHZj3F/lf9QX3uJkrtwE7qLK9joY82TANSb5yNP44Jr9ZDtj+LAd27vxV W3/ufy+GsVHIlxF63/OaUj4InNRGBTYGPr5sx0BUR2B8lUuQ8d1duPewxPSvN2P7eX yRDOzBB1x6P8+ILCPqIyQVTcjqe/SRuhxD2rp3imOzzBUmFwgRanw3YywTo9tQOKuW ZfPvIKebXbqh0e/zcGW3btINSdp27bSSA+RCT/qbOehNoUawXoPd0y2nNPTkPpkuOi +Pln20vS4TrEUSANKWX4ahsx3wW/yAVzk1XmmlKAZHH/EMPNfmUSLjqhayJFMIXstD 4VKUhMxucrnEw== Date: Mon, 7 Jun 2021 23:52:03 +0100 From: Will Deacon To: Valentin Schneider Cc: linux-arm-kernel@lists.infradead.org, linux-arch@vger.kernel.org, linux-kernel@vger.kernel.org, Catalin Marinas , Marc Zyngier , Greg Kroah-Hartman , Peter Zijlstra , Morten Rasmussen , Qais Yousef , Suren Baghdasaryan , Quentin Perret , Tejun Heo , Johannes Weiner , Ingo Molnar , Juri Lelli , Vincent Guittot , "Rafael J. Wysocki" , Dietmar Eggemann , Daniel Bristot de Oliveira , kernel-team@android.com Subject: Re: [PATCH v8 11/19] sched: Allow task CPU affinity to be restricted on asymmetric systems Message-ID: <20210607225202.GB8215@willie-the-truck> References: <20210602164719.31777-1-will@kernel.org> <20210602164719.31777-12-will@kernel.org> <87zgw5d05b.mognet@arm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <87zgw5d05b.mognet@arm.com> User-Agent: Mutt/1.10.1 (2018-07-13) Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Jun 04, 2021 at 06:12:32PM +0100, Valentin Schneider wrote: > On 02/06/21 17:47, Will Deacon wrote: > > +static int restrict_cpus_allowed_ptr(struct task_struct *p, > > + struct cpumask *new_mask, > > + const struct cpumask *subset_mask) > > +{ > > + struct rq_flags rf; > > + struct rq *rq; > > + int err; > > + struct cpumask *user_mask = NULL; > > + > > + if (!p->user_cpus_ptr) { > > + user_mask = kmalloc(cpumask_size(), GFP_KERNEL); > > + > > + if (!user_mask) > > + return -ENOMEM; > > + } > > + > > + rq = task_rq_lock(p, &rf); > > + > > + /* > > + * Forcefully restricting the affinity of a deadline task is > > + * likely to cause problems, so fail and noisily override the > > + * mask entirely. > > + */ > > + if (task_has_dl_policy(p) && dl_bandwidth_enabled()) { > > + err = -EPERM; > > + goto err_unlock; > > + } > > + > > + if (!cpumask_and(new_mask, &p->cpus_mask, subset_mask)) { > > + err = -EINVAL; > > + goto err_unlock; > > + } > > + > > + /* > > + * We're about to butcher the task affinity, so keep track of what > > + * the user asked for in case we're able to restore it later on. > > + */ > > + if (user_mask) { > > + cpumask_copy(user_mask, p->cpus_ptr); > > + p->user_cpus_ptr = user_mask; > > + } > > + > > Shouldn't that be done before any of the bailouts above, so we can > potentially restore the mask even if we end up forcefully expanding the > affinity? I don't think so. I deliberately only track the old mask if we've managed to take a subset for the 32-bit task. If we end up having to override the mask entirely, then I treat it the same way as an explicit affinity change (only with a warning printed) and don't then try to restore the old mask -- it feels like we'd be overriding the affinity twice if we tried to do that. > > + return __set_cpus_allowed_ptr_locked(p, new_mask, 0, rq, &rf); > > + > > +err_unlock: > > + task_rq_unlock(rq, p, &rf); > > + kfree(user_mask); > > + return err; > > +} > > + > > +/* > > + * Restrict the CPU affinity of task @p so that it is a subset of > > + * task_cpu_possible_mask() and point @p->user_cpu_ptr to a copy of the > > + * old affinity mask. If the resulting mask is empty, we warn and walk > > + * up the cpuset hierarchy until we find a suitable mask. > > + */ > > +void force_compatible_cpus_allowed_ptr(struct task_struct *p) > > +{ > > + cpumask_var_t new_mask; > > + const struct cpumask *override_mask = task_cpu_possible_mask(p); > > + > > + alloc_cpumask_var(&new_mask, GFP_KERNEL); > > + > > + /* > > + * __migrate_task() can fail silently in the face of concurrent > > + * offlining of the chosen destination CPU, so take the hotplug > > + * lock to ensure that the migration succeeds. > > + */ > > + cpus_read_lock(); > > I'm thinking this might not be required with: > > http://lore.kernel.org/r/20210526205751.842360-3-valentin.schneider@arm.com > > but then again this isn't merged yet :-) Agreed, if that patch does what it says on the tin ;) I need to digest your reply to me, as this is mind-bending stuff. > > +static int > > +__sched_setaffinity(struct task_struct *p, const struct cpumask *mask); > > + > > +/* > > + * Restore the affinity of a task @p which was previously restricted by a > > + * call to force_compatible_cpus_allowed_ptr(). This will clear (and free) > > + * @p->user_cpus_ptr. > > + */ > > +void relax_compatible_cpus_allowed_ptr(struct task_struct *p) > > +{ > > + unsigned long flags; > > + struct cpumask *mask = p->user_cpus_ptr; > > + > > + /* > > + * Try to restore the old affinity mask. If this fails, then > > + * we free the mask explicitly to avoid it being inherited across > > + * a subsequent fork(). > > + */ > > + if (!mask || !__sched_setaffinity(p, mask)) > > + return; > > + > > + raw_spin_lock_irqsave(&p->pi_lock, flags); > > + release_user_cpus_ptr(p); > > + raw_spin_unlock_irqrestore(&p->pi_lock, flags); > > AFAICT an affinity change can happen between __sched_setaffinity() and > reacquiring the ->pi_lock. Right now this can't be another > force_compatible_cpus_allowed_ptr() because this is only driven by > arch_setup_new_exec() against current, so we should be fine, but here be > dragons. That's a good point. I'll add a comment for now, since I'm not sure who else might end up using this in future. Generally it's pretty agnostic to how it's being used, but we're certainly relying on the serialisation of restrict/relax calls. Will