Received: by 2002:a05:6358:11c7:b0:104:8066:f915 with SMTP id i7csp726518rwl; Wed, 29 Mar 2023 07:43:24 -0700 (PDT) X-Google-Smtp-Source: AKy350YAbQvbEajzyntkgJzSURNVy+hapjemMfkYloH5yzNki22DfuA7KfnP1IwEEBsPVevOH0E5 X-Received: by 2002:aa7:cac4:0:b0:4fc:709f:7abd with SMTP id l4-20020aa7cac4000000b004fc709f7abdmr18485817edt.2.1680101004300; Wed, 29 Mar 2023 07:43:24 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1680101004; cv=none; d=google.com; s=arc-20160816; b=xyDNLSzq/+uxDxV5SKMX8Sd8mJha6RUYYi5LTqw2WN/O6dvsnt8sA8NlpL9HQ/c/y5 WnLx+rK9errbjAsh305QWpCBm9CSyN8c2JxWCce4KusLHq1x37tUf79I9+zemhw3Wm2s Qoz9s2uPHGMytP+ZxmrZqaUFCgMQWZ2wrFZqQKwdutDHwCryEPdz5h71Etz8HJYWuCWv VN/O/4baoi8qzN4pKlzSQ4wyL9xegAWnWkncnlRg4eTM0gp46RgSbTcNK6Vyd+3dkucZ dxOkDyobK575hru/RxPeT7bcXdou/TbVyjSKKQognqldmXvpupxZzxoc0WNSm8IQw8qe ctBA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:content-language:subject:user-agent:mime-version :date:message-id:dkim-signature; bh=FJE657/a/b3VekeiVjdoL+JS4Z4GUeEoDq18r7ombyg=; b=1BQeNjooNTMz2aneHRzJK9YMnCDn2zw4YgtGwDwbtxG7B2JnsiiYXWIihyDetwB3DX xivtFVwaSEnLE+nSfIczt3AgXNrYbK0aMXx0F/cHcfO2kHN810ty2QuqXz64gAF4bxDd LI/OhdnbFjE5JkJDbzPnmazd55K4hKfsgZAwt2I63yYpbdl74FMh9HJbte4MD+5oh3+f cDyxSO6mYm2r+i0/+liMPoaWJR4DsawBXoNg9lNcD6t+SJ/a8i4V1Lr62KEpsMJUWIuY YdAXulXioOV6RsMSP6oVXuDkyS6/u1ugEmn9UZ8Li6t6aYNtkwb9ccoRHvUSh8+aidMe ubkQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=EXTQLLkA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id z1-20020aa7c641000000b004fd2aaa62d0si33260080edr.242.2023.03.29.07.42.58; Wed, 29 Mar 2023 07:43:24 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=EXTQLLkA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230280AbjC2OhY (ORCPT + 99 others); Wed, 29 Mar 2023 10:37:24 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:56460 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230248AbjC2OhJ (ORCPT ); Wed, 29 Mar 2023 10:37:09 -0400 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E927F769B for ; Wed, 29 Mar 2023 07:32:18 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1680100278; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=FJE657/a/b3VekeiVjdoL+JS4Z4GUeEoDq18r7ombyg=; b=EXTQLLkAolm/bYnZn5K5/XQJwe0E78J2bdl6eYSCYo7+HkxB1lMgsIEnCBuclHQSHg/s53 Yq1RaIkQIzlIF62ENrCRd00wwbMvCEVAfV6Da//pNTkxn78JTNtCplvM+d7OZPQTT/fLSV s8ahRNPWEIuxVKINlOz8cjRvRdrr0IE= Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-588-H6RYuKwGNcWdbG0Ppbww6A-1; Wed, 29 Mar 2023 10:25:09 -0400 X-MC-Unique: H6RYuKwGNcWdbG0Ppbww6A-1 Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.rdu2.redhat.com [10.11.54.7]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 5344D85A5A3; Wed, 29 Mar 2023 14:25:08 +0000 (UTC) Received: from [10.22.34.224] (unknown [10.22.34.224]) by smtp.corp.redhat.com (Postfix) with ESMTP id 6924814171BB; Wed, 29 Mar 2023 14:25:07 +0000 (UTC) Message-ID: Date: Wed, 29 Mar 2023 10:25:07 -0400 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.7.1 Subject: Re: [PATCH 5/6] cgroup/cpuset: Free DL BW in case can_attach() fails Content-Language: en-US To: Juri Lelli , Peter Zijlstra , Ingo Molnar , Qais Yousef , Tejun Heo , Zefan Li , Johannes Weiner , Hao Luo Cc: Dietmar Eggemann , Steven Rostedt , linux-kernel@vger.kernel.org, luca.abeni@santannapisa.it, claudio@evidence.eu.com, tommaso.cucinotta@santannapisa.it, bristot@redhat.com, mathieu.poirier@linaro.org, cgroups@vger.kernel.org, Vincent Guittot , Wei Wang , Rick Yiu , Quentin Perret , Heiko Carstens , Vasily Gorbik , Alexander Gordeev , Sudeep Holla References: <20230329125558.255239-1-juri.lelli@redhat.com> <20230329125558.255239-6-juri.lelli@redhat.com> From: Waiman Long In-Reply-To: <20230329125558.255239-6-juri.lelli@redhat.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 3.1 on 10.11.54.7 X-Spam-Status: No, score=-0.2 required=5.0 tests=DKIMWL_WL_HIGH,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,NICE_REPLY_A,RCVD_IN_DNSWL_NONE, RCVD_IN_MSPIKE_H2,SPF_HELO_NONE,SPF_NONE autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 3/29/23 08:55, Juri Lelli wrote: > From: Dietmar Eggemann > > cpuset_can_attach() can fail. Postpone DL BW allocation until all tasks > have been checked. DL BW is not allocated per-task but as a sum over > all DL tasks migrating. > > If multiple controllers are attached to the cgroup next to the cuset Typo: : "cuset" => "cpuset" > controller a non-cpuset can_attach() can fail. In this case free DL BW > in cpuset_cancel_attach(). > > Finally, update cpuset DL task count (nr_deadline_tasks) only in > cpuset_attach(). > > Suggested-by: Waiman Long > Signed-off-by: Dietmar Eggemann > Signed-off-by: Juri Lelli > --- > include/linux/sched.h | 2 +- > kernel/cgroup/cpuset.c | 55 ++++++++++++++++++++++++++++++++++++++---- > kernel/sched/core.c | 17 ++----------- > 3 files changed, 53 insertions(+), 21 deletions(-) > > diff --git a/include/linux/sched.h b/include/linux/sched.h > index 6f3d84e0ed08..50cbbfefbe11 100644 > --- a/include/linux/sched.h > +++ b/include/linux/sched.h > @@ -1847,7 +1847,7 @@ current_restore_flags(unsigned long orig_flags, unsigned long flags) > } > > extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial); > -extern int task_can_attach(struct task_struct *p, const struct cpumask *cs_effective_cpus); > +extern int task_can_attach(struct task_struct *p); > extern int dl_bw_alloc(int cpu, u64 dl_bw); > extern void dl_bw_free(int cpu, u64 dl_bw); > #ifdef CONFIG_SMP > diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c > index eb0854ef9757..f8ebec66da51 100644 > --- a/kernel/cgroup/cpuset.c > +++ b/kernel/cgroup/cpuset.c > @@ -198,6 +198,8 @@ struct cpuset { > * know when to rebuild associated root domain bandwidth information. > */ > int nr_deadline_tasks; > + int nr_migrate_dl_tasks; > + u64 sum_migrate_dl_bw; > > /* Invalid partition error code, not lock protected */ > enum prs_errcode prs_err; > @@ -2464,16 +2466,23 @@ static int fmeter_getrate(struct fmeter *fmp) > > static struct cpuset *cpuset_attach_old_cs; > > +static void reset_migrate_dl_data(struct cpuset *cs) > +{ > + cs->nr_migrate_dl_tasks = 0; > + cs->sum_migrate_dl_bw = 0; > +} > + > /* Called by cgroups to determine if a cpuset is usable; cpuset_mutex held */ > static int cpuset_can_attach(struct cgroup_taskset *tset) > { > struct cgroup_subsys_state *css; > - struct cpuset *cs; > + struct cpuset *cs, *oldcs; > struct task_struct *task; > int ret; > > /* used later by cpuset_attach() */ > cpuset_attach_old_cs = task_cs(cgroup_taskset_first(tset, &css)); > + oldcs = cpuset_attach_old_cs; > cs = css_cs(css); > > mutex_lock(&cpuset_mutex); > @@ -2491,7 +2500,7 @@ static int cpuset_can_attach(struct cgroup_taskset *tset) > goto out_unlock; > > cgroup_taskset_for_each(task, css, tset) { > - ret = task_can_attach(task, cs->effective_cpus); > + ret = task_can_attach(task); > if (ret) > goto out_unlock; > ret = security_task_setscheduler(task); > @@ -2499,11 +2508,31 @@ static int cpuset_can_attach(struct cgroup_taskset *tset) > goto out_unlock; > > if (dl_task(task)) { > - cs->nr_deadline_tasks++; > - cpuset_attach_old_cs->nr_deadline_tasks--; > + cs->nr_migrate_dl_tasks++; > + cs->sum_migrate_dl_bw += task->dl.dl_bw; > + } > + } > + > + if (!cs->nr_migrate_dl_tasks) > + goto out_succes; > + > + if (!cpumask_intersects(oldcs->effective_cpus, cs->effective_cpus)) { > + int cpu = cpumask_any_and(cpu_active_mask, cs->effective_cpus); > + > + if (unlikely(cpu >= nr_cpu_ids)) { > + reset_migrate_dl_data(cs); > + ret = -EINVAL; > + goto out_unlock; > + } > + > + ret = dl_bw_alloc(cpu, cs->sum_migrate_dl_bw); > + if (ret) { > + reset_migrate_dl_data(cs); > + goto out_unlock; > } > } > > +out_succes: > /* > * Mark attach is in progress. This makes validate_change() fail > * changes which zero cpus/mems_allowed. > @@ -2518,11 +2547,21 @@ static int cpuset_can_attach(struct cgroup_taskset *tset) > static void cpuset_cancel_attach(struct cgroup_taskset *tset) > { > struct cgroup_subsys_state *css; > + struct cpuset *cs; > > cgroup_taskset_first(tset, &css); > + cs = css_cs(css); > > mutex_lock(&cpuset_mutex); > - css_cs(css)->attach_in_progress--; > + cs->attach_in_progress--; > + > + if (cs->nr_migrate_dl_tasks) { > + int cpu = cpumask_any(cs->effective_cpus); > + > + dl_bw_free(cpu, cs->sum_migrate_dl_bw); > + reset_migrate_dl_data(cs); > + } > + > mutex_unlock(&cpuset_mutex); > } > > @@ -2617,6 +2656,12 @@ static void cpuset_attach(struct cgroup_taskset *tset) > out: > cs->old_mems_allowed = cpuset_attach_nodemask_to; > > + if (cs->nr_migrate_dl_tasks) { > + cs->nr_deadline_tasks += cs->nr_migrate_dl_tasks; > + oldcs->nr_deadline_tasks -= cs->nr_migrate_dl_tasks; > + reset_migrate_dl_data(cs); > + } > + > cs->attach_in_progress--; > if (!cs->attach_in_progress) > wake_up(&cpuset_attach_wq); > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > index c83dae6b8586..10454980e830 100644 > --- a/kernel/sched/core.c > +++ b/kernel/sched/core.c > @@ -9269,8 +9269,7 @@ int cpuset_cpumask_can_shrink(const struct cpumask *cur, > return ret; > } > > -int task_can_attach(struct task_struct *p, > - const struct cpumask *cs_effective_cpus) > +int task_can_attach(struct task_struct *p) > { > int ret = 0; > > @@ -9283,21 +9282,9 @@ int task_can_attach(struct task_struct *p, > * success of set_cpus_allowed_ptr() on all attached tasks > * before cpus_mask may be changed. > */ > - if (p->flags & PF_NO_SETAFFINITY) { > + if (p->flags & PF_NO_SETAFFINITY) > ret = -EINVAL; > - goto out; > - } > - > - if (dl_task(p) && !cpumask_intersects(task_rq(p)->rd->span, > - cs_effective_cpus)) { > - int cpu = cpumask_any_and(cpu_active_mask, cs_effective_cpus); > > - if (unlikely(cpu >= nr_cpu_ids)) > - return -EINVAL; > - ret = dl_bw_alloc(cpu, p->dl.dl_bw); > - } > - > -out: > return ret; > } >