Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754790AbYGVOQT (ORCPT ); Tue, 22 Jul 2008 10:16:19 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1750798AbYGVOQA (ORCPT ); Tue, 22 Jul 2008 10:16:00 -0400 Received: from viefep18-int.chello.at ([213.46.255.22]:56810 "EHLO viefep19-int.chello.at" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1750869AbYGVOP7 (ORCPT ); Tue, 22 Jul 2008 10:15:59 -0400 X-SourceIP: 62.163.52.83 Subject: Re: [PATCH] cpu hotplug, sched:Introduce cpu_active_map and redoscheddomainmanagment (take 2) From: Peter Zijlstra To: Gregory Haskins Cc: Max Krasnyansky , mingo@elte.hu, dmitry.adamushko@gmail.com, torvalds@linux-foundation.org, pj@sgi.com, linux-kernel@vger.kernel.org In-Reply-To: <4885E952.8020708@novell.com> References: <1216122229-4865-1-git-send-email-maxk@qualcomm.com> <487DAD86.BA47.005A.0@novell.com> <487E6BD7.3020006@qualcomm.com> <487E7B6C.BA47.005A.0@novell.com> <487EF1E9.2040101@qualcomm.com> <487EFB71.BA47.005A.0@novell.com> <487F9509.9050802@qualcomm.com> <487F6972.BA47.005A.0@novell.com> <1216382024.28405.26.camel@twins> <488052D5.BA47.005A.0@novell.com> <48856BA9.6050609@qualcomm.com> <4885E952.8020708@novell.com> Content-Type: text/plain Date: Tue, 22 Jul 2008 16:16:00 +0200 Message-Id: <1216736160.7257.97.camel@twins> Mime-Version: 1.0 X-Mailer: Evolution 2.22.3.1 Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 4082 Lines: 112 On Tue, 2008-07-22 at 10:06 -0400, Gregory Haskins wrote: > Max Krasnyansky wrote: > > Greg, correct me if I'm wrong but we seem to have exact same issue with the > > rq->rq->online map. Lets take "cpu going down" for example. We're clearing > > rq->rd->online bit on DYING event, but nothing AFAICS prevents another cpu > > calling rebuild_sched_domains()->partition_sched_domains() in the middle of > > the hotplug sequence. > > partition_sched_domains() will happily reset rd->rq->online mask and things > > will fail. I'm talking about this path > > > > __build_sched_domains() -> cpu_attach_domain() -> rq_attach_root() > > ... > > cpu_set(rq->cpu, rd->span); > > if (cpu_isset(rq->cpu, cpu_online_map)) > > set_rq_online(rq); > > ... > > > > > > I think you are right, but wouldn't s/online/active above fix that as > well? The active_map didnt exist at the time that code went in initially ;) > > > -- > > > > btw Why didn't we convert sched*.c to use rq->rd->online when it was > > introduced ? ie Instead of using cpu_online_map directly. > > > I think things were converted where they made sense to convert. But we > also had a different goal at that time, so perhaps something was > missed. If you think something else should be converted, please point > it out. > > In the meantime, I would suggest we consider this patch on top of yours > (applies to tip/sched/devel): > > ---------------------- > > sched: Fully integrate cpus_active_map and root-domain code > > Signed-off-by: Gregory Haskins Right, makes sense. ACK > diff --git a/kernel/sched.c b/kernel/sched.c > index 62b1b8e..99ba70d 100644 > --- a/kernel/sched.c > +++ b/kernel/sched.c > @@ -6611,7 +6611,7 @@ static void rq_attach_root(struct rq *rq, struct > root_domain *rd) > rq->rd = rd; > > cpu_set(rq->cpu, rd->span); > - if (cpu_isset(rq->cpu, cpu_online_map)) > + if (cpu_isset(rq->cpu, cpu_active_map)) > set_rq_online(rq); > > spin_unlock_irqrestore(&rq->lock, flags); > diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c > index 7f70026..2bae8de 100644 > --- a/kernel/sched_fair.c > +++ b/kernel/sched_fair.c > @@ -1004,7 +1004,7 @@ static void yield_task_fair(struct rq *rq) > * search starts with cpus closest then further out as needed, > * so we always favor a closer, idle cpu. > * Domains may include CPUs that are not usable for migration, > - * hence we need to mask them out (cpu_active_map) > + * hence we need to mask them out (rq->rd->online) > * > * Returns the CPU we should wake onto. > */ > @@ -1032,7 +1032,7 @@ static int wake_idle(int cpu, struct task_struct *p) > || ((sd->flags & SD_WAKE_IDLE_FAR) > && !task_hot(p, task_rq(p)->clock, sd))) { > cpus_and(tmp, sd->span, p->cpus_allowed); > - cpus_and(tmp, tmp, cpu_active_map); > + cpus_and(tmp, tmp, task_rq(p)->rd->online); > for_each_cpu_mask(i, tmp) { > if (idle_cpu(i)) { > if (i != task_cpu(p)) { > diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c > index 24621ce..d93169d 100644 > --- a/kernel/sched_rt.c > +++ b/kernel/sched_rt.c > @@ -936,13 +936,6 @@ static int find_lowest_rq(struct task_struct *task) > return -1; /* No targets found */ > > /* > - * Only consider CPUs that are usable for migration. > - * I guess we might want to change cpupri_find() to ignore those > - * in the first place. > - */ > - cpus_and(*lowest_mask, *lowest_mask, cpu_active_map); > - > - /* > * At this point we have built a mask of cpus representing the > * lowest priority tasks in the system. Now we want to elect > * the best one based on our affinity and topology. > > -------------- > > Regards, > -Greg > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/