Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755953AbXJISSS (ORCPT ); Tue, 9 Oct 2007 14:18:18 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1753484AbXJISSH (ORCPT ); Tue, 9 Oct 2007 14:18:07 -0400 Received: from pentafluge.infradead.org ([213.146.154.40]:51085 "EHLO pentafluge.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753199AbXJISSF (ORCPT ); Tue, 9 Oct 2007 14:18:05 -0400 Subject: Re: [RFC PATCH RT] push waiting rt tasks to cpus with lower prios. From: Peter Zijlstra To: Steven Rostedt Cc: Gregory Haskins , Ingo Molnar , linux-rt-users , kravetz@us.ibm.com, LKML , pmorreale@novell.com, sdietrich@novell.com In-Reply-To: <1191952777.23198.8.camel@localhost.localdomain> References: <20071009142044.4941.65189.stgit@novell1.haskins.net> <1191944024.4281.72.camel@ghaskins-t60p.haskins.net> <1191952777.23198.8.camel@localhost.localdomain> Content-Type: text/plain Date: Tue, 09 Oct 2007 20:16:59 +0200 Message-Id: <1191953819.5797.9.camel@lappy> Mime-Version: 1.0 X-Mailer: Evolution 2.12.0 Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 6332 Lines: 219 On Tue, 2007-10-09 at 13:59 -0400, Steven Rostedt wrote: > This has been complied tested (and no more ;-) > > > The idea here is when we find a situation that we just scheduled in an > RT task and we either pushed a lesser RT task away or more than one RT > task was scheduled on this CPU before scheduling occurred. > > The answer that this patch does is to do a O(n) search of CPUs for the > CPU with the lowest prio task running. When that CPU is found the next > highest RT task is pushed to that CPU. > > Some notes: > > 1) no lock is taken while looking for the lowest priority CPU. When one > is found, only that CPU's lock is taken and after that a check is made > to see if it is still a candidate to push the RT task over. If not, we > try the search again, for a max of 3 tries. > > 2) I only do this for the second highest RT task on the CPU queue. This > can be easily changed to do it for all RT tasks until no more can be > pushed off to other CPUs. > > This is a simple approach right now, and is only being posted for > comments. I'm sure more can be done to make this more efficient or just > simply better. > > -- Steve Do we really want this PREEMPT_RT only? > Signed-off-by: Steven Rostedt > > Index: linux-2.6.23-rc9-rt2/kernel/sched.c > =================================================================== > --- linux-2.6.23-rc9-rt2.orig/kernel/sched.c > +++ linux-2.6.23-rc9-rt2/kernel/sched.c > @@ -304,6 +304,7 @@ struct rq { > #ifdef CONFIG_PREEMPT_RT > unsigned long rt_nr_running; > unsigned long rt_nr_uninterruptible; > + int curr_prio; > #endif > > unsigned long switch_timestamp; > @@ -1485,6 +1486,87 @@ next_in_queue: > static int double_lock_balance(struct rq *this_rq, struct rq *busiest); > > /* > + * If the current CPU has more than one RT task, see if the non > + * running task can migrate over to a CPU that is running a task > + * of lesser priority. > + */ > +static int push_rt_task(struct rq *this_rq) > +{ > + struct task_struct *next_task; > + struct rq *lowest_rq = NULL; > + int tries; > + int cpu; > + int dst_cpu = -1; > + int ret = 0; > + > + BUG_ON(!spin_is_locked(&this_rq->lock)); assert_spin_locked(&this_rq->lock); > + > + next_task = rt_next_highest_task(this_rq); > + if (!next_task) > + return 0; > + > + /* We might release this_rq lock */ > + get_task_struct(next_task); Can the rest of the code suffer this? (the caller that is) > + /* Only try this algorithm three times */ > + for (tries = 0; tries < 3; tries++) { magic numbers.. maybe a magic #define with a descriptive name? > + /* > + * Scan each rq for the lowest prio. > + */ > + for_each_cpu_mask(cpu, next_task->cpus_allowed) { > + struct rq *rq = &per_cpu(runqueues, cpu); > + > + if (cpu == smp_processor_id()) > + continue; > + > + /* no locking for now */ > + if (rq->curr_prio > next_task->prio && > + (!lowest_rq || rq->curr_prio < lowest_rq->curr_prio)) { > + dst_cpu = cpu; > + lowest_rq = rq; > + } > + } > + > + if (!lowest_rq) > + break; > + > + if (double_lock_balance(this_rq, lowest_rq)) { > + /* > + * We had to unlock the run queue. In > + * the mean time, next_task could have > + * migrated already or had its affinity changed. > + */ > + if (unlikely(task_rq(next_task) != this_rq || > + !cpu_isset(dst_cpu, next_task->cpus_allowed))) { > + spin_unlock(&lowest_rq->lock); > + break; > + } > + } > + > + /* if the prio of this runqueue changed, try again */ > + if (lowest_rq->curr_prio <= next_task->prio) { > + spin_unlock(&lowest_rq->lock); > + continue; > + } > + > + deactivate_task(this_rq, next_task, 0); > + set_task_cpu(next_task, dst_cpu); > + activate_task(lowest_rq, next_task, 0); > + > + set_tsk_need_resched(lowest_rq->curr); Use resched_task(), that will notify the remote cpu too. > + > + spin_unlock(&lowest_rq->lock); > + ret = 1; > + > + break; > + } > + > + put_task_struct(next_task); > + > + return ret; > +} > + > +/* > * Pull RT tasks from other CPUs in the RT-overload > * case. Interrupts are disabled, local rq is locked. > */ > @@ -2207,7 +2289,8 @@ static inline void finish_task_switch(st > * If we pushed an RT task off the runqueue, > * then kick other CPUs, they might run it: > */ > - if (unlikely(rt_task(current) && rq->rt_nr_running > 1)) { > + rq->curr_prio = current->prio; > + if (unlikely(rt_task(current) && push_rt_task(rq))) { > schedstat_inc(rq, rto_schedule); > smp_send_reschedule_allbutself_cpumask(current->cpus_allowed); Which will allow you to remove this thing. > } > Index: linux-2.6.23-rc9-rt2/kernel/sched_rt.c > =================================================================== > --- linux-2.6.23-rc9-rt2.orig/kernel/sched_rt.c > +++ linux-2.6.23-rc9-rt2/kernel/sched_rt.c > @@ -96,6 +96,48 @@ static struct task_struct *pick_next_tas > return next; > } > > +#ifdef CONFIG_PREEMPT_RT > +static struct task_struct *rt_next_highest_task(struct rq *rq) > +{ > + struct rt_prio_array *array = &rq->rt.active; > + struct task_struct *next; > + struct list_head *queue; > + int idx; > + > + if (likely (rq->rt_nr_running < 2)) > + return NULL; > + > + idx = sched_find_first_bit(array->bitmap); > + if (idx >= MAX_RT_PRIO) { > + WARN_ON(1); /* rt_nr__running is bad */ > + return NULL; > + } > + > + queue = array->queue + idx; > + if (queue->next->next != queue) { > + /* same prio task */ > + next = list_entry(queue->next->next, struct task_struct, run_list); > + goto out; > + } > + > + /* slower, but more flexible */ > + idx = find_next_bit(array->bitmap, MAX_RT_PRIO, idx+1); > + if (idx >= MAX_RT_PRIO) { > + WARN_ON(1); /* rt_nr_running was 2 and above! */ > + return NULL; > + } > + > + queue = array->queue + idx; > + next = list_entry(queue->next, struct task_struct, run_list); > + > + out: > + return next; > + > +} > +#else /* CONFIG_PREEMPT_RT */ > + > +#endif /* CONFIG_PREEMPT_RT */ > + > static void put_prev_task_rt(struct rq *rq, struct task_struct *p) > { > update_curr_rt(rq); > > - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/