Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756678AbYKIVYs (ORCPT ); Sun, 9 Nov 2008 16:24:48 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1755926AbYKIVYk (ORCPT ); Sun, 9 Nov 2008 16:24:40 -0500 Received: from casper.infradead.org ([85.118.1.10]:35424 "EHLO casper.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754243AbYKIVYi (ORCPT ); Sun, 9 Nov 2008 16:24:38 -0500 Date: Sun, 9 Nov 2008 13:24:38 -0800 From: Arjan van de Ven To: Lukas Hejtmanek Cc: "Rafael J. Wysocki" , Marcin Slusarz , linux-kernel@vger.kernel.org, Thomas Gleixner , torvalds@linux-foundation.org, corsac@debian.org, elendil@planet.nl Subject: Re: Bootup time regression from 2.6.27 to 2.6.28-rc3+ Message-ID: <20081109132438.5d81cd74@infradead.org> In-Reply-To: <20081109203432.GB4361@ics.muni.cz> References: <20081104173300.GF6134@ics.muni.cz> <20081109105159.GA5867@joi> <20081109113304.GE4353@ics.muni.cz> <200811091424.10943.rjw@sisk.pl> <20081109200918.GA4361@ics.muni.cz> <20081109122835.10f410fa@infradead.org> <20081109203432.GB4361@ics.muni.cz> Organization: Intel X-Mailer: Claws Mail 3.6.0 (GTK+ 2.14.4; i386-redhat-linux-gnu) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit X-SRS-Rewrite: SMTP reverse-path rewritten from by casper.infradead.org See http://www.infradead.org/rpr.html Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 28381 Lines: 959 On Sun, 9 Nov 2008 21:34:32 +0100 Lukas Hejtmanek wrote: > On Sun, Nov 09, 2008 at 12:28:35PM -0800, Arjan van de Ven wrote: > > for me, the plan is that we need to get Thomas' fixes tested by > > someone who can reproduce this very reliably. If they fix it, great. > > as for me, 100% of boots require a key hit to proceed in 2.6.28-rc3 > (more precisely, the current git head). > it would be greatly appreciated if those that see this issue can test the patch below (which is a combo patch of thomas' fixes, to make testing easier) and confirm if this fixes the hangs. If they do fix the issue we don't need to disable anything and get better power savings as bonus as well. Gautham R Shenoy (1): timers: handle HRTIMER_CB_IRQSAFE_UNLOCKED correctly from softirq context Ingo Molnar (2): irq: fix typo nohz: fix SMP race Thomas Gleixner (12): nohz: check broadcast on tick restart x86: apic reset counter on shutdown hrtimer: fix the accounting of hrtimer_peek_ahead() hrtimer: do not peek ahead when the hardware timer has fired already tick: clear broadcast mask when timer is reprogrammed irq: call __irq_enter() before calling the tick_idle_check x86: apic honour irq affinity which was set in early boot genirq: keep affinities set from userspace across free/request_irq() genirq: fix the affinity setting in setup_irq tick: cleanup idle checks ACPI: processor_idle: check for spurious wakeups ACPI: idle check for spurious wakeups accounting fix arch/x86/kernel/apic.c | 1 + arch/x86/kernel/io_apic.c | 26 ++++++++++-- drivers/acpi/processor_idle.c | 49 ++++++++++++++++++----- drivers/cpuidle/cpuidle.c | 5 ++- include/linux/hrtimer.h | 4 +- include/linux/irq.h | 8 +--- include/linux/tick.h | 5 ++ kernel/hrtimer.c | 81 +++++++++++++++++++++++++++++-------- kernel/irq/internals.h | 2 + kernel/irq/manage.c | 66 ++++++++++++++++++++++++------ kernel/irq/migration.c | 11 ----- kernel/irq/proc.c | 2 +- kernel/softirq.c | 7 ++- kernel/time/tick-broadcast.c | 27 ++++++++++++ kernel/time/tick-internal.h | 5 ++ kernel/time/tick-sched.c | 88 ++++++++++++++++++++++------------------ kernel/time/timer_list.c | 4 +- 17 files changed, 279 insertions(+), 112 deletions(-) diff --git a/arch/x86/kernel/apic.c b/arch/x86/kernel/apic.c index 04a7f96..70003b8 100644 --- a/arch/x86/kernel/apic.c +++ b/arch/x86/kernel/apic.c @@ -441,6 +441,7 @@ static void lapic_timer_setup(enum clock_event_mode mode, v = apic_read(APIC_LVTT); v |= (APIC_LVT_MASKED | LOCAL_TIMER_VECTOR); apic_write(APIC_LVTT, v); + apic_write(APIC_TMICT, 0xffffffff); break; case CLOCK_EVT_MODE_RESUME: /* Nothing to do here */ diff --git a/arch/x86/kernel/io_apic.c b/arch/x86/kernel/io_apic.c index 7a3f202..988ee89 100644 --- a/arch/x86/kernel/io_apic.c +++ b/arch/x86/kernel/io_apic.c @@ -3761,7 +3761,9 @@ int acpi_get_override_irq(int bus_irq, int *trigger, int *polarity) void __init setup_ioapic_dest(void) { int pin, ioapic, irq, irq_entry; + struct irq_desc *desc; struct irq_cfg *cfg; + cpumask_t mask; if (skip_ioapic_setup == 1) return; @@ -3778,16 +3780,30 @@ void __init setup_ioapic_dest(void) * cpu is online. */ cfg = irq_cfg(irq); - if (!cfg->vector) + if (!cfg->vector) { setup_IO_APIC_irq(ioapic, pin, irq, irq_trigger(irq_entry), irq_polarity(irq_entry)); + continue; + + } + + /* + * Honour affinities which have been set in early boot + */ + desc = irq_to_desc(irq); + if (desc->status & + (IRQ_NO_BALANCING | IRQ_AFFINITY_SET)) + mask = desc->affinity; + else + mask = TARGET_CPUS; + #ifdef CONFIG_INTR_REMAP - else if (intr_remapping_enabled) - set_ir_ioapic_affinity_irq(irq, TARGET_CPUS); -#endif + if (intr_remapping_enabled) + set_ir_ioapic_affinity_irq(irq, mask); else - set_ioapic_affinity_irq(irq, TARGET_CPUS); +#endif + set_ioapic_affinity_irq(irq, mask); } } diff --git a/drivers/acpi/processor_idle.c b/drivers/acpi/processor_idle.c index 81b40ed..371dc42 100644 --- a/drivers/acpi/processor_idle.c +++ b/drivers/acpi/processor_idle.c @@ -41,6 +41,7 @@ #include #include #include +#include /* * Include the apic definitions for x86 to have the APIC timer related defines @@ -1458,6 +1459,7 @@ static int acpi_idle_enter_c1(struct cpuidle_device *dev, u32 t1, t2; struct acpi_processor *pr; struct acpi_processor_cx *cx = cpuidle_get_statedata(state); + int sleep_ticks = 0; pr = __get_cpu_var(processors); @@ -1473,6 +1475,7 @@ static int acpi_idle_enter_c1(struct cpuidle_device *dev, return 0; } +again: if (pr->flags.bm_check) acpi_idle_update_bm_rld(pr, cx); @@ -1480,10 +1483,18 @@ static int acpi_idle_enter_c1(struct cpuidle_device *dev, acpi_idle_do_entry(cx); t2 = inl(acpi_gbl_FADT.xpm_timer_block.address); + sleep_ticks += ticks_elapsed(t1, t2); + local_irq_enable(); + + /* Check for spurious wakeup */ + if (check_idle_spurious_wakeup(pr->id) && !need_resched()) { + local_irq_disable(); + goto again; + } cx->usage++; - return ticks_elapsed_in_us(t1, t2); + return sleep_ticks; } /** @@ -1497,7 +1508,7 @@ static int acpi_idle_enter_simple(struct cpuidle_device *dev, struct acpi_processor *pr; struct acpi_processor_cx *cx = cpuidle_get_statedata(state); u32 t1, t2; - int sleep_ticks = 0; + int ticks, sleep_ticks = 0; pr = __get_cpu_var(processors); @@ -1527,6 +1538,7 @@ static int acpi_idle_enter_simple(struct cpuidle_device *dev, */ acpi_state_timer_broadcast(pr, cx, 1); +again: if (pr->flags.bm_check) acpi_idle_update_bm_rld(pr, cx); @@ -1544,12 +1556,19 @@ static int acpi_idle_enter_simple(struct cpuidle_device *dev, if (tsc_halts_in_c(cx->type)) mark_tsc_unstable("TSC halts in idle");; #endif - sleep_ticks = ticks_elapsed(t1, t2); - + ticks = ticks_elapsed(t1, t2); /* Tell the scheduler how much we idled: */ - sched_clock_idle_wakeup_event(sleep_ticks*PM_TIMER_TICK_NS); + sched_clock_idle_wakeup_event(ticks*PM_TIMER_TICK_NS); local_irq_enable(); + sleep_ticks += ticks; + + /* Check for spurious wakeup */ + if (check_idle_spurious_wakeup(pr->id) && !need_resched()) { + local_irq_disable(); + goto again; + } + current_thread_info()->status |= TS_POLLING; cx->usage++; @@ -1575,7 +1594,7 @@ static int acpi_idle_enter_bm(struct cpuidle_device *dev, struct acpi_processor *pr; struct acpi_processor_cx *cx = cpuidle_get_statedata(state); u32 t1, t2; - int sleep_ticks = 0; + int ticks, sleep_ticks = 0; pr = __get_cpu_var(processors); @@ -1613,14 +1632,16 @@ static int acpi_idle_enter_bm(struct cpuidle_device *dev, acpi_unlazy_tlb(smp_processor_id()); - /* Tell the scheduler that we are going deep-idle: */ - sched_clock_idle_sleep_event(); /* * Must be done before busmaster disable as we might need to * access HPET ! */ acpi_state_timer_broadcast(pr, cx, 1); +again: + /* Tell the scheduler that we are going deep-idle: */ + sched_clock_idle_sleep_event(); + acpi_idle_update_bm_rld(pr, cx); /* @@ -1661,11 +1682,19 @@ static int acpi_idle_enter_bm(struct cpuidle_device *dev, if (tsc_halts_in_c(ACPI_STATE_C3)) mark_tsc_unstable("TSC halts in idle"); #endif - sleep_ticks = ticks_elapsed(t1, t2); + ticks = ticks_elapsed(t1, t2); /* Tell the scheduler how much we idled: */ - sched_clock_idle_wakeup_event(sleep_ticks*PM_TIMER_TICK_NS); + sched_clock_idle_wakeup_event(ticks*PM_TIMER_TICK_NS); local_irq_enable(); + sleep_ticks += ticks; + + /* Check for spurious wakeup */ + if (check_idle_spurious_wakeup(pr->id) && !need_resched()) { + local_irq_disable(); + goto again; + } + current_thread_info()->status |= TS_POLLING; cx->usage++; diff --git a/drivers/cpuidle/cpuidle.c b/drivers/cpuidle/cpuidle.c index 5bed733..f3ca665 100644 --- a/drivers/cpuidle/cpuidle.c +++ b/drivers/cpuidle/cpuidle.c @@ -69,7 +69,10 @@ static void cpuidle_idle_call(void) * run any timers that can be run now, at this point * before calculating the idle duration etc. */ - hrtimer_peek_ahead_timers(); + if (hrtimer_peek_ahead_timers()) { + local_irq_enable(); + return; + } /* ask the governor for the next state */ next_state = cpuidle_curr_governor->select(dev); diff --git a/include/linux/hrtimer.h b/include/linux/hrtimer.h index 07e510a..e3d7d9f 100644 --- a/include/linux/hrtimer.h +++ b/include/linux/hrtimer.h @@ -294,7 +294,7 @@ static inline int hrtimer_is_hres_active(struct hrtimer *timer) return timer->base->cpu_base->hres_active; } -extern void hrtimer_peek_ahead_timers(void); +extern int hrtimer_peek_ahead_timers(void); /* * The resolution of the clocks. The resolution value is returned in @@ -318,7 +318,7 @@ extern void hrtimer_peek_ahead_timers(void); * is expired in the next softirq when the clock was advanced. */ static inline void clock_was_set(void) { } -static inline void hrtimer_peek_ahead_timers(void) { } +static inline int hrtimer_peek_ahead_timers(void) { return 0; } static inline void hres_timers_resume(void) { } diff --git a/include/linux/irq.h b/include/linux/irq.h index d058c57..36b186e 100644 --- a/include/linux/irq.h +++ b/include/linux/irq.h @@ -63,7 +63,8 @@ typedef void (*irq_flow_handler_t)(unsigned int irq, #define IRQ_MOVE_PENDING 0x00200000 /* need to re-target IRQ destination */ #define IRQ_NO_BALANCING 0x00400000 /* IRQ is excluded from balancing */ #define IRQ_SPURIOUS_DISABLED 0x00800000 /* IRQ was disabled by the spurious trap */ -#define IRQ_MOVE_PCNTXT 0x01000000 /* IRQ migration from process context */ +#define IRQ_MOVE_PCNTXT 0x01000000 /* IRQ migration from process context */ +#define IRQ_AFFINITY_SET 0x02000000 /* IRQ affinity was set from userspace*/ #ifdef CONFIG_IRQ_PER_CPU # define CHECK_IRQ_PER_CPU(var) ((var) & IRQ_PER_CPU) @@ -210,7 +211,6 @@ extern int setup_irq(unsigned int irq, struct irqaction *new); #ifdef CONFIG_GENERIC_PENDING_IRQ -void set_pending_irq(unsigned int irq, cpumask_t mask); void move_native_irq(int irq); void move_masked_irq(int irq); @@ -228,10 +228,6 @@ static inline void move_masked_irq(int irq) { } -static inline void set_pending_irq(unsigned int irq, cpumask_t mask) -{ -} - #endif /* CONFIG_GENERIC_PENDING_IRQ */ #else /* CONFIG_SMP */ diff --git a/include/linux/tick.h b/include/linux/tick.h index b6ec818..ff202ac 100644 --- a/include/linux/tick.h +++ b/include/linux/tick.h @@ -64,6 +64,9 @@ struct tick_sched { unsigned long last_jiffies; unsigned long next_jiffies; ktime_t idle_expires; + unsigned long irq_wakeups; + unsigned long irq_last_wakeups; + unsigned long spurious_wakeups; }; extern void __init tick_init(void); @@ -116,6 +119,7 @@ extern void tick_nohz_stop_sched_tick(int inidle); extern void tick_nohz_restart_sched_tick(void); extern ktime_t tick_nohz_get_sleep_length(void); extern u64 get_cpu_idle_time_us(int cpu, u64 *last_update_time); +extern int check_idle_spurious_wakeup(int cpu); # else static inline void tick_nohz_stop_sched_tick(int inidle) { } static inline void tick_nohz_restart_sched_tick(void) { } @@ -126,6 +130,7 @@ static inline ktime_t tick_nohz_get_sleep_length(void) return len; } static inline u64 get_cpu_idle_time_us(int cpu, u64 *unused) { return -1; } +static inline int check_idle_spurious_wakeup(int cpu) { return 0; } # endif /* !NO_HZ */ #endif diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index 2b465df..075bc8e 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c @@ -1209,6 +1209,7 @@ static void run_hrtimer_pending(struct hrtimer_cpu_base *cpu_base) enum hrtimer_restart (*fn)(struct hrtimer *); struct hrtimer *timer; int restart; + int emulate_hardirq_ctx = 0; timer = list_entry(cpu_base->cb_pending.next, struct hrtimer, cb_entry); @@ -1217,10 +1218,24 @@ static void run_hrtimer_pending(struct hrtimer_cpu_base *cpu_base) timer_stats_account_hrtimer(timer); fn = timer->function; + /* + * A timer might have been added to the cb_pending list + * when it was migrated during a cpu-offline operation. + * Emulate hardirq context for such timers. + */ + if (timer->cb_mode == HRTIMER_CB_IRQSAFE_PERCPU || + timer->cb_mode == HRTIMER_CB_IRQSAFE_UNLOCKED) + emulate_hardirq_ctx = 1; + __remove_hrtimer(timer, timer->base, HRTIMER_STATE_CALLBACK, 0); spin_unlock_irq(&cpu_base->lock); - restart = fn(timer); + if (unlikely(emulate_hardirq_ctx)) { + local_irq_disable(); + restart = fn(timer); + local_irq_enable(); + } else + restart = fn(timer); spin_lock_irq(&cpu_base->lock); @@ -1297,23 +1312,26 @@ static void __run_hrtimer(struct hrtimer *timer) #ifdef CONFIG_HIGH_RES_TIMERS /* - * High resolution timer interrupt - * Called with interrupts disabled + * High resolution timer interrupt internal worker function called + * with interrupts disabled either from hrtimer_interrupt() or from + * hrtimer_peek_ahead() */ -void hrtimer_interrupt(struct clock_event_device *dev) +int __hrtimer_interrupt(struct hrtimer_cpu_base *cpu_base, int peekahead) { - struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases); struct hrtimer_clock_base *base; ktime_t expires_next, now; int i, raise = 0; - BUG_ON(!cpu_base->hres_active); - cpu_base->nr_events++; - dev->next_event.tv64 = KTIME_MAX; - retry: now = ktime_get(); + /* + * In peek ahead mode bail out, if the hw interrupt is + * imminent: + */ + if (peekahead && cpu_base->expires_next.tv64 < now.tv64) + return -ETIME; + expires_next.tv64 = KTIME_MAX; base = cpu_base->clock_base; @@ -1370,17 +1388,49 @@ void hrtimer_interrupt(struct clock_event_device *dev) base++; } + /* + * We just peeked ahead. The hardware timer did not expire. So + * we can leave the timer armed. + */ + if (peekahead && cpu_base->expires_next.tv64 == expires_next.tv64) + goto out; + cpu_base->expires_next = expires_next; /* Reprogramming necessary ? */ if (expires_next.tv64 != KTIME_MAX) { + /* + * Clear the peeakahead flag once we decided to + * reprogram. Otherwise we break out in the check + * above. + */ + peekahead = 0; if (tick_program_event(expires_next, 0)) goto retry; } +out: /* Raise softirq ? */ if (raise) raise_softirq(HRTIMER_SOFTIRQ); + + return 0; +} + + +/* + * High resolution timer interrupt + * Called with interrupts disabled + */ +void hrtimer_interrupt(struct clock_event_device *dev) +{ + struct hrtimer_cpu_base *cpu_base = &__get_cpu_var(hrtimer_bases); + + BUG_ON(!cpu_base->hres_active); + cpu_base->nr_events++; + dev->next_event.tv64 = KTIME_MAX; + + __hrtimer_interrupt(cpu_base, 0); } /** @@ -1392,19 +1442,14 @@ void hrtimer_interrupt(struct clock_event_device *dev) * they are run immediately and then removed from the timer queue. * */ -void hrtimer_peek_ahead_timers(void) +int hrtimer_peek_ahead_timers(void) { - struct tick_device *td; - unsigned long flags; + WARN_ON_ONCE(!irqs_disabled()); if (!hrtimer_hres_active()) - return; + return 0; - local_irq_save(flags); - td = &__get_cpu_var(tick_cpu_device); - if (td && td->evtdev) - hrtimer_interrupt(td->evtdev); - local_irq_restore(flags); + return __hrtimer_interrupt(&__get_cpu_var(hrtimer_bases), 1); } static void run_hrtimer_softirq(struct softirq_action *h) diff --git a/kernel/irq/internals.h b/kernel/irq/internals.h index c9767e6..64c1c72 100644 --- a/kernel/irq/internals.h +++ b/kernel/irq/internals.h @@ -25,6 +25,8 @@ static inline void unregister_handler_proc(unsigned int irq, struct irqaction *action) { } #endif +extern int irq_select_affinity_usr(unsigned int irq); + /* * Debugging printout: */ diff --git a/kernel/irq/manage.c b/kernel/irq/manage.c index c498a1b..4358612 100644 --- a/kernel/irq/manage.c +++ b/kernel/irq/manage.c @@ -82,24 +82,27 @@ int irq_can_set_affinity(unsigned int irq) int irq_set_affinity(unsigned int irq, cpumask_t cpumask) { struct irq_desc *desc = irq_to_desc(irq); + unsigned long flags; if (!desc->chip->set_affinity) return -EINVAL; + spin_lock_irqsave(&desc->lock, flags); + #ifdef CONFIG_GENERIC_PENDING_IRQ if (desc->status & IRQ_MOVE_PCNTXT || desc->status & IRQ_DISABLED) { - unsigned long flags; - - spin_lock_irqsave(&desc->lock, flags); desc->affinity = cpumask; desc->chip->set_affinity(irq, cpumask); - spin_unlock_irqrestore(&desc->lock, flags); - } else - set_pending_irq(irq, cpumask); + } else { + desc->status |= IRQ_MOVE_PENDING; + desc->pending_mask = cpumask; + } #else desc->affinity = cpumask; desc->chip->set_affinity(irq, cpumask); #endif + desc->status |= IRQ_AFFINITY_SET; + spin_unlock_irqrestore(&desc->lock, flags); return 0; } @@ -107,24 +110,59 @@ int irq_set_affinity(unsigned int irq, cpumask_t cpumask) /* * Generic version of the affinity autoselector. */ -int irq_select_affinity(unsigned int irq) +int do_irq_select_affinity(unsigned int irq, struct irq_desc *desc) { cpumask_t mask; - struct irq_desc *desc; if (!irq_can_set_affinity(irq)) return 0; cpus_and(mask, cpu_online_map, irq_default_affinity); - desc = irq_to_desc(irq); + /* + * Preserve an userspace affinity setup, but make sure that + * one of the targets is online. + */ + if (desc->status & (IRQ_AFFINITY_SET | IRQ_NO_BALANCING)) { + if (cpus_intersects(desc->affinity, cpu_online_map)) + mask = desc->affinity; + else + desc->status &= ~IRQ_AFFINITY_SET; + } + desc->affinity = mask; desc->chip->set_affinity(irq, mask); return 0; } +#else +static inline int do_irq_select_affinity(unsigned int irq, struct irq_desc *d) +{ + return irq_select_affinity(irq); +} #endif +/* + * Called when affinity is set via /proc/irq + */ +int irq_select_affinity_usr(unsigned int irq) +{ + struct irq_desc *desc = irq_to_desc(irq); + unsigned long flags; + int ret; + + spin_lock_irqsave(&desc->lock, flags); + ret = do_irq_select_affinity(irq, desc); + spin_unlock_irqrestore(&desc->lock, flags); + + return ret; +} + +#else +static inline int do_irq_select_affinity(int irq, struct irq_desc *desc) +{ + return 0; +} #endif /** @@ -445,8 +483,12 @@ __setup_irq(unsigned int irq, struct irq_desc * desc, struct irqaction *new) /* Undo nested disables: */ desc->depth = 1; + /* Exclude IRQ from balancing if requested */ + if (new->flags & IRQF_NOBALANCING) + desc->status |= IRQ_NO_BALANCING; + /* Set default affinity mask once everything is setup */ - irq_select_affinity(irq); + do_irq_select_affinity(irq, desc); } else if ((new->flags & IRQF_TRIGGER_MASK) && (new->flags & IRQF_TRIGGER_MASK) @@ -459,10 +501,6 @@ __setup_irq(unsigned int irq, struct irq_desc * desc, struct irqaction *new) *p = new; - /* Exclude IRQ from balancing */ - if (new->flags & IRQF_NOBALANCING) - desc->status |= IRQ_NO_BALANCING; - /* Reset broken irq detection when installing new handler */ desc->irq_count = 0; desc->irqs_unhandled = 0; diff --git a/kernel/irq/migration.c b/kernel/irq/migration.c index 90b920d..9db681d 100644 --- a/kernel/irq/migration.c +++ b/kernel/irq/migration.c @@ -1,17 +1,6 @@ #include -void set_pending_irq(unsigned int irq, cpumask_t mask) -{ - struct irq_desc *desc = irq_to_desc(irq); - unsigned long flags; - - spin_lock_irqsave(&desc->lock, flags); - desc->status |= IRQ_MOVE_PENDING; - desc->pending_mask = mask; - spin_unlock_irqrestore(&desc->lock, flags); -} - void move_masked_irq(int irq) { struct irq_desc *desc = irq_to_desc(irq); diff --git a/kernel/irq/proc.c b/kernel/irq/proc.c index 4d161c7..d257e7d 100644 --- a/kernel/irq/proc.c +++ b/kernel/irq/proc.c @@ -62,7 +62,7 @@ static ssize_t irq_affinity_proc_write(struct file *file, if (!cpus_intersects(new_value, cpu_online_map)) /* Special case for empty set - allow the architecture code to set default SMP affinity. */ - return irq_select_affinity(irq) ? -EINVAL : count; + return irq_select_affinity_usr(irq) ? -EINVAL : count; irq_set_affinity(irq, new_value); diff --git a/kernel/softirq.c b/kernel/softirq.c index 7110dae..e7c69a7 100644 --- a/kernel/softirq.c +++ b/kernel/softirq.c @@ -269,10 +269,11 @@ void irq_enter(void) { int cpu = smp_processor_id(); - if (idle_cpu(cpu) && !in_interrupt()) + if (idle_cpu(cpu) && !in_interrupt()) { + __irq_enter(); tick_check_idle(cpu); - - __irq_enter(); + } else + __irq_enter(); } #ifdef __ARCH_IRQ_EXIT_IRQS_DISABLED diff --git a/kernel/time/tick-broadcast.c b/kernel/time/tick-broadcast.c index f98a1b7..ffb7252 100644 --- a/kernel/time/tick-broadcast.c +++ b/kernel/time/tick-broadcast.c @@ -397,6 +397,33 @@ void tick_check_oneshot_broadcast(int cpu) } /* + * Called from tick_nohz_kick_tick() to check, whether the next + * broadcast event is less than a tick_period away or already pending + * to avoid reprogramming of the per cpu device. + */ +int tick_check_oneshot_broadcast_wakeup(int cpu, ktime_t now) +{ + struct clock_event_device *bc; + ktime_t delta; + int res = 0; + + spin_lock(&tick_broadcast_lock); + bc = tick_broadcast_device.evtdev; + + if (bc) { + delta = ktime_sub(bc->next_event, now); + if (delta.tv64 <= tick_period.tv64) + res = 1; + else + cpu_clear(cpu, tick_broadcast_oneshot_mask); + } else + cpu_clear(cpu, tick_broadcast_oneshot_mask); + + spin_unlock(&tick_broadcast_lock); + return res; +} + +/* * Handle oneshot mode broadcasting */ static void tick_handle_oneshot_broadcast(struct clock_event_device *dev) diff --git a/kernel/time/tick-internal.h b/kernel/time/tick-internal.h index b1c05bf..b825c39 100644 --- a/kernel/time/tick-internal.h +++ b/kernel/time/tick-internal.h @@ -37,6 +37,7 @@ extern void tick_shutdown_broadcast_oneshot(unsigned int *cpup); extern int tick_resume_broadcast_oneshot(struct clock_event_device *bc); extern int tick_broadcast_oneshot_active(void); extern void tick_check_oneshot_broadcast(int cpu); +extern int tick_check_oneshot_broadcast_wakeup(int cpu, ktime_t now); # else /* BROADCAST */ static inline void tick_broadcast_setup_oneshot(struct clock_event_device *bc) { @@ -47,6 +48,10 @@ static inline void tick_broadcast_switch_to_oneshot(void) { } static inline void tick_shutdown_broadcast_oneshot(unsigned int *cpup) { } static inline int tick_broadcast_oneshot_active(void) { return 0; } static inline void tick_check_oneshot_broadcast(int cpu) { } +static inline int tick_check_oneshot_broadcast_wakeup(int cpu, ktime_t now) +{ + return 0; +} # endif /* !BROADCAST */ #else /* !ONESHOT */ diff --git a/kernel/time/tick-sched.c b/kernel/time/tick-sched.c index 5bbb104..35a775e 100644 --- a/kernel/time/tick-sched.c +++ b/kernel/time/tick-sched.c @@ -134,48 +134,31 @@ __setup("nohz=", setup_tick_nohz); * value. We do this unconditionally on any cpu, as we don't know whether the * cpu, which has the update task assigned is in a long sleep. */ -void tick_nohz_update_jiffies(void) +static void tick_nohz_update_jiffies(int cpu, struct tick_sched *ts, + ktime_t now) { - int cpu = smp_processor_id(); - struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); - unsigned long flags; - ktime_t now; - - if (!ts->tick_stopped) - return; - cpu_clear(cpu, nohz_cpu_mask); - now = ktime_get(); ts->idle_waketime = now; - local_irq_save(flags); tick_do_update_jiffies64(now); - local_irq_restore(flags); touch_softlockup_watchdog(); } -static void tick_nohz_stop_idle(int cpu) +static void tick_nohz_stop_idle(struct tick_sched *ts, ktime_t now) { - struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); - - if (ts->idle_active) { - ktime_t now, delta; - now = ktime_get(); - delta = ktime_sub(now, ts->idle_entrytime); - ts->idle_lastupdate = now; - ts->idle_sleeptime = ktime_add(ts->idle_sleeptime, delta); - ts->idle_active = 0; + ktime_t delta = ktime_sub(now, ts->idle_entrytime); - sched_clock_idle_wakeup_event(0); - } + ts->idle_lastupdate = now; + ts->idle_sleeptime = ktime_add(ts->idle_sleeptime, delta); + ts->idle_active = 0; + sched_clock_idle_wakeup_event(0); } static ktime_t tick_nohz_start_idle(struct tick_sched *ts) { - ktime_t now, delta; + ktime_t delta, now = ktime_get(); - now = ktime_get(); if (ts->idle_active) { delta = ktime_sub(now, ts->idle_entrytime); ts->idle_lastupdate = now; @@ -203,6 +186,21 @@ u64 get_cpu_idle_time_us(int cpu, u64 *last_update_time) } EXPORT_SYMBOL_GPL(get_cpu_idle_time_us); +int check_idle_spurious_wakeup(int cpu) +{ + struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); + int ret; + + local_irq_disable(); + ret= ts->irq_last_wakeups == ts->irq_wakeups; + + ts->irq_last_wakeups = ts->irq_wakeups; + ts->spurious_wakeups += ret; + local_irq_enable(); + return ret; +} +EXPORT_SYMBOL_GPL(check_idle_spurious_wakeup); + /** * tick_nohz_stop_sched_tick - stop the idle tick from the idle task * @@ -413,10 +411,10 @@ void tick_nohz_restart_sched_tick(void) int cpu = smp_processor_id(); struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); unsigned long ticks; - ktime_t now; + ktime_t now = ktime_get(); local_irq_disable(); - tick_nohz_stop_idle(cpu); + tick_nohz_stop_idle(ts, now); if (!ts->inidle || !ts->tick_stopped) { ts->inidle = 0; @@ -430,7 +428,6 @@ void tick_nohz_restart_sched_tick(void) /* Update jiffies first */ select_nohz_load_balancer(0); - now = ktime_get(); tick_do_update_jiffies64(now); cpu_clear(cpu, nohz_cpu_mask); @@ -566,24 +563,20 @@ static void tick_nohz_switch_to_nohz(void) * timer and do not touch the other magic bits which need to be done * when idle is left. */ -static void tick_nohz_kick_tick(int cpu) +static void tick_nohz_kick_tick(int cpu, struct tick_sched *ts, ktime_t now) { - struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); - ktime_t delta, now; - - if (!ts->tick_stopped) - return; + ktime_t delta; /* * Do not touch the tick device, when the next expiry is either * already reached or less/equal than the tick period. */ - now = ktime_get(); delta = ktime_sub(hrtimer_get_expires(&ts->sched_timer), now); if (delta.tv64 <= tick_period.tv64) return; - tick_nohz_restart(ts, now); + if (!tick_check_oneshot_broadcast_wakeup(cpu, now)) + tick_nohz_restart(ts, now); } #else @@ -597,11 +590,26 @@ static inline void tick_nohz_switch_to_nohz(void) { } */ void tick_check_idle(int cpu) { + struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); + ktime_t now; + + ts->irq_wakeups++; + tick_check_oneshot_broadcast(cpu); + #ifdef CONFIG_NO_HZ - tick_nohz_stop_idle(cpu); - tick_nohz_update_jiffies(); - tick_nohz_kick_tick(cpu); + if (!ts->tick_stopped && !ts->idle_active) + return; + + now = ktime_get(); + + if (ts->idle_active) + tick_nohz_stop_idle(ts, now); + + if (ts->tick_stopped) { + tick_nohz_update_jiffies(cpu, ts, now); + tick_nohz_kick_tick(cpu, ts, now); + } #endif } diff --git a/kernel/time/timer_list.c b/kernel/time/timer_list.c index a999b92..67c3d60 100644 --- a/kernel/time/timer_list.c +++ b/kernel/time/timer_list.c @@ -176,6 +176,8 @@ static void print_cpu(struct seq_file *m, int cpu, u64 now) P(last_jiffies); P(next_jiffies); P_ns(idle_expires); + P(irq_wakeups); + P(spurious_wakeups); SEQ_printf(m, "jiffies: %Lu\n", (unsigned long long)jiffies); } @@ -252,7 +254,7 @@ static int timer_list_show(struct seq_file *m, void *v) u64 now = ktime_to_ns(ktime_get()); int cpu; - SEQ_printf(m, "Timer List Version: v0.4\n"); + SEQ_printf(m, "Timer List Version: v0.5\n"); SEQ_printf(m, "HRTIMER_MAX_CLOCK_BASES: %d\n", HRTIMER_MAX_CLOCK_BASES); SEQ_printf(m, "now at %Ld nsecs\n", (unsigned long long)now); -- Arjan van de Ven Intel Open Source Technology Centre For development, discussion and tips for power savings, visit http://www.lesswatts.org -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/