Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753918AbZJHFyw (ORCPT ); Thu, 8 Oct 2009 01:54:52 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1753290AbZJHFyw (ORCPT ); Thu, 8 Oct 2009 01:54:52 -0400 Received: from e28smtp06.in.ibm.com ([59.145.155.6]:47727 "EHLO e28smtp06.in.ibm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752305AbZJHFyu (ORCPT ); Thu, 8 Oct 2009 01:54:50 -0400 Date: Thu, 8 Oct 2009 11:24:09 +0530 From: Arun R Bharadwaj To: Peter Zijlstra Cc: Joel Schopp , Benjamin Herrenschmidt , Paul Mackerras , Ingo Molnar , Vaidyanathan Srinivasan , Dipankar Sarma , Balbir Singh , Gautham R Shenoy , Venkatesh Pallipadi , linux-kernel@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-arch@vger.kernel.org, Arun Bharadwaj Subject: Re: [v7 PATCH 3/7]: x86: refactor x86 idle power management code and remove all instances of pm_idle. Message-ID: <20091008055409.GA32387@linux.vnet.ibm.com> Reply-To: arun@linux.vnet.ibm.com References: <20091006152421.GA7278@linux.vnet.ibm.com> <20091006153126.GA7358@linux.vnet.ibm.com> <1254926750.26976.253.camel@twins> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline In-Reply-To: <1254926750.26976.253.camel@twins> User-Agent: Mutt/1.5.18 (2008-05-17) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 5988 Lines: 166 * Peter Zijlstra [2009-10-07 16:45:50]: > On Tue, 2009-10-06 at 21:01 +0530, Arun R Bharadwaj wrote: > > +++ linux.trees.git/arch/x86/kernel/process.c > > @@ -9,6 +9,7 @@ > > #include > > #include > > #include > > +#include > > #include > > #include > > #include > > @@ -244,12 +245,6 @@ int sys_vfork(struct pt_regs *regs) > > unsigned long boot_option_idle_override = 0; > > EXPORT_SYMBOL(boot_option_idle_override); > > > > -/* > > - * Powermanagement idle function, if any.. > > - */ > > -void (*pm_idle)(void); > > -EXPORT_SYMBOL(pm_idle); > > - > > #ifdef CONFIG_X86_32 > > /* > > * This halt magic was a workaround for ancient floppy DMA > > @@ -329,17 +324,15 @@ static void do_nothing(void *unused) > > } > > > > /* > > - * cpu_idle_wait - Used to ensure that all the CPUs discard old value of > > - * pm_idle and update to new pm_idle value. Required while changing pm_idle > > - * handler on SMP systems. > > + * cpu_idle_wait - Required while changing idle routine handler on SMP systems. > > * > > - * Caller must have changed pm_idle to the new value before the call. Old > > - * pm_idle value will not be used by any CPU after the return of this function. > > + * Caller must have changed idle routine to the new value before the call. Old > > + * value will not be used by any CPU after the return of this function. > > */ > > void cpu_idle_wait(void) > > { > > smp_mb(); > > - /* kick all the CPUs so that they exit out of pm_idle */ > > + /* kick all the CPUs so that they exit out of idle loop */ > > smp_call_function(do_nothing, NULL, 1); > > } > > EXPORT_SYMBOL_GPL(cpu_idle_wait); > > @@ -518,15 +511,59 @@ static void c1e_idle(void) > > default_idle(); > > } > > > > +static void (*local_idle)(void); > > +DEFINE_PER_CPU(struct cpuidle_device, idle_devices); > > + > > +struct cpuidle_driver cpuidle_default_driver = { > > + .name = "cpuidle_default", > > +}; > > + > > +static int local_idle_loop(struct cpuidle_device *dev, struct cpuidle_state *st) > > +{ > > + ktime_t t1, t2; > > + s64 diff; > > + int ret; > > + > > + t1 = ktime_get(); > > + local_idle(); > > + t2 = ktime_get(); > > + > > + diff = ktime_to_us(ktime_sub(t2, t1)); > > + if (diff > INT_MAX) > > + diff = INT_MAX; > > + ret = (int) diff; > > + > > + return ret; > > +} > > + > > +static int setup_cpuidle_simple(void) > > +{ > > + struct cpuidle_device *dev; > > + int cpu; > > + > > + if (!cpuidle_curr_driver) > > + cpuidle_register_driver(&cpuidle_default_driver); > > + > > + for_each_online_cpu(cpu) { > > + dev = &per_cpu(idle_devices, cpu); > > + dev->cpu = cpu; > > + dev->states[0].enter = local_idle_loop; > > + dev->state_count = 1; > > + cpuidle_register_device(dev); > > + } > > + return 0; > > +} > > +device_initcall(setup_cpuidle_simple); > > + > > void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c) > > { > > #ifdef CONFIG_SMP > > - if (pm_idle == poll_idle && smp_num_siblings > 1) { > > + if (local_idle == poll_idle && smp_num_siblings > 1) { > > printk(KERN_WARNING "WARNING: polling idle and HT enabled," > > " performance may degrade.\n"); > > } > > #endif > > - if (pm_idle) > > + if (local_idle) > > return; > > > > if (cpu_has(c, X86_FEATURE_MWAIT) && mwait_usable(c)) { > > @@ -534,18 +571,20 @@ void __cpuinit select_idle_routine(const > > * One CPU supports mwait => All CPUs supports mwait > > */ > > printk(KERN_INFO "using mwait in idle threads.\n"); > > - pm_idle = mwait_idle; > > + local_idle = mwait_idle; > > } else if (check_c1e_idle(c)) { > > printk(KERN_INFO "using C1E aware idle routine\n"); > > - pm_idle = c1e_idle; > > + local_idle = c1e_idle; > > } else > > - pm_idle = default_idle; > > + local_idle = default_idle; > > + > > + return; > > } > > > > void __init init_c1e_mask(void) > > { > > /* If we're using c1e_idle, we need to allocate c1e_mask. */ > > - if (pm_idle == c1e_idle) > > + if (local_idle == c1e_idle) > > zalloc_cpumask_var(&c1e_mask, GFP_KERNEL); > > } > > > > @@ -556,7 +595,7 @@ static int __init idle_setup(char *str) > > > > if (!strcmp(str, "poll")) { > > printk("using polling idle threads.\n"); > > - pm_idle = poll_idle; > > + local_idle = poll_idle; > > } else if (!strcmp(str, "mwait")) > > force_mwait = 1; > > else if (!strcmp(str, "halt")) { > > @@ -567,7 +606,7 @@ static int __init idle_setup(char *str) > > * To continue to load the CPU idle driver, don't touch > > * the boot_option_idle_override. > > */ > > - pm_idle = default_idle; > > + local_idle = default_idle; > > idle_halt = 1; > > return 0; > > } else if (!strcmp(str, "nomwait")) { > > > What guarantees that the cpuidle bits actually select this > cpuidle_default driver when you do idle=poll? > > Also, cpuidle already has a poll loop in it, why duplicate that? > Yes, now i see it.. I'll get rid of the redundant poll_idle definition -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/