Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751379AbWHCUyZ (ORCPT ); Thu, 3 Aug 2006 16:54:25 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1751376AbWHCUyZ (ORCPT ); Thu, 3 Aug 2006 16:54:25 -0400 Received: from ms-smtp-02.nyroc.rr.com ([24.24.2.56]:24992 "EHLO ms-smtp-02.nyroc.rr.com") by vger.kernel.org with ESMTP id S1751379AbWHCUyY (ORCPT ); Thu, 3 Aug 2006 16:54:24 -0400 Subject: Re: Problems with 2.6.17-rt8 From: Steven Rostedt To: Bill Huey Cc: Robert Crocombe , linux-kernel@vger.kernel.org, Ingo Molnar , Thomas Gleixner In-Reply-To: <20060803202211.GA10720@gnuppy.monkey.org> References: <1154541079.25723.8.camel@localhost.localdomain> <1154615261.32264.6.camel@localhost.localdomain> <20060803202211.GA10720@gnuppy.monkey.org> Content-Type: text/plain Date: Thu, 03 Aug 2006 16:54:05 -0400 Message-Id: <1154638445.4655.11.camel@localhost.localdomain> Mime-Version: 1.0 X-Mailer: Evolution 2.6.2 Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2869 Lines: 66 On Thu, 2006-08-03 at 13:22 -0700, Bill Huey wrote: > On Thu, Aug 03, 2006 at 10:27:41AM -0400, Steven Rostedt wrote: > > The rest was probably caused as a side effect from above. The above is > > already broken! > > > > You have NUMA configured too, so this is also something to look at. > > > > I still wouldn't ignore the first bug message you got: > > > > ---- > > BUG: scheduling while atomic: udev_run_devd/0x00000001/1568 > > > > Call Trace: > > {__schedule+155} > > {_raw_spin_unlock_irqrestore+53} > > {task_blocks_on_rt_mutex+518} > > {free_pages_bulk+39} > > {free_pages_bulk+39} > > ... > > ---- > > > > This could also have a side effect that messes things up. > > > > Unfortunately, right now I'm assigned to other tasks and I cant spend > > much more time on this at the moment. So hopefully, Ingo, Thomas or > > Bill, or someone else can help you find the reason for this problem. > > free_pages_bulk is definitely being called inside of an atomic. > I force this stack trace when the in_atomic() test is true at the > beginning of the function. > > > [ 29.362863] Call Trace: > [ 29.367107] {free_pages_bulk+86} > [ 29.373122] {_raw_spin_unlock_irqrestore+44} > [ 29.380233] {__free_pages_ok+428} > [ 29.386336] {free_hot_page+25} > [ 29.392165] {__free_pages+41} > [ 29.397898] {__free_pages_bootmem+174} > [ 29.404457] {free_all_bootmem_core+253} > [ 29.411112] {free_all_bootmem_node+9} > [ 29.417574] {numa_free_all_bootmem+61} > [ 29.424122] {_etext+0} > [ 29.429224] {mem_init+128} > [ 29.434691] {start_kernel+377} > [ 29.440520] {_sinittext+667} > [ 29.446669] --------------------------- > [ 29.450963] | preempt count: 00000001 ] > [ 29.455257] | 1-level deep critical section nesting: > [ 29.460732] ---------------------------------------- > [ 29.466212] .. [] .... start_kernel+0x68/0x221 > [ 29.472815] .....[] .. ( <= _sinittext+0x29b/0x2a2) > [ 29.480056] Perhaps you could put in that in_atomic check at the start of each of these functions and point to where it is a problem. Perhaps a spinlock is taken that was real and not a mutex. -- Steve - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/