Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757415AbYCNX6W (ORCPT ); Fri, 14 Mar 2008 19:58:22 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1751546AbYCNX6N (ORCPT ); Fri, 14 Mar 2008 19:58:13 -0400 Received: from ns2.g-housing.de ([81.169.133.75]:35133 "EHLO mail.g-house.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751291AbYCNX6M (ORCPT ); Fri, 14 Mar 2008 19:58:12 -0400 Date: Sat, 15 Mar 2008 00:58:02 +0100 (CET) From: Christian Kujau X-X-Sender: evil@sheep.housecafe.de To: Milan Broz cc: David Chinner , LKML , xfs@oss.sgi.com, dm-devel@redhat.com Subject: Re: INFO: task mount:11202 blocked for more than 120 seconds In-Reply-To: <47DA44EB.8000307@redhat.com> Message-ID: References: <20080307224040.GV155259@sgi.com> <20080309213441.GQ155407@sgi.com> <47DA44EB.8000307@redhat.com> User-Agent: Alpine 1.00 (DEB 882 2007-12-20) MIME-Version: 1.0 Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3820 Lines: 79 On Fri, 14 Mar 2008, Milan Broz wrote: > Yes, there is bug in dm-crypt... > Please try if the patch here helps: http://lkml.org/lkml/2008/3/14/71 Hm, it seems to help the hangs, yes. Applied to today's -git a few hours ago, the hangs are gone. However, when doing lots of disk I/O, the machine locks up after a few (10-20) minutes. Sadly, netconsole got nothing :( After the first lockup I tried again and shortly after bootup I got: [ 866.681441] [ INFO: possible circular locking dependency detected ] [ 866.681876] 2.6.25-rc5 #1 [ 866.682203] ------------------------------------------------------- [ 866.682637] kswapd0/132 is trying to acquire lock: [ 866.683028] (&(&ip->i_iolock)->mr_lock){----}, at: [] xfs_ilock+0x96/0xb0 [ 866.683916] [ 866.683917] but task is already holding lock: [ 866.684582] (iprune_mutex){--..}, at: [] shrink_icache_memory+0x72/0x220 [ 866.685461] [ 866.685462] which lock already depends on the new lock. [ 866.685463] [ 866.686440] [ 866.686441] the existing dependency chain (in reverse order) is: [ 866.687151] [ 866.687152] -> #1 (iprune_mutex){--..}: [ 866.687339] [] add_lock_to_list+0x44/0xc0 [ 866.687339] [] __lock_acquire+0xc26/0x10b0 [ 866.687339] [] shrink_icache_memory+0x72/0x220 [ 866.687339] [] __lock_acquire+0x18f/0x10b0 [ 866.687339] [] lock_acquire+0x5e/0x80 [ 866.687339] [] shrink_icache_memory+0x72/0x220 [ 866.687339] [] mutex_lock_nested+0x89/0x240 [ 866.687339] [] shrink_icache_memory+0x72/0x220 [ 866.687339] [] shrink_icache_memory+0x72/0x220 [ 866.687339] [] shrink_icache_memory+0x72/0x220 [ 866.687339] [] shrink_slab+0x21/0x160 [ 866.687340] [] shrink_slab+0x101/0x160 [ 866.687340] [] try_to_free_pages+0x152/0x230 [ 866.687340] [] isolate_pages_global+0x0/0x60 [ 866.687340] [] __alloc_pages+0x14b/0x370 [ 866.687340] [] _read_unlock_irq+0x20/0x30 [ 866.687340] [] __grab_cache_page+0x81/0xc0 [ 866.687340] [] block_write_begin+0x76/0xe0 [ 866.687340] [] xfs_vm_write_begin+0x46/0x50 [ 866.687340] [] xfs_get_blocks+0x0/0x30 [ 866.687340] [] generic_file_buffered_write+0x117/0x650 [ 866.687340] [] xfs_ilock+0x6d/0xb0 [ 866.687340] [] xfs_write+0x7ac/0x8a0 [ 866.687340] [] core_sys_select+0x21/0x350 [ 866.687340] [] xfs_file_aio_write+0x5c/0x70 [ 866.687340] [] do_sync_write+0xd5/0x120 [ 866.687340] [] autoremove_wake_function+0x0/0x40 [ 866.687340] [] dnotify_parent+0x35/0x90 [ 866.687340] [] do_sync_write+0x0/0x120 [ 866.687340] [] vfs_write+0x9f/0x140 [ 866.687340] [] sys_write+0x41/0x70 [ 866.687340] [] sysenter_past_esp+0x5f/0xa5 [ 866.687340] [] 0xffffffff [ 866.687340] [ 866.687340] -> #0 (&(&ip->i_iolock)->mr_lock){----}: [ 866.687340] [] print_circular_bug_entry+0x40/0x50 The box was running fine then for ~20 minutes, then it locked up again. Full dmesg and .config: http://nerdbynature.de/bits/2.6.25-rc5/ Right now I'm back to 2.6.24.3... Thanks, Christian. -- BOFH excuse #350: paradigm shift...without a clutch -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/