Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp2556634pxk; Mon, 14 Sep 2020 17:24:24 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxGKUkLG1ZvEzgT/kyEwrG/r1HhvC7XWPTAWxQTh5C1i8tCIydSuufP2IXh4mVU4hEuUnEY X-Received: by 2002:a50:fd10:: with SMTP id i16mr20325652eds.54.1600129464329; Mon, 14 Sep 2020 17:24:24 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1600129464; cv=none; d=google.com; s=arc-20160816; b=McBgGkxeKaE4OnxaYmafBSUm3rAqAYsgBqsHRoe3Jnlf8CNQPX1rOaS6rP9Nd8r58a QaxTjc23tS0jurFlAZEF4WDra0HiOffgjgwp2bG6I2embij8whtRj7WHl71yN9nWkl+E yYQ9MJ04+J+rBlWh1ewSB09OvUq5nFDmdVZ11aoBYZrbTThkNNknhvaRbro+v8iRoiw2 wDnFPLj2O1+XhIH+pULVwlSMYP73KKGhDYT2KHsIDDpRGw1qRwKGar0zl+AxBgmDMTvx oZJtZ/buD54nVtmN4xDmS94J/g/9TyR5YwWBRhnMkWn5H8a5niCruQLCGsWNrgI3wYoS wbnQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:date:cc:to:from:subject:message-id :dkim-signature; bh=JIxbVL3SsLhU/KOArzoRLZaTcknM6wozcZMj8Pv2cYs=; b=Sx7OYpTpymETVpxmVnKionIzy76E2vpWEBkzlXZl5eUigRR55alxYs4jhE4alyFhkY IuoqH2jIJcgqzj8jgPVOtmGIY/o7BQB9Sf7PCOSFL7O/QSu5taeNLT3H5StxYEMsLecP rWLd3bEfau+RWmx/Q9s1GVvezJK2TuC5I11sPmUFCc1RQvKCL9XDlqjL0hLCljYmwF4e sGV7Nq29ozS+TYz4BLAOMxtBLD/wakeALmrdfBd7KjP75M4PrAhhjBdLka0wV9SoDzJI aJcwQ/mOXU1YC528J6NqffywgFEPz3gl1I2mUarge3HzsXecyjVuJ6TngsJElV0NIywP xDaA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=bc2XjpWN; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id d13si9540777edn.457.2020.09.14.17.24.01; Mon, 14 Sep 2020 17:24:24 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=bc2XjpWN; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726062AbgIOAVK (ORCPT + 99 others); Mon, 14 Sep 2020 20:21:10 -0400 Received: from us-smtp-1.mimecast.com ([205.139.110.61]:47145 "EHLO us-smtp-delivery-1.mimecast.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1725994AbgIOAVJ (ORCPT ); Mon, 14 Sep 2020 20:21:09 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1600129267; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=JIxbVL3SsLhU/KOArzoRLZaTcknM6wozcZMj8Pv2cYs=; b=bc2XjpWNOUEtUBBGZ67nVkxGFEBkKRXOMw7TKkF4jNltRUQOeAeTWdW7DfH/4h4GO+j/3Y xIwgmSSWZ82Y4Ss89xS3H1GV/vnlSMYxTnS1WNYZ0BQW3X/+WV9jcL/6I2Vhv8Lbbk1egf Y73iJ2I9iLmJzZQ0uBXzPhR1MD4z7Nw= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-92-QcTzheqtO-y6XW1X_6h32w-1; Mon, 14 Sep 2020 20:21:02 -0400 X-MC-Unique: QcTzheqtO-y6XW1X_6h32w-1 Received: from smtp.corp.redhat.com (int-mx08.intmail.prod.int.phx2.redhat.com [10.5.11.23]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id C9B1D1074642; Tue, 15 Sep 2020 00:21:00 +0000 (UTC) Received: from ovpn-113-249.rdu2.redhat.com (ovpn-113-249.rdu2.redhat.com [10.10.113.249]) by smtp.corp.redhat.com (Postfix) with ESMTP id D135727C2A; Tue, 15 Sep 2020 00:20:53 +0000 (UTC) Message-ID: <224bd11b533dd2acff3f6cce51ab4ca676eb4f9f.camel@redhat.com> Subject: Re: [PATCH v2 0/5] seqlock: Introduce PREEMPT_RT support From: Qian Cai To: "Ahmed S. Darwish" , Peter Zijlstra , Ingo Molnar , Will Deacon Cc: Thomas Gleixner , "Sebastian A. Siewior" , "Paul E. McKenney" , Steven Rostedt , LKML , Stephen Rothwell , linux-next@vger.kernel.org, Waiman Long , Boqun Feng Date: Mon, 14 Sep 2020 20:20:53 -0400 In-Reply-To: <20200904153231.11994-1-a.darwish@linutronix.de> References: <20200904153231.11994-1-a.darwish@linutronix.de> Content-Type: text/plain; charset="UTF-8" Mime-Version: 1.0 Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 2.84 on 10.5.11.23 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, 2020-09-04 at 17:32 +0200, Ahmed S. Darwish wrote: > Hi, > > Changelog-v2 > ============ > > - Standardize on seqcount_LOCKNAME_t as the canonical reference for > sequence counters with associated locks, instead of v1 > seqcount_LOCKTYPE_t. > > - Use unique prefix "seqprop_*" for all seqcount_t/seqcount_LOCKNAME_t > property accessors. > > - Touch-up the lock-unlock rationale for more clarity. Enforce writer > non-preemitiblity using "__seq_enforce_writer_non_preemptibility()". > > Cover letter (v1) > ================= > > https://lkml.kernel.org/r/20200828010710.5407-1-a.darwish@linutronix.de > > Preemption must be disabled before entering a sequence counter write > side critical section. Otherwise the read side section can preempt the > write side section and spin for the entire scheduler tick. If that > reader belongs to a real-time scheduling class, it can spin forever and > the kernel will livelock. > > Disabling preemption cannot be done for PREEMPT_RT though: it can lead > to higher latencies, and the write side sections will not be able to > acquire locks which become sleeping locks (e.g. spinlock_t). > > To remain preemptible, while avoiding a possible livelock caused by the > reader preempting the writer, use a different technique: let the reader > detect if a seqcount_LOCKNAME_t writer is in progress. If that's the > case, acquire then release the associated LOCKNAME writer serialization > lock. This will allow any possibly-preempted writer to make progress > until the end of its writer serialization lock critical section. > > Implement this lock-unlock technique for all seqcount_LOCKNAME_t with > an associated (PREEMPT_RT) sleeping lock, and for seqlock_t. Reverting this patchset [1] from today's linux-next fixed a splat below. The splat looks like a false positive anyway because the existing locking dependency chains from the task #1 here: &s->seqcount#2 ---> pidmap_lock [ 528.078061][ T7867] -> #1 (pidmap_lock){....}-{2:2}: [ 528.078078][ T7867] lock_acquire+0x10c/0x560 [ 528.078089][ T7867] _raw_spin_lock_irqsave+0x64/0xb0 [ 528.078108][ T7867] free_pid+0x5c/0x160 free_pid at kernel/pid.c:131 [ 528.078127][ T7867] release_task.part.40+0x59c/0x7f0 __unhash_process at kernel/exit.c:76 (inlined by) __exit_signal at kernel/exit.c:147 (inlined by) release_task at kernel/exit.c:198 [ 528.078145][ T7867] do_exit+0x77c/0xda0 exit_notify at kernel/exit.c:679 (inlined by) do_exit at kernel/exit.c:826 [ 528.078163][ T7867] kthread+0x148/0x1d0 [ 528.078182][ T7867] ret_from_kernel_thread+0x5c/0x80 It is write_seqlock(&sig->stats_lock) in __exit_signal(), but the &s->seqcount#2 in read_mems_allowed_begin() is read_seqcount_begin(¤t->mems_allowed_seq), so there should be no deadlock? [1] git revert --no-edit 0c9794c8b678..1909760f5fc3 [ 528.077900][ T7867] WARNING: possible circular locking dependency detected [ 528.077912][ T7867] 5.9.0-rc5-next-20200914 #1 Not tainted [ 528.077921][ T7867] ------------------------------------------------------ [ 528.077931][ T7867] runc:[1:CHILD]/7867 is trying to acquire lock: [ 528.077942][ T7867] c000001fce5570c8 (&s->seqcount#2){....}-{0:0}, at: __slab_alloc+0x34/0xf0 [ 528.077972][ T7867] [ 528.077972][ T7867] but task is already holding lock: [ 528.077983][ T7867] c0000000056b0198 (pidmap_lock){....}-{2:2}, at: alloc_pid+0x258/0x590 [ 528.078009][ T7867] [ 528.078009][ T7867] which lock already depends on the new lock. [ 528.078009][ T7867] [ 528.078031][ T7867] [ 528.078031][ T7867] the existing dependency chain (in reverse order) is: [ 528.078061][ T7867] [ 528.078061][ T7867] -> #1 (pidmap_lock){....}-{2:2}: [ 528.078078][ T7867] lock_acquire+0x10c/0x560 [ 528.078089][ T7867] _raw_spin_lock_irqsave+0x64/0xb0 [ 528.078108][ T7867] free_pid+0x5c/0x160 free_pid at kernel/pid.c:131 [ 528.078127][ T7867] release_task.part.40+0x59c/0x7f0 __unhash_process at kernel/exit.c:76 (inlined by) __exit_signal at kernel/exit.c:147 (inlined by) release_task at kernel/exit.c:198 [ 528.078145][ T7867] do_exit+0x77c/0xda0 exit_notify at kernel/exit.c:679 (inlined by) do_exit at kernel/exit.c:826 [ 528.078163][ T7867] kthread+0x148/0x1d0 [ 528.078182][ T7867] ret_from_kernel_thread+0x5c/0x80 [ 528.078208][ T7867] [ 528.078208][ T7867] -> #0 (&s->seqcount#2){....}-{0:0}: [ 528.078241][ T7867] check_prevs_add+0x1c4/0x1120 check_prev_add at kernel/locking/lockdep.c:2820 (inlined by) check_prevs_add at kernel/locking/lockdep.c:2944 [ 528.078260][ T7867] __lock_acquire+0x176c/0x1c00 validate_chain at kernel/locking/lockdep.c:3562 (inlined by) __lock_acquire at kernel/locking/lockdep.c:4796 [ 528.078278][ T7867] lock_acquire+0x10c/0x560 [ 528.078297][ T7867] ___slab_alloc+0xa40/0xb40 seqcount_lockdep_reader_access at include/linux/seqlock.h:103 (inlined by) read_mems_allowed_begin at include/linux/cpuset.h:135 (inlined by) get_any_partial at mm/slub.c:2035 (inlined by) get_partial at mm/slub.c:2078 (inlined by) new_slab_objects at mm/slub.c:2577 (inlined by) ___slab_alloc at mm/slub.c:2745 [ 528.078324][ T7867] __slab_alloc+0x34/0xf0 [ 528.078342][ T7867] kmem_cache_alloc+0x2d4/0x470 [ 528.078362][ T7867] create_object+0x74/0x430 [ 528.078381][ T7867] slab_post_alloc_hook+0xa4/0x670 [ 528.078399][ T7867] kmem_cache_alloc+0x1b4/0x470 [ 528.078418][ T7867] radix_tree_node_alloc.constprop.19+0xe4/0x160 [ 528.078438][ T7867] idr_get_free+0x298/0x360 [ 528.078456][ T7867] idr_alloc_u32+0x84/0x130 [ 528.078474][ T7867] idr_alloc_cyclic+0x7c/0x150 [ 528.078493][ T7867] alloc_pid+0x27c/0x590 [ 528.078511][ T7867] copy_process+0xc90/0x1930 copy_process at kernel/fork.c:2104 [ 528.078529][ T7867] kernel_clone+0x120/0xa10 [ 528.078546][ T7867] __do_sys_clone+0x88/0xd0 [ 528.078565][ T7867] system_call_exception+0xf8/0x1d0 [ 528.078592][ T7867] system_call_common+0xe8/0x218 [ 528.078609][ T7867] [ 528.078609][ T7867] other info that might help us debug this: [ 528.078609][ T7867] [ 528.078650][ T7867] Possible unsafe locking scenario: [ 528.078650][ T7867] [ 528.078670][ T7867] CPU0 CPU1 [ 528.078695][ T7867] ---- ---- [ 528.078713][ T7867] lock(pidmap_lock); [ 528.078730][ T7867] lock(&s->seqcount#2); [ 528.078751][ T7867] lock(pidmap_lock); [ 528.078770][ T7867] lock(&s->seqcount#2); [ 528.078788][ T7867] [ 528.078788][ T7867] *** DEADLOCK *** [ 528.078788][ T7867] [ 528.078800][ T7867] 2 locks held by runc:[1:CHILD]/7867: [ 528.078808][ T7867] #0: c000001ffea6f4f0 (lock#2){+.+.}-{2:2}, at: __radix_tree_preload+0x8/0x370 __radix_tree_preload at lib/radix-tree.c:322 [ 528.078844][ T7867] #1: c0000000056b0198 (pidmap_lock){....}-{2:2}, at: alloc_pid+0x258/0x590 [ 528.078870][ T7867] [ 528.078870][ T7867] stack backtrace: [ 528.078890][ T7867] CPU: 46 PID: 7867 Comm: runc:[1:CHILD] Not tainted 5.9.0-rc5-next-20200914 #1 [ 528.078921][ T7867] Call Trace: [ 528.078940][ T7867] [c000001ff07eefc0] [c00000000063f8c8] dump_stack+0xec/0x144 (unreliable) [ 528.078964][ T7867] [c000001ff07ef000] [c00000000013f44c] print_circular_bug.isra.43+0x2dc/0x350 [ 528.078978][ T7867] [c000001ff07ef0a0] [c00000000013f640] check_noncircular+0x180/0x1b0 [ 528.079000][ T7867] [c000001ff07ef170] [c000000000140b84] check_prevs_add+0x1c4/0x1120 [ 528.079022][ T7867] [c000001ff07ef280] [c0000000001446ec] __lock_acquire+0x176c/0x1c00 [ 528.079043][ T7867] [c000001ff07ef3a0] [c00000000014578c] lock_acquire+0x10c/0x560 [ 528.079066][ T7867] [c000001ff07ef490] [c0000000003565f0] ___slab_alloc+0xa40/0xb40 [ 528.079079][ T7867] [c000001ff07ef590] [c000000000356724] __slab_alloc+0x34/0xf0 [ 528.079100][ T7867] [c000001ff07ef5e0] [c000000000356ab4] kmem_cache_alloc+0x2d4/0x470 [ 528.079122][ T7867] [c000001ff07ef670] [c000000000397e14] create_object+0x74/0x430 [ 528.079144][ T7867] [c000001ff07ef720] [c000000000351944] slab_post_alloc_hook+0xa4/0x670 [ 528.079165][ T7867] [c000001ff07ef7e0] [c000000000356994] kmem_cache_alloc+0x1b4/0x470 [ 528.079187][ T7867] [c000001ff07ef870] [c00000000064e004] radix_tree_node_alloc.constprop.19+0xe4/0x160 radix_tree_node_alloc at lib/radix-tree.c:252 [ 528.079219][ T7867] [c000001ff07ef8e0] [c00000000064f2b8] idr_get_free+0x298/0x360 idr_get_free at lib/radix-tree.c:1507 [ 528.079249][ T7867] [c000001ff07ef970] [c000000000645db4] idr_alloc_u32+0x84/0x130 idr_alloc_u32 at lib/idr.c:46 (discriminator 4) [ 528.079271][ T7867] [c000001ff07ef9e0] [c000000000645f8c] idr_alloc_cyclic+0x7c/0x150 idr_alloc_cyclic at lib/idr.c:126 (discriminator 1) [ 528.079301][ T7867] [c000001ff07efa40] [c0000000000e48ac] alloc_pid+0x27c/0x590 [ 528.079342][ T7867] [c000001ff07efb20] [c0000000000acc60] copy_process+0xc90/0x1930 [ 528.079404][ T7867] [c000001ff07efc40] [c0000000000adc00] kernel_clone+0x120/0xa10 [ 528.079499][ T7867] [c000001ff07efd00] [c0000000000ae578] __do_sys_clone+0x88/0xd0 [ 528.079579][ T7867] [c000001ff07efdc0] [c000000000029c48] system_call_exception+0xf8/0x1d0 [ 528.079691][ T7867] [c000001ff07efe20] [c00000000000d0a8] system_call_common+0xe8/0x218 > > 8<-------------- > > Ahmed S. Darwish (5): > seqlock: seqcount_LOCKNAME_t: Standardize naming convention > seqlock: Use unique prefix for seqcount_t property accessors > seqlock: seqcount_t: Implement all read APIs as statement expressions > seqlock: seqcount_LOCKNAME_t: Introduce PREEMPT_RT support > seqlock: PREEMPT_RT: Do not starve seqlock_t writers > > include/linux/seqlock.h | 281 ++++++++++++++++++++++++---------------- > 1 file changed, 167 insertions(+), 114 deletions(-) > > base-commit: f75aef392f869018f78cfedf3c320a6b3fcfda6b > -- > 2.28.0