Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 30FEEC636CC for ; Tue, 31 Jan 2023 07:26:53 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230324AbjAaH0w (ORCPT ); Tue, 31 Jan 2023 02:26:52 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:39660 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229546AbjAaH0t (ORCPT ); Tue, 31 Jan 2023 02:26:49 -0500 Received: from dfw.source.kernel.org (dfw.source.kernel.org [IPv6:2604:1380:4641:c500::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D2A233029A for ; Mon, 30 Jan 2023 23:26:47 -0800 (PST) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id 6E0E661207 for ; Tue, 31 Jan 2023 07:26:47 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id C83CEC433A4 for ; Tue, 31 Jan 2023 07:26:46 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1675150006; bh=wCPmrzAAMLanhCFIoUR3SD8iT+eCR4/vmdDSq3T4MWo=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=BjTurvHkaP9YJDmkwPu7Y192VbL9gmT2gdX2GVxs804Dvs4P5bNTZ63eNi70QYFID C0vQQG01LuJHb0j7jbU9AGsh1th7gH+czwaQ/XE4+LePx5EElMqFOF8R95dn6Juy9H SGsMuPameGi/uSSwTQUoF5xM28EXdUFiHl9YhQVErZeisKUF5AQykap2KStO7QjJv4 nmkro+MouLGpMEYswCfK2NxYzDrGcbtbw7dvQgEeAvPt6J5wNE4zfXHONFAEggnACm f/wTtlrD8zPVcaTK/HoBPzjnwSfYWpD/BtZBNwvyQJjxw1gqa2t/8f7T0Bt4usHNMy Twe8hh5sez2VQ== Received: by mail-ej1-f42.google.com with SMTP id dr8so17337338ejc.12 for ; Mon, 30 Jan 2023 23:26:46 -0800 (PST) X-Gm-Message-State: AO0yUKUnoNPpWH2f1yBaSsfk890mszWlBiCsDrznVKbxpQFg5zpu8tcl LIGSo9NYm3Be+/gb/oNVxMdcyO/MyzEglH8W2pk= X-Google-Smtp-Source: AMrXdXsso7ne/wLS4njluI5Aika1v0KFEnmYhN6kcTrUEQpq6O0IYxPSJ/hURkytJOBENnR3a9lsB6vk6CxcX9dKLXE= X-Received: by 2002:a17:906:770d:b0:877:e1ef:e49a with SMTP id q13-20020a170906770d00b00877e1efe49amr6222746ejm.147.1675150004885; Mon, 30 Jan 2023 23:26:44 -0800 (PST) MIME-Version: 1.0 References: <20230130232659.3374212-1-changbin.du@huawei.com> In-Reply-To: From: Guo Ren Date: Tue, 31 Jan 2023 15:26:33 +0800 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [PATCH v3] riscv: patch: Fixup lockdep warning in stop_machine To: Conor Dooley Cc: Changbin Du , Paul Walmsley , Palmer Dabbelt , Albert Ou , Hui Wang , linux-riscv@lists.infradead.org, linux-kernel@vger.kernel.org, Changbin Du , Zong Li Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Jan 30, 2023 at 11:10 PM Conor Dooley wrote: > > Hey Changbin, > > On Tue, Jan 31, 2023 at 07:26:59AM +0800, Changbin Du wrote: > > From: Changbin Du > > > > The task of ftrace_arch_code_modify(_post)_prepare() caller is > > stop_machine, whose caller and work thread are of different tasks. The > > lockdep checker needs the same task context, or it's wrong. That means > > it's a bug here to use lockdep_assert_held because we don't guarantee > > the same task context. > > > > kernel/locking/lockdep.c: > > int __lock_is_held(const struct lockdep_map *lock, int read) > > { > > struct task_struct *curr = current; > > int i; > > > > for (i = 0; i < curr->lockdep_depth; i++) { > > ^^^^^^^^^^^^^^^^^^^ > > struct held_lock *hlock = curr->held_locks + i; > > ^^^^^^^^^^^^^^^^ > > if (match_held_lock(hlock, lock)) { > > if (read == -1 || !!hlock->read == read) > > return LOCK_STATE_HELD; > > > > The __lock_is_held depends on current held_locks records; if > > stop_machine makes the checker runing on another task, that's wrong. > > > > Here is the log: > > [ 15.761523] ------------[ cut here ]------------ > > [ 15.762125] WARNING: CPU: 0 PID: 15 at arch/riscv/kernel/patch.c:63 patch_insn_write+0x72/0x364 > > [ 15.763258] Modules linked in: > > [ 15.764154] CPU: 0 PID: 15 Comm: migration/0 Not tainted 6.1.0-rc1-00014-g66924be85884-dirty #377 > > [ 15.765339] Hardware name: riscv-virtio,qemu (DT) > > [ 15.765985] Stopper: multi_cpu_stop+0x0/0x192 <- stop_cpus.constprop.0+0x90/0xe2 > > [ 15.766711] epc : patch_insn_write+0x72/0x364 > > [ 15.767011] ra : patch_insn_write+0x70/0x364 > > [ 15.767276] epc : ffffffff8000721e ra : ffffffff8000721c sp : ff2000000067bca0 > > [ 15.767622] gp : ffffffff81603f90 tp : ff60000002432a00 t0 : 7300000000000000 > > [ 15.767919] t1 : 0000000000000000 t2 : 73695f6b636f6c5f s0 : ff2000000067bcf0 > > [ 15.768238] s1 : 0000000000000008 a0 : 0000000000000000 a1 : 0000000000000000 > > [ 15.768537] a2 : 0000000000000000 a3 : 0000000000000000 a4 : 0000000000000000 > > [ 15.768837] a5 : 0000000000000000 a6 : 0000000000000000 a7 : 0000000000000000 > > [ 15.769139] s2 : ffffffff80009faa s3 : ff2000000067bd10 s4 : ffffffffffffffff > > [ 15.769447] s5 : 0000000000000001 s6 : 0000000000000001 s7 : 0000000000000003 > > [ 15.769740] s8 : 0000000000000002 s9 : 0000000000000004 s10: 0000000000000003 > > [ 15.770027] s11: 0000000000000002 t3 : 0000000000000000 t4 : ffffffff819af097 > > [ 15.770323] t5 : ffffffff819af098 t6 : ff2000000067ba28 > > [ 15.770574] status: 0000000200000100 badaddr: 0000000000000000 cause: 0000000000000003 > > [ 15.771102] [] patch_text_nosync+0x10/0x3a > > [ 15.771421] [] ftrace_update_ftrace_func+0x74/0x10a > > [ 15.771704] [] ftrace_modify_all_code+0xb0/0x16c > > [ 15.771958] [] __ftrace_modify_code+0x12/0x1c > > [ 15.772196] [] multi_cpu_stop+0x14a/0x192 > > [ 15.772454] [] cpu_stopper_thread+0x96/0x14c > > [ 15.772699] [] smpboot_thread_fn+0xf8/0x1cc > > [ 15.772945] [] kthread+0xe2/0xf8 > > [ 15.773160] [] ret_from_exception+0x0/0x14 > > [ 15.773471] ---[ end trace 0000000000000000 ]--- > > FWIW, you can always crop the [15.192321] stuff out of commit messages, > as it just adds noise. > > > By the way, this also fixes the same issue for patch_text(). > > > > Fixes: 0ff7c3b33127 ("riscv: Use text_mutex instead of patch_lock") > > Co-developed-by: Guo Ren > > Signed-off-by: Guo Ren > > Cc: Zong Li > > Cc: Palmer Dabbelt > > Signed-off-by: Changbin Du > > --- > > Changes in v3: > > - denote this also fixes function patch_text(). > > > > Changes in v2: > > - Rewrite commit log with lockdep explanation [Guo Ren] > > - Rebase on v6.1 [Guo Ren] > > > > v1: > > https://lore.kernel.org/linux-riscv/20210417023532.354714-1-changbin.du@gmail.com/ > > --- > > arch/riscv/kernel/patch.c | 7 ------- > > 1 file changed, 7 deletions(-) > > > > diff --git a/arch/riscv/kernel/patch.c b/arch/riscv/kernel/patch.c > > index 765004b60513..8619706f8dfd 100644 > > --- a/arch/riscv/kernel/patch.c > > +++ b/arch/riscv/kernel/patch.c > > @@ -55,13 +55,6 @@ static int patch_insn_write(void *addr, const void *insn, size_t len) > > bool across_pages = (((uintptr_t) addr & ~PAGE_MASK) + len) > PAGE_SIZE; > > int ret; > > > > - /* > > - * Before reaching here, it was expected to lock the text_mutex > > - * already, so we don't need to give another lock here and could > > - * ensure that it was safe between each cores. > > - */ > > - lockdep_assert_held(&text_mutex); > > I must admit, patches like this do concern me a little, as a someone > unfamiliar with the world of probing and tracing. > Seeing an explicit check that the lock was held, leads me to believe > that the original author (Zong Li I think) thought that the text_mutex > lock was insufficient. > Do you think that their fear is unfounded? Explaining why it is safe to > remove this assertion in the commit message would go a long way towards > easing my anxiety! > > Also, why delete the comment altogether? The comment provides some > information that doesn't appear to become invalid, even with the > assertion removed? Stop_machine separated the mutex context and made a lockdep warning. So text_mutex can't be used here. We need to find another check solution. I agree with the patch. > > Thanks, > Conor. > > > - > > if (across_pages) > > patch_map(addr + len, FIX_TEXT_POKE1); > > > > -- > > 2.25.1 > > > > -- Best Regards Guo Ren