Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id F0E91C433EF for ; Mon, 29 Nov 2021 05:02:49 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230358AbhK2FGF (ORCPT ); Mon, 29 Nov 2021 00:06:05 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:56498 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231349AbhK2FEE (ORCPT ); Mon, 29 Nov 2021 00:04:04 -0500 Received: from dfw.source.kernel.org (dfw.source.kernel.org [IPv6:2604:1380:4641:c500::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7DB89C061746 for ; Sun, 28 Nov 2021 21:00:47 -0800 (PST) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id 49B19611E0 for ; Mon, 29 Nov 2021 05:00:46 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id D7D8DC004E1; Mon, 29 Nov 2021 05:00:42 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1638162045; bh=7uBpe6jzG2pC4cmnqrWZHw15hdx3/0c/20sYugqj7HM=; h=Date:From:To:Cc:Subject:In-Reply-To:References:From; b=GSJsUgCz/kPJAvhy0HF6VnfPhJKo+/d5SbUGEcuSIE3kOcN2m/gFU9ghDxQX23FS8 Dk42TX0db7uOjkK1bul+7Ck6hhma+c/42SjdYgy4tSmBb9qqxyNzf1bepyHiEQaxR9 NnASZ5KXlcdaa5wEVvW0gI945QOZG+8QB2OGHcanoQdu1DoGJQbs/pDh+eLRxuy8A3 yIDIoF73cYKt1n6DGYF/OvRS6RXUXGirGXoiKegPRQtg7sZARgwEyQ9xHyCml3ueGx cLDMRGjjaXBySUP8Y1Fq78sQYU4JsAI33hX5+8LVxry5QvmpJd+LGSm1Vg092LLXVA 1A2i6c5udU33Q== Date: Mon, 29 Nov 2021 14:00:40 +0900 From: Masami Hiramatsu To: "liuqi (BA)" Cc: Mark Rutland , , , , , , , , , , , , Subject: Re: [PATCH v4 2/2] arm64: kprobe: Enable OPTPROBE for arm64 Message-Id: <20211129140040.87c5f423a72c95c90602c2c6@kernel.org> In-Reply-To: <4998f219-eb47-a07c-b3ed-c2ae46a77230@huawei.com> References: <20210818073336.59678-1-liuqi115@huawei.com> <20210818073336.59678-3-liuqi115@huawei.com> <20210824105001.GA96738@C02TD0UTHF1T.local> <20211127212302.f71345c34e5a62e5e779adb2@kernel.org> <4998f219-eb47-a07c-b3ed-c2ae46a77230@huawei.com> X-Mailer: Sylpheed 3.7.0 (GTK+ 2.24.32; x86_64-pc-linux-gnu) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, 29 Nov 2021 09:40:30 +0800 "liuqi (BA)" wrote: > > > On 2021/11/27 20:23, Masami Hiramatsu wrote: > > On Fri, 26 Nov 2021 18:31:06 +0800 > > "liuqi (BA)" wrote: > > > >> > >> > >> On 2021/8/24 18:50, Mark Rutland wrote: > >>>> diff --git a/arch/arm64/kernel/probes/optprobe_trampoline.S b/arch/arm64/kernel/probes/optprobe_trampoline.S > >>>> new file mode 100644 > >>>> index 000000000000..24d713d400cd > >>>> --- /dev/null > >>>> +++ b/arch/arm64/kernel/probes/optprobe_trampoline.S > >>>> @@ -0,0 +1,37 @@ > >>>> +/* SPDX-License-Identifier: GPL-2.0 */ > >>>> +/* > >>>> + * trampoline entry and return code for optprobes. > >>>> + */ > >>>> + > >>>> +#include > >>>> +#include > >>>> +#include > >>>> + > >>>> + .global optprobe_template_entry > >>>> +optprobe_template_entry: > >>> Please use SYM_*(); see arch/arm64/kernel/entry-ftrace.S for examples of > >>> how to use that for trampolines. > >>> > >>> This should be: > >>> > >>> SYM_CODE_START(optprobe_template) > >>> > >> Hi all, > >> > >> I meet a problem when I use SYM_CODE_START(optprobe_template) to replace > >> optprobe_template_entry. > >> > >> If SYM_CODE_START is used, all optprobe will share one trampoline space. > >> Under this circumstances, if user register two optprobes, trampoline > >> will be overwritten by the newer one, and this will cause kernel panic > >> when the old optprobe is trigger. > > > > Hm, this is curious, because the template should be copied to the > > trampoline buffer for each optprobe and be modified. > > > >> > >> Using optprobe_template_entry will not have this problem, as each > >> optprobe has its own trampoline space (alloced in get_opinsn_slot()). > > > > Yes, it is designed to do so. > > > > Thank you, > > > > Hi Masami, > > Thanks for your reply. But I also met a problem when using > get_opinsn_slot() to alloc trampoline buffer. > > As module_alloc(like x86) is used to alloc buffer, trampoline is in > module space, so if origin insn is in kernel space, the range between > origin insn and trampoline is out of 128M. > > As module PLT cannot used here, I have no idea to achieve long jump in > this situation. Do you have any good idea? One possible solution is to use pre-allocated trampoline space in the text area, as same as ppc64 does. (See arch/powerpc/kernel/optprobes_head.S, it embeds a space at "optinsn_slot") Also, the trampoline can be minimized, since what we need is the probed address (and the address of struct optprobe). A single trampoline entry will do the following; 1. push lr and a victim register (here, x0) 2. load the address of optprobe to x0 3. call(br) common-optprobe asm code 4. pop lr and x0 5. jump back to (next to) the original place Here the common-optprobe asm code does; c1. push all registers on the stack (like save_all_base_regs) for making struct pt_regs. c2. set the pt_regs address to x1. c3. call optimized_callback() c4. return Since arm64 will emulate the probed instruction, we can do this. (On the other hand, x86 needs to run the probed insn in trampoline code, it will do that between step 4 and 5) The trampoline entry code is just 5 instructions (but may need an immediate value (&optprobe) needs to be embedded). Thank you, > > Thanks, > Qi > > >> > >> So how to reuse SYM_CODE_START in this situation, does anyone has a > >> good idea? > >> > >> Thanks, > >> Qi > >>> ... and note the matching end below. > >>> > >>>> + sub sp, sp, #PT_REGS_SIZE > >>>> + save_all_base_regs > >>>> + /* Get parameters to optimized_callback() */ > >>>> + ldr x0, 1f > >>>> + mov x1, sp > >>>> + /* Branch to optimized_callback() */ > >>>> + .global optprobe_template_call > >>>> +optprobe_template_call: > >>> SYM_INNER_LABEL(optprobe_template_call, SYM_L_GLOBAL) > >>> > >>> ...and likewise for all the other labels. > >>> > >>>> + nop > >>>> + restore_all_base_regs > >>>> + ldr lr, [sp, #S_LR] > >>>> + add sp, sp, #PT_REGS_SIZE > >>>> + .global optprobe_template_restore_orig_insn > >>>> +optprobe_template_restore_orig_insn: > >>>> + nop > >>>> + .global optprobe_template_restore_end > >>>> +optprobe_template_restore_end: > >>>> + nop > >>>> + .global optprobe_template_end > >>>> +optprobe_template_end: > >>>> + .global optprobe_template_val > >>>> +optprobe_template_val: > >>>> + 1: .long 0 > >>>> + .long 0 > >>>> + .global optprobe_template_max_length > >>>> +optprobe_template_max_length: > >>> SYM_INNER_LABEL(optprobe_template_end, SYM_L_GLOBAL) > >>> SYM_CODE_END(optprobe_template) > >>> > >>> Thanks, > >>> Mark. > >>> > >>>> -- > > > > -- Masami Hiramatsu