Received: by 2002:a05:6358:11c7:b0:104:8066:f915 with SMTP id i7csp1828186rwl; Thu, 30 Mar 2023 01:59:18 -0700 (PDT) X-Google-Smtp-Source: AKy350Y/2J9SPGEjB0SVDOD5JPDS9u1IgBSryOJ6PU7B7efLMo1DCi8dBCTIBmI7enUIRuIC8vQr X-Received: by 2002:aa7:8d0f:0:b0:626:62f:38d8 with SMTP id j15-20020aa78d0f000000b00626062f38d8mr1547861pfe.3.1680166758405; Thu, 30 Mar 2023 01:59:18 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1680166758; cv=none; d=google.com; s=arc-20160816; b=XyM2TyWqk7rjEWN9+lwD1HzGhBX7aGmLoWOJqdxnOB6mp+FQPoVJpuPFlAJLrsDtZZ KjwLM6wOx3cQlSfpChTrIdL/3H+s2XYeOXmEDX9jDBxaj1QUrXsEjKG6pyenTB9NYgH6 5y/5m9p71vZN6Zd8BpboyWdappWBRvxU11yylvtKjTXvaoX/YLDPRcfnvhvVcGp2aewh YJfDLQZlvmbz2mzmR0DrWm8uetaSXot2xrmGhfzfVWcyT/vqOZzI3xaCnMpFotDffCfl 1gq2xkfvP3F2wWa6V6JtDiKVM2QNTxCgxFLqQ7QVH+FRtChPGBuJym9VLHaM7F+R5mA5 dp8A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject; bh=XiIFc/kJxGUgDO+jUaMUhtl9YzCpioTA0ikTFaf5d+8=; b=UfJMwtIuSmhBj33awcHPvMU1pfvO9XPre/mEkfAPUSXYOZ2b1p8jMynYs9rLV94aIT z23YfM56n/J/JHApXMbNSxhXPjbv/7bintcKvjrlLgUxmZnEgGW8jH1bUPBbJYSNfZhR io5gjwmE4vv6rg4Ly0yx1HFkIIlDVo7FY5l/4Sz1oi+VcOS1d54WhbJby+GFYu1x6ztO BMkSNgn0Rp0NqXSnl378Uaa0AQVg9ysryFIIwIA2Me7dQmaWSBWfdKuudUasY9/wnzB3 V/1srsqKZkoZ0N3fV1lSCsBK3PXrEkX6o47tFApxK7bsCEJBpX8FlJ63wiST540IscMT JLUg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id u28-20020a056a00099c00b005a7ea9c6d7csi35496497pfg.41.2023.03.30.01.59.06; Thu, 30 Mar 2023 01:59:18 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229694AbjC3IxU (ORCPT + 99 others); Thu, 30 Mar 2023 04:53:20 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54142 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229624AbjC3IxP (ORCPT ); Thu, 30 Mar 2023 04:53:15 -0400 Received: from loongson.cn (mail.loongson.cn [114.242.206.163]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id CD9BB7286; Thu, 30 Mar 2023 01:53:09 -0700 (PDT) Received: from loongson.cn (unknown [113.200.148.30]) by gateway (Coremail) with SMTP id _____8AxJ_D0TSVkcE8UAA--.31210S3; Thu, 30 Mar 2023 16:53:08 +0800 (CST) Received: from [10.130.0.102] (unknown [113.200.148.30]) by localhost.localdomain (Coremail) with SMTP id AQAAf8Axnr7wTSVkvhMRAA--.43030S3; Thu, 30 Mar 2023 16:53:06 +0800 (CST) Subject: Re: [PATCH] LoongArch: Add kernel address sanitizer support To: Youling Tang Cc: Andrey Ryabinin , Jonathan Corbet , Huacai Chen , Andrew Morton , Alexander Potapenko , Andrey Konovalov , Dmitry Vyukov , Vincenzo Frascino , WANG Xuerui , Jiaxun Yang , kasan-dev@googlegroups.com, linux-doc@vger.kernel.org, linux-mm@kvack.org, loongarch@lists.linux.dev, linux-kernel@vger.kernel.org, linux-hardening@vger.kernel.org References: <20230328111714.2056-1-zhangqing@loongson.cn> <4ad7dfe6-160a-d4a8-e262-1fb13a395510@loongson.cn> From: Qing Zhang Message-ID: <72bc516f-0ee7-fbf1-7814-8501335b4246@loongson.cn> Date: Thu, 30 Mar 2023 16:53:03 +0800 User-Agent: Mozilla/5.0 (X11; Linux mips64; rv:68.0) Gecko/20100101 Thunderbird/68.7.0 MIME-Version: 1.0 In-Reply-To: <4ad7dfe6-160a-d4a8-e262-1fb13a395510@loongson.cn> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 8bit X-CM-TRANSID: AQAAf8Axnr7wTSVkvhMRAA--.43030S3 X-CM-SenderInfo: x2kd0wptlqwqxorr0wxvrqhubq/ X-Coremail-Antispam: 1Uk129KBjvAXoW3tF13JFyDCFW3Xr45Xw4rGrg_yoW8XFykJo W5KF13tr4rJw47Krs8Xw1DJry5Jr1UCrs7A3y7WryxJF1xAF15C3yUtrWaq3y3JrykGr13 G3yUGryrAFy8Zrn8n29KB7ZKAUJUUUUk529EdanIXcx71UUUUU7KY7ZEXasCq-sGcSsGvf J3Ic02F40EFcxC0VAKzVAqx4xG6I80ebIjqfuFe4nvWSU5nxnvy29KBjDU0xBIdaVrnRJU UUPIb4IE77IF4wAFF20E14v26r1j6r4UM7CY07I20VC2zVCF04k26cxKx2IYs7xG6rWj6s 0DM7CIcVAFz4kK6r1Y6r17M28lY4IEw2IIxxk0rwA2F7IY1VAKz4vEj48ve4kI8wA2z4x0 Y4vE2Ix0cI8IcVAFwI0_Gr0_Xr1l84ACjcxK6xIIjxv20xvEc7CjxVAFwI0_Gr0_Cr1l84 ACjcxK6I8E87Iv67AKxVW8Jr0_Cr1UM28EF7xvwVC2z280aVCY1x0267AKxVWxJr0_GcWl n4kS14v26r1q6r43M2AIxVAIcxkEcVAq07x20xvEncxIr21l57IF6xkI12xvs2x26I8E6x ACxx1l5I8CrVACY4xI64kE6c02F40Ex7xfMcIj6xIIjxv20xvE14v26r1q6rW5McIj6I8E 87Iv67AKxVW8JVWxJwAm72CE4IkC6x0Yz7v_Jr0_Gr1lF7xvr2IY64vIr41lc7I2V7IY0V AS07AlzVAYIcxG8wCY1x0262kKe7AKxVWUtVW8ZwCF04k20xvY0x0EwIxGrwCFx2IqxVCF s4IE7xkEbVWUJVW8JwCFI7km07C267AKxVWUtVW8ZwC20s026c02F40E14v26r1j6r18MI 8I3I0E7480Y4vE14v26r106r1rMI8E67AF67kF1VAFwI0_GFv_WrylIxkGc2Ij64vIr41l IxAIcVC0I7IYx2IY67AKxVWUCVW8JwCI42IY6xIIjxv20xvEc7CjxVAFwI0_Gr0_Cr1lIx AIcVCF04k26cxKx2IYs7xG6r1j6r1xMIIF0xvEx4A2jsIE14v26r4j6F4UMIIF0xvEx4A2 jsIEc7CjxVAFwI0_Gr0_Gr1UYxBIdaVFxhVjvjDU0xZFpf9x07jTq2NUUUUU= X-Spam-Status: No, score=-0.0 required=5.0 tests=NICE_REPLY_A,SPF_HELO_PASS, SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, Youling On 2023/3/30 下午2:50, Youling Tang wrote: > Hi, Qing > > On 03/28/2023 07:17 PM, Qing Zhang wrote: >> 1/8 of kernel addresses reserved for shadow memory. But for LoongArch, >> There are a lot of holes between different segments and valid address >> space(256T available) is insufficient to map all these segments to kasan >> shadow memory with the common formula provided by kasan core, saying >> addr >> KASAN_SHADOW_SCALE_SHIFT) + KASAN_SHADOW_OFFSET >> >> So Loongarch has a ARCH specific mapping formula,different segments >> are mapped individually, and only limited length of space of that >> specific segment is mapped to shadow. >> >> At early boot stage the whole shadow region populated with just >> one physical page (kasan_early_shadow_page). Later, this page is >> reused as readonly zero shadow for some memory that Kasan currently >> don't track. >> After mapping the physical memory, pages for shadow memory are >> allocated and mapped. >> >> Functions like memset/memmove/memcpy do a lot of memory accesses. >> If bad pointer passed to one of these function it is important >> to catch this. Compiler's instrumentation cannot do this since >> these functions are written in assembly. >> KASan replaces memory functions with manually instrumented variants. >> Original functions declared as weak symbols so strong definitions >> in mm/kasan/kasan.c could replace them. Original functions have aliases >> with '__' prefix in name, so we could call non-instrumented variant >> if needed. >> >> Signed-off-by: Qing Zhang >> --- >>  Documentation/dev-tools/kasan.rst             |   4 +- >>  .../features/debug/KASAN/arch-support.txt     |   2 +- >>  arch/loongarch/Kconfig                        |   7 + >>  arch/loongarch/include/asm/kasan.h            | 120 +++++++++ >>  arch/loongarch/include/asm/pgtable.h          |   7 + >>  arch/loongarch/include/asm/setup.h            |   2 +- >>  arch/loongarch/include/asm/string.h           |  20 ++ >>  arch/loongarch/kernel/Makefile                |   3 + >>  arch/loongarch/kernel/head.S                  |  14 +- >>  arch/loongarch/kernel/relocate.c              |   8 +- >>  arch/loongarch/kernel/setup.c                 |   4 + >>  arch/loongarch/lib/memcpy.S                   |   4 +- >>  arch/loongarch/lib/memmove.S                  |  13 +- >>  arch/loongarch/lib/memset.S                   |   4 +- >>  arch/loongarch/mm/Makefile                    |   2 + >>  arch/loongarch/mm/kasan_init.c                | 255 ++++++++++++++++++ >>  arch/loongarch/vdso/Makefile                  |   4 + >>  include/linux/kasan.h                         |   2 + >>  mm/kasan/generic.c                            |   5 + >>  mm/kasan/init.c                               |  10 +- >>  mm/kasan/kasan.h                              |   6 + >>  21 files changed, 470 insertions(+), 26 deletions(-) >>  create mode 100644 arch/loongarch/include/asm/kasan.h >>  create mode 100644 arch/loongarch/mm/kasan_init.c >> >> diff --git a/Documentation/dev-tools/kasan.rst >> b/Documentation/dev-tools/kasan.rst >> index e66916a483cd..ee91f2872767 100644 >> --- a/Documentation/dev-tools/kasan.rst >> +++ b/Documentation/dev-tools/kasan.rst >> @@ -41,8 +41,8 @@ Support >>  Architectures >>  ~~~~~~~~~~~~~ >> >> -Generic KASAN is supported on x86_64, arm, arm64, powerpc, riscv, >> s390, and >> -xtensa, and the tag-based KASAN modes are supported only on arm64. >> +Generic KASAN is supported on x86_64, arm, arm64, powerpc, riscv, >> s390, xtensa, >> +and loongarch, and the tag-based KASAN modes are supported only on >> arm64. >> >>  Compilers >>  ~~~~~~~~~ >> diff --git a/Documentation/features/debug/KASAN/arch-support.txt >> b/Documentation/features/debug/KASAN/arch-support.txt >> index bf0124fae643..c4581c2edb28 100644 >> --- a/Documentation/features/debug/KASAN/arch-support.txt >> +++ b/Documentation/features/debug/KASAN/arch-support.txt >> @@ -13,7 +13,7 @@ >>      |        csky: | TODO | >>      |     hexagon: | TODO | >>      |        ia64: | TODO | >> -    |   loongarch: | TODO | >> +    |   loongarch: |  ok  | >>      |        m68k: | TODO | >>      |  microblaze: | TODO | >>      |        mips: | TODO | >> diff --git a/arch/loongarch/Kconfig b/arch/loongarch/Kconfig >> index 72dd00f48b8c..61f883c51045 100644 >> --- a/arch/loongarch/Kconfig >> +++ b/arch/loongarch/Kconfig >> @@ -7,6 +7,7 @@ config LOONGARCH >>      select ACPI_MCFG if ACPI >>      select ACPI_SYSTEM_POWER_STATES_SUPPORT    if ACPI >>      select ARCH_BINFMT_ELF_STATE >> +    select ARCH_DISABLE_KASAN_INLINE >>      select ARCH_ENABLE_MEMORY_HOTPLUG >>      select ARCH_ENABLE_MEMORY_HOTREMOVE >>      select ARCH_HAS_ACPI_TABLE_UPGRADE    if ACPI >> @@ -83,6 +84,7 @@ config LOONGARCH >>      select HAVE_ARCH_AUDITSYSCALL >>      select HAVE_ARCH_MMAP_RND_BITS if MMU >>      select HAVE_ARCH_SECCOMP_FILTER >> +    select HAVE_ARCH_KASAN if 64BIT >>      select HAVE_ARCH_TRACEHOOK >>      select HAVE_ARCH_TRANSPARENT_HUGEPAGE >>      select HAVE_ASM_MODVERSIONS >> @@ -626,6 +628,11 @@ config ARCH_MMAP_RND_BITS_MIN >>  config ARCH_MMAP_RND_BITS_MAX >>      default 18 >> >> +config KASAN_SHADOW_OFFSET >> +    hex >> +    default 0x0 >> +    depends on KASAN >> + >>  menu "Power management options" >> >>  config ARCH_SUSPEND_POSSIBLE >> diff --git a/arch/loongarch/include/asm/kasan.h >> b/arch/loongarch/include/asm/kasan.h >> new file mode 100644 >> index 000000000000..582bcded311e >> --- /dev/null >> +++ b/arch/loongarch/include/asm/kasan.h >> @@ -0,0 +1,120 @@ >> +/* SPDX-License-Identifier: GPL-2.0 */ >> +#ifndef __ASM_KASAN_H >> +#define __ASM_KASAN_H >> + >> +#ifndef __ASSEMBLY__ >> + >> +#include >> +#include >> +#include >> +#include >> +#include >> + >> +#define __HAVE_ARCH_SHADOW_MAP >> + >> +#define KASAN_SHADOW_SCALE_SHIFT 3 >> +#define KASAN_SHADOW_OFFSET    _AC(CONFIG_KASAN_SHADOW_OFFSET, UL) >> + >> +#define XRANGE_SHIFT (48) >> + >> +/* Valid address length */ >> +#define XRANGE_SHADOW_SHIFT    (PGDIR_SHIFT + PAGE_SHIFT - 3) >> +/* Used for taking out the valid address */ >> +#define XRANGE_SHADOW_MASK    GENMASK_ULL(XRANGE_SHADOW_SHIFT - 1, 0) >> +/* One segment whole address space size */ >> +#define    XRANGE_SIZE        (XRANGE_SHADOW_MASK + 1) >> + >> +/* 64-bit segment value. */ >> +#define XKPRANGE_UC_SEG        (0x8000) >> +#define XKPRANGE_CC_SEG        (0x9000) >> +#define XKVRANGE_VC_SEG        (0xffff) >> + >> +/* Cached */ >> +#define XKPRANGE_CC_START        CACHE_BASE >> +#define XKPRANGE_CC_SIZE        XRANGE_SIZE >> +#define XKPRANGE_CC_KASAN_OFFSET    (0) >> +#define XKPRANGE_CC_SHADOW_SIZE        (XKPRANGE_CC_SIZE >> >> KASAN_SHADOW_SCALE_SHIFT) >> +#define XKPRANGE_CC_SHADOW_END        (XKPRANGE_CC_KASAN_OFFSET + >> XKPRANGE_CC_SHADOW_SIZE) >> + >> +/* UnCached */ >> +#define XKPRANGE_UC_START        UNCACHE_BASE >> +#define XKPRANGE_UC_SIZE        XRANGE_SIZE >> +#define XKPRANGE_UC_KASAN_OFFSET    XKPRANGE_CC_SHADOW_END >> +#define XKPRANGE_UC_SHADOW_SIZE        (XKPRANGE_UC_SIZE >> >> KASAN_SHADOW_SCALE_SHIFT) >> +#define XKPRANGE_UC_SHADOW_END        (XKPRANGE_UC_KASAN_OFFSET + >> XKPRANGE_UC_SHADOW_SIZE) >> + >> +/* VMALLOC (Cached or UnCached)  */ >> +#define XKVRANGE_VC_START        MODULES_VADDR >> +#define XKVRANGE_VC_SIZE        round_up(VMEMMAP_END - MODULES_VADDR >> + 1, PGDIR_SIZE) >> +#define XKVRANGE_VC_KASAN_OFFSET    XKPRANGE_UC_SHADOW_END >> +#define XKVRANGE_VC_SHADOW_SIZE        (XKVRANGE_VC_SIZE >> >> KASAN_SHADOW_SCALE_SHIFT) >> +#define XKVRANGE_VC_SHADOW_END        (XKVRANGE_VC_KASAN_OFFSET + >> XKVRANGE_VC_SHADOW_SIZE) >> + >> +/* Kasan shadow memory start right after vmalloc. */ >> +#define KASAN_SHADOW_START        round_up(VMEMMAP_END, PGDIR_SIZE) >> +#define KASAN_SHADOW_SIZE        (XKVRANGE_VC_SHADOW_END - >> XKPRANGE_CC_KASAN_OFFSET) >> +#define KASAN_SHADOW_END        round_up(KASAN_SHADOW_START + >> KASAN_SHADOW_SIZE, PGDIR_SIZE) >> + >> +#define XKPRANGE_CC_SHADOW_OFFSET    (KASAN_SHADOW_START + >> XKPRANGE_CC_KASAN_OFFSET) >> +#define XKPRANGE_UC_SHADOW_OFFSET    (KASAN_SHADOW_START + >> XKPRANGE_UC_KASAN_OFFSET) >> +#define XKVRANGE_VC_SHADOW_OFFSET    (KASAN_SHADOW_START + >> XKVRANGE_VC_KASAN_OFFSET) >> + >> +extern bool kasan_early_stage; >> +extern unsigned char kasan_early_shadow_page[PAGE_SIZE]; >> + >> +static inline void *kasan_mem_to_shadow(const void *addr) >> +{ >> +    if (kasan_early_stage) { >> +        return (void *)(kasan_early_shadow_page); >> +    } else { >> +        unsigned long maddr = (unsigned long)addr; >> +        unsigned long xrange = (maddr >> XRANGE_SHIFT) & 0xffff; >> +        unsigned long offset = 0; >> + >> +        maddr &= XRANGE_SHADOW_MASK; >> +        switch (xrange) { >> +        case XKPRANGE_CC_SEG: >> +            offset = XKPRANGE_CC_SHADOW_OFFSET; >> +            break; >> +        case XKPRANGE_UC_SEG: >> +            offset = XKPRANGE_UC_SHADOW_OFFSET; >> +            break; >> +        case XKVRANGE_VC_SEG: >> +            offset = XKVRANGE_VC_SHADOW_OFFSET; >> +            break; >> +        default: >> +            WARN_ON(1); >> +            return NULL; >> +        } >> + >> +        return (void *)((maddr >> KASAN_SHADOW_SCALE_SHIFT) + offset); >> +    } >> +} >> + >> +static inline const void *kasan_shadow_to_mem(const void *shadow_addr) >> +{ >> +    unsigned long addr = (unsigned long)shadow_addr; >> + >> +    if (unlikely(addr > KASAN_SHADOW_END) || >> +        unlikely(addr < KASAN_SHADOW_START)) { >> +        WARN_ON(1); >> +        return NULL; >> +    } >> + >> +    if (addr >= XKVRANGE_VC_SHADOW_OFFSET) >> +        return (void *)(((addr - XKVRANGE_VC_SHADOW_OFFSET) << >> KASAN_SHADOW_SCALE_SHIFT) + XKVRANGE_VC_START); >> +    else if (addr >= XKPRANGE_UC_SHADOW_OFFSET) >> +        return (void *)(((addr - XKPRANGE_UC_SHADOW_OFFSET) << >> KASAN_SHADOW_SCALE_SHIFT) + XKPRANGE_UC_START); >> +    else if (addr >= XKPRANGE_CC_SHADOW_OFFSET) >> +        return (void *)(((addr - XKPRANGE_CC_SHADOW_OFFSET) << >> KASAN_SHADOW_SCALE_SHIFT) + XKPRANGE_CC_START); >> +    else { >> +        WARN_ON(1); >> +        return NULL; >> +    } >> +} >> + >> +void kasan_init(void); >> +asmlinkage void kasan_early_init(void); >> + >> +#endif >> +#endif >> diff --git a/arch/loongarch/include/asm/pgtable.h >> b/arch/loongarch/include/asm/pgtable.h >> index d28fb9dbec59..5cfdf79b287e 100644 >> --- a/arch/loongarch/include/asm/pgtable.h >> +++ b/arch/loongarch/include/asm/pgtable.h >> @@ -86,9 +86,16 @@ extern unsigned long zero_page_mask; >>  #define MODULES_END    (MODULES_VADDR + SZ_256M) >> >>  #define VMALLOC_START    MODULES_END >> + >> +#ifndef CONFIG_KASAN >>  #define VMALLOC_END    \ >>      (vm_map_base +    \ >>       min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * >> PAGE_SIZE, (1UL << cpu_vabits)) - PMD_SIZE - VMEMMAP_SIZE) >> +#else >> +#define VMALLOC_END    \ >> +    (vm_map_base +    \ >> +     min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * >> PAGE_SIZE, (1UL << cpu_vabits) / 2) - PMD_SIZE - VMEMMAP_SIZE) >> +#endif >> >>  #define vmemmap        ((struct page *)((VMALLOC_END + PMD_SIZE) & >> PMD_MASK)) >>  #define VMEMMAP_END    ((unsigned long)vmemmap + VMEMMAP_SIZE - 1) >> diff --git a/arch/loongarch/include/asm/setup.h >> b/arch/loongarch/include/asm/setup.h >> index be05c0e706a2..2dca0d1dd90a 100644 >> --- a/arch/loongarch/include/asm/setup.h >> +++ b/arch/loongarch/include/asm/setup.h >> @@ -33,7 +33,7 @@ extern long __la_abs_end; >>  extern long __rela_dyn_begin; >>  extern long __rela_dyn_end; >> >> -extern void * __init relocate_kernel(void); >> +extern unsigned long __init relocate_kernel(void); >> >>  #endif >> >> diff --git a/arch/loongarch/include/asm/string.h >> b/arch/loongarch/include/asm/string.h >> index 7b29cc9c70aa..5bb5a90d2681 100644 >> --- a/arch/loongarch/include/asm/string.h >> +++ b/arch/loongarch/include/asm/string.h >> @@ -7,11 +7,31 @@ >> >>  #define __HAVE_ARCH_MEMSET >>  extern void *memset(void *__s, int __c, size_t __count); >> +extern void *__memset(void *__s, int __c, size_t __count); >> >>  #define __HAVE_ARCH_MEMCPY >>  extern void *memcpy(void *__to, __const__ void *__from, size_t __n); >> +extern void *__memcpy(void *__to, __const__ void *__from, size_t __n); >> >>  #define __HAVE_ARCH_MEMMOVE >>  extern void *memmove(void *__dest, __const__ void *__src, size_t __n); >> +extern void *__memmove(void *__dest, __const__ void *__src, size_t __n); >> + >> +#if defined(CONFIG_KASAN) && !defined(__SANITIZE_ADDRESS__) >> + >> +/* >> + * For files that are not instrumented (e.g. mm/slub.c) we >> + * should use not instrumented version of mem* functions. >> + */ >> + >> +#define memset(s, c, n) __memset(s, c, n) >> +#define memcpy(dst, src, len) __memcpy(dst, src, len) >> +#define memmove(dst, src, len) __memmove(dst, src, len) >> + >> +#ifndef __NO_FORTIFY >> +#define __NO_FORTIFY /* FORTIFY_SOURCE uses __builtin_memcpy, etc. */ > > If we need to add __NO_FORTIFY processing, please add > ARCH_HAS_FORTIFY_SOURCE in Kconfig, and submit a separate patch like > 9530141455c9 ("riscv: Add ARCH_HAS_FORTIFY_SOURCE"). Ok, I will add separate patch in v2. Thanks, -Qing > > Youling.