Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp5160524pxj; Wed, 26 May 2021 04:24:57 -0700 (PDT) X-Google-Smtp-Source: ABdhPJy+MMbGqPSO+gyG1cN+1JUdsFvuP466qhYW/4y25FO+t2snQqQz6UUb4UVMTcuL3Btze86i X-Received: by 2002:a92:b102:: with SMTP id t2mr26147876ilh.102.1622028296878; Wed, 26 May 2021 04:24:56 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1622028296; cv=none; d=google.com; s=arc-20160816; b=r596M2lc7gLLuaih1cBlzzKmoUn+FhtMp3SUeXfpVvlB7FkkUPW51UYhL/4dHFYD1O uiYBl5fPMWYnpvWZI1Ubjf276ufeRA9TOxkgkupnXE+PrmhPtuIJXV1OKd6kivtYd01r aB7Uw+N0Wk2fABXkYejiLxR92JnQYp7EP3TlDHopB1Z0j6LfKrDKpPXkIpamAWnF4/sD 6MQqm9yHfuAo4ghiNKnfQcPnsKXi9G8n+dg96eKb/WXrIIONdSRP8f+3AF//bi4+fsG3 JoDiYT+AS8N5r95CIF6caASJXkmmqljXNKrP7jKfUpHvrj4yktEu7+TvWEdFd2VYmBtG ggww== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:references:in-reply-to:message-id:date:subject :cc:to:from:dkim-signature; bh=2fg62J3tlvfkRTLHkvaQcfd0Wm5t5m69G8mg3YJgTeE=; b=vqAQZFHHF1P1kgHNZg9C8KhOqcOeYcmz/PLiZZXOcAeSImQ6RpCT6U3b8I8/4VxFou Cv76anKJ5/ELrxd3ievA67xQFqyu3k/gdn4b1fJdCmeIQAVAZHrZUMqpHogD/7JxTPc6 oy3l0ppnPAkMP5JgAtdBrBFFyJrXGPRVmUtmGffyWsvWpiy2Rsdh/pEkVaczBv5VXKSi 5yx8Lr3Jgzm+zR8UVmPk1WfFZ2JR+PtPfzaaXL2YRiW0vXAoCcJPfTxJ8LDStAFZwwde 4GHTAuA406dd07oLJTvds4FoBwPyVPPC7Kdu3IHwNCNqZacNkKiyzeDdaL4RBn5G9cXD OE+A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=CFOJR0R4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id o21si20450418jat.63.2021.05.26.04.24.42; Wed, 26 May 2021 04:24:56 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=CFOJR0R4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232372AbhEZFwD (ORCPT + 99 others); Wed, 26 May 2021 01:52:03 -0400 Received: from mail.kernel.org ([198.145.29.99]:47670 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232197AbhEZFvv (ORCPT ); Wed, 26 May 2021 01:51:51 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id 3E90961378; Wed, 26 May 2021 05:50:10 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1622008212; bh=oUqQsjWk/eKlS41GlTMA5rHbD9Z00zBqEiyeD2OkOOI=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=CFOJR0R4L8whZpMOkACO7IePTirWiWjitUw3GYvC12hzH9cpqTfWy1PeCSsP+7UVy GGjlNNMNSaqJBDMZVw11lBoucnJMQPWUNb5KSPdUb1eyDIkz6ymR2H2jlt0gdwioko G2QYj6X5zXTp8v0uxJSWltdfJsMyuTz+5AeHmI8z1HlFScuHkMjgLLISj/Kh9QBLF0 7kHkf9hCnGm5tSgsOgiSb/uYHpOSeXyS8KhxWdt4n3G4zj3hlRyIVyDqm8QXaPI4R7 QpWxZExXbAsOPEMiLgVTqYeV6Zb3SX1/hVl5a87ChGET+7cHURTKychDFYn5Nh+P84 y0N9il5PQeroA== From: guoren@kernel.org To: guoren@kernel.org, anup.patel@wdc.com, palmerdabbelt@google.com, arnd@arndb.de, hch@lst.de Cc: linux-riscv@lists.infradead.org, linux-kernel@vger.kernel.org, linux-arch@vger.kernel.org, linux-sunxi@lists.linux.dev, Guo Ren Subject: [PATCH V4 2/2] riscv: Use use_asid_allocator flush TLB Date: Wed, 26 May 2021 05:49:21 +0000 Message-Id: <1622008161-41451-3-git-send-email-guoren@kernel.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1622008161-41451-1-git-send-email-guoren@kernel.org> References: <1622008161-41451-1-git-send-email-guoren@kernel.org> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Guo Ren Use static_branch_unlikely(&use_asid_allocator) to keep the origin tlb flush style, so it's no effect on the existing machine. Here are the optimized functions: - flush_tlb_mm - flush_tlb_page - flush_tlb_range All above are based on the below new implement functions: - __sbi_tlb_flush_range_asid - local_flush_tlb_range_asid These functions are based on ASID instead of previous non-ASID tlb_flush implementation which invalidates more useful tlb entries. Signed-off-by: Guo Ren Reviewed-by: Anup Patel Cc: Palmer Dabbelt Cc: Christoph Hellwig --- arch/riscv/include/asm/mmu_context.h | 2 ++ arch/riscv/include/asm/tlbflush.h | 23 ++++++++++++++++++ arch/riscv/mm/context.c | 2 +- arch/riscv/mm/tlbflush.c | 46 +++++++++++++++++++++++++++++++++--- 4 files changed, 69 insertions(+), 4 deletions(-) diff --git a/arch/riscv/include/asm/mmu_context.h b/arch/riscv/include/asm/mmu_context.h index b065941..7030837 100644 --- a/arch/riscv/include/asm/mmu_context.h +++ b/arch/riscv/include/asm/mmu_context.h @@ -33,6 +33,8 @@ static inline int init_new_context(struct task_struct *tsk, return 0; } +DECLARE_STATIC_KEY_FALSE(use_asid_allocator); + #include #endif /* _ASM_RISCV_MMU_CONTEXT_H */ diff --git a/arch/riscv/include/asm/tlbflush.h b/arch/riscv/include/asm/tlbflush.h index c84218a..cee476b 100644 --- a/arch/riscv/include/asm/tlbflush.h +++ b/arch/riscv/include/asm/tlbflush.h @@ -22,9 +22,32 @@ static inline void local_flush_tlb_page(unsigned long addr) { ALT_FLUSH_TLB_PAGE(__asm__ __volatile__ ("sfence.vma %0" : : "r" (addr) : "memory")); } + +static inline void local_flush_tlb_all_asid(unsigned long asid) +{ + __asm__ __volatile__ ("sfence.vma x0, %0" + : + : "r" (asid) + : "memory"); +} + +static inline void local_flush_tlb_range_asid(unsigned long start, + unsigned long size, unsigned long asid) +{ + unsigned long tmp, end = ALIGN(start + size, PAGE_SIZE); + + for (tmp = start & PAGE_MASK; tmp < end; tmp += PAGE_SIZE) { + __asm__ __volatile__ ("sfence.vma %0, %1" + : + : "r" (tmp), "r" (asid) + : "memory"); + tmp += PAGE_SIZE; + } +} #else /* CONFIG_MMU */ #define local_flush_tlb_all() do { } while (0) #define local_flush_tlb_page(addr) do { } while (0) +#define local_flush_tlb_range_asid(addr) do { } while (0) #endif /* CONFIG_MMU */ #if defined(CONFIG_SMP) && defined(CONFIG_MMU) diff --git a/arch/riscv/mm/context.c b/arch/riscv/mm/context.c index 68aa312..45c1b04 100644 --- a/arch/riscv/mm/context.c +++ b/arch/riscv/mm/context.c @@ -18,7 +18,7 @@ #ifdef CONFIG_MMU -static DEFINE_STATIC_KEY_FALSE(use_asid_allocator); +DEFINE_STATIC_KEY_FALSE(use_asid_allocator); static unsigned long asid_bits; static unsigned long num_asids; diff --git a/arch/riscv/mm/tlbflush.c b/arch/riscv/mm/tlbflush.c index 720b443..87b4e52 100644 --- a/arch/riscv/mm/tlbflush.c +++ b/arch/riscv/mm/tlbflush.c @@ -4,6 +4,7 @@ #include #include #include +#include void flush_tlb_all(void) { @@ -39,18 +40,57 @@ static void __sbi_tlb_flush_range(struct cpumask *cmask, unsigned long start, put_cpu(); } +static void __sbi_tlb_flush_range_asid(struct cpumask *cmask, + unsigned long start, + unsigned long size, + unsigned long asid) +{ + struct cpumask hmask; + unsigned int cpuid; + + if (cpumask_empty(cmask)) + return; + + cpuid = get_cpu(); + + if (cpumask_any_but(cmask, cpuid) >= nr_cpu_ids) { + if (size == -1) + local_flush_tlb_all_asid(asid); + else + local_flush_tlb_range_asid(start, size, asid); + } else { + riscv_cpuid_to_hartid_mask(cmask, &hmask); + sbi_remote_sfence_vma_asid(cpumask_bits(&hmask), + start, size, asid); + } + + put_cpu(); +} + void flush_tlb_mm(struct mm_struct *mm) { - __sbi_tlb_flush_range(mm_cpumask(mm), 0, -1); + if (static_branch_unlikely(&use_asid_allocator)) + __sbi_tlb_flush_range_asid(mm_cpumask(mm), 0, -1, + atomic_long_read(&mm->context.id)); + else + __sbi_tlb_flush_range(mm_cpumask(mm), 0, -1); } void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) { - __sbi_tlb_flush_range(mm_cpumask(vma->vm_mm), addr, PAGE_SIZE); + if (static_branch_unlikely(&use_asid_allocator)) + __sbi_tlb_flush_range_asid(mm_cpumask(vma->vm_mm), addr, PAGE_SIZE, + atomic_long_read(&vma->vm_mm->context.id)); + else + __sbi_tlb_flush_range(mm_cpumask(vma->vm_mm), addr, PAGE_SIZE); } void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) { - __sbi_tlb_flush_range(mm_cpumask(vma->vm_mm), start, end - start); + if (static_branch_unlikely(&use_asid_allocator)) + __sbi_tlb_flush_range_asid(mm_cpumask(vma->vm_mm), start, end - start, + atomic_long_read(&vma->vm_mm->context.id)); + else + __sbi_tlb_flush_range(mm_cpumask(vma->vm_mm), start, end - start); } -- 2.7.4