Received: by 2002:ab2:784b:0:b0:1fd:adc2:8405 with SMTP id m11csp214255lqp; Mon, 10 Jun 2024 01:24:04 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCXdte1O8Edc3YcOZSACA5GHZDzW8ACBeSjPaqsmg0BWpb0jlzQ7bLCHMjsqRZnw6ACpzF7Cm4KBOUpDCKGPF5kAYXoclPu8TE74qb5WKg== X-Google-Smtp-Source: AGHT+IEgsNo7WNesCgZ2UL9hNQB3+lwq+AwEKKUeH7/GyfyuKg0PLxE/Vd8sYWXuPm/eZXk5ggL7 X-Received: by 2002:a05:620a:4553:b0:795:5122:7ce4 with SMTP id af79cd13be357-79551228550mr876415985a.14.1718007844403; Mon, 10 Jun 2024 01:24:04 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1718007844; cv=pass; d=google.com; s=arc-20160816; b=mhxqETXXhttf9sm5ckVr+wEiMhc/rryj5jbwdKxdpFAGUhnbvmJi4NOBw0H1IvvcO/ nKgH48uBCxJRwKNs04SoOFOUTrVtoLbBgemXg2Gf8gq3qCmi9SthZEF6v5f/+YtmkpKH qAaVNPC2yQFDfTHwfmRdi+J2u453B/xFexTVQFai1jhxyBwBEzsmDHhA++ZCY/ZU4n3m NShV+OMWxWOxfkmT956Ozd3mrOPju+lEGMF5VhtsdP6Vgx1fxhd9EJd8LZf2IuCrU2TJ K3FC7Rj2pEFFp8yilgJlwZOEo4Fnz6yZU0DANjtdmkfClKi9wr3tGpZS6n4XcPG1LMwp nsCQ== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:content-language :references:cc:to:subject:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id:dkim-signature; bh=jV53rnFbdfepoDhe2KzgK5UUmjiUI69pJ1VKH45G2YM=; fh=ZYbKgjpEEVgX235/AAmj2MPzQYUolCG8c5n7EIj6F4A=; b=HIxR6a04Eaw6rpTSRUxLilzjl8Nr1JvPRLNA5JXlL+vkhOqRDkn1eEzle8DjGTtz34 gG3/d3Hg3DOmAhDWAzigUXXGFnDNm3W/ymaKoDN2yx3Y4w/paTFITtCO8PiKuDAiWtJP juLoa+hDu2y6L5suUcvALu+9GChkvymDuvb5XMsID59CaD5GoAIeucbkI8fLUWCuXVFz 5F/m7Cxi8cgKGE0I9FKN7GMIrrUE2zpYH5LwWlH+WcqHGGERzuAWifF9l7KTs+hz1k4j s3ImQNgxfnPPpApwJyBGZsxRKup3WlAElb2eQBvlXYzNHRm/MAzljYRio2qB5Yqrqlmy 4AYA==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@rivosinc-com.20230601.gappssmtp.com header.s=20230601 header.b=su4QM2L8; arc=pass (i=1 spf=pass spfdomain=rivosinc.com dkim=pass dkdomain=rivosinc-com.20230601.gappssmtp.com); spf=pass (google.com: domain of linux-kernel+bounces-207840-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-207840-linux.lists.archive=gmail.com@vger.kernel.org" Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [2604:1380:45d1:ec00::1]) by mx.google.com with ESMTPS id af79cd13be357-7955b5dd0b9si432023185a.506.2024.06.10.01.24.04 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 10 Jun 2024 01:24:04 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-207840-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) client-ip=2604:1380:45d1:ec00::1; Authentication-Results: mx.google.com; dkim=pass header.i=@rivosinc-com.20230601.gappssmtp.com header.s=20230601 header.b=su4QM2L8; arc=pass (i=1 spf=pass spfdomain=rivosinc.com dkim=pass dkdomain=rivosinc-com.20230601.gappssmtp.com); spf=pass (google.com: domain of linux-kernel+bounces-207840-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:45d1:ec00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-207840-linux.lists.archive=gmail.com@vger.kernel.org" Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id 167431C21008 for ; Mon, 10 Jun 2024 08:24:04 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 936BA626CB; Mon, 10 Jun 2024 08:23:57 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=rivosinc-com.20230601.gappssmtp.com header.i=@rivosinc-com.20230601.gappssmtp.com header.b="su4QM2L8" Received: from mail-wm1-f53.google.com (mail-wm1-f53.google.com [209.85.128.53]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 8DA276F2E0 for ; Mon, 10 Jun 2024 08:23:53 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.128.53 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1718007836; cv=none; b=LsSl+1b9TRROSbOWWMRtErh8rqKWRILH9NP4T9MA0kVOowuN0lVMwt5qvls7kug79IN3gYNdRWYxfXtRfE4p8rkim7YmPkeMqavHUJwamNqbe/1cv7JqbjDC8vH32uT5GDNbS7hKVcMypLNX3eBN2TdJUGMZQ2pwJ8Bk6TLul+U= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1718007836; c=relaxed/simple; bh=dhGGFAdgIvbTY7rnpW674JFvZpHeDN/sHnrCwrvSh3M=; h=Message-ID:Date:MIME-Version:Subject:To:Cc:References:From: In-Reply-To:Content-Type; b=aU7/q6j3EdY9yZosAtrRAJIpH2zEKiWAkcDtz8JFIvLeKYxPVcwYdyNhKmYjLj7oNJKa2KaOUpewQIeA+8Qid+0vCorkd2r/gZqb/R3UemmWIKZIDdUoIrML/jy6Edkc6X5vpQfZy/0XgLjnsxPkIJU2KQi+JIGI5btnanjiuRU= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=rivosinc.com; spf=pass smtp.mailfrom=rivosinc.com; dkim=pass (2048-bit key) header.d=rivosinc-com.20230601.gappssmtp.com header.i=@rivosinc-com.20230601.gappssmtp.com header.b=su4QM2L8; arc=none smtp.client-ip=209.85.128.53 Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=rivosinc.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=rivosinc.com Received: by mail-wm1-f53.google.com with SMTP id 5b1f17b1804b1-4217f072531so1446475e9.1 for ; Mon, 10 Jun 2024 01:23:53 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=rivosinc-com.20230601.gappssmtp.com; s=20230601; t=1718007832; x=1718612632; darn=vger.kernel.org; h=content-transfer-encoding:in-reply-to:from:content-language :references:cc:to:subject:user-agent:mime-version:date:message-id :from:to:cc:subject:date:message-id:reply-to; bh=jV53rnFbdfepoDhe2KzgK5UUmjiUI69pJ1VKH45G2YM=; b=su4QM2L8L/l7PMScP5rhDUmlq3t2Z0+oCUxAmOntEHhH8wS9rUrmLlEnuD5nhRJjvJ /4aC0hkgTOqiAEp36CZaalFT36mJ5uyUiEpHP4rbKB5cS52iPda9DcceNHM+6nuvPLdl um0vyjHr6VOYKIyhNm7NrzNyZ252ZowO7coF7zZCOL511P+ao20EsxyZyx81isb6kLvc ugt73fiqkYooLW3dyFFdW3GZwh5rbyfj+kZtVLZDGol7vIr+LQ1/iMWj4GfyerZcxfid Xw5vFZoLpd/IiwQ4z4m07/nE/a7uwxRYwIc8iAiVITXLIwVzXMvZlOgkcKrRme3n+KzT weSg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1718007832; x=1718612632; h=content-transfer-encoding:in-reply-to:from:content-language :references:cc:to:subject:user-agent:mime-version:date:message-id :x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=jV53rnFbdfepoDhe2KzgK5UUmjiUI69pJ1VKH45G2YM=; b=UxUfgNH6JMNyOv4k0D/KMAsMeuvc+8Kzw7T7PWAwfuPQbFmcwsXKBOM06AshY0/I8Y cegTGYz1xsJIj7xe64tZJUASCpSeqD/MBmvCMemyESnzQOw4BqS8cQBG3qgT9WbG5Yh1 8/UBMeX164wouQREqgGVqyT/5YyU0TlNSdl8yKetFyoX5GmKfDFbpiGuqSYa1bJIcgUb as8C1CjItBb1mwDiKjRNW7nY3ODyiUfi2vHyDZ3nekbJ1TDlnWVD/aRR6bWYNORRuqA0 LbzBIHcINhAh6G1SZkj/CqoQQNQrZjOPb1VREmX723QEoltTOz8bqOCcFvMnBW8La67H krDw== X-Forwarded-Encrypted: i=1; AJvYcCXkL35GYuRJJQmX89JlIpHrNH4bmwSLKQ8A47UCllB0GacW3WMybvbhrqPMJxccQsUKM+IypMrSTro6CE/mJZKBgp3rvpSt9eELDfh7 X-Gm-Message-State: AOJu0Yy5i6kx5gUFytq+7S9peusyXuSS2nxWGJVeUmGF5HsWEbXtvWGb hwoDnO7vc3htdVGCP4xj/SvhWwTSQZIHi0Qgd1ZgF983LsJ+LU2BL6HuPcJKj9g= X-Received: by 2002:a05:600c:1c0e:b0:422:78c:82f6 with SMTP id 5b1f17b1804b1-422078c8571mr6188125e9.3.1718007831643; Mon, 10 Jun 2024 01:23:51 -0700 (PDT) Received: from ?IPV6:2a01:e0a:999:a3a0:deb:d879:7179:d010? ([2a01:e0a:999:a3a0:deb:d879:7179:d010]) by smtp.gmail.com with ESMTPSA id 5b1f17b1804b1-4220f668387sm7215725e9.4.2024.06.10.01.23.50 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Mon, 10 Jun 2024 01:23:51 -0700 (PDT) Message-ID: <5284ad23-9870-41cd-893e-f17170af90f8@rivosinc.com> Date: Mon, 10 Jun 2024 10:23:49 +0200 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH 2/3] RISC-V: Detect unaligned vector accesses supported. To: Charlie Jenkins , Jesse Taube Cc: linux-riscv@lists.infradead.org, Paul Walmsley , Palmer Dabbelt , Albert Ou , Conor Dooley , Evan Green , Andrew Jones , Xiao Wang , Andy Chiu , Greentime Hu , Heiko Stuebner , Guo Ren , =?UTF-8?B?QmrDtnJuIFTDtnBlbA==?= , Costa Shulyupin , Andrew Morton , Baoquan He , Sami Tolvanen , Zong Li , Ben Dooks , Erick Archer , Vincent Chen , Joel Granados , linux-kernel@vger.kernel.org References: <20240606183215.416829-1-jesse@rivosinc.com> <20240606183215.416829-2-jesse@rivosinc.com> Content-Language: en-US From: =?UTF-8?B?Q2zDqW1lbnQgTMOpZ2Vy?= In-Reply-To: Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit On 06/06/2024 23:29, Charlie Jenkins wrote: > On Thu, Jun 06, 2024 at 02:32:14PM -0400, Jesse Taube wrote: >> Run a unaligned vector access to test if the system supports >> vector unaligned access. Add the result to a new key in hwprobe. >> This is useful for usermode to know if vector misaligned accesses are >> supported and if they are faster or slower than equivalent byte accesses. >> >> Signed-off-by: Jesse Taube >> --- >> arch/riscv/include/asm/cpufeature.h | 2 + >> arch/riscv/include/asm/hwprobe.h | 2 +- >> arch/riscv/include/asm/vector.h | 1 + >> arch/riscv/include/uapi/asm/hwprobe.h | 6 ++ >> arch/riscv/kernel/sys_hwprobe.c | 34 +++++++++ >> arch/riscv/kernel/traps_misaligned.c | 84 ++++++++++++++++++++-- >> arch/riscv/kernel/unaligned_access_speed.c | 4 ++ >> arch/riscv/kernel/vector.c | 2 +- >> 8 files changed, 129 insertions(+), 6 deletions(-) >> >> diff --git a/arch/riscv/include/asm/cpufeature.h b/arch/riscv/include/asm/cpufeature.h >> index 347805446151..a012c8490a27 100644 >> --- a/arch/riscv/include/asm/cpufeature.h >> +++ b/arch/riscv/include/asm/cpufeature.h >> @@ -35,9 +35,11 @@ void riscv_user_isa_enable(void); >> >> #if defined(CONFIG_RISCV_MISALIGNED) >> bool check_unaligned_access_emulated_all_cpus(void); >> +bool check_vector_unaligned_access_all_cpus(void); >> void unaligned_emulation_finish(void); >> bool unaligned_ctl_available(void); >> DECLARE_PER_CPU(long, misaligned_access_speed); >> +DECLARE_PER_CPU(long, vector_misaligned_access); >> #else >> static inline bool unaligned_ctl_available(void) >> { >> diff --git a/arch/riscv/include/asm/hwprobe.h b/arch/riscv/include/asm/hwprobe.h >> index 630507dff5ea..150a9877b0af 100644 >> --- a/arch/riscv/include/asm/hwprobe.h >> +++ b/arch/riscv/include/asm/hwprobe.h >> @@ -8,7 +8,7 @@ >> >> #include >> >> -#define RISCV_HWPROBE_MAX_KEY 6 >> +#define RISCV_HWPROBE_MAX_KEY 7 >> >> static inline bool riscv_hwprobe_key_is_valid(__s64 key) >> { >> diff --git a/arch/riscv/include/asm/vector.h b/arch/riscv/include/asm/vector.h >> index 731dcd0ed4de..776af9b37e23 100644 >> --- a/arch/riscv/include/asm/vector.h >> +++ b/arch/riscv/include/asm/vector.h >> @@ -21,6 +21,7 @@ >> >> extern unsigned long riscv_v_vsize; >> int riscv_v_setup_vsize(void); >> +bool insn_is_vector(u32 insn_buf); >> bool riscv_v_first_use_handler(struct pt_regs *regs); >> void kernel_vector_begin(void); >> void kernel_vector_end(void); >> diff --git a/arch/riscv/include/uapi/asm/hwprobe.h b/arch/riscv/include/uapi/asm/hwprobe.h >> index 060212331a03..ebacff86f134 100644 >> --- a/arch/riscv/include/uapi/asm/hwprobe.h >> +++ b/arch/riscv/include/uapi/asm/hwprobe.h >> @@ -68,6 +68,12 @@ struct riscv_hwprobe { >> #define RISCV_HWPROBE_MISALIGNED_UNSUPPORTED (4 << 0) >> #define RISCV_HWPROBE_MISALIGNED_MASK (7 << 0) >> #define RISCV_HWPROBE_KEY_ZICBOZ_BLOCK_SIZE 6 >> +#define RISCV_HWPROBE_VEC_KEY_MISALIGNED_PERF 7 >> +#define RISCV_HWPROBE_VEC_MISALIGNED_UNKNOWN 0 >> +#define RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED 1 >> +#define RISCV_HWPROBE_VEC_MISALIGNED_SLOW 2 >> +#define RISCV_HWPROBE_VEC_MISALIGNED_FAST 3 >> +#define RISCV_HWPROBE_VEC_MISALIGNED_UNSUPPORTED 4 >> /* Increase RISCV_HWPROBE_MAX_KEY when adding items. */ >> >> /* Flags */ >> diff --git a/arch/riscv/kernel/sys_hwprobe.c b/arch/riscv/kernel/sys_hwprobe.c >> index b286b73e763e..ce641cc6e47a 100644 >> --- a/arch/riscv/kernel/sys_hwprobe.c >> +++ b/arch/riscv/kernel/sys_hwprobe.c >> @@ -184,6 +184,36 @@ static u64 hwprobe_misaligned(const struct cpumask *cpus) >> } >> #endif >> >> +#if defined(CONFIG_RISCV_PROBE_UNALIGNED_ACCESS) >> +static u64 hwprobe_vec_misaligned(const struct cpumask *cpus) >> +{ >> + int cpu; >> + u64 perf = -1ULL; >> + >> + for_each_cpu(cpu, cpus) { >> + int this_perf = per_cpu(vector_misaligned_access, cpu); >> + >> + if (perf == -1ULL) >> + perf = this_perf; >> + >> + if (perf != this_perf) { >> + perf = RISCV_HWPROBE_VEC_MISALIGNED_UNKNOWN; >> + break; >> + } >> + } >> + >> + if (perf == -1ULL) >> + return RISCV_HWPROBE_VEC_MISALIGNED_UNKNOWN; >> + >> + return perf; >> +} >> +#else >> +static u64 hwprobe_vec_misaligned(const struct cpumask *cpus) >> +{ >> + return RISCV_HWPROBE_VEC_MISALIGNED_UNKNOWN; >> +} >> +#endif >> + >> static void hwprobe_one_pair(struct riscv_hwprobe *pair, >> const struct cpumask *cpus) >> { >> @@ -211,6 +241,10 @@ static void hwprobe_one_pair(struct riscv_hwprobe *pair, >> pair->value = hwprobe_misaligned(cpus); >> break; >> >> + case RISCV_HWPROBE_VEC_KEY_MISALIGNED_PERF: >> + pair->value = hwprobe_vec_misaligned(cpus); >> + break; >> + >> case RISCV_HWPROBE_KEY_ZICBOZ_BLOCK_SIZE: >> pair->value = 0; >> if (hwprobe_ext0_has(cpus, RISCV_HWPROBE_EXT_ZICBOZ)) >> diff --git a/arch/riscv/kernel/traps_misaligned.c b/arch/riscv/kernel/traps_misaligned.c >> index 2adb7c3e4dd5..8f26c3d92230 100644 >> --- a/arch/riscv/kernel/traps_misaligned.c >> +++ b/arch/riscv/kernel/traps_misaligned.c >> @@ -16,6 +16,7 @@ >> #include >> #include >> #include >> +#include >> >> #define INSN_MATCH_LB 0x3 >> #define INSN_MASK_LB 0x707f >> @@ -413,10 +414,6 @@ int handle_misaligned_load(struct pt_regs *regs) >> >> perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, addr); >> >> -#ifdef CONFIG_RISCV_PROBE_UNALIGNED_ACCESS >> - *this_cpu_ptr(&misaligned_access_speed) = RISCV_HWPROBE_MISALIGNED_EMULATED; >> -#endif >> - >> if (!unaligned_enabled) >> return -1; >> >> @@ -426,6 +423,17 @@ int handle_misaligned_load(struct pt_regs *regs) >> if (get_insn(regs, epc, &insn)) >> return -1; >> >> +#ifdef CONFIG_RISCV_PROBE_UNALIGNED_ACCESS >> + if (insn_is_vector(insn) && >> + *this_cpu_ptr(&vector_misaligned_access) == RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED) { >> + *this_cpu_ptr(&vector_misaligned_access) = RISCV_HWPROBE_VEC_MISALIGNED_UNSUPPORTED; >> + regs->epc = epc + INSN_LEN(insn); >> + return 0; >> + } >> + >> + *this_cpu_ptr(&misaligned_access_speed) = RISCV_HWPROBE_MISALIGNED_EMULATED; > > This unconditionally sets scalar unaligned accesses even if the > unaligned access is caused by vector. Scalar unaligned accesses should > only be set to emulated if this function is entered from a scalar > unaligned load. > > The rest of this function handles how scalar unaligned accesses are > emulated, and the equivalent needs to happen for vector. You need to add > routines that manually load the data from the memory address into the > vector register. When Clément did this for scalar, he provided a test > case to help reviewers [1]. Please add onto these test cases or make > your own for vector. > > Link: https://github.com/clementleger/unaligned_test [1] Hi Jesse, I already mentionned that in a previous message and got no answer [1]. Please address all feedback before sending another version. Adding a cover letter for such series would also be appreciated. Thanks, Clément Link :https://lore.kernel.org/all/2c355b76-c768-46b1-9f37 fd1991a54bfd@rivosinc.com/ [1] > >> +#endif >> + >> regs->epc = 0; >> >> if ((insn & INSN_MASK_LW) == INSN_MATCH_LW) { >> @@ -625,6 +633,74 @@ static bool check_unaligned_access_emulated(int cpu) >> return misaligned_emu_detected; >> } >> >> +#ifdef CONFIG_RISCV_ISA_V >> +static void check_vector_unaligned_access(struct work_struct *unused) > > Can you standardize this name with the scalar version by writing > emulated in it? > > "check_vector_unaligned_access_emulated_all_cpus" > >> +{ >> + int cpu = smp_processor_id(); >> + long *mas_ptr = this_cpu_ptr(&vector_misaligned_access); >> + unsigned long tmp_var; >> + >> + if (!riscv_isa_extension_available(hart_isa[cpu].isa, v)) >> + return; >> + >> + *mas_ptr = RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED; >> + >> + local_irq_enable(); >> + kernel_vector_begin(); >> + __asm__ __volatile__ ( >> + ".balign 4\n\t" >> + ".option push\n\t" >> + ".option arch, +v\n\t" >> + " vsetivli zero, 1, e16, m1, ta, ma\n\t" // Vectors of 16b >> + " vle16.v v0, (%[ptr])\n\t" // Load bytes >> + ".option pop\n\t" >> + : : [ptr] "r" ((u8 *)&tmp_var + 1) : "v0", "memory"); > > memory is being read from, but not written to, so there is no need to > have a memory clobber. > >> + kernel_vector_end(); >> + >> + if (*mas_ptr == RISCV_HWPROBE_VEC_MISALIGNED_UNKNOWN) >> + *mas_ptr = RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED; >> +} >> + >> +bool check_vector_unaligned_access_all_cpus(void) >> +{ >> + int cpu; >> + bool ret = true; >> + >> + for_each_online_cpu(cpu) >> + if (riscv_isa_extension_available(hart_isa[cpu].isa, ZICCLSM)) > > zicclsm is not specific to vector so it can be extracted out of this > vector specific function. Assuming that hardware properly reports the > extension, if zicclsm is present then it is known that both vector and > scalar unaligned accesses are supported. > >> + per_cpu(vector_misaligned_access, cpu) = RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED; > > Please use the exising UNKNOWN terminology instead of renaming to > SUPPORTED. Any option that is not UNSUPPORTED implies that unaligned > accesses are supported. > >> + else >> + ret = false; >> + >> + >> + if (ret) >> + return true; >> + ret = true; >> + >> + schedule_on_each_cpu(check_vector_unaligned_access); >> + >> + for_each_online_cpu(cpu) >> + if (per_cpu(vector_misaligned_access, cpu) >> + != RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED) >> + return false; >> + >> + return ret; >> +} >> +#else > > If CONFIG_RISCV_ISA_V is not set, there is no value in checking if > vector unaligned accesses are supported because userspace will not be > allowed to use vector instructions anyway. > > - Charlie > >> +bool check_vector_unaligned_access_all_cpus(void) >> +{ >> + int cpu; >> + >> + for_each_online_cpu(cpu) >> + if (riscv_isa_extension_available(hart_isa[cpu].isa, ZICCLSM)) >> + per_cpu(vector_misaligned_access, cpu) = RISCV_HWPROBE_VEC_MISALIGNED_SUPPORTED; >> + else >> + per_cpu(vector_misaligned_access, cpu) = RISCV_HWPROBE_VEC_MISALIGNED_UNSUPPORTED; >> + >> + return false; >> +} >> +#endif >> + >> bool check_unaligned_access_emulated_all_cpus(void) >> { >> int cpu; >> diff --git a/arch/riscv/kernel/unaligned_access_speed.c b/arch/riscv/kernel/unaligned_access_speed.c >> index a9a6bcb02acf..92a84239beaa 100644 >> --- a/arch/riscv/kernel/unaligned_access_speed.c >> +++ b/arch/riscv/kernel/unaligned_access_speed.c >> @@ -20,6 +20,7 @@ >> #define MISALIGNED_COPY_SIZE ((MISALIGNED_BUFFER_SIZE / 2) - 0x80) >> >> DEFINE_PER_CPU(long, misaligned_access_speed); >> +DEFINE_PER_CPU(long, vector_misaligned_access) = RISCV_HWPROBE_VEC_MISALIGNED_UNKNOWN; >> >> #ifdef CONFIG_RISCV_PROBE_UNALIGNED_ACCESS >> static cpumask_t fast_misaligned_access; >> @@ -264,6 +265,8 @@ static int check_unaligned_access_all_cpus(void) >> { >> bool all_cpus_emulated = check_unaligned_access_emulated_all_cpus(); >> >> + check_vector_unaligned_access_all_cpus(); >> + >> if (!all_cpus_emulated) >> return check_unaligned_access_speed_all_cpus(); >> >> @@ -273,6 +276,7 @@ static int check_unaligned_access_all_cpus(void) >> static int check_unaligned_access_all_cpus(void) >> { >> check_unaligned_access_emulated_all_cpus(); >> + check_vector_unaligned_access_all_cpus(); >> >> return 0; >> } >> diff --git a/arch/riscv/kernel/vector.c b/arch/riscv/kernel/vector.c >> index 6727d1d3b8f2..2cceab739b2c 100644 >> --- a/arch/riscv/kernel/vector.c >> +++ b/arch/riscv/kernel/vector.c >> @@ -66,7 +66,7 @@ void __init riscv_v_setup_ctx_cache(void) >> #endif >> } >> >> -static bool insn_is_vector(u32 insn_buf) >> +bool insn_is_vector(u32 insn_buf) >> { >> u32 opcode = insn_buf & __INSN_OPCODE_MASK; >> u32 width, csr; >> -- >> 2.43.0 >>