2022-06-01 18:46:28

by Jiaxun Yang

[permalink] [raw]
Subject: Re: [PATCH V12 19/24] LoongArch: Add some library functions



在2022年6月1日六月 上午11:00,Huacai Chen写道:
> Add some library functions for LoongArch, including: delay, memset,
> memcpy, memmove, copy_user, strncpy_user, strnlen_user and tlb dump
> functions.
>
> Reviewed-by: WANG Xuerui <[email protected]>
> Signed-off-by: Huacai Chen <[email protected]>

Reviewed-by: Jiaxun Yang <[email protected]>

I particularly love tlbdump function for MIPS. It’s a handy tool for mm debugging.
Good job for LA one.

Thanks.

> ---
> arch/loongarch/include/asm/delay.h | 26 +++++++
> arch/loongarch/include/asm/string.h | 12 +++
> arch/loongarch/lib/clear_user.S | 43 +++++++++++
> arch/loongarch/lib/copy_user.S | 47 ++++++++++++
> arch/loongarch/lib/delay.c | 43 +++++++++++
> arch/loongarch/lib/dump_tlb.c | 111 ++++++++++++++++++++++++++++
> 6 files changed, 282 insertions(+)
> create mode 100644 arch/loongarch/include/asm/delay.h
> create mode 100644 arch/loongarch/include/asm/string.h
> create mode 100644 arch/loongarch/lib/clear_user.S
> create mode 100644 arch/loongarch/lib/copy_user.S
> create mode 100644 arch/loongarch/lib/delay.c
> create mode 100644 arch/loongarch/lib/dump_tlb.c
>
> diff --git a/arch/loongarch/include/asm/delay.h
> b/arch/loongarch/include/asm/delay.h
> new file mode 100644
> index 000000000000..36d775191310
> --- /dev/null
> +++ b/arch/loongarch/include/asm/delay.h
> @@ -0,0 +1,26 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +#ifndef _ASM_DELAY_H
> +#define _ASM_DELAY_H
> +
> +#include <linux/param.h>
> +
> +extern void __delay(unsigned long cycles);
> +extern void __ndelay(unsigned long ns);
> +extern void __udelay(unsigned long us);
> +
> +#define ndelay(ns) __ndelay(ns)
> +#define udelay(us) __udelay(us)
> +
> +/* make sure "usecs *= ..." in udelay do not overflow. */
> +#if HZ >= 1000
> +#define MAX_UDELAY_MS 1
> +#elif HZ <= 200
> +#define MAX_UDELAY_MS 5
> +#else
> +#define MAX_UDELAY_MS (1000 / HZ)
> +#endif
> +
> +#endif /* _ASM_DELAY_H */
> diff --git a/arch/loongarch/include/asm/string.h
> b/arch/loongarch/include/asm/string.h
> new file mode 100644
> index 000000000000..b07e60ded957
> --- /dev/null
> +++ b/arch/loongarch/include/asm/string.h
> @@ -0,0 +1,12 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +#ifndef _ASM_STRING_H
> +#define _ASM_STRING_H
> +
> +extern void *memset(void *__s, int __c, size_t __count);
> +extern void *memcpy(void *__to, __const__ void *__from, size_t __n);
> +extern void *memmove(void *__dest, __const__ void *__src, size_t __n);
> +
> +#endif /* _ASM_STRING_H */
> diff --git a/arch/loongarch/lib/clear_user.S
> b/arch/loongarch/lib/clear_user.S
> new file mode 100644
> index 000000000000..25d9be5fbb19
> --- /dev/null
> +++ b/arch/loongarch/lib/clear_user.S
> @@ -0,0 +1,43 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asm.h>
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +.macro fixup_ex from, to, offset, fix
> +.if \fix
> + .section .fixup, "ax"
> +\to: addi.d a0, a1, \offset
> + jr ra
> + .previous
> +.endif
> + .section __ex_table, "a"
> + PTR \from\()b, \to\()b
> + .previous
> +.endm
> +
> +/*
> + * unsigned long __clear_user(void *addr, size_t size)
> + *
> + * a0: addr
> + * a1: size
> + */
> +SYM_FUNC_START(__clear_user)
> + beqz a1, 2f
> +
> +1: st.b zero, a0, 0
> + addi.d a0, a0, 1
> + addi.d a1, a1, -1
> + bgt a1, zero, 1b
> +
> +2: move a0, a1
> + jr ra
> +
> + fixup_ex 1, 3, 0, 1
> +SYM_FUNC_END(__clear_user)
> +
> +EXPORT_SYMBOL(__clear_user)
> diff --git a/arch/loongarch/lib/copy_user.S
> b/arch/loongarch/lib/copy_user.S
> new file mode 100644
> index 000000000000..9ae507f851b5
> --- /dev/null
> +++ b/arch/loongarch/lib/copy_user.S
> @@ -0,0 +1,47 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +
> +#include <asm/asm.h>
> +#include <asm/asmmacro.h>
> +#include <asm/export.h>
> +#include <asm/regdef.h>
> +
> +.macro fixup_ex from, to, offset, fix
> +.if \fix
> + .section .fixup, "ax"
> +\to: addi.d a0, a2, \offset
> + jr ra
> + .previous
> +.endif
> + .section __ex_table, "a"
> + PTR \from\()b, \to\()b
> + .previous
> +.endm
> +
> +/*
> + * unsigned long __copy_user(void *to, const void *from, size_t n)
> + *
> + * a0: to
> + * a1: from
> + * a2: n
> + */
> +SYM_FUNC_START(__copy_user)
> + beqz a2, 3f
> +
> +1: ld.b t0, a1, 0
> +2: st.b t0, a0, 0
> + addi.d a0, a0, 1
> + addi.d a1, a1, 1
> + addi.d a2, a2, -1
> + bgt a2, zero, 1b
> +
> +3: move a0, a2
> + jr ra
> +
> + fixup_ex 1, 4, 0, 1
> + fixup_ex 2, 4, 0, 0
> +SYM_FUNC_END(__copy_user)
> +
> +EXPORT_SYMBOL(__copy_user)
> diff --git a/arch/loongarch/lib/delay.c b/arch/loongarch/lib/delay.c
> new file mode 100644
> index 000000000000..5d856694fcfe
> --- /dev/null
> +++ b/arch/loongarch/lib/delay.c
> @@ -0,0 +1,43 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + */
> +#include <linux/delay.h>
> +#include <linux/export.h>
> +#include <linux/smp.h>
> +#include <linux/timex.h>
> +
> +#include <asm/compiler.h>
> +#include <asm/processor.h>
> +
> +void __delay(unsigned long cycles)
> +{
> + u64 t0 = get_cycles();
> +
> + while ((unsigned long)(get_cycles() - t0) < cycles)
> + cpu_relax();
> +}
> +EXPORT_SYMBOL(__delay);
> +
> +/*
> + * Division by multiplication: you don't have to worry about
> + * loss of precision.
> + *
> + * Use only for very small delays ( < 1 msec). Should probably use a
> + * lookup table, really, as the multiplications take much too long with
> + * short delays. This is a "reasonable" implementation, though (and
> the
> + * first constant multiplications gets optimized away if the delay is
> + * a constant)
> + */
> +
> +void __udelay(unsigned long us)
> +{
> + __delay((us * 0x000010c7ull * HZ * lpj_fine) >> 32);
> +}
> +EXPORT_SYMBOL(__udelay);
> +
> +void __ndelay(unsigned long ns)
> +{
> + __delay((ns * 0x00000005ull * HZ * lpj_fine) >> 32);
> +}
> +EXPORT_SYMBOL(__ndelay);
> diff --git a/arch/loongarch/lib/dump_tlb.c
> b/arch/loongarch/lib/dump_tlb.c
> new file mode 100644
> index 000000000000..cda2c6bc7f09
> --- /dev/null
> +++ b/arch/loongarch/lib/dump_tlb.c
> @@ -0,0 +1,111 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
> + *
> + * Derived from MIPS:
> + * Copyright (C) 1994, 1995 by Waldorf Electronics, written by Ralf
> Baechle.
> + * Copyright (C) 1999 by Silicon Graphics, Inc.
> + */
> +#include <linux/kernel.h>
> +#include <linux/mm.h>
> +
> +#include <asm/loongarch.h>
> +#include <asm/page.h>
> +#include <asm/pgtable.h>
> +#include <asm/tlb.h>
> +
> +void dump_tlb_regs(void)
> +{
> + const int field = 2 * sizeof(unsigned long);
> +
> + pr_info("Index : %0x\n", read_csr_tlbidx());
> + pr_info("PageSize : %0x\n", read_csr_pagesize());
> + pr_info("EntryHi : %0*llx\n", field, read_csr_entryhi());
> + pr_info("EntryLo0 : %0*llx\n", field, read_csr_entrylo0());
> + pr_info("EntryLo1 : %0*llx\n", field, read_csr_entrylo1());
> +}
> +
> +static void dump_tlb(int first, int last)
> +{
> + unsigned long s_entryhi, entryhi, asid;
> + unsigned long long entrylo0, entrylo1, pa;
> + unsigned int index;
> + unsigned int s_index, s_asid;
> + unsigned int pagesize, c0, c1, i;
> + unsigned long asidmask = cpu_asid_mask(&current_cpu_data);
> + int pwidth = 11;
> + int vwidth = 11;
> + int asidwidth = DIV_ROUND_UP(ilog2(asidmask) + 1, 4);
> +
> + s_entryhi = read_csr_entryhi();
> + s_index = read_csr_tlbidx();
> + s_asid = read_csr_asid();
> +
> + for (i = first; i <= last; i++) {
> + write_csr_index(i);
> + tlb_read();
> + pagesize = read_csr_pagesize();
> + entryhi = read_csr_entryhi();
> + entrylo0 = read_csr_entrylo0();
> + entrylo1 = read_csr_entrylo1();
> + index = read_csr_tlbidx();
> + asid = read_csr_asid();
> +
> + /* EHINV bit marks entire entry as invalid */
> + if (index & CSR_TLBIDX_EHINV)
> + continue;
> + /*
> + * ASID takes effect in absence of G (global) bit.
> + */
> + if (!((entrylo0 | entrylo1) & ENTRYLO_G) &&
> + asid != s_asid)
> + continue;
> +
> + /*
> + * Only print entries in use
> + */
> + pr_info("Index: %2d pgsize=%x ", i, (1 << pagesize));
> +
> + c0 = (entrylo0 & ENTRYLO_C) >> ENTRYLO_C_SHIFT;
> + c1 = (entrylo1 & ENTRYLO_C) >> ENTRYLO_C_SHIFT;
> +
> + pr_cont("va=%0*lx asid=%0*lx",
> + vwidth, (entryhi & ~0x1fffUL), asidwidth, asid & asidmask);
> +
> + /* NR/NX are in awkward places, so mask them off separately */
> + pa = entrylo0 & ~(ENTRYLO_NR | ENTRYLO_NX);
> + pa = pa & PAGE_MASK;
> + pr_cont("\n\t[");
> + pr_cont("ri=%d xi=%d ",
> + (entrylo0 & ENTRYLO_NR) ? 1 : 0,
> + (entrylo0 & ENTRYLO_NX) ? 1 : 0);
> + pr_cont("pa=%0*llx c=%d d=%d v=%d g=%d plv=%lld] [",
> + pwidth, pa, c0,
> + (entrylo0 & ENTRYLO_D) ? 1 : 0,
> + (entrylo0 & ENTRYLO_V) ? 1 : 0,
> + (entrylo0 & ENTRYLO_G) ? 1 : 0,
> + (entrylo0 & ENTRYLO_PLV) >> ENTRYLO_PLV_SHIFT);
> + /* NR/NX are in awkward places, so mask them off separately */
> + pa = entrylo1 & ~(ENTRYLO_NR | ENTRYLO_NX);
> + pa = pa & PAGE_MASK;
> + pr_cont("ri=%d xi=%d ",
> + (entrylo1 & ENTRYLO_NR) ? 1 : 0,
> + (entrylo1 & ENTRYLO_NX) ? 1 : 0);
> + pr_cont("pa=%0*llx c=%d d=%d v=%d g=%d plv=%lld]\n",
> + pwidth, pa, c1,
> + (entrylo1 & ENTRYLO_D) ? 1 : 0,
> + (entrylo1 & ENTRYLO_V) ? 1 : 0,
> + (entrylo1 & ENTRYLO_G) ? 1 : 0,
> + (entrylo1 & ENTRYLO_PLV) >> ENTRYLO_PLV_SHIFT);
> + }
> + pr_info("\n");
> +
> + write_csr_entryhi(s_entryhi);
> + write_csr_tlbidx(s_index);
> + write_csr_asid(s_asid);
> +}
> +
> +void dump_tlb_all(void)
> +{
> + dump_tlb(0, current_cpu_data.tlbsize - 1);
> +}
> --
> 2.27.0

--
- Jiaxun