module_layout manages different types of memory (text, data, rodata, etc.)
in one allocation, which is problematic for some reasons:
1. It is hard to enable CONFIG_STRICT_MODULE_RWX.
2. It is hard to use huge pages in modules (and not break strict rwx).
3. Many archs uses module_layout for arch-specific data, but it is not
obvious how these data are used (are they RO, RX, or RW?)
Improve the scenario by replacing 2 (or 3) module_layout per module with
up to 7 module_memory per module:
MOD_TEXT,
MOD_DATA,
MOD_RODATA,
MOD_RO_AFTER_INIT,
MOD_INIT_TEXT,
MOD_INIT_DATA,
MOD_INIT_RODATA,
and allocating them separately. This adds slightly more entries to
mod_tree (from up to 3 entries per module, to up to 7 entries per
module). However, this at most adds a small constant overhead to
__module_address(), which is expected to be fast.
Various archs use module_layout for different data. These data are put
into different module_memory based on their location in module_layout.
IOW, data that used to go with text is allocated with MOD_MEM_TYPE_TEXT;
data that used to go with data is allocated with MOD_MEM_TYPE_DATA, etc.
module_memory simplifies quite some of the module code. For example,
ARCH_WANTS_MODULES_DATA_IN_VMALLOC is a lot cleaner, as it just uses a
different allocator for the data. kernel/module/strict_rwx.c is also
much cleaner with module_memory.
Signed-off-by: Song Liu <[email protected]>
Cc: Luis Chamberlain <[email protected]>
Cc: Thomas Gleixner <[email protected]>
Cc: Peter Zijlstra <[email protected]>
Cc: Guenter Roeck <[email protected]>
Cc: Christophe Leroy <[email protected]>
---
This is the preparation work for the type aware module_alloc() discussed
in [1]. While this work is not covered much in the discussion, it is a
critical step of the effort.
As this part grows pretty big (~1000 lines, + and -), I would like get
some feedback on it, so that I know it is on the right track.
Please share your comments. Thanks!
Test coverage: Tested on x86_64.
Build tested by kernel test bot in [2]. The only regression in [2] was a
typo in parisc, which is also fixed.
[1] https://lore.kernel.org/linux-mm/[email protected]/T/#u
[2] https://lore.kernel.org/linux-raid/63b8827e.clJQX2wg+I+tiX7m%[email protected]/T/#u
Changes v1 => v2:
1. Add data_addr_[min|max] for CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
case.
Changes v2 => v3:
1. Fix and remove the last use of INIT_OFFSET_MASK.
2. Add more information in the commit log. (Luis Chamberlain)
3. Rebase and fix issues in x86/calltrunks.
4. Minor cleanup.
Changes v3 => v4:
1. Shorten enum/variable names, so that the code are easier to read.
(Christophe Leroy)
2. Remove an used variable. (Guenter Roeck, Christophe Leroy)
Changes v4 => v5:
1. Simplify some code some code. (Peter Zijlstra, Christophe Leroy)
2. Remove module_check_misalignment(), which is not useful any more.
Changes v5 => v6:
1. Improve mod_mem_type_is_* and for_*mod_mem_type marcos.
(Peter Zijlstra).
---
arch/alpha/kernel/module.c | 2 +-
arch/arc/kernel/unwind.c | 9 +-
arch/arm/kernel/module-plts.c | 2 +-
arch/arm64/kernel/module-plts.c | 2 +-
arch/ia64/kernel/module.c | 24 ++-
arch/mips/kernel/vpe.c | 11 +-
arch/parisc/kernel/module.c | 51 ++---
arch/powerpc/kernel/module_32.c | 7 +-
arch/s390/kernel/module.c | 26 +--
arch/x86/kernel/callthunks.c | 4 +-
arch/x86/kernel/module.c | 4 +-
include/linux/module.h | 86 +++++---
kernel/module/internal.h | 40 ++--
kernel/module/kallsyms.c | 56 ++---
kernel/module/kdb.c | 17 +-
kernel/module/main.c | 355 ++++++++++++++++++--------------
kernel/module/procfs.c | 16 +-
kernel/module/strict_rwx.c | 100 ++-------
kernel/module/tree_lookup.c | 39 ++--
19 files changed, 432 insertions(+), 419 deletions(-)
diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c
index 5b60c248de9e..9109213abc09 100644
--- a/arch/alpha/kernel/module.c
+++ b/arch/alpha/kernel/module.c
@@ -148,7 +148,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab,
/* The small sections were sorted to the end of the segment.
The following should definitely cover them. */
- gp = (u64)me->core_layout.base + me->core_layout.size - 0x8000;
+ gp = (u64)me->mem[MOD_DATA].base + me->mem[MOD_DATA].size - 0x8000;
got = sechdrs[me->arch.gotsecindex].sh_addr;
for (i = 0; i < n; i++) {
diff --git a/arch/arc/kernel/unwind.c b/arch/arc/kernel/unwind.c
index 200270a94558..933451f4494f 100644
--- a/arch/arc/kernel/unwind.c
+++ b/arch/arc/kernel/unwind.c
@@ -369,6 +369,8 @@ void *unwind_add_table(struct module *module, const void *table_start,
unsigned long table_size)
{
struct unwind_table *table;
+ struct module_memory *mod_mem_core_text;
+ struct module_memory *mod_mem_init_text;
if (table_size <= 0)
return NULL;
@@ -377,9 +379,12 @@ void *unwind_add_table(struct module *module, const void *table_start,
if (!table)
return NULL;
+ mod_mem_core_text = &module->mem[MOD_TEXT];
+ mod_mem_init_text = &module->mem[MOD_INIT_TEXT];
+
init_unwind_table(table, module->name,
- module->core_layout.base, module->core_layout.size,
- module->init_layout.base, module->init_layout.size,
+ mod_mem_core_text->base, mod_mem_core_text->size,
+ mod_mem_init_text->base, mod_mem_init_text->size,
table_start, table_size,
NULL, 0);
diff --git a/arch/arm/kernel/module-plts.c b/arch/arm/kernel/module-plts.c
index af7c322ebed6..9d4ecb6b1412 100644
--- a/arch/arm/kernel/module-plts.c
+++ b/arch/arm/kernel/module-plts.c
@@ -30,7 +30,7 @@ static const u32 fixed_plts[] = {
static bool in_init(const struct module *mod, unsigned long loc)
{
- return loc - (u32)mod->init_layout.base < mod->init_layout.size;
+ return within_module_init(loc, mod);
}
static void prealloc_fixed(struct mod_plt_sec *pltsec, struct plt_entries *plt)
diff --git a/arch/arm64/kernel/module-plts.c b/arch/arm64/kernel/module-plts.c
index 5a0a8f552a61..fedffa83a283 100644
--- a/arch/arm64/kernel/module-plts.c
+++ b/arch/arm64/kernel/module-plts.c
@@ -67,7 +67,7 @@ static bool plt_entries_equal(const struct plt_entry *a,
static bool in_init(const struct module *mod, void *loc)
{
- return (u64)loc - (u64)mod->init_layout.base < mod->init_layout.size;
+ return within_module_init(loc, mod);
}
u64 module_emit_plt_entry(struct module *mod, Elf64_Shdr *sechdrs,
diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c
index 8f62cf97f691..3661135da9d9 100644
--- a/arch/ia64/kernel/module.c
+++ b/arch/ia64/kernel/module.c
@@ -485,19 +485,19 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
return 0;
}
-static inline int
+static inline bool
in_init (const struct module *mod, uint64_t addr)
{
- return addr - (uint64_t) mod->init_layout.base < mod->init_layout.size;
+ return within_module_init(addr, mod);
}
-static inline int
+static inline bool
in_core (const struct module *mod, uint64_t addr)
{
- return addr - (uint64_t) mod->core_layout.base < mod->core_layout.size;
+ return within_module_core(addr, mod);
}
-static inline int
+static inline bool
is_internal (const struct module *mod, uint64_t value)
{
return in_init(mod, value) || in_core(mod, value);
@@ -677,7 +677,8 @@ do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend,
break;
case RV_BDREL:
- val -= (uint64_t) (in_init(mod, val) ? mod->init_layout.base : mod->core_layout.base);
+ val -= (uint64_t) (in_init(mod, val) ? mod->mem[MOD_INIT_TEXT].base :
+ mod->mem[MOD_TEXT].base);
break;
case RV_LTV:
@@ -812,15 +813,18 @@ apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind
* addresses have been selected...
*/
uint64_t gp;
- if (mod->core_layout.size > MAX_LTOFF)
+ struct module_memory *mod_mem;
+
+ mod_mem = &mod->mem[MOD_DATA];
+ if (mod_mem->size > MAX_LTOFF)
/*
* This takes advantage of fact that SHF_ARCH_SMALL gets allocated
* at the end of the module.
*/
- gp = mod->core_layout.size - MAX_LTOFF / 2;
+ gp = mod_mem->size - MAX_LTOFF / 2;
else
- gp = mod->core_layout.size / 2;
- gp = (uint64_t) mod->core_layout.base + ((gp + 7) & -8);
+ gp = mod_mem->size / 2;
+ gp = (uint64_t) mod_mem->base + ((gp + 7) & -8);
mod->arch.gp = gp;
DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
}
diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c
index 13294972707b..ab114c686f9d 100644
--- a/arch/mips/kernel/vpe.c
+++ b/arch/mips/kernel/vpe.c
@@ -199,18 +199,17 @@ static void layout_sections(struct module *mod, const Elf_Ehdr *hdr,
for (m = 0; m < ARRAY_SIZE(masks); ++m) {
for (i = 0; i < hdr->e_shnum; ++i) {
Elf_Shdr *s = &sechdrs[i];
+ struct module_memory *mod_mem;
+
+ mod_mem = &mod->mem[MOD_TEXT];
if ((s->sh_flags & masks[m][0]) != masks[m][0]
|| (s->sh_flags & masks[m][1])
|| s->sh_entsize != ~0UL)
continue;
s->sh_entsize =
- get_offset((unsigned long *)&mod->core_layout.size, s);
+ get_offset((unsigned long *)&mod_mem->size, s);
}
-
- if (m == 0)
- mod->core_layout.text_size = mod->core_layout.size;
-
}
}
@@ -641,7 +640,7 @@ static int vpe_elfload(struct vpe *v)
layout_sections(&mod, hdr, sechdrs, secstrings);
}
- v->load_addr = alloc_progmem(mod.core_layout.size);
+ v->load_addr = alloc_progmem(mod.mod_mem[MOD_TEXT].size);
if (!v->load_addr)
return -ENOMEM;
diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c
index 7df140545b22..f6e38c4d3904 100644
--- a/arch/parisc/kernel/module.c
+++ b/arch/parisc/kernel/module.c
@@ -27,9 +27,9 @@
* We are not doing SEGREL32 handling correctly. According to the ABI, we
* should do a value offset, like this:
* if (in_init(me, (void *)val))
- * val -= (uint32_t)me->init_layout.base;
+ * val -= (uint32_t)me->mem[MOD_INIT_TEXT].base;
* else
- * val -= (uint32_t)me->core_layout.base;
+ * val -= (uint32_t)me->mem[MOD_TEXT].base;
* However, SEGREL32 is used only for PARISC unwind entries, and we want
* those entries to have an absolute address, and not just an offset.
*
@@ -76,25 +76,6 @@
* allows us to allocate up to 4095 GOT entries. */
#define MAX_GOTS 4095
-/* three functions to determine where in the module core
- * or init pieces the location is */
-static inline int in_init(struct module *me, void *loc)
-{
- return (loc >= me->init_layout.base &&
- loc <= (me->init_layout.base + me->init_layout.size));
-}
-
-static inline int in_core(struct module *me, void *loc)
-{
- return (loc >= me->core_layout.base &&
- loc <= (me->core_layout.base + me->core_layout.size));
-}
-
-static inline int in_local(struct module *me, void *loc)
-{
- return in_init(me, loc) || in_core(me, loc);
-}
-
#ifndef CONFIG_64BIT
struct got_entry {
Elf32_Addr addr;
@@ -302,6 +283,7 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
{
unsigned long gots = 0, fdescs = 0, len;
unsigned int i;
+ struct module_memory *mod_mem;
len = hdr->e_shnum * sizeof(me->arch.section[0]);
me->arch.section = kzalloc(len, GFP_KERNEL);
@@ -346,14 +328,15 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
me->arch.section[s].stub_entries += count;
}
+ mod_mem = &me->mem[MOD_TEXT];
/* align things a bit */
- me->core_layout.size = ALIGN(me->core_layout.size, 16);
- me->arch.got_offset = me->core_layout.size;
- me->core_layout.size += gots * sizeof(struct got_entry);
+ mod_mem->size = ALIGN(mod_mem->size, 16);
+ me->arch.got_offset = mod_mem->size;
+ mod_mem->size += gots * sizeof(struct got_entry);
- me->core_layout.size = ALIGN(me->core_layout.size, 16);
- me->arch.fdesc_offset = me->core_layout.size;
- me->core_layout.size += fdescs * sizeof(Elf_Fdesc);
+ mod_mem->size = ALIGN(mod_mem->size, 16);
+ me->arch.fdesc_offset = mod_mem->size;
+ mod_mem->size += fdescs * sizeof(Elf_Fdesc);
me->arch.got_max = gots;
me->arch.fdesc_max = fdescs;
@@ -371,7 +354,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
BUG_ON(value == 0);
- got = me->core_layout.base + me->arch.got_offset;
+ got = me->mem[MOD_TEXT].base + me->arch.got_offset;
for (i = 0; got[i].addr; i++)
if (got[i].addr == value)
goto out;
@@ -389,7 +372,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
#ifdef CONFIG_64BIT
static Elf_Addr get_fdesc(struct module *me, unsigned long value)
{
- Elf_Fdesc *fdesc = me->core_layout.base + me->arch.fdesc_offset;
+ Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
if (!value) {
printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
@@ -407,7 +390,7 @@ static Elf_Addr get_fdesc(struct module *me, unsigned long value)
/* Create new one */
fdesc->addr = value;
- fdesc->gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
+ fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
return (Elf_Addr)fdesc;
}
#endif /* CONFIG_64BIT */
@@ -742,7 +725,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
loc, val);
val += addend;
/* can we reach it locally? */
- if (in_local(me, (void *)val)) {
+ if (within_module(val, me)) {
/* this is the case where the symbol is local
* to the module, but in a different section,
* so stub the jump in case it's more than 22
@@ -801,7 +784,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
break;
case R_PARISC_FPTR64:
/* 64-bit function address */
- if(in_local(me, (void *)(val + addend))) {
+ if (within_module(val + addend, me)) {
*loc64 = get_fdesc(me, val+addend);
pr_debug("FDESC for %s at %llx points to %llx\n",
strtab + sym->st_name, *loc64,
@@ -839,7 +822,7 @@ register_unwind_table(struct module *me,
table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
end = table + sechdrs[me->arch.unwind_section].sh_size;
- gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
+ gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
pr_debug("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
me->arch.unwind_section, table, end, gp);
@@ -977,7 +960,7 @@ void module_arch_cleanup(struct module *mod)
#ifdef CONFIG_64BIT
void *dereference_module_function_descriptor(struct module *mod, void *ptr)
{
- unsigned long start_opd = (Elf64_Addr)mod->core_layout.base +
+ unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
mod->arch.fdesc_offset;
unsigned long end_opd = start_opd +
mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c
index ea6536171778..2729179e9dc0 100644
--- a/arch/powerpc/kernel/module_32.c
+++ b/arch/powerpc/kernel/module_32.c
@@ -163,8 +163,7 @@ static uint32_t do_plt_call(void *location,
pr_debug("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
/* Init, or core PLT? */
- if (location >= mod->core_layout.base
- && location < mod->core_layout.base + mod->core_layout.size)
+ if (within_module_init(location, mod))
entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
else
entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
@@ -322,14 +321,14 @@ notrace int module_trampoline_target(struct module *mod, unsigned long addr,
int module_finalize_ftrace(struct module *module, const Elf_Shdr *sechdrs)
{
- module->arch.tramp = do_plt_call(module->core_layout.base,
+ module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
(unsigned long)ftrace_caller,
sechdrs, module);
if (!module->arch.tramp)
return -ENOENT;
#ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
- module->arch.tramp_regs = do_plt_call(module->core_layout.base,
+ module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
(unsigned long)ftrace_regs_caller,
sechdrs, module);
if (!module->arch.tramp_regs)
diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c
index 2d159b32885b..3ed455b785ad 100644
--- a/arch/s390/kernel/module.c
+++ b/arch/s390/kernel/module.c
@@ -126,6 +126,7 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
Elf_Rela *rela;
char *strings;
int nrela, i, j;
+ struct module_memory *mod_mem;
/* Find symbol table and string table. */
symtab = NULL;
@@ -173,14 +174,15 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
/* Increase core size by size of got & plt and set start
offsets for got and plt. */
- me->core_layout.size = ALIGN(me->core_layout.size, 4);
- me->arch.got_offset = me->core_layout.size;
- me->core_layout.size += me->arch.got_size;
- me->arch.plt_offset = me->core_layout.size;
+ mod_mem = &me->mem[MOD_TEXT];
+ mod_mem->size = ALIGN(mod_mem->size, 4);
+ me->arch.got_offset = mod_mem->size;
+ mod_mem->size += me->arch.got_size;
+ me->arch.plt_offset = mod_mem->size;
if (me->arch.plt_size) {
if (IS_ENABLED(CONFIG_EXPOLINE) && !nospec_disable)
me->arch.plt_size += PLT_ENTRY_SIZE;
- me->core_layout.size += me->arch.plt_size;
+ mod_mem->size += me->arch.plt_size;
}
return 0;
}
@@ -304,7 +306,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
case R_390_GOTPLT64: /* 64 bit offset to jump slot. */
case R_390_GOTPLTENT: /* 32 bit rel. offset to jump slot >> 1. */
if (info->got_initialized == 0) {
- Elf_Addr *gotent = me->core_layout.base +
+ Elf_Addr *gotent = me->mem[MOD_TEXT].base +
me->arch.got_offset +
info->got_offset;
@@ -329,7 +331,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
rc = apply_rela_bits(loc, val, 0, 64, 0, write);
else if (r_type == R_390_GOTENT ||
r_type == R_390_GOTPLTENT) {
- val += (Elf_Addr) me->core_layout.base - loc;
+ val += (Elf_Addr) me->mem[MOD_TEXT].base - loc;
rc = apply_rela_bits(loc, val, 1, 32, 1, write);
}
break;
@@ -345,7 +347,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
char *plt_base;
char *ip;
- plt_base = me->core_layout.base + me->arch.plt_offset;
+ plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
ip = plt_base + info->plt_offset;
*(int *)insn = 0x0d10e310; /* basr 1,0 */
*(int *)&insn[4] = 0x100c0004; /* lg 1,12(1) */
@@ -375,7 +377,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
val - loc + 0xffffUL < 0x1ffffeUL) ||
(r_type == R_390_PLT32DBL &&
val - loc + 0xffffffffULL < 0x1fffffffeULL)))
- val = (Elf_Addr) me->core_layout.base +
+ val = (Elf_Addr) me->mem[MOD_TEXT].base +
me->arch.plt_offset +
info->plt_offset;
val += rela->r_addend - loc;
@@ -397,7 +399,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
case R_390_GOTOFF32: /* 32 bit offset to GOT. */
case R_390_GOTOFF64: /* 64 bit offset to GOT. */
val = val + rela->r_addend -
- ((Elf_Addr) me->core_layout.base + me->arch.got_offset);
+ ((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
if (r_type == R_390_GOTOFF16)
rc = apply_rela_bits(loc, val, 0, 16, 0, write);
else if (r_type == R_390_GOTOFF32)
@@ -407,7 +409,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
break;
case R_390_GOTPC: /* 32 bit PC relative offset to GOT. */
case R_390_GOTPCDBL: /* 32 bit PC rel. off. to GOT shifted by 1. */
- val = (Elf_Addr) me->core_layout.base + me->arch.got_offset +
+ val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
rela->r_addend - loc;
if (r_type == R_390_GOTPC)
rc = apply_rela_bits(loc, val, 1, 32, 0, write);
@@ -515,7 +517,7 @@ int module_finalize(const Elf_Ehdr *hdr,
!nospec_disable && me->arch.plt_size) {
unsigned int *ij;
- ij = me->core_layout.base + me->arch.plt_offset +
+ ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
me->arch.plt_size - PLT_ENTRY_SIZE;
ij[0] = 0xc6000000; /* exrl %r0,.+10 */
ij[1] = 0x0005a7f4; /* j . */
diff --git a/arch/x86/kernel/callthunks.c b/arch/x86/kernel/callthunks.c
index ffea98f9064b..22ab13966427 100644
--- a/arch/x86/kernel/callthunks.c
+++ b/arch/x86/kernel/callthunks.c
@@ -330,8 +330,8 @@ void noinline callthunks_patch_module_calls(struct callthunk_sites *cs,
struct module *mod)
{
struct core_text ct = {
- .base = (unsigned long)mod->core_layout.base,
- .end = (unsigned long)mod->core_layout.base + mod->core_layout.size,
+ .base = (unsigned long)mod->mem[MOD_TEXT].base,
+ .end = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
.name = mod->name,
};
diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c
index 705fb2a41d7d..dbc8e226c464 100644
--- a/arch/x86/kernel/module.c
+++ b/arch/x86/kernel/module.c
@@ -339,8 +339,8 @@ int module_finalize(const Elf_Ehdr *hdr,
}
if (locks) {
void *lseg = (void *)locks->sh_addr;
- void *text = me->core_layout.base;
- void *text_end = text + me->core_layout.text_size;
+ void *text = me->mem[MOD_TEXT].base;
+ void *text_end = text + me->mem[MOD_TEXT].size;
alternatives_smp_module_add(me, me->name,
lseg, lseg + locks->sh_size,
text, text_end);
diff --git a/include/linux/module.h b/include/linux/module.h
index 8c5909c0076c..ccc9de8a93db 100644
--- a/include/linux/module.h
+++ b/include/linux/module.h
@@ -320,17 +320,47 @@ struct mod_tree_node {
struct latch_tree_node node;
};
-struct module_layout {
- /* The actual code + data. */
+enum mod_mem_type {
+ MOD_TEXT = 0,
+ MOD_DATA,
+ MOD_RODATA,
+ MOD_RO_AFTER_INIT,
+ MOD_INIT_TEXT,
+ MOD_INIT_DATA,
+ MOD_INIT_RODATA,
+
+ MOD_MEM_NUM_TYPES,
+ MOD_INVALID = -1,
+};
+
+#define mod_mem_type_is_init(type) \
+ ((type) == MOD_INIT_TEXT || \
+ (type) == MOD_INIT_DATA || \
+ (type) == MOD_INIT_RODATA)
+
+#define mod_mem_type_is_core(type) (!mod_mem_type_is_init(type))
+
+#define mod_mem_type_is_text(type) \
+ ((type) == MOD_TEXT || \
+ (type) == MOD_INIT_TEXT)
+
+#define mod_mem_type_is_data(type) (!mod_mem_type_is_text(type))
+
+#define mod_mem_type_is_core_data(type) \
+ (mod_mem_type_is_core(type) && \
+ mod_mem_type_is_data(type))
+
+#define for_each_mod_mem_type(type) \
+ for (enum mod_mem_type (type) = 0; \
+ (type) < MOD_MEM_NUM_TYPES; (type)++)
+
+#define for_class_mod_mem_type(type, class) \
+ for_each_mod_mem_type(type) \
+ if (mod_mem_type_is_##class(type))
+
+struct module_memory {
void *base;
- /* Total size. */
unsigned int size;
- /* The size of the executable code. */
- unsigned int text_size;
- /* Size of RO section of the module (text+rodata) */
- unsigned int ro_size;
- /* Size of RO after init section */
- unsigned int ro_after_init_size;
#ifdef CONFIG_MODULES_TREE_LOOKUP
struct mod_tree_node mtn;
@@ -339,9 +369,9 @@ struct module_layout {
#ifdef CONFIG_MODULES_TREE_LOOKUP
/* Only touch one cacheline for common rbtree-for-core-layout case. */
-#define __module_layout_align ____cacheline_aligned
+#define __module_memory_align ____cacheline_aligned
#else
-#define __module_layout_align
+#define __module_memory_align
#endif
struct mod_kallsyms {
@@ -418,12 +448,8 @@ struct module {
/* Startup function. */
int (*init)(void);
- /* Core layout: rbtree is accessed frequently, so keep together. */
- struct module_layout core_layout __module_layout_align;
- struct module_layout init_layout;
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- struct module_layout data_layout;
-#endif
+ /* rbtree is accessed frequently, so keep together. */
+ struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
/* Arch-specific module values */
struct mod_arch_specific arch;
@@ -573,23 +599,33 @@ bool __is_module_percpu_address(unsigned long addr, unsigned long *can_addr);
bool is_module_percpu_address(unsigned long addr);
bool is_module_text_address(unsigned long addr);
+static inline bool within_module_mem_type(unsigned long addr,
+ const struct module *mod,
+ enum mod_mem_type type)
+{
+ unsigned long base, size;
+
+ base = (unsigned long)mod->mem[type].base;
+ size = mod->mem[type].size;
+ return addr - base < size;
+}
+
static inline bool within_module_core(unsigned long addr,
const struct module *mod)
{
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- if ((unsigned long)mod->data_layout.base <= addr &&
- addr < (unsigned long)mod->data_layout.base + mod->data_layout.size)
+ for_class_mod_mem_type(type, core)
+ if (within_module_mem_type(addr, mod, type))
return true;
-#endif
- return (unsigned long)mod->core_layout.base <= addr &&
- addr < (unsigned long)mod->core_layout.base + mod->core_layout.size;
+ return false;
}
static inline bool within_module_init(unsigned long addr,
const struct module *mod)
{
- return (unsigned long)mod->init_layout.base <= addr &&
- addr < (unsigned long)mod->init_layout.base + mod->init_layout.size;
+ for_class_mod_mem_type(type, init)
+ if (within_module_mem_type(addr, mod, type))
+ return true;
+ return false;
}
static inline bool within_module(unsigned long addr, const struct module *mod)
diff --git a/kernel/module/internal.h b/kernel/module/internal.h
index 2e2bf236f558..c6c44ceca07a 100644
--- a/kernel/module/internal.h
+++ b/kernel/module/internal.h
@@ -17,27 +17,19 @@
#define ARCH_SHF_SMALL 0
#endif
-/* If this is set, the section belongs in the init part of the module */
-#define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG - 1))
-/* Maximum number of characters written by module_flags() */
-#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
-
-#ifndef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-#define data_layout core_layout
-#endif
-
/*
- * Modules' sections will be aligned on page boundaries
- * to ensure complete separation of code and data, but
- * only when CONFIG_STRICT_MODULE_RWX=y
+ * Use highest 4 bits of sh_entsize to store the mod_mem_type of this
+ * section. This leaves 28 bits for offset on 32-bit systems, which is
+ * about 256 MiB (WARN_ON_ONCE if we exceed that).
*/
-static inline unsigned int strict_align(unsigned int size)
-{
- if (IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
- return PAGE_ALIGN(size);
- else
- return size;
-}
+
+#define SH_ENTSIZE_TYPE_BITS 4
+#define SH_ENTSIZE_TYPE_SHIFT (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)
+#define SH_ENTSIZE_TYPE_MASK ((1UL << SH_ENTSIZE_TYPE_BITS) - 1)
+#define SH_ENTSIZE_OFFSET_MASK ((1UL << (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)) - 1)
+
+/* Maximum number of characters written by module_flags() */
+#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
extern struct mutex module_mutex;
extern struct list_head modules;
@@ -101,8 +93,8 @@ int try_to_force_load(struct module *mod, const char *reason);
bool find_symbol(struct find_symbol_arg *fsa);
struct module *find_module_all(const char *name, size_t len, bool even_unformed);
int cmp_name(const void *name, const void *sym);
-long module_get_offset(struct module *mod, unsigned int *size, Elf_Shdr *sechdr,
- unsigned int section);
+long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
+ Elf_Shdr *sechdr, unsigned int section);
char *module_flags(struct module *mod, char *buf, bool show_state);
size_t module_flags_taint(unsigned long taints, char *buf);
@@ -190,10 +182,13 @@ struct mod_tree_root {
#endif
unsigned long addr_min;
unsigned long addr_max;
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+ unsigned long data_addr_min;
+ unsigned long data_addr_max;
+#endif
};
extern struct mod_tree_root mod_tree;
-extern struct mod_tree_root mod_data_tree;
#ifdef CONFIG_MODULES_TREE_LOOKUP
void mod_tree_insert(struct module *mod);
@@ -224,7 +219,6 @@ void module_enable_nx(const struct module *mod);
void module_enable_x(const struct module *mod);
int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
char *secstrings, struct module *mod);
-bool module_check_misalignment(const struct module *mod);
#ifdef CONFIG_MODULE_SIG
int module_sig_check(struct load_info *info, int flags);
diff --git a/kernel/module/kallsyms.c b/kernel/module/kallsyms.c
index 4523f99b0358..acdca7692961 100644
--- a/kernel/module/kallsyms.c
+++ b/kernel/module/kallsyms.c
@@ -78,6 +78,7 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
unsigned int shnum, unsigned int pcpundx)
{
const Elf_Shdr *sec;
+ enum mod_mem_type type;
if (src->st_shndx == SHN_UNDEF ||
src->st_shndx >= shnum ||
@@ -90,11 +91,12 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
#endif
sec = sechdrs + src->st_shndx;
+ type = sec->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
if (!(sec->sh_flags & SHF_ALLOC)
#ifndef CONFIG_KALLSYMS_ALL
|| !(sec->sh_flags & SHF_EXECINSTR)
#endif
- || (sec->sh_entsize & INIT_OFFSET_MASK))
+ || mod_mem_type_is_init(type))
return false;
return true;
@@ -113,11 +115,13 @@ void layout_symtab(struct module *mod, struct load_info *info)
Elf_Shdr *strsect = info->sechdrs + info->index.str;
const Elf_Sym *src;
unsigned int i, nsrc, ndst, strtab_size = 0;
+ struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
+ struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
/* Put symbol section at end of init part of module. */
symsect->sh_flags |= SHF_ALLOC;
- symsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, symsect,
- info->index.sym) | INIT_OFFSET_MASK;
+ symsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
+ symsect, info->index.sym);
pr_debug("\t%s\n", info->secstrings + symsect->sh_name);
src = (void *)info->hdr + symsect->sh_offset;
@@ -134,28 +138,27 @@ void layout_symtab(struct module *mod, struct load_info *info)
}
/* Append room for core symbols at end of core part. */
- info->symoffs = ALIGN(mod->data_layout.size, symsect->sh_addralign ?: 1);
- info->stroffs = mod->data_layout.size = info->symoffs + ndst * sizeof(Elf_Sym);
- mod->data_layout.size += strtab_size;
+ info->symoffs = ALIGN(mod_mem_data->size, symsect->sh_addralign ?: 1);
+ info->stroffs = mod_mem_data->size = info->symoffs + ndst * sizeof(Elf_Sym);
+ mod_mem_data->size += strtab_size;
/* Note add_kallsyms() computes strtab_size as core_typeoffs - stroffs */
- info->core_typeoffs = mod->data_layout.size;
- mod->data_layout.size += ndst * sizeof(char);
- mod->data_layout.size = strict_align(mod->data_layout.size);
+ info->core_typeoffs = mod_mem_data->size;
+ mod_mem_data->size += ndst * sizeof(char);
/* Put string table section at end of init part of module. */
strsect->sh_flags |= SHF_ALLOC;
- strsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, strsect,
- info->index.str) | INIT_OFFSET_MASK;
+ strsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
+ strsect, info->index.str);
pr_debug("\t%s\n", info->secstrings + strsect->sh_name);
/* We'll tack temporary mod_kallsyms on the end. */
- mod->init_layout.size = ALIGN(mod->init_layout.size,
+ mod_mem_init_data->size = ALIGN(mod_mem_init_data->size,
__alignof__(struct mod_kallsyms));
- info->mod_kallsyms_init_off = mod->init_layout.size;
- mod->init_layout.size += sizeof(struct mod_kallsyms);
- info->init_typeoffs = mod->init_layout.size;
- mod->init_layout.size += nsrc * sizeof(char);
- mod->init_layout.size = strict_align(mod->init_layout.size);
+ info->mod_kallsyms_init_off = mod_mem_init_data->size;
+
+ mod_mem_init_data->size += sizeof(struct mod_kallsyms);
+ info->init_typeoffs = mod_mem_init_data->size;
+ mod_mem_init_data->size += nsrc * sizeof(char);
}
/*
@@ -171,9 +174,11 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
char *s;
Elf_Shdr *symsec = &info->sechdrs[info->index.sym];
unsigned long strtab_size;
+ void *data_base = mod->mem[MOD_DATA].base;
+ void *init_data_base = mod->mem[MOD_INIT_DATA].base;
/* Set up to point into init section. */
- mod->kallsyms = (void __rcu *)mod->init_layout.base +
+ mod->kallsyms = (void __rcu *)init_data_base +
info->mod_kallsyms_init_off;
rcu_read_lock();
@@ -183,15 +188,15 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
/* Make sure we get permanent strtab: don't use info->strtab. */
rcu_dereference(mod->kallsyms)->strtab =
(void *)info->sechdrs[info->index.str].sh_addr;
- rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs;
+ rcu_dereference(mod->kallsyms)->typetab = init_data_base + info->init_typeoffs;
/*
* Now populate the cut down core kallsyms for after init
* and set types up while we still have access to sections.
*/
- mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs;
- mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs;
- mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs;
+ mod->core_kallsyms.symtab = dst = data_base + info->symoffs;
+ mod->core_kallsyms.strtab = s = data_base + info->stroffs;
+ mod->core_kallsyms.typetab = data_base + info->core_typeoffs;
strtab_size = info->core_typeoffs - info->stroffs;
src = rcu_dereference(mod->kallsyms)->symtab;
for (ndst = i = 0; i < rcu_dereference(mod->kallsyms)->num_symtab; i++) {
@@ -267,12 +272,15 @@ static const char *find_kallsyms_symbol(struct module *mod,
unsigned int i, best = 0;
unsigned long nextval, bestval;
struct mod_kallsyms *kallsyms = rcu_dereference_sched(mod->kallsyms);
+ struct module_memory *mod_mem;
/* At worse, next value is at end of module */
if (within_module_init(addr, mod))
- nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size;
+ mod_mem = &mod->mem[MOD_INIT_TEXT];
else
- nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size;
+ mod_mem = &mod->mem[MOD_TEXT];
+
+ nextval = (unsigned long)mod_mem->base + mod_mem->size;
bestval = kallsyms_symbol_value(&kallsyms->symtab[best]);
diff --git a/kernel/module/kdb.c b/kernel/module/kdb.c
index f4317f92e189..995c32d3698f 100644
--- a/kernel/module/kdb.c
+++ b/kernel/module/kdb.c
@@ -26,10 +26,11 @@ int kdb_lsmod(int argc, const char **argv)
if (mod->state == MODULE_STATE_UNFORMED)
continue;
- kdb_printf("%-20s%8u", mod->name, mod->core_layout.size);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- kdb_printf("/%8u", mod->data_layout.size);
-#endif
+ kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
+ kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
+ kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
+ kdb_printf("/%8u", mod->mem[MOD_DATA].size);
+
kdb_printf(" 0x%px ", (void *)mod);
#ifdef CONFIG_MODULE_UNLOAD
kdb_printf("%4d ", module_refcount(mod));
@@ -40,10 +41,10 @@ int kdb_lsmod(int argc, const char **argv)
kdb_printf(" (Loading)");
else
kdb_printf(" (Live)");
- kdb_printf(" 0x%px", mod->core_layout.base);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- kdb_printf("/0x%px", mod->data_layout.base);
-#endif
+ kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
+ kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
+ kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
+ kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
#ifdef CONFIG_MODULE_UNLOAD
{
diff --git a/kernel/module/main.c b/kernel/module/main.c
index d3be89de706d..9f890d14072c 100644
--- a/kernel/module/main.c
+++ b/kernel/module/main.c
@@ -80,12 +80,6 @@ struct mod_tree_root mod_tree __cacheline_aligned = {
.addr_min = -1UL,
};
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
-struct mod_tree_root mod_data_tree __cacheline_aligned = {
- .addr_min = -1UL,
-};
-#endif
-
struct symsearch {
const struct kernel_symbol *start, *stop;
const s32 *crcs;
@@ -93,14 +87,24 @@ struct symsearch {
};
/*
- * Bounds of module text, for speeding up __module_address.
+ * Bounds of module memory, for speeding up __module_address.
* Protected by module_mutex.
*/
-static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree)
+static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base,
+ unsigned int size, struct mod_tree_root *tree)
{
unsigned long min = (unsigned long)base;
unsigned long max = min + size;
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+ if (mod_mem_type_is_core_data(type)) {
+ if (min < tree->data_addr_min)
+ tree->data_addr_min = min;
+ if (max > tree->data_addr_max)
+ tree->data_addr_max = max;
+ return;
+ }
+#endif
if (min < tree->addr_min)
tree->addr_min = min;
if (max > tree->addr_max)
@@ -109,12 +113,12 @@ static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_r
static void mod_update_bounds(struct module *mod)
{
- __mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree);
- if (mod->init_layout.size)
- __mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- __mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree);
-#endif
+ for_each_mod_mem_type(type) {
+ struct module_memory *mod_mem = &mod->mem[type];
+
+ if (mod_mem->size)
+ __mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree);
+ }
}
/* Block module loading/unloading? */
@@ -923,12 +927,27 @@ static ssize_t store_uevent(struct module_attribute *mattr,
struct module_attribute module_uevent =
__ATTR(uevent, 0200, NULL, store_uevent);
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+
static ssize_t show_coresize(struct module_attribute *mattr,
struct module_kobject *mk, char *buffer)
{
- return sprintf(buffer, "%u\n", mk->mod->core_layout.size);
+ return sprintf(buffer, "%u\n", mk->mod->mem[MOD_TEXT].size);
}
+#else
+
+static ssize_t show_coresize(struct module_attribute *mattr,
+ struct module_kobject *mk, char *buffer)
+{
+ unsigned int size = 0;
+
+ for_class_mod_mem_type(type, core)
+ size += mk->mod->mem[type].size;
+ return sprintf(buffer, "%u\n", size);
+}
+#endif
+
static struct module_attribute modinfo_coresize =
__ATTR(coresize, 0444, show_coresize, NULL);
@@ -936,7 +955,14 @@ static struct module_attribute modinfo_coresize =
static ssize_t show_datasize(struct module_attribute *mattr,
struct module_kobject *mk, char *buffer)
{
- return sprintf(buffer, "%u\n", mk->mod->data_layout.size);
+ unsigned int size = 0;
+
+ for_class_mod_mem_type(type, core) {
+ if (type == MOD_TEXT)
+ continue;
+ size += mk->mod->mem[type].size;
+ }
+ return sprintf(buffer, "%u\n", size);
}
static struct module_attribute modinfo_datasize =
@@ -946,7 +972,11 @@ static struct module_attribute modinfo_datasize =
static ssize_t show_initsize(struct module_attribute *mattr,
struct module_kobject *mk, char *buffer)
{
- return sprintf(buffer, "%u\n", mk->mod->init_layout.size);
+ unsigned int size = 0;
+
+ for_class_mod_mem_type(type, init)
+ size += mk->mod->mem[type].size;
+ return sprintf(buffer, "%u\n", size);
}
static struct module_attribute modinfo_initsize =
@@ -1143,6 +1173,63 @@ void __weak module_arch_freeing_init(struct module *mod)
{
}
+#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
+static bool mod_mem_use_vmalloc(enum mod_mem_type type)
+{
+ return mod_mem_is_core_data(type);
+}
+#else
+static bool mod_mem_use_vmalloc(enum mod_mem_type type)
+{
+ return false;
+}
+#endif
+
+static void *module_memory_alloc(unsigned int size, enum mod_mem_type type)
+{
+ if (mod_mem_use_vmalloc(type))
+ return vzalloc(size);
+ return module_alloc(size);
+}
+
+static void module_memory_free(void *ptr, enum mod_mem_type type)
+{
+ if (mod_mem_use_vmalloc(type))
+ vfree(ptr);
+ else
+ module_memfree(ptr);
+}
+
+static void free_mod_mem(struct module *mod)
+{
+ /* free the memory in the right order to avoid use-after-free */
+ static enum mod_mem_type mod_mem_free_order[MOD_MEM_NUM_TYPES] = {
+ /* first free init sections */
+ MOD_INIT_TEXT,
+ MOD_INIT_DATA,
+ MOD_INIT_RODATA,
+
+ /* then core sections, except rw data */
+ MOD_TEXT,
+ MOD_RODATA,
+ MOD_RO_AFTER_INIT,
+
+ /* last, rw data */
+ MOD_DATA,
+ };
+ int i;
+
+ for (i = 0; i < MOD_MEM_NUM_TYPES; i++) {
+ enum mod_mem_type type = mod_mem_free_order[i];
+ struct module_memory *mod_mem = &mod->mem[type];
+
+ /* Free lock-classes; relies on the preceding sync_rcu(). */
+ lockdep_free_key_range(mod_mem->base, mod_mem->size);
+ if (mod_mem->size)
+ module_memory_free(mod_mem->base, type);
+ }
+}
+
/* Free a module, remove from lists, etc. */
static void free_module(struct module *mod)
{
@@ -1189,18 +1276,10 @@ static void free_module(struct module *mod)
/* This may be empty, but that's OK */
module_arch_freeing_init(mod);
- module_memfree(mod->init_layout.base);
kfree(mod->args);
percpu_modfree(mod);
- /* Free lock-classes; relies on the preceding sync_rcu(). */
- lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
-
- /* Finally, free the core (containing the module structure) */
- module_memfree(mod->core_layout.base);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- vfree(mod->data_layout.base);
-#endif
+ free_mod_mem(mod);
}
void *__symbol_get(const char *symbol)
@@ -1387,16 +1466,18 @@ unsigned int __weak arch_mod_section_prepend(struct module *mod,
return 0;
}
-/* Update size with this section: return offset. */
-long module_get_offset(struct module *mod, unsigned int *size,
+long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
Elf_Shdr *sechdr, unsigned int section)
{
- long ret;
+ long offset;
+ long mask = ((unsigned long)(type) & SH_ENTSIZE_TYPE_MASK) << SH_ENTSIZE_TYPE_SHIFT;
- *size += arch_mod_section_prepend(mod, section);
- ret = ALIGN(*size, sechdr->sh_addralign ?: 1);
- *size = ret + sechdr->sh_size;
- return ret;
+ mod->mem[type].size += arch_mod_section_prepend(mod, section);
+ offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
+ mod->mem[type].size = offset + sechdr->sh_size;
+
+ WARN_ON_ONCE(offset & mask);
+ return offset | mask;
}
static bool module_init_layout_section(const char *sname)
@@ -1428,6 +1509,20 @@ static void layout_sections(struct module *mod, struct load_info *info)
{ SHF_WRITE | SHF_ALLOC, ARCH_SHF_SMALL },
{ ARCH_SHF_SMALL | SHF_ALLOC, 0 }
};
+ static int core_m_to_mem_type[] = {
+ MOD_TEXT,
+ MOD_RODATA,
+ MOD_RO_AFTER_INIT,
+ MOD_DATA,
+ MOD_INVALID,
+ };
+ static int init_m_to_mem_type[] = {
+ MOD_INIT_TEXT,
+ MOD_INIT_RODATA,
+ MOD_INVALID,
+ MOD_INIT_DATA,
+ MOD_INVALID,
+ };
unsigned int m, i;
for (i = 0; i < info->hdr->e_shnum; i++)
@@ -1435,41 +1530,30 @@ static void layout_sections(struct module *mod, struct load_info *info)
pr_debug("Core section allocation order:\n");
for (m = 0; m < ARRAY_SIZE(masks); ++m) {
+ enum mod_mem_type type = core_m_to_mem_type[m];
+
for (i = 0; i < info->hdr->e_shnum; ++i) {
Elf_Shdr *s = &info->sechdrs[i];
const char *sname = info->secstrings + s->sh_name;
- unsigned int *sizep;
if ((s->sh_flags & masks[m][0]) != masks[m][0]
|| (s->sh_flags & masks[m][1])
|| s->sh_entsize != ~0UL
|| module_init_layout_section(sname))
continue;
- sizep = m ? &mod->data_layout.size : &mod->core_layout.size;
- s->sh_entsize = module_get_offset(mod, sizep, s, i);
+
+ if (WARN_ON_ONCE(type == MOD_INVALID))
+ continue;
+
+ s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
pr_debug("\t%s\n", sname);
}
- switch (m) {
- case 0: /* executable */
- mod->core_layout.size = strict_align(mod->core_layout.size);
- mod->core_layout.text_size = mod->core_layout.size;
- break;
- case 1: /* RO: text and ro-data */
- mod->data_layout.size = strict_align(mod->data_layout.size);
- mod->data_layout.ro_size = mod->data_layout.size;
- break;
- case 2: /* RO after init */
- mod->data_layout.size = strict_align(mod->data_layout.size);
- mod->data_layout.ro_after_init_size = mod->data_layout.size;
- break;
- case 4: /* whole core */
- mod->data_layout.size = strict_align(mod->data_layout.size);
- break;
- }
}
pr_debug("Init section allocation order:\n");
for (m = 0; m < ARRAY_SIZE(masks); ++m) {
+ enum mod_mem_type type = init_m_to_mem_type[m];
+
for (i = 0; i < info->hdr->e_shnum; ++i) {
Elf_Shdr *s = &info->sechdrs[i];
const char *sname = info->secstrings + s->sh_name;
@@ -1479,30 +1563,13 @@ static void layout_sections(struct module *mod, struct load_info *info)
|| s->sh_entsize != ~0UL
|| !module_init_layout_section(sname))
continue;
- s->sh_entsize = (module_get_offset(mod, &mod->init_layout.size, s, i)
- | INIT_OFFSET_MASK);
+
+ if (WARN_ON_ONCE(type == MOD_INVALID))
+ continue;
+
+ s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
pr_debug("\t%s\n", sname);
}
- switch (m) {
- case 0: /* executable */
- mod->init_layout.size = strict_align(mod->init_layout.size);
- mod->init_layout.text_size = mod->init_layout.size;
- break;
- case 1: /* RO: text and ro-data */
- mod->init_layout.size = strict_align(mod->init_layout.size);
- mod->init_layout.ro_size = mod->init_layout.size;
- break;
- case 2:
- /*
- * RO after init doesn't apply to init_layout (only
- * core_layout), so it just takes the value of ro_size.
- */
- mod->init_layout.ro_after_init_size = mod->init_layout.ro_size;
- break;
- case 4: /* whole init */
- mod->init_layout.size = strict_align(mod->init_layout.size);
- break;
- }
}
}
@@ -2122,72 +2189,42 @@ static int move_module(struct module *mod, struct load_info *info)
{
int i;
void *ptr;
+ enum mod_mem_type t;
- /* Do the allocs. */
- ptr = module_alloc(mod->core_layout.size);
- /*
- * The pointer to this block is stored in the module structure
- * which is inside the block. Just mark it as not being a
- * leak.
- */
- kmemleak_not_leak(ptr);
- if (!ptr)
- return -ENOMEM;
-
- memset(ptr, 0, mod->core_layout.size);
- mod->core_layout.base = ptr;
-
- if (mod->init_layout.size) {
- ptr = module_alloc(mod->init_layout.size);
- /*
- * The pointer to this block is stored in the module structure
- * which is inside the block. This block doesn't need to be
- * scanned as it contains data and code that will be freed
- * after the module is initialized.
- */
- kmemleak_ignore(ptr);
- if (!ptr) {
- module_memfree(mod->core_layout.base);
- return -ENOMEM;
+ for_each_mod_mem_type(type) {
+ if (!mod->mem[type].size) {
+ mod->mem[type].base = NULL;
+ continue;
}
- memset(ptr, 0, mod->init_layout.size);
- mod->init_layout.base = ptr;
- } else
- mod->init_layout.base = NULL;
+ mod->mem[type].size = PAGE_ALIGN(mod->mem[type].size);
+ ptr = module_memory_alloc(mod->mem[type].size, type);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- /* Do the allocs. */
- ptr = vzalloc(mod->data_layout.size);
/*
* The pointer to this block is stored in the module structure
* which is inside the block. Just mark it as not being a
* leak.
*/
- kmemleak_not_leak(ptr);
+ kmemleak_ignore(ptr);
if (!ptr) {
- module_memfree(mod->core_layout.base);
- module_memfree(mod->init_layout.base);
- return -ENOMEM;
+ t = type;
+ goto out_enomem;
+ }
+ memset(ptr, 0, mod->mem[type].size);
+ mod->mem[type].base = ptr;
}
- mod->data_layout.base = ptr;
-#endif
/* Transfer each section which specifies SHF_ALLOC */
pr_debug("final section addresses:\n");
for (i = 0; i < info->hdr->e_shnum; i++) {
void *dest;
Elf_Shdr *shdr = &info->sechdrs[i];
+ enum mod_mem_type type = shdr->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
if (!(shdr->sh_flags & SHF_ALLOC))
continue;
- if (shdr->sh_entsize & INIT_OFFSET_MASK)
- dest = mod->init_layout.base
- + (shdr->sh_entsize & ~INIT_OFFSET_MASK);
- else if (!(shdr->sh_flags & SHF_EXECINSTR))
- dest = mod->data_layout.base + shdr->sh_entsize;
- else
- dest = mod->core_layout.base + shdr->sh_entsize;
+ dest = mod->mem[type].base +
+ (shdr->sh_entsize & SH_ENTSIZE_OFFSET_MASK);
if (shdr->sh_type != SHT_NOBITS)
memcpy(dest, (void *)shdr->sh_addr, shdr->sh_size);
@@ -2198,6 +2235,10 @@ static int move_module(struct module *mod, struct load_info *info)
}
return 0;
+out_enomem:
+ for (t--; t >= 0; t--)
+ module_memory_free(mod->mem[t].base, t);
+ return -ENOMEM;
}
static int check_module_license_and_versions(struct module *mod)
@@ -2242,12 +2283,13 @@ static void flush_module_icache(const struct module *mod)
* Do it before processing of module parameters, so the module
* can provide parameter accessor functions of its own.
*/
- if (mod->init_layout.base)
- flush_icache_range((unsigned long)mod->init_layout.base,
- (unsigned long)mod->init_layout.base
- + mod->init_layout.size);
- flush_icache_range((unsigned long)mod->core_layout.base,
- (unsigned long)mod->core_layout.base + mod->core_layout.size);
+ for_each_mod_mem_type(type) {
+ const struct module_memory *mod_mem = &mod->mem[type];
+
+ if (mod_mem->size)
+ flush_icache_range((unsigned long)mod_mem->base,
+ (unsigned long)mod_mem->base + mod_mem->size);
+ }
}
bool __weak module_elf_check_arch(Elf_Ehdr *hdr)
@@ -2350,11 +2392,8 @@ static void module_deallocate(struct module *mod, struct load_info *info)
{
percpu_modfree(mod);
module_arch_freeing_init(mod);
- module_memfree(mod->init_layout.base);
- module_memfree(mod->core_layout.base);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- vfree(mod->data_layout.base);
-#endif
+
+ free_mod_mem(mod);
}
int __weak module_finalize(const Elf_Ehdr *hdr,
@@ -2415,7 +2454,9 @@ static void do_mod_ctors(struct module *mod)
/* For freeing module_init on success, in case kallsyms traversing */
struct mod_initfree {
struct llist_node node;
- void *module_init;
+ void *init_text;
+ void *init_data;
+ void *init_rodata;
};
static void do_free_init(struct work_struct *w)
@@ -2429,7 +2470,9 @@ static void do_free_init(struct work_struct *w)
llist_for_each_safe(pos, n, list) {
initfree = container_of(pos, struct mod_initfree, node);
- module_memfree(initfree->module_init);
+ module_memfree(initfree->init_text);
+ module_memfree(initfree->init_data);
+ module_memfree(initfree->init_rodata);
kfree(initfree);
}
}
@@ -2456,7 +2499,9 @@ static noinline int do_init_module(struct module *mod)
ret = -ENOMEM;
goto fail;
}
- freeinit->module_init = mod->init_layout.base;
+ freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
+ freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
+ freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
do_mod_ctors(mod);
/* Start the module */
@@ -2492,8 +2537,8 @@ static noinline int do_init_module(struct module *mod)
if (!mod->async_probe_requested)
async_synchronize_full();
- ftrace_free_mem(mod, mod->init_layout.base, mod->init_layout.base +
- mod->init_layout.size);
+ ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
+ mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
mutex_lock(&module_mutex);
/* Drop initial reference. */
module_put(mod);
@@ -2505,11 +2550,10 @@ static noinline int do_init_module(struct module *mod)
module_enable_ro(mod, true);
mod_tree_remove_init(mod);
module_arch_freeing_init(mod);
- mod->init_layout.base = NULL;
- mod->init_layout.size = 0;
- mod->init_layout.ro_size = 0;
- mod->init_layout.ro_after_init_size = 0;
- mod->init_layout.text_size = 0;
+ for_class_mod_mem_type(type, init) {
+ mod->mem[type].base = NULL;
+ mod->mem[type].size = 0;
+ }
#ifdef CONFIG_DEBUG_INFO_BTF_MODULES
/* .BTF is not SHF_ALLOC and will get removed, so sanitize pointer */
mod->btf_data = NULL;
@@ -2628,9 +2672,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
module_bug_finalize(info->hdr, info->sechdrs, mod);
module_cfi_finalize(info->hdr, info->sechdrs, mod);
- if (module_check_misalignment(mod))
- goto out_misaligned;
-
module_enable_ro(mod, false);
module_enable_nx(mod);
module_enable_x(mod);
@@ -2644,8 +2685,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
return 0;
-out_misaligned:
- err = -EINVAL;
out:
mutex_unlock(&module_mutex);
return err;
@@ -2909,7 +2948,10 @@ static int load_module(struct load_info *info, const char __user *uargs,
mutex_unlock(&module_mutex);
free_module:
/* Free lock-classes; relies on the preceding sync_rcu() */
- lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
+ for_class_mod_mem_type(type, core) {
+ lockdep_free_key_range(mod->mem[type].base,
+ mod->mem[type].size);
+ }
module_deallocate(mod, info);
free_copy:
@@ -3060,20 +3102,21 @@ bool is_module_address(unsigned long addr)
struct module *__module_address(unsigned long addr)
{
struct module *mod;
- struct mod_tree_root *tree;
if (addr >= mod_tree.addr_min && addr <= mod_tree.addr_max)
- tree = &mod_tree;
+ goto lookup;
+
#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- else if (addr >= mod_data_tree.addr_min && addr <= mod_data_tree.addr_max)
- tree = &mod_data_tree;
+ if (addr >= mod_tree.data_addr_min && addr <= mod_tree.data_addr_max)
+ goto lookup;
#endif
- else
+
return NULL;
+lookup:
module_assert_mutex_or_preempt();
- mod = mod_find(addr, tree);
+ mod = mod_find(addr, &mod_tree);
if (mod) {
BUG_ON(!within_module(addr, mod));
if (mod->state == MODULE_STATE_UNFORMED)
@@ -3113,8 +3156,8 @@ struct module *__module_text_address(unsigned long addr)
struct module *mod = __module_address(addr);
if (mod) {
/* Make sure it's within the text section. */
- if (!within(addr, mod->init_layout.base, mod->init_layout.text_size)
- && !within(addr, mod->core_layout.base, mod->core_layout.text_size))
+ if (!within_module_mem_type(addr, mod, MOD_TEXT) &&
+ !within_module_mem_type(addr, mod, MOD_INIT_TEXT))
mod = NULL;
}
return mod;
diff --git a/kernel/module/procfs.c b/kernel/module/procfs.c
index cf5b9f1e6ec4..0a4841e88adb 100644
--- a/kernel/module/procfs.c
+++ b/kernel/module/procfs.c
@@ -62,6 +62,15 @@ static void m_stop(struct seq_file *m, void *p)
mutex_unlock(&module_mutex);
}
+static unsigned int module_total_size(struct module *mod)
+{
+ int size = 0;
+
+ for_each_mod_mem_type(type)
+ size += mod->mem[type].size;
+ return size;
+}
+
static int m_show(struct seq_file *m, void *p)
{
struct module *mod = list_entry(p, struct module, list);
@@ -73,10 +82,7 @@ static int m_show(struct seq_file *m, void *p)
if (mod->state == MODULE_STATE_UNFORMED)
return 0;
- size = mod->init_layout.size + mod->core_layout.size;
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- size += mod->data_layout.size;
-#endif
+ size = module_total_size(mod);
seq_printf(m, "%s %u", mod->name, size);
print_unload_info(m, mod);
@@ -86,7 +92,7 @@ static int m_show(struct seq_file *m, void *p)
mod->state == MODULE_STATE_COMING ? "Loading" :
"Live");
/* Used by oprofile and other similar tools. */
- value = m->private ? NULL : mod->core_layout.base;
+ value = m->private ? NULL : mod->mem[MOD_TEXT].base;
seq_printf(m, " 0x%px", value);
/* Taints info */
diff --git a/kernel/module/strict_rwx.c b/kernel/module/strict_rwx.c
index 14fbea66f12f..2e79a77f40eb 100644
--- a/kernel/module/strict_rwx.c
+++ b/kernel/module/strict_rwx.c
@@ -11,82 +11,26 @@
#include <linux/set_memory.h>
#include "internal.h"
-/*
- * LKM RO/NX protection: protect module's text/ro-data
- * from modification and any data from execution.
- *
- * General layout of module is:
- * [text] [read-only-data] [ro-after-init] [writable data]
- * text_size -----^ ^ ^ ^
- * ro_size ------------------------| | |
- * ro_after_init_size -----------------------------| |
- * size -----------------------------------------------------------|
- *
- * These values are always page-aligned (as is base) when
- * CONFIG_STRICT_MODULE_RWX is set.
- */
+static void module_set_memory(
+ const struct module *mod, enum mod_mem_type type,
+ int (*set_memory)(unsigned long start, int num_pages))
+{
+ const struct module_memory *mod_mem = &mod->mem[type];
+
+ set_vm_flush_reset_perms(mod_mem->base);
+ set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
+}
/*
* Since some arches are moving towards PAGE_KERNEL module allocations instead
- * of PAGE_KERNEL_EXEC, keep frob_text() and module_enable_x() independent of
+ * of PAGE_KERNEL_EXEC, keep module_enable_x() independent of
* CONFIG_STRICT_MODULE_RWX because they are needed regardless of whether we
* are strict.
*/
-static void frob_text(const struct module_layout *layout,
- int (*set_memory)(unsigned long start, int num_pages))
-{
- set_memory((unsigned long)layout->base,
- PAGE_ALIGN(layout->text_size) >> PAGE_SHIFT);
-}
-
-static void frob_rodata(const struct module_layout *layout,
- int (*set_memory)(unsigned long start, int num_pages))
-{
- set_memory((unsigned long)layout->base + layout->text_size,
- (layout->ro_size - layout->text_size) >> PAGE_SHIFT);
-}
-
-static void frob_ro_after_init(const struct module_layout *layout,
- int (*set_memory)(unsigned long start, int num_pages))
-{
- set_memory((unsigned long)layout->base + layout->ro_size,
- (layout->ro_after_init_size - layout->ro_size) >> PAGE_SHIFT);
-}
-
-static void frob_writable_data(const struct module_layout *layout,
- int (*set_memory)(unsigned long start, int num_pages))
-{
- set_memory((unsigned long)layout->base + layout->ro_after_init_size,
- (layout->size - layout->ro_after_init_size) >> PAGE_SHIFT);
-}
-
-static bool layout_check_misalignment(const struct module_layout *layout)
-{
- return WARN_ON(!PAGE_ALIGNED(layout->base)) ||
- WARN_ON(!PAGE_ALIGNED(layout->text_size)) ||
- WARN_ON(!PAGE_ALIGNED(layout->ro_size)) ||
- WARN_ON(!PAGE_ALIGNED(layout->ro_after_init_size)) ||
- WARN_ON(!PAGE_ALIGNED(layout->size));
-}
-
-bool module_check_misalignment(const struct module *mod)
-{
- if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
- return false;
-
- return layout_check_misalignment(&mod->core_layout) ||
- layout_check_misalignment(&mod->data_layout) ||
- layout_check_misalignment(&mod->init_layout);
-}
-
void module_enable_x(const struct module *mod)
{
- if (!PAGE_ALIGNED(mod->core_layout.base) ||
- !PAGE_ALIGNED(mod->init_layout.base))
- return;
-
- frob_text(&mod->core_layout, set_memory_x);
- frob_text(&mod->init_layout, set_memory_x);
+ for_class_mod_mem_type(type, text)
+ module_set_memory(mod, type, set_memory_x);
}
void module_enable_ro(const struct module *mod, bool after_init)
@@ -98,16 +42,13 @@ void module_enable_ro(const struct module *mod, bool after_init)
return;
#endif
- set_vm_flush_reset_perms(mod->core_layout.base);
- set_vm_flush_reset_perms(mod->init_layout.base);
- frob_text(&mod->core_layout, set_memory_ro);
-
- frob_rodata(&mod->data_layout, set_memory_ro);
- frob_text(&mod->init_layout, set_memory_ro);
- frob_rodata(&mod->init_layout, set_memory_ro);
+ module_set_memory(mod, MOD_TEXT, set_memory_ro);
+ module_set_memory(mod, MOD_INIT_TEXT, set_memory_ro);
+ module_set_memory(mod, MOD_RODATA, set_memory_ro);
+ module_set_memory(mod, MOD_INIT_RODATA, set_memory_ro);
if (after_init)
- frob_ro_after_init(&mod->data_layout, set_memory_ro);
+ module_set_memory(mod, MOD_RO_AFTER_INIT, set_memory_ro);
}
void module_enable_nx(const struct module *mod)
@@ -115,11 +56,8 @@ void module_enable_nx(const struct module *mod)
if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
return;
- frob_rodata(&mod->data_layout, set_memory_nx);
- frob_ro_after_init(&mod->data_layout, set_memory_nx);
- frob_writable_data(&mod->data_layout, set_memory_nx);
- frob_rodata(&mod->init_layout, set_memory_nx);
- frob_writable_data(&mod->init_layout, set_memory_nx);
+ for_class_mod_mem_type(type, data)
+ module_set_memory(mod, type, set_memory_nx);
}
int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
diff --git a/kernel/module/tree_lookup.c b/kernel/module/tree_lookup.c
index 8ec5cfd60496..277197977d43 100644
--- a/kernel/module/tree_lookup.c
+++ b/kernel/module/tree_lookup.c
@@ -21,16 +21,16 @@
static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n)
{
- struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
+ struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
- return (unsigned long)layout->base;
+ return (unsigned long)mod_mem->base;
}
static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n)
{
- struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
+ struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
- return (unsigned long)layout->size;
+ return (unsigned long)mod_mem->size;
}
static __always_inline bool
@@ -77,32 +77,27 @@ static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *
*/
void mod_tree_insert(struct module *mod)
{
- mod->core_layout.mtn.mod = mod;
- mod->init_layout.mtn.mod = mod;
-
- __mod_tree_insert(&mod->core_layout.mtn, &mod_tree);
- if (mod->init_layout.size)
- __mod_tree_insert(&mod->init_layout.mtn, &mod_tree);
-
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- mod->data_layout.mtn.mod = mod;
- __mod_tree_insert(&mod->data_layout.mtn, &mod_data_tree);
-#endif
+ for_each_mod_mem_type(type) {
+ mod->mem[type].mtn.mod = mod;
+ if (mod->mem[type].size)
+ __mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
+ }
}
void mod_tree_remove_init(struct module *mod)
{
- if (mod->init_layout.size)
- __mod_tree_remove(&mod->init_layout.mtn, &mod_tree);
+ for_class_mod_mem_type(type, init) {
+ if (mod->mem[type].size)
+ __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
+ }
}
void mod_tree_remove(struct module *mod)
{
- __mod_tree_remove(&mod->core_layout.mtn, &mod_tree);
- mod_tree_remove_init(mod);
-#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
- __mod_tree_remove(&mod->data_layout.mtn, &mod_data_tree);
-#endif
+ for_each_mod_mem_type(type) {
+ if (mod->mem[type].size)
+ __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
+ }
}
struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)
--
2.30.2
Hi,
On what tree / branch does this patch applies ?
I have tried with modules/modules-next, linus/master, with v6.1. It
conflicts on all of them.
'git am -3' even says that 'It does not apply to blobs recorded in its
index.'
Christophe
Le 01/02/2023 à 18:36, Song Liu a écrit :
> module_layout manages different types of memory (text, data, rodata, etc.)
> in one allocation, which is problematic for some reasons:
>
> 1. It is hard to enable CONFIG_STRICT_MODULE_RWX.
> 2. It is hard to use huge pages in modules (and not break strict rwx).
> 3. Many archs uses module_layout for arch-specific data, but it is not
> obvious how these data are used (are they RO, RX, or RW?)
>
> Improve the scenario by replacing 2 (or 3) module_layout per module with
> up to 7 module_memory per module:
>
> MOD_TEXT,
> MOD_DATA,
> MOD_RODATA,
> MOD_RO_AFTER_INIT,
> MOD_INIT_TEXT,
> MOD_INIT_DATA,
> MOD_INIT_RODATA,
>
> and allocating them separately. This adds slightly more entries to
> mod_tree (from up to 3 entries per module, to up to 7 entries per
> module). However, this at most adds a small constant overhead to
> __module_address(), which is expected to be fast.
>
> Various archs use module_layout for different data. These data are put
> into different module_memory based on their location in module_layout.
> IOW, data that used to go with text is allocated with MOD_MEM_TYPE_TEXT;
> data that used to go with data is allocated with MOD_MEM_TYPE_DATA, etc.
>
> module_memory simplifies quite some of the module code. For example,
> ARCH_WANTS_MODULES_DATA_IN_VMALLOC is a lot cleaner, as it just uses a
> different allocator for the data. kernel/module/strict_rwx.c is also
> much cleaner with module_memory.
>
> Signed-off-by: Song Liu <[email protected]>
> Cc: Luis Chamberlain <[email protected]>
> Cc: Thomas Gleixner <[email protected]>
> Cc: Peter Zijlstra <[email protected]>
> Cc: Guenter Roeck <[email protected]>
> Cc: Christophe Leroy <[email protected]>
>
> ---
>
> This is the preparation work for the type aware module_alloc() discussed
> in [1]. While this work is not covered much in the discussion, it is a
> critical step of the effort.
>
> As this part grows pretty big (~1000 lines, + and -), I would like get
> some feedback on it, so that I know it is on the right track.
>
> Please share your comments. Thanks!
>
> Test coverage: Tested on x86_64.
> Build tested by kernel test bot in [2]. The only regression in [2] was a
> typo in parisc, which is also fixed.
>
> [1] https://lore.kernel.org/linux-mm/[email protected]/T/#u
> [2] https://lore.kernel.org/linux-raid/63b8827e.clJQX2wg+I+tiX7m%[email protected]/T/#u
>
> Changes v1 => v2:
> 1. Add data_addr_[min|max] for CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> case.
>
> Changes v2 => v3:
> 1. Fix and remove the last use of INIT_OFFSET_MASK.
> 2. Add more information in the commit log. (Luis Chamberlain)
> 3. Rebase and fix issues in x86/calltrunks.
> 4. Minor cleanup.
>
> Changes v3 => v4:
> 1. Shorten enum/variable names, so that the code are easier to read.
> (Christophe Leroy)
> 2. Remove an used variable. (Guenter Roeck, Christophe Leroy)
>
> Changes v4 => v5:
> 1. Simplify some code some code. (Peter Zijlstra, Christophe Leroy)
> 2. Remove module_check_misalignment(), which is not useful any more.
>
> Changes v5 => v6:
> 1. Improve mod_mem_type_is_* and for_*mod_mem_type marcos.
> (Peter Zijlstra).
> ---
> arch/alpha/kernel/module.c | 2 +-
> arch/arc/kernel/unwind.c | 9 +-
> arch/arm/kernel/module-plts.c | 2 +-
> arch/arm64/kernel/module-plts.c | 2 +-
> arch/ia64/kernel/module.c | 24 ++-
> arch/mips/kernel/vpe.c | 11 +-
> arch/parisc/kernel/module.c | 51 ++---
> arch/powerpc/kernel/module_32.c | 7 +-
> arch/s390/kernel/module.c | 26 +--
> arch/x86/kernel/callthunks.c | 4 +-
> arch/x86/kernel/module.c | 4 +-
> include/linux/module.h | 86 +++++---
> kernel/module/internal.h | 40 ++--
> kernel/module/kallsyms.c | 56 ++---
> kernel/module/kdb.c | 17 +-
> kernel/module/main.c | 355 ++++++++++++++++++--------------
> kernel/module/procfs.c | 16 +-
> kernel/module/strict_rwx.c | 100 ++-------
> kernel/module/tree_lookup.c | 39 ++--
> 19 files changed, 432 insertions(+), 419 deletions(-)
>
> diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c
> index 5b60c248de9e..9109213abc09 100644
> --- a/arch/alpha/kernel/module.c
> +++ b/arch/alpha/kernel/module.c
> @@ -148,7 +148,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab,
>
> /* The small sections were sorted to the end of the segment.
> The following should definitely cover them. */
> - gp = (u64)me->core_layout.base + me->core_layout.size - 0x8000;
> + gp = (u64)me->mem[MOD_DATA].base + me->mem[MOD_DATA].size - 0x8000;
> got = sechdrs[me->arch.gotsecindex].sh_addr;
>
> for (i = 0; i < n; i++) {
> diff --git a/arch/arc/kernel/unwind.c b/arch/arc/kernel/unwind.c
> index 200270a94558..933451f4494f 100644
> --- a/arch/arc/kernel/unwind.c
> +++ b/arch/arc/kernel/unwind.c
> @@ -369,6 +369,8 @@ void *unwind_add_table(struct module *module, const void *table_start,
> unsigned long table_size)
> {
> struct unwind_table *table;
> + struct module_memory *mod_mem_core_text;
> + struct module_memory *mod_mem_init_text;
>
> if (table_size <= 0)
> return NULL;
> @@ -377,9 +379,12 @@ void *unwind_add_table(struct module *module, const void *table_start,
> if (!table)
> return NULL;
>
> + mod_mem_core_text = &module->mem[MOD_TEXT];
> + mod_mem_init_text = &module->mem[MOD_INIT_TEXT];
> +
> init_unwind_table(table, module->name,
> - module->core_layout.base, module->core_layout.size,
> - module->init_layout.base, module->init_layout.size,
> + mod_mem_core_text->base, mod_mem_core_text->size,
> + mod_mem_init_text->base, mod_mem_init_text->size,
> table_start, table_size,
> NULL, 0);
>
> diff --git a/arch/arm/kernel/module-plts.c b/arch/arm/kernel/module-plts.c
> index af7c322ebed6..9d4ecb6b1412 100644
> --- a/arch/arm/kernel/module-plts.c
> +++ b/arch/arm/kernel/module-plts.c
> @@ -30,7 +30,7 @@ static const u32 fixed_plts[] = {
>
> static bool in_init(const struct module *mod, unsigned long loc)
> {
> - return loc - (u32)mod->init_layout.base < mod->init_layout.size;
> + return within_module_init(loc, mod);
> }
>
> static void prealloc_fixed(struct mod_plt_sec *pltsec, struct plt_entries *plt)
> diff --git a/arch/arm64/kernel/module-plts.c b/arch/arm64/kernel/module-plts.c
> index 5a0a8f552a61..fedffa83a283 100644
> --- a/arch/arm64/kernel/module-plts.c
> +++ b/arch/arm64/kernel/module-plts.c
> @@ -67,7 +67,7 @@ static bool plt_entries_equal(const struct plt_entry *a,
>
> static bool in_init(const struct module *mod, void *loc)
> {
> - return (u64)loc - (u64)mod->init_layout.base < mod->init_layout.size;
> + return within_module_init(loc, mod);
> }
>
> u64 module_emit_plt_entry(struct module *mod, Elf64_Shdr *sechdrs,
> diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c
> index 8f62cf97f691..3661135da9d9 100644
> --- a/arch/ia64/kernel/module.c
> +++ b/arch/ia64/kernel/module.c
> @@ -485,19 +485,19 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
> return 0;
> }
>
> -static inline int
> +static inline bool
> in_init (const struct module *mod, uint64_t addr)
> {
> - return addr - (uint64_t) mod->init_layout.base < mod->init_layout.size;
> + return within_module_init(addr, mod);
> }
>
> -static inline int
> +static inline bool
> in_core (const struct module *mod, uint64_t addr)
> {
> - return addr - (uint64_t) mod->core_layout.base < mod->core_layout.size;
> + return within_module_core(addr, mod);
> }
>
> -static inline int
> +static inline bool
> is_internal (const struct module *mod, uint64_t value)
> {
> return in_init(mod, value) || in_core(mod, value);
> @@ -677,7 +677,8 @@ do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend,
> break;
>
> case RV_BDREL:
> - val -= (uint64_t) (in_init(mod, val) ? mod->init_layout.base : mod->core_layout.base);
> + val -= (uint64_t) (in_init(mod, val) ? mod->mem[MOD_INIT_TEXT].base :
> + mod->mem[MOD_TEXT].base);
> break;
>
> case RV_LTV:
> @@ -812,15 +813,18 @@ apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind
> * addresses have been selected...
> */
> uint64_t gp;
> - if (mod->core_layout.size > MAX_LTOFF)
> + struct module_memory *mod_mem;
> +
> + mod_mem = &mod->mem[MOD_DATA];
> + if (mod_mem->size > MAX_LTOFF)
> /*
> * This takes advantage of fact that SHF_ARCH_SMALL gets allocated
> * at the end of the module.
> */
> - gp = mod->core_layout.size - MAX_LTOFF / 2;
> + gp = mod_mem->size - MAX_LTOFF / 2;
> else
> - gp = mod->core_layout.size / 2;
> - gp = (uint64_t) mod->core_layout.base + ((gp + 7) & -8);
> + gp = mod_mem->size / 2;
> + gp = (uint64_t) mod_mem->base + ((gp + 7) & -8);
> mod->arch.gp = gp;
> DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
> }
> diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c
> index 13294972707b..ab114c686f9d 100644
> --- a/arch/mips/kernel/vpe.c
> +++ b/arch/mips/kernel/vpe.c
> @@ -199,18 +199,17 @@ static void layout_sections(struct module *mod, const Elf_Ehdr *hdr,
> for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> for (i = 0; i < hdr->e_shnum; ++i) {
> Elf_Shdr *s = &sechdrs[i];
> + struct module_memory *mod_mem;
> +
> + mod_mem = &mod->mem[MOD_TEXT];
>
> if ((s->sh_flags & masks[m][0]) != masks[m][0]
> || (s->sh_flags & masks[m][1])
> || s->sh_entsize != ~0UL)
> continue;
> s->sh_entsize =
> - get_offset((unsigned long *)&mod->core_layout.size, s);
> + get_offset((unsigned long *)&mod_mem->size, s);
> }
> -
> - if (m == 0)
> - mod->core_layout.text_size = mod->core_layout.size;
> -
> }
> }
>
> @@ -641,7 +640,7 @@ static int vpe_elfload(struct vpe *v)
> layout_sections(&mod, hdr, sechdrs, secstrings);
> }
>
> - v->load_addr = alloc_progmem(mod.core_layout.size);
> + v->load_addr = alloc_progmem(mod.mod_mem[MOD_TEXT].size);
> if (!v->load_addr)
> return -ENOMEM;
>
> diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c
> index 7df140545b22..f6e38c4d3904 100644
> --- a/arch/parisc/kernel/module.c
> +++ b/arch/parisc/kernel/module.c
> @@ -27,9 +27,9 @@
> * We are not doing SEGREL32 handling correctly. According to the ABI, we
> * should do a value offset, like this:
> * if (in_init(me, (void *)val))
> - * val -= (uint32_t)me->init_layout.base;
> + * val -= (uint32_t)me->mem[MOD_INIT_TEXT].base;
> * else
> - * val -= (uint32_t)me->core_layout.base;
> + * val -= (uint32_t)me->mem[MOD_TEXT].base;
> * However, SEGREL32 is used only for PARISC unwind entries, and we want
> * those entries to have an absolute address, and not just an offset.
> *
> @@ -76,25 +76,6 @@
> * allows us to allocate up to 4095 GOT entries. */
> #define MAX_GOTS 4095
>
> -/* three functions to determine where in the module core
> - * or init pieces the location is */
> -static inline int in_init(struct module *me, void *loc)
> -{
> - return (loc >= me->init_layout.base &&
> - loc <= (me->init_layout.base + me->init_layout.size));
> -}
> -
> -static inline int in_core(struct module *me, void *loc)
> -{
> - return (loc >= me->core_layout.base &&
> - loc <= (me->core_layout.base + me->core_layout.size));
> -}
> -
> -static inline int in_local(struct module *me, void *loc)
> -{
> - return in_init(me, loc) || in_core(me, loc);
> -}
> -
> #ifndef CONFIG_64BIT
> struct got_entry {
> Elf32_Addr addr;
> @@ -302,6 +283,7 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
> {
> unsigned long gots = 0, fdescs = 0, len;
> unsigned int i;
> + struct module_memory *mod_mem;
>
> len = hdr->e_shnum * sizeof(me->arch.section[0]);
> me->arch.section = kzalloc(len, GFP_KERNEL);
> @@ -346,14 +328,15 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
> me->arch.section[s].stub_entries += count;
> }
>
> + mod_mem = &me->mem[MOD_TEXT];
> /* align things a bit */
> - me->core_layout.size = ALIGN(me->core_layout.size, 16);
> - me->arch.got_offset = me->core_layout.size;
> - me->core_layout.size += gots * sizeof(struct got_entry);
> + mod_mem->size = ALIGN(mod_mem->size, 16);
> + me->arch.got_offset = mod_mem->size;
> + mod_mem->size += gots * sizeof(struct got_entry);
>
> - me->core_layout.size = ALIGN(me->core_layout.size, 16);
> - me->arch.fdesc_offset = me->core_layout.size;
> - me->core_layout.size += fdescs * sizeof(Elf_Fdesc);
> + mod_mem->size = ALIGN(mod_mem->size, 16);
> + me->arch.fdesc_offset = mod_mem->size;
> + mod_mem->size += fdescs * sizeof(Elf_Fdesc);
>
> me->arch.got_max = gots;
> me->arch.fdesc_max = fdescs;
> @@ -371,7 +354,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
>
> BUG_ON(value == 0);
>
> - got = me->core_layout.base + me->arch.got_offset;
> + got = me->mem[MOD_TEXT].base + me->arch.got_offset;
> for (i = 0; got[i].addr; i++)
> if (got[i].addr == value)
> goto out;
> @@ -389,7 +372,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
> #ifdef CONFIG_64BIT
> static Elf_Addr get_fdesc(struct module *me, unsigned long value)
> {
> - Elf_Fdesc *fdesc = me->core_layout.base + me->arch.fdesc_offset;
> + Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
>
> if (!value) {
> printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
> @@ -407,7 +390,7 @@ static Elf_Addr get_fdesc(struct module *me, unsigned long value)
>
> /* Create new one */
> fdesc->addr = value;
> - fdesc->gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> + fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
> return (Elf_Addr)fdesc;
> }
> #endif /* CONFIG_64BIT */
> @@ -742,7 +725,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
> loc, val);
> val += addend;
> /* can we reach it locally? */
> - if (in_local(me, (void *)val)) {
> + if (within_module(val, me)) {
> /* this is the case where the symbol is local
> * to the module, but in a different section,
> * so stub the jump in case it's more than 22
> @@ -801,7 +784,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
> break;
> case R_PARISC_FPTR64:
> /* 64-bit function address */
> - if(in_local(me, (void *)(val + addend))) {
> + if (within_module(val + addend, me)) {
> *loc64 = get_fdesc(me, val+addend);
> pr_debug("FDESC for %s at %llx points to %llx\n",
> strtab + sym->st_name, *loc64,
> @@ -839,7 +822,7 @@ register_unwind_table(struct module *me,
>
> table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
> end = table + sechdrs[me->arch.unwind_section].sh_size;
> - gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> + gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
>
> pr_debug("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
> me->arch.unwind_section, table, end, gp);
> @@ -977,7 +960,7 @@ void module_arch_cleanup(struct module *mod)
> #ifdef CONFIG_64BIT
> void *dereference_module_function_descriptor(struct module *mod, void *ptr)
> {
> - unsigned long start_opd = (Elf64_Addr)mod->core_layout.base +
> + unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
> mod->arch.fdesc_offset;
> unsigned long end_opd = start_opd +
> mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
> diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c
> index ea6536171778..2729179e9dc0 100644
> --- a/arch/powerpc/kernel/module_32.c
> +++ b/arch/powerpc/kernel/module_32.c
> @@ -163,8 +163,7 @@ static uint32_t do_plt_call(void *location,
>
> pr_debug("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
> /* Init, or core PLT? */
> - if (location >= mod->core_layout.base
> - && location < mod->core_layout.base + mod->core_layout.size)
> + if (within_module_init(location, mod))
> entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
> else
> entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
> @@ -322,14 +321,14 @@ notrace int module_trampoline_target(struct module *mod, unsigned long addr,
>
> int module_finalize_ftrace(struct module *module, const Elf_Shdr *sechdrs)
> {
> - module->arch.tramp = do_plt_call(module->core_layout.base,
> + module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
> (unsigned long)ftrace_caller,
> sechdrs, module);
> if (!module->arch.tramp)
> return -ENOENT;
>
> #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
> - module->arch.tramp_regs = do_plt_call(module->core_layout.base,
> + module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
> (unsigned long)ftrace_regs_caller,
> sechdrs, module);
> if (!module->arch.tramp_regs)
> diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c
> index 2d159b32885b..3ed455b785ad 100644
> --- a/arch/s390/kernel/module.c
> +++ b/arch/s390/kernel/module.c
> @@ -126,6 +126,7 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> Elf_Rela *rela;
> char *strings;
> int nrela, i, j;
> + struct module_memory *mod_mem;
>
> /* Find symbol table and string table. */
> symtab = NULL;
> @@ -173,14 +174,15 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
>
> /* Increase core size by size of got & plt and set start
> offsets for got and plt. */
> - me->core_layout.size = ALIGN(me->core_layout.size, 4);
> - me->arch.got_offset = me->core_layout.size;
> - me->core_layout.size += me->arch.got_size;
> - me->arch.plt_offset = me->core_layout.size;
> + mod_mem = &me->mem[MOD_TEXT];
> + mod_mem->size = ALIGN(mod_mem->size, 4);
> + me->arch.got_offset = mod_mem->size;
> + mod_mem->size += me->arch.got_size;
> + me->arch.plt_offset = mod_mem->size;
> if (me->arch.plt_size) {
> if (IS_ENABLED(CONFIG_EXPOLINE) && !nospec_disable)
> me->arch.plt_size += PLT_ENTRY_SIZE;
> - me->core_layout.size += me->arch.plt_size;
> + mod_mem->size += me->arch.plt_size;
> }
> return 0;
> }
> @@ -304,7 +306,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> case R_390_GOTPLT64: /* 64 bit offset to jump slot. */
> case R_390_GOTPLTENT: /* 32 bit rel. offset to jump slot >> 1. */
> if (info->got_initialized == 0) {
> - Elf_Addr *gotent = me->core_layout.base +
> + Elf_Addr *gotent = me->mem[MOD_TEXT].base +
> me->arch.got_offset +
> info->got_offset;
>
> @@ -329,7 +331,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> rc = apply_rela_bits(loc, val, 0, 64, 0, write);
> else if (r_type == R_390_GOTENT ||
> r_type == R_390_GOTPLTENT) {
> - val += (Elf_Addr) me->core_layout.base - loc;
> + val += (Elf_Addr) me->mem[MOD_TEXT].base - loc;
> rc = apply_rela_bits(loc, val, 1, 32, 1, write);
> }
> break;
> @@ -345,7 +347,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> char *plt_base;
> char *ip;
>
> - plt_base = me->core_layout.base + me->arch.plt_offset;
> + plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
> ip = plt_base + info->plt_offset;
> *(int *)insn = 0x0d10e310; /* basr 1,0 */
> *(int *)&insn[4] = 0x100c0004; /* lg 1,12(1) */
> @@ -375,7 +377,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> val - loc + 0xffffUL < 0x1ffffeUL) ||
> (r_type == R_390_PLT32DBL &&
> val - loc + 0xffffffffULL < 0x1fffffffeULL)))
> - val = (Elf_Addr) me->core_layout.base +
> + val = (Elf_Addr) me->mem[MOD_TEXT].base +
> me->arch.plt_offset +
> info->plt_offset;
> val += rela->r_addend - loc;
> @@ -397,7 +399,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> case R_390_GOTOFF32: /* 32 bit offset to GOT. */
> case R_390_GOTOFF64: /* 64 bit offset to GOT. */
> val = val + rela->r_addend -
> - ((Elf_Addr) me->core_layout.base + me->arch.got_offset);
> + ((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
> if (r_type == R_390_GOTOFF16)
> rc = apply_rela_bits(loc, val, 0, 16, 0, write);
> else if (r_type == R_390_GOTOFF32)
> @@ -407,7 +409,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> break;
> case R_390_GOTPC: /* 32 bit PC relative offset to GOT. */
> case R_390_GOTPCDBL: /* 32 bit PC rel. off. to GOT shifted by 1. */
> - val = (Elf_Addr) me->core_layout.base + me->arch.got_offset +
> + val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
> rela->r_addend - loc;
> if (r_type == R_390_GOTPC)
> rc = apply_rela_bits(loc, val, 1, 32, 0, write);
> @@ -515,7 +517,7 @@ int module_finalize(const Elf_Ehdr *hdr,
> !nospec_disable && me->arch.plt_size) {
> unsigned int *ij;
>
> - ij = me->core_layout.base + me->arch.plt_offset +
> + ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
> me->arch.plt_size - PLT_ENTRY_SIZE;
> ij[0] = 0xc6000000; /* exrl %r0,.+10 */
> ij[1] = 0x0005a7f4; /* j . */
> diff --git a/arch/x86/kernel/callthunks.c b/arch/x86/kernel/callthunks.c
> index ffea98f9064b..22ab13966427 100644
> --- a/arch/x86/kernel/callthunks.c
> +++ b/arch/x86/kernel/callthunks.c
> @@ -330,8 +330,8 @@ void noinline callthunks_patch_module_calls(struct callthunk_sites *cs,
> struct module *mod)
> {
> struct core_text ct = {
> - .base = (unsigned long)mod->core_layout.base,
> - .end = (unsigned long)mod->core_layout.base + mod->core_layout.size,
> + .base = (unsigned long)mod->mem[MOD_TEXT].base,
> + .end = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
> .name = mod->name,
> };
>
> diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c
> index 705fb2a41d7d..dbc8e226c464 100644
> --- a/arch/x86/kernel/module.c
> +++ b/arch/x86/kernel/module.c
> @@ -339,8 +339,8 @@ int module_finalize(const Elf_Ehdr *hdr,
> }
> if (locks) {
> void *lseg = (void *)locks->sh_addr;
> - void *text = me->core_layout.base;
> - void *text_end = text + me->core_layout.text_size;
> + void *text = me->mem[MOD_TEXT].base;
> + void *text_end = text + me->mem[MOD_TEXT].size;
> alternatives_smp_module_add(me, me->name,
> lseg, lseg + locks->sh_size,
> text, text_end);
> diff --git a/include/linux/module.h b/include/linux/module.h
> index 8c5909c0076c..ccc9de8a93db 100644
> --- a/include/linux/module.h
> +++ b/include/linux/module.h
> @@ -320,17 +320,47 @@ struct mod_tree_node {
> struct latch_tree_node node;
> };
>
> -struct module_layout {
> - /* The actual code + data. */
> +enum mod_mem_type {
> + MOD_TEXT = 0,
> + MOD_DATA,
> + MOD_RODATA,
> + MOD_RO_AFTER_INIT,
> + MOD_INIT_TEXT,
> + MOD_INIT_DATA,
> + MOD_INIT_RODATA,
> +
> + MOD_MEM_NUM_TYPES,
> + MOD_INVALID = -1,
> +};
> +
> +#define mod_mem_type_is_init(type) \
> + ((type) == MOD_INIT_TEXT || \
> + (type) == MOD_INIT_DATA || \
> + (type) == MOD_INIT_RODATA)
> +
> +#define mod_mem_type_is_core(type) (!mod_mem_type_is_init(type))
> +
> +#define mod_mem_type_is_text(type) \
> + ((type) == MOD_TEXT || \
> + (type) == MOD_INIT_TEXT)
> +
> +#define mod_mem_type_is_data(type) (!mod_mem_type_is_text(type))
> +
> +#define mod_mem_type_is_core_data(type) \
> + (mod_mem_type_is_core(type) && \
> + mod_mem_type_is_data(type))
> +
> +#define for_each_mod_mem_type(type) \
> + for (enum mod_mem_type (type) = 0; \
> + (type) < MOD_MEM_NUM_TYPES; (type)++)
> +
> +#define for_class_mod_mem_type(type, class) \
> + for_each_mod_mem_type(type) \
> + if (mod_mem_type_is_##class(type))
> +
> +struct module_memory {
> void *base;
> - /* Total size. */
> unsigned int size;
> - /* The size of the executable code. */
> - unsigned int text_size;
> - /* Size of RO section of the module (text+rodata) */
> - unsigned int ro_size;
> - /* Size of RO after init section */
> - unsigned int ro_after_init_size;
>
> #ifdef CONFIG_MODULES_TREE_LOOKUP
> struct mod_tree_node mtn;
> @@ -339,9 +369,9 @@ struct module_layout {
>
> #ifdef CONFIG_MODULES_TREE_LOOKUP
> /* Only touch one cacheline for common rbtree-for-core-layout case. */
> -#define __module_layout_align ____cacheline_aligned
> +#define __module_memory_align ____cacheline_aligned
> #else
> -#define __module_layout_align
> +#define __module_memory_align
> #endif
>
> struct mod_kallsyms {
> @@ -418,12 +448,8 @@ struct module {
> /* Startup function. */
> int (*init)(void);
>
> - /* Core layout: rbtree is accessed frequently, so keep together. */
> - struct module_layout core_layout __module_layout_align;
> - struct module_layout init_layout;
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - struct module_layout data_layout;
> -#endif
> + /* rbtree is accessed frequently, so keep together. */
> + struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
>
> /* Arch-specific module values */
> struct mod_arch_specific arch;
> @@ -573,23 +599,33 @@ bool __is_module_percpu_address(unsigned long addr, unsigned long *can_addr);
> bool is_module_percpu_address(unsigned long addr);
> bool is_module_text_address(unsigned long addr);
>
> +static inline bool within_module_mem_type(unsigned long addr,
> + const struct module *mod,
> + enum mod_mem_type type)
> +{
> + unsigned long base, size;
> +
> + base = (unsigned long)mod->mem[type].base;
> + size = mod->mem[type].size;
> + return addr - base < size;
> +}
> +
> static inline bool within_module_core(unsigned long addr,
> const struct module *mod)
> {
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - if ((unsigned long)mod->data_layout.base <= addr &&
> - addr < (unsigned long)mod->data_layout.base + mod->data_layout.size)
> + for_class_mod_mem_type(type, core)
> + if (within_module_mem_type(addr, mod, type))
> return true;
> -#endif
> - return (unsigned long)mod->core_layout.base <= addr &&
> - addr < (unsigned long)mod->core_layout.base + mod->core_layout.size;
> + return false;
> }
>
> static inline bool within_module_init(unsigned long addr,
> const struct module *mod)
> {
> - return (unsigned long)mod->init_layout.base <= addr &&
> - addr < (unsigned long)mod->init_layout.base + mod->init_layout.size;
> + for_class_mod_mem_type(type, init)
> + if (within_module_mem_type(addr, mod, type))
> + return true;
> + return false;
> }
>
> static inline bool within_module(unsigned long addr, const struct module *mod)
> diff --git a/kernel/module/internal.h b/kernel/module/internal.h
> index 2e2bf236f558..c6c44ceca07a 100644
> --- a/kernel/module/internal.h
> +++ b/kernel/module/internal.h
> @@ -17,27 +17,19 @@
> #define ARCH_SHF_SMALL 0
> #endif
>
> -/* If this is set, the section belongs in the init part of the module */
> -#define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG - 1))
> -/* Maximum number of characters written by module_flags() */
> -#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
> -
> -#ifndef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -#define data_layout core_layout
> -#endif
> -
> /*
> - * Modules' sections will be aligned on page boundaries
> - * to ensure complete separation of code and data, but
> - * only when CONFIG_STRICT_MODULE_RWX=y
> + * Use highest 4 bits of sh_entsize to store the mod_mem_type of this
> + * section. This leaves 28 bits for offset on 32-bit systems, which is
> + * about 256 MiB (WARN_ON_ONCE if we exceed that).
> */
> -static inline unsigned int strict_align(unsigned int size)
> -{
> - if (IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> - return PAGE_ALIGN(size);
> - else
> - return size;
> -}
> +
> +#define SH_ENTSIZE_TYPE_BITS 4
> +#define SH_ENTSIZE_TYPE_SHIFT (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)
> +#define SH_ENTSIZE_TYPE_MASK ((1UL << SH_ENTSIZE_TYPE_BITS) - 1)
> +#define SH_ENTSIZE_OFFSET_MASK ((1UL << (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)) - 1)
> +
> +/* Maximum number of characters written by module_flags() */
> +#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
>
> extern struct mutex module_mutex;
> extern struct list_head modules;
> @@ -101,8 +93,8 @@ int try_to_force_load(struct module *mod, const char *reason);
> bool find_symbol(struct find_symbol_arg *fsa);
> struct module *find_module_all(const char *name, size_t len, bool even_unformed);
> int cmp_name(const void *name, const void *sym);
> -long module_get_offset(struct module *mod, unsigned int *size, Elf_Shdr *sechdr,
> - unsigned int section);
> +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> + Elf_Shdr *sechdr, unsigned int section);
> char *module_flags(struct module *mod, char *buf, bool show_state);
> size_t module_flags_taint(unsigned long taints, char *buf);
>
> @@ -190,10 +182,13 @@ struct mod_tree_root {
> #endif
> unsigned long addr_min;
> unsigned long addr_max;
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> + unsigned long data_addr_min;
> + unsigned long data_addr_max;
> +#endif
> };
>
> extern struct mod_tree_root mod_tree;
> -extern struct mod_tree_root mod_data_tree;
>
> #ifdef CONFIG_MODULES_TREE_LOOKUP
> void mod_tree_insert(struct module *mod);
> @@ -224,7 +219,6 @@ void module_enable_nx(const struct module *mod);
> void module_enable_x(const struct module *mod);
> int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> char *secstrings, struct module *mod);
> -bool module_check_misalignment(const struct module *mod);
>
> #ifdef CONFIG_MODULE_SIG
> int module_sig_check(struct load_info *info, int flags);
> diff --git a/kernel/module/kallsyms.c b/kernel/module/kallsyms.c
> index 4523f99b0358..acdca7692961 100644
> --- a/kernel/module/kallsyms.c
> +++ b/kernel/module/kallsyms.c
> @@ -78,6 +78,7 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
> unsigned int shnum, unsigned int pcpundx)
> {
> const Elf_Shdr *sec;
> + enum mod_mem_type type;
>
> if (src->st_shndx == SHN_UNDEF ||
> src->st_shndx >= shnum ||
> @@ -90,11 +91,12 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
> #endif
>
> sec = sechdrs + src->st_shndx;
> + type = sec->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
> if (!(sec->sh_flags & SHF_ALLOC)
> #ifndef CONFIG_KALLSYMS_ALL
> || !(sec->sh_flags & SHF_EXECINSTR)
> #endif
> - || (sec->sh_entsize & INIT_OFFSET_MASK))
> + || mod_mem_type_is_init(type))
> return false;
>
> return true;
> @@ -113,11 +115,13 @@ void layout_symtab(struct module *mod, struct load_info *info)
> Elf_Shdr *strsect = info->sechdrs + info->index.str;
> const Elf_Sym *src;
> unsigned int i, nsrc, ndst, strtab_size = 0;
> + struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
> + struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
>
> /* Put symbol section at end of init part of module. */
> symsect->sh_flags |= SHF_ALLOC;
> - symsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, symsect,
> - info->index.sym) | INIT_OFFSET_MASK;
> + symsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> + symsect, info->index.sym);
> pr_debug("\t%s\n", info->secstrings + symsect->sh_name);
>
> src = (void *)info->hdr + symsect->sh_offset;
> @@ -134,28 +138,27 @@ void layout_symtab(struct module *mod, struct load_info *info)
> }
>
> /* Append room for core symbols at end of core part. */
> - info->symoffs = ALIGN(mod->data_layout.size, symsect->sh_addralign ?: 1);
> - info->stroffs = mod->data_layout.size = info->symoffs + ndst * sizeof(Elf_Sym);
> - mod->data_layout.size += strtab_size;
> + info->symoffs = ALIGN(mod_mem_data->size, symsect->sh_addralign ?: 1);
> + info->stroffs = mod_mem_data->size = info->symoffs + ndst * sizeof(Elf_Sym);
> + mod_mem_data->size += strtab_size;
> /* Note add_kallsyms() computes strtab_size as core_typeoffs - stroffs */
> - info->core_typeoffs = mod->data_layout.size;
> - mod->data_layout.size += ndst * sizeof(char);
> - mod->data_layout.size = strict_align(mod->data_layout.size);
> + info->core_typeoffs = mod_mem_data->size;
> + mod_mem_data->size += ndst * sizeof(char);
>
> /* Put string table section at end of init part of module. */
> strsect->sh_flags |= SHF_ALLOC;
> - strsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, strsect,
> - info->index.str) | INIT_OFFSET_MASK;
> + strsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> + strsect, info->index.str);
> pr_debug("\t%s\n", info->secstrings + strsect->sh_name);
>
> /* We'll tack temporary mod_kallsyms on the end. */
> - mod->init_layout.size = ALIGN(mod->init_layout.size,
> + mod_mem_init_data->size = ALIGN(mod_mem_init_data->size,
> __alignof__(struct mod_kallsyms));
> - info->mod_kallsyms_init_off = mod->init_layout.size;
> - mod->init_layout.size += sizeof(struct mod_kallsyms);
> - info->init_typeoffs = mod->init_layout.size;
> - mod->init_layout.size += nsrc * sizeof(char);
> - mod->init_layout.size = strict_align(mod->init_layout.size);
> + info->mod_kallsyms_init_off = mod_mem_init_data->size;
> +
> + mod_mem_init_data->size += sizeof(struct mod_kallsyms);
> + info->init_typeoffs = mod_mem_init_data->size;
> + mod_mem_init_data->size += nsrc * sizeof(char);
> }
>
> /*
> @@ -171,9 +174,11 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
> char *s;
> Elf_Shdr *symsec = &info->sechdrs[info->index.sym];
> unsigned long strtab_size;
> + void *data_base = mod->mem[MOD_DATA].base;
> + void *init_data_base = mod->mem[MOD_INIT_DATA].base;
>
> /* Set up to point into init section. */
> - mod->kallsyms = (void __rcu *)mod->init_layout.base +
> + mod->kallsyms = (void __rcu *)init_data_base +
> info->mod_kallsyms_init_off;
>
> rcu_read_lock();
> @@ -183,15 +188,15 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
> /* Make sure we get permanent strtab: don't use info->strtab. */
> rcu_dereference(mod->kallsyms)->strtab =
> (void *)info->sechdrs[info->index.str].sh_addr;
> - rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs;
> + rcu_dereference(mod->kallsyms)->typetab = init_data_base + info->init_typeoffs;
>
> /*
> * Now populate the cut down core kallsyms for after init
> * and set types up while we still have access to sections.
> */
> - mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs;
> - mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs;
> - mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs;
> + mod->core_kallsyms.symtab = dst = data_base + info->symoffs;
> + mod->core_kallsyms.strtab = s = data_base + info->stroffs;
> + mod->core_kallsyms.typetab = data_base + info->core_typeoffs;
> strtab_size = info->core_typeoffs - info->stroffs;
> src = rcu_dereference(mod->kallsyms)->symtab;
> for (ndst = i = 0; i < rcu_dereference(mod->kallsyms)->num_symtab; i++) {
> @@ -267,12 +272,15 @@ static const char *find_kallsyms_symbol(struct module *mod,
> unsigned int i, best = 0;
> unsigned long nextval, bestval;
> struct mod_kallsyms *kallsyms = rcu_dereference_sched(mod->kallsyms);
> + struct module_memory *mod_mem;
>
> /* At worse, next value is at end of module */
> if (within_module_init(addr, mod))
> - nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size;
> + mod_mem = &mod->mem[MOD_INIT_TEXT];
> else
> - nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size;
> + mod_mem = &mod->mem[MOD_TEXT];
> +
> + nextval = (unsigned long)mod_mem->base + mod_mem->size;
>
> bestval = kallsyms_symbol_value(&kallsyms->symtab[best]);
>
> diff --git a/kernel/module/kdb.c b/kernel/module/kdb.c
> index f4317f92e189..995c32d3698f 100644
> --- a/kernel/module/kdb.c
> +++ b/kernel/module/kdb.c
> @@ -26,10 +26,11 @@ int kdb_lsmod(int argc, const char **argv)
> if (mod->state == MODULE_STATE_UNFORMED)
> continue;
>
> - kdb_printf("%-20s%8u", mod->name, mod->core_layout.size);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - kdb_printf("/%8u", mod->data_layout.size);
> -#endif
> + kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
> + kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
> + kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
> + kdb_printf("/%8u", mod->mem[MOD_DATA].size);
> +
> kdb_printf(" 0x%px ", (void *)mod);
> #ifdef CONFIG_MODULE_UNLOAD
> kdb_printf("%4d ", module_refcount(mod));
> @@ -40,10 +41,10 @@ int kdb_lsmod(int argc, const char **argv)
> kdb_printf(" (Loading)");
> else
> kdb_printf(" (Live)");
> - kdb_printf(" 0x%px", mod->core_layout.base);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - kdb_printf("/0x%px", mod->data_layout.base);
> -#endif
> + kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
> + kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
> + kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
> + kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
>
> #ifdef CONFIG_MODULE_UNLOAD
> {
> diff --git a/kernel/module/main.c b/kernel/module/main.c
> index d3be89de706d..9f890d14072c 100644
> --- a/kernel/module/main.c
> +++ b/kernel/module/main.c
> @@ -80,12 +80,6 @@ struct mod_tree_root mod_tree __cacheline_aligned = {
> .addr_min = -1UL,
> };
>
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> -struct mod_tree_root mod_data_tree __cacheline_aligned = {
> - .addr_min = -1UL,
> -};
> -#endif
> -
> struct symsearch {
> const struct kernel_symbol *start, *stop;
> const s32 *crcs;
> @@ -93,14 +87,24 @@ struct symsearch {
> };
>
> /*
> - * Bounds of module text, for speeding up __module_address.
> + * Bounds of module memory, for speeding up __module_address.
> * Protected by module_mutex.
> */
> -static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree)
> +static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base,
> + unsigned int size, struct mod_tree_root *tree)
> {
> unsigned long min = (unsigned long)base;
> unsigned long max = min + size;
>
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> + if (mod_mem_type_is_core_data(type)) {
> + if (min < tree->data_addr_min)
> + tree->data_addr_min = min;
> + if (max > tree->data_addr_max)
> + tree->data_addr_max = max;
> + return;
> + }
> +#endif
> if (min < tree->addr_min)
> tree->addr_min = min;
> if (max > tree->addr_max)
> @@ -109,12 +113,12 @@ static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_r
>
> static void mod_update_bounds(struct module *mod)
> {
> - __mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree);
> - if (mod->init_layout.size)
> - __mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - __mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree);
> -#endif
> + for_each_mod_mem_type(type) {
> + struct module_memory *mod_mem = &mod->mem[type];
> +
> + if (mod_mem->size)
> + __mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree);
> + }
> }
>
> /* Block module loading/unloading? */
> @@ -923,12 +927,27 @@ static ssize_t store_uevent(struct module_attribute *mattr,
> struct module_attribute module_uevent =
> __ATTR(uevent, 0200, NULL, store_uevent);
>
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> +
> static ssize_t show_coresize(struct module_attribute *mattr,
> struct module_kobject *mk, char *buffer)
> {
> - return sprintf(buffer, "%u\n", mk->mod->core_layout.size);
> + return sprintf(buffer, "%u\n", mk->mod->mem[MOD_TEXT].size);
> }
>
> +#else
> +
> +static ssize_t show_coresize(struct module_attribute *mattr,
> + struct module_kobject *mk, char *buffer)
> +{
> + unsigned int size = 0;
> +
> + for_class_mod_mem_type(type, core)
> + size += mk->mod->mem[type].size;
> + return sprintf(buffer, "%u\n", size);
> +}
> +#endif
> +
> static struct module_attribute modinfo_coresize =
> __ATTR(coresize, 0444, show_coresize, NULL);
>
> @@ -936,7 +955,14 @@ static struct module_attribute modinfo_coresize =
> static ssize_t show_datasize(struct module_attribute *mattr,
> struct module_kobject *mk, char *buffer)
> {
> - return sprintf(buffer, "%u\n", mk->mod->data_layout.size);
> + unsigned int size = 0;
> +
> + for_class_mod_mem_type(type, core) {
> + if (type == MOD_TEXT)
> + continue;
> + size += mk->mod->mem[type].size;
> + }
> + return sprintf(buffer, "%u\n", size);
> }
>
> static struct module_attribute modinfo_datasize =
> @@ -946,7 +972,11 @@ static struct module_attribute modinfo_datasize =
> static ssize_t show_initsize(struct module_attribute *mattr,
> struct module_kobject *mk, char *buffer)
> {
> - return sprintf(buffer, "%u\n", mk->mod->init_layout.size);
> + unsigned int size = 0;
> +
> + for_class_mod_mem_type(type, init)
> + size += mk->mod->mem[type].size;
> + return sprintf(buffer, "%u\n", size);
> }
>
> static struct module_attribute modinfo_initsize =
> @@ -1143,6 +1173,63 @@ void __weak module_arch_freeing_init(struct module *mod)
> {
> }
>
> +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> +{
> + return mod_mem_is_core_data(type);
> +}
> +#else
> +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> +{
> + return false;
> +}
> +#endif
> +
> +static void *module_memory_alloc(unsigned int size, enum mod_mem_type type)
> +{
> + if (mod_mem_use_vmalloc(type))
> + return vzalloc(size);
> + return module_alloc(size);
> +}
> +
> +static void module_memory_free(void *ptr, enum mod_mem_type type)
> +{
> + if (mod_mem_use_vmalloc(type))
> + vfree(ptr);
> + else
> + module_memfree(ptr);
> +}
> +
> +static void free_mod_mem(struct module *mod)
> +{
> + /* free the memory in the right order to avoid use-after-free */
> + static enum mod_mem_type mod_mem_free_order[MOD_MEM_NUM_TYPES] = {
> + /* first free init sections */
> + MOD_INIT_TEXT,
> + MOD_INIT_DATA,
> + MOD_INIT_RODATA,
> +
> + /* then core sections, except rw data */
> + MOD_TEXT,
> + MOD_RODATA,
> + MOD_RO_AFTER_INIT,
> +
> + /* last, rw data */
> + MOD_DATA,
> + };
> + int i;
> +
> + for (i = 0; i < MOD_MEM_NUM_TYPES; i++) {
> + enum mod_mem_type type = mod_mem_free_order[i];
> + struct module_memory *mod_mem = &mod->mem[type];
> +
> + /* Free lock-classes; relies on the preceding sync_rcu(). */
> + lockdep_free_key_range(mod_mem->base, mod_mem->size);
> + if (mod_mem->size)
> + module_memory_free(mod_mem->base, type);
> + }
> +}
> +
> /* Free a module, remove from lists, etc. */
> static void free_module(struct module *mod)
> {
> @@ -1189,18 +1276,10 @@ static void free_module(struct module *mod)
>
> /* This may be empty, but that's OK */
> module_arch_freeing_init(mod);
> - module_memfree(mod->init_layout.base);
> kfree(mod->args);
> percpu_modfree(mod);
>
> - /* Free lock-classes; relies on the preceding sync_rcu(). */
> - lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> -
> - /* Finally, free the core (containing the module structure) */
> - module_memfree(mod->core_layout.base);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - vfree(mod->data_layout.base);
> -#endif
> + free_mod_mem(mod);
> }
>
> void *__symbol_get(const char *symbol)
> @@ -1387,16 +1466,18 @@ unsigned int __weak arch_mod_section_prepend(struct module *mod,
> return 0;
> }
>
> -/* Update size with this section: return offset. */
> -long module_get_offset(struct module *mod, unsigned int *size,
> +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> Elf_Shdr *sechdr, unsigned int section)
> {
> - long ret;
> + long offset;
> + long mask = ((unsigned long)(type) & SH_ENTSIZE_TYPE_MASK) << SH_ENTSIZE_TYPE_SHIFT;
>
> - *size += arch_mod_section_prepend(mod, section);
> - ret = ALIGN(*size, sechdr->sh_addralign ?: 1);
> - *size = ret + sechdr->sh_size;
> - return ret;
> + mod->mem[type].size += arch_mod_section_prepend(mod, section);
> + offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
> + mod->mem[type].size = offset + sechdr->sh_size;
> +
> + WARN_ON_ONCE(offset & mask);
> + return offset | mask;
> }
>
> static bool module_init_layout_section(const char *sname)
> @@ -1428,6 +1509,20 @@ static void layout_sections(struct module *mod, struct load_info *info)
> { SHF_WRITE | SHF_ALLOC, ARCH_SHF_SMALL },
> { ARCH_SHF_SMALL | SHF_ALLOC, 0 }
> };
> + static int core_m_to_mem_type[] = {
> + MOD_TEXT,
> + MOD_RODATA,
> + MOD_RO_AFTER_INIT,
> + MOD_DATA,
> + MOD_INVALID,
> + };
> + static int init_m_to_mem_type[] = {
> + MOD_INIT_TEXT,
> + MOD_INIT_RODATA,
> + MOD_INVALID,
> + MOD_INIT_DATA,
> + MOD_INVALID,
> + };
> unsigned int m, i;
>
> for (i = 0; i < info->hdr->e_shnum; i++)
> @@ -1435,41 +1530,30 @@ static void layout_sections(struct module *mod, struct load_info *info)
>
> pr_debug("Core section allocation order:\n");
> for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> + enum mod_mem_type type = core_m_to_mem_type[m];
> +
> for (i = 0; i < info->hdr->e_shnum; ++i) {
> Elf_Shdr *s = &info->sechdrs[i];
> const char *sname = info->secstrings + s->sh_name;
> - unsigned int *sizep;
>
> if ((s->sh_flags & masks[m][0]) != masks[m][0]
> || (s->sh_flags & masks[m][1])
> || s->sh_entsize != ~0UL
> || module_init_layout_section(sname))
> continue;
> - sizep = m ? &mod->data_layout.size : &mod->core_layout.size;
> - s->sh_entsize = module_get_offset(mod, sizep, s, i);
> +
> + if (WARN_ON_ONCE(type == MOD_INVALID))
> + continue;
> +
> + s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
> pr_debug("\t%s\n", sname);
> }
> - switch (m) {
> - case 0: /* executable */
> - mod->core_layout.size = strict_align(mod->core_layout.size);
> - mod->core_layout.text_size = mod->core_layout.size;
> - break;
> - case 1: /* RO: text and ro-data */
> - mod->data_layout.size = strict_align(mod->data_layout.size);
> - mod->data_layout.ro_size = mod->data_layout.size;
> - break;
> - case 2: /* RO after init */
> - mod->data_layout.size = strict_align(mod->data_layout.size);
> - mod->data_layout.ro_after_init_size = mod->data_layout.size;
> - break;
> - case 4: /* whole core */
> - mod->data_layout.size = strict_align(mod->data_layout.size);
> - break;
> - }
> }
>
> pr_debug("Init section allocation order:\n");
> for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> + enum mod_mem_type type = init_m_to_mem_type[m];
> +
> for (i = 0; i < info->hdr->e_shnum; ++i) {
> Elf_Shdr *s = &info->sechdrs[i];
> const char *sname = info->secstrings + s->sh_name;
> @@ -1479,30 +1563,13 @@ static void layout_sections(struct module *mod, struct load_info *info)
> || s->sh_entsize != ~0UL
> || !module_init_layout_section(sname))
> continue;
> - s->sh_entsize = (module_get_offset(mod, &mod->init_layout.size, s, i)
> - | INIT_OFFSET_MASK);
> +
> + if (WARN_ON_ONCE(type == MOD_INVALID))
> + continue;
> +
> + s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
> pr_debug("\t%s\n", sname);
> }
> - switch (m) {
> - case 0: /* executable */
> - mod->init_layout.size = strict_align(mod->init_layout.size);
> - mod->init_layout.text_size = mod->init_layout.size;
> - break;
> - case 1: /* RO: text and ro-data */
> - mod->init_layout.size = strict_align(mod->init_layout.size);
> - mod->init_layout.ro_size = mod->init_layout.size;
> - break;
> - case 2:
> - /*
> - * RO after init doesn't apply to init_layout (only
> - * core_layout), so it just takes the value of ro_size.
> - */
> - mod->init_layout.ro_after_init_size = mod->init_layout.ro_size;
> - break;
> - case 4: /* whole init */
> - mod->init_layout.size = strict_align(mod->init_layout.size);
> - break;
> - }
> }
> }
>
> @@ -2122,72 +2189,42 @@ static int move_module(struct module *mod, struct load_info *info)
> {
> int i;
> void *ptr;
> + enum mod_mem_type t;
>
> - /* Do the allocs. */
> - ptr = module_alloc(mod->core_layout.size);
> - /*
> - * The pointer to this block is stored in the module structure
> - * which is inside the block. Just mark it as not being a
> - * leak.
> - */
> - kmemleak_not_leak(ptr);
> - if (!ptr)
> - return -ENOMEM;
> -
> - memset(ptr, 0, mod->core_layout.size);
> - mod->core_layout.base = ptr;
> -
> - if (mod->init_layout.size) {
> - ptr = module_alloc(mod->init_layout.size);
> - /*
> - * The pointer to this block is stored in the module structure
> - * which is inside the block. This block doesn't need to be
> - * scanned as it contains data and code that will be freed
> - * after the module is initialized.
> - */
> - kmemleak_ignore(ptr);
> - if (!ptr) {
> - module_memfree(mod->core_layout.base);
> - return -ENOMEM;
> + for_each_mod_mem_type(type) {
> + if (!mod->mem[type].size) {
> + mod->mem[type].base = NULL;
> + continue;
> }
> - memset(ptr, 0, mod->init_layout.size);
> - mod->init_layout.base = ptr;
> - } else
> - mod->init_layout.base = NULL;
> + mod->mem[type].size = PAGE_ALIGN(mod->mem[type].size);
> + ptr = module_memory_alloc(mod->mem[type].size, type);
>
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - /* Do the allocs. */
> - ptr = vzalloc(mod->data_layout.size);
> /*
> * The pointer to this block is stored in the module structure
> * which is inside the block. Just mark it as not being a
> * leak.
> */
> - kmemleak_not_leak(ptr);
> + kmemleak_ignore(ptr);
> if (!ptr) {
> - module_memfree(mod->core_layout.base);
> - module_memfree(mod->init_layout.base);
> - return -ENOMEM;
> + t = type;
> + goto out_enomem;
> + }
> + memset(ptr, 0, mod->mem[type].size);
> + mod->mem[type].base = ptr;
> }
>
> - mod->data_layout.base = ptr;
> -#endif
> /* Transfer each section which specifies SHF_ALLOC */
> pr_debug("final section addresses:\n");
> for (i = 0; i < info->hdr->e_shnum; i++) {
> void *dest;
> Elf_Shdr *shdr = &info->sechdrs[i];
> + enum mod_mem_type type = shdr->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
>
> if (!(shdr->sh_flags & SHF_ALLOC))
> continue;
>
> - if (shdr->sh_entsize & INIT_OFFSET_MASK)
> - dest = mod->init_layout.base
> - + (shdr->sh_entsize & ~INIT_OFFSET_MASK);
> - else if (!(shdr->sh_flags & SHF_EXECINSTR))
> - dest = mod->data_layout.base + shdr->sh_entsize;
> - else
> - dest = mod->core_layout.base + shdr->sh_entsize;
> + dest = mod->mem[type].base +
> + (shdr->sh_entsize & SH_ENTSIZE_OFFSET_MASK);
>
> if (shdr->sh_type != SHT_NOBITS)
> memcpy(dest, (void *)shdr->sh_addr, shdr->sh_size);
> @@ -2198,6 +2235,10 @@ static int move_module(struct module *mod, struct load_info *info)
> }
>
> return 0;
> +out_enomem:
> + for (t--; t >= 0; t--)
> + module_memory_free(mod->mem[t].base, t);
> + return -ENOMEM;
> }
>
> static int check_module_license_and_versions(struct module *mod)
> @@ -2242,12 +2283,13 @@ static void flush_module_icache(const struct module *mod)
> * Do it before processing of module parameters, so the module
> * can provide parameter accessor functions of its own.
> */
> - if (mod->init_layout.base)
> - flush_icache_range((unsigned long)mod->init_layout.base,
> - (unsigned long)mod->init_layout.base
> - + mod->init_layout.size);
> - flush_icache_range((unsigned long)mod->core_layout.base,
> - (unsigned long)mod->core_layout.base + mod->core_layout.size);
> + for_each_mod_mem_type(type) {
> + const struct module_memory *mod_mem = &mod->mem[type];
> +
> + if (mod_mem->size)
> + flush_icache_range((unsigned long)mod_mem->base,
> + (unsigned long)mod_mem->base + mod_mem->size);
> + }
> }
>
> bool __weak module_elf_check_arch(Elf_Ehdr *hdr)
> @@ -2350,11 +2392,8 @@ static void module_deallocate(struct module *mod, struct load_info *info)
> {
> percpu_modfree(mod);
> module_arch_freeing_init(mod);
> - module_memfree(mod->init_layout.base);
> - module_memfree(mod->core_layout.base);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - vfree(mod->data_layout.base);
> -#endif
> +
> + free_mod_mem(mod);
> }
>
> int __weak module_finalize(const Elf_Ehdr *hdr,
> @@ -2415,7 +2454,9 @@ static void do_mod_ctors(struct module *mod)
> /* For freeing module_init on success, in case kallsyms traversing */
> struct mod_initfree {
> struct llist_node node;
> - void *module_init;
> + void *init_text;
> + void *init_data;
> + void *init_rodata;
> };
>
> static void do_free_init(struct work_struct *w)
> @@ -2429,7 +2470,9 @@ static void do_free_init(struct work_struct *w)
>
> llist_for_each_safe(pos, n, list) {
> initfree = container_of(pos, struct mod_initfree, node);
> - module_memfree(initfree->module_init);
> + module_memfree(initfree->init_text);
> + module_memfree(initfree->init_data);
> + module_memfree(initfree->init_rodata);
> kfree(initfree);
> }
> }
> @@ -2456,7 +2499,9 @@ static noinline int do_init_module(struct module *mod)
> ret = -ENOMEM;
> goto fail;
> }
> - freeinit->module_init = mod->init_layout.base;
> + freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
> + freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
> + freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
>
> do_mod_ctors(mod);
> /* Start the module */
> @@ -2492,8 +2537,8 @@ static noinline int do_init_module(struct module *mod)
> if (!mod->async_probe_requested)
> async_synchronize_full();
>
> - ftrace_free_mem(mod, mod->init_layout.base, mod->init_layout.base +
> - mod->init_layout.size);
> + ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
> + mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
> mutex_lock(&module_mutex);
> /* Drop initial reference. */
> module_put(mod);
> @@ -2505,11 +2550,10 @@ static noinline int do_init_module(struct module *mod)
> module_enable_ro(mod, true);
> mod_tree_remove_init(mod);
> module_arch_freeing_init(mod);
> - mod->init_layout.base = NULL;
> - mod->init_layout.size = 0;
> - mod->init_layout.ro_size = 0;
> - mod->init_layout.ro_after_init_size = 0;
> - mod->init_layout.text_size = 0;
> + for_class_mod_mem_type(type, init) {
> + mod->mem[type].base = NULL;
> + mod->mem[type].size = 0;
> + }
> #ifdef CONFIG_DEBUG_INFO_BTF_MODULES
> /* .BTF is not SHF_ALLOC and will get removed, so sanitize pointer */
> mod->btf_data = NULL;
> @@ -2628,9 +2672,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
> module_bug_finalize(info->hdr, info->sechdrs, mod);
> module_cfi_finalize(info->hdr, info->sechdrs, mod);
>
> - if (module_check_misalignment(mod))
> - goto out_misaligned;
> -
> module_enable_ro(mod, false);
> module_enable_nx(mod);
> module_enable_x(mod);
> @@ -2644,8 +2685,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
>
> return 0;
>
> -out_misaligned:
> - err = -EINVAL;
> out:
> mutex_unlock(&module_mutex);
> return err;
> @@ -2909,7 +2948,10 @@ static int load_module(struct load_info *info, const char __user *uargs,
> mutex_unlock(&module_mutex);
> free_module:
> /* Free lock-classes; relies on the preceding sync_rcu() */
> - lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> + for_class_mod_mem_type(type, core) {
> + lockdep_free_key_range(mod->mem[type].base,
> + mod->mem[type].size);
> + }
>
> module_deallocate(mod, info);
> free_copy:
> @@ -3060,20 +3102,21 @@ bool is_module_address(unsigned long addr)
> struct module *__module_address(unsigned long addr)
> {
> struct module *mod;
> - struct mod_tree_root *tree;
>
> if (addr >= mod_tree.addr_min && addr <= mod_tree.addr_max)
> - tree = &mod_tree;
> + goto lookup;
> +
> #ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - else if (addr >= mod_data_tree.addr_min && addr <= mod_data_tree.addr_max)
> - tree = &mod_data_tree;
> + if (addr >= mod_tree.data_addr_min && addr <= mod_tree.data_addr_max)
> + goto lookup;
> #endif
> - else
> +
> return NULL;
>
> +lookup:
> module_assert_mutex_or_preempt();
>
> - mod = mod_find(addr, tree);
> + mod = mod_find(addr, &mod_tree);
> if (mod) {
> BUG_ON(!within_module(addr, mod));
> if (mod->state == MODULE_STATE_UNFORMED)
> @@ -3113,8 +3156,8 @@ struct module *__module_text_address(unsigned long addr)
> struct module *mod = __module_address(addr);
> if (mod) {
> /* Make sure it's within the text section. */
> - if (!within(addr, mod->init_layout.base, mod->init_layout.text_size)
> - && !within(addr, mod->core_layout.base, mod->core_layout.text_size))
> + if (!within_module_mem_type(addr, mod, MOD_TEXT) &&
> + !within_module_mem_type(addr, mod, MOD_INIT_TEXT))
> mod = NULL;
> }
> return mod;
> diff --git a/kernel/module/procfs.c b/kernel/module/procfs.c
> index cf5b9f1e6ec4..0a4841e88adb 100644
> --- a/kernel/module/procfs.c
> +++ b/kernel/module/procfs.c
> @@ -62,6 +62,15 @@ static void m_stop(struct seq_file *m, void *p)
> mutex_unlock(&module_mutex);
> }
>
> +static unsigned int module_total_size(struct module *mod)
> +{
> + int size = 0;
> +
> + for_each_mod_mem_type(type)
> + size += mod->mem[type].size;
> + return size;
> +}
> +
> static int m_show(struct seq_file *m, void *p)
> {
> struct module *mod = list_entry(p, struct module, list);
> @@ -73,10 +82,7 @@ static int m_show(struct seq_file *m, void *p)
> if (mod->state == MODULE_STATE_UNFORMED)
> return 0;
>
> - size = mod->init_layout.size + mod->core_layout.size;
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - size += mod->data_layout.size;
> -#endif
> + size = module_total_size(mod);
> seq_printf(m, "%s %u", mod->name, size);
> print_unload_info(m, mod);
>
> @@ -86,7 +92,7 @@ static int m_show(struct seq_file *m, void *p)
> mod->state == MODULE_STATE_COMING ? "Loading" :
> "Live");
> /* Used by oprofile and other similar tools. */
> - value = m->private ? NULL : mod->core_layout.base;
> + value = m->private ? NULL : mod->mem[MOD_TEXT].base;
> seq_printf(m, " 0x%px", value);
>
> /* Taints info */
> diff --git a/kernel/module/strict_rwx.c b/kernel/module/strict_rwx.c
> index 14fbea66f12f..2e79a77f40eb 100644
> --- a/kernel/module/strict_rwx.c
> +++ b/kernel/module/strict_rwx.c
> @@ -11,82 +11,26 @@
> #include <linux/set_memory.h>
> #include "internal.h"
>
> -/*
> - * LKM RO/NX protection: protect module's text/ro-data
> - * from modification and any data from execution.
> - *
> - * General layout of module is:
> - * [text] [read-only-data] [ro-after-init] [writable data]
> - * text_size -----^ ^ ^ ^
> - * ro_size ------------------------| | |
> - * ro_after_init_size -----------------------------| |
> - * size -----------------------------------------------------------|
> - *
> - * These values are always page-aligned (as is base) when
> - * CONFIG_STRICT_MODULE_RWX is set.
> - */
> +static void module_set_memory(
> + const struct module *mod, enum mod_mem_type type,
> + int (*set_memory)(unsigned long start, int num_pages))
> +{
> + const struct module_memory *mod_mem = &mod->mem[type];
> +
> + set_vm_flush_reset_perms(mod_mem->base);
> + set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
> +}
>
> /*
> * Since some arches are moving towards PAGE_KERNEL module allocations instead
> - * of PAGE_KERNEL_EXEC, keep frob_text() and module_enable_x() independent of
> + * of PAGE_KERNEL_EXEC, keep module_enable_x() independent of
> * CONFIG_STRICT_MODULE_RWX because they are needed regardless of whether we
> * are strict.
> */
> -static void frob_text(const struct module_layout *layout,
> - int (*set_memory)(unsigned long start, int num_pages))
> -{
> - set_memory((unsigned long)layout->base,
> - PAGE_ALIGN(layout->text_size) >> PAGE_SHIFT);
> -}
> -
> -static void frob_rodata(const struct module_layout *layout,
> - int (*set_memory)(unsigned long start, int num_pages))
> -{
> - set_memory((unsigned long)layout->base + layout->text_size,
> - (layout->ro_size - layout->text_size) >> PAGE_SHIFT);
> -}
> -
> -static void frob_ro_after_init(const struct module_layout *layout,
> - int (*set_memory)(unsigned long start, int num_pages))
> -{
> - set_memory((unsigned long)layout->base + layout->ro_size,
> - (layout->ro_after_init_size - layout->ro_size) >> PAGE_SHIFT);
> -}
> -
> -static void frob_writable_data(const struct module_layout *layout,
> - int (*set_memory)(unsigned long start, int num_pages))
> -{
> - set_memory((unsigned long)layout->base + layout->ro_after_init_size,
> - (layout->size - layout->ro_after_init_size) >> PAGE_SHIFT);
> -}
> -
> -static bool layout_check_misalignment(const struct module_layout *layout)
> -{
> - return WARN_ON(!PAGE_ALIGNED(layout->base)) ||
> - WARN_ON(!PAGE_ALIGNED(layout->text_size)) ||
> - WARN_ON(!PAGE_ALIGNED(layout->ro_size)) ||
> - WARN_ON(!PAGE_ALIGNED(layout->ro_after_init_size)) ||
> - WARN_ON(!PAGE_ALIGNED(layout->size));
> -}
> -
> -bool module_check_misalignment(const struct module *mod)
> -{
> - if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> - return false;
> -
> - return layout_check_misalignment(&mod->core_layout) ||
> - layout_check_misalignment(&mod->data_layout) ||
> - layout_check_misalignment(&mod->init_layout);
> -}
> -
> void module_enable_x(const struct module *mod)
> {
> - if (!PAGE_ALIGNED(mod->core_layout.base) ||
> - !PAGE_ALIGNED(mod->init_layout.base))
> - return;
> -
> - frob_text(&mod->core_layout, set_memory_x);
> - frob_text(&mod->init_layout, set_memory_x);
> + for_class_mod_mem_type(type, text)
> + module_set_memory(mod, type, set_memory_x);
> }
>
> void module_enable_ro(const struct module *mod, bool after_init)
> @@ -98,16 +42,13 @@ void module_enable_ro(const struct module *mod, bool after_init)
> return;
> #endif
>
> - set_vm_flush_reset_perms(mod->core_layout.base);
> - set_vm_flush_reset_perms(mod->init_layout.base);
> - frob_text(&mod->core_layout, set_memory_ro);
> -
> - frob_rodata(&mod->data_layout, set_memory_ro);
> - frob_text(&mod->init_layout, set_memory_ro);
> - frob_rodata(&mod->init_layout, set_memory_ro);
> + module_set_memory(mod, MOD_TEXT, set_memory_ro);
> + module_set_memory(mod, MOD_INIT_TEXT, set_memory_ro);
> + module_set_memory(mod, MOD_RODATA, set_memory_ro);
> + module_set_memory(mod, MOD_INIT_RODATA, set_memory_ro);
>
> if (after_init)
> - frob_ro_after_init(&mod->data_layout, set_memory_ro);
> + module_set_memory(mod, MOD_RO_AFTER_INIT, set_memory_ro);
> }
>
> void module_enable_nx(const struct module *mod)
> @@ -115,11 +56,8 @@ void module_enable_nx(const struct module *mod)
> if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> return;
>
> - frob_rodata(&mod->data_layout, set_memory_nx);
> - frob_ro_after_init(&mod->data_layout, set_memory_nx);
> - frob_writable_data(&mod->data_layout, set_memory_nx);
> - frob_rodata(&mod->init_layout, set_memory_nx);
> - frob_writable_data(&mod->init_layout, set_memory_nx);
> + for_class_mod_mem_type(type, data)
> + module_set_memory(mod, type, set_memory_nx);
> }
>
> int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> diff --git a/kernel/module/tree_lookup.c b/kernel/module/tree_lookup.c
> index 8ec5cfd60496..277197977d43 100644
> --- a/kernel/module/tree_lookup.c
> +++ b/kernel/module/tree_lookup.c
> @@ -21,16 +21,16 @@
>
> static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n)
> {
> - struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> + struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
>
> - return (unsigned long)layout->base;
> + return (unsigned long)mod_mem->base;
> }
>
> static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n)
> {
> - struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> + struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
>
> - return (unsigned long)layout->size;
> + return (unsigned long)mod_mem->size;
> }
>
> static __always_inline bool
> @@ -77,32 +77,27 @@ static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *
> */
> void mod_tree_insert(struct module *mod)
> {
> - mod->core_layout.mtn.mod = mod;
> - mod->init_layout.mtn.mod = mod;
> -
> - __mod_tree_insert(&mod->core_layout.mtn, &mod_tree);
> - if (mod->init_layout.size)
> - __mod_tree_insert(&mod->init_layout.mtn, &mod_tree);
> -
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - mod->data_layout.mtn.mod = mod;
> - __mod_tree_insert(&mod->data_layout.mtn, &mod_data_tree);
> -#endif
> + for_each_mod_mem_type(type) {
> + mod->mem[type].mtn.mod = mod;
> + if (mod->mem[type].size)
> + __mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
> + }
> }
>
> void mod_tree_remove_init(struct module *mod)
> {
> - if (mod->init_layout.size)
> - __mod_tree_remove(&mod->init_layout.mtn, &mod_tree);
> + for_class_mod_mem_type(type, init) {
> + if (mod->mem[type].size)
> + __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> + }
> }
>
> void mod_tree_remove(struct module *mod)
> {
> - __mod_tree_remove(&mod->core_layout.mtn, &mod_tree);
> - mod_tree_remove_init(mod);
> -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> - __mod_tree_remove(&mod->data_layout.mtn, &mod_data_tree);
> -#endif
> + for_each_mod_mem_type(type) {
> + if (mod->mem[type].size)
> + __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> + }
> }
>
> struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)
On Wed, Feb 1, 2023 at 10:35 AM Christophe Leroy
<[email protected]> wrote:
>
> Hi,
>
> On what tree / branch does this patch applies ?
>
> I have tried with modules/modules-next, linus/master, with v6.1. It
> conflicts on all of them.
>
> 'git am -3' even says that 'It does not apply to blobs recorded in its
> index.'
Hmmm... I was using "git format-patch -b" to skip space changes. I
guess I should use --ignore-space-at-eol instead.
Resending it.
Thanks,
Song
PS: it was based on modules-next.
>
> Christophe
>
> Le 01/02/2023 à 18:36, Song Liu a écrit :
> > module_layout manages different types of memory (text, data, rodata, etc.)
> > in one allocation, which is problematic for some reasons:
> >
> > 1. It is hard to enable CONFIG_STRICT_MODULE_RWX.
> > 2. It is hard to use huge pages in modules (and not break strict rwx).
> > 3. Many archs uses module_layout for arch-specific data, but it is not
> > obvious how these data are used (are they RO, RX, or RW?)
> >
> > Improve the scenario by replacing 2 (or 3) module_layout per module with
> > up to 7 module_memory per module:
> >
> > MOD_TEXT,
> > MOD_DATA,
> > MOD_RODATA,
> > MOD_RO_AFTER_INIT,
> > MOD_INIT_TEXT,
> > MOD_INIT_DATA,
> > MOD_INIT_RODATA,
> >
> > and allocating them separately. This adds slightly more entries to
> > mod_tree (from up to 3 entries per module, to up to 7 entries per
> > module). However, this at most adds a small constant overhead to
> > __module_address(), which is expected to be fast.
> >
> > Various archs use module_layout for different data. These data are put
> > into different module_memory based on their location in module_layout.
> > IOW, data that used to go with text is allocated with MOD_MEM_TYPE_TEXT;
> > data that used to go with data is allocated with MOD_MEM_TYPE_DATA, etc.
> >
> > module_memory simplifies quite some of the module code. For example,
> > ARCH_WANTS_MODULES_DATA_IN_VMALLOC is a lot cleaner, as it just uses a
> > different allocator for the data. kernel/module/strict_rwx.c is also
> > much cleaner with module_memory.
> >
> > Signed-off-by: Song Liu <[email protected]>
> > Cc: Luis Chamberlain <[email protected]>
> > Cc: Thomas Gleixner <[email protected]>
> > Cc: Peter Zijlstra <[email protected]>
> > Cc: Guenter Roeck <[email protected]>
> > Cc: Christophe Leroy <[email protected]>
> >
> > ---
> >
> > This is the preparation work for the type aware module_alloc() discussed
> > in [1]. While this work is not covered much in the discussion, it is a
> > critical step of the effort.
> >
> > As this part grows pretty big (~1000 lines, + and -), I would like get
> > some feedback on it, so that I know it is on the right track.
> >
> > Please share your comments. Thanks!
> >
> > Test coverage: Tested on x86_64.
> > Build tested by kernel test bot in [2]. The only regression in [2] was a
> > typo in parisc, which is also fixed.
> >
> > [1] https://lore.kernel.org/linux-mm/[email protected]/T/#u
> > [2] https://lore.kernel.org/linux-raid/63b8827e.clJQX2wg+I+tiX7m%[email protected]/T/#u
> >
> > Changes v1 => v2:
> > 1. Add data_addr_[min|max] for CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > case.
> >
> > Changes v2 => v3:
> > 1. Fix and remove the last use of INIT_OFFSET_MASK.
> > 2. Add more information in the commit log. (Luis Chamberlain)
> > 3. Rebase and fix issues in x86/calltrunks.
> > 4. Minor cleanup.
> >
> > Changes v3 => v4:
> > 1. Shorten enum/variable names, so that the code are easier to read.
> > (Christophe Leroy)
> > 2. Remove an used variable. (Guenter Roeck, Christophe Leroy)
> >
> > Changes v4 => v5:
> > 1. Simplify some code some code. (Peter Zijlstra, Christophe Leroy)
> > 2. Remove module_check_misalignment(), which is not useful any more.
> >
> > Changes v5 => v6:
> > 1. Improve mod_mem_type_is_* and for_*mod_mem_type marcos.
> > (Peter Zijlstra).
> > ---
> > arch/alpha/kernel/module.c | 2 +-
> > arch/arc/kernel/unwind.c | 9 +-
> > arch/arm/kernel/module-plts.c | 2 +-
> > arch/arm64/kernel/module-plts.c | 2 +-
> > arch/ia64/kernel/module.c | 24 ++-
> > arch/mips/kernel/vpe.c | 11 +-
> > arch/parisc/kernel/module.c | 51 ++---
> > arch/powerpc/kernel/module_32.c | 7 +-
> > arch/s390/kernel/module.c | 26 +--
> > arch/x86/kernel/callthunks.c | 4 +-
> > arch/x86/kernel/module.c | 4 +-
> > include/linux/module.h | 86 +++++---
> > kernel/module/internal.h | 40 ++--
> > kernel/module/kallsyms.c | 56 ++---
> > kernel/module/kdb.c | 17 +-
> > kernel/module/main.c | 355 ++++++++++++++++++--------------
> > kernel/module/procfs.c | 16 +-
> > kernel/module/strict_rwx.c | 100 ++-------
> > kernel/module/tree_lookup.c | 39 ++--
> > 19 files changed, 432 insertions(+), 419 deletions(-)
> >
> > diff --git a/arch/alpha/kernel/module.c b/arch/alpha/kernel/module.c
> > index 5b60c248de9e..9109213abc09 100644
> > --- a/arch/alpha/kernel/module.c
> > +++ b/arch/alpha/kernel/module.c
> > @@ -148,7 +148,7 @@ apply_relocate_add(Elf64_Shdr *sechdrs, const char *strtab,
> >
> > /* The small sections were sorted to the end of the segment.
> > The following should definitely cover them. */
> > - gp = (u64)me->core_layout.base + me->core_layout.size - 0x8000;
> > + gp = (u64)me->mem[MOD_DATA].base + me->mem[MOD_DATA].size - 0x8000;
> > got = sechdrs[me->arch.gotsecindex].sh_addr;
> >
> > for (i = 0; i < n; i++) {
> > diff --git a/arch/arc/kernel/unwind.c b/arch/arc/kernel/unwind.c
> > index 200270a94558..933451f4494f 100644
> > --- a/arch/arc/kernel/unwind.c
> > +++ b/arch/arc/kernel/unwind.c
> > @@ -369,6 +369,8 @@ void *unwind_add_table(struct module *module, const void *table_start,
> > unsigned long table_size)
> > {
> > struct unwind_table *table;
> > + struct module_memory *mod_mem_core_text;
> > + struct module_memory *mod_mem_init_text;
> >
> > if (table_size <= 0)
> > return NULL;
> > @@ -377,9 +379,12 @@ void *unwind_add_table(struct module *module, const void *table_start,
> > if (!table)
> > return NULL;
> >
> > + mod_mem_core_text = &module->mem[MOD_TEXT];
> > + mod_mem_init_text = &module->mem[MOD_INIT_TEXT];
> > +
> > init_unwind_table(table, module->name,
> > - module->core_layout.base, module->core_layout.size,
> > - module->init_layout.base, module->init_layout.size,
> > + mod_mem_core_text->base, mod_mem_core_text->size,
> > + mod_mem_init_text->base, mod_mem_init_text->size,
> > table_start, table_size,
> > NULL, 0);
> >
> > diff --git a/arch/arm/kernel/module-plts.c b/arch/arm/kernel/module-plts.c
> > index af7c322ebed6..9d4ecb6b1412 100644
> > --- a/arch/arm/kernel/module-plts.c
> > +++ b/arch/arm/kernel/module-plts.c
> > @@ -30,7 +30,7 @@ static const u32 fixed_plts[] = {
> >
> > static bool in_init(const struct module *mod, unsigned long loc)
> > {
> > - return loc - (u32)mod->init_layout.base < mod->init_layout.size;
> > + return within_module_init(loc, mod);
> > }
> >
> > static void prealloc_fixed(struct mod_plt_sec *pltsec, struct plt_entries *plt)
> > diff --git a/arch/arm64/kernel/module-plts.c b/arch/arm64/kernel/module-plts.c
> > index 5a0a8f552a61..fedffa83a283 100644
> > --- a/arch/arm64/kernel/module-plts.c
> > +++ b/arch/arm64/kernel/module-plts.c
> > @@ -67,7 +67,7 @@ static bool plt_entries_equal(const struct plt_entry *a,
> >
> > static bool in_init(const struct module *mod, void *loc)
> > {
> > - return (u64)loc - (u64)mod->init_layout.base < mod->init_layout.size;
> > + return within_module_init(loc, mod);
> > }
> >
> > u64 module_emit_plt_entry(struct module *mod, Elf64_Shdr *sechdrs,
> > diff --git a/arch/ia64/kernel/module.c b/arch/ia64/kernel/module.c
> > index 8f62cf97f691..3661135da9d9 100644
> > --- a/arch/ia64/kernel/module.c
> > +++ b/arch/ia64/kernel/module.c
> > @@ -485,19 +485,19 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
> > return 0;
> > }
> >
> > -static inline int
> > +static inline bool
> > in_init (const struct module *mod, uint64_t addr)
> > {
> > - return addr - (uint64_t) mod->init_layout.base < mod->init_layout.size;
> > + return within_module_init(addr, mod);
> > }
> >
> > -static inline int
> > +static inline bool
> > in_core (const struct module *mod, uint64_t addr)
> > {
> > - return addr - (uint64_t) mod->core_layout.base < mod->core_layout.size;
> > + return within_module_core(addr, mod);
> > }
> >
> > -static inline int
> > +static inline bool
> > is_internal (const struct module *mod, uint64_t value)
> > {
> > return in_init(mod, value) || in_core(mod, value);
> > @@ -677,7 +677,8 @@ do_reloc (struct module *mod, uint8_t r_type, Elf64_Sym *sym, uint64_t addend,
> > break;
> >
> > case RV_BDREL:
> > - val -= (uint64_t) (in_init(mod, val) ? mod->init_layout.base : mod->core_layout.base);
> > + val -= (uint64_t) (in_init(mod, val) ? mod->mem[MOD_INIT_TEXT].base :
> > + mod->mem[MOD_TEXT].base);
> > break;
> >
> > case RV_LTV:
> > @@ -812,15 +813,18 @@ apply_relocate_add (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symind
> > * addresses have been selected...
> > */
> > uint64_t gp;
> > - if (mod->core_layout.size > MAX_LTOFF)
> > + struct module_memory *mod_mem;
> > +
> > + mod_mem = &mod->mem[MOD_DATA];
> > + if (mod_mem->size > MAX_LTOFF)
> > /*
> > * This takes advantage of fact that SHF_ARCH_SMALL gets allocated
> > * at the end of the module.
> > */
> > - gp = mod->core_layout.size - MAX_LTOFF / 2;
> > + gp = mod_mem->size - MAX_LTOFF / 2;
> > else
> > - gp = mod->core_layout.size / 2;
> > - gp = (uint64_t) mod->core_layout.base + ((gp + 7) & -8);
> > + gp = mod_mem->size / 2;
> > + gp = (uint64_t) mod_mem->base + ((gp + 7) & -8);
> > mod->arch.gp = gp;
> > DEBUGP("%s: placing gp at 0x%lx\n", __func__, gp);
> > }
> > diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c
> > index 13294972707b..ab114c686f9d 100644
> > --- a/arch/mips/kernel/vpe.c
> > +++ b/arch/mips/kernel/vpe.c
> > @@ -199,18 +199,17 @@ static void layout_sections(struct module *mod, const Elf_Ehdr *hdr,
> > for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> > for (i = 0; i < hdr->e_shnum; ++i) {
> > Elf_Shdr *s = &sechdrs[i];
> > + struct module_memory *mod_mem;
> > +
> > + mod_mem = &mod->mem[MOD_TEXT];
> >
> > if ((s->sh_flags & masks[m][0]) != masks[m][0]
> > || (s->sh_flags & masks[m][1])
> > || s->sh_entsize != ~0UL)
> > continue;
> > s->sh_entsize =
> > - get_offset((unsigned long *)&mod->core_layout.size, s);
> > + get_offset((unsigned long *)&mod_mem->size, s);
> > }
> > -
> > - if (m == 0)
> > - mod->core_layout.text_size = mod->core_layout.size;
> > -
> > }
> > }
> >
> > @@ -641,7 +640,7 @@ static int vpe_elfload(struct vpe *v)
> > layout_sections(&mod, hdr, sechdrs, secstrings);
> > }
> >
> > - v->load_addr = alloc_progmem(mod.core_layout.size);
> > + v->load_addr = alloc_progmem(mod.mod_mem[MOD_TEXT].size);
> > if (!v->load_addr)
> > return -ENOMEM;
> >
> > diff --git a/arch/parisc/kernel/module.c b/arch/parisc/kernel/module.c
> > index 7df140545b22..f6e38c4d3904 100644
> > --- a/arch/parisc/kernel/module.c
> > +++ b/arch/parisc/kernel/module.c
> > @@ -27,9 +27,9 @@
> > * We are not doing SEGREL32 handling correctly. According to the ABI, we
> > * should do a value offset, like this:
> > * if (in_init(me, (void *)val))
> > - * val -= (uint32_t)me->init_layout.base;
> > + * val -= (uint32_t)me->mem[MOD_INIT_TEXT].base;
> > * else
> > - * val -= (uint32_t)me->core_layout.base;
> > + * val -= (uint32_t)me->mem[MOD_TEXT].base;
> > * However, SEGREL32 is used only for PARISC unwind entries, and we want
> > * those entries to have an absolute address, and not just an offset.
> > *
> > @@ -76,25 +76,6 @@
> > * allows us to allocate up to 4095 GOT entries. */
> > #define MAX_GOTS 4095
> >
> > -/* three functions to determine where in the module core
> > - * or init pieces the location is */
> > -static inline int in_init(struct module *me, void *loc)
> > -{
> > - return (loc >= me->init_layout.base &&
> > - loc <= (me->init_layout.base + me->init_layout.size));
> > -}
> > -
> > -static inline int in_core(struct module *me, void *loc)
> > -{
> > - return (loc >= me->core_layout.base &&
> > - loc <= (me->core_layout.base + me->core_layout.size));
> > -}
> > -
> > -static inline int in_local(struct module *me, void *loc)
> > -{
> > - return in_init(me, loc) || in_core(me, loc);
> > -}
> > -
> > #ifndef CONFIG_64BIT
> > struct got_entry {
> > Elf32_Addr addr;
> > @@ -302,6 +283,7 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
> > {
> > unsigned long gots = 0, fdescs = 0, len;
> > unsigned int i;
> > + struct module_memory *mod_mem;
> >
> > len = hdr->e_shnum * sizeof(me->arch.section[0]);
> > me->arch.section = kzalloc(len, GFP_KERNEL);
> > @@ -346,14 +328,15 @@ int module_frob_arch_sections(CONST Elf_Ehdr *hdr,
> > me->arch.section[s].stub_entries += count;
> > }
> >
> > + mod_mem = &me->mem[MOD_TEXT];
> > /* align things a bit */
> > - me->core_layout.size = ALIGN(me->core_layout.size, 16);
> > - me->arch.got_offset = me->core_layout.size;
> > - me->core_layout.size += gots * sizeof(struct got_entry);
> > + mod_mem->size = ALIGN(mod_mem->size, 16);
> > + me->arch.got_offset = mod_mem->size;
> > + mod_mem->size += gots * sizeof(struct got_entry);
> >
> > - me->core_layout.size = ALIGN(me->core_layout.size, 16);
> > - me->arch.fdesc_offset = me->core_layout.size;
> > - me->core_layout.size += fdescs * sizeof(Elf_Fdesc);
> > + mod_mem->size = ALIGN(mod_mem->size, 16);
> > + me->arch.fdesc_offset = mod_mem->size;
> > + mod_mem->size += fdescs * sizeof(Elf_Fdesc);
> >
> > me->arch.got_max = gots;
> > me->arch.fdesc_max = fdescs;
> > @@ -371,7 +354,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
> >
> > BUG_ON(value == 0);
> >
> > - got = me->core_layout.base + me->arch.got_offset;
> > + got = me->mem[MOD_TEXT].base + me->arch.got_offset;
> > for (i = 0; got[i].addr; i++)
> > if (got[i].addr == value)
> > goto out;
> > @@ -389,7 +372,7 @@ static Elf64_Word get_got(struct module *me, unsigned long value, long addend)
> > #ifdef CONFIG_64BIT
> > static Elf_Addr get_fdesc(struct module *me, unsigned long value)
> > {
> > - Elf_Fdesc *fdesc = me->core_layout.base + me->arch.fdesc_offset;
> > + Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
> >
> > if (!value) {
> > printk(KERN_ERR "%s: zero OPD requested!\n", me->name);
> > @@ -407,7 +390,7 @@ static Elf_Addr get_fdesc(struct module *me, unsigned long value)
> >
> > /* Create new one */
> > fdesc->addr = value;
> > - fdesc->gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> > + fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
> > return (Elf_Addr)fdesc;
> > }
> > #endif /* CONFIG_64BIT */
> > @@ -742,7 +725,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
> > loc, val);
> > val += addend;
> > /* can we reach it locally? */
> > - if (in_local(me, (void *)val)) {
> > + if (within_module(val, me)) {
> > /* this is the case where the symbol is local
> > * to the module, but in a different section,
> > * so stub the jump in case it's more than 22
> > @@ -801,7 +784,7 @@ int apply_relocate_add(Elf_Shdr *sechdrs,
> > break;
> > case R_PARISC_FPTR64:
> > /* 64-bit function address */
> > - if(in_local(me, (void *)(val + addend))) {
> > + if (within_module(val + addend, me)) {
> > *loc64 = get_fdesc(me, val+addend);
> > pr_debug("FDESC for %s at %llx points to %llx\n",
> > strtab + sym->st_name, *loc64,
> > @@ -839,7 +822,7 @@ register_unwind_table(struct module *me,
> >
> > table = (unsigned char *)sechdrs[me->arch.unwind_section].sh_addr;
> > end = table + sechdrs[me->arch.unwind_section].sh_size;
> > - gp = (Elf_Addr)me->core_layout.base + me->arch.got_offset;
> > + gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
> >
> > pr_debug("register_unwind_table(), sect = %d at 0x%p - 0x%p (gp=0x%lx)\n",
> > me->arch.unwind_section, table, end, gp);
> > @@ -977,7 +960,7 @@ void module_arch_cleanup(struct module *mod)
> > #ifdef CONFIG_64BIT
> > void *dereference_module_function_descriptor(struct module *mod, void *ptr)
> > {
> > - unsigned long start_opd = (Elf64_Addr)mod->core_layout.base +
> > + unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
> > mod->arch.fdesc_offset;
> > unsigned long end_opd = start_opd +
> > mod->arch.fdesc_count * sizeof(Elf64_Fdesc);
> > diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c
> > index ea6536171778..2729179e9dc0 100644
> > --- a/arch/powerpc/kernel/module_32.c
> > +++ b/arch/powerpc/kernel/module_32.c
> > @@ -163,8 +163,7 @@ static uint32_t do_plt_call(void *location,
> >
> > pr_debug("Doing plt for call to 0x%x at 0x%x\n", val, (unsigned int)location);
> > /* Init, or core PLT? */
> > - if (location >= mod->core_layout.base
> > - && location < mod->core_layout.base + mod->core_layout.size)
> > + if (within_module_init(location, mod))
> > entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr;
> > else
> > entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr;
> > @@ -322,14 +321,14 @@ notrace int module_trampoline_target(struct module *mod, unsigned long addr,
> >
> > int module_finalize_ftrace(struct module *module, const Elf_Shdr *sechdrs)
> > {
> > - module->arch.tramp = do_plt_call(module->core_layout.base,
> > + module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
> > (unsigned long)ftrace_caller,
> > sechdrs, module);
> > if (!module->arch.tramp)
> > return -ENOENT;
> >
> > #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
> > - module->arch.tramp_regs = do_plt_call(module->core_layout.base,
> > + module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
> > (unsigned long)ftrace_regs_caller,
> > sechdrs, module);
> > if (!module->arch.tramp_regs)
> > diff --git a/arch/s390/kernel/module.c b/arch/s390/kernel/module.c
> > index 2d159b32885b..3ed455b785ad 100644
> > --- a/arch/s390/kernel/module.c
> > +++ b/arch/s390/kernel/module.c
> > @@ -126,6 +126,7 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> > Elf_Rela *rela;
> > char *strings;
> > int nrela, i, j;
> > + struct module_memory *mod_mem;
> >
> > /* Find symbol table and string table. */
> > symtab = NULL;
> > @@ -173,14 +174,15 @@ int module_frob_arch_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> >
> > /* Increase core size by size of got & plt and set start
> > offsets for got and plt. */
> > - me->core_layout.size = ALIGN(me->core_layout.size, 4);
> > - me->arch.got_offset = me->core_layout.size;
> > - me->core_layout.size += me->arch.got_size;
> > - me->arch.plt_offset = me->core_layout.size;
> > + mod_mem = &me->mem[MOD_TEXT];
> > + mod_mem->size = ALIGN(mod_mem->size, 4);
> > + me->arch.got_offset = mod_mem->size;
> > + mod_mem->size += me->arch.got_size;
> > + me->arch.plt_offset = mod_mem->size;
> > if (me->arch.plt_size) {
> > if (IS_ENABLED(CONFIG_EXPOLINE) && !nospec_disable)
> > me->arch.plt_size += PLT_ENTRY_SIZE;
> > - me->core_layout.size += me->arch.plt_size;
> > + mod_mem->size += me->arch.plt_size;
> > }
> > return 0;
> > }
> > @@ -304,7 +306,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> > case R_390_GOTPLT64: /* 64 bit offset to jump slot. */
> > case R_390_GOTPLTENT: /* 32 bit rel. offset to jump slot >> 1. */
> > if (info->got_initialized == 0) {
> > - Elf_Addr *gotent = me->core_layout.base +
> > + Elf_Addr *gotent = me->mem[MOD_TEXT].base +
> > me->arch.got_offset +
> > info->got_offset;
> >
> > @@ -329,7 +331,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> > rc = apply_rela_bits(loc, val, 0, 64, 0, write);
> > else if (r_type == R_390_GOTENT ||
> > r_type == R_390_GOTPLTENT) {
> > - val += (Elf_Addr) me->core_layout.base - loc;
> > + val += (Elf_Addr) me->mem[MOD_TEXT].base - loc;
> > rc = apply_rela_bits(loc, val, 1, 32, 1, write);
> > }
> > break;
> > @@ -345,7 +347,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> > char *plt_base;
> > char *ip;
> >
> > - plt_base = me->core_layout.base + me->arch.plt_offset;
> > + plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
> > ip = plt_base + info->plt_offset;
> > *(int *)insn = 0x0d10e310; /* basr 1,0 */
> > *(int *)&insn[4] = 0x100c0004; /* lg 1,12(1) */
> > @@ -375,7 +377,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> > val - loc + 0xffffUL < 0x1ffffeUL) ||
> > (r_type == R_390_PLT32DBL &&
> > val - loc + 0xffffffffULL < 0x1fffffffeULL)))
> > - val = (Elf_Addr) me->core_layout.base +
> > + val = (Elf_Addr) me->mem[MOD_TEXT].base +
> > me->arch.plt_offset +
> > info->plt_offset;
> > val += rela->r_addend - loc;
> > @@ -397,7 +399,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> > case R_390_GOTOFF32: /* 32 bit offset to GOT. */
> > case R_390_GOTOFF64: /* 64 bit offset to GOT. */
> > val = val + rela->r_addend -
> > - ((Elf_Addr) me->core_layout.base + me->arch.got_offset);
> > + ((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
> > if (r_type == R_390_GOTOFF16)
> > rc = apply_rela_bits(loc, val, 0, 16, 0, write);
> > else if (r_type == R_390_GOTOFF32)
> > @@ -407,7 +409,7 @@ static int apply_rela(Elf_Rela *rela, Elf_Addr base, Elf_Sym *symtab,
> > break;
> > case R_390_GOTPC: /* 32 bit PC relative offset to GOT. */
> > case R_390_GOTPCDBL: /* 32 bit PC rel. off. to GOT shifted by 1. */
> > - val = (Elf_Addr) me->core_layout.base + me->arch.got_offset +
> > + val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
> > rela->r_addend - loc;
> > if (r_type == R_390_GOTPC)
> > rc = apply_rela_bits(loc, val, 1, 32, 0, write);
> > @@ -515,7 +517,7 @@ int module_finalize(const Elf_Ehdr *hdr,
> > !nospec_disable && me->arch.plt_size) {
> > unsigned int *ij;
> >
> > - ij = me->core_layout.base + me->arch.plt_offset +
> > + ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
> > me->arch.plt_size - PLT_ENTRY_SIZE;
> > ij[0] = 0xc6000000; /* exrl %r0,.+10 */
> > ij[1] = 0x0005a7f4; /* j . */
> > diff --git a/arch/x86/kernel/callthunks.c b/arch/x86/kernel/callthunks.c
> > index ffea98f9064b..22ab13966427 100644
> > --- a/arch/x86/kernel/callthunks.c
> > +++ b/arch/x86/kernel/callthunks.c
> > @@ -330,8 +330,8 @@ void noinline callthunks_patch_module_calls(struct callthunk_sites *cs,
> > struct module *mod)
> > {
> > struct core_text ct = {
> > - .base = (unsigned long)mod->core_layout.base,
> > - .end = (unsigned long)mod->core_layout.base + mod->core_layout.size,
> > + .base = (unsigned long)mod->mem[MOD_TEXT].base,
> > + .end = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
> > .name = mod->name,
> > };
> >
> > diff --git a/arch/x86/kernel/module.c b/arch/x86/kernel/module.c
> > index 705fb2a41d7d..dbc8e226c464 100644
> > --- a/arch/x86/kernel/module.c
> > +++ b/arch/x86/kernel/module.c
> > @@ -339,8 +339,8 @@ int module_finalize(const Elf_Ehdr *hdr,
> > }
> > if (locks) {
> > void *lseg = (void *)locks->sh_addr;
> > - void *text = me->core_layout.base;
> > - void *text_end = text + me->core_layout.text_size;
> > + void *text = me->mem[MOD_TEXT].base;
> > + void *text_end = text + me->mem[MOD_TEXT].size;
> > alternatives_smp_module_add(me, me->name,
> > lseg, lseg + locks->sh_size,
> > text, text_end);
> > diff --git a/include/linux/module.h b/include/linux/module.h
> > index 8c5909c0076c..ccc9de8a93db 100644
> > --- a/include/linux/module.h
> > +++ b/include/linux/module.h
> > @@ -320,17 +320,47 @@ struct mod_tree_node {
> > struct latch_tree_node node;
> > };
> >
> > -struct module_layout {
> > - /* The actual code + data. */
> > +enum mod_mem_type {
> > + MOD_TEXT = 0,
> > + MOD_DATA,
> > + MOD_RODATA,
> > + MOD_RO_AFTER_INIT,
> > + MOD_INIT_TEXT,
> > + MOD_INIT_DATA,
> > + MOD_INIT_RODATA,
> > +
> > + MOD_MEM_NUM_TYPES,
> > + MOD_INVALID = -1,
> > +};
> > +
> > +#define mod_mem_type_is_init(type) \
> > + ((type) == MOD_INIT_TEXT || \
> > + (type) == MOD_INIT_DATA || \
> > + (type) == MOD_INIT_RODATA)
> > +
> > +#define mod_mem_type_is_core(type) (!mod_mem_type_is_init(type))
> > +
> > +#define mod_mem_type_is_text(type) \
> > + ((type) == MOD_TEXT || \
> > + (type) == MOD_INIT_TEXT)
> > +
> > +#define mod_mem_type_is_data(type) (!mod_mem_type_is_text(type))
> > +
> > +#define mod_mem_type_is_core_data(type) \
> > + (mod_mem_type_is_core(type) && \
> > + mod_mem_type_is_data(type))
> > +
> > +#define for_each_mod_mem_type(type) \
> > + for (enum mod_mem_type (type) = 0; \
> > + (type) < MOD_MEM_NUM_TYPES; (type)++)
> > +
> > +#define for_class_mod_mem_type(type, class) \
> > + for_each_mod_mem_type(type) \
> > + if (mod_mem_type_is_##class(type))
> > +
> > +struct module_memory {
> > void *base;
> > - /* Total size. */
> > unsigned int size;
> > - /* The size of the executable code. */
> > - unsigned int text_size;
> > - /* Size of RO section of the module (text+rodata) */
> > - unsigned int ro_size;
> > - /* Size of RO after init section */
> > - unsigned int ro_after_init_size;
> >
> > #ifdef CONFIG_MODULES_TREE_LOOKUP
> > struct mod_tree_node mtn;
> > @@ -339,9 +369,9 @@ struct module_layout {
> >
> > #ifdef CONFIG_MODULES_TREE_LOOKUP
> > /* Only touch one cacheline for common rbtree-for-core-layout case. */
> > -#define __module_layout_align ____cacheline_aligned
> > +#define __module_memory_align ____cacheline_aligned
> > #else
> > -#define __module_layout_align
> > +#define __module_memory_align
> > #endif
> >
> > struct mod_kallsyms {
> > @@ -418,12 +448,8 @@ struct module {
> > /* Startup function. */
> > int (*init)(void);
> >
> > - /* Core layout: rbtree is accessed frequently, so keep together. */
> > - struct module_layout core_layout __module_layout_align;
> > - struct module_layout init_layout;
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - struct module_layout data_layout;
> > -#endif
> > + /* rbtree is accessed frequently, so keep together. */
> > + struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
> >
> > /* Arch-specific module values */
> > struct mod_arch_specific arch;
> > @@ -573,23 +599,33 @@ bool __is_module_percpu_address(unsigned long addr, unsigned long *can_addr);
> > bool is_module_percpu_address(unsigned long addr);
> > bool is_module_text_address(unsigned long addr);
> >
> > +static inline bool within_module_mem_type(unsigned long addr,
> > + const struct module *mod,
> > + enum mod_mem_type type)
> > +{
> > + unsigned long base, size;
> > +
> > + base = (unsigned long)mod->mem[type].base;
> > + size = mod->mem[type].size;
> > + return addr - base < size;
> > +}
> > +
> > static inline bool within_module_core(unsigned long addr,
> > const struct module *mod)
> > {
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - if ((unsigned long)mod->data_layout.base <= addr &&
> > - addr < (unsigned long)mod->data_layout.base + mod->data_layout.size)
> > + for_class_mod_mem_type(type, core)
> > + if (within_module_mem_type(addr, mod, type))
> > return true;
> > -#endif
> > - return (unsigned long)mod->core_layout.base <= addr &&
> > - addr < (unsigned long)mod->core_layout.base + mod->core_layout.size;
> > + return false;
> > }
> >
> > static inline bool within_module_init(unsigned long addr,
> > const struct module *mod)
> > {
> > - return (unsigned long)mod->init_layout.base <= addr &&
> > - addr < (unsigned long)mod->init_layout.base + mod->init_layout.size;
> > + for_class_mod_mem_type(type, init)
> > + if (within_module_mem_type(addr, mod, type))
> > + return true;
> > + return false;
> > }
> >
> > static inline bool within_module(unsigned long addr, const struct module *mod)
> > diff --git a/kernel/module/internal.h b/kernel/module/internal.h
> > index 2e2bf236f558..c6c44ceca07a 100644
> > --- a/kernel/module/internal.h
> > +++ b/kernel/module/internal.h
> > @@ -17,27 +17,19 @@
> > #define ARCH_SHF_SMALL 0
> > #endif
> >
> > -/* If this is set, the section belongs in the init part of the module */
> > -#define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG - 1))
> > -/* Maximum number of characters written by module_flags() */
> > -#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
> > -
> > -#ifndef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -#define data_layout core_layout
> > -#endif
> > -
> > /*
> > - * Modules' sections will be aligned on page boundaries
> > - * to ensure complete separation of code and data, but
> > - * only when CONFIG_STRICT_MODULE_RWX=y
> > + * Use highest 4 bits of sh_entsize to store the mod_mem_type of this
> > + * section. This leaves 28 bits for offset on 32-bit systems, which is
> > + * about 256 MiB (WARN_ON_ONCE if we exceed that).
> > */
> > -static inline unsigned int strict_align(unsigned int size)
> > -{
> > - if (IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> > - return PAGE_ALIGN(size);
> > - else
> > - return size;
> > -}
> > +
> > +#define SH_ENTSIZE_TYPE_BITS 4
> > +#define SH_ENTSIZE_TYPE_SHIFT (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)
> > +#define SH_ENTSIZE_TYPE_MASK ((1UL << SH_ENTSIZE_TYPE_BITS) - 1)
> > +#define SH_ENTSIZE_OFFSET_MASK ((1UL << (BITS_PER_LONG - SH_ENTSIZE_TYPE_BITS)) - 1)
> > +
> > +/* Maximum number of characters written by module_flags() */
> > +#define MODULE_FLAGS_BUF_SIZE (TAINT_FLAGS_COUNT + 4)
> >
> > extern struct mutex module_mutex;
> > extern struct list_head modules;
> > @@ -101,8 +93,8 @@ int try_to_force_load(struct module *mod, const char *reason);
> > bool find_symbol(struct find_symbol_arg *fsa);
> > struct module *find_module_all(const char *name, size_t len, bool even_unformed);
> > int cmp_name(const void *name, const void *sym);
> > -long module_get_offset(struct module *mod, unsigned int *size, Elf_Shdr *sechdr,
> > - unsigned int section);
> > +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> > + Elf_Shdr *sechdr, unsigned int section);
> > char *module_flags(struct module *mod, char *buf, bool show_state);
> > size_t module_flags_taint(unsigned long taints, char *buf);
> >
> > @@ -190,10 +182,13 @@ struct mod_tree_root {
> > #endif
> > unsigned long addr_min;
> > unsigned long addr_max;
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > + unsigned long data_addr_min;
> > + unsigned long data_addr_max;
> > +#endif
> > };
> >
> > extern struct mod_tree_root mod_tree;
> > -extern struct mod_tree_root mod_data_tree;
> >
> > #ifdef CONFIG_MODULES_TREE_LOOKUP
> > void mod_tree_insert(struct module *mod);
> > @@ -224,7 +219,6 @@ void module_enable_nx(const struct module *mod);
> > void module_enable_x(const struct module *mod);
> > int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> > char *secstrings, struct module *mod);
> > -bool module_check_misalignment(const struct module *mod);
> >
> > #ifdef CONFIG_MODULE_SIG
> > int module_sig_check(struct load_info *info, int flags);
> > diff --git a/kernel/module/kallsyms.c b/kernel/module/kallsyms.c
> > index 4523f99b0358..acdca7692961 100644
> > --- a/kernel/module/kallsyms.c
> > +++ b/kernel/module/kallsyms.c
> > @@ -78,6 +78,7 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
> > unsigned int shnum, unsigned int pcpundx)
> > {
> > const Elf_Shdr *sec;
> > + enum mod_mem_type type;
> >
> > if (src->st_shndx == SHN_UNDEF ||
> > src->st_shndx >= shnum ||
> > @@ -90,11 +91,12 @@ static bool is_core_symbol(const Elf_Sym *src, const Elf_Shdr *sechdrs,
> > #endif
> >
> > sec = sechdrs + src->st_shndx;
> > + type = sec->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
> > if (!(sec->sh_flags & SHF_ALLOC)
> > #ifndef CONFIG_KALLSYMS_ALL
> > || !(sec->sh_flags & SHF_EXECINSTR)
> > #endif
> > - || (sec->sh_entsize & INIT_OFFSET_MASK))
> > + || mod_mem_type_is_init(type))
> > return false;
> >
> > return true;
> > @@ -113,11 +115,13 @@ void layout_symtab(struct module *mod, struct load_info *info)
> > Elf_Shdr *strsect = info->sechdrs + info->index.str;
> > const Elf_Sym *src;
> > unsigned int i, nsrc, ndst, strtab_size = 0;
> > + struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
> > + struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
> >
> > /* Put symbol section at end of init part of module. */
> > symsect->sh_flags |= SHF_ALLOC;
> > - symsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, symsect,
> > - info->index.sym) | INIT_OFFSET_MASK;
> > + symsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> > + symsect, info->index.sym);
> > pr_debug("\t%s\n", info->secstrings + symsect->sh_name);
> >
> > src = (void *)info->hdr + symsect->sh_offset;
> > @@ -134,28 +138,27 @@ void layout_symtab(struct module *mod, struct load_info *info)
> > }
> >
> > /* Append room for core symbols at end of core part. */
> > - info->symoffs = ALIGN(mod->data_layout.size, symsect->sh_addralign ?: 1);
> > - info->stroffs = mod->data_layout.size = info->symoffs + ndst * sizeof(Elf_Sym);
> > - mod->data_layout.size += strtab_size;
> > + info->symoffs = ALIGN(mod_mem_data->size, symsect->sh_addralign ?: 1);
> > + info->stroffs = mod_mem_data->size = info->symoffs + ndst * sizeof(Elf_Sym);
> > + mod_mem_data->size += strtab_size;
> > /* Note add_kallsyms() computes strtab_size as core_typeoffs - stroffs */
> > - info->core_typeoffs = mod->data_layout.size;
> > - mod->data_layout.size += ndst * sizeof(char);
> > - mod->data_layout.size = strict_align(mod->data_layout.size);
> > + info->core_typeoffs = mod_mem_data->size;
> > + mod_mem_data->size += ndst * sizeof(char);
> >
> > /* Put string table section at end of init part of module. */
> > strsect->sh_flags |= SHF_ALLOC;
> > - strsect->sh_entsize = module_get_offset(mod, &mod->init_layout.size, strsect,
> > - info->index.str) | INIT_OFFSET_MASK;
> > + strsect->sh_entsize = module_get_offset_and_type(mod, MOD_INIT_DATA,
> > + strsect, info->index.str);
> > pr_debug("\t%s\n", info->secstrings + strsect->sh_name);
> >
> > /* We'll tack temporary mod_kallsyms on the end. */
> > - mod->init_layout.size = ALIGN(mod->init_layout.size,
> > + mod_mem_init_data->size = ALIGN(mod_mem_init_data->size,
> > __alignof__(struct mod_kallsyms));
> > - info->mod_kallsyms_init_off = mod->init_layout.size;
> > - mod->init_layout.size += sizeof(struct mod_kallsyms);
> > - info->init_typeoffs = mod->init_layout.size;
> > - mod->init_layout.size += nsrc * sizeof(char);
> > - mod->init_layout.size = strict_align(mod->init_layout.size);
> > + info->mod_kallsyms_init_off = mod_mem_init_data->size;
> > +
> > + mod_mem_init_data->size += sizeof(struct mod_kallsyms);
> > + info->init_typeoffs = mod_mem_init_data->size;
> > + mod_mem_init_data->size += nsrc * sizeof(char);
> > }
> >
> > /*
> > @@ -171,9 +174,11 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
> > char *s;
> > Elf_Shdr *symsec = &info->sechdrs[info->index.sym];
> > unsigned long strtab_size;
> > + void *data_base = mod->mem[MOD_DATA].base;
> > + void *init_data_base = mod->mem[MOD_INIT_DATA].base;
> >
> > /* Set up to point into init section. */
> > - mod->kallsyms = (void __rcu *)mod->init_layout.base +
> > + mod->kallsyms = (void __rcu *)init_data_base +
> > info->mod_kallsyms_init_off;
> >
> > rcu_read_lock();
> > @@ -183,15 +188,15 @@ void add_kallsyms(struct module *mod, const struct load_info *info)
> > /* Make sure we get permanent strtab: don't use info->strtab. */
> > rcu_dereference(mod->kallsyms)->strtab =
> > (void *)info->sechdrs[info->index.str].sh_addr;
> > - rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs;
> > + rcu_dereference(mod->kallsyms)->typetab = init_data_base + info->init_typeoffs;
> >
> > /*
> > * Now populate the cut down core kallsyms for after init
> > * and set types up while we still have access to sections.
> > */
> > - mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs;
> > - mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs;
> > - mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs;
> > + mod->core_kallsyms.symtab = dst = data_base + info->symoffs;
> > + mod->core_kallsyms.strtab = s = data_base + info->stroffs;
> > + mod->core_kallsyms.typetab = data_base + info->core_typeoffs;
> > strtab_size = info->core_typeoffs - info->stroffs;
> > src = rcu_dereference(mod->kallsyms)->symtab;
> > for (ndst = i = 0; i < rcu_dereference(mod->kallsyms)->num_symtab; i++) {
> > @@ -267,12 +272,15 @@ static const char *find_kallsyms_symbol(struct module *mod,
> > unsigned int i, best = 0;
> > unsigned long nextval, bestval;
> > struct mod_kallsyms *kallsyms = rcu_dereference_sched(mod->kallsyms);
> > + struct module_memory *mod_mem;
> >
> > /* At worse, next value is at end of module */
> > if (within_module_init(addr, mod))
> > - nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size;
> > + mod_mem = &mod->mem[MOD_INIT_TEXT];
> > else
> > - nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size;
> > + mod_mem = &mod->mem[MOD_TEXT];
> > +
> > + nextval = (unsigned long)mod_mem->base + mod_mem->size;
> >
> > bestval = kallsyms_symbol_value(&kallsyms->symtab[best]);
> >
> > diff --git a/kernel/module/kdb.c b/kernel/module/kdb.c
> > index f4317f92e189..995c32d3698f 100644
> > --- a/kernel/module/kdb.c
> > +++ b/kernel/module/kdb.c
> > @@ -26,10 +26,11 @@ int kdb_lsmod(int argc, const char **argv)
> > if (mod->state == MODULE_STATE_UNFORMED)
> > continue;
> >
> > - kdb_printf("%-20s%8u", mod->name, mod->core_layout.size);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - kdb_printf("/%8u", mod->data_layout.size);
> > -#endif
> > + kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
> > + kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
> > + kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
> > + kdb_printf("/%8u", mod->mem[MOD_DATA].size);
> > +
> > kdb_printf(" 0x%px ", (void *)mod);
> > #ifdef CONFIG_MODULE_UNLOAD
> > kdb_printf("%4d ", module_refcount(mod));
> > @@ -40,10 +41,10 @@ int kdb_lsmod(int argc, const char **argv)
> > kdb_printf(" (Loading)");
> > else
> > kdb_printf(" (Live)");
> > - kdb_printf(" 0x%px", mod->core_layout.base);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - kdb_printf("/0x%px", mod->data_layout.base);
> > -#endif
> > + kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
> > + kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
> > + kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
> > + kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
> >
> > #ifdef CONFIG_MODULE_UNLOAD
> > {
> > diff --git a/kernel/module/main.c b/kernel/module/main.c
> > index d3be89de706d..9f890d14072c 100644
> > --- a/kernel/module/main.c
> > +++ b/kernel/module/main.c
> > @@ -80,12 +80,6 @@ struct mod_tree_root mod_tree __cacheline_aligned = {
> > .addr_min = -1UL,
> > };
> >
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > -struct mod_tree_root mod_data_tree __cacheline_aligned = {
> > - .addr_min = -1UL,
> > -};
> > -#endif
> > -
> > struct symsearch {
> > const struct kernel_symbol *start, *stop;
> > const s32 *crcs;
> > @@ -93,14 +87,24 @@ struct symsearch {
> > };
> >
> > /*
> > - * Bounds of module text, for speeding up __module_address.
> > + * Bounds of module memory, for speeding up __module_address.
> > * Protected by module_mutex.
> > */
> > -static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree)
> > +static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base,
> > + unsigned int size, struct mod_tree_root *tree)
> > {
> > unsigned long min = (unsigned long)base;
> > unsigned long max = min + size;
> >
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > + if (mod_mem_type_is_core_data(type)) {
> > + if (min < tree->data_addr_min)
> > + tree->data_addr_min = min;
> > + if (max > tree->data_addr_max)
> > + tree->data_addr_max = max;
> > + return;
> > + }
> > +#endif
> > if (min < tree->addr_min)
> > tree->addr_min = min;
> > if (max > tree->addr_max)
> > @@ -109,12 +113,12 @@ static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_r
> >
> > static void mod_update_bounds(struct module *mod)
> > {
> > - __mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree);
> > - if (mod->init_layout.size)
> > - __mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - __mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree);
> > -#endif
> > + for_each_mod_mem_type(type) {
> > + struct module_memory *mod_mem = &mod->mem[type];
> > +
> > + if (mod_mem->size)
> > + __mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree);
> > + }
> > }
> >
> > /* Block module loading/unloading? */
> > @@ -923,12 +927,27 @@ static ssize_t store_uevent(struct module_attribute *mattr,
> > struct module_attribute module_uevent =
> > __ATTR(uevent, 0200, NULL, store_uevent);
> >
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > +
> > static ssize_t show_coresize(struct module_attribute *mattr,
> > struct module_kobject *mk, char *buffer)
> > {
> > - return sprintf(buffer, "%u\n", mk->mod->core_layout.size);
> > + return sprintf(buffer, "%u\n", mk->mod->mem[MOD_TEXT].size);
> > }
> >
> > +#else
> > +
> > +static ssize_t show_coresize(struct module_attribute *mattr,
> > + struct module_kobject *mk, char *buffer)
> > +{
> > + unsigned int size = 0;
> > +
> > + for_class_mod_mem_type(type, core)
> > + size += mk->mod->mem[type].size;
> > + return sprintf(buffer, "%u\n", size);
> > +}
> > +#endif
> > +
> > static struct module_attribute modinfo_coresize =
> > __ATTR(coresize, 0444, show_coresize, NULL);
> >
> > @@ -936,7 +955,14 @@ static struct module_attribute modinfo_coresize =
> > static ssize_t show_datasize(struct module_attribute *mattr,
> > struct module_kobject *mk, char *buffer)
> > {
> > - return sprintf(buffer, "%u\n", mk->mod->data_layout.size);
> > + unsigned int size = 0;
> > +
> > + for_class_mod_mem_type(type, core) {
> > + if (type == MOD_TEXT)
> > + continue;
> > + size += mk->mod->mem[type].size;
> > + }
> > + return sprintf(buffer, "%u\n", size);
> > }
> >
> > static struct module_attribute modinfo_datasize =
> > @@ -946,7 +972,11 @@ static struct module_attribute modinfo_datasize =
> > static ssize_t show_initsize(struct module_attribute *mattr,
> > struct module_kobject *mk, char *buffer)
> > {
> > - return sprintf(buffer, "%u\n", mk->mod->init_layout.size);
> > + unsigned int size = 0;
> > +
> > + for_class_mod_mem_type(type, init)
> > + size += mk->mod->mem[type].size;
> > + return sprintf(buffer, "%u\n", size);
> > }
> >
> > static struct module_attribute modinfo_initsize =
> > @@ -1143,6 +1173,63 @@ void __weak module_arch_freeing_init(struct module *mod)
> > {
> > }
> >
> > +#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> > +{
> > + return mod_mem_is_core_data(type);
> > +}
> > +#else
> > +static bool mod_mem_use_vmalloc(enum mod_mem_type type)
> > +{
> > + return false;
> > +}
> > +#endif
> > +
> > +static void *module_memory_alloc(unsigned int size, enum mod_mem_type type)
> > +{
> > + if (mod_mem_use_vmalloc(type))
> > + return vzalloc(size);
> > + return module_alloc(size);
> > +}
> > +
> > +static void module_memory_free(void *ptr, enum mod_mem_type type)
> > +{
> > + if (mod_mem_use_vmalloc(type))
> > + vfree(ptr);
> > + else
> > + module_memfree(ptr);
> > +}
> > +
> > +static void free_mod_mem(struct module *mod)
> > +{
> > + /* free the memory in the right order to avoid use-after-free */
> > + static enum mod_mem_type mod_mem_free_order[MOD_MEM_NUM_TYPES] = {
> > + /* first free init sections */
> > + MOD_INIT_TEXT,
> > + MOD_INIT_DATA,
> > + MOD_INIT_RODATA,
> > +
> > + /* then core sections, except rw data */
> > + MOD_TEXT,
> > + MOD_RODATA,
> > + MOD_RO_AFTER_INIT,
> > +
> > + /* last, rw data */
> > + MOD_DATA,
> > + };
> > + int i;
> > +
> > + for (i = 0; i < MOD_MEM_NUM_TYPES; i++) {
> > + enum mod_mem_type type = mod_mem_free_order[i];
> > + struct module_memory *mod_mem = &mod->mem[type];
> > +
> > + /* Free lock-classes; relies on the preceding sync_rcu(). */
> > + lockdep_free_key_range(mod_mem->base, mod_mem->size);
> > + if (mod_mem->size)
> > + module_memory_free(mod_mem->base, type);
> > + }
> > +}
> > +
> > /* Free a module, remove from lists, etc. */
> > static void free_module(struct module *mod)
> > {
> > @@ -1189,18 +1276,10 @@ static void free_module(struct module *mod)
> >
> > /* This may be empty, but that's OK */
> > module_arch_freeing_init(mod);
> > - module_memfree(mod->init_layout.base);
> > kfree(mod->args);
> > percpu_modfree(mod);
> >
> > - /* Free lock-classes; relies on the preceding sync_rcu(). */
> > - lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> > -
> > - /* Finally, free the core (containing the module structure) */
> > - module_memfree(mod->core_layout.base);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - vfree(mod->data_layout.base);
> > -#endif
> > + free_mod_mem(mod);
> > }
> >
> > void *__symbol_get(const char *symbol)
> > @@ -1387,16 +1466,18 @@ unsigned int __weak arch_mod_section_prepend(struct module *mod,
> > return 0;
> > }
> >
> > -/* Update size with this section: return offset. */
> > -long module_get_offset(struct module *mod, unsigned int *size,
> > +long module_get_offset_and_type(struct module *mod, enum mod_mem_type type,
> > Elf_Shdr *sechdr, unsigned int section)
> > {
> > - long ret;
> > + long offset;
> > + long mask = ((unsigned long)(type) & SH_ENTSIZE_TYPE_MASK) << SH_ENTSIZE_TYPE_SHIFT;
> >
> > - *size += arch_mod_section_prepend(mod, section);
> > - ret = ALIGN(*size, sechdr->sh_addralign ?: 1);
> > - *size = ret + sechdr->sh_size;
> > - return ret;
> > + mod->mem[type].size += arch_mod_section_prepend(mod, section);
> > + offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
> > + mod->mem[type].size = offset + sechdr->sh_size;
> > +
> > + WARN_ON_ONCE(offset & mask);
> > + return offset | mask;
> > }
> >
> > static bool module_init_layout_section(const char *sname)
> > @@ -1428,6 +1509,20 @@ static void layout_sections(struct module *mod, struct load_info *info)
> > { SHF_WRITE | SHF_ALLOC, ARCH_SHF_SMALL },
> > { ARCH_SHF_SMALL | SHF_ALLOC, 0 }
> > };
> > + static int core_m_to_mem_type[] = {
> > + MOD_TEXT,
> > + MOD_RODATA,
> > + MOD_RO_AFTER_INIT,
> > + MOD_DATA,
> > + MOD_INVALID,
> > + };
> > + static int init_m_to_mem_type[] = {
> > + MOD_INIT_TEXT,
> > + MOD_INIT_RODATA,
> > + MOD_INVALID,
> > + MOD_INIT_DATA,
> > + MOD_INVALID,
> > + };
> > unsigned int m, i;
> >
> > for (i = 0; i < info->hdr->e_shnum; i++)
> > @@ -1435,41 +1530,30 @@ static void layout_sections(struct module *mod, struct load_info *info)
> >
> > pr_debug("Core section allocation order:\n");
> > for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> > + enum mod_mem_type type = core_m_to_mem_type[m];
> > +
> > for (i = 0; i < info->hdr->e_shnum; ++i) {
> > Elf_Shdr *s = &info->sechdrs[i];
> > const char *sname = info->secstrings + s->sh_name;
> > - unsigned int *sizep;
> >
> > if ((s->sh_flags & masks[m][0]) != masks[m][0]
> > || (s->sh_flags & masks[m][1])
> > || s->sh_entsize != ~0UL
> > || module_init_layout_section(sname))
> > continue;
> > - sizep = m ? &mod->data_layout.size : &mod->core_layout.size;
> > - s->sh_entsize = module_get_offset(mod, sizep, s, i);
> > +
> > + if (WARN_ON_ONCE(type == MOD_INVALID))
> > + continue;
> > +
> > + s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
> > pr_debug("\t%s\n", sname);
> > }
> > - switch (m) {
> > - case 0: /* executable */
> > - mod->core_layout.size = strict_align(mod->core_layout.size);
> > - mod->core_layout.text_size = mod->core_layout.size;
> > - break;
> > - case 1: /* RO: text and ro-data */
> > - mod->data_layout.size = strict_align(mod->data_layout.size);
> > - mod->data_layout.ro_size = mod->data_layout.size;
> > - break;
> > - case 2: /* RO after init */
> > - mod->data_layout.size = strict_align(mod->data_layout.size);
> > - mod->data_layout.ro_after_init_size = mod->data_layout.size;
> > - break;
> > - case 4: /* whole core */
> > - mod->data_layout.size = strict_align(mod->data_layout.size);
> > - break;
> > - }
> > }
> >
> > pr_debug("Init section allocation order:\n");
> > for (m = 0; m < ARRAY_SIZE(masks); ++m) {
> > + enum mod_mem_type type = init_m_to_mem_type[m];
> > +
> > for (i = 0; i < info->hdr->e_shnum; ++i) {
> > Elf_Shdr *s = &info->sechdrs[i];
> > const char *sname = info->secstrings + s->sh_name;
> > @@ -1479,30 +1563,13 @@ static void layout_sections(struct module *mod, struct load_info *info)
> > || s->sh_entsize != ~0UL
> > || !module_init_layout_section(sname))
> > continue;
> > - s->sh_entsize = (module_get_offset(mod, &mod->init_layout.size, s, i)
> > - | INIT_OFFSET_MASK);
> > +
> > + if (WARN_ON_ONCE(type == MOD_INVALID))
> > + continue;
> > +
> > + s->sh_entsize = module_get_offset_and_type(mod, type, s, i);
> > pr_debug("\t%s\n", sname);
> > }
> > - switch (m) {
> > - case 0: /* executable */
> > - mod->init_layout.size = strict_align(mod->init_layout.size);
> > - mod->init_layout.text_size = mod->init_layout.size;
> > - break;
> > - case 1: /* RO: text and ro-data */
> > - mod->init_layout.size = strict_align(mod->init_layout.size);
> > - mod->init_layout.ro_size = mod->init_layout.size;
> > - break;
> > - case 2:
> > - /*
> > - * RO after init doesn't apply to init_layout (only
> > - * core_layout), so it just takes the value of ro_size.
> > - */
> > - mod->init_layout.ro_after_init_size = mod->init_layout.ro_size;
> > - break;
> > - case 4: /* whole init */
> > - mod->init_layout.size = strict_align(mod->init_layout.size);
> > - break;
> > - }
> > }
> > }
> >
> > @@ -2122,72 +2189,42 @@ static int move_module(struct module *mod, struct load_info *info)
> > {
> > int i;
> > void *ptr;
> > + enum mod_mem_type t;
> >
> > - /* Do the allocs. */
> > - ptr = module_alloc(mod->core_layout.size);
> > - /*
> > - * The pointer to this block is stored in the module structure
> > - * which is inside the block. Just mark it as not being a
> > - * leak.
> > - */
> > - kmemleak_not_leak(ptr);
> > - if (!ptr)
> > - return -ENOMEM;
> > -
> > - memset(ptr, 0, mod->core_layout.size);
> > - mod->core_layout.base = ptr;
> > -
> > - if (mod->init_layout.size) {
> > - ptr = module_alloc(mod->init_layout.size);
> > - /*
> > - * The pointer to this block is stored in the module structure
> > - * which is inside the block. This block doesn't need to be
> > - * scanned as it contains data and code that will be freed
> > - * after the module is initialized.
> > - */
> > - kmemleak_ignore(ptr);
> > - if (!ptr) {
> > - module_memfree(mod->core_layout.base);
> > - return -ENOMEM;
> > + for_each_mod_mem_type(type) {
> > + if (!mod->mem[type].size) {
> > + mod->mem[type].base = NULL;
> > + continue;
> > }
> > - memset(ptr, 0, mod->init_layout.size);
> > - mod->init_layout.base = ptr;
> > - } else
> > - mod->init_layout.base = NULL;
> > + mod->mem[type].size = PAGE_ALIGN(mod->mem[type].size);
> > + ptr = module_memory_alloc(mod->mem[type].size, type);
> >
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - /* Do the allocs. */
> > - ptr = vzalloc(mod->data_layout.size);
> > /*
> > * The pointer to this block is stored in the module structure
> > * which is inside the block. Just mark it as not being a
> > * leak.
> > */
> > - kmemleak_not_leak(ptr);
> > + kmemleak_ignore(ptr);
> > if (!ptr) {
> > - module_memfree(mod->core_layout.base);
> > - module_memfree(mod->init_layout.base);
> > - return -ENOMEM;
> > + t = type;
> > + goto out_enomem;
> > + }
> > + memset(ptr, 0, mod->mem[type].size);
> > + mod->mem[type].base = ptr;
> > }
> >
> > - mod->data_layout.base = ptr;
> > -#endif
> > /* Transfer each section which specifies SHF_ALLOC */
> > pr_debug("final section addresses:\n");
> > for (i = 0; i < info->hdr->e_shnum; i++) {
> > void *dest;
> > Elf_Shdr *shdr = &info->sechdrs[i];
> > + enum mod_mem_type type = shdr->sh_entsize >> SH_ENTSIZE_TYPE_SHIFT;
> >
> > if (!(shdr->sh_flags & SHF_ALLOC))
> > continue;
> >
> > - if (shdr->sh_entsize & INIT_OFFSET_MASK)
> > - dest = mod->init_layout.base
> > - + (shdr->sh_entsize & ~INIT_OFFSET_MASK);
> > - else if (!(shdr->sh_flags & SHF_EXECINSTR))
> > - dest = mod->data_layout.base + shdr->sh_entsize;
> > - else
> > - dest = mod->core_layout.base + shdr->sh_entsize;
> > + dest = mod->mem[type].base +
> > + (shdr->sh_entsize & SH_ENTSIZE_OFFSET_MASK);
> >
> > if (shdr->sh_type != SHT_NOBITS)
> > memcpy(dest, (void *)shdr->sh_addr, shdr->sh_size);
> > @@ -2198,6 +2235,10 @@ static int move_module(struct module *mod, struct load_info *info)
> > }
> >
> > return 0;
> > +out_enomem:
> > + for (t--; t >= 0; t--)
> > + module_memory_free(mod->mem[t].base, t);
> > + return -ENOMEM;
> > }
> >
> > static int check_module_license_and_versions(struct module *mod)
> > @@ -2242,12 +2283,13 @@ static void flush_module_icache(const struct module *mod)
> > * Do it before processing of module parameters, so the module
> > * can provide parameter accessor functions of its own.
> > */
> > - if (mod->init_layout.base)
> > - flush_icache_range((unsigned long)mod->init_layout.base,
> > - (unsigned long)mod->init_layout.base
> > - + mod->init_layout.size);
> > - flush_icache_range((unsigned long)mod->core_layout.base,
> > - (unsigned long)mod->core_layout.base + mod->core_layout.size);
> > + for_each_mod_mem_type(type) {
> > + const struct module_memory *mod_mem = &mod->mem[type];
> > +
> > + if (mod_mem->size)
> > + flush_icache_range((unsigned long)mod_mem->base,
> > + (unsigned long)mod_mem->base + mod_mem->size);
> > + }
> > }
> >
> > bool __weak module_elf_check_arch(Elf_Ehdr *hdr)
> > @@ -2350,11 +2392,8 @@ static void module_deallocate(struct module *mod, struct load_info *info)
> > {
> > percpu_modfree(mod);
> > module_arch_freeing_init(mod);
> > - module_memfree(mod->init_layout.base);
> > - module_memfree(mod->core_layout.base);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - vfree(mod->data_layout.base);
> > -#endif
> > +
> > + free_mod_mem(mod);
> > }
> >
> > int __weak module_finalize(const Elf_Ehdr *hdr,
> > @@ -2415,7 +2454,9 @@ static void do_mod_ctors(struct module *mod)
> > /* For freeing module_init on success, in case kallsyms traversing */
> > struct mod_initfree {
> > struct llist_node node;
> > - void *module_init;
> > + void *init_text;
> > + void *init_data;
> > + void *init_rodata;
> > };
> >
> > static void do_free_init(struct work_struct *w)
> > @@ -2429,7 +2470,9 @@ static void do_free_init(struct work_struct *w)
> >
> > llist_for_each_safe(pos, n, list) {
> > initfree = container_of(pos, struct mod_initfree, node);
> > - module_memfree(initfree->module_init);
> > + module_memfree(initfree->init_text);
> > + module_memfree(initfree->init_data);
> > + module_memfree(initfree->init_rodata);
> > kfree(initfree);
> > }
> > }
> > @@ -2456,7 +2499,9 @@ static noinline int do_init_module(struct module *mod)
> > ret = -ENOMEM;
> > goto fail;
> > }
> > - freeinit->module_init = mod->init_layout.base;
> > + freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
> > + freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
> > + freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
> >
> > do_mod_ctors(mod);
> > /* Start the module */
> > @@ -2492,8 +2537,8 @@ static noinline int do_init_module(struct module *mod)
> > if (!mod->async_probe_requested)
> > async_synchronize_full();
> >
> > - ftrace_free_mem(mod, mod->init_layout.base, mod->init_layout.base +
> > - mod->init_layout.size);
> > + ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
> > + mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
> > mutex_lock(&module_mutex);
> > /* Drop initial reference. */
> > module_put(mod);
> > @@ -2505,11 +2550,10 @@ static noinline int do_init_module(struct module *mod)
> > module_enable_ro(mod, true);
> > mod_tree_remove_init(mod);
> > module_arch_freeing_init(mod);
> > - mod->init_layout.base = NULL;
> > - mod->init_layout.size = 0;
> > - mod->init_layout.ro_size = 0;
> > - mod->init_layout.ro_after_init_size = 0;
> > - mod->init_layout.text_size = 0;
> > + for_class_mod_mem_type(type, init) {
> > + mod->mem[type].base = NULL;
> > + mod->mem[type].size = 0;
> > + }
> > #ifdef CONFIG_DEBUG_INFO_BTF_MODULES
> > /* .BTF is not SHF_ALLOC and will get removed, so sanitize pointer */
> > mod->btf_data = NULL;
> > @@ -2628,9 +2672,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
> > module_bug_finalize(info->hdr, info->sechdrs, mod);
> > module_cfi_finalize(info->hdr, info->sechdrs, mod);
> >
> > - if (module_check_misalignment(mod))
> > - goto out_misaligned;
> > -
> > module_enable_ro(mod, false);
> > module_enable_nx(mod);
> > module_enable_x(mod);
> > @@ -2644,8 +2685,6 @@ static int complete_formation(struct module *mod, struct load_info *info)
> >
> > return 0;
> >
> > -out_misaligned:
> > - err = -EINVAL;
> > out:
> > mutex_unlock(&module_mutex);
> > return err;
> > @@ -2909,7 +2948,10 @@ static int load_module(struct load_info *info, const char __user *uargs,
> > mutex_unlock(&module_mutex);
> > free_module:
> > /* Free lock-classes; relies on the preceding sync_rcu() */
> > - lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size);
> > + for_class_mod_mem_type(type, core) {
> > + lockdep_free_key_range(mod->mem[type].base,
> > + mod->mem[type].size);
> > + }
> >
> > module_deallocate(mod, info);
> > free_copy:
> > @@ -3060,20 +3102,21 @@ bool is_module_address(unsigned long addr)
> > struct module *__module_address(unsigned long addr)
> > {
> > struct module *mod;
> > - struct mod_tree_root *tree;
> >
> > if (addr >= mod_tree.addr_min && addr <= mod_tree.addr_max)
> > - tree = &mod_tree;
> > + goto lookup;
> > +
> > #ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - else if (addr >= mod_data_tree.addr_min && addr <= mod_data_tree.addr_max)
> > - tree = &mod_data_tree;
> > + if (addr >= mod_tree.data_addr_min && addr <= mod_tree.data_addr_max)
> > + goto lookup;
> > #endif
> > - else
> > +
> > return NULL;
> >
> > +lookup:
> > module_assert_mutex_or_preempt();
> >
> > - mod = mod_find(addr, tree);
> > + mod = mod_find(addr, &mod_tree);
> > if (mod) {
> > BUG_ON(!within_module(addr, mod));
> > if (mod->state == MODULE_STATE_UNFORMED)
> > @@ -3113,8 +3156,8 @@ struct module *__module_text_address(unsigned long addr)
> > struct module *mod = __module_address(addr);
> > if (mod) {
> > /* Make sure it's within the text section. */
> > - if (!within(addr, mod->init_layout.base, mod->init_layout.text_size)
> > - && !within(addr, mod->core_layout.base, mod->core_layout.text_size))
> > + if (!within_module_mem_type(addr, mod, MOD_TEXT) &&
> > + !within_module_mem_type(addr, mod, MOD_INIT_TEXT))
> > mod = NULL;
> > }
> > return mod;
> > diff --git a/kernel/module/procfs.c b/kernel/module/procfs.c
> > index cf5b9f1e6ec4..0a4841e88adb 100644
> > --- a/kernel/module/procfs.c
> > +++ b/kernel/module/procfs.c
> > @@ -62,6 +62,15 @@ static void m_stop(struct seq_file *m, void *p)
> > mutex_unlock(&module_mutex);
> > }
> >
> > +static unsigned int module_total_size(struct module *mod)
> > +{
> > + int size = 0;
> > +
> > + for_each_mod_mem_type(type)
> > + size += mod->mem[type].size;
> > + return size;
> > +}
> > +
> > static int m_show(struct seq_file *m, void *p)
> > {
> > struct module *mod = list_entry(p, struct module, list);
> > @@ -73,10 +82,7 @@ static int m_show(struct seq_file *m, void *p)
> > if (mod->state == MODULE_STATE_UNFORMED)
> > return 0;
> >
> > - size = mod->init_layout.size + mod->core_layout.size;
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - size += mod->data_layout.size;
> > -#endif
> > + size = module_total_size(mod);
> > seq_printf(m, "%s %u", mod->name, size);
> > print_unload_info(m, mod);
> >
> > @@ -86,7 +92,7 @@ static int m_show(struct seq_file *m, void *p)
> > mod->state == MODULE_STATE_COMING ? "Loading" :
> > "Live");
> > /* Used by oprofile and other similar tools. */
> > - value = m->private ? NULL : mod->core_layout.base;
> > + value = m->private ? NULL : mod->mem[MOD_TEXT].base;
> > seq_printf(m, " 0x%px", value);
> >
> > /* Taints info */
> > diff --git a/kernel/module/strict_rwx.c b/kernel/module/strict_rwx.c
> > index 14fbea66f12f..2e79a77f40eb 100644
> > --- a/kernel/module/strict_rwx.c
> > +++ b/kernel/module/strict_rwx.c
> > @@ -11,82 +11,26 @@
> > #include <linux/set_memory.h>
> > #include "internal.h"
> >
> > -/*
> > - * LKM RO/NX protection: protect module's text/ro-data
> > - * from modification and any data from execution.
> > - *
> > - * General layout of module is:
> > - * [text] [read-only-data] [ro-after-init] [writable data]
> > - * text_size -----^ ^ ^ ^
> > - * ro_size ------------------------| | |
> > - * ro_after_init_size -----------------------------| |
> > - * size -----------------------------------------------------------|
> > - *
> > - * These values are always page-aligned (as is base) when
> > - * CONFIG_STRICT_MODULE_RWX is set.
> > - */
> > +static void module_set_memory(
> > + const struct module *mod, enum mod_mem_type type,
> > + int (*set_memory)(unsigned long start, int num_pages))
> > +{
> > + const struct module_memory *mod_mem = &mod->mem[type];
> > +
> > + set_vm_flush_reset_perms(mod_mem->base);
> > + set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT);
> > +}
> >
> > /*
> > * Since some arches are moving towards PAGE_KERNEL module allocations instead
> > - * of PAGE_KERNEL_EXEC, keep frob_text() and module_enable_x() independent of
> > + * of PAGE_KERNEL_EXEC, keep module_enable_x() independent of
> > * CONFIG_STRICT_MODULE_RWX because they are needed regardless of whether we
> > * are strict.
> > */
> > -static void frob_text(const struct module_layout *layout,
> > - int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > - set_memory((unsigned long)layout->base,
> > - PAGE_ALIGN(layout->text_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static void frob_rodata(const struct module_layout *layout,
> > - int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > - set_memory((unsigned long)layout->base + layout->text_size,
> > - (layout->ro_size - layout->text_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static void frob_ro_after_init(const struct module_layout *layout,
> > - int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > - set_memory((unsigned long)layout->base + layout->ro_size,
> > - (layout->ro_after_init_size - layout->ro_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static void frob_writable_data(const struct module_layout *layout,
> > - int (*set_memory)(unsigned long start, int num_pages))
> > -{
> > - set_memory((unsigned long)layout->base + layout->ro_after_init_size,
> > - (layout->size - layout->ro_after_init_size) >> PAGE_SHIFT);
> > -}
> > -
> > -static bool layout_check_misalignment(const struct module_layout *layout)
> > -{
> > - return WARN_ON(!PAGE_ALIGNED(layout->base)) ||
> > - WARN_ON(!PAGE_ALIGNED(layout->text_size)) ||
> > - WARN_ON(!PAGE_ALIGNED(layout->ro_size)) ||
> > - WARN_ON(!PAGE_ALIGNED(layout->ro_after_init_size)) ||
> > - WARN_ON(!PAGE_ALIGNED(layout->size));
> > -}
> > -
> > -bool module_check_misalignment(const struct module *mod)
> > -{
> > - if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> > - return false;
> > -
> > - return layout_check_misalignment(&mod->core_layout) ||
> > - layout_check_misalignment(&mod->data_layout) ||
> > - layout_check_misalignment(&mod->init_layout);
> > -}
> > -
> > void module_enable_x(const struct module *mod)
> > {
> > - if (!PAGE_ALIGNED(mod->core_layout.base) ||
> > - !PAGE_ALIGNED(mod->init_layout.base))
> > - return;
> > -
> > - frob_text(&mod->core_layout, set_memory_x);
> > - frob_text(&mod->init_layout, set_memory_x);
> > + for_class_mod_mem_type(type, text)
> > + module_set_memory(mod, type, set_memory_x);
> > }
> >
> > void module_enable_ro(const struct module *mod, bool after_init)
> > @@ -98,16 +42,13 @@ void module_enable_ro(const struct module *mod, bool after_init)
> > return;
> > #endif
> >
> > - set_vm_flush_reset_perms(mod->core_layout.base);
> > - set_vm_flush_reset_perms(mod->init_layout.base);
> > - frob_text(&mod->core_layout, set_memory_ro);
> > -
> > - frob_rodata(&mod->data_layout, set_memory_ro);
> > - frob_text(&mod->init_layout, set_memory_ro);
> > - frob_rodata(&mod->init_layout, set_memory_ro);
> > + module_set_memory(mod, MOD_TEXT, set_memory_ro);
> > + module_set_memory(mod, MOD_INIT_TEXT, set_memory_ro);
> > + module_set_memory(mod, MOD_RODATA, set_memory_ro);
> > + module_set_memory(mod, MOD_INIT_RODATA, set_memory_ro);
> >
> > if (after_init)
> > - frob_ro_after_init(&mod->data_layout, set_memory_ro);
> > + module_set_memory(mod, MOD_RO_AFTER_INIT, set_memory_ro);
> > }
> >
> > void module_enable_nx(const struct module *mod)
> > @@ -115,11 +56,8 @@ void module_enable_nx(const struct module *mod)
> > if (!IS_ENABLED(CONFIG_STRICT_MODULE_RWX))
> > return;
> >
> > - frob_rodata(&mod->data_layout, set_memory_nx);
> > - frob_ro_after_init(&mod->data_layout, set_memory_nx);
> > - frob_writable_data(&mod->data_layout, set_memory_nx);
> > - frob_rodata(&mod->init_layout, set_memory_nx);
> > - frob_writable_data(&mod->init_layout, set_memory_nx);
> > + for_class_mod_mem_type(type, data)
> > + module_set_memory(mod, type, set_memory_nx);
> > }
> >
> > int module_enforce_rwx_sections(Elf_Ehdr *hdr, Elf_Shdr *sechdrs,
> > diff --git a/kernel/module/tree_lookup.c b/kernel/module/tree_lookup.c
> > index 8ec5cfd60496..277197977d43 100644
> > --- a/kernel/module/tree_lookup.c
> > +++ b/kernel/module/tree_lookup.c
> > @@ -21,16 +21,16 @@
> >
> > static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n)
> > {
> > - struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> > + struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
> >
> > - return (unsigned long)layout->base;
> > + return (unsigned long)mod_mem->base;
> > }
> >
> > static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n)
> > {
> > - struct module_layout *layout = container_of(n, struct module_layout, mtn.node);
> > + struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node);
> >
> > - return (unsigned long)layout->size;
> > + return (unsigned long)mod_mem->size;
> > }
> >
> > static __always_inline bool
> > @@ -77,32 +77,27 @@ static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *
> > */
> > void mod_tree_insert(struct module *mod)
> > {
> > - mod->core_layout.mtn.mod = mod;
> > - mod->init_layout.mtn.mod = mod;
> > -
> > - __mod_tree_insert(&mod->core_layout.mtn, &mod_tree);
> > - if (mod->init_layout.size)
> > - __mod_tree_insert(&mod->init_layout.mtn, &mod_tree);
> > -
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - mod->data_layout.mtn.mod = mod;
> > - __mod_tree_insert(&mod->data_layout.mtn, &mod_data_tree);
> > -#endif
> > + for_each_mod_mem_type(type) {
> > + mod->mem[type].mtn.mod = mod;
> > + if (mod->mem[type].size)
> > + __mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
> > + }
> > }
> >
> > void mod_tree_remove_init(struct module *mod)
> > {
> > - if (mod->init_layout.size)
> > - __mod_tree_remove(&mod->init_layout.mtn, &mod_tree);
> > + for_class_mod_mem_type(type, init) {
> > + if (mod->mem[type].size)
> > + __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> > + }
> > }
> >
> > void mod_tree_remove(struct module *mod)
> > {
> > - __mod_tree_remove(&mod->core_layout.mtn, &mod_tree);
> > - mod_tree_remove_init(mod);
> > -#ifdef CONFIG_ARCH_WANTS_MODULES_DATA_IN_VMALLOC
> > - __mod_tree_remove(&mod->data_layout.mtn, &mod_data_tree);
> > -#endif
> > + for_each_mod_mem_type(type) {
> > + if (mod->mem[type].size)
> > + __mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
> > + }
> > }
> >
> > struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)