2022-04-06 12:44:23

by Hyeonggon Yoo

[permalink] [raw]
Subject: Re: [mm/slub] 555b8c8cb3: WARNING:at_lib/stackdepot.c:#stack_depot_fetch

On Tue, Apr 05, 2022 at 01:07:53PM +0200, Marco Elver wrote:
> On Tue, Apr 05, 2022 at 11:00AM +0900, Hyeonggon Yoo wrote:
> > On Mon, Apr 04, 2022 at 05:18:16PM +0200, Marco Elver wrote:
> > > On Mon, 4 Apr 2022 at 16:20, Vlastimil Babka <[email protected]> wrote:
> [...]
> > > > But here we are in mem_dump_obj() -> kmem_dump_obj() -> kmem_obj_info().
> > > > Because kmem_valid_obj() returned true, fooled by folio_test_slab()
> > > > returning true because of the /* Set required slab fields. */ code.
> > > > Yet the illusion is not perfect and we read garbage instead of a valid
> > > > stackdepot handle.
> > > >
> > > > IMHO we should e.g. add the appropriate is_kfence_address() test into
> > > > kmem_valid_obj(), to exclude kfence-allocated objects? Sounds much simpler
> > > > than trying to extend the illusion further to make kmem_dump_obj() work?
> > > > Instead kfence could add its own specific handler to mem_dump_obj() to print
> > > > its debugging data?
> > >
> > > I think this explanation makes sense! Indeed, KFENCE already records
> > > allocation stacks internally anyway, so it should be straightforward
> > > to convince it to just print that.
> > >
> >
> > Thank you both! Yeah the explanation makes sense... thats why KASAN/KCSAN couldn't yield anything -- it was not overwritten.
> >
> > I'm writing a fix and will test if the bug disappears.
> > This may take few days.
>

I did check the bug is not reproduced after simple fix. (reproduced 0 of 373)
This approach was right.

> The below should fix it -- I'd like to make kmem_obj_info() do something
> useful for KFENCE objects.
>

Agreed.

> I lightly tested it by calling mem_dump_obj() on a KFENCE object, and
> prior to the below patch it'd produce garbage data.
>
> Does that look reasonable to you?
>
> Thanks,
> -- Marco
>
> ------ >8 ------
>
> From 09f32964284110846ded8ade9a1a2bfcb17dc58e Mon Sep 17 00:00:00 2001
> From: Marco Elver <[email protected]>
> Date: Tue, 5 Apr 2022 12:43:48 +0200
> Subject: [PATCH RFC] kfence, slab, slub: support kmem_obj_info() with KFENCE
> objects
>
> Calling kmem_obj_info() on KFENCE objects has been producing garbage
> data due to the object not actually being maintained by SLAB or SLUB.
>
> Fix this by asking KFENCE to copy missing KFENCE-specific information to
> struct kmem_obj_info when the object was allocated by KFENCE.
>
> Link: https://lore.kernel.org/all/20220323090520.GG16885@xsang-OptiPlex-9020/
> Fixes: b89fb5ef0ce6 ("mm, kfence: insert KFENCE hooks for SLUB")
> Fixes: d3fb45f370d9 ("mm, kfence: insert KFENCE hooks for SLAB")
> Reported-by: kernel test robot <[email protected]>
> Signed-off-by: Marco Elver <[email protected]>
> ---
> include/linux/kfence.h | 22 ++++++++++++++++++++++
> mm/kfence/core.c | 21 ---------------------
> mm/kfence/kfence.h | 21 +++++++++++++++++++++
> mm/kfence/report.c | 34 ++++++++++++++++++++++++++++++++++
> mm/slab.c | 4 ++++
> mm/slub.c | 4 ++++
> 6 files changed, 85 insertions(+), 21 deletions(-)
>
> diff --git a/include/linux/kfence.h b/include/linux/kfence.h
> index f49e64222628..4a7c633cb219 100644
> --- a/include/linux/kfence.h
> +++ b/include/linux/kfence.h
> @@ -204,6 +204,23 @@ static __always_inline __must_check bool kfence_free(void *addr)
> */
> bool __must_check kfence_handle_page_fault(unsigned long addr, bool is_write, struct pt_regs *regs);
>
> +#ifdef CONFIG_PRINTK
> +struct kmem_obj_info;
> +/**
> + * kfence_kmem_obj_info() - fill kmem_obj_info struct
> + * @kpp: kmem_obj_info to be filled
> + * @object: the object
> + *
> + * Return:
> + * * false - not a KFENCE object
> + * * true - a KFENCE object and filled @kpp
> + *
> + * Copies information to @kpp that kmem_obj_info() is unable to populate for
> + * KFENCE objects.
> + */
> +bool kfence_kmem_obj_info(struct kmem_obj_info *kpp, void *object);
> +#endif
> +
> #else /* CONFIG_KFENCE */
>
> static inline bool is_kfence_address(const void *addr) { return false; }
> @@ -221,6 +238,11 @@ static inline bool __must_check kfence_handle_page_fault(unsigned long addr, boo
> return false;
> }
>
> +#ifdef CONFIG_PRINTK
> +struct kmem_obj_info;
> +static inline bool kfence_kmem_obj_info(struct kmem_obj_info *kpp, void *object) { return false; }
> +#endif
> +
> #endif
>
> #endif /* _LINUX_KFENCE_H */
> diff --git a/mm/kfence/core.c b/mm/kfence/core.c
> index a203747ad2c0..9b2b5f56f4ae 100644
> --- a/mm/kfence/core.c
> +++ b/mm/kfence/core.c
> @@ -231,27 +231,6 @@ static bool kfence_unprotect(unsigned long addr)
> return !KFENCE_WARN_ON(!kfence_protect_page(ALIGN_DOWN(addr, PAGE_SIZE), false));
> }
>
> -static inline struct kfence_metadata *addr_to_metadata(unsigned long addr)
> -{
> - long index;
> -
> - /* The checks do not affect performance; only called from slow-paths. */
> -
> - if (!is_kfence_address((void *)addr))
> - return NULL;
> -
> - /*
> - * May be an invalid index if called with an address at the edge of
> - * __kfence_pool, in which case we would report an "invalid access"
> - * error.
> - */
> - index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1;
> - if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS)
> - return NULL;
> -
> - return &kfence_metadata[index];
> -}
> -
> static inline unsigned long metadata_to_pageaddr(const struct kfence_metadata *meta)
> {
> unsigned long offset = (meta - kfence_metadata + 1) * PAGE_SIZE * 2;
> diff --git a/mm/kfence/kfence.h b/mm/kfence/kfence.h
> index 9a6c4b1b12a8..600f2e2431d6 100644
> --- a/mm/kfence/kfence.h
> +++ b/mm/kfence/kfence.h
> @@ -96,6 +96,27 @@ struct kfence_metadata {
>
> extern struct kfence_metadata kfence_metadata[CONFIG_KFENCE_NUM_OBJECTS];
>
> +static inline struct kfence_metadata *addr_to_metadata(unsigned long addr)
> +{
> + long index;
> +
> + /* The checks do not affect performance; only called from slow-paths. */
> +
> + if (!is_kfence_address((void *)addr))
> + return NULL;
> +
> + /*
> + * May be an invalid index if called with an address at the edge of
> + * __kfence_pool, in which case we would report an "invalid access"
> + * error.
> + */
> + index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1;
> + if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS)
> + return NULL;
> +
> + return &kfence_metadata[index];
> +}
> +
> /* KFENCE error types for report generation. */
> enum kfence_error_type {
> KFENCE_ERROR_OOB, /* Detected a out-of-bounds access. */
> diff --git a/mm/kfence/report.c b/mm/kfence/report.c
> index f93a7b2a338b..5887fa610c9d 100644
> --- a/mm/kfence/report.c
> +++ b/mm/kfence/report.c
> @@ -273,3 +273,37 @@ void kfence_report_error(unsigned long address, bool is_write, struct pt_regs *r
> /* We encountered a memory safety error, taint the kernel! */
> add_taint(TAINT_BAD_PAGE, LOCKDEP_STILL_OK);
> }
> +
> +#ifdef CONFIG_PRINTK
> +static void kfence_to_kp_stack(const struct kfence_track *track, void **kp_stack)
> +{
> + int i, j;
> +
> + i = get_stack_skipnr(track->stack_entries, track->num_stack_entries, NULL);
> + for (j = 0; i < track->num_stack_entries && j < KS_ADDRS_COUNT - 1; ++i, ++j)

why KS_ADDRS_COUNT - 1 instead of KS_ADDRS_COUNT?

> + kp_stack[j] = (void *)track->stack_entries[i];
> + kp_stack[j] = NULL;
> +}
> +
> +bool kfence_kmem_obj_info(struct kmem_obj_info *kpp, void *object)
> +{
> + const struct kfence_metadata *meta = addr_to_metadata((unsigned long)object);
> +
> + if (!meta)
> + return false;
> +
> + /* Requesting info an a never-used object is almost certainly a bug. */
> + if (WARN_ON(meta->state == KFENCE_OBJECT_UNUSED))
> + return true;
> +
> + kpp->kp_objp = (void *)meta->addr;
> +

no need to take meta->lock here?

> + kfence_to_kp_stack(&meta->alloc_track, kpp->kp_stack);
> + if (meta->state == KFENCE_OBJECT_FREED)
> + kfence_to_kp_stack(&meta->free_track, kpp->kp_free_stack);
> + /* get_stack_skipnr() ensures the first entry is outside allocator. */
> + kpp->kp_ret = kpp->kp_stack[0];
> +
> + return true;
> +}

kfence_kmem_obj_info() does not set kp_data_offset. kp_data_offset
may not be zero when e.g.) mem_dump_obj(&rhp->func); in rcutorture case.

BTW, I would prefer implementing something like kfence_obj_info()
(called by kmem_dump_obj() and called instead of kmem_obj_info())
for better readability.

And when mem_dump_obj() is called, I guess it's for debugging purpose.
I think it would be better to let user know the object is allocated
from kfence pool. maybe adding if (is_kfence_address(object)) pr_cont(" kfence");
in kmem_dump_obj() would be enough?

Thanks!
Hyeonggon

> +#endif
> diff --git a/mm/slab.c b/mm/slab.c
> index b04e40078bdf..4d44b094e0ab 100644
> --- a/mm/slab.c
> +++ b/mm/slab.c
> @@ -3675,6 +3675,10 @@ void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> kpp->kp_slab = slab;
> cachep = slab->slab_cache;
> kpp->kp_slab_cache = cachep;
> +
> + if (kfence_kmem_obj_info(kpp, object))
> + return;
> +
> objp = object - obj_offset(cachep);
> kpp->kp_data_offset = obj_offset(cachep);
> slab = virt_to_slab(objp);
> diff --git a/mm/slub.c b/mm/slub.c
> index 74d92aa4a3a2..c7d2cfd60b87 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -4325,6 +4325,10 @@ void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> kpp->kp_ptr = object;
> kpp->kp_slab = slab;
> kpp->kp_slab_cache = s;
> +
> + if (kfence_kmem_obj_info(kpp, object))
> + return;
> +
> base = slab_address(slab);
> objp0 = kasan_reset_tag(object);
> #ifdef CONFIG_SLUB_DEBUG
> --
> 2.35.1.1094.g7c7d902a7c-goog
>


2022-04-06 14:27:26

by Marco Elver

[permalink] [raw]
Subject: Re: [mm/slub] 555b8c8cb3: WARNING:at_lib/stackdepot.c:#stack_depot_fetch

On Wed, Apr 06, 2022 at 07:47AM +0900, Hyeonggon Yoo wrote:
> On Tue, Apr 05, 2022 at 01:07:53PM +0200, Marco Elver wrote:
> > On Tue, Apr 05, 2022 at 11:00AM +0900, Hyeonggon Yoo wrote:
> > > On Mon, Apr 04, 2022 at 05:18:16PM +0200, Marco Elver wrote:
> > > > On Mon, 4 Apr 2022 at 16:20, Vlastimil Babka <[email protected]> wrote:
> > [...]
> > > > > But here we are in mem_dump_obj() -> kmem_dump_obj() -> kmem_obj_info().
> > > > > Because kmem_valid_obj() returned true, fooled by folio_test_slab()
> > > > > returning true because of the /* Set required slab fields. */ code.
> > > > > Yet the illusion is not perfect and we read garbage instead of a valid
> > > > > stackdepot handle.
> > > > >
> > > > > IMHO we should e.g. add the appropriate is_kfence_address() test into
> > > > > kmem_valid_obj(), to exclude kfence-allocated objects? Sounds much simpler
> > > > > than trying to extend the illusion further to make kmem_dump_obj() work?
> > > > > Instead kfence could add its own specific handler to mem_dump_obj() to print
> > > > > its debugging data?
> > > >
> > > > I think this explanation makes sense! Indeed, KFENCE already records
> > > > allocation stacks internally anyway, so it should be straightforward
> > > > to convince it to just print that.
> > > >
> > >
> > > Thank you both! Yeah the explanation makes sense... thats why KASAN/KCSAN couldn't yield anything -- it was not overwritten.
> > >
> > > I'm writing a fix and will test if the bug disappears.
> > > This may take few days.
> >
>
> I did check the bug is not reproduced after simple fix. (reproduced 0 of 373)
> This approach was right.
>
> > The below should fix it -- I'd like to make kmem_obj_info() do something
> > useful for KFENCE objects.
> >
>
> Agreed.
>
[...]
> > + i = get_stack_skipnr(track->stack_entries, track->num_stack_entries, NULL);
> > + for (j = 0; i < track->num_stack_entries && j < KS_ADDRS_COUNT - 1; ++i, ++j)
>
> why KS_ADDRS_COUNT - 1 instead of KS_ADDRS_COUNT?

For `kp_stack[j] = NULL` because KFENCE's stack_entries does not have a
NULL-delimiter (we have num_stack_entries). But it seems for kp_stack
it's only added if `j < KS_ADDR_COUNT`, so I've fixed that.

> > + kp_stack[j] = (void *)track->stack_entries[i];
> > + kp_stack[j] = NULL;
[...]
> > + kpp->kp_objp = (void *)meta->addr;
> > +
>
> no need to take meta->lock here?

Yes, in case state is KFENCE_OBJECT_FREED there could be a race.

> > + kfence_to_kp_stack(&meta->alloc_track, kpp->kp_stack);
> > + if (meta->state == KFENCE_OBJECT_FREED)
> > + kfence_to_kp_stack(&meta->free_track, kpp->kp_free_stack);
> > + /* get_stack_skipnr() ensures the first entry is outside allocator. */
> > + kpp->kp_ret = kpp->kp_stack[0];
> > +
> > + return true;
> > +}
>
> kfence_kmem_obj_info() does not set kp_data_offset. kp_data_offset
> may not be zero when e.g.) mem_dump_obj(&rhp->func); in rcutorture case.

kp_data_offset is the offset e.g. when SLUB has added a redzone:

| objp0 = kasan_reset_tag(object);
| #ifdef CONFIG_SLUB_DEBUG
| objp = restore_red_left(s, objp0);
| #else
| objp = objp0;
| #endif
| objnr = obj_to_index(s, slab, objp);
| kpp->kp_data_offset = (unsigned long)((char *)objp0 - (char *)objp);

In !CONFIG_SLUB_DEBUG and !(s->flags & SLAB_RED_ZONE) cases it's always
0, and otherwise it's

`objp0 - restore_red_left(objp0)` ==
`object - (object - s->red_left_pad)` ==
`s->red_left_pad`.

This matters if kp_objp is not the object start accessible by the user.
But in the KFENCE case this is always the case so kp_data_offset=0.

> BTW, I would prefer implementing something like kfence_obj_info()
> (called by kmem_dump_obj() and called instead of kmem_obj_info())
> for better readability.

Hmm, I guess that saves us from having to fix up both slab.c/slub.c. But
it makes kmem_obj_info() error-prone to use. What if someone calls
kmem_obj_info() in future somewhere else? That caller then would have to
remember to also call kfence_obj_info().

I'd prefer fixing it as close to the root-cause (in kmem_obj_info()) to
avoid that.

What do you prefer?

> And when mem_dump_obj() is called, I guess it's for debugging purpose.
> I think it would be better to let user know the object is allocated
> from kfence pool. maybe adding if (is_kfence_address(object)) pr_cont(" kfence");
> in kmem_dump_obj() would be enough?

We can add that.

Thanks,
-- Marco

2022-04-06 15:45:24

by Hyeonggon Yoo

[permalink] [raw]
Subject: Re: [mm/slub] 555b8c8cb3: WARNING:at_lib/stackdepot.c:#stack_depot_fetch

On Wed, Apr 06, 2022 at 10:34:07AM +0200, Marco Elver wrote:
> On Wed, Apr 06, 2022 at 07:47AM +0900, Hyeonggon Yoo wrote:
> > On Tue, Apr 05, 2022 at 01:07:53PM +0200, Marco Elver wrote:
> > > On Tue, Apr 05, 2022 at 11:00AM +0900, Hyeonggon Yoo wrote:
> > > > On Mon, Apr 04, 2022 at 05:18:16PM +0200, Marco Elver wrote:
> > > > > On Mon, 4 Apr 2022 at 16:20, Vlastimil Babka <[email protected]> wrote:
> > > [...]
> > > > > > But here we are in mem_dump_obj() -> kmem_dump_obj() -> kmem_obj_info().
> > > > > > Because kmem_valid_obj() returned true, fooled by folio_test_slab()
> > > > > > returning true because of the /* Set required slab fields. */ code.
> > > > > > Yet the illusion is not perfect and we read garbage instead of a valid
> > > > > > stackdepot handle.
> > > > > >
> > > > > > IMHO we should e.g. add the appropriate is_kfence_address() test into
> > > > > > kmem_valid_obj(), to exclude kfence-allocated objects? Sounds much simpler
> > > > > > than trying to extend the illusion further to make kmem_dump_obj() work?
> > > > > > Instead kfence could add its own specific handler to mem_dump_obj() to print
> > > > > > its debugging data?
> > > > >
> > > > > I think this explanation makes sense! Indeed, KFENCE already records
> > > > > allocation stacks internally anyway, so it should be straightforward
> > > > > to convince it to just print that.
> > > > >
> > > >
> > > > Thank you both! Yeah the explanation makes sense... thats why KASAN/KCSAN couldn't yield anything -- it was not overwritten.
> > > >
> > > > I'm writing a fix and will test if the bug disappears.
> > > > This may take few days.
> > >
> >
> > I did check the bug is not reproduced after simple fix. (reproduced 0 of 373)
> > This approach was right.
> >
> > > The below should fix it -- I'd like to make kmem_obj_info() do something
> > > useful for KFENCE objects.
> > >
> >
> > Agreed.
> >
> [...]
> > > + i = get_stack_skipnr(track->stack_entries, track->num_stack_entries, NULL);
> > > + for (j = 0; i < track->num_stack_entries && j < KS_ADDRS_COUNT - 1; ++i, ++j)
> >
> > why KS_ADDRS_COUNT - 1 instead of KS_ADDRS_COUNT?
>
> For `kp_stack[j] = NULL` because KFENCE's stack_entries does not have a
> NULL-delimiter (we have num_stack_entries). But it seems for kp_stack
> it's only added if `j < KS_ADDR_COUNT`, so I've fixed that.
>

Okay.

> > > + kp_stack[j] = (void *)track->stack_entries[i];
> > > + kp_stack[j] = NULL;
> [...]
> > > + kpp->kp_objp = (void *)meta->addr;
> > > +
> >
> > no need to take meta->lock here?
>
> Yes, in case state is KFENCE_OBJECT_FREED there could be a race.
>

dumping object that is freed sounds like a bug?
but seems it's better to avoid a race in that case too.

> > > + kfence_to_kp_stack(&meta->alloc_track, kpp->kp_stack);
> > > + if (meta->state == KFENCE_OBJECT_FREED)
> > > + kfence_to_kp_stack(&meta->free_track, kpp->kp_free_stack);
> > > + /* get_stack_skipnr() ensures the first entry is outside allocator. */
> > > + kpp->kp_ret = kpp->kp_stack[0];
> > > +
> > > + return true;
> > > +}
> >
> > kfence_kmem_obj_info() does not set kp_data_offset. kp_data_offset
> > may not be zero when e.g.) mem_dump_obj(&rhp->func); in rcutorture case.
>
> kp_data_offset is the offset e.g. when SLUB has added a redzone:
>
> | objp0 = kasan_reset_tag(object);
> | #ifdef CONFIG_SLUB_DEBUG
> | objp = restore_red_left(s, objp0);
> | #else
> | objp = objp0;
> | #endif
> | objnr = obj_to_index(s, slab, objp);
> | kpp->kp_data_offset = (unsigned long)((char *)objp0 - (char *)objp);
>
> In !CONFIG_SLUB_DEBUG and !(s->flags & SLAB_RED_ZONE) cases it's always
> 0, and otherwise it's
>
> `objp0 - restore_red_left(objp0)` ==
> `object - (object - s->red_left_pad)` ==
> `s->red_left_pad`.
>
> This matters if kp_objp is not the object start accessible by the user.
> But in the KFENCE case this is always the case so kp_data_offset=0.
>

Ah, right. confused "pointer offset" with "data offset".
it's always 0 in kfence case.

> > BTW, I would prefer implementing something like kfence_obj_info()
> > (called by kmem_dump_obj() and called instead of kmem_obj_info())
> > for better readability.
>
> Hmm, I guess that saves us from having to fix up both slab.c/slub.c. But
> it makes kmem_obj_info() error-prone to use. What if someone calls
> kmem_obj_info() in future somewhere else? That caller then would have to
> remember to also call kfence_obj_info().
>

Valid point.

> I'd prefer fixing it as close to the root-cause (in kmem_obj_info()) to
> avoid that.
>
> What do you prefer?

Then what about something like this?

kmem_obj_info(object) {
if (object is from kfence)
__kfence_obj_info()
else
__kmem_obj_info()
}

>
> > And when mem_dump_obj() is called, I guess it's for debugging purpose.
> > I think it would be better to let user know the object is allocated
> > from kfence pool. maybe adding if (is_kfence_address(object)) pr_cont(" kfence");
> > in kmem_dump_obj() would be enough?
>
> We can add that.
>
> Thanks,
> -- Marco

--
Thanks,
Hyeonggon

2022-04-06 16:17:09

by Marco Elver

[permalink] [raw]
Subject: Re: [mm/slub] 555b8c8cb3: WARNING:at_lib/stackdepot.c:#stack_depot_fetch

On Wed, Apr 06, 2022 at 08:50PM +0900, Hyeonggon Yoo wrote:
[...]
> > I'd prefer fixing it as close to the root-cause (in kmem_obj_info()) to
> > avoid that.
> >
> > What do you prefer?
>
> Then what about something like this?
>
> kmem_obj_info(object) {
> if (object is from kfence)
> __kfence_obj_info()
> else
> __kmem_obj_info()
> }

How about the below?

Thanks,
-- Marco

------ >8 ------

From b1c756f522cfecf6f5016581839c693eccaa5827 Mon Sep 17 00:00:00 2001
From: Marco Elver <[email protected]>
Date: Tue, 5 Apr 2022 12:43:48 +0200
Subject: [PATCH RFC] kfence, slab, slub: support kmem_obj_info() with KFENCE
objects

Calling kmem_obj_info() on KFENCE objects has been producing garbage
data due to the object not actually being maintained by SLAB or SLUB.

Fix this by asking KFENCE to copy missing KFENCE-specific information to
struct kmem_obj_info when the object was allocated by KFENCE. For
completeness, make kmem_dump_obj() show if it's a KFENCE object.

Link: https://lore.kernel.org/all/20220323090520.GG16885@xsang-OptiPlex-9020/
Fixes: b89fb5ef0ce6 ("mm, kfence: insert KFENCE hooks for SLUB")
Fixes: d3fb45f370d9 ("mm, kfence: insert KFENCE hooks for SLAB")
Reported-by: kernel test robot <[email protected]>
Signed-off-by: Marco Elver <[email protected]>
---
include/linux/kfence.h | 24 +++++++++++++++++++++
mm/kfence/core.c | 21 -------------------
mm/kfence/kfence.h | 21 +++++++++++++++++++
mm/kfence/report.c | 47 ++++++++++++++++++++++++++++++++++++++++++
mm/slab.c | 2 +-
mm/slab.h | 2 +-
mm/slab_common.c | 9 ++++++++
mm/slob.c | 2 +-
mm/slub.c | 2 +-
9 files changed, 105 insertions(+), 25 deletions(-)

diff --git a/include/linux/kfence.h b/include/linux/kfence.h
index f49e64222628..726857a4b680 100644
--- a/include/linux/kfence.h
+++ b/include/linux/kfence.h
@@ -204,6 +204,22 @@ static __always_inline __must_check bool kfence_free(void *addr)
*/
bool __must_check kfence_handle_page_fault(unsigned long addr, bool is_write, struct pt_regs *regs);

+#ifdef CONFIG_PRINTK
+struct kmem_obj_info;
+/**
+ * __kfence_obj_info() - fill kmem_obj_info struct
+ * @kpp: kmem_obj_info to be filled
+ * @object: the object
+ *
+ * Return:
+ * * false - not a KFENCE object
+ * * true - a KFENCE object, filled @kpp
+ *
+ * Copies information to @kpp for KFENCE objects.
+ */
+bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
+#endif
+
#else /* CONFIG_KFENCE */

static inline bool is_kfence_address(const void *addr) { return false; }
@@ -221,6 +237,14 @@ static inline bool __must_check kfence_handle_page_fault(unsigned long addr, boo
return false;
}

+#ifdef CONFIG_PRINTK
+struct kmem_obj_info;
+static inline bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
+{
+ return false;
+}
+#endif
+
#endif

#endif /* _LINUX_KFENCE_H */
diff --git a/mm/kfence/core.c b/mm/kfence/core.c
index a203747ad2c0..9b2b5f56f4ae 100644
--- a/mm/kfence/core.c
+++ b/mm/kfence/core.c
@@ -231,27 +231,6 @@ static bool kfence_unprotect(unsigned long addr)
return !KFENCE_WARN_ON(!kfence_protect_page(ALIGN_DOWN(addr, PAGE_SIZE), false));
}

-static inline struct kfence_metadata *addr_to_metadata(unsigned long addr)
-{
- long index;
-
- /* The checks do not affect performance; only called from slow-paths. */
-
- if (!is_kfence_address((void *)addr))
- return NULL;
-
- /*
- * May be an invalid index if called with an address at the edge of
- * __kfence_pool, in which case we would report an "invalid access"
- * error.
- */
- index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1;
- if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS)
- return NULL;
-
- return &kfence_metadata[index];
-}
-
static inline unsigned long metadata_to_pageaddr(const struct kfence_metadata *meta)
{
unsigned long offset = (meta - kfence_metadata + 1) * PAGE_SIZE * 2;
diff --git a/mm/kfence/kfence.h b/mm/kfence/kfence.h
index 9a6c4b1b12a8..600f2e2431d6 100644
--- a/mm/kfence/kfence.h
+++ b/mm/kfence/kfence.h
@@ -96,6 +96,27 @@ struct kfence_metadata {

extern struct kfence_metadata kfence_metadata[CONFIG_KFENCE_NUM_OBJECTS];

+static inline struct kfence_metadata *addr_to_metadata(unsigned long addr)
+{
+ long index;
+
+ /* The checks do not affect performance; only called from slow-paths. */
+
+ if (!is_kfence_address((void *)addr))
+ return NULL;
+
+ /*
+ * May be an invalid index if called with an address at the edge of
+ * __kfence_pool, in which case we would report an "invalid access"
+ * error.
+ */
+ index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1;
+ if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS)
+ return NULL;
+
+ return &kfence_metadata[index];
+}
+
/* KFENCE error types for report generation. */
enum kfence_error_type {
KFENCE_ERROR_OOB, /* Detected a out-of-bounds access. */
diff --git a/mm/kfence/report.c b/mm/kfence/report.c
index f93a7b2a338b..f5a6d8ba3e21 100644
--- a/mm/kfence/report.c
+++ b/mm/kfence/report.c
@@ -273,3 +273,50 @@ void kfence_report_error(unsigned long address, bool is_write, struct pt_regs *r
/* We encountered a memory safety error, taint the kernel! */
add_taint(TAINT_BAD_PAGE, LOCKDEP_STILL_OK);
}
+
+#ifdef CONFIG_PRINTK
+static void kfence_to_kp_stack(const struct kfence_track *track, void **kp_stack)
+{
+ int i, j;
+
+ i = get_stack_skipnr(track->stack_entries, track->num_stack_entries, NULL);
+ for (j = 0; i < track->num_stack_entries && j < KS_ADDRS_COUNT; ++i, ++j)
+ kp_stack[j] = (void *)track->stack_entries[i];
+ if (j < KS_ADDRS_COUNT)
+ kp_stack[j] = NULL;
+}
+
+bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
+{
+ struct kfence_metadata *meta = addr_to_metadata((unsigned long)object);
+ unsigned long flags;
+
+ if (!meta)
+ return false;
+
+ /*
+ * If state is UNUSED at least show the pointer requested; the rest
+ * would be garbage data.
+ */
+ kpp->kp_ptr = object;
+
+ /* Requesting info an a never-used object is almost certainly a bug. */
+ if (WARN_ON(meta->state == KFENCE_OBJECT_UNUSED))
+ return true;
+
+ raw_spin_lock_irqsave(&meta->lock, flags);
+
+ kpp->kp_slab = slab;
+ kpp->kp_slab_cache = meta->cache;
+ kpp->kp_objp = (void *)meta->addr;
+ kfence_to_kp_stack(&meta->alloc_track, kpp->kp_stack);
+ if (meta->state == KFENCE_OBJECT_FREED)
+ kfence_to_kp_stack(&meta->free_track, kpp->kp_free_stack);
+ /* get_stack_skipnr() ensures the first entry is outside allocator. */
+ kpp->kp_ret = kpp->kp_stack[0];
+
+ raw_spin_unlock_irqrestore(&meta->lock, flags);
+
+ return true;
+}
+#endif
diff --git a/mm/slab.c b/mm/slab.c
index b04e40078bdf..0edb474edef1 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -3665,7 +3665,7 @@ EXPORT_SYMBOL(__kmalloc_node_track_caller);
#endif /* CONFIG_NUMA */

#ifdef CONFIG_PRINTK
-void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
+void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
{
struct kmem_cache *cachep;
unsigned int objnr;
diff --git a/mm/slab.h b/mm/slab.h
index fd7ae2024897..95eb34174c1b 100644
--- a/mm/slab.h
+++ b/mm/slab.h
@@ -868,7 +868,7 @@ struct kmem_obj_info {
void *kp_stack[KS_ADDRS_COUNT];
void *kp_free_stack[KS_ADDRS_COUNT];
};
-void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
+void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
#endif

#ifdef CONFIG_HAVE_HARDENED_USERCOPY_ALLOCATOR
diff --git a/mm/slab_common.c b/mm/slab_common.c
index 6ee64d6208b3..2b3206a2c3b5 100644
--- a/mm/slab_common.c
+++ b/mm/slab_common.c
@@ -555,6 +555,13 @@ bool kmem_valid_obj(void *object)
}
EXPORT_SYMBOL_GPL(kmem_valid_obj);

+static void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
+{
+ if (__kfence_obj_info(kpp, object, slab))
+ return;
+ __kmem_obj_info(kpp, object, slab);
+}
+
/**
* kmem_dump_obj - Print available slab provenance information
* @object: slab object for which to find provenance information.
@@ -590,6 +597,8 @@ void kmem_dump_obj(void *object)
pr_cont(" slab%s %s", cp, kp.kp_slab_cache->name);
else
pr_cont(" slab%s", cp);
+ if (is_kfence_address(object))
+ pr_cont(" (kfence)");
if (kp.kp_objp)
pr_cont(" start %px", kp.kp_objp);
if (kp.kp_data_offset)
diff --git a/mm/slob.c b/mm/slob.c
index dfa6808dff36..40ea6e2d4ccd 100644
--- a/mm/slob.c
+++ b/mm/slob.c
@@ -463,7 +463,7 @@ static void slob_free(void *block, int size)
}

#ifdef CONFIG_PRINTK
-void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
+void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
{
kpp->kp_ptr = object;
kpp->kp_slab = slab;
diff --git a/mm/slub.c b/mm/slub.c
index 74d92aa4a3a2..ed5c2c03a47a 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4312,7 +4312,7 @@ int __kmem_cache_shutdown(struct kmem_cache *s)
}

#ifdef CONFIG_PRINTK
-void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
+void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
{
void *base;
int __maybe_unused i;
--
2.35.1.1094.g7c7d902a7c-goog

2022-04-06 16:22:25

by Hyeonggon Yoo

[permalink] [raw]
Subject: Re: [mm/slub] 555b8c8cb3: WARNING:at_lib/stackdepot.c:#stack_depot_fetch

On Wed, Apr 06, 2022 at 02:15:35PM +0200, Marco Elver wrote:
> On Wed, Apr 06, 2022 at 08:50PM +0900, Hyeonggon Yoo wrote:
> [...]
> > > I'd prefer fixing it as close to the root-cause (in kmem_obj_info()) to
> > > avoid that.
> > >
> > > What do you prefer?
> >
> > Then what about something like this?
> >
> > kmem_obj_info(object) {
> > if (object is from kfence)
> > __kfence_obj_info()
> > else
> > __kmem_obj_info()
> > }
>
> How about the below?
>
> Thanks,
> -- Marco
>
> ------ >8 ------
>
> From b1c756f522cfecf6f5016581839c693eccaa5827 Mon Sep 17 00:00:00 2001
> From: Marco Elver <[email protected]>
> Date: Tue, 5 Apr 2022 12:43:48 +0200
> Subject: [PATCH RFC] kfence, slab, slub: support kmem_obj_info() with KFENCE
> objects
>
> Calling kmem_obj_info() on KFENCE objects has been producing garbage
> data due to the object not actually being maintained by SLAB or SLUB.
>
> Fix this by asking KFENCE to copy missing KFENCE-specific information to
> struct kmem_obj_info when the object was allocated by KFENCE. For
> completeness, make kmem_dump_obj() show if it's a KFENCE object.
>

This changelog need to be updated.
Otherwise the code looks good to me.

Reviewed-by: Hyeonggon Yoo <[email protected]>

> Link: https://lore.kernel.org/all/20220323090520.GG16885@xsang-OptiPlex-9020/
> Fixes: b89fb5ef0ce6 ("mm, kfence: insert KFENCE hooks for SLUB")
> Fixes: d3fb45f370d9 ("mm, kfence: insert KFENCE hooks for SLAB")
> Reported-by: kernel test robot <[email protected]>
> Signed-off-by: Marco Elver <[email protected]>
> ---
> include/linux/kfence.h | 24 +++++++++++++++++++++
> mm/kfence/core.c | 21 -------------------
> mm/kfence/kfence.h | 21 +++++++++++++++++++
> mm/kfence/report.c | 47 ++++++++++++++++++++++++++++++++++++++++++
> mm/slab.c | 2 +-
> mm/slab.h | 2 +-
> mm/slab_common.c | 9 ++++++++
> mm/slob.c | 2 +-
> mm/slub.c | 2 +-
> 9 files changed, 105 insertions(+), 25 deletions(-)
>
> diff --git a/include/linux/kfence.h b/include/linux/kfence.h
> index f49e64222628..726857a4b680 100644
> --- a/include/linux/kfence.h
> +++ b/include/linux/kfence.h
> @@ -204,6 +204,22 @@ static __always_inline __must_check bool kfence_free(void *addr)
> */
> bool __must_check kfence_handle_page_fault(unsigned long addr, bool is_write, struct pt_regs *regs);
>
> +#ifdef CONFIG_PRINTK
> +struct kmem_obj_info;
> +/**
> + * __kfence_obj_info() - fill kmem_obj_info struct
> + * @kpp: kmem_obj_info to be filled
> + * @object: the object
> + *
> + * Return:
> + * * false - not a KFENCE object
> + * * true - a KFENCE object, filled @kpp
> + *
> + * Copies information to @kpp for KFENCE objects.
> + */
> +bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
> +#endif
> +
> #else /* CONFIG_KFENCE */
>
> static inline bool is_kfence_address(const void *addr) { return false; }
> @@ -221,6 +237,14 @@ static inline bool __must_check kfence_handle_page_fault(unsigned long addr, boo
> return false;
> }
>
> +#ifdef CONFIG_PRINTK
> +struct kmem_obj_info;
> +static inline bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> +{
> + return false;
> +}
> +#endif
> +
> #endif
>
> #endif /* _LINUX_KFENCE_H */
> diff --git a/mm/kfence/core.c b/mm/kfence/core.c
> index a203747ad2c0..9b2b5f56f4ae 100644
> --- a/mm/kfence/core.c
> +++ b/mm/kfence/core.c
> @@ -231,27 +231,6 @@ static bool kfence_unprotect(unsigned long addr)
> return !KFENCE_WARN_ON(!kfence_protect_page(ALIGN_DOWN(addr, PAGE_SIZE), false));
> }
>
> -static inline struct kfence_metadata *addr_to_metadata(unsigned long addr)
> -{
> - long index;
> -
> - /* The checks do not affect performance; only called from slow-paths. */
> -
> - if (!is_kfence_address((void *)addr))
> - return NULL;
> -
> - /*
> - * May be an invalid index if called with an address at the edge of
> - * __kfence_pool, in which case we would report an "invalid access"
> - * error.
> - */
> - index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1;
> - if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS)
> - return NULL;
> -
> - return &kfence_metadata[index];
> -}
> -
> static inline unsigned long metadata_to_pageaddr(const struct kfence_metadata *meta)
> {
> unsigned long offset = (meta - kfence_metadata + 1) * PAGE_SIZE * 2;
> diff --git a/mm/kfence/kfence.h b/mm/kfence/kfence.h
> index 9a6c4b1b12a8..600f2e2431d6 100644
> --- a/mm/kfence/kfence.h
> +++ b/mm/kfence/kfence.h
> @@ -96,6 +96,27 @@ struct kfence_metadata {
>
> extern struct kfence_metadata kfence_metadata[CONFIG_KFENCE_NUM_OBJECTS];
>
> +static inline struct kfence_metadata *addr_to_metadata(unsigned long addr)
> +{
> + long index;
> +
> + /* The checks do not affect performance; only called from slow-paths. */
> +
> + if (!is_kfence_address((void *)addr))
> + return NULL;
> +
> + /*
> + * May be an invalid index if called with an address at the edge of
> + * __kfence_pool, in which case we would report an "invalid access"
> + * error.
> + */
> + index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1;
> + if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS)
> + return NULL;
> +
> + return &kfence_metadata[index];
> +}
> +
> /* KFENCE error types for report generation. */
> enum kfence_error_type {
> KFENCE_ERROR_OOB, /* Detected a out-of-bounds access. */
> diff --git a/mm/kfence/report.c b/mm/kfence/report.c
> index f93a7b2a338b..f5a6d8ba3e21 100644
> --- a/mm/kfence/report.c
> +++ b/mm/kfence/report.c
> @@ -273,3 +273,50 @@ void kfence_report_error(unsigned long address, bool is_write, struct pt_regs *r
> /* We encountered a memory safety error, taint the kernel! */
> add_taint(TAINT_BAD_PAGE, LOCKDEP_STILL_OK);
> }
> +
> +#ifdef CONFIG_PRINTK
> +static void kfence_to_kp_stack(const struct kfence_track *track, void **kp_stack)
> +{
> + int i, j;
> +
> + i = get_stack_skipnr(track->stack_entries, track->num_stack_entries, NULL);
> + for (j = 0; i < track->num_stack_entries && j < KS_ADDRS_COUNT; ++i, ++j)
> + kp_stack[j] = (void *)track->stack_entries[i];
> + if (j < KS_ADDRS_COUNT)
> + kp_stack[j] = NULL;
> +}
> +
> +bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> +{
> + struct kfence_metadata *meta = addr_to_metadata((unsigned long)object);
> + unsigned long flags;
> +
> + if (!meta)
> + return false;
> +
> + /*
> + * If state is UNUSED at least show the pointer requested; the rest
> + * would be garbage data.
> + */
> + kpp->kp_ptr = object;
> +
> + /* Requesting info an a never-used object is almost certainly a bug. */
> + if (WARN_ON(meta->state == KFENCE_OBJECT_UNUSED))
> + return true;
> +
> + raw_spin_lock_irqsave(&meta->lock, flags);
> +
> + kpp->kp_slab = slab;
> + kpp->kp_slab_cache = meta->cache;
> + kpp->kp_objp = (void *)meta->addr;
> + kfence_to_kp_stack(&meta->alloc_track, kpp->kp_stack);
> + if (meta->state == KFENCE_OBJECT_FREED)
> + kfence_to_kp_stack(&meta->free_track, kpp->kp_free_stack);
> + /* get_stack_skipnr() ensures the first entry is outside allocator. */
> + kpp->kp_ret = kpp->kp_stack[0];
> +
> + raw_spin_unlock_irqrestore(&meta->lock, flags);
> +
> + return true;
> +}
> +#endif
> diff --git a/mm/slab.c b/mm/slab.c
> index b04e40078bdf..0edb474edef1 100644
> --- a/mm/slab.c
> +++ b/mm/slab.c
> @@ -3665,7 +3665,7 @@ EXPORT_SYMBOL(__kmalloc_node_track_caller);
> #endif /* CONFIG_NUMA */
>
> #ifdef CONFIG_PRINTK
> -void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> +void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> {
> struct kmem_cache *cachep;
> unsigned int objnr;
> diff --git a/mm/slab.h b/mm/slab.h
> index fd7ae2024897..95eb34174c1b 100644
> --- a/mm/slab.h
> +++ b/mm/slab.h
> @@ -868,7 +868,7 @@ struct kmem_obj_info {
> void *kp_stack[KS_ADDRS_COUNT];
> void *kp_free_stack[KS_ADDRS_COUNT];
> };
> -void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
> +void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
> #endif
>
> #ifdef CONFIG_HAVE_HARDENED_USERCOPY_ALLOCATOR
> diff --git a/mm/slab_common.c b/mm/slab_common.c
> index 6ee64d6208b3..2b3206a2c3b5 100644
> --- a/mm/slab_common.c
> +++ b/mm/slab_common.c
> @@ -555,6 +555,13 @@ bool kmem_valid_obj(void *object)
> }
> EXPORT_SYMBOL_GPL(kmem_valid_obj);
>
> +static void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> +{
> + if (__kfence_obj_info(kpp, object, slab))
> + return;
> + __kmem_obj_info(kpp, object, slab);
> +}
> +
> /**
> * kmem_dump_obj - Print available slab provenance information
> * @object: slab object for which to find provenance information.
> @@ -590,6 +597,8 @@ void kmem_dump_obj(void *object)
> pr_cont(" slab%s %s", cp, kp.kp_slab_cache->name);
> else
> pr_cont(" slab%s", cp);
> + if (is_kfence_address(object))
> + pr_cont(" (kfence)");
> if (kp.kp_objp)
> pr_cont(" start %px", kp.kp_objp);
> if (kp.kp_data_offset)
> diff --git a/mm/slob.c b/mm/slob.c
> index dfa6808dff36..40ea6e2d4ccd 100644
> --- a/mm/slob.c
> +++ b/mm/slob.c
> @@ -463,7 +463,7 @@ static void slob_free(void *block, int size)
> }
>
> #ifdef CONFIG_PRINTK
> -void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> +void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> {
> kpp->kp_ptr = object;
> kpp->kp_slab = slab;
> diff --git a/mm/slub.c b/mm/slub.c
> index 74d92aa4a3a2..ed5c2c03a47a 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -4312,7 +4312,7 @@ int __kmem_cache_shutdown(struct kmem_cache *s)
> }
>
> #ifdef CONFIG_PRINTK
> -void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> +void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
> {
> void *base;
> int __maybe_unused i;
> --
> 2.35.1.1094.g7c7d902a7c-goog
>

--
Thanks,
Hyeonggon