Received: by 2002:a05:6830:16d2:b0:61c:ac69:ca1b with SMTP id l18csp2175188otr; Mon, 25 Jul 2022 09:05:44 -0700 (PDT) X-Google-Smtp-Source: AGRyM1uZDGiI2beI+vCW0DyP//AnoFyPMPtyAZi9h+xvF0VyNuuCW1QO3/f8HRvzmcgdB8Nx0rIz X-Received: by 2002:a17:906:dc8d:b0:72f:2993:ed84 with SMTP id cs13-20020a170906dc8d00b0072f2993ed84mr10326382ejc.251.1658765144385; Mon, 25 Jul 2022 09:05:44 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1658765144; cv=none; d=google.com; s=arc-20160816; b=uCme+0kGpLXCjdbB7duWJslx8dvBok7qbtD4SQrbMhURnyQIYUa8cNCIURvzHqk50H uuCq3uuNyOI9/fW1XwXhvzjDigo4ECm7hhcX7DcOuWQgrkuv+REta4ttJnM7mfhIeCzY 2CIV5foQ90kshPxoi2ZsP+wAEFUXA5MGuSD7VRF/+ISFVWjR2ZiEnDWYfSqzy1v0o59B LRYHoOWDMItM9RbC+xnSIL5meMFFoRQtPRzAqKMCL5idm2+uF0r2+17O2zmP2j/q0HzR 3jYJmQjvhLyzyz/LEXn/1ejkcQjuJYw/5W8iMQwLKtYjA0H+bQVHpKLrHa6LjqTWnWJX L8Iw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:content-language:subject:user-agent:mime-version :date:message-id:dkim-signature:dkim-signature; bh=FUYHd99k0Lv/L/KMfN0swd4Iugm7o/OREB4uGos35AQ=; b=fSSPO0U2kROgeAc29O+CRTd3b5n8NhSy2FbbBzczipr5OuGgBsQUW4+EhBANfV3ur4 9xeUcESHtXBLVrWaC6z07ju/LK8S+GPF2FKk/5h2hDko6B8xAVpTG2vciT7npCapzbUc JsgiV4PdwnFGXWhYmNg8vHP5zljhHeYNDQGmgCAU1Z3/BeBI5vajDLyXb1f6ShvjRsnP 8vOL6cDyX70gOnRbqTGcICg9UhqImpelXLI9DrAdIfAhoezoglOU7z9iVphPbIDmYcV6 MogocrRxJkpslIs9uyg2xeAnnE+nbTYHYYrCD9kaTobefgvENFekrwot1w/v1C0C/pxm DnYA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@suse.cz header.s=susede2_rsa header.b=Ty3jGMhn; dkim=neutral (no key) header.i=@suse.cz header.s=susede2_ed25519; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id j7-20020a05640211c700b0043c2f94a3adsi725445edw.592.2022.07.25.09.05.18; Mon, 25 Jul 2022 09:05:44 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@suse.cz header.s=susede2_rsa header.b=Ty3jGMhn; dkim=neutral (no key) header.i=@suse.cz header.s=susede2_ed25519; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233504AbiGYP4l (ORCPT + 99 others); Mon, 25 Jul 2022 11:56:41 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:39098 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231321AbiGYP4j (ORCPT ); Mon, 25 Jul 2022 11:56:39 -0400 Received: from smtp-out2.suse.de (smtp-out2.suse.de [IPv6:2001:67c:2178:6::1d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 0ECC5DFF4 for ; Mon, 25 Jul 2022 08:56:37 -0700 (PDT) Received: from imap2.suse-dmz.suse.de (imap2.suse-dmz.suse.de [192.168.254.74]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature ECDSA (P-521) server-digest SHA512) (No client certificate requested) by smtp-out2.suse.de (Postfix) with ESMTPS id B4BCC1FDDF; Mon, 25 Jul 2022 15:56:35 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=suse.cz; s=susede2_rsa; t=1658764595; h=from:from:reply-to:date:date:message-id:message-id:to:to:cc:cc: mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=FUYHd99k0Lv/L/KMfN0swd4Iugm7o/OREB4uGos35AQ=; b=Ty3jGMhnqs+4oS3xvDcDdV6IZIKaDcMZopVdchI20On4aQScwlAE00JomJIHnwiCk4Pd/U KxqNNClqMUDN+uUtVJVO3ZWMkDqa1Z4ie0YbbmMtMHOZEePQ267VS681hEXX0JjHvbYNqF jfFnhNXIMkXAZRgCqw4lk5jkPIMhhi8= DKIM-Signature: v=1; a=ed25519-sha256; c=relaxed/relaxed; d=suse.cz; s=susede2_ed25519; t=1658764595; h=from:from:reply-to:date:date:message-id:message-id:to:to:cc:cc: mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=FUYHd99k0Lv/L/KMfN0swd4Iugm7o/OREB4uGos35AQ=; b=DT5mX9oP8x1EfK8CSNVS8AfDGHmZioYQK4RM16JYrCDx38Mh1I3KqBGalvL0FK3sd+A63I L9dvuljhLsD+YTDQ== Received: from imap2.suse-dmz.suse.de (imap2.suse-dmz.suse.de [192.168.254.74]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature ECDSA (P-521) server-digest SHA512) (No client certificate requested) by imap2.suse-dmz.suse.de (Postfix) with ESMTPS id 8A55E13ABB; Mon, 25 Jul 2022 15:56:35 +0000 (UTC) Received: from dovecot-director2.suse.de ([192.168.254.65]) by imap2.suse-dmz.suse.de with ESMTPSA id 72M2ITO93mJ7PwAAMHmgww (envelope-from ); Mon, 25 Jul 2022 15:56:35 +0000 Message-ID: Date: Mon, 25 Jul 2022 17:56:35 +0200 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:91.0) Gecko/20100101 Thunderbird/91.11.0 Subject: Re: [PATCH v2 1/2] mm/slub: enable debugging memory wasting of kmalloc Content-Language: en-US To: Feng Tang , Andrew Morton , Christoph Lameter , Pekka Enberg , David Rientjes , Joonsoo Kim , Roman Gushchin , Hyeonggon Yoo <42.hyeyoo@gmail.com>, linux-mm@kvack.org, linux-kernel@vger.kernel.org Cc: Dave Hansen , Robin Murphy , John Garry References: <20220725112025.22625-1-feng.tang@intel.com> From: Vlastimil Babka In-Reply-To: <20220725112025.22625-1-feng.tang@intel.com> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 7bit X-Spam-Status: No, score=-1.4 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,NICE_REPLY_A,SPF_HELO_NONE, SPF_SOFTFAIL autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, please rebase next version on https://git.kernel.org/pub/scm/linux/kernel/git/vbabka/slab.git branch for-next as patch 2/2 didn't apply cleanly. On 7/25/22 13:20, Feng Tang wrote: > kmalloc's API family is critical for mm, with one shortcoming that > its object size is fixed to be power of 2. When user requests memory > for '2^n + 1' bytes, actually 2^(n+1) bytes will be allocated, so > in worst case, there is around 50% memory space waste. > > We've met a kernel boot OOM panic (v5.10), and from the dumped slab info: > > [ 26.062145] kmalloc-2k 814056KB 814056KB > > From debug we found there are huge number of 'struct iova_magazine', > whose size is 1032 bytes (1024 + 8), so each allocation will waste > 1016 bytes. Though the issue was solved by giving the right (bigger) > size of RAM, it is still nice to optimize the size (either use a > kmalloc friendly size or create a dedicated slab for it). > > And from lkml archive, there was another crash kernel OOM case [1] > back in 2019, which seems to be related with the similar slab waste > situation, as the log is similar: > > [ 4.332648] iommu: Adding device 0000:20:02.0 to group 16 > [ 4.338946] swapper/0 invoked oom-killer: gfp_mask=0x6040c0(GFP_KERNEL|__GFP_COMP), nodemask=(null), order=0, oom_score_adj=0 > ... > [ 4.857565] kmalloc-2048 59164KB 59164KB > > The crash kernel only has 256M memory, and 59M is pretty big here. > (Note: the related code has been changed and optimised in recent > kernel [2], these logs are picked just to demo the problem) > > So add an way to track each kmalloc's memory waste info, and leverage > the existing SLUB debug framework to show its call stack info, so > that user can evaluate the waste situation, identify some hot spots > and optimize accordingly, for a better utilization of memory. > > The waste info is integrated into existing interface: > /sys/kernel/debug/slab/kmalloc-xx/alloc_traces, one example of > 'kmalloc-4k' after boot is: > > 126 ixgbe_alloc_q_vector+0xa5/0x4a0 [ixgbe] waste=233856/1856 age=1493302/1493830/1494358 pid=1284 cpus=32 nodes=1 > __slab_alloc.isra.86+0x52/0x80 > __kmalloc_node+0x143/0x350 > ixgbe_alloc_q_vector+0xa5/0x4a0 [ixgbe] > ixgbe_init_interrupt_scheme+0x1a6/0x730 [ixgbe] > ixgbe_probe+0xc8e/0x10d0 [ixgbe] > local_pci_probe+0x42/0x80 > work_for_cpu_fn+0x13/0x20 > process_one_work+0x1c5/0x390 > > which means in 'kmalloc-4k' slab, there are 126 requests of > 2240 bytes which got a 4KB space (wasting 1856 bytes each > and 233856 bytes in total). And when system starts some real "233856/1856" is not the most self-explanatory way of conveying this, but I don't immediately see a better one that's not too verbose, hm... > workload like multiple docker instances, there are more > severe waste. > > [1]. https://lkml.org/lkml/2019/8/12/266 > [2]. https://lore.kernel.org/lkml/2920df89-9975-5785-f79b-257d3052dfaf@huawei.com/ > > [Thanks Hyeonggon for pointing out several bugs about sorting/format] > [Thanks Vlastimil for suggesting way to reduce memory usage of > orig_size and keep it only for kmalloc objects] > > Signed-off-by: Feng Tang > --- > since v1: > * limit the 'orig_size' to kmalloc objects only, and save > it after track in metadata (Vlastimil Babka) > * fix a offset calculation problem in print_trailer > > since RFC: > * fix problems in kmem_cache_alloc_bulk() and records sorting, > improve the print format (Hyeonggon Yoo) > * fix a compiling issue found by 0Day bot > * update the commit log based info from iova developers > > > > include/linux/slab.h | 2 + > mm/slub.c | 96 ++++++++++++++++++++++++++++++++++++-------- > 2 files changed, 82 insertions(+), 16 deletions(-) > > diff --git a/include/linux/slab.h b/include/linux/slab.h > index 0fefdf528e0d..a713b0e5bbcd 100644 > --- a/include/linux/slab.h > +++ b/include/linux/slab.h > @@ -29,6 +29,8 @@ > #define SLAB_RED_ZONE ((slab_flags_t __force)0x00000400U) > /* DEBUG: Poison objects */ > #define SLAB_POISON ((slab_flags_t __force)0x00000800U) > +/* Indicate a kmalloc slab */ > +#define SLAB_KMALLOC ((slab_flags_t __force)0x00001000U) > /* Align objs on cache lines */ > #define SLAB_HWCACHE_ALIGN ((slab_flags_t __force)0x00002000U) > /* Use GFP_DMA memory */ > diff --git a/mm/slub.c b/mm/slub.c > index b1281b8654bd..9763a38bc4f0 100644 > --- a/mm/slub.c > +++ b/mm/slub.c > @@ -191,6 +191,12 @@ static inline bool kmem_cache_debug(struct kmem_cache *s) > return kmem_cache_debug_flags(s, SLAB_DEBUG_FLAGS); > } > > +static inline bool slub_debug_orig_size(struct kmem_cache *s) > +{ > + return (s->flags & SLAB_KMALLOC && > + kmem_cache_debug_flags(s, SLAB_STORE_USER)); > +} Besides what Kefen wrote, we seem to be calling set_orig_size/slub_debug_orig_size() always right after a check for SLAB_STORE_USER, so it would be better to piggyback to those, check only SLAB_KMALLOC and drop the slub_debug_orig_size() completely. In detail... > void *fixup_red_left(struct kmem_cache *s, void *p) > { > if (kmem_cache_debug_flags(s, SLAB_RED_ZONE)) > @@ -814,6 +820,36 @@ static void print_slab_info(const struct slab *slab) > pr_err("Slab 0x%p objects=%u used=%u fp=0x%p flags=%pGp\n", > slab, slab->objects, slab->inuse, slab->freelist, > folio_flags(folio, 0)); > + > +} > +static inline void set_orig_size(struct kmem_cache *s, > + void *object, unsigned int orig_size) > +{ > + void *p = kasan_reset_tag(object); > + > + if (!slub_debug_orig_size(s)) > + return; You could check only for SLAB_KMALLOC here and assume when set_orig_size() is called, SLAB_STORE_USER is true already. > + > + p = object + get_info_end(s); > + > + if (s->flags & SLAB_STORE_USER) Then this check can be also assumed true. > + p += sizeof(struct track) * 2; > + > + *(unsigned int *)p = orig_size; > +} > + > +static unsigned int get_orig_size(struct kmem_cache *s, void *object) > +{ > + void *p = kasan_reset_tag(object); > + > + if (!slub_debug_orig_size(s)) As this would be the only remaining user of slub_debug_orig_size() it could be opencoded > + return s->object_size; > + > + p = object + get_info_end(s); > + if (s->flags & SLAB_STORE_USER) Again, if we are here then it already should be true. > + p += sizeof(struct track) * 2; > + > + return *(unsigned int *)p; > } > > static void slab_bug(struct kmem_cache *s, char *fmt, ...) > @@ -875,6 +911,9 @@ static void print_trailer(struct kmem_cache *s, struct slab *slab, u8 *p) > if (s->flags & SLAB_STORE_USER) > off += 2 * sizeof(struct track); > > + if (slub_debug_orig_size(s)) > + off += sizeof(unsigned int); > + > off += kasan_metadata_size(s); > > if (off != size_from_object(s)) > @@ -1030,6 +1069,9 @@ static int check_pad_bytes(struct kmem_cache *s, struct slab *slab, u8 *p) > /* We also have user information there */ > off += 2 * sizeof(struct track); ^ Here is inside a SLAB_STORE_USER check. so just add check for SLAB_KMALLOC here. We never store orig_size without the other SLAB_STORE_USER info. > > + if (slub_debug_orig_size(s)) > + off += sizeof(unsigned int); > + > off += kasan_metadata_size(s); > > if (size_from_object(s) == off) > @@ -1325,7 +1367,8 @@ static inline int alloc_consistency_checks(struct kmem_cache *s, > > static noinline int alloc_debug_processing(struct kmem_cache *s, > struct slab *slab, > - void *object, unsigned long addr) > + void *object, unsigned long addr, > + unsigned int orig_size) > { > if (s->flags & SLAB_CONSISTENCY_CHECKS) { > if (!alloc_consistency_checks(s, slab, object)) > @@ -1335,6 +1378,9 @@ static noinline int alloc_debug_processing(struct kmem_cache *s, > /* Success perform special debug activities for allocs */ > if (s->flags & SLAB_STORE_USER) > set_track(s, object, TRACK_ALLOC, addr); ^ similar here > + > + set_orig_size(s, object, orig_size); > + > trace(s, slab, object, 1); > init_object(s, object, SLUB_RED_ACTIVE); > return 1; > @@ -1661,7 +1707,8 @@ static inline > void setup_slab_debug(struct kmem_cache *s, struct slab *slab, void *addr) {} > > static inline int alloc_debug_processing(struct kmem_cache *s, > - struct slab *slab, void *object, unsigned long addr) { return 0; } > + struct slab *slab, void *object, unsigned long addr, > + unsigned int orig_size) { return 0; } > > static inline int free_debug_processing( > struct kmem_cache *s, struct slab *slab, > @@ -2905,7 +2952,7 @@ static inline void *get_freelist(struct kmem_cache *s, struct slab *slab) > * already disabled (which is the case for bulk allocation). > */ > static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, > - unsigned long addr, struct kmem_cache_cpu *c) > + unsigned long addr, struct kmem_cache_cpu *c, unsigned int orig_size) > { > void *freelist; > struct slab *slab; > @@ -3048,7 +3095,7 @@ static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, > check_new_slab: > > if (kmem_cache_debug(s)) { > - if (!alloc_debug_processing(s, slab, freelist, addr)) { > + if (!alloc_debug_processing(s, slab, freelist, addr, orig_size)) { > /* Slab failed checks. Next slab needed */ > goto new_slab; > } else { > @@ -3102,7 +3149,7 @@ static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, > * pointer. > */ > static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, > - unsigned long addr, struct kmem_cache_cpu *c) > + unsigned long addr, struct kmem_cache_cpu *c, unsigned int orig_size) > { > void *p; > > @@ -3115,7 +3162,7 @@ static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, > c = slub_get_cpu_ptr(s->cpu_slab); > #endif > > - p = ___slab_alloc(s, gfpflags, node, addr, c); > + p = ___slab_alloc(s, gfpflags, node, addr, c, orig_size); > #ifdef CONFIG_PREEMPT_COUNT > slub_put_cpu_ptr(s->cpu_slab); > #endif > @@ -3206,7 +3253,7 @@ static __always_inline void *slab_alloc_node(struct kmem_cache *s, struct list_l > */ > if (IS_ENABLED(CONFIG_PREEMPT_RT) || > unlikely(!object || !slab || !node_match(slab, node))) { > - object = __slab_alloc(s, gfpflags, node, addr, c); > + object = __slab_alloc(s, gfpflags, node, addr, c, orig_size); > } else { > void *next_object = get_freepointer_safe(s, object); > > @@ -3731,7 +3778,7 @@ int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size, > * of re-populating per CPU c->freelist > */ > p[i] = ___slab_alloc(s, flags, NUMA_NO_NODE, > - _RET_IP_, c); > + _RET_IP_, c, s->object_size); > if (unlikely(!p[i])) > goto error; > > @@ -4140,6 +4187,10 @@ static int calculate_sizes(struct kmem_cache *s) > * the object. > */ > size += 2 * sizeof(struct track); ^ and here > + > + /* Save the original requested kmalloc size */ > + if (slub_debug_orig_size(s)) > + size += sizeof(unsigned int); > #endif > > kasan_cache_create(s, &size, &s->flags); > @@ -4864,7 +4915,7 @@ void __init kmem_cache_init(void) > > /* Now we can use the kmem_cache to allocate kmalloc slabs */ > setup_kmalloc_cache_index_table(); > - create_kmalloc_caches(0); > + create_kmalloc_caches(SLAB_KMALLOC); > > /* Setup random freelists for each cache */ > init_freelist_randomization(); > @@ -5092,6 +5143,7 @@ struct location { > depot_stack_handle_t handle; > unsigned long count; > unsigned long addr; > + unsigned long waste; > long long sum_time; > long min_time; > long max_time; > @@ -5138,13 +5190,15 @@ static int alloc_loc_track(struct loc_track *t, unsigned long max, gfp_t flags) > } > > static int add_location(struct loc_track *t, struct kmem_cache *s, > - const struct track *track) > + const struct track *track, > + unsigned int orig_size) > { > long start, end, pos; > struct location *l; > - unsigned long caddr, chandle; > + unsigned long caddr, chandle, cwaste; > unsigned long age = jiffies - track->when; > depot_stack_handle_t handle = 0; > + unsigned int waste = s->object_size - orig_size; > > #ifdef CONFIG_STACKDEPOT > handle = READ_ONCE(track->handle); > @@ -5162,11 +5216,13 @@ static int add_location(struct loc_track *t, struct kmem_cache *s, > if (pos == end) > break; > > - caddr = t->loc[pos].addr; > - chandle = t->loc[pos].handle; > - if ((track->addr == caddr) && (handle == chandle)) { > + l = &t->loc[pos]; > + caddr = l->addr; > + chandle = l->handle; > + cwaste = l->waste; > + if ((track->addr == caddr) && (handle == chandle) && > + (waste == cwaste)) { > > - l = &t->loc[pos]; > l->count++; > if (track->when) { > l->sum_time += age; > @@ -5191,6 +5247,9 @@ static int add_location(struct loc_track *t, struct kmem_cache *s, > end = pos; > else if (track->addr == caddr && handle < chandle) > end = pos; > + else if (track->addr == caddr && handle == chandle && > + waste < cwaste) > + end = pos; > else > start = pos; > } > @@ -5214,6 +5273,7 @@ static int add_location(struct loc_track *t, struct kmem_cache *s, > l->min_pid = track->pid; > l->max_pid = track->pid; > l->handle = handle; > + l->waste = waste; > cpumask_clear(to_cpumask(l->cpus)); > cpumask_set_cpu(track->cpu, to_cpumask(l->cpus)); > nodes_clear(l->nodes); > @@ -5232,7 +5292,7 @@ static void process_slab(struct loc_track *t, struct kmem_cache *s, > > for_each_object(p, s, addr, slab->objects) > if (!test_bit(__obj_to_index(s, addr, p), obj_map)) > - add_location(t, s, get_track(s, p, alloc)); > + add_location(t, s, get_track(s, p, alloc), get_orig_size(s, p)); > } > #endif /* CONFIG_DEBUG_FS */ > #endif /* CONFIG_SLUB_DEBUG */ > @@ -6102,6 +6162,10 @@ static int slab_debugfs_show(struct seq_file *seq, void *v) > else > seq_puts(seq, ""); > > + if (l->waste) > + seq_printf(seq, " waste=%lu/%lu", > + l->count * l->waste, l->waste); > + > if (l->sum_time != l->min_time) { > seq_printf(seq, " age=%ld/%llu/%ld", > l->min_time, div_u64(l->sum_time, l->count),