Received: by 2002:a05:6358:1087:b0:cb:c9d3:cd90 with SMTP id j7csp1579952rwi; Fri, 14 Oct 2022 22:42:09 -0700 (PDT) X-Google-Smtp-Source: AMsMyM5N4FpfZkb4jTSJDo1t6vO4ncbHKyWTn5CrCz1sRbjwovMbV8k42qMPDtBEmGAjl9ovDZ42 X-Received: by 2002:a17:907:16a6:b0:78e:f140:a9b9 with SMTP id hc38-20020a17090716a600b0078ef140a9b9mr544240ejc.502.1665812529005; Fri, 14 Oct 2022 22:42:09 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1665812528; cv=none; d=google.com; s=arc-20160816; b=B9kzOVoICDFU70D+VyUWjiz0cnjieORv3STCDVdju2nrqRUX6VCNmc3qQFCCHFUx8H fbt7L7wLa/k4BGbxtZGCmWB0HNSGaqdaN0OlWiuN05d9ohNGc3EHObooTW1JfVEQn7T2 I59nmOMec0aW/oOMxGTo8zFMi6lPhqq6FNLXU2oZtkTUOxKGMbHAKnBC80yRfBCf4stY C7PIoBFbc7tAkBaxQEzBby0Zrlmns099x9UURrt0MhVjf/2wOITVXEIJOIkGPK9PvfR6 10m85pvga8iTvIeo9hK09nwk6/TFwDA4OAU0bBZJhYXidxz98JKLsFhVx1dqJYhrXIld 6DkQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :message-id:subject:cc:to:from:date:dkim-signature; bh=TPDhjKmk3ApEu3jvJvKzYDjHE8g/IRJ8TwRkyx2nzjs=; b=Eei539nrlTty4aoElKjE78R42vuK9TwHUPvXIEmjoo7p9WzpNtjFZC8XRafDjgg11n v1H02mN0rWc2d1oiAGI8u0cdfa272x7a7db2kHabgvdf3dokA13E4OMb/miJA7jcCZPg mNUYRiT0vdg6JhUIA+7TAqW306yuYESxyg9Wh5Gh+Pdp8fIHjef4pb33A7aeDehnnP/K TDB6csk4EsemzndIrBg6XaVa4b11SNLgYM5SPpzEXeyO2QXw/3ZunuC8qv3pVBIiqpVc fQfDNQhn7eOzhwKkUfqix0bD7Xvtxcym2OL6TcAtHSme8LdOqRbKAfmkmogkefZ3zmDL Zqow== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=HkV+O9se; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id sa24-20020a1709076d1800b0078deaf6b447si4157556ejc.624.2022.10.14.22.41.44; Fri, 14 Oct 2022 22:42:08 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=HkV+O9se; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229594AbiJOEen (ORCPT + 99 others); Sat, 15 Oct 2022 00:34:43 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:52146 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229581AbiJOEel (ORCPT ); Sat, 15 Oct 2022 00:34:41 -0400 Received: from mail-pj1-x102d.google.com (mail-pj1-x102d.google.com [IPv6:2607:f8b0:4864:20::102d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E0A899E for ; Fri, 14 Oct 2022 21:34:35 -0700 (PDT) Received: by mail-pj1-x102d.google.com with SMTP id gf8so6606995pjb.5 for ; Fri, 14 Oct 2022 21:34:35 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=in-reply-to:content-disposition:mime-version:message-id:subject:cc :to:from:date:from:to:cc:subject:date:message-id:reply-to; bh=TPDhjKmk3ApEu3jvJvKzYDjHE8g/IRJ8TwRkyx2nzjs=; b=HkV+O9seq0CJgkveVWp+jhhU3Frhx8ed3q7VPN7T7t+LSwRYfLEdktC8MT7SydK3E+ s05PH1scaQT4RHsuTElXFlfY/B3DD6fJ4txF39LzDBF20GcUIY4FF3s9wSX95sc90hWx nOE9yKjgMl8VVBPBg5l2RXAKIXOSg3zuvyd9LtlSZNZkaDR8xf+xNysEmIcvBg07vsI0 pdDWzio85rtm0aBLEN66vpHVySQDRkZUoGnxSB93bJxGtvjUM/R8I2zVE0FMlVpS1c99 vsiVUkGszV4iF+Pr6vFcMP3raFRj8MSxdwqPske/Hh+HUITESZYWd/XJCx3EmmncPRO5 O6mA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=in-reply-to:content-disposition:mime-version:message-id:subject:cc :to:from:date:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=TPDhjKmk3ApEu3jvJvKzYDjHE8g/IRJ8TwRkyx2nzjs=; b=eo23AhIi4wVwC60yDaA2xxC/68vThgHBctzcpfwQjMErjJJtqTayRiZQDYhp6LLTD+ TfZiXnZQoL6pt//9NGggcvTi/H/f/lP5o5KQ4ORY6/BAPYGA7r/O6ab/dUkdi3bQged+ t8rKFImJ+0EPNGIGiaiDRxzflw4CSgsfvdKJxf3tYfwSUraWHU1dqOp8UTfddfiGR/b8 TyuIU/XMDIgKV9//OoACwTGvLsad8vi0GDCPWAOS/gMpV9mx1Xpx/JdX2dY3TU2b2bmo q3rMWAAYYRh5sdW2SKMV5EtP66/7KF66QbwZ/gkwDm6vkqzPRUeyyILmAK9yHC9QPsT/ zSGA== X-Gm-Message-State: ACrzQf23XquB+ETkPyp29Wwd0KV1X/vKmOPTX72+H7qlSR9KhtOKsWL+ UlGVVfIacH6u1U4icnXI63w= X-Received: by 2002:a17:903:234f:b0:17f:6711:1f9f with SMTP id c15-20020a170903234f00b0017f67111f9fmr1204421plh.32.1665808475299; Fri, 14 Oct 2022 21:34:35 -0700 (PDT) Received: from hyeyoo ([114.29.91.56]) by smtp.gmail.com with ESMTPSA id e1-20020a170902784100b00172b87d9770sm2487852pln.81.2022.10.14.21.34.32 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 14 Oct 2022 21:34:34 -0700 (PDT) Date: Sat, 15 Oct 2022 13:34:29 +0900 From: Hyeonggon Yoo <42.hyeyoo@gmail.com> To: Guenter Roeck Cc: Christoph Lameter , Pekka Enberg , David Rientjes , Joonsoo Kim , Andrew Morton , Vlastimil Babka , Roman Gushchin , linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH] mm/slab: use kmalloc_node() for off slab freelist_idx_t array allocation Message-ID: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20221014205818.GA1428667@roeck-us.net> X-Spam-Status: No, score=-0.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_FROM,HK_RANDOM_ENVFROM, HK_RANDOM_FROM,RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org After commit d6a71648dbc0 ("mm/slab: kmalloc: pass requests larger than order-1 page to page allocator"), SLAB passes large ( > PAGE_SIZE * 2) requests to buddy like SLUB does. SLAB has been using kmalloc caches to allocate freelist_idx_t array for off slab caches. But after the commit, freelist_size can be bigger than KMALLOC_MAX_CACHE_SIZE. Instead of using pointer to kmalloc cache, use kmalloc_node() and only check if the kmalloc cache is off slab during calculate_slab_order(). If freelist_size > KMALLOC_MAX_CACHE_SIZE, no looping condition happens as it allocates freelist_idx_t array directly from buddy. Reported-by: Guenter Roeck Fixes: d6a71648dbc0 ("mm/slab: kmalloc: pass requests larger than order-1 page to page allocator") Signed-off-by: Hyeonggon Yoo <42.hyeyoo@gmail.com> --- @Guenter: This fixes the issue on my emulation. Can you please test this on your environment? include/linux/slab_def.h | 1 - mm/slab.c | 37 +++++++++++++++++++------------------ 2 files changed, 19 insertions(+), 19 deletions(-) diff --git a/include/linux/slab_def.h b/include/linux/slab_def.h index e24c9aff6fed..f0ffad6a3365 100644 --- a/include/linux/slab_def.h +++ b/include/linux/slab_def.h @@ -33,7 +33,6 @@ struct kmem_cache { size_t colour; /* cache colouring range */ unsigned int colour_off; /* colour offset */ - struct kmem_cache *freelist_cache; unsigned int freelist_size; /* constructor func */ diff --git a/mm/slab.c b/mm/slab.c index a5486ff8362a..d1f6e2c64c2e 100644 --- a/mm/slab.c +++ b/mm/slab.c @@ -1619,7 +1619,7 @@ static void slab_destroy(struct kmem_cache *cachep, struct slab *slab) * although actual page can be freed in rcu context */ if (OFF_SLAB(cachep)) - kmem_cache_free(cachep->freelist_cache, freelist); + kfree(freelist); } /* @@ -1671,21 +1671,27 @@ static size_t calculate_slab_order(struct kmem_cache *cachep, if (flags & CFLGS_OFF_SLAB) { struct kmem_cache *freelist_cache; size_t freelist_size; + size_t freelist_cache_size; freelist_size = num * sizeof(freelist_idx_t); - freelist_cache = kmalloc_slab(freelist_size, 0u); - if (!freelist_cache) - continue; - - /* - * Needed to avoid possible looping condition - * in cache_grow_begin() - */ - if (OFF_SLAB(freelist_cache)) - continue; + if (freelist_size > KMALLOC_MAX_CACHE_SIZE) { + freelist_cache_size = PAGE_SIZE << get_order(freelist_size); + } else { + freelist_cache = kmalloc_slab(freelist_size, 0u); + if (!freelist_cache) + continue; + freelist_cache_size = freelist_cache->size; + + /* + * Needed to avoid possible looping condition + * in cache_grow_begin() + */ + if (OFF_SLAB(freelist_cache)) + continue; + } /* check if off slab has enough benefit */ - if (freelist_cache->size > cachep->size / 2) + if (freelist_cache_size > cachep->size / 2) continue; } @@ -2061,11 +2067,6 @@ int __kmem_cache_create(struct kmem_cache *cachep, slab_flags_t flags) cachep->flags &= ~(SLAB_RED_ZONE | SLAB_STORE_USER); #endif - if (OFF_SLAB(cachep)) { - cachep->freelist_cache = - kmalloc_slab(cachep->freelist_size, 0u); - } - err = setup_cpu_cache(cachep, gfp); if (err) { __kmem_cache_release(cachep); @@ -2292,7 +2293,7 @@ static void *alloc_slabmgmt(struct kmem_cache *cachep, freelist = NULL; else if (OFF_SLAB(cachep)) { /* Slab management obj is off-slab. */ - freelist = kmem_cache_alloc_node(cachep->freelist_cache, + freelist = kmalloc_node(cachep->freelist_size, local_flags, nodeid); } else { /* We will use last bytes at the slab for freelist */ -- 2.32.0