Received: by 2002:a05:7412:2a8c:b0:e2:908c:2ebd with SMTP id u12csp3817619rdh; Fri, 29 Sep 2023 03:20:49 -0700 (PDT) X-Google-Smtp-Source: AGHT+IGxLdpmDmJHacJp5ztjUKgO9/8UWinIxfbgt3vXDu3An8vkC9Z5YiZc/yYl7pmIx6tC9VtS X-Received: by 2002:a05:6a00:1952:b0:693:3cbc:3d8e with SMTP id s18-20020a056a00195200b006933cbc3d8emr4052706pfk.0.1695982848887; Fri, 29 Sep 2023 03:20:48 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1695982848; cv=none; d=google.com; s=arc-20160816; b=VpOQMnZflXVkLrCIKDP/y+ZHvHjUQ6IFXJZCy0elP9tJoNdFfDTvpv5fEwmP8bUecL vZtnuf2DOlc+kIeO3eEp8V+p7kQfGvEr1dBzLZ7ehzFUPvBK8AYHLl0SO6os+kYVfahK SUa/rMbVW5ACnUeFbkiUXJWeWlgkS+BabE7cEPH0F3ipTaXy3L5waqxwPdHbdJCMyfKo +0laTlzr0hsgy9RyF3uG+vc8etnNvKScnpXdpWPgeuNTWZy/bES7EV1hPUHLm+dHIxN+ LnGP4o5p4Hlz3qWKhr1DpM8cly7X4RBpL0ex7hJeAl5isYp9iyD9GlCZA7U9dQdrvL6E VxFA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:content-language:subject:mime-version:date :dkim-signature:message-id; bh=4NPhOnnySortdrMr8daiwxvE+Wm91JEasajS1w18cuI=; fh=u2XZE6gbz+UGQ+2IUoDW4Il1P+aJbPJXY+MrbT03Bc0=; b=O7CHV98LFag+SL9c9qhPSmdCh8IiXAE+vX2JpAHC+g5b8ws2Z3SW2gb/TISBBVWjAc SMgv+42rnDHMGn+G1lbYtsiOEbVNHduvnamAN0PDwryAA5i0Dh+FGdzxkrpYzpZROG+e ljXtULbr9c6ReuQdMJiY8Q3wID6DqK+rmb6pap/YZixYrTaM40+LnyETURqPhlhL50A5 eE9YAUSx/IUSmWSapDvz5lLKEdFu1T/ufnZrAcALvcnS8zA5YevwVktwukPN8/q1LI8p prkTxb2kt8j+hH4GtztCloUr7YUngC14gZuSsA/4P57EnTcdePZJbl1Nr5Wowh4HVz2A KJ3A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linux.dev header.s=key1 header.b="Z/qpUVqD"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:3 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linux.dev Return-Path: Received: from lipwig.vger.email (lipwig.vger.email. [2620:137:e000::3:3]) by mx.google.com with ESMTPS id eg6-20020a056a00800600b0068fefb0c039si20779166pfb.99.2023.09.29.03.20.48 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 29 Sep 2023 03:20:48 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:3 as permitted sender) client-ip=2620:137:e000::3:3; Authentication-Results: mx.google.com; dkim=pass header.i=@linux.dev header.s=key1 header.b="Z/qpUVqD"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:3 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linux.dev Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by lipwig.vger.email (Postfix) with ESMTP id 8214C804C66D; Fri, 29 Sep 2023 03:11:15 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at lipwig.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232814AbjI2KLB (ORCPT + 99 others); Fri, 29 Sep 2023 06:11:01 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:55956 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232779AbjI2KLA (ORCPT ); Fri, 29 Sep 2023 06:11:00 -0400 Received: from out-198.mta1.migadu.com (out-198.mta1.migadu.com [95.215.58.198]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id CD98A199 for ; Fri, 29 Sep 2023 03:10:57 -0700 (PDT) Message-ID: <812f0818-9658-3107-3a45-a913b7afc3c3@linux.dev> DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1695982256; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=4NPhOnnySortdrMr8daiwxvE+Wm91JEasajS1w18cuI=; b=Z/qpUVqDaEMVJM5ZwhqyFzq/iw9Ib5PpyuRnMxttsDrvzsIYbKoYqOa2V5VmZac41ZKaDU iOhSmGFCF3SENv/oA7e2MlBfIFp1OwW7z7WhzVtm21M3Xcs6nm3AX9jAlxay5UL4hyDinO rI5Ktk8G9sq4PXQcRCL3QlFPDQUlJow= Date: Fri, 29 Sep 2023 18:10:50 +0800 MIME-Version: 1.0 Subject: Re: [PATCH] memblock: don't run loop in memblock_add_range() twice Content-Language: en-US To: Mike Rapoport Cc: akpm@linux-foundation.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org References: <20230927013752.2515238-1-yajun.deng@linux.dev> <20230928061619.GS3303@kernel.org> <3ee9c8e4-870c-4ab0-906a-7d214031d1a6@linux.dev> <20230929090406.GV3303@kernel.org> X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. From: Yajun Deng In-Reply-To: <20230929090406.GV3303@kernel.org> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Migadu-Flow: FLOW_OUT X-Spam-Status: No, score=-0.9 required=5.0 tests=DKIM_SIGNED,DKIM_VALID, DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI, SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lipwig.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (lipwig.vger.email [0.0.0.0]); Fri, 29 Sep 2023 03:11:15 -0700 (PDT) On 2023/9/29 17:04, Mike Rapoport wrote: > On Thu, Sep 28, 2023 at 04:47:59PM +0800, Yajun Deng wrote: >> On 2023/9/28 14:16, Mike Rapoport wrote: >>> On Wed, Sep 27, 2023 at 09:37:52AM +0800, Yajun Deng wrote: >>>> There is round twice in memblock_add_range(). The first counts the number >>>> of regions needed to accommodate the new area. The second actually inserts >>>> them. But the first round isn't really needed, we just need to check the >>>> counts before inserting them. >>>> >>>> Check the count before calling memblock_insert_region(). If the count is >>>> equal to the maximum value, it needs to resize the array. Otherwise, >>>> insert it directly. >>>> >>>> To avoid nested calls to memblock_add_range(), we need to call >>>> memblock_reserve() out of memblock_double_array(). >>> memblock_add_range() does an extra loop once in a while, but I don't think >>> removing it will have any actual effect on the boot time. >> >> Yes, it has no obvious actual effect on the boot time,  but it does reduce >> the number of unnecessary loop. >> >> The actual effect on the boot time should not be the only criterion for >> whether a patch is accepted or not. >> >> Since the comment in the previous code, it tells the user that it would be >> executed twice, this can be misleading to users. >> >> So the new code will be simpler and clearer. It not just change the code, >> but also remove the comment > Adding return-by-pointer parameters to memblock_double_array() and pulling > memblock_reserve() out of this function is in no way simpler and clearer > that having an extra loop. If memblock_reserve() in memblock_double_array(),  there will be nested calls to memblock_add_range(). memblock_add_range(A)->memblock_double_array(A)->memblock_reserve(B)->memblock_add_range(B) ->memblock_insert_region(B)->memblock_merge_regions(B)->memblock_insert_region(A)->memblock_merge_regions(A) It's hard to see that and debug. If memblock_reserve() out of memblock_double_array(),  there wouldn't have a nested calls. memblock_add_range(A)->memblock_double_array(A)->memblock_insert_region(A)->memblock_merge_regions(A)-> memblock_reserve(B)->memblock_add_range(B)->memblock_insert_region(B)->memblock_merge_regions(B) We should make memblock_add_range is done, and do another memblock_add_range. > If the comment is wrong, just fix the comment. > >> about "executed twice",  it obviously tells the user only resize the array >> if it is equal to the maximum value >> >> and doesn't need to be executed twice. >> >>>> Signed-off-by: Yajun Deng >>>> --- >>>> mm/memblock.c | 117 ++++++++++++++++++++++++-------------------------- >>>> 1 file changed, 57 insertions(+), 60 deletions(-) >>>> >>>> diff --git a/mm/memblock.c b/mm/memblock.c >>>> index 5a88d6d24d79..3f44c84f5d0b 100644 >>>> --- a/mm/memblock.c >>>> +++ b/mm/memblock.c >>>> @@ -400,6 +400,8 @@ void __init memblock_discard(void) >>>> * @type: memblock type of the regions array being doubled >>>> * @new_area_start: starting address of memory range to avoid overlap with >>>> * @new_area_size: size of memory range to avoid overlap with >>>> + * @new_reserve_base: starting address of new array >>>> + * @new_reserve_size: size of new array >>>> * >>>> * Double the size of the @type regions array. If memblock is being used to >>>> * allocate memory for a new reserved regions array and there is a previously >>>> @@ -412,7 +414,9 @@ void __init memblock_discard(void) >>>> */ >>>> static int __init_memblock memblock_double_array(struct memblock_type *type, >>>> phys_addr_t new_area_start, >>>> - phys_addr_t new_area_size) >>>> + phys_addr_t new_area_size, >>>> + phys_addr_t *new_reserve_base, >>>> + phys_addr_t *new_reserve_size) >>>> { >>>> struct memblock_region *new_array, *old_array; >>>> phys_addr_t old_alloc_size, new_alloc_size; >>>> @@ -490,11 +494,13 @@ static int __init_memblock memblock_double_array(struct memblock_type *type, >>>> memblock_free(old_array, old_alloc_size); >>>> /* >>>> - * Reserve the new array if that comes from the memblock. Otherwise, we >>>> - * needn't do it >>>> + * Keep the address and size if that comes from the memblock. Otherwise, >>>> + * we needn't do it. >>>> */ >>>> - if (!use_slab) >>>> - BUG_ON(memblock_reserve(addr, new_alloc_size)); >>>> + if (!use_slab) { >>>> + *new_reserve_base = addr; >>>> + *new_reserve_size = new_alloc_size; >>>> + } >>>> /* Update slab flag */ >>>> *in_slab = use_slab; >>>> @@ -588,11 +594,12 @@ static int __init_memblock memblock_add_range(struct memblock_type *type, >>>> phys_addr_t base, phys_addr_t size, >>>> int nid, enum memblock_flags flags) >>>> { >>>> - bool insert = false; >>>> phys_addr_t obase = base; >>>> phys_addr_t end = base + memblock_cap_size(base, &size); >>>> - int idx, nr_new, start_rgn = -1, end_rgn; >>>> + phys_addr_t new_base = 0, new_size; >>>> + int idx, start_rgn = -1, end_rgn; >>>> struct memblock_region *rgn; >>>> + unsigned long ocnt = type->cnt; >>>> if (!size) >>>> return 0; >>>> @@ -608,25 +615,6 @@ static int __init_memblock memblock_add_range(struct memblock_type *type, >>>> return 0; >>>> } >>>> - /* >>>> - * The worst case is when new range overlaps all existing regions, >>>> - * then we'll need type->cnt + 1 empty regions in @type. So if >>>> - * type->cnt * 2 + 1 is less than or equal to type->max, we know >>>> - * that there is enough empty regions in @type, and we can insert >>>> - * regions directly. >>>> - */ >>>> - if (type->cnt * 2 + 1 <= type->max) >>>> - insert = true; >>>> - >>>> -repeat: >>>> - /* >>>> - * The following is executed twice. Once with %false @insert and >>>> - * then with %true. The first counts the number of regions needed >>>> - * to accommodate the new area. The second actually inserts them. >>>> - */ >>>> - base = obase; >>>> - nr_new = 0; >>>> - >>>> for_each_memblock_type(idx, type, rgn) { >>>> phys_addr_t rbase = rgn->base; >>>> phys_addr_t rend = rbase + rgn->size; >>>> @@ -644,15 +632,23 @@ static int __init_memblock memblock_add_range(struct memblock_type *type, >>>> WARN_ON(nid != memblock_get_region_node(rgn)); >>>> #endif >>>> WARN_ON(flags != rgn->flags); >>>> - nr_new++; >>>> - if (insert) { >>>> - if (start_rgn == -1) >>>> - start_rgn = idx; >>>> - end_rgn = idx + 1; >>>> - memblock_insert_region(type, idx++, base, >>>> - rbase - base, nid, >>>> - flags); >>>> - } >>>> + >>>> + /* >>>> + * If type->cnt is equal to type->max, it means there's >>>> + * not enough empty region and the array needs to be >>>> + * resized. Otherwise, insert it directly. >>>> + */ >>>> + if ((type->cnt == type->max) && >>>> + memblock_double_array(type, obase, size, >>>> + &new_base, &new_size)) >>>> + return -ENOMEM; >>>> + >>>> + if (start_rgn == -1) >>>> + start_rgn = idx; >>>> + end_rgn = idx + 1; >>>> + memblock_insert_region(type, idx++, base, >>>> + rbase - base, nid, >>>> + flags); >>>> } >>>> /* area below @rend is dealt with, forget about it */ >>>> base = min(rend, end); >>>> @@ -660,33 +656,28 @@ static int __init_memblock memblock_add_range(struct memblock_type *type, >>>> /* insert the remaining portion */ >>>> if (base < end) { >>>> - nr_new++; >>>> - if (insert) { >>>> - if (start_rgn == -1) >>>> - start_rgn = idx; >>>> - end_rgn = idx + 1; >>>> - memblock_insert_region(type, idx, base, end - base, >>>> - nid, flags); >>>> - } >>>> + if ((type->cnt == type->max) && >>>> + memblock_double_array(type, obase, size, >>>> + &new_base, &new_size)) >>>> + return -ENOMEM; >>>> + >>>> + if (start_rgn == -1) >>>> + start_rgn = idx; >>>> + end_rgn = idx + 1; >>>> + memblock_insert_region(type, idx, base, end - base, >>>> + nid, flags); >>>> } >>>> - if (!nr_new) >>>> + if (ocnt == type->cnt) >>>> return 0; >>>> - /* >>>> - * If this was the first round, resize array and repeat for actual >>>> - * insertions; otherwise, merge and return. >>>> - */ >>>> - if (!insert) { >>>> - while (type->cnt + nr_new > type->max) >>>> - if (memblock_double_array(type, obase, size) < 0) >>>> - return -ENOMEM; >>>> - insert = true; >>>> - goto repeat; >>>> - } else { >>>> - memblock_merge_regions(type, start_rgn, end_rgn); >>>> - return 0; >>>> - } >>>> + memblock_merge_regions(type, start_rgn, end_rgn); >>>> + >>>> + /* Reserve the new array */ >>>> + if (new_base) >>>> + memblock_reserve(new_base, new_size); >>>> + >>>> + return 0; >>>> } >>>> /** >>>> @@ -755,6 +746,7 @@ static int __init_memblock memblock_isolate_range(struct memblock_type *type, >>>> int *start_rgn, int *end_rgn) >>>> { >>>> phys_addr_t end = base + memblock_cap_size(base, &size); >>>> + phys_addr_t new_base = 0, new_size; >>>> int idx; >>>> struct memblock_region *rgn; >>>> @@ -764,10 +756,15 @@ static int __init_memblock memblock_isolate_range(struct memblock_type *type, >>>> return 0; >>>> /* we'll create at most two more regions */ >>>> - while (type->cnt + 2 > type->max) >>>> - if (memblock_double_array(type, base, size) < 0) >>>> + if (type->cnt + 2 > type->max) { >>>> + if (memblock_double_array(type, base, size, >>>> + &new_base, &new_size)) >>>> return -ENOMEM; >>>> + if (new_base) >>>> + memblock_reserve(new_base, new_size); >>>> + } >>>> + >>>> for_each_memblock_type(idx, type, rgn) { >>>> phys_addr_t rbase = rgn->base; >>>> phys_addr_t rend = rbase + rgn->size; >>>> -- >>>> 2.25.1 >>>>