Received: by 2002:a05:6a10:9e8c:0:0:0:0 with SMTP id y12csp601710pxx; Thu, 29 Oct 2020 09:53:10 -0700 (PDT) X-Google-Smtp-Source: ABdhPJw+5NDSrocraVGDNF0X0dkgc4Ip0lfsrIqrw126OPBkZXFEkqSoBCweJmSAK/DGWKaxQoYD X-Received: by 2002:a17:906:d1c3:: with SMTP id bs3mr4758424ejb.246.1603990390471; Thu, 29 Oct 2020 09:53:10 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1603990390; cv=none; d=google.com; s=arc-20160816; b=STd2OdNRXkS2j9knTTvOt5Em3KwPQEbMBLufErbFDMhDErZNopYCyxrolosNcH9mCQ gtgKWE6BHaH66RfAmvnZVXhujGK6EaSHsS0/zdkVqVK8zMSV4hZf1ZmHE6zPiKyTBDPq ix9DfNFpXxQsEncCDFs2IQvR1jN2lIYk8LURV+r/3pQ5g40YFDQ1NIg06tykkEK33zkU ePQJTWUChp06Mg46hL3GPTWB2gcP7WuRP8YtUw4y9eBZoYiHp08gMF+QUXxzdPxlVNrp 6wnNHhKn/LOJ4zo8P9Y88rpSuPPVAZlT5T0TbAA0SgajhFfY4sgIfGfIsj1UPicjJ7DP Kv9Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature; bh=IpUt4mT/PoPxhoGI6GDvlTPnelKagQoJc6qRPSScVQw=; b=o+BrFAEhk5gNYY7FNnATZtKYAh7OYSl0Y+SFH+nlQZmCWDyRg81Rx+pDRDn+gQpTLG n30gYBmpGrmaC7MIDrTPoJ4NU1XXXGNiSLotz6v6Oae4os5JtehNRYQAntxdlmKzFBis IuNv34DJomzQ4Kuqx2hoZCEmXjSNXcYYsfex72CnZT+MEBZ2lC2oK9shnTxRVkNj3U3U x9+LcJiFdhkJxIEUEGtGsAvdHopgbnwrPfhC1FLq6zCf0+UxyVnyDt5+bEPPn0tDE09V bebjWvTM+7iBSG5zMUb2jr/BQlS75FAiU/9/RXzbIXadrmDGkl5UBx4zmLS9+NW9xQ2P TfTA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=ct29Iqn0; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id qc4si1985645ejb.200.2020.10.29.09.52.46; Thu, 29 Oct 2020 09:53:10 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=ct29Iqn0; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726709AbgJ2QvG (ORCPT + 99 others); Thu, 29 Oct 2020 12:51:06 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:59206 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727952AbgJ2Qu6 (ORCPT ); Thu, 29 Oct 2020 12:50:58 -0400 Received: from mail-lf1-x141.google.com (mail-lf1-x141.google.com [IPv6:2a00:1450:4864:20::141]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 8A0A8C0613D6; Thu, 29 Oct 2020 09:50:57 -0700 (PDT) Received: by mail-lf1-x141.google.com with SMTP id h6so4203271lfj.3; Thu, 29 Oct 2020 09:50:57 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=IpUt4mT/PoPxhoGI6GDvlTPnelKagQoJc6qRPSScVQw=; b=ct29Iqn0rHM+D3y4UNzDZVIr2ySU3BuWjzNTeAFChBXKf5anlkR3IdPMzwYAiYnCMe mg/cfnhwsMCTmDCGSnu/KDN2bhnB/sWNWR6jW9iJHBCs9BN3JagW7Ft7B2vsvHpiftTz 7Z53zagyI9udlqcNcOtK04b/PLfbVjgJZ5yOhm2daJKEPvSBwvaXqpR/u6QShRWc1x19 MDo2M61seHBths5M+cACp4YYj3Ma4eEWZUJj2rl6O0cQ7BfUpJzA58Ykz+6JgfgtXZyT KW+xsUxZ7ojjNd5LLdDDytqepJlz9V+sYzTvJDKsdkAsMvWDzlB+KBpT+t4r3CO2KbMf cgKA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=IpUt4mT/PoPxhoGI6GDvlTPnelKagQoJc6qRPSScVQw=; b=csmtNyCyEXqnxF6PNiF6LWGs/SMfADokPi+DHDIQxSjB/NOlRT3IQ/gk7+e4BB9gMb Xm79p0g8BuKCIznEbWU2GUFcY7m+fH/V811ga1Zg9kk7AafQqqTBiBb/1FtZfWZu+owJ Zc9gjNuxbaSRMyr//L9Ss/BSfl39K3uZyUA/C465MvdR847WxRnCexg6IPv52IGLB5oK SV+kft7Z/oiyCXBwBqhyXwtZ70vE0TI7/xpdf1kS4dQir6xHECWDia4VnHw7GK6wOqT7 WiePr4kGy9RQueNF5c2Ty+cPT6KDi1ZgxwNwH8mb0AXkK2R5sMv/0rN6Dd4kmbWGl5By rgtw== X-Gm-Message-State: AOAM532p4vRbfX6yvGX9qUzLtlnj8sYFrDvS0NzAq6JE9e3+XJFdgrI7 b+hQl6x4UtDBxqI/TcytfIKdLPcwQGCEyA== X-Received: by 2002:ac2:515b:: with SMTP id q27mr2011709lfd.123.1603990255361; Thu, 29 Oct 2020 09:50:55 -0700 (PDT) Received: from pc638.lan (h5ef52e31.seluork.dyn.perspektivbredband.net. [94.245.46.49]) by smtp.gmail.com with ESMTPSA id s1sm331832lfd.236.2020.10.29.09.50.53 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 29 Oct 2020 09:50:54 -0700 (PDT) From: "Uladzislau Rezki (Sony)" To: LKML , RCU , "Paul E . McKenney" Cc: Andrew Morton , Peter Zijlstra , Michal Hocko , Thomas Gleixner , "Theodore Y . Ts'o" , Joel Fernandes , Sebastian Andrzej Siewior , Uladzislau Rezki , Oleksiy Avramchenko Subject: [PATCH 15/16] rcu/tree: Allocate a page when caller is preemptible Date: Thu, 29 Oct 2020 17:50:18 +0100 Message-Id: <20201029165019.14218-15-urezki@gmail.com> X-Mailer: git-send-email 2.20.1 In-Reply-To: <20201029165019.14218-1-urezki@gmail.com> References: <20201029165019.14218-1-urezki@gmail.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Given that CONFIG_PREEMPT_COUNT is unconditionally enabled by the earlier commits in this series, the preemptible() macro now properly detects preempt-disable code regions even in kernels built with CONFIG_PREEMPT_NONE. This commit therefore uses preemptible() to determine whether allocation is possible at all for double-argument kvfree_rcu(). If !preemptible(), then allocation is not possible, and kvfree_rcu() falls back to using the less cache-friendly rcu_head approach. Even when preemptible(), the caller might be involved in reclaim, so the GFP_ flags used by double-argument kvfree_rcu() must avoid invoking reclaim processing. Note that single-argument kvfree_rcu() must be invoked in sleepable contexts, and that its fallback is the relatively high latency synchronize_rcu(). Single-argument kvfree_rcu() therefore uses GFP_KERNEL|__GFP_RETRY_MAYFAIL to allow limited sleeping within the memory allocator. [ paulmck: Add add_ptr_to_bulk_krc_lock header comment per Michal Hocko. ] Signed-off-by: Uladzislau Rezki (Sony) Signed-off-by: Paul E. McKenney --- kernel/rcu/tree.c | 48 ++++++++++++++++++++++++++++++----------------- 1 file changed, 31 insertions(+), 17 deletions(-) diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c index f2da2a1cc716..3f9b016a44dc 100644 --- a/kernel/rcu/tree.c +++ b/kernel/rcu/tree.c @@ -3406,37 +3406,55 @@ run_page_cache_worker(struct kfree_rcu_cpu *krcp) } } +// Record ptr in a page managed by krcp, with the pre-krc_this_cpu_lock() +// state specified by flags. If can_sleep is true, the caller must +// be schedulable and not be holding any locks or mutexes that might be +// acquired by the memory allocator or anything that it might invoke. +// If !can_sleep, then if !preemptible() no allocation will be undertaken, +// otherwise the allocation will use GFP_ATOMIC to avoid the remainder of +// the aforementioned deadlock possibilities. Returns true if ptr was +// successfully recorded, else the caller must use a fallback. static inline bool -kvfree_call_rcu_add_ptr_to_bulk(struct kfree_rcu_cpu *krcp, void *ptr) +add_ptr_to_bulk_krc_lock(struct kfree_rcu_cpu **krcp, + unsigned long *flags, void *ptr, bool can_sleep) { struct kvfree_rcu_bulk_data *bnode; + bool can_alloc_page = preemptible(); + gfp_t gfp = (can_sleep ? GFP_KERNEL | __GFP_RETRY_MAYFAIL : + GFP_ATOMIC) | __GFP_NOWARN; int idx; - if (unlikely(!krcp->initialized)) + *krcp = krc_this_cpu_lock(flags); + if (unlikely(!(*krcp)->initialized)) return false; - lockdep_assert_held(&krcp->lock); idx = !!is_vmalloc_addr(ptr); /* Check if a new block is required. */ - if (!krcp->bkvhead[idx] || - krcp->bkvhead[idx]->nr_records == KVFREE_BULK_MAX_ENTR) { - bnode = get_cached_bnode(krcp); - /* Switch to emergency path. */ + if (!(*krcp)->bkvhead[idx] || + (*krcp)->bkvhead[idx]->nr_records == KVFREE_BULK_MAX_ENTR) { + bnode = get_cached_bnode(*krcp); + if (!bnode && can_alloc_page) { + krc_this_cpu_unlock(*krcp, *flags); + bnode = (struct kvfree_rcu_bulk_data *) + __get_free_page(gfp); + *krcp = krc_this_cpu_lock(flags); + } + if (!bnode) return false; /* Initialize the new block. */ bnode->nr_records = 0; - bnode->next = krcp->bkvhead[idx]; + bnode->next = (*krcp)->bkvhead[idx]; /* Attach it to the head. */ - krcp->bkvhead[idx] = bnode; + (*krcp)->bkvhead[idx] = bnode; } /* Finally insert. */ - krcp->bkvhead[idx]->records - [krcp->bkvhead[idx]->nr_records++] = ptr; + (*krcp)->bkvhead[idx]->records + [(*krcp)->bkvhead[idx]->nr_records++] = ptr; return true; } @@ -3474,20 +3492,16 @@ void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func) ptr = (unsigned long *) func; } - krcp = krc_this_cpu_lock(&flags); - // Queue the object but don't yet schedule the batch. if (debug_rcu_head_queue(ptr)) { // Probable double kfree_rcu(), just leak. WARN_ONCE(1, "%s(): Double-freed call. rcu_head %p\n", __func__, head); - // Mark as success and leave. - success = true; - goto unlock_return; + return; } - success = kvfree_call_rcu_add_ptr_to_bulk(krcp, ptr); + success = add_ptr_to_bulk_krc_lock(&krcp, &flags, ptr, !head); if (!success) { run_page_cache_worker(krcp); -- 2.20.1