Received: by 2002:a05:6902:102b:0:0:0:0 with SMTP id x11csp76160ybt; Tue, 30 Jun 2020 15:10:04 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz6+H2upDxlm1edJuDU07rId7KiXqevh23cjVMXjlzQ3q5/lns47gWP8LGrWusMAUfB2GnR X-Received: by 2002:aa7:d44f:: with SMTP id q15mr24305734edr.340.1593554555197; Tue, 30 Jun 2020 15:02:35 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1593554555; cv=none; d=google.com; s=arc-20160816; b=DZgca7LLg3WIq6rhoDLKni6NDXLRGXFUfyoMeQLbmD1Tnva85Hpj2+8a92D5bo00MK J/yCvg/7ENgFuvQdvP4BW4iZAJuALvdfuEDhMCDgisGyfWALwWOOfeok/LUXxLcOP+bX MznNDLy5BAjT46b/XnHD/+HjWPE1ogffmuTtpLAUZ7DOqUUIx7F+A4A26nrRidaZvwPH WPimwS3/fmqDYKwUchEDBeFlYreZ9+iJTHDHulZwTUY2gcaxedRdY3/5Rp9jkxkuLTU1 9FGoaV6GiNyMztuGcCCSf5YSw/fXm3+knO7PaXCtg2duHwwOQkF0Ui6rA+WXl7G3mlKC NXPg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :ironport-sdr:ironport-sdr; bh=BmsvS2tfjpcwIz9PTjfxLQNkwwc5Y2P9npK6smpefNY=; b=I7E+/OzxBxoNTaA91HBbFdd0t7EWR7jTKDjZKPXsFRiQS+7zlnFuJZh7PqB9k8IMVa cRGsHShKo5oB6N+mASJ5hpAmMBZJdEYWxRfrdeEzfmjquN064/cZ0wiPIlEGcWeDSWa7 y2Yh/WjOPXIIX8FMQ3DlNLrNmFnwMo8K5E6FX8HsSy28JqOqKvjY5FkuBQ1VGqLHvn+D tWnHzzoF6qkHY25fFvRkHY8BGMMq+Fpv0VginfGtUP5okYtDNFyRtMN7zyPnof+b4ZZ2 v25Qe1kXzVXWWQOcTnQ5tbOx9+IohGl1FcX8jV3GC7qCwrD1rfYra2CJnWlXIcq1qrP0 FA6g== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id e13si2292490ejx.153.2020.06.30.15.02.09; Tue, 30 Jun 2020 15:02:35 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730570AbgF3VZk (ORCPT + 99 others); Tue, 30 Jun 2020 17:25:40 -0400 Received: from mga02.intel.com ([134.134.136.20]:43379 "EHLO mga02.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1730387AbgF3VZ3 (ORCPT ); Tue, 30 Jun 2020 17:25:29 -0400 IronPort-SDR: GtymgMFP5i6yMeX8gG5Ra6C2B0R50exf7aD6daghq7/Xa8l542mMynwLMvshXb0k85Zn4JPAXF YqB9t15LPOWQ== X-IronPort-AV: E=McAfee;i="6000,8403,9668"; a="134676821" X-IronPort-AV: E=Sophos;i="5.75,298,1589266800"; d="scan'208";a="134676821" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga005.fm.intel.com ([10.253.24.32]) by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 30 Jun 2020 14:25:26 -0700 IronPort-SDR: qWxOIhL7nJJh0zS00TYWJugW1j+mNLWMxg4HVF+3r0kt5FQiYEUuz4ti33qSn8YqMi0x1CXO0f 6FGbE6q+/+EA== X-IronPort-AV: E=Sophos;i="5.75,298,1589266800"; d="scan'208";a="481336293" Received: from schittin-mobl.amr.corp.intel.com (HELO bwidawsk-mobl5.local) ([10.252.132.42]) by fmsmga005-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 30 Jun 2020 14:25:26 -0700 From: Ben Widawsky To: linux-mm , linux-kernel@vger.kernel.org Cc: Michal Hocko , Dave Hansen , Ben Widawsky , Andrea Arcangeli , Andrew Morton , David Rientjes , Mel Gorman , Vlastimil Babka Subject: [PATCH 10/12] mm/mempolicy: VMA allocation for many preferred Date: Tue, 30 Jun 2020 14:25:15 -0700 Message-Id: <20200630212517.308045-11-ben.widawsky@intel.com> X-Mailer: git-send-email 2.27.0 In-Reply-To: <20200630212517.308045-1-ben.widawsky@intel.com> References: <20200630212517.308045-1-ben.widawsky@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org This patch implements MPOL_PREFERRED_MANY for alloc_pages_vma(). Like alloc_pages_current(), alloc_pages_vma() needs to support policy based decisions if they've been configured via mbind(2). The temporary "hack" of treating MPOL_PREFERRED and MPOL_PREFERRED_MANY can now be removed with this, too. All the actual machinery to make this work was part of ("mm/mempolicy: Create a page allocator for policy") Cc: Andrea Arcangeli Cc: Andrew Morton Cc: David Rientjes Cc: Mel Gorman Cc: Michal Hocko Cc: Vlastimil Babka Signed-off-by: Ben Widawsky --- mm/mempolicy.c | 29 +++++++++++++++++++++-------- 1 file changed, 21 insertions(+), 8 deletions(-) diff --git a/mm/mempolicy.c b/mm/mempolicy.c index 5fb70e6599a6..51ac0d4a2eda 100644 --- a/mm/mempolicy.c +++ b/mm/mempolicy.c @@ -2281,8 +2281,6 @@ alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, { struct mempolicy *pol; struct page *page; - int preferred_nid; - nodemask_t *nmask; pol = get_vma_policy(vma, addr); @@ -2296,6 +2294,7 @@ alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, } if (unlikely(IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) && hugepage)) { + nodemask_t *nmask; int hpage_node = node; /* @@ -2309,10 +2308,26 @@ alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, * does not allow the current node in its nodemask, we allocate * the standard way. */ - if ((pol->mode == MPOL_PREFERRED || - pol->mode == MPOL_PREFERRED_MANY) && - !(pol->flags & MPOL_F_LOCAL)) + if (pol->mode == MPOL_PREFERRED || !(pol->flags & MPOL_F_LOCAL)) { hpage_node = first_node(pol->nodes); + } else if (pol->mode == MPOL_PREFERRED_MANY) { + struct zoneref *z; + + /* + * In this policy, with direct reclaim, the normal + * policy based allocation will do the right thing - try + * twice using the preferred nodes first, and all nodes + * second. + */ + if (gfp & __GFP_DIRECT_RECLAIM) { + page = alloc_pages_policy(pol, gfp, order, NUMA_NO_NODE); + goto out; + } + + z = first_zones_zonelist(node_zonelist(numa_node_id(), GFP_HIGHUSER), + gfp_zone(GFP_HIGHUSER), &pol->nodes); + hpage_node = zone_to_nid(z->zone); + } nmask = policy_nodemask(gfp, pol); if (!nmask || node_isset(hpage_node, *nmask)) { @@ -2338,9 +2353,7 @@ alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, } } - nmask = policy_nodemask(gfp, pol); - preferred_nid = policy_node(gfp, pol, node); - page = __alloc_pages_nodemask(gfp, order, preferred_nid, nmask); + page = alloc_pages_policy(pol, gfp, order, NUMA_NO_NODE); mpol_cond_put(pol); out: return page; -- 2.27.0