Received: by 10.223.185.116 with SMTP id b49csp26187wrg; Tue, 13 Feb 2018 15:52:41 -0800 (PST) X-Google-Smtp-Source: AH8x225EZZsMf7xI8d89Dn/vhFZwWqOWrv5Vg48yhtT3BWS/cZwsSrH3U3CJFNpe4lcqzBV5LZZc X-Received: by 2002:a17:902:6783:: with SMTP id g3-v6mr564882plk.179.1518565961777; Tue, 13 Feb 2018 15:52:41 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1518565961; cv=none; d=google.com; s=arc-20160816; b=MET/Nqkkzu7vpwi9NKFy3P+2d3AKOcPFJDmscqvFRnLyVAwEBMwBVluqaC5f7AeHqn T95MIWtY8XcurxwCAv7eJxc4WMGFHhEkFUGCVPvL7U7KbVIvj77w+Felq75E5B40EORA OaOkq7pH31u+oQjsiKogYFrp1j9iIKnlxvzHCCxvR56s7BBN8gKUB2qhJxyNVh8TYaG1 GFkE7qo3U0tNOg/6Mm91FPZItZieOWBsnliSOgXxn1QZBzYDTi3s1YyLx3kofSI0Chhl ybcETvmrLSA7x0Y3G9XCS/uksTsiBlzWH0OX50+Ce2wR7giRQ831Ob/o4MLbqHzoS+/Q ARUw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:references :in-reply-to:message-id:date:subject:cc:to:from :arc-authentication-results; bh=NgDWFjTgZ6bfvfgrcH3KqW3pU9TVFU6Zrs6ocgOh9ec=; b=Thzax8XubqaIpQoZQSwDdIZrGYc9/kBwzF00+1jVBZgwN0hz6ANAkykYoxYvf1k3RM BehqN81Jmb48t+KVAtCQKgS1d2QBC7V6m0CX2CF4bN7Nq7hD0GYzoKJInyayyWbggIFe PLzBisC5J/xP7/ghcAhHa7GJx95eaFqkVAyyw9wKZmjb1CaQ8trS62fr+05bFFfYqlH+ j9oL0BEl0kzYPUgoQ7YzYdkgkgG1sOuWA32qq+ZSmc36uVHF8vX2CnbzPafBJw7mrryS 2rPDzJXErMT+cqHHXK2EyJQhCJgCfJoD/flEbcYxaCkkBdnwwaefqtBJsP9RwdHtAiVP Kc/w== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id e1-v6si320582pli.384.2018.02.13.15.52.26; Tue, 13 Feb 2018 15:52:41 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S966314AbeBMXuy (ORCPT + 99 others); Tue, 13 Feb 2018 18:50:54 -0500 Received: from mga17.intel.com ([192.55.52.151]:53404 "EHLO mga17.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S966236AbeBMXtm (ORCPT ); Tue, 13 Feb 2018 18:49:42 -0500 X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by fmsmga107.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 13 Feb 2018 15:49:41 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.46,509,1511856000"; d="scan'208";a="29822378" Received: from rchatre-s.jf.intel.com ([10.54.70.76]) by fmsmga004.fm.intel.com with ESMTP; 13 Feb 2018 15:49:41 -0800 From: Reinette Chatre To: tglx@linutronix.de, fenghua.yu@intel.com, tony.luck@intel.com Cc: gavin.hindman@intel.com, vikas.shivappa@linux.intel.com, dave.hansen@intel.com, mingo@redhat.com, hpa@zytor.com, x86@kernel.org, linux-kernel@vger.kernel.org, Reinette Chatre Subject: [RFC PATCH V2 22/22] x86/intel_rdt: Support contiguous memory of all sizes Date: Tue, 13 Feb 2018 07:47:06 -0800 Message-Id: <1f495b99c766db9d7c942962bca138e695f4651c.1518443616.git.reinette.chatre@intel.com> X-Mailer: git-send-email 2.13.6 In-Reply-To: References: In-Reply-To: References: Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Through "mm/hugetlb: Enable large allocations through gigantic page API" we are able to allocate contiguous memory regions larger than what the SLAB allocators can support. Use the alloc_gigantic_page/free_gigantic_page API to support allocation of large contiguous memory regions in order to support pseudo-locked regions larger than 4MB. Signed-off-by: Reinette Chatre --- arch/x86/kernel/cpu/intel_rdt_pseudo_lock.c | 89 ++++++++++++++++++++++------- 1 file changed, 68 insertions(+), 21 deletions(-) diff --git a/arch/x86/kernel/cpu/intel_rdt_pseudo_lock.c b/arch/x86/kernel/cpu/intel_rdt_pseudo_lock.c index 90f040166fcd..99918943a98a 100644 --- a/arch/x86/kernel/cpu/intel_rdt_pseudo_lock.c +++ b/arch/x86/kernel/cpu/intel_rdt_pseudo_lock.c @@ -23,6 +23,7 @@ #include #include #include +#include #include #include #include @@ -136,7 +137,7 @@ struct pseudo_lock_region { bool locked; struct kref refcount; bool deleted; - void *kmem; + struct page *kmem; #ifdef CONFIG_INTEL_RDT_DEBUGFS struct dentry *debugfs_dir; #endif @@ -202,12 +203,69 @@ static int pseudo_lock_clos_set(struct pseudo_lock_region *plr, return ret; } +/** + * contig_mem_alloc - Allocate contiguous memory for pseudo-locked region + * @plr: pseudo-locked region for which memory is requested + * + * In an effort to ensure best coverage of cache with allocated memory + * (fewest conflicting physical addresses) allocate contiguous memory + * that will be pseudo-locked. The SLAB allocators are restricted wrt + * the maximum memory it can allocate. If more memory is required than + * what can be requested from the SLAB allocators a gigantic page is + * requested instead. + */ +static int contig_mem_alloc(struct pseudo_lock_region *plr) +{ + void *kmem; + + /* + * We should not be allocating from the slab cache - we need whole + * pages. + */ + if (plr->size < KMALLOC_MAX_CACHE_SIZE) { + rdt_last_cmd_puts("requested region smaller than page size\n"); + return -EINVAL; + } + + if (plr->size > KMALLOC_MAX_SIZE) { + plr->kmem = alloc_gigantic_page(cpu_to_node(plr->cpu), + get_order(plr->size), + GFP_KERNEL | __GFP_ZERO); + if (!plr->kmem) { + rdt_last_cmd_puts("unable to allocate gigantic page\n"); + return -ENOMEM; + } + } else { + kmem = kzalloc(plr->size, GFP_KERNEL); + if (!kmem) { + rdt_last_cmd_puts("unable to allocate memory\n"); + return -ENOMEM; + } + + if (!PAGE_ALIGNED(kmem)) { + rdt_last_cmd_puts("received unaligned memory\n"); + kfree(kmem); + return -ENOMEM; + } + plr->kmem = virt_to_page(kmem); + } + return 0; +} + +static void contig_mem_free(struct pseudo_lock_region *plr) +{ + if (plr->size > KMALLOC_MAX_SIZE) + free_gigantic_page(plr->kmem, get_order(plr->size)); + else + kfree(page_to_virt(plr->kmem)); +} + static void pseudo_lock_region_clear(struct pseudo_lock_region *plr) { - plr->size = 0; plr->line_size = 0; - kfree(plr->kmem); + contig_mem_free(plr); plr->kmem = NULL; + plr->size = 0; plr->r = NULL; plr->d = NULL; } @@ -444,7 +502,7 @@ static int measure_cycles_hist_fn(void *_plr) * local register variable used for memory pointer. */ __wrmsr(MSR_MISC_FEATURE_CONTROL, prefetch_disable_bits, 0x0); - mem_r = plr->kmem; + mem_r = page_to_virt(plr->kmem); for (i = 0; i < plr->size; i += 32) { start = rdtsc_ordered(); asm volatile("mov (%0,%1,1), %%eax\n\t" @@ -568,7 +626,7 @@ static int measure_cycles_perf_fn(void *_plr) pseudo_wrmsrl_notrace(MSR_ARCH_PERFMON_EVENTSEL0 + 3, l3_miss_bits); } - mem_r = plr->kmem; + mem_r = page_to_virt(plr->kmem); size = plr->size; line_size = plr->line_size; for (i = 0; i < size; i += line_size) { @@ -912,20 +970,9 @@ static int pseudo_lock_region_init(struct pseudo_lock_region *plr, return -ENOSPC; } - /* - * We do not yet support contiguous regions larger than - * KMALLOC_MAX_SIZE - */ - if (plr->size > KMALLOC_MAX_SIZE) { - rdt_last_cmd_puts("requested region exceeds maximum size\n"); - return -E2BIG; - } - - plr->kmem = kzalloc(plr->size, GFP_KERNEL); - if (!plr->kmem) { - rdt_last_cmd_puts("unable to allocate memory\n"); + ret = contig_mem_alloc(plr); + if (ret < 0) return -ENOMEM; - } plr->r = r; plr->d = d; @@ -996,7 +1043,7 @@ static int pseudo_lock_fn(void *_plr) __wrmsr(MSR_MISC_FEATURE_CONTROL, prefetch_disable_bits, 0x0); closid_p = this_cpu_read(pqr_state.cur_closid); rmid_p = this_cpu_read(pqr_state.cur_rmid); - mem_r = plr->kmem; + mem_r = page_to_virt(plr->kmem); size = plr->size; line_size = plr->line_size; __wrmsr(IA32_PQR_ASSOC, rmid_p, plr->closid); @@ -1630,7 +1677,7 @@ static int pseudo_lock_dev_mmap(struct file *file, struct vm_area_struct *vma) return -EINVAL; } - physical = __pa(plr->kmem) >> PAGE_SHIFT; + physical = page_to_phys(plr->kmem) >> PAGE_SHIFT; psize = plr->size - off; if (off > plr->size) { @@ -1652,7 +1699,7 @@ static int pseudo_lock_dev_mmap(struct file *file, struct vm_area_struct *vma) return -ENOSPC; } - memset(plr->kmem + off, 0, vsize); + memset(page_to_virt(plr->kmem) + off, 0, vsize); if (remap_pfn_range(vma, vma->vm_start, physical + vma->vm_pgoff, vsize, vma->vm_page_prot)) { -- 2.13.6