Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757979AbYACP3u (ORCPT ); Thu, 3 Jan 2008 10:29:50 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1754592AbYACPYl (ORCPT ); Thu, 3 Jan 2008 10:24:41 -0500 Received: from ns2.suse.de ([195.135.220.15]:43028 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754412AbYACPYf (ORCPT ); Thu, 3 Jan 2008 10:24:35 -0500 From: Andi Kleen References: <20080103424.989432000@suse.de> In-Reply-To: <20080103424.989432000@suse.de> To: linux-kernel@vger.kernel.org Subject: [PATCH CPA] [19/28] CPA: Limit cache flushing to pages that really change caching Message-Id: <20080103152434.30E6614E23@wotan.suse.de> Date: Thu, 3 Jan 2008 16:24:34 +0100 (CET) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 10155 Lines: 309 Previously change_page_attr always flushed caches even for pages that only change a non caching related attribute (like RO for read/write protection). This changes the flush code to only flush the cache when the caching attributes actually change. I made some effort to already handle reprogrammed PAT bits, although this is not strictly needed right now by the core kernel (but that will probably change soon) This will only make a difference on AMD CPUs or older Intel CPUs, because all newer Intel CPUs support "self-snoop" and do not require this cache flushing anyways. Another advantage of this patch is that it prevents recursive slab calls with slab debugging. Signed-off-by: Andi Kleen --- arch/x86/mm/pageattr_32.c | 43 +++++++++++++++++++++++++++++++++---------- arch/x86/mm/pageattr_64.c | 42 ++++++++++++++++++++++++++++++++++-------- include/asm-x86/pgtable_32.h | 8 ++++++++ include/asm-x86/pgtable_64.h | 8 ++++++++ 4 files changed, 83 insertions(+), 18 deletions(-) Index: linux/arch/x86/mm/pageattr_64.c =================================================================== --- linux.orig/arch/x86/mm/pageattr_64.c +++ linux/arch/x86/mm/pageattr_64.c @@ -13,7 +13,10 @@ #include #include +enum flush_mode { FLUSH_NONE, FLUSH_CACHE, FLUSH_TLB }; + struct flush { + enum flush_mode mode; struct list_head l; unsigned long addr; }; @@ -76,7 +79,7 @@ static struct page *split_large_page(uns } struct flush_arg { - int full_flush; + enum flush_mode full_flush; struct list_head l; }; @@ -91,14 +94,19 @@ static void flush_kernel_map(void *arg) { struct flush_arg *a = (struct flush_arg *)arg; struct flush *f; + int cache_flush = a->full_flush == FLUSH_CACHE; /* When clflush is available always use it because it is much cheaper than WBINVD. */ list_for_each_entry(f, &a->l, l) { if (!a->full_flush) __flush_tlb_one(f->addr); - if (!a->full_flush && !cpu_has_ss) - clflush_cache_range((void *)f->addr, PAGE_SIZE); + if (f->mode == FLUSH_CACHE && !cpu_has_ss) { + if (cpu_has_clflush) + clflush_cache_range((void *)f->addr, PAGE_SIZE); + else + cache_flush++; + } } if (a->full_flush) __flush_tlb_all(); @@ -108,12 +116,12 @@ static void flush_kernel_map(void *arg) * here and in the loop. But it is moot on Self-Snoop CPUs anyways. */ - if ((!cpu_has_clflush || a->full_flush) && !cpu_has_ss) + if (cache_flush > 0 && !cpu_has_ss) wbinvd(); } /* both protected by init_mm.mmap_sem */ -static int full_flush; +static enum flush_mode full_flush; static LIST_HEAD(deferred_pages); static LIST_HEAD(flush_pages); @@ -155,17 +163,35 @@ static void revert_page(unsigned long ad * data structure to keep track of the flush. This has the added bonus that it will * work for MMIO holes without mem_map too. */ -static void set_tlb_flush(unsigned long address) +static void set_tlb_flush(unsigned long address, int cache) { + enum flush_mode mode = cache ? FLUSH_CACHE : FLUSH_TLB; struct flush *f = kmalloc(sizeof(struct flush), GFP_KERNEL); if (!f) { - full_flush = 1; + if (full_flush < mode) + full_flush = mode; return; } + f->addr = address; + f->mode = mode; list_add_tail(&f->l, &flush_pages); } +static unsigned short pat_bit[5] = { + [4] = _PAGE_PAT, + [3] = _PAGE_PAT_LARGE, +}; + +static int cache_attr_changed(pte_t pte, pgprot_t prot, int level) +{ + unsigned a = pte_val(pte) & (_PAGE_PCD|_PAGE_PWT); + /* Correct for PAT bit being in different position on large pages */ + if (pte_val(pte) & pat_bit[level]) + a |= _PAGE_PAT; + return a != (pgprot_val(prot) & _PAGE_CACHE); +} + static int __change_page_attr(unsigned long address, unsigned long pfn, pgprot_t prot, pgprot_t ref_prot) @@ -181,7 +207,7 @@ __change_page_attr(unsigned long address BUG_ON(PageCompound(kpte_page)); BUG_ON(PageLRU(kpte_page)); - set_tlb_flush(address); + set_tlb_flush(address, cache_attr_changed(*kpte, prot, level)); if (pgprot_val(prot) != pgprot_val(ref_prot)) { if (level == 4) { Index: linux/include/asm-x86/pgtable_64.h =================================================================== --- linux.orig/include/asm-x86/pgtable_64.h +++ linux/include/asm-x86/pgtable_64.h @@ -150,10 +150,15 @@ static inline pte_t ptep_get_and_clear_f #define _PAGE_BIT_ACCESSED 5 #define _PAGE_BIT_DIRTY 6 #define _PAGE_BIT_PSE 7 /* 4 MB (or 2MB) page */ +#define _PAGE_BIT_PAT 7 /* Only on 4kb pages */ +#define _PAGE_BIT_PAT_LARGE 12 /* Only on large pages */ #define _PAGE_BIT_FILE 6 #define _PAGE_BIT_GLOBAL 8 /* Global TLB entry PPro+ */ #define _PAGE_BIT_NX 63 /* No execute: only valid after cpuid check */ +/* Needs special handling for large pages */ +#define _PAGE_CACHE (_PAGE_PCD|_PAGE_PWT|_PAGE_PAT) + #define _PAGE_PRESENT (_AC(1, UL)<<_PAGE_BIT_PRESENT) #define _PAGE_RW (_AC(1, UL)<<_PAGE_BIT_RW) #define _PAGE_USER (_AC(1, UL)<<_PAGE_BIT_USER) @@ -171,6 +176,9 @@ static inline pte_t ptep_get_and_clear_f #define _PAGE_PROTNONE 0x080 /* If not present */ #define _PAGE_NX (_AC(1, UL)<<_PAGE_BIT_NX) +#define _PAGE_PAT (_AC(1,UL) << _PAGE_BIT_PAT) +#define _PAGE_PAT_LARGE (_AC(1,UL) << _PAGE_BIT_PAT_LARGE) + #define _PAGE_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | _PAGE_ACCESSED | _PAGE_DIRTY) #define _KERNPG_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_ACCESSED | _PAGE_DIRTY) Index: linux/arch/x86/mm/pageattr_32.c =================================================================== --- linux.orig/arch/x86/mm/pageattr_32.c +++ linux/arch/x86/mm/pageattr_32.c @@ -21,12 +21,15 @@ #define ClearPageDeferred(p) clear_bit(PG_deferred, &(p)->flags) #define TestSetPageDeferred(p) test_and_set_bit(PG_deferred, &(p)->flags) +enum flush_mode { FLUSH_NONE, FLUSH_CACHE, FLUSH_TLB }; + /* Variables protected by init_mm.mmap_sem */ -static int full_flush; +static enum flush_mode full_flush; static struct list_head df_list = LIST_HEAD_INIT(df_list); static LIST_HEAD(flush_pages); struct flush { + enum flush_mode mode; struct list_head l; unsigned long addr; }; @@ -81,7 +84,7 @@ static struct page *split_large_page(uns } struct flush_arg { - int full_flush; + enum flush_mode full_flush; struct list_head l; }; @@ -96,12 +99,17 @@ static void flush_kernel_map(void *arg) { struct flush_arg *a = (struct flush_arg *)arg; struct flush *f; + int cache_flush = a->full_flush == FLUSH_CACHE; list_for_each_entry(f, &a->l, l) { - if (!a->full_flush && !cpu_has_ss) - clflush_cache_range((void *)f->addr, PAGE_SIZE); if (!a->full_flush) __flush_tlb_one(f->addr); + if (f->mode == FLUSH_CACHE && !cpu_has_ss) { + if (cpu_has_clflush) + clflush_cache_range((void *)f->addr, PAGE_SIZE); + else + cache_flush++; + } } /* Handle errata with flushing large pages in early Athlons */ @@ -115,10 +123,8 @@ static void flush_kernel_map(void *arg) * here and in the loop. But it is moot on Self-Snoop CPUs anyways. */ - if ((!cpu_has_clflush || a->full_flush) && - !cpu_has_ss && boot_cpu_data.x86_model >= 4) + if (cache_flush > 0 && !cpu_has_ss && boot_cpu_data.x86_model >= 4) wbinvd(); - } static void set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte) @@ -169,6 +175,20 @@ static inline void save_page(struct page list_add(&kpte_page->lru, &df_list); } +static unsigned short pat_bit[4] = { + [3] = _PAGE_PAT, + [2] = _PAGE_PAT_LARGE, +}; + +static int cache_attr_changed(pte_t pte, pgprot_t prot, int level) +{ + unsigned long a = pte_val(pte) & (_PAGE_PCD|_PAGE_PWT); + /* Correct for PAT bit being in different position on large pages */ + if (pte_val(pte) & pat_bit[level]) + a |= _PAGE_PAT; + return a != (pgprot_val(prot) & _PAGE_CACHE); +} + /* * Mark the address for flushing later in global_tlb_flush(). * @@ -177,14 +197,17 @@ static inline void save_page(struct page * data structure to keep track of the flush. This has the added bonus that it will * work for MMIO holes without mem_map too. */ -static void set_tlb_flush(unsigned long address) +static void set_tlb_flush(unsigned long address, int cache) { + enum flush_mode mode = cache ? FLUSH_CACHE : FLUSH_TLB; struct flush *f = kmalloc(sizeof(struct flush), GFP_KERNEL); if (!f) { - full_flush = 1; + if (full_flush < mode) + full_flush = mode; return; } f->addr = address; + f->mode = mode; list_add_tail(&f->l, &flush_pages); } @@ -206,7 +229,7 @@ __change_page_attr(struct page *page, pg BUG_ON(PageLRU(kpte_page)); BUG_ON(PageCompound(kpte_page)); - set_tlb_flush(address); + set_tlb_flush(address, cache_attr_changed(*kpte, prot, level)); if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) { if (level == 3) { Index: linux/include/asm-x86/pgtable_32.h =================================================================== --- linux.orig/include/asm-x86/pgtable_32.h +++ linux/include/asm-x86/pgtable_32.h @@ -97,6 +97,8 @@ void paging_init(void); #define _PAGE_BIT_ACCESSED 5 #define _PAGE_BIT_DIRTY 6 #define _PAGE_BIT_PSE 7 /* 4 MB (or 2MB) page, Pentium+, if present.. */ +#define _PAGE_BIT_PAT 7 /* only on 4K pages */ +#define _PAGE_BIT_PAT_LARGE 12 /* only on large pages */ #define _PAGE_BIT_GLOBAL 8 /* Global TLB entry PPro+ */ #define _PAGE_BIT_UNUSED1 9 /* available for programmer */ #define _PAGE_BIT_UNUSED2 10 @@ -126,6 +128,12 @@ void paging_init(void); #define _PAGE_NX 0 #endif +#define _PAGE_PAT (1U << _PAGE_BIT_PAT) +#define _PAGE_PAT_LARGE (1U << _PAGE_BIT_PAT_LARGE) + +/* Needs special handling for large pages */ +#define _PAGE_CACHE (_PAGE_PWT|_PAGE_PCD|_PAGE_PAT) + #define _PAGE_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | _PAGE_ACCESSED | _PAGE_DIRTY) #define _KERNPG_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_ACCESSED | _PAGE_DIRTY) #define _PAGE_CHG_MASK (PTE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/