Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756915AbYANWWW (ORCPT ); Mon, 14 Jan 2008 17:22:22 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1755883AbYANWQ5 (ORCPT ); Mon, 14 Jan 2008 17:16:57 -0500 Received: from ns1.suse.de ([195.135.220.2]:59829 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755672AbYANWQu (ORCPT ); Mon, 14 Jan 2008 17:16:50 -0500 From: Andi Kleen References: <200801141116.534682000@suse.de> In-Reply-To: <200801141116.534682000@suse.de> To: linux-kernel@vger.kernel.org, jbeulich@novell.com, mingo@elte.hu, tglx@linutronix.de Subject: [PATCH] [17/31] CPA: Reorder TLB / cache flushes to follow Intel recommendation Message-Id: <20080114221649.D0D1014F83@wotan.suse.de> Date: Mon, 14 Jan 2008 23:16:49 +0100 (CET) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2963 Lines: 87 Intel recommends to first flush the TLBs and then the caches on caching attribute changes. c_p_a() previously did it the other way round. Reorder that. The procedure is still not fully compliant to the Intel documentation because Intel recommends a all CPU synchronization step between the TLB flushes and the cache flushes. However on all new Intel CPUs this is now meaningless anyways because they support Self-Snoop and can skip the cache flush step anyways. Signed-off-by: Andi Kleen --- arch/x86/mm/pageattr_32.c | 13 ++++++++++--- arch/x86/mm/pageattr_64.c | 14 ++++++++++---- 2 files changed, 20 insertions(+), 7 deletions(-) Index: linux/arch/x86/mm/pageattr_32.c =================================================================== --- linux.orig/arch/x86/mm/pageattr_32.c +++ linux/arch/x86/mm/pageattr_32.c @@ -97,9 +97,6 @@ static void flush_kernel_map(void *arg) struct flush_arg *a = (struct flush_arg *)arg; struct flush *f; - if ((!cpu_has_clflush || a->full_flush) && boot_cpu_data.x86_model >= 4 && - !cpu_has_ss) - wbinvd(); list_for_each_entry(f, &a->l, l) { if (!a->full_flush && !cpu_has_ss) clflush_cache_range((void *)f->addr, PAGE_SIZE); @@ -112,6 +109,16 @@ static void flush_kernel_map(void *arg) (boot_cpu_data.x86_vendor == X86_VENDOR_AMD && boot_cpu_data.x86 == 7)) __flush_tlb_all(); + + /* + * RED-PEN: Intel documentation ask for a CPU synchronization step + * here and in the loop. But it is moot on Self-Snoop CPUs anyways. + */ + + if ((!cpu_has_clflush || a->full_flush) && + !cpu_has_ss && boot_cpu_data.x86_model >= 4) + wbinvd(); + } static void set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte) Index: linux/arch/x86/mm/pageattr_64.c =================================================================== --- linux.orig/arch/x86/mm/pageattr_64.c +++ linux/arch/x86/mm/pageattr_64.c @@ -94,16 +94,22 @@ static void flush_kernel_map(void *arg) /* When clflush is available always use it because it is much cheaper than WBINVD. */ - if ((a->full_flush || !cpu_has_clflush) && !cpu_has_ss) - wbinvd(); list_for_each_entry(f, &a->l, l) { - if (!a->full_flush && !cpu_has_ss) - clflush_cache_range((void *)f->addr, PAGE_SIZE); if (!a->full_flush) __flush_tlb_one(f->addr); + if (!a->full_flush && !cpu_has_ss) + clflush_cache_range((void *)f->addr, PAGE_SIZE); } if (a->full_flush) __flush_tlb_all(); + + /* + * RED-PEN: Intel documentation ask for a CPU synchronization step + * here and in the loop. But it is moot on Self-Snoop CPUs anyways. + */ + + if ((!cpu_has_clflush || a->full_flush) && !cpu_has_ss) + wbinvd(); } /* both protected by init_mm.mmap_sem */ -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/