Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754573AbdDGKvY (ORCPT ); Fri, 7 Apr 2017 06:51:24 -0400 Received: from mail-yw0-f196.google.com ([209.85.161.196]:33132 "EHLO mail-yw0-f196.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751759AbdDGKvR (ORCPT ); Fri, 7 Apr 2017 06:51:17 -0400 MIME-Version: 1.0 In-Reply-To: References: <1490811363-93944-1-git-send-email-keescook@chromium.org> <1490811363-93944-5-git-send-email-keescook@chromium.org> From: Mathias Krause Date: Fri, 7 Apr 2017 12:51:15 +0200 Message-ID: Subject: Re: [kernel-hardening] Re: [RFC v2][PATCH 04/11] x86: Implement __arch_rare_write_begin/unmap() To: Thomas Gleixner Cc: Andy Lutomirski , Kees Cook , Andy Lutomirski , "kernel-hardening@lists.openwall.com" , Mark Rutland , Hoeun Ryu , PaX Team , Emese Revfy , Russell King , X86 ML , "linux-kernel@vger.kernel.org" , "linux-arm-kernel@lists.infradead.org" , Peter Zijlstra Content-Type: text/plain; charset=UTF-8 Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2574 Lines: 71 On 7 April 2017 at 11:46, Thomas Gleixner wrote: > On Fri, 7 Apr 2017, Mathias Krause wrote: >> On 6 April 2017 at 17:59, Andy Lutomirski wrote: >> > On Wed, Apr 5, 2017 at 5:14 PM, Kees Cook wrote: >> >> static __always_inline rare_write_begin(void) >> >> { >> >> preempt_disable(); >> >> local_irq_disable(); >> >> barrier(); >> >> __arch_rare_write_begin(); >> >> barrier(); >> >> } >> > >> > Looks good, except you don't need preempt_disable(). >> > local_irq_disable() also disables preemption. You might need to use >> > local_irq_save(), though, depending on whether any callers already >> > have IRQs off. >> >> Well, doesn't look good to me. NMIs will still be able to interrupt >> this code and will run with CR0.WP = 0. >> >> Shouldn't you instead question yourself why PaX can do it "just" with >> preempt_disable() instead?! > > That's silly. Just because PaX does it, doesn't mean it's correct. To be > honest, playing games with the CR0.WP bit is outright stupid to begin with. Why that? It allows fast and CPU local modifications of r/o memory. OTOH, an approach that needs to fiddle with page table entries requires global synchronization to keep the individual TLB states in sync. Hmm.. Not that fast, I'd say. > Whether protected by preempt_disable or local_irq_disable, to make that > work it needs CR0 handling in the exception entry/exit at the lowest > level. And that's just a nightmare maintainence wise as it's prone to be > broken over time. It seems to be working fine for more than a decade now in PaX. So it can't be such a big maintenance nightmare ;) > Aside of that it's pointless overhead for the normal case. > > The proper solution is: > > write_rare(ptr, val) > { > mp = map_shadow_rw(ptr); > *mp = val; > unmap_shadow_rw(mp); > } > > map_shadow_rw() is essentially the same thing as we do in the highmem case > where the kernel creates a shadow mapping of the user space pages via > kmap_atomic(). The "proper solution" seems to be much slower compared to just toggling CR0.WP (which is costly in itself, already) because of the TLB invalidation / synchronisation involved. > It's valid (at least on x86) to have a shadow map with the same page > attributes but write enabled. That does not require any fixups of CR0 and > just works. "Just works", sure -- but it's not as tightly focused as the PaX solution which is CPU local, while your proposed solution is globally visible. Cheers, Mathias