Received: by 2002:a5b:505:0:0:0:0:0 with SMTP id o5csp1358466ybp; Fri, 11 Oct 2019 13:00:35 -0700 (PDT) X-Google-Smtp-Source: APXvYqzNx4iPzuvX5Hf0r92cP4pKC5WBt9knJnoPBqEdJNMt+vXiFU+9APoTuvguv8VclY2HoDCs X-Received: by 2002:aa7:d44d:: with SMTP id q13mr15518777edr.48.1570824035431; Fri, 11 Oct 2019 13:00:35 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1570824035; cv=none; d=google.com; s=arc-20160816; b=PV69/FYZLy99dptBcgi+UhQaLsLlDQepSgXRhGX2m5B+wA6Pvq84rlLbGW/aD+r7I2 0ZH3bGwMyCa9XbqKTsNJfnyprN7kAfjwpUzBPU4uxqEb4VajwAe2D/w0GWhEt4jJEI4V f6OZXnCscpPRFVj7UUwNuYHYBv0vs0HI1viLVTdcCgJ+RTBw/5un1Ydvh62cmuHBMLUC 0aKwVd/DWxovx60imc0LYeYLXyO6E2AqwQp/C1BnRS9R6j6dKNa25Aodvd242xu1oUlM wLhivGUhqKgN/Cm6ztOvywflZaOBdyD3wSzl205eaR8HQKMWVt07DG+WSmFkdVMpHjXP Pwsg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=R/hHiMZW1/sCbcWTSv3/PUDpwhmhjmnFIf+gkWTJHT8=; b=X/l7ZQ/0NkUEdUnQrsb0nXJvBHHKlGOnk1zulK9RhB8dKFUU+JCkCEMA9ym/qydQJL SAzgzTUk8kAufbSLS3JcsFk36a7m0V74KynyUWTC0fpydbc2Q6hU8emuyhIwWUYneV51 6y6XhbJXPyiW0YG/6T/F27gV7JIuaf0WR1BtuYmIbciaiLsX6j/SaFob6L9FUH0m6K8r u1/gXjDvXTnzbZgPeevi48ctPLV8bBrwvS418kK5gOTqRpnvpkKv6RQxyli2kTVqdstS sakRgeoQ/+mJGJzGLDKBsgt6ul44DE8cJon26Uydmok6bm2f1UgUZnTCwxVXjT2S5sXm I9KA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=virtuozzo.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id z6si5653594edx.139.2019.10.11.13.00.10; Fri, 11 Oct 2019 13:00:35 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=virtuozzo.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729055AbfJKT55 (ORCPT + 99 others); Fri, 11 Oct 2019 15:57:57 -0400 Received: from relay.sw.ru ([185.231.240.75]:44562 "EHLO relay.sw.ru" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728783AbfJKT55 (ORCPT ); Fri, 11 Oct 2019 15:57:57 -0400 Received: from [172.16.25.5] by relay.sw.ru with esmtp (Exim 4.92.2) (envelope-from ) id 1iJ12k-00055b-MW; Fri, 11 Oct 2019 22:57:42 +0300 Subject: Re: [PATCH v8 1/5] kasan: support backing vmalloc space with real shadow memory To: Daniel Axtens , kasan-dev@googlegroups.com, linux-mm@kvack.org, x86@kernel.org, glider@google.com, luto@kernel.org, linux-kernel@vger.kernel.org, mark.rutland@arm.com, dvyukov@google.com, christophe.leroy@c-s.fr Cc: linuxppc-dev@lists.ozlabs.org, gor@linux.ibm.com References: <20191001065834.8880-1-dja@axtens.net> <20191001065834.8880-2-dja@axtens.net> From: Andrey Ryabinin Message-ID: <352cb4fa-2e57-7e3b-23af-898e113bbe22@virtuozzo.com> Date: Fri, 11 Oct 2019 22:57:28 +0300 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.9.0 MIME-Version: 1.0 In-Reply-To: <20191001065834.8880-2-dja@axtens.net> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 10/1/19 9:58 AM, Daniel Axtens wrote: > core_initcall(kasan_memhotplug_init); > #endif > + > +#ifdef CONFIG_KASAN_VMALLOC > +static int kasan_populate_vmalloc_pte(pte_t *ptep, unsigned long addr, > + void *unused) > +{ > + unsigned long page; > + pte_t pte; > + > + if (likely(!pte_none(*ptep))) > + return 0; > + > + page = __get_free_page(GFP_KERNEL); > + if (!page) > + return -ENOMEM; > + > + memset((void *)page, KASAN_VMALLOC_INVALID, PAGE_SIZE); > + pte = pfn_pte(PFN_DOWN(__pa(page)), PAGE_KERNEL); > + > + /* > + * Ensure poisoning is visible before the shadow is made visible > + * to other CPUs. > + */ > + smp_wmb(); I'm not quite understand what this barrier do and why it needed. And if it's really needed there should be a pairing barrier on the other side which I don't see. > + > + spin_lock(&init_mm.page_table_lock); > + if (likely(pte_none(*ptep))) { > + set_pte_at(&init_mm, addr, ptep, pte); > + page = 0; > + } > + spin_unlock(&init_mm.page_table_lock); > + if (page) > + free_page(page); > + return 0; > +} > + ... > @@ -754,6 +769,8 @@ merge_or_add_vmap_area(struct vmap_area *va, > } > > insert: > + kasan_release_vmalloc(orig_start, orig_end, va->va_start, va->va_end); > + > if (!merged) { > link_va(va, root, parent, link, head); > augment_tree_propagate_from(va); > @@ -2068,6 +2085,22 @@ static struct vm_struct *__get_vm_area_node(unsigned long size, > > setup_vmalloc_vm(area, va, flags, caller); > > + /* > + * For KASAN, if we are in vmalloc space, we need to cover the shadow > + * area with real memory. If we come here through VM_ALLOC, this is > + * done by a higher level function that has access to the true size, > + * which might not be a full page. > + * > + * We assume module space comes via VM_ALLOC path. > + */ > + if (is_vmalloc_addr(area->addr) && !(area->flags & VM_ALLOC)) { > + if (kasan_populate_vmalloc(area->size, area)) { > + unmap_vmap_area(va); > + kfree(area); > + return NULL; > + } > + } > + > return area; > } > > @@ -2245,6 +2278,9 @@ static void __vunmap(const void *addr, int deallocate_pages) > debug_check_no_locks_freed(area->addr, get_vm_area_size(area)); > debug_check_no_obj_freed(area->addr, get_vm_area_size(area)); > > + if (area->flags & VM_KASAN) > + kasan_poison_vmalloc(area->addr, area->size); > + > vm_remove_mappings(area, deallocate_pages); > > if (deallocate_pages) { > @@ -2497,6 +2533,9 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align, > if (!addr) > return NULL; > > + if (kasan_populate_vmalloc(real_size, area)) > + return NULL; > + KASAN itself uses __vmalloc_node_range() to allocate and map shadow in memory online callback. So we should either skip non-vmalloc and non-module addresses here or teach kasan's memory online/offline callbacks to not use __vmalloc_node_range() (do something similar to kasan_populate_vmalloc() perhaps?).