Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp847824pxb; Thu, 28 Jan 2021 01:30:17 -0800 (PST) X-Google-Smtp-Source: ABdhPJzMK6WQFS4tMh0LRTkKU8BLrGIjzGn7GNjJksEflxdUsNh/XkHGQVBaGEu608xn8nWsn6lE X-Received: by 2002:a17:906:9499:: with SMTP id t25mr9799275ejx.339.1611826216963; Thu, 28 Jan 2021 01:30:16 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1611826216; cv=none; d=google.com; s=arc-20160816; b=OTel2er3McVvnsGFtlchPrdmM7B9Lh6VR+fWGUkvUnTiW7qhw4HEgCpzrgR9z59gA+ 90xlwxMnVDs9MRQVNknJmztkfkXSHBFbGsRmaN1mm0tbvjaqVzF4T3SvpYWnbRMosyWl KOQyKD0iKZbD19+4kpubVM4VDsVhGA79hwtsgdj0G8Z/qHhs3DG5/bfJrGVdg1tGEA5L yyJrf31VhB/C/RdCDnLT1RKl9BFyN8gdLkK/BpfBVJW7ZPpsvcjQ/IdnJ1AUuwoAFTBk tPeEYY0ElfnYpL+swu3Xyp5nkgkx7m/Lyoe6vVDhPO94LipYUo91juReV2eVSN9oYdi9 GYkA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=0x/LyLEktMzoXWwEnfYi0lE3LU/1DwvLkIET2dICnJc=; b=KdnC5pYiUBdCz5zLvc0XqqCLbxCuFpFva4kSG/u0Vc8oIdnsZd9MjbBNwsAQUhykYA +QaVz+AT9PRYyYGmwi+17wCjcCcwa4WCgDJIwzO1o0sZ/t3H50b/0EGir16RvwJ19ZGU 5/P287Hwq9Mg/7y2tYBbpZTqz3mieajDkdRGXU5BcYAqVOEFXEcaw2URzSu0AICw1eNX eof7azUvFQqTvjLmxXSL68ZYRRYJQ22Nht257BhvsW5fmXd6Sw/NPsMRXtcmzgrWDWtG e/FAz9Cti4yDqMVoSCoVSEiL4DqfZ7KA1mchyztasSic082B7am0/PaEUYTg1fR09a1V VSjg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=rxsJaJjU; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id d11si2149768ejr.349.2021.01.28.01.29.52; Thu, 28 Jan 2021 01:30:16 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=rxsJaJjU; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231721AbhA1J0r (ORCPT + 99 others); Thu, 28 Jan 2021 04:26:47 -0500 Received: from mail.kernel.org ([198.145.29.99]:52198 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232046AbhA1JYA (ORCPT ); Thu, 28 Jan 2021 04:24:00 -0500 Received: by mail.kernel.org (Postfix) with ESMTPSA id BDB2664DBD; Thu, 28 Jan 2021 09:23:04 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1611825798; bh=wFG5EhdlVFCwH+afpQDALNPmvIOm44hDKteic5pytDI=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=rxsJaJjU3sN48PyjZmsBwjDAMzpV1mSDumbq6Vyhy1ylEqh17Y4BWGasGmEKM8wGb FFiXgqi9YKa2UW585N4J/UJUUiMSsvRqdthQ8lWnluP+x5F7LzK94JTif6Q9PlN4Qp 0vTd4DdZTyxQs/VSO/swnV8HFYZgo19Bwx/Pip182+f3Fb3AxlbTFLjCdj8TdrjWSX h/qZFSgYAfnmiXoEsb6t5X8uFqRtKIMlLzCTMEcmb/dlZipJDPSJ67u+u2XAzNKw+U FwH99yq1Pt4VruScF6VwyzEABEVF5UHEXyz20348tQsQVXLyUoh3LOrNAXJp3tqz74 76PCZTZpxL6Hg== Date: Thu, 28 Jan 2021 11:22:59 +0200 From: Mike Rapoport To: Michal Hocko Cc: David Hildenbrand , Andrew Morton , Alexander Viro , Andy Lutomirski , Arnd Bergmann , Borislav Petkov , Catalin Marinas , Christopher Lameter , Dan Williams , Dave Hansen , Elena Reshetova , "H. Peter Anvin" , Ingo Molnar , James Bottomley , "Kirill A. Shutemov" , Matthew Wilcox , Mark Rutland , Mike Rapoport , Michael Kerrisk , Palmer Dabbelt , Paul Walmsley , Peter Zijlstra , Rick Edgecombe , Roman Gushchin , Shakeel Butt , Shuah Khan , Thomas Gleixner , Tycho Andersen , Will Deacon , linux-api@vger.kernel.org, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-kselftest@vger.kernel.org, linux-nvdimm@lists.01.org, linux-riscv@lists.infradead.org, x86@kernel.org, Hagen Paul Pfeifer , Palmer Dabbelt Subject: Re: [PATCH v16 07/11] secretmem: use PMD-size pages to amortize direct map fragmentation Message-ID: <20210128092259.GB242749@kernel.org> References: <20210121122723.3446-1-rppt@kernel.org> <20210121122723.3446-8-rppt@kernel.org> <20210126114657.GL827@dhcp22.suse.cz> <303f348d-e494-e386-d1f5-14505b5da254@redhat.com> <20210126120823.GM827@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20210126120823.GM827@dhcp22.suse.cz> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Jan 26, 2021 at 01:08:23PM +0100, Michal Hocko wrote: > On Tue 26-01-21 12:56:48, David Hildenbrand wrote: > > On 26.01.21 12:46, Michal Hocko wrote: > > > On Thu 21-01-21 14:27:19, Mike Rapoport wrote: > > > > From: Mike Rapoport > > > > > > > > Removing a PAGE_SIZE page from the direct map every time such page is > > > > allocated for a secret memory mapping will cause severe fragmentation of > > > > the direct map. This fragmentation can be reduced by using PMD-size pages > > > > as a pool for small pages for secret memory mappings. > > > > > > > > Add a gen_pool per secretmem inode and lazily populate this pool with > > > > PMD-size pages. > > > > > > > > As pages allocated by secretmem become unmovable, use CMA to back large > > > > page caches so that page allocator won't be surprised by failing attempt to > > > > migrate these pages. > > > > > > > > The CMA area used by secretmem is controlled by the "secretmem=" kernel > > > > parameter. This allows explicit control over the memory available for > > > > secretmem and provides upper hard limit for secretmem consumption. > > > > > > OK, so I have finally had a look at this closer and this is really not > > > acceptable. I have already mentioned that in a response to other patch > > > but any task is able to deprive access to secret memory to other tasks > > > and cause OOM killer which wouldn't really recover ever and potentially > > > panic the system. Now you could be less drastic and only make SIGBUS on > > > fault but that would be still quite terrible. There is a very good > > > reason why hugetlb implements is non-trivial reservation system to avoid > > > exactly these problems. So, if I understand your concerns correct this implementation has two issues: 1) allocation failure at page fault that causes unrecoverable OOM and 2) a possibility for an unprivileged user to deplete secretmem pool and cause (1) to others I'm not really familiar with OOM internals, but when I simulated an allocation failure in my testing only the allocating process and it's parent were OOM-killed and then the system continued normally. You are right, it would be better if we SIGBUS instead of OOM but I don't agree SIGBUS is terrible. As we started to draw parallels with hugetlbfs even despite it's complex reservation system, hugetlb_fault() may fail to allocate pages from CMA and this still will cause SIGBUS. And hugetlb pools may be also depleted by anybody by calling mmap(MAP_HUGETLB) and there is no any limiting knob for this, while secretmem has RLIMIT_MEMLOCK. That said, simply replacing VM_FAULT_OOM with VM_FAULT_SIGBUS makes secretmem at least as controllable and robust than hugeltbfs even without complex reservation at mmap() time. > > > So unless I am really misreading the code > > > Nacked-by: Michal Hocko > > > > > > That doesn't mean I reject the whole idea. There are some details to > > > sort out as mentioned elsewhere but you cannot really depend on > > > pre-allocated pool which can fail at a fault time like that. > > > > So, to do it similar to hugetlbfs (e.g., with CMA), there would have to be a > > mechanism to actually try pre-reserving (e.g., from the CMA area), at which > > point in time the pages would get moved to the secretmem pool, and a > > mechanism for mmap() etc. to "reserve" from these secretmem pool, such that > > there are guarantees at fault time? > > yes, reserve at mmap time and use during the fault. But this all sounds > like a self inflicted problem to me. Sure you can have a pre-allocated > or more dynamic pool to reduce the direct mapping fragmentation but you > can always fall back to regular allocatios. In other ways have the pool > as an optimization rather than a hard requirement. With a careful access > control this sounds like a manageable solution to me. I'd really wish we had this discussion for earlier spins of this series, but since this didn't happen let's refresh the history a bit. One of the major pushbacks on the first RFC [1] of the concept was about the direct map fragmentation. I tried really hard to find data that shows what is the performance difference with different page sizes in the direct map and I didn't find anything. So presuming that large pages do provide advantage the first implementation of secretmem used PMD_ORDER allocations to amortise the effect of the direct map fragmentation and then handed out 4k pages at each fault. In addition there was an option to reserve a finite pool at boot time and limit secretmem allocations only to that pool. At some point David suggested to use CMA to improve overall flexibility [3], so I switched secretmem to use CMA. Now, with the data we have at hand (my benchmarks and Intel's report David mentioned) I'm even not sure this whole pooling even required. I like the idea to have a pool as an optimization rather than a hard requirement but I don't see why would it need a careful access control. As the direct map fragmentation is not necessarily degrades the performance (and even sometimes it actually improves it) and even then the degradation is small, trying a PMD_ORDER allocation for a pool and then falling back to 4K page may be just fine. I think we could have something like this (error handling is mostly omitted): static int secretmem_pool_increase(struct secretmem_ctx *ctx, gfp_t gfp) { struct page *page = alloc_pages(gfp, PMD_PAGE_ORDER); if (!page) return -ENOMEM; /* add large page to pool */ return 0; } static struct page *secretmem_alloc_page(struct secretmem_ctx *ctx, gfp_t gfp) { struct page *page; ... if (gen_pool_avail(pool) < PAGE_SIZE) { err = secretmem_pool_increase(ctx, gfp); if (!err) { addr = gen_pool_alloc(pool, PAGE_SIZE); if (addr) page = virt_to_page(addr); } } if (!page) page = alloc_page(gfp); return page; } [1] https://lore.kernel.org/lkml/1572171452-7958-1-git-send-email-rppt@kernel.org/ [2] https://lore.kernel.org/lkml/20200720092435.17469-1-rppt@kernel.org/ [3] https://lore.kernel.org/lkml/03ec586d-c00c-c57e-3118-7186acb7b823@redhat.com/#t -- Sincerely yours, Mike.