Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp8913278imu; Tue, 4 Dec 2018 16:58:53 -0800 (PST) X-Google-Smtp-Source: AFSGD/UEf6QQZonrSX/vFvbySGwpRg7Fw3st5j2bmIwy+QyCptY1j/bTqgZrZqu9OXORlOP5zs7T X-Received: by 2002:a17:902:59d6:: with SMTP id d22mr22693319plj.10.1543971533763; Tue, 04 Dec 2018 16:58:53 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1543971533; cv=none; d=google.com; s=arc-20160816; b=L0LQwkt7QRGLY/Ecc2WyY8Hd1UEN2pq6fsaJH8UNB7b5V7Jlh6GC00GpPjMTzGD49q ww6GppGR/W7WergLs16ajcA9et3q8gdNasZUchTVOXfGlBqVF6zi0TlFD/wLUO88yJli N80DiVI0awYBwtMelqfmgQ5lw6LHiVG3P2AYEFPdW9lglWSZR4BQmzs2uSBbNqongLKR NGoM1XJxFotiM0TE68+1ay990IJ4cdWaf36Dx7GbHI63skRhw8yj+ngfvPJLL0CEmAsR pxAtc/O9++HTqyjOGWi9/xVpGeMPUmmBKGKGdJ77+CZO6kc4UnU+Vi+1itb8HkcRCfkc XqsQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:dkim-signature:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=lrUuyTiGNwmNjr0ajjBtduhiimG9rX1Xq4MGTUwm7LU=; b=kdvr2XY8BUE8MC/com0jZzijQcbiTDOVIcmXNUZBZafJ0PJj76EQ0FjRYeQw/tURzE yvZiORlB7Bx3Zwv4iYvGiMkH87e0qVIdU4ey/6GVcJxcJbSRTgt4dBAcDPlk8NyRoMxY yBeuTdpHQ8JsaHZdbJaahwTrxrlBjJtsFJeZdMv7nObTWnC6p1nuAKHHHbncaJfne6qI ISBFEt/9h7I10SkY/8gYVYG1nnAPV770wmm7wPwKANymR55qxuVYrVPF0erhAmxGfPN4 lzYL1S913SAUZeG735JmMO3HU2aKc361SXIA8xkkNP20hMDCPjzfb6wFXDHXg9ttQLnO m/bg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=E1LyPrYk; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id s12si15481758pgg.188.2018.12.04.16.58.38; Tue, 04 Dec 2018 16:58:53 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=E1LyPrYk; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726677AbeLEA6F (ORCPT + 99 others); Tue, 4 Dec 2018 19:58:05 -0500 Received: from hqemgate15.nvidia.com ([216.228.121.64]:18719 "EHLO hqemgate15.nvidia.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725979AbeLEA6E (ORCPT ); Tue, 4 Dec 2018 19:58:04 -0500 Received: from hqpgpgate101.nvidia.com (Not Verified[216.228.121.13]) by hqemgate15.nvidia.com (using TLS: TLSv1.2, DES-CBC3-SHA) id ; Tue, 04 Dec 2018 16:57:59 -0800 Received: from hqmail.nvidia.com ([172.20.161.6]) by hqpgpgate101.nvidia.com (PGP Universal service); Tue, 04 Dec 2018 16:58:02 -0800 X-PGP-Universal: processed; by hqpgpgate101.nvidia.com on Tue, 04 Dec 2018 16:58:02 -0800 Received: from [10.110.48.28] (10.124.1.5) by HQMAIL101.nvidia.com (172.20.187.10) with Microsoft SMTP Server (TLS) id 15.0.1395.4; Wed, 5 Dec 2018 00:58:01 +0000 Subject: Re: [PATCH 1/2] mm: introduce put_user_page*(), placeholder versions To: Dan Williams CC: John Hubbard , Andrew Morton , Linux MM , Jan Kara , , Al Viro , , Christoph Hellwig , Christopher Lameter , "Dalessandro, Dennis" , Doug Ledford , Jason Gunthorpe , =?UTF-8?B?SsOpcsO0bWUgR2xpc3Nl?= , Matthew Wilcox , Michal Hocko , , , Linux Kernel Mailing List , linux-fsdevel References: <20181204001720.26138-1-jhubbard@nvidia.com> <20181204001720.26138-2-jhubbard@nvidia.com> X-Nvconfidentiality: public From: John Hubbard Message-ID: <3c91d335-921c-4704-d159-2975ff3a5f20@nvidia.com> Date: Tue, 4 Dec 2018 16:58:01 -0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.3.2 MIME-Version: 1.0 In-Reply-To: X-Originating-IP: [10.124.1.5] X-ClientProxiedBy: HQMAIL108.nvidia.com (172.18.146.13) To HQMAIL101.nvidia.com (172.20.187.10) Content-Type: text/plain; charset="utf-8" Content-Language: en-US-large Content-Transfer-Encoding: 7bit DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1543971479; bh=lrUuyTiGNwmNjr0ajjBtduhiimG9rX1Xq4MGTUwm7LU=; h=X-PGP-Universal:Subject:To:CC:References:X-Nvconfidentiality:From: Message-ID:Date:User-Agent:MIME-Version:In-Reply-To: X-Originating-IP:X-ClientProxiedBy:Content-Type:Content-Language: Content-Transfer-Encoding; b=E1LyPrYkQ90wd637wQ+a1KGiFou9/xMidEUEKy5RP2BXmhOGz1a43szrx0nKwCFDm 9XIs4zS+KmF4QZxnCAGdQzo6QjZ8oNWb8mNukNKExTRTF0FhImjG7nLMLivdwRkxYB 50acOMeFHCkG9yH2femHm7NYre5/JmvJQoHNKShovpLl3bSR5arNzlhdTaTi9c0In1 QiXxmVx3p3fyakXvO8TV8X54rFGq/qRoqej7BYtprJZDVlQZUB4eJeekYenoXZX1cq bBgU6tuijBPWiK0tM6UbNCjFzZRXUHcxdZqyTqJys+G/ClhG7BohOE70ehO44QZILK vvZZJR2c0ZnuQ== Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 12/4/18 3:03 PM, Dan Williams wrote: > On Tue, Dec 4, 2018 at 1:56 PM John Hubbard wrote: >> >> On 12/4/18 12:28 PM, Dan Williams wrote: >>> On Mon, Dec 3, 2018 at 4:17 PM wrote: >>>> >>>> From: John Hubbard >>>> >>>> Introduces put_user_page(), which simply calls put_page(). >>>> This provides a way to update all get_user_pages*() callers, >>>> so that they call put_user_page(), instead of put_page(). >>>> >>>> Also introduces put_user_pages(), and a few dirty/locked variations, >>>> as a replacement for release_pages(), and also as a replacement >>>> for open-coded loops that release multiple pages. >>>> These may be used for subsequent performance improvements, >>>> via batching of pages to be released. >>>> >>>> This is the first step of fixing the problem described in [1]. The steps >>>> are: >>>> >>>> 1) (This patch): provide put_user_page*() routines, intended to be used >>>> for releasing pages that were pinned via get_user_pages*(). >>>> >>>> 2) Convert all of the call sites for get_user_pages*(), to >>>> invoke put_user_page*(), instead of put_page(). This involves dozens of >>>> call sites, and will take some time. >>>> >>>> 3) After (2) is complete, use get_user_pages*() and put_user_page*() to >>>> implement tracking of these pages. This tracking will be separate from >>>> the existing struct page refcounting. >>>> >>>> 4) Use the tracking and identification of these pages, to implement >>>> special handling (especially in writeback paths) when the pages are >>>> backed by a filesystem. Again, [1] provides details as to why that is >>>> desirable. >>> >>> I thought at Plumbers we talked about using a page bit to tag pages >>> that have had their reference count elevated by get_user_pages()? That >>> way there is no need to distinguish put_page() from put_user_page() it >>> just happens internally to put_page(). At the conference Matthew was >>> offering to free up a page bit for this purpose. >>> >> >> ...but then, upon further discussion in that same session, we realized that >> that doesn't help. You need a reference count. Otherwise a random put_page >> could affect your dma-pinned pages, etc, etc. > > Ok, sorry, I mis-remembered. So, you're effectively trying to capture > the end of the page pin event separate from the final 'put' of the > page? Makes sense. > Yes, that's it exactly. >> I was not able to actually find any place where a single additional page >> bit would help our situation, which is why this still uses LRU fields for >> both the two bits required (the RFC [1] still applies), and the dma_pinned_count. > > Except the LRU fields are already in use for ZONE_DEVICE pages... how > does this proposal interact with those? Very badly: page->pgmap and page->hmm_data both get corrupted. Is there an entire use case I'm missing: calling get_user_pages() on ZONE_DEVICE pages? Said another way: is it reasonable to disallow calling get_user_pages() on ZONE_DEVICE pages? If we have to support get_user_pages() on ZONE_DEVICE pages, then the whole LRU field approach is unusable. thanks, -- John Hubbard NVIDIA