Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp13123pxk; Wed, 2 Sep 2020 12:59:50 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwep5rtmlFuGrPct+PHKVO+3bzhXdbMWEKY9vR18I2RX8Dt0G0Igeli91HB/D0ieDabx5cJ X-Received: by 2002:a17:906:36d7:: with SMTP id b23mr1670282ejc.149.1599076790423; Wed, 02 Sep 2020 12:59:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1599076790; cv=none; d=google.com; s=arc-20160816; b=cw+1yfXeNHWyscwoEU5tNrzwrQRe8G/sQUHy6vCCtcRSGWkoouU8UzXPPZNEJy5ud9 7WVo4wB3oP10BmZQe402UmJJRwzrzIio0cNXSUMXFqV7XKsJ7yJ8iY8xZjLFcmSajB49 N2YG1GQCaauWJXUsIiKgt8YSgH2c4CpHBkwlacEpsRdzlbZR3x15dfXw/dL8nnkAlOjq PD02t7ew9DMNj7hIfxdht4h2V6Jfv85hfaQZouH9B3eRGrwlA5NDvKxdj4BPfYZTGqu4 81bziwIK8Bj5W9ME5HdR4yXCZFVEQXDbW5+yAaltFkCTSobB0M7jK8pBCPGvXDtABv+2 kfhg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:dkim-signature:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=mW6r3UlWM6C03eov4+W2txxZIW7sPG+T+/8G7NJyv2U=; b=ONZY6VXiMJWVKkuZYK9D8Ekgjg+e0CYj9JbCGLcVovLjikAAC3WFWFru27RmVq1itg K7EexF1QCyKBLwXk4ceZDuZoUEzYS0MWMfd/b+avvgMry5mJr6y01kuaaxIg5NVS8bjb RrIba3zSKafO/kfMcv8plkkrqvvFeT8C1pSAKNZFMm+1j/j/mTxBYnFKVqo8dUp97L80 fzQ17PYct6yJEvOipSMKeVJuLoZZjkVZ58ICzPv92KJJJkibvEAidkb0LVvTEwhcJ8fN agUHozX003tqaRalrpaO3drI9aAwpkq3OkDe+X33h/+TpTV8d6hZmbPMd2fZu983XrS6 BqAg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b="q1gaRsr/"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id ap10si371419ejc.106.2020.09.02.12.59.26; Wed, 02 Sep 2020 12:59:50 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b="q1gaRsr/"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726594AbgIBT62 (ORCPT + 99 others); Wed, 2 Sep 2020 15:58:28 -0400 Received: from hqnvemgate26.nvidia.com ([216.228.121.65]:17974 "EHLO hqnvemgate26.nvidia.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726489AbgIBT61 (ORCPT ); Wed, 2 Sep 2020 15:58:27 -0400 Received: from hqpgpgate101.nvidia.com (Not Verified[216.228.121.13]) by hqnvemgate26.nvidia.com (using TLS: TLSv1.2, DES-CBC3-SHA) id ; Wed, 02 Sep 2020 12:58:10 -0700 Received: from hqmail.nvidia.com ([172.20.161.6]) by hqpgpgate101.nvidia.com (PGP Universal service); Wed, 02 Sep 2020 12:58:24 -0700 X-PGP-Universal: processed; by hqpgpgate101.nvidia.com on Wed, 02 Sep 2020 12:58:24 -0700 Received: from rcampbell-dev.nvidia.com (10.124.1.5) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1473.3; Wed, 2 Sep 2020 19:58:19 +0000 Subject: Re: [PATCH] mm/doc: editorial pass on page migration To: Randy Dunlap , , , CC: Anshuman Khandual , Jonathan Corbet , Andrew Morton References: <20200902190603.13484-1-rcampbell@nvidia.com> <1d17188f-e947-70c4-9bf7-2426c8429314@infradead.org> From: Ralph Campbell X-Nvconfidentiality: public Message-ID: Date: Wed, 2 Sep 2020 12:58:19 -0700 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.2.2 MIME-Version: 1.0 In-Reply-To: <1d17188f-e947-70c4-9bf7-2426c8429314@infradead.org> X-Originating-IP: [10.124.1.5] X-ClientProxiedBy: HQMAIL101.nvidia.com (172.20.187.10) To HQMAIL107.nvidia.com (172.20.187.13) Content-Type: text/plain; charset="utf-8"; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1599076691; bh=mW6r3UlWM6C03eov4+W2txxZIW7sPG+T+/8G7NJyv2U=; h=X-PGP-Universal:Subject:To:CC:References:From:X-Nvconfidentiality: Message-ID:Date:User-Agent:MIME-Version:In-Reply-To: X-Originating-IP:X-ClientProxiedBy:Content-Type:Content-Language: Content-Transfer-Encoding; b=q1gaRsr/x4vVNSizZXRDm5dYOP6VBFXkLuGTZoNnqEwo2GavB5Cux+yYoOYTXXVas jpmYUksA+uGA3jEt0DTU/2SCdgJnG2XFQLlJH9DiUxb4OmIz0elxIJnQzfa/NVr62C pCW6qEzsZ2lIw+YXyZ5AA11uf8quF3Jj7ZDI5VqFTxQEt0JAmHkjmprZXxY9jpuiRW dZceFnbnb8tOJuRK7YKsUMRZeF+bJ3spsQDwyi/T2gy/wazcocIonDQxNGcpMDi6J0 1I2o4Ym2HzZhQoCLjoPW+T0NgUa66SVv6K4Hhbw+IZQlG8YE6wyeYJ0BmCq4POK4mj ayTZ3dOl+sCnA== Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 9/2/20 12:41 PM, Randy Dunlap wrote: > Hey Ralph, > > Thanks for the update/corrections. Nice job. > > A few nits/comments below: > > On 9/2/20 12:06 PM, Ralph Campbell wrote: >> Add Sphinx reference links to HMM and CPUSETS, and numerous small >> editorial changes to make the page_migration.rst document more readable. >> >> Signed-off-by: Ralph Campbell >> --- >> .../admin-guide/cgroup-v1/cpusets.rst | 2 + >> Documentation/vm/hmm.rst | 2 +- >> Documentation/vm/page_migration.rst | 150 +++++++++--------- >> 3 files changed, 80 insertions(+), 74 deletions(-) >> > >> diff --git a/Documentation/vm/page_migration.rst b/Documentation/vm/page_migration.rst >> index 68883ac485fa..bde21cd2f21f 100644 >> --- a/Documentation/vm/page_migration.rst >> +++ b/Documentation/vm/page_migration.rst >> @@ -4,25 +4,28 @@ >> Page migration >> ============== >> >> -Page migration allows the moving of the physical location of pages between >> -nodes in a numa system while the process is running. This means that the >> +Page migration allows moving the physical location of pages between >> +nodes in a NUMA system while the process is running. This means that the >> virtual addresses that the process sees do not change. However, the >> system rearranges the physical location of those pages. >> >> -The main intend of page migration is to reduce the latency of memory access >> +Also see :ref:`Heterogeneous Memory Management (HMM) ` >> +for migrating pages to or from device private memory. >> + >> +The main intent of page migration is to reduce the latency of memory accesses >> by moving pages near to the processor where the process accessing that memory >> is running. >> >> Page migration allows a process to manually relocate the node on which its >> pages are located through the MF_MOVE and MF_MOVE_ALL options while setting >> -a new memory policy via mbind(). The pages of process can also be relocated >> +a new memory policy via mbind(). The pages of a process can also be relocated >> from another process using the sys_migrate_pages() function call. The >> -migrate_pages function call takes two sets of nodes and moves pages of a >> +migrate_pages() function call takes two sets of nodes and moves pages of a >> process that are located on the from nodes to the destination nodes. >> Page migration functions are provided by the numactl package by Andi Kleen >> (a version later than 0.9.3 is required. Get it from >> ftp://oss.sgi.com/www/projects/libnuma/download/). numactl provides libnuma > > URL not valid/working AFAICT. I'll update it to https://github.com/numactl/numactl.git >> -which provides an interface similar to other numa functionality for page >> +which provides an interface similar to other NUMA functionality for page >> migration. cat ``/proc//numa_maps`` allows an easy review of where the >> pages of a process are located. See also the numa_maps documentation in the >> proc(5) man page. >> @@ -30,19 +33,19 @@ proc(5) man page. >> Manual migration is useful if for example the scheduler has relocated >> a process to a processor on a distant node. A batch scheduler or an >> administrator may detect the situation and move the pages of the process >> -nearer to the new processor. The kernel itself does only provide >> +nearer to the new processor. The kernel itself only provides >> manual page migration support. Automatic page migration may be implemented >> through user space processes that move pages. A special function call >> "move_pages" allows the moving of individual pages within a process. >> -A NUMA profiler may f.e. obtain a log showing frequent off node >> +For example, A NUMA profiler may obtain a log showing frequent off node > > nit only: off-node OK >> accesses and may use the result to move pages to more advantageous >> locations. >> >> Larger installations usually partition the system using cpusets into >> sections of nodes. Paul Jackson has equipped cpusets with the ability to >> move pages when a task is moved to another cpuset (See >> -Documentation/admin-guide/cgroup-v1/cpusets.rst). >> -Cpusets allows the automation of process locality. If a task is moved to >> +:ref:`CPUSETS `). >> +Cpusets allow the automation of process locality. If a task is moved to >> a new cpuset then also all its pages are moved with it so that the >> performance of the process does not sink dramatically. Also the pages >> of processes in a cpuset are moved if the allowed memory nodes of a >> @@ -67,9 +70,9 @@ In kernel use of migrate_pages() >> Lists of pages to be migrated are generated by scanning over >> pages and moving them into lists. This is done by >> calling isolate_lru_page(). >> - Calling isolate_lru_page increases the references to the page >> + Calling isolate_lru_page() increases the references to the page >> so that it cannot vanish while the page migration occurs. >> - It also prevents the swapper or other scans to encounter >> + It also prevents the swapper or other scans from encountering >> the page. >> >> 2. We need to have a function of type new_page_t that can be >> @@ -91,23 +94,24 @@ is increased so that the page cannot be freed while page migration occurs. >> >> Steps: >> >> -1. Lock the page to be migrated >> +1. Lock the page to be migrated. >> >> 2. Ensure that writeback is complete. >> >> 3. Lock the new page that we want to move to. It is locked so that accesses to >> - this (not yet uptodate) page immediately lock while the move is in progress. >> + this (not yet uptodate) page immediately block while the move is in progress. >> >> 4. All the page table references to the page are converted to migration >> entries. This decreases the mapcount of a page. If the resulting >> mapcount is not zero then we do not migrate the page. All user space >> - processes that attempt to access the page will now wait on the page lock. >> + processes that attempt to access the page will now wait on the page lock >> + or wait for the migration page table entry to be removed. >> >> 5. The i_pages lock is taken. This will cause all processes trying >> to access the page via the mapping to block on the spinlock. >> >> -6. The refcount of the page is examined and we back out if references remain >> - otherwise we know that we are the only one referencing this page. >> +6. The refcount of the page is examined and we back out if references remain. >> + Otherwise, we know that we are the only one referencing this page. >> >> 7. The radix tree is checked and if it does not contain the pointer to this >> page then we back out because someone else modified the radix tree. >> @@ -134,22 +138,22 @@ Steps: >> >> 15. Queued up writeback on the new page is triggered. >> >> -16. If migration entries were page then replace them with real ptes. Doing >> - so will enable access for user space processes not already waiting for >> - the page lock. >> +16. If migration entries were inserted into the page table, then replace them >> + with real ptes. Doing so will enable access for user space processes not >> + already waiting for the page lock. >> >> -19. The page locks are dropped from the old and new page. >> +17. The page locks are dropped from the old and new page. >> Processes waiting on the page lock will redo their page faults >> and will reach the new page. >> >> -20. The new page is moved to the LRU and can be scanned by the swapper >> - etc again. >> +18. The new page is moved to the LRU and can be scanned by the swapper, >> + etc. again. >> >> Non-LRU page migration >> ====================== >> >> -Although original migration aimed for reducing the latency of memory access >> -for NUMA, compaction who want to create high-order page is also main customer. >> +Although migration originally aimed for reducing the latency of memory accesses >> +for NUMA, compaction also uses migration to create high-order pages. >> >> Current problem of the implementation is that it is designed to migrate only >> *LRU* pages. However, there are potential non-lru pages which can be migrated >> @@ -158,46 +162,47 @@ in drivers, for example, zsmalloc, virtio-balloon pages. >> For virtio-balloon pages, some parts of migration code path have been hooked >> up and added virtio-balloon specific functions to intercept migration logics. >> It's too specific to a driver so other drivers who want to make their pages >> -movable would have to add own specific hooks in migration path. >> +movable would have to add their own specific hooks in the migration path. >> >> -To overclome the problem, VM supports non-LRU page migration which provides >> +To overcome the problem, VM supports non-LRU page migration which provides >> generic functions for non-LRU movable pages without driver specific hooks >> -migration path. >> +in the migration path. >> >> -If a driver want to make own pages movable, it should define three functions >> +If a driver wants to make its pages movable, it should define three functions >> which are function pointers of struct address_space_operations. >> >> 1. ``bool (*isolate_page) (struct page *page, isolate_mode_t mode);`` >> >> - What VM expects on isolate_page function of driver is to return *true* >> - if driver isolates page successfully. On returing true, VM marks the page >> + What VM expects on isolate_page() function of driver is to return *true* > > maybe of {or from} from sounds good to me. >> + if driver isolates the page successfully. On returning true, VM marks the page >> as PG_isolated so concurrent isolation in several CPUs skip the page >> for isolation. If a driver cannot isolate the page, it should return *false*. >> >> Once page is successfully isolated, VM uses page.lru fields so driver >> - shouldn't expect to preserve values in that fields. >> + shouldn't expect to preserve values in those fields. >> >> 2. ``int (*migratepage) (struct address_space *mapping,`` >> | ``struct page *newpage, struct page *oldpage, enum migrate_mode);`` >> >> - After isolation, VM calls migratepage of driver with isolated page. >> - The function of migratepage is to move content of the old page to new page >> + After isolation, VM calls migratepage() of driver with the isolated page. >> + The function of migratepage() is to move the contents of the old page to the >> + new page >> and set up fields of struct page newpage. Keep in mind that you should >> indicate to the VM the oldpage is no longer movable via __ClearPageMovable() >> - under page_lock if you migrated the oldpage successfully and returns >> + under page_lock if you migrated the oldpage successfully and returned >> MIGRATEPAGE_SUCCESS. If driver cannot migrate the page at the moment, driver >> can return -EAGAIN. On -EAGAIN, VM will retry page migration in a short time >> - because VM interprets -EAGAIN as "temporal migration failure". On returning >> - any error except -EAGAIN, VM will give up the page migration without retrying >> - in this time. >> + because VM interprets -EAGAIN as "temporary migration failure". On returning >> + any error except -EAGAIN, VM will give up the page migration without >> + retrying. >> >> - Driver shouldn't touch page.lru field VM using in the functions. >> + Driver shouldn't touch the page.lru field while in the migratepage() function. >> >> 3. ``void (*putback_page)(struct page *);`` >> >> - If migration fails on isolated page, VM should return the isolated page >> - to the driver so VM calls driver's putback_page with migration failed page. >> - In this function, driver should put the isolated page back to the own data >> + If migration fails on the isolated page, VM should return the isolated page >> + to the driver so VM calls the driver's putback_page() with the isolated page. >> + In this function, the driver should put the isolated page back into its own data >> structure. >> >> 4. non-lru movable page flags >> @@ -206,52 +211,51 @@ which are function pointers of struct address_space_operations. >> >> * PG_movable >> >> - Driver should use the below function to make page movable under page_lock:: >> + Driver should use the function below to make page movable under page_lock:: >> >> void __SetPageMovable(struct page *page, struct address_space *mapping) >> >> It needs argument of address_space for registering migration >> family functions which will be called by VM. Exactly speaking, >> - PG_movable is not a real flag of struct page. Rather than, VM >> - reuses page->mapping's lower bits to represent it. >> + PG_movable is not a real flag of struct page. Rather, VM >> + reuses the page->mapping's lower bits to represent it:: >> >> -:: >> #define PAGE_MAPPING_MOVABLE 0x2 >> page->mapping = page->mapping | PAGE_MAPPING_MOVABLE; >> >> so driver shouldn't access page->mapping directly. Instead, driver should >> - use page_mapping which mask off the low two bits of page->mapping under >> - page lock so it can get right struct address_space. >> - >> - For testing of non-lru movable page, VM supports __PageMovable function. >> - However, it doesn't guarantee to identify non-lru movable page because >> - page->mapping field is unified with other variables in struct page. >> - As well, if driver releases the page after isolation by VM, page->mapping >> - doesn't have stable value although it has PAGE_MAPPING_MOVABLE >> - (Look at __ClearPageMovable). But __PageMovable is cheap to catch whether >> - page is LRU or non-lru movable once the page has been isolated. Because >> - LRU pages never can have PAGE_MAPPING_MOVABLE in page->mapping. It is also >> + use page_mapping() which masks off the low two bits of page->mapping under >> + page lock so it can get the right struct address_space. >> + >> + For testing of non-lru movable pages, VM supports __PageMovable() function. > > non-LRU I'll globally replace non-lru with non-LRU. >> + However, it doesn't guarantee to identify non-lru movable pages because > > non-LRU > >> + the page->mapping field is unified with other variables in struct page. >> + If the driver releases the page after isolation by VM, page->mapping >> + doesn't have a stable value although it has PAGE_MAPPING_MOVABLE set >> + (look at __ClearPageMovable). But __PageMovable() is cheap to call whether >> + page is LRU or non-lru movable once the page has been isolated because LRU > > non-LRU > >> + pages can never have PAGE_MAPPING_MOVABLE set in page->mapping. It is also >> good for just peeking to test non-lru movable pages before more expensive >> - checking with lock_page in pfn scanning to select victim. >> + checking with lock_page() in pfn scanning to select a victim. >> >> - For guaranteeing non-lru movable page, VM provides PageMovable function. >> - Unlike __PageMovable, PageMovable functions validates page->mapping and >> - mapping->a_ops->isolate_page under lock_page. The lock_page prevents sudden >> - destroying of page->mapping. >> + For guaranteeing non-lru movable page, VM provides PageMovable() function. > > non-LRU > >> + Unlike __PageMovable(), PageMovable() validates page->mapping and >> + mapping->a_ops->isolate_page under lock_page(). The lock_page() prevents >> + sudden destroying of page->mapping. >> >> - Driver using __SetPageMovable should clear the flag via __ClearMovablePage >> - under page_lock before the releasing the page. >> + Drivers using __SetPageMovable() should clear the flag via >> + __ClearMovablePage() under page_lock() before the releasing the page. >> >> * PG_isolated >> >> To prevent concurrent isolation among several CPUs, VM marks isolated page >> - as PG_isolated under lock_page. So if a CPU encounters PG_isolated non-lru >> - movable page, it can skip it. Driver doesn't need to manipulate the flag >> - because VM will set/clear it automatically. Keep in mind that if driver >> - sees PG_isolated page, it means the page have been isolated by VM so it >> - shouldn't touch page.lru field. >> - PG_isolated is alias with PG_reclaim flag so driver shouldn't use the flag >> - for own purpose. >> + as PG_isolated under lock_page(). So if a CPU encounters PG_isolated >> + non-lru movable page, it can skip it. Driver doesn't need to manipulate the > > non-LRU > >> + flag because VM will set/clear it automatically. Keep in mind that if the >> + driver sees a PG_isolated page, it means the page has been isolated by the >> + VM so it shouldn't touch the page.lru field. >> + The PG_isolated flag is aliased with the PG_reclaim flag so drivers >> + shouldn't use PG_isolated for its own purposes. >> >> Monitoring Migration >> ===================== >> @@ -266,8 +270,8 @@ The following events (counters) can be used to monitor page migration. >> 512. >> >> 2. PGMIGRATE_FAIL: Normal page migration failure. Same counting rules as for >> - _SUCCESS, above: this will be increased by the number of subpages, if it was >> - a THP. >> + PGMIGRATE_SUCCESS, above: this will be increased by the number of subpages, >> + if it was a THP. >> >> 3. THP_MIGRATION_SUCCESS: A THP was migrated without being split. >> Thanks for the suggestions. I'll post a v2 with the above changes. I also forgot to mention which tree these should go to. It applies cleanly to the latest linux or linux-mm so I don't think there is a merge conflict reason pushing this to a particular tree. Maybe Andrew can pick this up for linux-mm since it is VM related.