Received: by 2002:ac0:aa62:0:0:0:0:0 with SMTP id w31-v6csp1140336ima; Wed, 24 Oct 2018 15:19:34 -0700 (PDT) X-Google-Smtp-Source: AJdET5erEX0GUh5wiK1Q46iQfPEFtPPwwDpreleH4OJjFwaS6HGypXALbAjIoRNeNe6AzuePtIZ3 X-Received: by 2002:a62:e048:: with SMTP id f69-v6mr4325160pfh.208.1540419574193; Wed, 24 Oct 2018 15:19:34 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1540419574; cv=none; d=google.com; s=arc-20160816; b=vr5LvzqLrc2AaGaJJX63yLZsAEi2vsHWYnplSnlz3XHXVqZXuhpm3r3MRqdTOWpf3O AzYke3NwuDRg8IA2TuKOV4fujxwte6soI8xoy2qnvGMBckpF3kQ/6RqXEtK5YT4N0xrW vWwkJYYjASRqagvjIlDmQYwSpNJ2rDVL/yGtlhKONNeMmEqRX2l1cVrra3IPYTk4Rsyo vFP8cbH4OAZCjL/m+C9F1CeJLxipNbkDu9UOYV6u5OdZ4T+CDcidfGuJZy3jpuQvDXEf 5z4vvVqmS6870QUBUV0Co9SJ6K/8iE7I5FXIjtTH0D8g/v6ExmTvjWVYjFno9B86Kcx0 wF/A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:subject:cc:to:from:date; bh=tqgZTZZTYjTRohL9fhydYfadSybiyMlC2nWpw1EJBcM=; b=DNM/uPzXlyfh6fPv8Sgw3Nx4EPd2HpkxSIcgvJ4r2VitdlxRojA3TvJYrOCaIKaD/5 T6oLAjJr9Dcudbic9IzuXjIYS0710kM0aJ3BY2GJYI5mbCXQSlGjGAzfx7NuKiX2XUJl RZBbP8/jCFObIm9SN73XCWIHFVcs5ZVo7NElLjjB8mmDwlwMT48G1jrEvW0zVAOygc0Y FneZGX+daaYeUEpypjeaKOvdK2P3wEb18MrXiiR/NGMm9v4B7xUaY+OLe9ySnlXm/W00 ug98SupiLxYhn4/1WbZCgTWFQqLdqS4Si/+QQC/2BLrP+WVNLbKLfrPnQZOwD1uK1/ZL pNXg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id d16-v6si6111220pgd.555.2018.10.24.15.19.17; Wed, 24 Oct 2018 15:19:34 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726678AbeJYGst (ORCPT + 99 others); Thu, 25 Oct 2018 02:48:49 -0400 Received: from mail.linuxfoundation.org ([140.211.169.12]:45936 "EHLO mail.linuxfoundation.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726297AbeJYGst (ORCPT ); Thu, 25 Oct 2018 02:48:49 -0400 Received: from akpm3.svl.corp.google.com (unknown [104.133.8.65]) by mail.linuxfoundation.org (Postfix) with ESMTPSA id E55711721; Wed, 24 Oct 2018 22:18:55 +0000 (UTC) Date: Wed, 24 Oct 2018 15:18:53 -0700 From: Andrew Morton To: Roman Gushchin Cc: "linux-mm@kvack.org" , "linux-kernel@vger.kernel.org" , Kernel Team , Michal Hocko , Rik van Riel , Randy Dunlap Subject: Re: [RFC PATCH] mm: don't reclaim inodes with many attached pages Message-Id: <20181024151853.3edd9097400b0d52edff1f16@linux-foundation.org> In-Reply-To: <20181023164302.20436-1-guro@fb.com> References: <20181023164302.20436-1-guro@fb.com> X-Mailer: Sylpheed 3.6.0 (GTK+ 2.24.31; x86_64-pc-linux-gnu) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, 23 Oct 2018 16:43:29 +0000 Roman Gushchin wrote: > Spock reported that the commit 172b06c32b94 ("mm: slowly shrink slabs > with a relatively small number of objects") leads to a regression on > his setup: periodically the majority of the pagecache is evicted > without an obvious reason, while before the change the amount of free > memory was balancing around the watermark. > > The reason behind is that the mentioned above change created some > minimal background pressure on the inode cache. The problem is that > if an inode is considered to be reclaimed, all belonging pagecache > page are stripped, no matter how many of them are there. So, if a huge > multi-gigabyte file is cached in the memory, and the goal is to > reclaim only few slab objects (unused inodes), we still can eventually > evict all gigabytes of the pagecache at once. > > The workload described by Spock has few large non-mapped files in the > pagecache, so it's especially noticeable. > > To solve the problem let's postpone the reclaim of inodes, which have > more than 1 attached page. Let's wait until the pagecache pages will > be evicted naturally by scanning the corresponding LRU lists, and only > then reclaim the inode structure. > > ... > > --- a/fs/inode.c > +++ b/fs/inode.c > @@ -730,8 +730,11 @@ static enum lru_status inode_lru_isolate(struct list_head *item, > return LRU_REMOVED; > } > > - /* recently referenced inodes get one more pass */ > - if (inode->i_state & I_REFERENCED) { > + /* > + * Recently referenced inodes and inodes with many attached pages > + * get one more pass. > + */ > + if (inode->i_state & I_REFERENCED || inode->i_data.nrpages > 1) { > inode->i_state &= ~I_REFERENCED; > spin_unlock(&inode->i_lock); > return LRU_ROTATE; hm, why "1"? I guess one could argue that this will encompass long symlinks, but I just made that up to make "1" appear more justifiable ;)