Received: by 2002:a25:7ec1:0:0:0:0:0 with SMTP id z184csp572497ybc; Tue, 12 Nov 2019 06:07:50 -0800 (PST) X-Google-Smtp-Source: APXvYqyf00VAsbA0tr8JBisPIFlhozPS5hqZE8DEL4AKW7WbKNggVTYS+BcIn245pce6K9XsEKEJ X-Received: by 2002:aa7:d0c1:: with SMTP id u1mr32953836edo.27.1573567670720; Tue, 12 Nov 2019 06:07:50 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1573567670; cv=none; d=google.com; s=arc-20160816; b=KC8Ono3URjbY4qXqQC9UH24tIC0cL32BlJjJYASxmCo7qjjPGkTSVS0AXrfZ7b/DaG IrlJ0FP3j0RCTzHFGgBkS8We8YlpC9erf3uj3z8/kFMTOPMZOspFi6nGROPff27bdImJ kVUjTau3RC4suThX4JEUo7DstXBRb4ibY7EMVOKaSFfHxnaKBF8NwooT9xilVuaa8t7W D4EWz2ausynsqC9MHsJ8zZikiV6vBASO4KQy1Vvf6MFoXay1IDeY/4wl62I1/4T2SCe7 JX1cWnEsmm+4AF2EG1DauGI/y6DTjSHkJNNqxjHUhOhf4UjBwWPgsEubh4fr7Sx1S8Be yLig== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:message-id:date:subject:to:from; bh=jPEYYr6+ZG0qkJp9GN+XLPP2KgrcXmBco99rQmsP6Co=; b=mGEfNrhOlVSh/0tjFs0W4YVJmDlOAjVuMRZZpWzvCfdGRuByeqjOM54jBtXUFdWf0P 5Pzg/cArEw/G3769puzbRG6v5AgQ9NjyqttJURWJMQ7iVCFC7x669Ooeqw306xiFYi26 qKrM8bi6hWow89VEGrKAhW8a9CPRx92Cd2RhGTbodfkqHk3xk86ZaAXwI5B91J54NH2Q ldKSpqz/IAXXZ6uJQHvJC0nGDe+QDr/j1P8gPy4WJF7beNG3p01FD5ETY/7njubEZewX 3vCphZhUgglpRX67fFbiEPUuNh8lKtS9wnqHQD5dqGAm9nd62o1Ox7O100+eZUQk2d8H ZrAA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id n2si23857ejy.60.2019.11.12.06.07.25; Tue, 12 Nov 2019 06:07:50 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727384AbfKLOGm (ORCPT + 99 others); Tue, 12 Nov 2019 09:06:42 -0500 Received: from out30-133.freemail.mail.aliyun.com ([115.124.30.133]:46872 "EHLO out30-133.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726497AbfKLOGm (ORCPT ); Tue, 12 Nov 2019 09:06:42 -0500 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R101e4;CH=green;DM=||false|;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e01422;MF=alex.shi@linux.alibaba.com;NM=1;PH=DS;RN=11;SR=0;TI=SMTPD_---0Thubem1_1573567598; Received: from localhost(mailfrom:alex.shi@linux.alibaba.com fp:SMTPD_---0Thubem1_1573567598) by smtp.aliyun-inc.com(127.0.0.1); Tue, 12 Nov 2019 22:06:38 +0800 From: Alex Shi To: alex.shi@linux.alibaba.com, cgroups@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, akpm@linux-foundation.org, mgorman@techsingularity.net, tj@kernel.org, hughd@google.com, khlebnikov@yandex-team.ru, daniel.m.jordan@oracle.com, yang.shi@linux.alibaba.com Subject: [PATCH v2 0/8] per lruvec lru_lock for memcg Date: Tue, 12 Nov 2019 22:06:20 +0800 Message-Id: <1573567588-47048-1-git-send-email-alex.shi@linux.alibaba.com> X-Mailer: git-send-email 1.8.3.1 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi all, This patchset move lru_lock into lruvec, give a lru_lock for each of lruvec, thus bring a lru_lock for each of memcg per node. According to Daniel Jordan's suggestion, I run 64 'dd' with on 32 containers on my 2s* 8 core * HT box with the modefied case: https://git.kernel.org/pub/scm/linux/kernel/git/wfg/vm-scalability.git/tree/case-lru-file-readtwice With this change above lru_lock censitive testing improved 17% with multiple containers scenario. And no performance lose w/o mem_cgroup. Thanks Hugh Dickins and Konstantin Khlebnikov, they both bring the same idea 7 years ago. I don't know why they didn't go further, but according to my testing, and google internal usage. This feathre is clearly benefit multi-container user. So I like to introduce it here. v2: bypass a performance regression bug and fix some function issues --- Documentation/admin-guide/cgroup-v1/memcg_test.rst | 15 +++------------ Documentation/admin-guide/cgroup-v1/memory.rst | 6 +++--- Documentation/trace/events-kmem.rst | 2 +- Documentation/vm/unevictable-lru.rst | 22 ++++++++-------------- include/linux/memcontrol.h | 67 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ include/linux/mm_types.h | 2 +- include/linux/mmzone.h | 7 +++++-- mm/compaction.c | 62 ++++++++++++++++++++++++++++++++++++++++++-------------------- mm/filemap.c | 4 ++-- mm/huge_memory.c | 16 ++++++---------- mm/memcontrol.c | 64 +++++++++++++++++++++++++++++++++++++++++++++++++++------------- mm/mlock.c | 27 ++++++++++++++------------- mm/mmzone.c | 1 + mm/page_alloc.c | 1 - mm/page_idle.c | 5 +++-- mm/rmap.c | 2 +- mm/swap.c | 77 +++++++++++++++++++++++++++++++---------------------------------------------- mm/vmscan.c | 74 ++++++++++++++++++++++++++++++++++++++------------------------------------ 18 files changed, 277 insertions(+), 177 deletions(-) [PATCH v2 1/8] mm/lru: add per lruvec lock for memcg [PATCH v2 2/8] mm/lruvec: add irqsave flags into lruvec struct [PATCH v2 3/8] mm/lru: replace pgdat lru_lock with lruvec lock [PATCH v2 4/8] mm/lru: only change the lru_lock iff page's lruvec is [PATCH v2 5/8] mm/pgdat: remove pgdat lru_lock [PATCH v2 6/8] mm/lru: remove rcu_read_lock to fix performance [PATCH v2 7/8] mm/lru: likely enhancement [PATCH v2 8/8] mm/lru: revise the comments of lru_lock