Received: by 2002:a05:6358:5282:b0:b5:90e7:25cb with SMTP id g2csp1780577rwa; Sun, 21 Aug 2022 17:18:56 -0700 (PDT) X-Google-Smtp-Source: AA6agR5iKvNtsgSidWTG3Y5kqd9MBgxvGjC8u637CpX6WGZtSGHl9WPxh3raMbyTUexkNYCZDeQ8 X-Received: by 2002:a05:6a00:1705:b0:52f:6028:5c33 with SMTP id h5-20020a056a00170500b0052f60285c33mr18320888pfc.29.1661127536466; Sun, 21 Aug 2022 17:18:56 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1661127536; cv=none; d=google.com; s=arc-20160816; b=MKYNuOEQULbORwtGDJpVMYGe2s8x5oFZt76xr+nrS7vJLByaFr5WkINcGx8LNSzB2D DylKHNVDEIXiheorhq2PscdBl0X2+6YzTwmS5Y1T5dffbB+XkhosQrIavCEIUmUHDdeL zzoVyKVnPOBfJQS7gAJuG61e3D8GhK3SAZA91D4+AHnQ1xw4kDDGHkTaCd1TSrdj7bvt cgUX5eDNXhqa1rJpW6F8yzKVFHX9+JpkVc79S5OFrNKD5qgBvrBMF/Q3prT1i8fJjUTe cNHWjNueCOzlGsJH7Rzkl3m4gFK+A4ksvZnpCkKIW4rX8BSEeTFivmDzKMeNMuekndRD 1ewg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:from:subject:mime-version:message-id:date :dkim-signature; bh=eVYfwZxwoaftgj4urkXWqJfl1yf2BkP0ai3v6afDBzY=; b=SY9G7Sk2byFa/GSc/3RcEuuH15lgYYxIk/vYadFgQ5mGstbOUtP9n+AupqQqJi3k68 SqW2ghgRJS1iYuhier3YQu2opSCh3vrAR9vhNLQ5Gf037sZKrF+T7WV8VWE89S05c9vN GWsxyTYpvyyTYHFRHQK7L1uulCw01QLNKk/D4PALTua72dDs3UVpBEmRHLqr/s8F3aVx um5Fv1EmhboOzctx8vKQD0ZLomzVf3QHpE3spjtVqSvGURH08O1gRaokVmm/Puu5TP0T b+4QYReXYetNZTc3zMG8Qo1AkygxHtbE2Wlw6Y65pKhgQrAXbotyzczYRkEMc2EM0Eg2 2pQw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=M1l5wTrP; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id 29-20020a17090a001d00b001f5073da33fsi13220393pja.131.2022.08.21.17.18.45; Sun, 21 Aug 2022 17:18:56 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=M1l5wTrP; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231923AbiHVARr (ORCPT + 99 others); Sun, 21 Aug 2022 20:17:47 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33442 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229541AbiHVARp (ORCPT ); Sun, 21 Aug 2022 20:17:45 -0400 Received: from mail-yw1-x1149.google.com (mail-yw1-x1149.google.com [IPv6:2607:f8b0:4864:20::1149]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id EF8B811444 for ; Sun, 21 Aug 2022 17:17:44 -0700 (PDT) Received: by mail-yw1-x1149.google.com with SMTP id 00721157ae682-3339532b6a8so160523457b3.1 for ; Sun, 21 Aug 2022 17:17:44 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=cc:to:from:subject:mime-version:message-id:date:from:to:cc; bh=eVYfwZxwoaftgj4urkXWqJfl1yf2BkP0ai3v6afDBzY=; b=M1l5wTrPM/Bn69v9qPWUy2ze89j/C2QhgC3uGWDIOLN0WzyTcSLHJG+RMJMatiXqzX 27XcKNJabca81sRVdeO2CT1uL6V3eUrxJu8cAS5kLO6npO5tc6V6+FSswOr5rvF6RaE5 NnZaglMI2t5Fcdxoo3js4AEbpcjU5+NawpoYbCoq0mblDVlFiOzpAKZf/doM3EvCHCLx 4Q3+m22Td0kDdin0LFVj23XqtyJX1saMxHUxHMwrnzPDoQvY1ZsDuHKa9Kh07T/Mylwk qX7Y7PhcDJVc31z1kgWb/ztxtLAbJS62kPbXsESgW/3GkZHCTfqGGUoWHVPWeD15A1hl kqqA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:from:subject:mime-version:message-id:date:x-gm-message-state :from:to:cc; bh=eVYfwZxwoaftgj4urkXWqJfl1yf2BkP0ai3v6afDBzY=; b=z9ZWwjnoNRjeWdqbtBoyg/YJ/FmNgzxP8yWMQhxLkVbwN+z09Tc3sscRVKqstXscaF KZcMC9KJLVjBdAFXcYW6MK53GODifIS/SasbJJJtDi05f6qXz7UAVvkXEwzixPsSnm5M 3knlRLSjH74MLhe99lAVstALEWAj2l6zX+u20a71Z0LbPxwoDinlsFUa6K8bZKxK4NMQ hkPITC/tvMpO0kwNgMmc7IE5Wj1JWeFtnAHMpmFyGt5Ymx1ecE7JYSfM4nnxATqyF/al 9Y4nqpDh2pLygCxCh0R83UOns4YEgD4OplDloqw2sgbS3pTDTXu06zN81RDok1/wi5+A bPgw== X-Gm-Message-State: ACgBeo3wDOkBE7qLR5FHSqkRUuDgMIoPirexPO3J5SzRCek0gF0gPGU+ LUEAHwUUdaNgkMii1U8gfS/dAkPrbRC1TA== X-Received: from shakeelb.c.googlers.com ([fda3:e722:ac3:cc00:20:ed76:c0a8:28b]) (user=shakeelb job=sendgmr) by 2002:a25:7304:0:b0:693:bc0d:fc6c with SMTP id o4-20020a257304000000b00693bc0dfc6cmr15772624ybc.375.1661127464253; Sun, 21 Aug 2022 17:17:44 -0700 (PDT) Date: Mon, 22 Aug 2022 00:17:34 +0000 Message-Id: <20220822001737.4120417-1-shakeelb@google.com> Mime-Version: 1.0 X-Mailer: git-send-email 2.37.1.595.g718a3a8f04-goog Subject: [PATCH 0/3] memcg: optimizatize charge codepath From: Shakeel Butt To: Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: "=?UTF-8?q?Michal=20Koutn=C3=BD?=" , Eric Dumazet , Soheil Hassas Yeganeh , Feng Tang , Oliver Sang , Andrew Morton , lkp@lists.01.org, cgroups@vger.kernel.org, linux-mm@kvack.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, Shakeel Butt Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-9.6 required=5.0 tests=BAYES_00,DKIMWL_WL_MED, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE,USER_IN_DEF_DKIM_WL autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Recently Linux networking stack has moved from a very old per socket pre-charge caching to per-cpu caching to avoid pre-charge fragmentation and unwarranted OOMs. One impact of this change is that for network traffic workloads, memcg charging codepath can become a bottleneck. The kernel test robot has also reported this regression. This patch series tries to improve the memcg charging for such workloads. This patch series implement three optimizations: (A) Reduce atomic ops in page counter update path. (B) Change layout of struct page_counter to eliminate false sharing between usage and high. (C) Increase the memcg charge batch to 64. To evaluate the impact of these optimizations, on a 72 CPUs machine, we ran the following workload in root memcg and then compared with scenario where the workload is run in a three level of cgroup hierarchy with top level having min and low setup appropriately. $ netserver -6 # 36 instances of netperf with following params $ netperf -6 -H ::1 -l 60 -t TCP_SENDFILE -- -m 10K Results (average throughput of netperf): 1. root memcg 21694.8 2. 6.0-rc1 10482.7 (-51.6%) 3. 6.0-rc1 + (A) 14542.5 (-32.9%) 4. 6.0-rc1 + (B) 12413.7 (-42.7%) 5. 6.0-rc1 + (C) 17063.7 (-21.3%) 6. 6.0-rc1 + (A+B+C) 20120.3 (-7.2%) With all three optimizations, the memcg overhead of this workload has been reduced from 51.6% to just 7.2%. Shakeel Butt (3): mm: page_counter: remove unneeded atomic ops for low/min mm: page_counter: rearrange struct page_counter fields memcg: increase MEMCG_CHARGE_BATCH to 64 include/linux/memcontrol.h | 7 ++++--- include/linux/page_counter.h | 34 +++++++++++++++++++++++----------- mm/page_counter.c | 13 ++++++------- 3 files changed, 33 insertions(+), 21 deletions(-) -- 2.37.1.595.g718a3a8f04-goog