Received: by 2002:a05:6358:45e:b0:b5:b6eb:e1f9 with SMTP id 30csp64995rwe; Wed, 24 Aug 2022 17:06:49 -0700 (PDT) X-Google-Smtp-Source: AA6agR6Sk1kEa/2JRBJu+y9jphfdxlqErivd7drV3zttorJyR6F7seXwbEKyPzMoCFJoZTMrBtAL X-Received: by 2002:a63:9042:0:b0:41c:cdd4:ae66 with SMTP id a63-20020a639042000000b0041ccdd4ae66mr1059069pge.47.1661386008795; Wed, 24 Aug 2022 17:06:48 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1661386008; cv=none; d=google.com; s=arc-20160816; b=OySllkYDPorrcNh1HrCg6F5I9KySru8DLneIALB9IBBx/OCfrgrgoz0gvQhctbn+Ht gPaLtT3sGeT6rue28r2iBWilanOpjQ8w+Fs26t/3XUM2a4x3kchvwb5euFBnCFgMweyX aGHl18u2JQSCX17teKcEyOQQSUg/YiFY7MNUnbxqBOSAMdBKy99b5ASKk6I/EcT7Xacy H12IptQ0MJ0BEXCsMlVeTdTF91ZzfOzkzc0EmfEImtyHrc3rHZyJbIWuw9Ut05wHRxg8 /4I1PuNz/Qp5Y8BmY7lBRWDfpjBOBSv4Bh/2INVgPpa5Hs3OoRKv3gI0XFd8cgCWyoIV jpJw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:from:subject:mime-version:message-id:date :dkim-signature; bh=vgA35UGCl2OHwAbiBhc19zCa8HdxydG1D9TkQ3os1vo=; b=o8W3rH7b8qBu2pex5mvwUI08ZEkQ+QJoL2wrSorw94xfeo/XI3kqAqtEC/IzBk01Th zCnPmoVGTdc2LgKT9n9NhZBE9PJoN5+/2mTViT2MKDLgxP7vokScdl53nayo30/zLnZ+ LjcBt+/MxpnJYnht6tdNDAk2vxshuihAEZ+KF1puaUS6uYFOEnD2PFxfRy9m+gM9nduc 4chvDm4Pe/RkEiCKMzfXqQA1ulK8pkilfGxksbU83iRxBMRt/dta1zVgdQeC8+T7HcUs AIDtlmyF1HBkKyUbzi9jc7NxvWdpi4rH8wgE5fdxIl5Qpmh/mfQH8ugErF09O6LBg85k Ly7w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=Ge7brMID; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id k19-20020a170902ba9300b0017312cc5263si2863977pls.613.2022.08.24.17.06.36; Wed, 24 Aug 2022 17:06:48 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=Ge7brMID; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230234AbiHYAF1 (ORCPT + 99 others); Wed, 24 Aug 2022 20:05:27 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:59180 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230338AbiHYAFY (ORCPT ); Wed, 24 Aug 2022 20:05:24 -0400 Received: from mail-pj1-x1049.google.com (mail-pj1-x1049.google.com [IPv6:2607:f8b0:4864:20::1049]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1D94A61D79 for ; Wed, 24 Aug 2022 17:05:22 -0700 (PDT) Received: by mail-pj1-x1049.google.com with SMTP id a16-20020a17090abe1000b001fad8c29b0bso1821162pjs.2 for ; Wed, 24 Aug 2022 17:05:22 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=cc:to:from:subject:mime-version:message-id:date:from:to:cc; bh=vgA35UGCl2OHwAbiBhc19zCa8HdxydG1D9TkQ3os1vo=; b=Ge7brMIDek9qbeisJgfHL6rmBn4FcJKYA9NOFZ0ssAyFWYzyQc4kXS30ZHP2em4Z1h P3TuTjpB+zzcwiH6oROwzlRsWero2l5QxfYmurrc9hJBSB8kUcwtU4kRAIqOTRDxWiI9 FnZosPt+RBbfPjyDFj14FpKbfQ3sV+YnH3jHUhIvfJSXyA0AUA63cdDg11e8xX4/Myfg EqjDHsQ+yqVJdJoUnYtFmb24rYbR/drHoWA/7uWQxfNfnukE7muUDFSfj7q7MxclK3cA TIrcuHrDD3wmMumrUkRreEIKBtZbjTxfTYXMiVtJ9KUxV0mLM3E70ZKhnC+Rx7IbBro5 vwJg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:from:subject:mime-version:message-id:date:x-gm-message-state :from:to:cc; bh=vgA35UGCl2OHwAbiBhc19zCa8HdxydG1D9TkQ3os1vo=; b=NlwchtE8Si1hLO5XAFk5hxDr4SopXn22P0daH8V/4cRWuu0wIchr1hstoqSziGbOz6 QMm7h1iUYHVIdx2P7rTtq9nBlpSSlIk9kzv/n1H1U3nJ9i4roJoMuT8eRO+82DmClRio A9DGT5IxBE5yjzpmylq1Fa3AN3m3pYYnyD8IABEluHuEh3v8XnXb8Ps57eQD2Wg+Qz0a 3gF4Y/UhZerK5x4EcI2qKLcB3mPJdNC/edXHmpzrdZVEWRF8d8m+1Rm/iI2bNJV+GJ3j RXJf32x2mDoFXunSdeQ3FbvJO4Ww6YsSyNdmsC6VFruPdT4Tc7jdA51Xq9+5DQLVcaAu FLUw== X-Gm-Message-State: ACgBeo2w37baEX1Z7NaM7VndV7zEXifpDEns6WJBRUQCzazpPqjNF/xK BZ1bqSmdOoDsMwMlfVDZPPEgxj27R/i7yQ== X-Received: from shakeelb.c.googlers.com ([fda3:e722:ac3:cc00:7f:e700:c0a8:262e]) (user=shakeelb job=sendgmr) by 2002:a62:1649:0:b0:536:55af:1f4d with SMTP id 70-20020a621649000000b0053655af1f4dmr1405863pfw.61.1661385921563; Wed, 24 Aug 2022 17:05:21 -0700 (PDT) Date: Thu, 25 Aug 2022 00:05:03 +0000 Message-Id: <20220825000506.239406-1-shakeelb@google.com> Mime-Version: 1.0 X-Mailer: git-send-email 2.37.1.595.g718a3a8f04-goog Subject: [PATCH v2 0/3] memcg: optimize charge codepath From: Shakeel Butt To: Johannes Weiner , Michal Hocko , Roman Gushchin , Muchun Song Cc: "=?UTF-8?q?Michal=20Koutn=C3=BD?=" , Eric Dumazet , Soheil Hassas Yeganeh , Feng Tang , Oliver Sang , Andrew Morton , lkp@lists.01.org, cgroups@vger.kernel.org, linux-mm@kvack.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, Shakeel Butt Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-9.6 required=5.0 tests=BAYES_00,DKIMWL_WL_MED, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE,USER_IN_DEF_DKIM_WL autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Recently Linux networking stack has moved from a very old per socket pre-charge caching to per-cpu caching to avoid pre-charge fragmentation and unwarranted OOMs. One impact of this change is that for network traffic workloads, memcg charging codepath can become a bottleneck. The kernel test robot has also reported this regression[1]. This patch series tries to improve the memcg charging for such workloads. This patch series implement three optimizations: (A) Reduce atomic ops in page counter update path. (B) Change layout of struct page_counter to eliminate false sharing between usage and high. (C) Increase the memcg charge batch to 64. To evaluate the impact of these optimizations, on a 72 CPUs machine, we ran the following workload in root memcg and then compared with scenario where the workload is run in a three level of cgroup hierarchy with top level having min and low setup appropriately. $ netserver -6 # 36 instances of netperf with following params $ netperf -6 -H ::1 -l 60 -t TCP_SENDFILE -- -m 10K Results (average throughput of netperf): 1. root memcg 21694.8 Mbps 2. 6.0-rc1 10482.7 Mbps (-51.6%) 3. 6.0-rc1 + (A) 14542.5 Mbps (-32.9%) 4. 6.0-rc1 + (B) 12413.7 Mbps (-42.7%) 5. 6.0-rc1 + (C) 17063.7 Mbps (-21.3%) 6. 6.0-rc1 + (A+B+C) 20120.3 Mbps (-7.2%) With all three optimizations, the memcg overhead of this workload has been reduced from 51.6% to just 7.2%. [1] https://lore.kernel.org/linux-mm/20220619150456.GB34471@xsang-OptiPlex-9020/ Changes since v1: - Commit message updates - Instead of explicit padding add align compiler option with struct Shakeel Butt (3): mm: page_counter: remove unneeded atomic ops for low/min mm: page_counter: rearrange struct page_counter fields memcg: increase MEMCG_CHARGE_BATCH to 64 include/linux/memcontrol.h | 7 ++++--- include/linux/page_counter.h | 34 +++++++++++++++++++++++----------- mm/page_counter.c | 13 ++++++------- 3 files changed, 33 insertions(+), 21 deletions(-) -- 2.37.1.595.g718a3a8f04-goog