Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp5651996rwd; Wed, 24 May 2023 05:07:55 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ51dPp76BXx7WY46FVd8ypsYg9IIZ0G1pUiUhM4eQtzdgJrZIAbm2+zIUB/1aAjVEUAa31d X-Received: by 2002:aa7:88d0:0:b0:64d:6251:4ee5 with SMTP id k16-20020aa788d0000000b0064d62514ee5mr2639977pff.22.1684930075215; Wed, 24 May 2023 05:07:55 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1684930075; cv=none; d=google.com; s=arc-20160816; b=oN7/VCeWuyjIV2/CrZ4aJ+bTiV5u4A/P0KCRzZk9fh64I4U0XFLd2fmv47BQIMWsXA irtxzkj63EUKfvlTHV9TG3kDTuTl6V6Lf593TZADqoYdc4XYu8I1IgO5jSnJwf/tVEzw JzvNvMiORXpcuAC+aF6Gri33cm3DoSydVcwlFHQMcECHpRPRPHHMYY1/WryRsHDIUjP5 Txq+//YzBVu+NgcULSQ8JLBTiF5n3midhirzwsgIAybsyfXmhPoUfjjUF5Hri2WindZs VUde21pIakxbfPcJxblpq6rCRn05tWT5IHeaOfpKZQdfUaNX/38uKlva4AKYeHlV3C1P GSHQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:references :cc:to:from:content-language:subject:mime-version:date :dkim-signature:message-id; bh=uMsexn7FypO2MxTw7kOkZKMP6/q4BmJjDQCpRviZwTk=; b=W7NGqTcRv6jnqVkLP13hImN2ECpASml4XSrj6WYuKRsZLVQqGGuSHC7RAnYYQ8wzTp 21Wx9FGzEyXj7ZZVYHqPyGBpbVh+HLzVAxfCSJxUWXekLB+HJxER/lSLzKh6YEQcjXPE Qf+DlhdScGynonEE97MrVwiMX3zaM4G75NhKKi10LciMLGppwQXmERVHHdmXOa4sVxDy AiAzoL1Mj+3f6SO/QkbvrUhrxIq5H0ZLkzE6WtBvxxfB1brEQfj5xM0mk/wZTjsrh8c5 kNY5CxBlCGdKjvPs1+WNU5FMVSBTaaNtPAwrDomGSinlUKOUnHv/C+gAqRYsMffHzUK/ 3dbQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linux.dev header.s=key1 header.b=WiHpg5RN; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linux.dev Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id l65-20020a622544000000b0064d3a475274si8135247pfl.75.2023.05.24.05.07.40; Wed, 24 May 2023 05:07:55 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@linux.dev header.s=key1 header.b=WiHpg5RN; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linux.dev Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229919AbjEXL4s (ORCPT + 99 others); Wed, 24 May 2023 07:56:48 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:37022 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231704AbjEXL4q (ORCPT ); Wed, 24 May 2023 07:56:46 -0400 Received: from out-59.mta1.migadu.com (out-59.mta1.migadu.com [IPv6:2001:41d0:203:375::3b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id F2D7C189 for ; Wed, 24 May 2023 04:56:41 -0700 (PDT) Message-ID: DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.dev; s=key1; t=1684929400; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=uMsexn7FypO2MxTw7kOkZKMP6/q4BmJjDQCpRviZwTk=; b=WiHpg5RNA4GTwK2SDkHLvlaHzfo3/HpO34N4XoPdsW0f4IrTrk2Utap9fkZOHfoZnzWEUM 0CNPosbjasx5hfvCkRzmQcfaYfYXeLk2MywMuAOKAjaRsOv+4Vn4L3rjgUgWuewCVawufe gIoLMAnY89tKHEm9SxQWeeC/G6kpSrM= Date: Wed, 24 May 2023 19:56:30 +0800 MIME-Version: 1.0 Subject: Re: [linus:master] [mm] f95bdb700b: stress-ng.ramfs.ops_per_sec -88.8% regression Content-Language: en-US X-Report-Abuse: Please report any abuse attempt to abuse@migadu.com and include these headers. From: Qi Zheng To: RCU Cc: oe-lkp@lists.linux.dev, lkp@intel.com, linux-kernel@vger.kernel.org, Andrew Morton , Vlastimil Babka , Kirill Tkhai , Roman Gushchin , =?UTF-8?Q?Christian_K=c3=b6nig?= , David Hildenbrand , Davidlohr Bueso , Johannes Weiner , Michal Hocko , Muchun Song , "Paul E. McKenney" , Shakeel Butt , Yang Shi , linux-mm@kvack.org, ying.huang@intel.com, feng.tang@intel.com, fengwei.yin@intel.com, Yujie Liu References: <202305230837.db2c233f-yujie.liu@intel.com> <896bbb09-d400-ec73-ba3a-b64c6e9bbe46@linux.dev> In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Migadu-Flow: FLOW_OUT X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,SPF_HELO_NONE,SPF_PASS, T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2023/5/24 19:08, Qi Zheng wrote: [...] > > Well, I just ran the following command and reproduced the result: > > stress-ng --timeout 60 --times --verify --metrics-brief --ramfs 9 & > > 1) with commit 42c9db3970483: > > stress-ng: info:  [11023] setting to a 60 second run per stressor > stress-ng: info:  [11023] dispatching hogs: 9 ramfs > stress-ng: info:  [11023] stressor       bogo ops real time  usr time > sys time   bogo ops/s     bogo ops/s > stress-ng: info:  [11023]                           (secs)    (secs) > (secs)   (real time) (usr+sys time) > stress-ng: info:  [11023] ramfs            774966     60.00     10.18 > 169.45     12915.89        4314.26 > stress-ng: info:  [11023] for a 60.00s run time: > stress-ng: info:  [11023]    1920.11s available CPU time > stress-ng: info:  [11023]      10.18s user time   (  0.53%) > stress-ng: info:  [11023]     169.44s system time (  8.82%) > stress-ng: info:  [11023]     179.62s total time  (  9.35%) > stress-ng: info:  [11023] load average: 8.99 2.69 0.93 > stress-ng: info:  [11023] successful run completed in 60.00s (1 min, > 0.00 secs) > > 2) with commit f95bdb700bc6b: > > stress-ng: info:  [37676] dispatching hogs: 9 ramfs > stress-ng: info:  [37676] stressor       bogo ops real time  usr time > sys time   bogo ops/s     bogo ops/s > stress-ng: info:  [37676]                           (secs)    (secs) > (secs)   (real time) (usr+sys time) > stress-ng: info:  [37676] ramfs            168673     60.00      1.61 >  39.66      2811.08        4087.47 > stress-ng: info:  [37676] for a 60.10s run time: > stress-ng: info:  [37676]    1923.36s available CPU time > stress-ng: info:  [37676]       1.60s user time   (  0.08%) > stress-ng: info:  [37676]      39.66s system time (  2.06%) > stress-ng: info:  [37676]      41.26s total time  (  2.15%) > stress-ng: info:  [37676] load average: 7.69 3.63 2.36 > stress-ng: info:  [37676] successful run completed in 60.10s (1 min, > 0.10 secs) > > The bogo ops/s (real time) did drop significantly. > > And the memory reclaimation was not triggered in the whole process. so > theoretically no one is in the read critical section of shrinker_srcu. > > Then I found that some stress-ng-ramfs processes were in > TASK_UNINTERRUPTIBLE state for a long time: > > root       42313  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42314  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42315  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42316  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42317  7.8  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42318  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42319  7.8  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42320  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42321  7.8  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42322  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42323  7.8  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42324  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42325  7.8  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42326  0.0  0.0  69592  2068 pts/0    S    19:00   0:00 > stress-ng-ramfs [run] > root       42327  7.9  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42328  7.9  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42329  7.9  0.0  69592  1812 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > root       42330  7.9  0.0  69592  1556 pts/0    D    19:00   0:02 > stress-ng-ramfs [run] > > Their call stack is as follows: > > cat /proc/42330/stack > > [<0>] __synchronize_srcu.part.21+0x83/0xb0 > [<0>] unregister_shrinker+0x85/0xb0 > [<0>] deactivate_locked_super+0x27/0x70 > [<0>] cleanup_mnt+0xb8/0x140 > [<0>] task_work_run+0x65/0x90 > [<0>] exit_to_user_mode_prepare+0x1ba/0x1c0 > [<0>] syscall_exit_to_user_mode+0x1b/0x40 > [<0>] do_syscall_64+0x44/0x80 > [<0>] entry_SYSCALL_64_after_hwframe+0x63/0xcd > > + RCU folks, Is this result as expected? I would have thought that > synchronize_srcu() should return quickly if no one is in the read > critical section. :( > With the following changes, ops/s can return to previous levels: diff --git a/mm/vmscan.c b/mm/vmscan.c index db2ed6e08f67..90f541b07cd1 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -763,7 +763,7 @@ void unregister_shrinker(struct shrinker *shrinker) debugfs_entry = shrinker_debugfs_remove(shrinker); up_write(&shrinker_rwsem); - synchronize_srcu(&shrinker_srcu); + synchronize_srcu_expedited(&shrinker_srcu); debugfs_remove_recursive(debugfs_entry); stress-ng: info: [13159] dispatching hogs: 9 ramfs stress-ng: info: [13159] stressor bogo ops real time usr time sys time bogo ops/s bogo ops/s stress-ng: info: [13159] (secs) (secs) (secs) (real time) (usr+sys time) stress-ng: info: [13159] ramfs 710062 60.00 9.63 157.26 11834.18 4254.75 stress-ng: info: [13159] for a 60.00s run time: stress-ng: info: [13159] 1920.14s available CPU time stress-ng: info: [13159] 9.62s user time ( 0.50%) stress-ng: info: [13159] 157.26s system time ( 8.19%) stress-ng: info: [13159] 166.88s total time ( 8.69%) stress-ng: info: [13159] load average: 9.49 4.02 1.65 stress-ng: info: [13159] successful run completed in 60.00s (1 min, 0.00 secs) Can we make synchronize_srcu() call synchronize_srcu_expedited() when no one is in the read critical section? > -- Thanks, Qi