Received: by 2002:ad5:4acb:0:0:0:0:0 with SMTP id n11csp3446425imw; Thu, 7 Jul 2022 01:46:41 -0700 (PDT) X-Google-Smtp-Source: AGRyM1vBxfN02xODGYn+zR2AfzpsTouwiZykgM3HwFLV17ydPtuZjNLVDvVCBcNiLRj4MCwCZkpj X-Received: by 2002:a05:6402:354b:b0:437:60fd:4891 with SMTP id f11-20020a056402354b00b0043760fd4891mr61526280edd.344.1657183601601; Thu, 07 Jul 2022 01:46:41 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1657183601; cv=none; d=google.com; s=arc-20160816; b=wWfP7KUD1LmYDEhIvdQNJe3usWrNyBQPqTdeahXG/LNeVKar7Qaxka6w4pIhXI4HSs vkPcbxPXqunQV1o8gY5HL804ns9cX5LOd4yQxQI654vsjRBz8CJTHauS1MCDi1QL3rjU Ob0WNjk8mtdH6lfItms4SXLIFoehc45dQbBW5jFOzSTbtd23XglopqdqHPmlvPuWxlKJ lNmGwjLd0TObDeKrmqUogv4GdazGQieMD+zyXfqMgaLwdgAfQWPCPzGnzRgnN7PCyz1m RrhrJtpEQ/aELE6jdV2M2i0OmmpO0AFhkiXfyKPQ2FEx2EtFBlD8vdT5HVrnpS4lJL3B 2SjQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=2G3yGl1PipdwgXS3LUPNMuy6sfc53mWmHA9ICTmAWFA=; b=t1z8lGaAapzTOTybOBOmrIl+b5S33NoBnc7kR0YCevuzLdjWyNa5J+63Gy6LF31YWt bySeA/eaKtelW0HoVVWFGcrPemiid1YmS6zeZ2Rxb4Cd//o8Fimb4GD/kGzVxOrvn6hc W9u0gHjV53Gln3n5XFqw6KT1Uw2rkgSIKJDsQcsj6j13J29QIw8ENlrb9duQC++tiYy+ NVlkeb17vI/lNrojlHOVsuDxz5QZLrhDn/1BtJWx43gVl/rB+CVGF3BDuoOxwLRQ1wQe cRewvREn9hVESrncWH8RS7kNem2DgtoCIIAVRUVqU+b0LmNx6VS0euf0WWYN95ovW6lx qwRQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=BPoGseHI; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id hs40-20020a1709073ea800b0072aa56cd0a3si27430342ejc.28.2022.07.07.01.46.08; Thu, 07 Jul 2022 01:46:41 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=BPoGseHI; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234370AbiGGIW5 (ORCPT + 99 others); Thu, 7 Jul 2022 04:22:57 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:39060 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235592AbiGGIWq (ORCPT ); Thu, 7 Jul 2022 04:22:46 -0400 Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 30E341138 for ; Thu, 7 Jul 2022 01:22:45 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1657182165; x=1688718165; h=date:from:to:cc:subject:message-id:references: mime-version:in-reply-to; bh=PfJWDdzaRrnLREcfNPr+BcDTs6O3+NQxca0cViqrlxU=; b=BPoGseHIPB46yyONY6HGaoKrZGOZ3p/x/jZ0+MEBr591pLoSWDr1mhDK iTEDIUGkCOcjzjsXgbHnk2ZLNf9wF4EsuyaG2P5qAlwzUducBHVpFWTdM SHJbvSJkuiuJe0dE8OYp1T+lpW6AReZ2AbvffCeLMmY0QjusPNY4nXPpW O7vK1yoiBJZBRE7MHh3nBK5YBlnb7U8yQKKVsCQfBmroO2dc/YEivc1D0 k6twakMeGkpMujwsCd9+mslXzBiMpvjPRPJwJCRf9Ke/yf5wL9RWPuhg8 MZF5j807y7Ce75U+eFbk8Vo3nxr4LPsE+si6ywCg9GtuisMX7dlQLF7y8 A==; X-IronPort-AV: E=McAfee;i="6400,9594,10400"; a="282717760" X-IronPort-AV: E=Sophos;i="5.92,252,1650956400"; d="scan'208";a="282717760" Received: from fmsmga008.fm.intel.com ([10.253.24.58]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 07 Jul 2022 01:22:42 -0700 X-IronPort-AV: E=Sophos;i="5.92,252,1650956400"; d="scan'208";a="651039795" Received: from xsang-optiplex-9020.sh.intel.com (HELO xsang-OptiPlex-9020) ([10.239.159.143]) by fmsmga008-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 07 Jul 2022 01:22:39 -0700 Date: Thu, 7 Jul 2022 16:22:24 +0800 From: Oliver Sang To: Mel Gorman Cc: Andrew Morton , 0day robot , LKML , linux-mm@kvack.org, lkp@lists.01.org, Nicolas Saenz Julienne , Marcelo Tosatti , Vlastimil Babka , Michal Hocko , Hugh Dickins Subject: Re: [mm/page_alloc] 2bd8eec68f: BUG:sleeping_function_called_from_invalid_context_at_mm/gup.c Message-ID: References: <20220613125622.18628-8-mgorman@techsingularity.net> <20220703132209.875b823d1cb7169a8d51d56d@linux-foundation.org> <20220706095535.GD27531@techsingularity.net> <20220706115328.GE27531@techsingularity.net> <20220706145241.GG27531@techsingularity.net> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20220706145241.GG27531@techsingularity.net> X-Spam-Status: No, score=-5.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_MED, SPF_HELO_NONE,SPF_NONE,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Mel Gorman, On Wed, Jul 06, 2022 at 03:52:41PM +0100, Mel Gorman wrote: > On Wed, Jul 06, 2022 at 10:21:36PM +0800, Oliver Sang wrote: > > > I tried reproducing this on a 2-socket machine with Xeon > > > Gold Gold 5218R CPUs. It was necessary to set timeouts in both > > > vm/settings and kselftest/runner.sh to avoid timeouts. Testing with > > > a standard config on my original 5.19-rc3 baseline and the baseline > > > b13baccc3850ca8b8cccbf8ed9912dbaa0fdf7f3 both passed. I tried your kernel > > > config with i915 disabled (would not build) and necessary storage drivers > > > and network drivers enabled (for boot and access). The kernel log shows > > > a bunch of warnings related to USBAN during boot and during some of the > > > tests but otherwise compaction_test completed successfully as well as > > > the other VM tests. > > > > > > Is this always reproducible? > > > > not always but high rate. > > we actually also observed other dmesgs stats for both 2bd8eec68f74 and its > > parent > > Ok, it's unclear what the "other dmesg stats" are but given that it happens > for the parent. Does 5.19-rc2 (your baseline) have the same messages as > 2bd8eec68f74^? yeah, 5.19-rc2 has similar results as 2bd8eec68f74^ by multi-runs, while 2bd8eec68f74 looks quite similar to '-fix' commit which we applied it as 292baeb4c714. take the 'BUG:sleeping_function_called_from_invalid_context_at_mm/gup.c' we reported as example: v5.19-rc2 eec0ff5df2945d19039d16841b9 2bd8eec68f740608db5ea58ecff 292baeb4c7149ac2cb844137481 ---------------- --------------------------- --------------------------- --------------------------- fail:runs %reproduction fail:runs %reproduction fail:runs %reproduction fail:runs | | | | | | | :31 0% :20 55% 11:20 65% 13:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/gup.c the 'fail:runs' means we observed the issue 'fail' times while running 'runs' times. for v5.19-rc2, " :31", so we run the same jobs upon v5.19-rc2 31 times, but never see this "dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/gup.c" for eec0ff5df2 (2bd8eec68f74^), also clean on 20 runs. but for both 2bd8eec68f74 ("mm/page_alloc: Replace local_lock with normal spinlock") 292baeb4c714 ("mm/page_alloc: replace local_lock with normal spinlock -fix") it almost can reproduce in half of runs (11 out of 20 runs, 13 out of 31 runs respectively) the full comparison of these 4 commits are as [1] generally, for those dmesg.BUG:sleeping_function_called_from_invalid***, quite clean on v5.19-rc2 2bd8eec68f74^, but have similar rate on 2bd8eec68f74 & 292baeb4c714 but we also obversed other issues, such like "dmesg.RIP:rcu_eqs_exit", almost always happen on all 4 commits (this is what I said 'other dmesg stats', sorry for confusion, and I will avoid using this kind of 'internal' words in the future) > Does the kselftests vm suite always pass but sometimes > fails with 2bd8eec68f74? below is results of kselftests vm suite, so it's really like what you said, sometimes fails with 2bd8eec68f74 (also 292baeb4c714) one example is kernel-selftests.vm.run_vmtests.sh../userfaultfd_anon_20_16 if always pass on v5.19-rc2 and 2bd8eec68f74^ but fail 6 times out of 20 runs on 2bd8eec68f74, and fail 5 times out of 21 runs on 292baeb4c714 but since this rate seems not match with above issues, so not sure if they are related? v5.19-rc2 eec0ff5df2945d19039d16841b9 2bd8eec68f740608db5ea58ecff 292baeb4c7149ac2cb844137481 ---------------- --------------------------- --------------------------- --------------------------- fail:runs %reproduction fail:runs %reproduction fail:runs %reproduction fail:runs | | | | | | | 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.madv_populate.fail 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../gup_test_a.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../gup_test_ct_F_0x1_0_19_0x1000.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../gup_test_u.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../hugepage_mmap.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../hugepage_mremap_./huge/huge_mremap.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../hugepage_shm.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../hugepage_vmemmap.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../hugetlb_madvise_./huge/madvise_test.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../map_fixed_noreplace.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.run_vmtests.sh../map_hugetlb.pass 31:31 -35% 20:20 -30% 14:20 -20% 16:21 kernel-selftests.vm.run_vmtests.sh../userfaultfd_anon_20_16.pass 31:31 -35% 20:20 -30% 14:20 -20% 16:21 kernel-selftests.vm.run_vmtests.sh../userfaultfd_hugetlb_256_32.pass 29:31 -32% 19:20 -29% 13:20 -34% 12:21 kernel-selftests.vm.run_vmtests.sh../userfaultfd_shmem_20_16.pass 31:31 -35% 20:20 -35% 13:20 -25% 15:21 kernel-selftests.vm.run_vmtests.sh.fail 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.soft-dirty.pass 31:31 -35% 20:20 0% 20:20 5% 21:21 kernel-selftests.vm.split_huge_page_test.pass > > > but those dmesg.BUG:sleeping_function_called_from_invalid_context_at* > > seem only happen on 2bd8eec68f74 as well as the '-fix' commit. > > > > And roughly how often does it happen? I'm running it in a loop now to > see if I can trigger it locally. just as above, the 'BUG:sleeping_function_called_from_invalid_context_at_mm/gup.c' has around 50% rate to happen upon 2bd8eec68f74 & 292baeb4c714 BTW, will test that patch you mentioned in another mail later and update you with the results. > > -- > Mel Gorman > SUSE Labs [1] v5.19-rc2 eec0ff5df2945d19039d16841b9 2bd8eec68f740608db5ea58ecff 292baeb4c7149ac2cb844137481 ---------------- --------------------------- --------------------------- --------------------------- fail:runs %reproduction fail:runs %reproduction fail:runs %reproduction fail:runs | | | | | | | :31 0% :20 75% 15:20 70% 14:21 dmesg.BUG:scheduling_while_atomic :31 0% :20 5% 1:20 0% :21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_fs/binfmt_elf.c :31 0% :20 5% 1:20 10% 2:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_fs/dcache.c :31 0% :20 5% 1:20 5% 1:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_include/linux/freezer.h :31 0% :20 10% 2:20 25% 5:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_include/linux/mmu_notifier.h :31 0% :20 5% 1:20 0% :21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_include/linux/percpu-rwsem.h :31 0% :20 40% 8:20 40% 8:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_include/linux/sched/mm.h :31 0% :20 10% 2:20 0% :21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_kernel/locking/mutex.c :31 0% :20 10% 2:20 10% 2:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_lib/strncpy_from_user.c :31 0% :20 55% 11:20 65% 13:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/gup.c :31 0% :20 15% 3:20 5% 1:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/memory.c :31 0% :20 60% 12:20 55% 11:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/migrate.c :31 0% :20 5% 1:20 5% 1:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/page_alloc.c :31 0% :20 0% :20 5% 1:21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/rmap.c :31 0% :20 15% 3:20 0% :21 dmesg.BUG:sleeping_function_called_from_invalid_context_at_mm/vmalloc.c :31 0% :20 45% 9:20 45% 9:21 dmesg.BUG:workqueue_leaked_lock_or_atomic :31 0% :20 25% 5:20 15% 3:21 dmesg.Kernel_panic-not_syncing:Attempted_to_kill_init!exitcode= :31 0% :20 5% 1:20 0% :21 dmesg.RIP:__clear_user 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.RIP:rcu_eqs_exit 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.RIP:sched_clock_tick :31 0% :20 5% 1:20 0% :21 dmesg.RIP:smp_call_function_many_cond 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.WARNING:at_kernel/rcu/tree.c:#rcu_eqs_exit 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.WARNING:at_kernel/sched/clock.c:#sched_clock_tick :31 0% :20 5% 1:20 0% :21 dmesg.WARNING:at_kernel/smp.c:#smp_call_function_many_cond 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.WARNING:suspicious_RCU_usage 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.boot_failures 11:31 -6% 9:20 -5% 6:20 5% 8:21 dmesg.include/linux/rcupdate.h:#rcu_read_lock()used_illegally_while_idle 11:31 -6% 9:20 -5% 6:20 5% 8:21 dmesg.include/linux/rcupdate.h:#rcu_read_unlock()used_illegally_while_idle 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.include/trace/events/error_report.h:#suspicious_rcu_dereference_check()usage 29:31 -29% 20:20 6% 20:20 11% 21:21 dmesg.include/trace/events/lock.h:#suspicious_rcu_dereference_check()usage