Received: by 2002:a05:6358:9144:b0:117:f937:c515 with SMTP id r4csp10124105rwr; Fri, 12 May 2023 04:16:57 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ4zTFydoqZ2X9nJ9QWsivqsJbZFuaSMW270OYIPdHfm6WfLPT3wSuTs3JSAZbf/v8VBDwwa X-Received: by 2002:a17:90a:a681:b0:237:62f7:3106 with SMTP id d1-20020a17090aa68100b0023762f73106mr24270507pjq.17.1683890217753; Fri, 12 May 2023 04:16:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1683890217; cv=none; d=google.com; s=arc-20160816; b=NJSOCTzSM5KlnlH8OkujLhMDUuSYICxUz0FDqxbEAz9s6ktnpMo/LSjGCrD+q5rRW9 XkS9DJE/qyEy+4IhS0kCeHGsOwPLVT7QkjmcMXprDULo25JEy09c0aoMPexmxFaYRgwT 1H+j5NQuEB0TPa8HzpaeNwFYLXimJFvHTU4iqBC0vZJBezCvq2OYyA7OGrKspe6Mlitt NHMPAP2QeObE7jT4I9YAZVYoVHy0b98eStU2l3AzPDNXwnLsntn1POIugiR22ZPITh9P OfgftF37e8u455uk2fDYLOCxoEPflklw+z6T3XypGYXSVmqzGVTNJ7B3BbptAEtX90KA rt9w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :references:cc:to:content-language:subject:user-agent:mime-version :date:message-id; bh=HZFKgQVLNFmJ0Qa+9NBzc7TeL98znok498nULDXd8wI=; b=JrjShOrGInhKhWMgLsO0KSIEOFQ548Zzn2By5SlOHXKIOEpxTXGvXLvTqTDCZePj/T 5LRLdob0hIhmF+a+8mV7Homlh9Tk9kDzMJV4w0XXTW4tihvlgj/v2GWe8ejFoW+lc7vx VgrBwQ66mfcK8jlq7+nRwT7IeGCiraKAZAyoibwqI4PYlIFAzZk2DBn8jKF0Y0viQeLJ rEUBHsto08caSAEwieMKA7Om0sJ610N0DS/K1QlhxcSSj3teqKqcEjMtg5TBIV8YO06X MeZr6ocXvIT1Yt5MYmmpWsDQ7TB87z4iGcosb+jAW1WJ7BctF9kKPy2aI8KESbp1Fdfv nuWw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id w4-20020a656944000000b00514477c4416si8966619pgq.801.2023.05.12.04.16.43; Fri, 12 May 2023 04:16:57 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S240926AbjELK5n (ORCPT + 99 others); Fri, 12 May 2023 06:57:43 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:55250 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S240219AbjELK5g (ORCPT ); Fri, 12 May 2023 06:57:36 -0400 Received: from www262.sakura.ne.jp (www262.sakura.ne.jp [202.181.97.72]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D3B163592 for ; Fri, 12 May 2023 03:57:33 -0700 (PDT) Received: from fsav415.sakura.ne.jp (fsav415.sakura.ne.jp [133.242.250.114]) by www262.sakura.ne.jp (8.15.2/8.15.2) with ESMTP id 34CAv7ww070055; Fri, 12 May 2023 19:57:07 +0900 (JST) (envelope-from penguin-kernel@I-love.SAKURA.ne.jp) Received: from www262.sakura.ne.jp (202.181.97.72) by fsav415.sakura.ne.jp (F-Secure/fsigk_smtp/550/fsav415.sakura.ne.jp); Fri, 12 May 2023 19:57:07 +0900 (JST) X-Virus-Status: clean(F-Secure/fsigk_smtp/550/fsav415.sakura.ne.jp) Received: from [192.168.1.6] (M106072142033.v4.enabler.ne.jp [106.72.142.33]) (authenticated bits=0) by www262.sakura.ne.jp (8.15.2/8.15.2) with ESMTPSA id 34CAv6iv070050 (version=TLSv1.2 cipher=AES256-GCM-SHA384 bits=256 verify=NO); Fri, 12 May 2023 19:57:06 +0900 (JST) (envelope-from penguin-kernel@I-love.SAKURA.ne.jp) Message-ID: Date: Fri, 12 May 2023 19:57:07 +0900 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Thunderbird/102.10.1 Subject: Re: [PATCH] debugobject: don't wake up kswapd from fill_pool() Content-Language: en-US To: Andrew Morton Cc: syzbot , syzkaller-bugs@googlegroups.com, Ingo Molnar , Thomas Gleixner , linux-kernel@vger.kernel.org, linux-mm References: <000000000000008ddb05fb5e2576@google.com> <6577e1fa-b6ee-f2be-2414-a2b51b1c5e30@I-love.SAKURA.ne.jp> <20230511204458.819f9009d2ef8b46cc163191@linux-foundation.org> From: Tetsuo Handa In-Reply-To: <20230511204458.819f9009d2ef8b46cc163191@linux-foundation.org> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 7bit X-Spam-Status: No, score=-3.8 required=5.0 tests=BAYES_00,NICE_REPLY_A, SPF_HELO_NONE,SPF_NONE,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2023/05/12 12:44, Andrew Morton wrote: > On Thu, 11 May 2023 22:47:32 +0900 Tetsuo Handa wrote: > >> syzbot is reporting lockdep warning in fill_pool(), for GFP_ATOMIC is >> (__GFP_HIGH | __GFP_KSWAPD_RECLAIM) which wakes up kswapd. >> Since fill_pool() might be called with arbitrary locks held, >> fill_pool() should not assume that holding pgdat->kswapd_wait is safe. > > hm. But many GFP_ATOMIC allocation attempts are made with locks held. > Why aren't all such callers buggy, by trying to wake kswapd with locks > held? What's special about this one? Because debugobject cannot know what locks are held when fill_pool() does GFP_ATOMIC allocation. syzbot is reporting base->lock => pgdat->kswapd_wait dependency add_timer() { __mod_timer() { base = lock_timer_base(timer, &flags); new_base = get_target_base(base, timer->flags); if (base != new_base) { raw_spin_unlock(&base->lock); base = new_base; raw_spin_lock(&base->lock); } debug_timer_activate(timer) { debug_object_activate(timer, &timer_debug_descr) { debug_objects_fill_pool() { fill_pool() { kmem_cache_zalloc(GFP_ATOMIC | __GFP_NORETRY | __GFP_NOWARN) { // wakes kswapd } } } } } raw_spin_unlock_irqrestore(&base->lock, flags); } } when pgdat->kswapd_wait => p->pi_lock dependency __alloc_pages() { get_page_from_freelist() { rmqueue() { wakeup_kswapd() { wake_up_interruptible(&pgdat->kswapd_wait) { __wake_up_common_lock() { spin_lock_irqsave(&pgdat->kswapd_wait.lock, flags); __wake_up_common() { autoremove_wake_function() { try_to_wake_up() { raw_spin_lock_irqsave(&p->pi_lock, flags); raw_spin_unlock_irqrestore(&p->pi_lock, flags); } } } spin_unlock_irqrestore(&pgdat->kswapd_wait.lock, flags); } } } } } } and p->pi_lock => rq->__lock => base->lock dependency wake_up_new_task() { raw_spin_lock_irqsave(&p->pi_lock, rf.flags); rq = __task_rq_lock(p, &rf); // acquires rq->lock activate_task(rq, p, ENQUEUE_NOCLOCK) { enqueue_task() { psi_enqueue() { psi_task_change() { queue_delayed_work_on() { __queue_delayed_work() { add_timer() { __mod_timer() { base = lock_timer_base(timer, &flags); // acquires base->lock debug_timer_activate(timer); // possible base->lock => pgdat->kswapd_wait => p->pi_lock dependency raw_spin_unlock_irqrestore(&base->lock, flags); } } } } } } } } task_rq_unlock(rq, p, &rf); } exists. All GFP_ATOMIC allocation users are supposed to be aware of what locks are held, and are supposed to explicitly remove __GFP_KSWAPD_RECLAIM if waking up kswapd can cause deadlock. But reality is that we can't be careful enough to error-free. Who would imagine GFP_ATOMIC allocation while base->lock is held can form circular locking dependency? > >> Also, __GFP_NORETRY is pointless for !__GFP_DIRECT_RECLAIM allocation __GFP_NORETRY is not checked by !__GFP_DIRECT_RECLAIM allocation. GFP_ATOMIC - __GFP_KSWAPD_RECLAIM is __GFP_HIGH. >> >> @@ -126,7 +126,7 @@ static const char *obj_states[ODEBUG_STATE_MAX] = { >> >> static void fill_pool(void) >> { >> - gfp_t gfp = GFP_ATOMIC | __GFP_NORETRY | __GFP_NOWARN; >> + gfp_t gfp = __GFP_HIGH | __GFP_NOWARN; > > Does this weaken fill_pool()'s allocation attempt more than necessary? > We can still pass __GFP_HIGH? What do you mean? I think that killing base->lock => pgdat->kswapd_wait by removing __GFP_KSWAPD_RECLAIM is the right fix. This weakening is needed for avoiding base->lock => pgdat->kswapd_wait dependency from debugobject code. For locking dependency safety, I wish that GFP_ATOMIC / GFP_NOWAIT do not imply __GFP_KSWAPD_RECLAIM. Such allocations should not try to allocate as many pages as even __GFP_HIGH fails. And if such allocations try to allocate as many pages as even __GFP_HIGH fails, they likely already failed before background kswapd reclaim finds some reusable pages....