Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp643269pxk; Wed, 2 Sep 2020 10:54:10 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwmpdbpp/mCmvdvwyu5dgid/bjJddeGelbpUO/0omv/PhT5lkH6ttuguTmnc7AFJhHPoMpD X-Received: by 2002:a17:906:914b:: with SMTP id y11mr1204403ejw.145.1599069250151; Wed, 02 Sep 2020 10:54:10 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1599069250; cv=none; d=google.com; s=arc-20160816; b=w2iIHg4AggNpBJcBCujI8jYIAqPsCd4T53T9csVe269DoLYdvrzzcNZwlr3Ahya9QG nefoyR+Rpwl5HpE5ozblKBrvK8Y8hd0ewWUyyKhjZ29bK1NnUq9Gm+gr/XVs9coRqatB YmFjTbzojHPjohXuuRtjsAb6MVSd7KORGXbCewJ392vbXeB26JTpgRTjN5U3Z3BHQ2lN oMZqxcBdR9qIiT9Qj8v1kT/2FvNpeK9AxQnsgeWs+335jz3IUlufmpH6KFS75AdtfvYI y5/akZXRsNTOBB3PIrs86BVHcLvQoiCYt7xgiU9iEUoMLZdE5sU84xcUOiBp9XATKRVj LI6Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=y6kG/b6ekeiSr/StNdGkASFtMIPC2oZFHfvfSHsNR2M=; b=nUPp4ZAioU2dWUky0DZpr1Kx/qagOnJHQnuuMkfYgBDFNApvk+477G5ah6PE1JdqQ5 pQWn0OV/OYgBHnRf9uti5ET7kv2dhasu6fKnuiNmmxJ/MLThHPKy55scEW8CADWOBDv/ xBHYfwRJ2/k082/Pf3uUL4b57qJ16T3lmKBIVPUENQrPu/zbeCrh0KWsFuuqAZNc3MQh 97TUHDIwKcrRNKrZcf0pHCqW5YawreyWmeZqiaJ9Qlb+sp88t1p9P+1sZMZ7ZwXVJ1II 48YR95NfKiZ+TmD9QiAAOf4AQb21swCWuZsCEv9Zm17YHVhoFLisCrYTVSqKx4cp+JZW GiZw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id h25si187845eje.30.2020.09.02.10.53.46; Wed, 02 Sep 2020 10:54:10 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727823AbgIBRvs (ORCPT + 99 others); Wed, 2 Sep 2020 13:51:48 -0400 Received: from mx2.suse.de ([195.135.220.15]:46328 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726247AbgIBRvr (ORCPT ); Wed, 2 Sep 2020 13:51:47 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.221.27]) by mx2.suse.de (Postfix) with ESMTP id DD109AD5C; Wed, 2 Sep 2020 17:51:46 +0000 (UTC) Subject: Re: [PATCH] mm/memory_hotplug: drain per-cpu pages again during memory offline To: Michal Hocko Cc: Pavel Tatashin , LKML , Andrew Morton , linux-mm References: <20200901124615.137200-1-pasha.tatashin@soleen.com> <20200902140851.GJ4617@dhcp22.suse.cz> <74f2341a-7834-3e37-0346-7fbc48d74df3@suse.cz> <20200902151306.GL4617@dhcp22.suse.cz> From: Vlastimil Babka Message-ID: Date: Wed, 2 Sep 2020 19:51:45 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.11.0 MIME-Version: 1.0 In-Reply-To: <20200902151306.GL4617@dhcp22.suse.cz> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 9/2/20 5:13 PM, Michal Hocko wrote: > On Wed 02-09-20 16:55:05, Vlastimil Babka wrote: >> On 9/2/20 4:26 PM, Pavel Tatashin wrote: >> > On Wed, Sep 2, 2020 at 10:08 AM Michal Hocko wrote: >> >> >> >> > >> >> > Thread#1 - continue >> >> > free_unref_page_commit >> >> > migratetype = get_pcppage_migratetype(page); >> >> > // get old migration type >> >> > list_add(&page->lru, &pcp->lists[migratetype]); >> >> > // add new page to already drained pcp list >> >> > >> >> > Thread#2 >> >> > Never drains pcp again, and therefore gets stuck in the loop. >> >> > >> >> > The fix is to try to drain per-cpu lists again after >> >> > check_pages_isolated_cb() fails. >> >> >> >> But this means that the page is not isolated and so it could be reused >> >> for something else. No? >> > >> > The page is in a movable zone, has zero references, and the section is >> > isolated (i.e. set_pageblock_migratetype(page, MIGRATE_ISOLATE);) is >> > set. The page should be offlinable, but it is lost in a pcp list as >> > that list is never drained again after the first failure to migrate >> > all pages in the range. >> >> Yeah. To answer Michal's "it could be reused for something else" - yes, somebody >> could allocate it from the pcplist before we do the extra drain. But then it >> becomes "visible again" and the loop in __offline_pages() should catch it by >> scan_movable_pages() - do_migrate_range(). And this time the pageblock is >> already marked as isolated, so the page (freed by migration) won't end up on the >> pcplist again. > > So the page block is marked MIGRATE_ISOLATE but the allocation itself > could be used for non migrateable objects. Or does anything prevent that > from happening? In a movable zone, the allocation should not be used for non migrateable objects. E.g. if the zone was not ZONE_MOVABLE, the offlining could fail regardless of this race (analogically for migrating away from CMA pageblocks). > We really do depend on isolation to not allow reuse when offlining. This is not really different than if the page on pcplist was allocated just a moment before the offlining, thus isolation started. We ultimately rely on being able to migrate any allocated pages away during the isolation. This "freeing to pcplists" race doesn't fundamentally change anything in this regard. We just have to guarantee that pages on pcplists will be eventually flushed, to make forward progress, and there was a bug in this aspect.