Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp7808256imu; Thu, 15 Nov 2018 01:43:48 -0800 (PST) X-Google-Smtp-Source: AJdET5cexN7RW/at6VUAtmaO/gTW6B5RfNv4AYmsj1t89ErKiTtWw170zFS3MUWLLxiZuafR696q X-Received: by 2002:a62:4c6:: with SMTP id 189-v6mr5889249pfe.110.1542275028199; Thu, 15 Nov 2018 01:43:48 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1542275028; cv=none; d=google.com; s=arc-20160816; b=WS8UfRUicP5PvbXeoPZ7Z5GrRbgv9YJz8qzCUPddT6cqmH8hibTFcYlYZTwAf0Pg4c bqlO2D5wykFRtcPYajALuzZidOIJoYpZCx+6zpzYi8I8mFejueAEMsx72ujylEdZyse9 ExvWgKHOAF14Gjd6gu9cxlxxgQFXRM0SuQh/qLlhDQYxt78km3N43NAnAZ/zCxLdyYav sx2Q8A7QcPyqUsN0YrYIDALQV8Lb++0lNCHciQixXOJ4JQuaKe5+HB8PJ80eUaA0d8TC eZzPp+4RCqC8c/r1BW6p1+nh2PCjhGwCO0+661aseaDcchW26iFzVVLaEa7ELMMSGGCL nyGQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:organization:autocrypt:openpgp:from:references:cc:to :subject; bh=QloI1g5MmCsPPz0RNJZ/UA2M4bH1toPXcjMDgZulfg4=; b=on+lrwNoIVIzOnzVbjzxDo2wKOTQYEmM0lIt047Ubf6ZPBVTLJ192xIds0HijIEvtV z6z3udOEOrKX850VCuZZrh0LiW0wPSfRFzDKiZYZYbqUc8l1Iw0A2EsZzy6yEqYau5G9 raitLFf36F/uCqThGSTgS/YolYWPi4nWc3Kv7fi/48nH95BUUGI52nZkxXfDWQSRSiFJ 02pKo9LixFSLFwBdS6Q86yQV4U/al10oyyZVsIIIc5M5C4yRQK59GOEbaOAxKuGWBJhZ tePfO6BUtLbBnP3rqgoPCuCGjQ0JIpAKieeNlt3KH2sQCZouyH/B9soCRQtN4nxmSLuw mufg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id b15si6700972plm.431.2018.11.15.01.43.34; Thu, 15 Nov 2018 01:43:48 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2387943AbeKOTty (ORCPT + 99 others); Thu, 15 Nov 2018 14:49:54 -0500 Received: from mx1.redhat.com ([209.132.183.28]:46656 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2387911AbeKOTtx (ORCPT ); Thu, 15 Nov 2018 14:49:53 -0500 Received: from smtp.corp.redhat.com (int-mx05.intmail.prod.int.phx2.redhat.com [10.5.11.15]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 2D71186673; Thu, 15 Nov 2018 09:42:49 +0000 (UTC) Received: from [10.36.117.224] (ovpn-117-224.ams2.redhat.com [10.36.117.224]) by smtp.corp.redhat.com (Postfix) with ESMTP id B5E3A5D759; Thu, 15 Nov 2018 09:42:47 +0000 (UTC) Subject: Re: Memory hotplug softlock issue To: Michal Hocko , Baoquan He Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, akpm@linux-foundation.org, aarcange@redhat.com References: <20181114070909.GB2653@MiWiFi-R3L-srv> <5a6c6d6b-ebcd-8bfa-d6e0-4312bfe86586@redhat.com> <20181114090134.GG23419@dhcp22.suse.cz> <20181114145250.GE2653@MiWiFi-R3L-srv> <20181114150029.GY23419@dhcp22.suse.cz> <20181115051034.GK2653@MiWiFi-R3L-srv> <20181115073052.GA23831@dhcp22.suse.cz> <20181115075349.GL2653@MiWiFi-R3L-srv> <20181115083055.GD23831@dhcp22.suse.cz> From: David Hildenbrand Openpgp: preference=signencrypt Autocrypt: addr=david@redhat.com; prefer-encrypt=mutual; keydata= xsFNBFXLn5EBEAC+zYvAFJxCBY9Tr1xZgcESmxVNI/0ffzE/ZQOiHJl6mGkmA1R7/uUpiCjJ dBrn+lhhOYjjNefFQou6478faXE6o2AhmebqT4KiQoUQFV4R7y1KMEKoSyy8hQaK1umALTdL QZLQMzNE74ap+GDK0wnacPQFpcG1AE9RMq3aeErY5tujekBS32jfC/7AnH7I0v1v1TbbK3Gp XNeiN4QroO+5qaSr0ID2sz5jtBLRb15RMre27E1ImpaIv2Jw8NJgW0k/D1RyKCwaTsgRdwuK Kx/Y91XuSBdz0uOyU/S8kM1+ag0wvsGlpBVxRR/xw/E8M7TEwuCZQArqqTCmkG6HGcXFT0V9 PXFNNgV5jXMQRwU0O/ztJIQqsE5LsUomE//bLwzj9IVsaQpKDqW6TAPjcdBDPLHvriq7kGjt WhVhdl0qEYB8lkBEU7V2Yb+SYhmhpDrti9Fq1EsmhiHSkxJcGREoMK/63r9WLZYI3+4W2rAc UucZa4OT27U5ZISjNg3Ev0rxU5UH2/pT4wJCfxwocmqaRr6UYmrtZmND89X0KigoFD/XSeVv jwBRNjPAubK9/k5NoRrYqztM9W6sJqrH8+UWZ1Idd/DdmogJh0gNC0+N42Za9yBRURfIdKSb B3JfpUqcWwE7vUaYrHG1nw54pLUoPG6sAA7Mehl3nd4pZUALHwARAQABzSREYXZpZCBIaWxk ZW5icmFuZCA8ZGF2aWRAcmVkaGF0LmNvbT7CwX4EEwECACgFAljj9eoCGwMFCQlmAYAGCwkI BwMCBhUIAgkKCwQWAgMBAh4BAheAAAoJEE3eEPcA/4Na5IIP/3T/FIQMxIfNzZshIq687qgG 8UbspuE/YSUDdv7r5szYTK6KPTlqN8NAcSfheywbuYD9A4ZeSBWD3/NAVUdrCaRP2IvFyELj xoMvfJccbq45BxzgEspg/bVahNbyuBpLBVjVWwRtFCUEXkyazksSv8pdTMAs9IucChvFmmq3 jJ2vlaz9lYt/lxN246fIVceckPMiUveimngvXZw21VOAhfQ+/sofXF8JCFv2mFcBDoa7eYob s0FLpmqFaeNRHAlzMWgSsP80qx5nWWEvRLdKWi533N2vC/EyunN3HcBwVrXH4hxRBMco3jvM m8VKLKao9wKj82qSivUnkPIwsAGNPdFoPbgghCQiBjBe6A75Z2xHFrzo7t1jg7nQfIyNC7ez MZBJ59sqA9EDMEJPlLNIeJmqslXPjmMFnE7Mby/+335WJYDulsRybN+W5rLT5aMvhC6x6POK z55fMNKrMASCzBJum2Fwjf/VnuGRYkhKCqqZ8gJ3OvmR50tInDV2jZ1DQgc3i550T5JDpToh dPBxZocIhzg+MBSRDXcJmHOx/7nQm3iQ6iLuwmXsRC6f5FbFefk9EjuTKcLMvBsEx+2DEx0E UnmJ4hVg7u1PQ+2Oy+Lh/opK/BDiqlQ8Pz2jiXv5xkECvr/3Sv59hlOCZMOaiLTTjtOIU7Tq 7ut6OL64oAq+zsFNBFXLn5EBEADn1959INH2cwYJv0tsxf5MUCghCj/CA/lc/LMthqQ773ga uB9mN+F1rE9cyyXb6jyOGn+GUjMbnq1o121Vm0+neKHUCBtHyseBfDXHA6m4B3mUTWo13nid 0e4AM71r0DS8+KYh6zvweLX/LL5kQS9GQeT+QNroXcC1NzWbitts6TZ+IrPOwT1hfB4WNC+X 2n4AzDqp3+ILiVST2DT4VBc11Gz6jijpC/KI5Al8ZDhRwG47LUiuQmt3yqrmN63V9wzaPhC+ xbwIsNZlLUvuRnmBPkTJwwrFRZvwu5GPHNndBjVpAfaSTOfppyKBTccu2AXJXWAE1Xjh6GOC 8mlFjZwLxWFqdPHR1n2aPVgoiTLk34LR/bXO+e0GpzFXT7enwyvFFFyAS0Nk1q/7EChPcbRb hJqEBpRNZemxmg55zC3GLvgLKd5A09MOM2BrMea+l0FUR+PuTenh2YmnmLRTro6eZ/qYwWkC u8FFIw4pT0OUDMyLgi+GI1aMpVogTZJ70FgV0pUAlpmrzk/bLbRkF3TwgucpyPtcpmQtTkWS gDS50QG9DR/1As3LLLcNkwJBZzBG6PWbvcOyrwMQUF1nl4SSPV0LLH63+BrrHasfJzxKXzqg rW28CTAE2x8qi7e/6M/+XXhrsMYG+uaViM7n2je3qKe7ofum3s4vq7oFCPsOgwARAQABwsFl BBgBAgAPBQJVy5+RAhsMBQkJZgGAAAoJEE3eEPcA/4NagOsP/jPoIBb/iXVbM+fmSHOjEshl KMwEl/m5iLj3iHnHPVLBUWrXPdS7iQijJA/VLxjnFknhaS60hkUNWexDMxVVP/6lbOrs4bDZ NEWDMktAeqJaFtxackPszlcpRVkAs6Msn9tu8hlvB517pyUgvuD7ZS9gGOMmYwFQDyytpepo YApVV00P0u3AaE0Cj/o71STqGJKZxcVhPaZ+LR+UCBZOyKfEyq+ZN311VpOJZ1IvTExf+S/5 lqnciDtbO3I4Wq0ArLX1gs1q1XlXLaVaA3yVqeC8E7kOchDNinD3hJS4OX0e1gdsx/e6COvy qNg5aL5n0Kl4fcVqM0LdIhsubVs4eiNCa5XMSYpXmVi3HAuFyg9dN+x8thSwI836FoMASwOl C7tHsTjnSGufB+D7F7ZBT61BffNBBIm1KdMxcxqLUVXpBQHHlGkbwI+3Ye+nE6HmZH7IwLwV W+Ajl7oYF+jeKaH4DZFtgLYGLtZ1LDwKPjX7VAsa4Yx7S5+EBAaZGxK510MjIx6SGrZWBrrV TEvdV00F2MnQoeXKzD7O4WFbL55hhyGgfWTHwZ457iN9SgYi1JLPqWkZB0JRXIEtjd4JEQcx +8Umfre0Xt4713VxMygW0PnQt5aSQdMD58jHFxTk092mU+yIHj5LeYgvwSgZN4airXk5yRXl SE+xAvmumFBY Organization: Red Hat GmbH Message-ID: Date: Thu, 15 Nov 2018 10:42:46 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.3.0 MIME-Version: 1.0 In-Reply-To: <20181115083055.GD23831@dhcp22.suse.cz> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 2.79 on 10.5.11.15 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.26]); Thu, 15 Nov 2018 09:42:49 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 15.11.18 09:30, Michal Hocko wrote: > On Thu 15-11-18 15:53:56, Baoquan He wrote: >> On 11/15/18 at 08:30am, Michal Hocko wrote: >>> On Thu 15-11-18 13:10:34, Baoquan He wrote: >>>> On 11/14/18 at 04:00pm, Michal Hocko wrote: >>>>> On Wed 14-11-18 22:52:50, Baoquan He wrote: >>>>>> On 11/14/18 at 10:01am, Michal Hocko wrote: >>>>>>> I have seen an issue when the migration cannot make a forward progress >>>>>>> because of a glibc page with a reference count bumping up and down. Most >>>>>>> probable explanation is the faultaround code. I am working on this and >>>>>>> will post a patch soon. In any case the migration should converge and if >>>>>>> it doesn't do then there is a bug lurking somewhere. >>>>>>> >>>>>>> Failing on ENOMEM is a questionable thing. I haven't seen that happening >>>>>>> wildly but if it is a case then I wouldn't be opposed. >>>>>> >>>>>> Applied your debugging patches, it helps a lot to printing message. >>>>>> >>>>>> Below is the dmesg log about the migrating failure. It can't pass >>>>>> migrate_pages() and loop forever. >>>>>> >>>>>> [ +0.083841] migrating pfn 10fff7d0 failed >>>>>> [ +0.000005] page:ffffea043ffdf400 count:208 mapcount:201 mapping:ffff888dff4bdda8 index:0x2 >>>>>> [ +0.012689] xfs_address_space_operations [xfs] >>>>>> [ +0.000030] name:"stress" >>>>>> [ +0.004556] flags: 0x5fffffc0000004(uptodate) >>>>>> [ +0.007339] raw: 005fffffc0000004 ffffc900000e3d80 ffffc900000e3d80 ffff888dff4bdda8 >>>>>> [ +0.009488] raw: 0000000000000002 0000000000000000 000000cb000000c8 ffff888e7353d000 >>>>>> [ +0.007726] page->mem_cgroup:ffff888e7353d000 >>>>>> [ +0.084538] migrating pfn 10fff7d0 failed >>>>>> [ +0.000006] page:ffffea043ffdf400 count:210 mapcount:201 mapping:ffff888dff4bdda8 index:0x2 >>>>>> [ +0.012798] xfs_address_space_operations [xfs] >>>>>> [ +0.000034] name:"stress" >>>>>> [ +0.004524] flags: 0x5fffffc0000004(uptodate) >>>>>> [ +0.007068] raw: 005fffffc0000004 ffffc900000e3d80 ffffc900000e3d80 ffff888dff4bdda8 >>>>>> [ +0.009359] raw: 0000000000000002 0000000000000000 000000cb000000c8 ffff888e7353d000 >>>>>> [ +0.007728] page->mem_cgroup:ffff888e7353d000 >>>>> >>>>> I wouldn't be surprised if this was a similar/same issue I've been >>>>> chasing recently. Could you try to disable faultaround to see if that >>>>> helps. It seems that it helped in my particular case but I am still >>>>> waiting for the final good-to-go to post the patch as I do not own the >>>>> workload which triggered that issue. >>>> >>>> Tried, still stuck in last block sometime. Usually after several times >>>> of hotplug/unplug. If stop stress program, the last block will be >>>> offlined immediately. >>> >>> Is the pattern still the same? I mean failing over few pages with >>> reference count jumping up and down between attempts? >> >> ->count jumping up and down, mapcount stays the same value. >> >>> >>>> [root@ ~]# cat /sys/kernel/debug/fault_around_bytes >>>> 4096 >>> >>> Can you make it 0? >> >> I executed 'echo 0 > fault_around_bytes', value less than one page size >> will round up to one page. > > OK, I have missed that. So then there must be a different source of the > page count volatility. Is it always the same file? > > I think we can rule out memory reclaim because that depends on the page > lock. Is the stress test hitting on memory compaction? In other words, > are > grep compact /proc/vmstat > counters changing during the offline test heavily? I am asking because I > do not see compaction pfn walkers skipping over MIGRATE_ISOLATE > pageblocks. But I might be missing something easily. > > It would be also good to find out whether this is fs specific. E.g. does > it make any difference if you use a different one for your stress > testing? > I am wondering why it is always the last memory block of that device (and even that node). Coincidence? -- Thanks, David / dhildenb