Received: by 2002:a25:ca44:0:0:0:0:0 with SMTP id a65csp1207429ybg; Wed, 29 Jul 2020 08:21:09 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwjZxvf45YyjoEX5c07lmKiQt9Kx0kgBXnAIl1KRh2kTJbYreT1rL6Wo2RSBvDJpPwk+0/v X-Received: by 2002:a17:906:1a54:: with SMTP id j20mr7620128ejf.102.1596036069601; Wed, 29 Jul 2020 08:21:09 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1596036069; cv=none; d=google.com; s=arc-20160816; b=kM7bNZZ1oImrFU2kzLXhuy+GWoT0Rjq15+N/+d5Wrc9w5QjJIT9RrGa5FtPI7yy9DK vBlzgK1FZuSbUJ2grGnzuhYvvyR4KoUwSa2oD5r/T0oQ9Cj3YJT37fXCbNCR0Uf0BssK H68I2MaT+s+/I4nWcfhvRRKHzQcsXqyR19PcKhNM1Fzno+qP8hcHwJ6w3r/EYl4u46cE SBwIiqnwgfgXUYGuqjqamZmMs0Th4m9hedvNsss+rxCxrleEcEttk4C3b4DrLyBuim16 BAmqtv7rkjmzinoP0ETsfh3ADNxPaEiobJqSfYSuy2WnvYMynnVbM2F0x+JLQ8vKgz5h Ms4A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=VVFRx5RFHzlv+1Sn1fr6JII3C4Kn4sq7np/Pe4wqaBo=; b=ceco2AOw4VkVf5qHJnTovkVW70Bin6DVDeqUkSRcvSKmSA63BZ0bHgs6XoRahM6h/C JYCE5fI4TQePTkfUidnLdbbTsXeSTnpVgTXLdMyUi7VsiwRlj49bX4LHry2Kj6OQC/vC BWvd+WvsTaqmi9/5pkK42e0xC8Fu9WjiM1oF+rvS2cGpluXv+lWMRpkrgCDRi8LpD8Kp bQX+L1FHK0pGZo471nu/BqFwN/JypFWid96qF21b42J80riB+Sjm9sfpqiA58MtErXwZ CrQ+VVjBxTqnxKUuCr1mfK7XHgGQf8fgeXessEhNc0jdQnJ7uju87uWzFx2qSbT5a0v8 QlNA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id ck3si548612edb.324.2020.07.29.08.20.46; Wed, 29 Jul 2020 08:21:09 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726509AbgG2PUg (ORCPT + 99 others); Wed, 29 Jul 2020 11:20:36 -0400 Received: from mail.kernel.org ([198.145.29.99]:51340 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726054AbgG2PUg (ORCPT ); Wed, 29 Jul 2020 11:20:36 -0400 Received: from gaia (unknown [95.146.230.158]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 6E2E920809; Wed, 29 Jul 2020 15:20:31 +0000 (UTC) Date: Wed, 29 Jul 2020 16:20:29 +0100 From: Catalin Marinas To: chenzhou Cc: tglx@linutronix.de, mingo@redhat.com, dyoung@redhat.com, bhe@redhat.com, will@kernel.org, james.morse@arm.com, robh+dt@kernel.org, arnd@arndb.de, John.P.donnelly@oracle.com, prabhakar.pkin@gmail.com, nsaenzjulienne@suse.de, corbet@lwn.net, bhsharma@redhat.com, horms@verge.net.au, guohanjun@huawei.com, xiexiuqi@huawei.com, huawei.libin@huawei.com, linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org, kexec@lists.infradead.org, linux-doc@vger.kernel.org Subject: Re: [PATCH v10 4/5] arm64: kdump: fix kdump broken with ZONE_DMA reintroduced Message-ID: <20200729152028.GE5524@gaia> References: <20200703035816.31289-1-chenzhou10@huawei.com> <20200703035816.31289-5-chenzhou10@huawei.com> <20200727173014.GL13938@gaia> <20200729115851.GC5524@gaia> <217004f5-dd8e-d04c-038b-c88b132d5495@huawei.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <217004f5-dd8e-d04c-038b-c88b132d5495@huawei.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Jul 29, 2020 at 10:14:32PM +0800, chenzhou wrote: > On 2020/7/29 19:58, Catalin Marinas wrote: > > On Wed, Jul 29, 2020 at 11:52:39AM +0800, chenzhou wrote: > >> How about like this: > >> 1. For ZONE_DMA issue, use Bhupesh's solution, keep the crashkernel= > >> behaviour to ZONE_DMA allocations. > >> 2. For this patch series, make the reserve_crashkernel_low() to > >> ZONE_DMA allocations. > > > > So you mean rebasing your series on top of Bhupesh's? I guess you can > > combine the two, I really don't care which way as long as we fix both > > issues and agree on the crashkernel= semantics. I think with some tweaks > > we can go with your series alone. > > > > IIUC from the x86 code (especially the part you #ifdef'ed out for > > arm64), if ",low" is not passed (so just standard crashkernel=X), it > > still allocates sufficient low memory for the swiotlb in ZONE_DMA. The > > rest can go in a high region. Why can't we do something similar on > > arm64? Of course, you can keep the ",low" argument for explicit > > allocation but I don't want to mandate it. > > It is a good idea to combine the two. > > For parameter crashkernel=X, we do like this: > 1. allocate some low memory in ZONE_DMA(or ZONE_DMA32 if CONFIG_ZONE_DMA=n) > 2. allocate X size memory in a high region > > ",low" argument can be used to specify the low memory. > > Do i understand correctly? Yes, although we could follow the x86 approach: 1. Try low (ZONE_DMA for arm64) allocation, fallback to high allocation if it fails. 2. If crash_base is outside ZONE_DMA, call reserve_crashkernel_low() which either honours the ,low option or allocates some small amount in ZONE_DMA. If at some point we have platforms failing step 2, we'll look at changing ZONE_DMA to the full 4GB on non-RPi4 platforms. It looks to me like x86 ignores the ,low option if the first step managed to get some low memory. Shall we do the same on arm64? > > So with an implicit ZONE_DMA allocation similar to the x86 one, we > > probably don't need Bhupesh's series at all. In addition, we can limit > > crashkernel= to the first 4G with a fall-back to high like x86 (not sure > > if memblock_find_in_range() is guaranteed to search in ascending order). > > I don't think we need an explicit ",high" annotation. > > > > So with the above, just a crashkernel=1G gives you at least 256MB in > > ZONE_DMA followed by the rest anywhere, with a preference for > > ZONE_DMA32. This way we can also keep the reserve_crashkernel_low() > > mostly intact from x86 (less #ifdef's). > > Yes. We can let crashkernel=X try to reserve low memory and fall back to use high memory > if failing to find a low range. The only question is whether we need to preserve some more ZONE_DMA on the current system. If for example we pass a crashkernel=512M and some cma=, we may end up with very little free memory in ZONE_DMA. That's mostly an issue for RPi4 since other platforms would work with ZONE_DMA32. We could add a threshold and go for high allocation directly if the required size is too large. > About the function reserve_crashkernel_low(), if we put it in arch/arm64, there is some common > code with x86_64. Some suggestions about this? If we can use this function almost intact, just move it in a common place. But if it gets sprinkled with #ifdef CONFIG_ARM64, I'd rather duplicate it. I'd still prefer to move it to a common place if possible. You can go a step further and also move the x86 reserve_crashkernel() to common code. I don't think there a significant difference between arm64 and x86 here. You'd have to define arch-specific specific CRASH_ADDR_LOW_MAX etc. Also patches moving code should not have any functional change. The CRASH_ALIGN change from 16M to 2M on x86 should be a separate patch as it needs to be acked by the x86 maintainers (IIRC, Ingo only acked the function move if there was no functional change; CRASH_ALIGN is used for the start address, not just alignment, on x86). -- Catalin