Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1760242AbdDSGUf (ORCPT ); Wed, 19 Apr 2017 02:20:35 -0400 Received: from mx0a-001b2d01.pphosted.com ([148.163.156.1]:55145 "EHLO mx0a-001b2d01.pphosted.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1759732AbdDSGUc (ORCPT ); Wed, 19 Apr 2017 02:20:32 -0400 From: "Aneesh Kumar K.V" To: Anshuman Khandual , linux-kernel@vger.kernel.org, linux-mm@kvack.org Cc: n-horiguchi@ah.jp.nec.com, akpm@linux-foundation.org Subject: Re: [RFC] mm/madvise: Enable (soft|hard) offline of HugeTLB pages at PGD level In-Reply-To: <20170419032759.29700-1-khandual@linux.vnet.ibm.com> References: <20170419032759.29700-1-khandual@linux.vnet.ibm.com> Date: Wed, 19 Apr 2017 11:50:24 +0530 MIME-Version: 1.0 Content-Type: text/plain X-TM-AS-MML: disable x-cbid: 17041906-0012-0000-0000-000003E4F6E7 X-IBM-AV-DETECTION: SAVI=unused REMOTE=unused XFE=unused x-cbparentid: 17041906-0013-0000-0000-00001B6BAF71 Message-Id: <877f2ghqaf.fsf@skywalker.in.ibm.com> X-Proofpoint-Virus-Version: vendor=fsecure engine=2.50.10432:,, definitions=2017-04-19_04:,, signatures=0 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 spamscore=0 suspectscore=1 malwarescore=0 phishscore=0 adultscore=0 bulkscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.0.1-1703280000 definitions=main-1704190058 Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 962 Lines: 24 Anshuman Khandual writes: > Though migrating gigantic HugeTLB pages does not sound much like real > world use case, they can be affected by memory errors. Hence migration > at the PGD level HugeTLB pages should be supported just to enable soft > and hard offline use cases. In that case do we want to isolated the entire 16GB range ? Should we just dequeue the page from hugepage pool convert them to regular 64K pages and then isolate the 64K that had memory error ? > > While allocating the new gigantic HugeTLB page, it should not matter > whether new page comes from the same node or not. There would be very > few gigantic pages on the system afterall, we should not be bothered > about node locality when trying to save a big page from crashing. > > This introduces a new HugeTLB allocator called alloc_gigantic_page() > which will scan over all online nodes on the system and allocate a > single HugeTLB page. > -aneesh