Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp2376881pxk; Mon, 14 Sep 2020 11:37:11 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwcbKbBsF0UFhqElQiBeVc6nONWVU9aPW9x0sNOjMCerABQGp4c8ptnc6nffE4u7ezrCGIb X-Received: by 2002:aa7:cb44:: with SMTP id w4mr18849263edt.139.1600108631369; Mon, 14 Sep 2020 11:37:11 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1600108631; cv=none; d=google.com; s=arc-20160816; b=dvstUSV44QNNwMuiH2IYF4Nh2ni9oRqLxdI5JHeoBkg+A5xCwTG9GRi2iCnpAgFFDO g0j+TbloFmEFmtSBfZccmLaoNpz5sgwdfIIhXQnp/voAbwlFGQPU1BUxqwXyhvKfD7og pdTOtBlrt2vcqocxc8qd1mm2lN0dXyS69Sjp6CApFoKBDTIckViF8zbO1AnxNdUNyZp3 qbgMCb4Nsj22/doSQ9Cq9qqA3F0UPuyPeCiURvWXBz47gQOPyNqpFi8hNh3OVjDeR2/8 Lzk6/KjHMGwJGqrlOtJTMgLUK01ucHnSfBVFAgiWweThETBN5A54mh5ueODohn+01EdJ w9Og== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:message-id:references :in-reply-to:subject:cc:to:from:date:content-transfer-encoding :mime-version:dkim-signature; bh=n5jUUJ7Y5eFQeaeymKJYtour4RQWDP8UQVgNUhiT16A=; b=raXNl3RVjojm+jVAK1JgvE0tPR47RVLVU1aqn13TxC4onu2mIrGM1FXB1GpiGxQhpY /IbDOcEDK4n1VXsPqBFIi7Otv99ySEaseYWW88du2H/uAMpji6Rnwxpy9yzhaFKE7Dbe 1dBXhzDeh5fSDjbqgMAxwoV2ROhyWAh1IkS9qZtqhQiCcHVcEvVoRB5/OSx1ICW8WSoM 9YIXaTm1lVdkNhlCeRF2X9mRaxWAtyboIU5jLqXs6FLfSMf65mSXAWEH1RN9V6tzV4pQ Eu21+N9gBEh24AH1aQiJRBml5/l2iUvdN5ErDR0kuM3jb9VMl2f0D51r3fMx8teZI6DV Z0kg== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@mg.codeaurora.org header.s=smtp header.b=ldLL3ajZ; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id bh20si7204480ejb.294.2020.09.14.11.36.48; Mon, 14 Sep 2020 11:37:11 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=fail header.i=@mg.codeaurora.org header.s=smtp header.b=ldLL3ajZ; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1725951AbgINSdk (ORCPT + 99 others); Mon, 14 Sep 2020 14:33:40 -0400 Received: from mail29.static.mailgun.info ([104.130.122.29]:47340 "EHLO mail29.static.mailgun.info" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726069AbgINSdd (ORCPT ); Mon, 14 Sep 2020 14:33:33 -0400 DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mg.codeaurora.org; q=dns/txt; s=smtp; t=1600108412; h=Message-ID: References: In-Reply-To: Subject: Cc: To: From: Date: Content-Transfer-Encoding: Content-Type: MIME-Version: Sender; bh=n5jUUJ7Y5eFQeaeymKJYtour4RQWDP8UQVgNUhiT16A=; b=ldLL3ajZ4/a1Nf+PmAMC158Mmrk8qLYVWf2QnFBRcSa6yIUSQ7E1fH0DBXaZlp5oUVXGFM0r oIXEHf8g9iQfPxnmxuVouq7RETt9Oj8N5ihb4oPOdcyKlZFuNqnUxtD8l6dtEGph8zhrYraU rqwzmc5I5HnOOxDlCcubpYmFm58= X-Mailgun-Sending-Ip: 104.130.122.29 X-Mailgun-Sid: WyI0MWYwYSIsICJsaW51eC1rZXJuZWxAdmdlci5rZXJuZWwub3JnIiwgImJlOWU0YSJd Received: from smtp.codeaurora.org (ec2-35-166-182-171.us-west-2.compute.amazonaws.com [35.166.182.171]) by smtp-out-n04.prod.us-east-1.postgun.com with SMTP id 5f5fb76673afa3417e8e9cf5 (version=TLS1.2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256); Mon, 14 Sep 2020 18:33:10 GMT Received: by smtp.codeaurora.org (Postfix, from userid 1001) id 65FF0C433FE; Mon, 14 Sep 2020 18:33:09 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-caf-mail-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.9 required=2.0 tests=ALL_TRUSTED,BAYES_00, URIBL_BLOCKED autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.codeaurora.org (localhost.localdomain [127.0.0.1]) (using TLSv1 with cipher ECDHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) (Authenticated sender: cgoldswo) by smtp.codeaurora.org (Postfix) with ESMTPSA id 3085FC433CA; Mon, 14 Sep 2020 18:33:08 +0000 (UTC) MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII; format=flowed Content-Transfer-Encoding: 7bit Date: Mon, 14 Sep 2020 11:33:08 -0700 From: Chris Goldsworthy To: David Hildenbrand Cc: akpm@linux-foundation.org, linux-mm@kvack.org, linux-arm-msm@vger.kernel.org, linux-kernel@vger.kernel.org, pratikp@codeaurora.org, pdaly@codeaurora.org, sudraja@codeaurora.org, iamjoonsoo.kim@lge.com, linux-arm-msm-owner@vger.kernel.org, Vinayak Menon , linux-kernel-owner@vger.kernel.org Subject: Re: [PATCH v2] mm: cma: indefinitely retry allocations in cma_alloc In-Reply-To: References: <06489716814387e7f147cf53d1b185a8@codeaurora.org> <1599851809-4342-1-git-send-email-cgoldswo@codeaurora.org> <010101747e998731-e49f209f-8232-4496-a9fc-2465334e70d7-000000@us-west-2.amazonses.com> Message-ID: <72ae0f361df527cf70946992e4ab1eb3@codeaurora.org> X-Sender: cgoldswo@codeaurora.org User-Agent: Roundcube Webmail/1.3.9 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2020-09-14 02:31, David Hildenbrand wrote: > On 11.09.20 21:17, Chris Goldsworthy wrote: >> >> So, inside of cma_alloc(), instead of giving up when >> alloc_contig_range() >> returns -EBUSY after having scanned a whole CMA-region bitmap, perform >> retries indefinitely, with sleeps, to give the system an opportunity >> to >> unpin any pinned pages. >> >> Signed-off-by: Chris Goldsworthy >> Co-developed-by: Vinayak Menon >> Signed-off-by: Vinayak Menon >> --- >> mm/cma.c | 25 +++++++++++++++++++++++-- >> 1 file changed, 23 insertions(+), 2 deletions(-) >> >> diff --git a/mm/cma.c b/mm/cma.c >> index 7f415d7..90bb505 100644 >> --- a/mm/cma.c >> +++ b/mm/cma.c >> @@ -442,8 +443,28 @@ struct page *cma_alloc(struct cma *cma, size_t >> count, unsigned int align, >> bitmap_maxno, start, bitmap_count, mask, >> offset); >> if (bitmap_no >= bitmap_maxno) { >> - mutex_unlock(&cma->lock); >> - break; >> + if (ret == -EBUSY) { >> + mutex_unlock(&cma->lock); >> + >> + /* >> + * Page may be momentarily pinned by some other >> + * process which has been scheduled out, e.g. >> + * in exit path, during unmap call, or process >> + * fork and so cannot be freed there. Sleep >> + * for 100ms and retry the allocation. >> + */ >> + start = 0; >> + ret = -ENOMEM; >> + msleep(100); >> + continue; >> + } else { >> + /* >> + * ret == -ENOMEM - all bits in cma->bitmap are >> + * set, so we break accordingly. >> + */ >> + mutex_unlock(&cma->lock); >> + break; >> + } >> } >> bitmap_set(cma->bitmap, bitmap_no, bitmap_count); >> /* >> > > What about long-term pinnings? IIRC, that can happen easily e.g., with > vfio (and I remember there is a way via vmsplice). > > Not convinced trying forever is a sane approach in the general case ... Hi David, I've botched the threading, so there are discussions with respect to the previous patch-set that is missing on this thread, which I will summarize below: V1: [1] https://lkml.org/lkml/2020/8/5/1097 [2] https://lkml.org/lkml/2020/8/6/1040 [3] https://lkml.org/lkml/2020/8/11/893 [4] https://lkml.org/lkml/2020/8/21/1490 [5] https://lkml.org/lkml/2020/9/11/1072 [1] features version of the patch featured a finite number of retries, which has been stable for our kernels. In [2], Andrew questioned whether we could actually find a way of solving the problem on the grounds that doing a finite number of retries doesn't actually fix the problem (more importantly, in [4] Andrew indicated that he would prefer not to merge the patch as it doesn't solve the issue). In [3], I suggest one actual fix for this, which is to use preempt_disable/enable() to prevent context switches from occurring during the periods in copy_one_pte() and exit_mmap() (I forgot to mention this case in the commit text) in which _refcount > _mapcount for a page - you would also need to prevent interrupts from occurring to if we were to fully prevent the issue from occurring. I think this would be acceptable for the copy_one_pte() case, since there _refcount > _mapcount for little time. For the exit_mmap() case, however, _refcount is greater than _mapcount whilst the page-tables are being torn down for a process - that could be too long for disabling preemption / interrupts. So, in [4], Andrew asks about two alternatives to see if they're viable: (1) acquiring locks on the exit_mmap path and migration paths, (2) retrying indefinitely. In [5], I discuss how using locks could increase the time it takes to perform a CMA allocation, such that a retry approach would avoid increased CMA allocation times. I'm also uncertain about how the locking scheme could be implemented effectively without introducing a new per-page lock that will be used specifically to solve this issue, and I'm not sure this would be accepted. We're fine with doing indefinite retries, on the grounds that if there is some long-term pinning that occurs when alloc_contig_range returns -EBUSY, that it should be debugged and fixed. Would it be possible to make this infinite-retrying something that could be enabled or disabled by a defconfig option? Thanks, Chris. -- The Qualcomm Innovation Center, Inc. The Qualcomm Innovation Center, Inc. is a member of the Code Aurora Forum, a Linux Foundation Collaborative Project