Received: by 2002:a25:23cc:0:0:0:0:0 with SMTP id j195csp958578ybj; Thu, 7 May 2020 11:24:12 -0700 (PDT) X-Google-Smtp-Source: APiQypKbnI6l2PnLo7FGmOtyoFgllXqdyuOTHzNllrk1M/tuXCC7245WQhszOqmW9PWRaElxY8TY X-Received: by 2002:aa7:cdd9:: with SMTP id h25mr13672598edw.17.1588875852688; Thu, 07 May 2020 11:24:12 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588875852; cv=none; d=google.com; s=arc-20160816; b=NDMUpb16dpaEnDXYt2slBYBMluUSK4Xv4xjWsYIaCwsEXPM+FrU/3DAozy+likM3IU Lau4T9aoXLJTSsbU1PUhIxUXPxdsunP4zYZUg0anbHKvoNEkMPo+mFQbRmdDQXJn72D1 038UhhGhDeVlP8n8pz+Hwi0LagNvE0sxaXeIiCQcQTnQZJ2CK2FRW+9uCuizdgLzkJuD TjxrELrkIriPUZsXzFet6p7fy2dGLPbrt4zEDmdgIHXAnilZKaPZFfDkQTGhJAiaLI1D 67ZwdXeUWgJ/2IfSMPOcLgkE1gihlS3j9A4nnZ1FGKnHK5MlWvzoEbntHcpPBQjAwtkL r9Ig== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :references:in-reply-to:mime-version:dkim-signature; bh=Hio9A8K2n+RsdYoc8pl4frI4S8NgtDDOmOIDiCLHFEg=; b=W5elGti8O5zY2917bVPsOesUKAxRWT5lTncnwnyRJk1H4eYIMPd+ZwVlrtAfeBt7v1 LkRP4TIDkhGivLOtjyL6OHFKrT6YaGsWUPBLEoPwxItZvtk4BlVp3vPOsfgYdgCRa4U5 ZMVaTzHyEbT1lxOT9HAfb/HokvDCS1HLZzIeLfVyJAVJ4CLGu7bhYn+jEyVCeY1jYX5K GrdSTw3hKZBRKyY7vaAO4FiCIXyojWH5cb4EDShJJVgJPWt4sns9AhEgYofF0ff65FV1 6/2Z+nSoVBvR80s59nJ7rWGob0dLe9fYp0r5AZ2qSrAoRUGuCAwO/OyWbTC3auu614af DXvw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=r83N9YXN; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id pv24si3905411ejb.244.2020.05.07.11.23.47; Thu, 07 May 2020 11:24:12 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=r83N9YXN; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728045AbgEGSWO (ORCPT + 99 others); Thu, 7 May 2020 14:22:14 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:57430 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-FAIL-OK-FAIL) by vger.kernel.org with ESMTP id S1726367AbgEGSWN (ORCPT ); Thu, 7 May 2020 14:22:13 -0400 Received: from mail-yb1-xb43.google.com (mail-yb1-xb43.google.com [IPv6:2607:f8b0:4864:20::b43]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id C1A80C05BD43 for ; Thu, 7 May 2020 11:22:13 -0700 (PDT) Received: by mail-yb1-xb43.google.com with SMTP id v9so3397663ybq.13 for ; Thu, 07 May 2020 11:22:13 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :cc; bh=Hio9A8K2n+RsdYoc8pl4frI4S8NgtDDOmOIDiCLHFEg=; b=r83N9YXNHAuysfUCfvdi8Rl5DkYgLHkygL38sXx7PdanrqM5Lnz7xMFP85L4IbZCoi vCyj3Pl++/OcRr0eTL0YpewVfULtTz5to5XRjmXnOAmKmcIPNGdxUdhWDEn2382KDGWo f2uN10wq86ASCiOn1IVDTVnUvLF8VMzW3K0NrfRT2+nDaYUI+coSWuaLXUXRZdz5xcRU dL+SMJaFxZjMNS6kr5Jytapxopu2rlnWRCIHTYbAWxUyuoX4ndxHD2pqvVAz9KDU5Ng+ zlcdQ7riCkAtGifkj4MT3XGpPUyT9YxVyZVOHL37IzTh0gQOsy9dd8Ti5XG+XlrKuzIJ HxXA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to:cc; bh=Hio9A8K2n+RsdYoc8pl4frI4S8NgtDDOmOIDiCLHFEg=; b=NMiF5N1MJMWFXJ52SuArKF8SsAtcHWTyaiNn41EVk0RPVTHV97OlYyeuStXaR2/rwK 8ln/Sz+ueIRr7MX8erikdYWtibpRrKNyIdoVDNgLQycdLcCMiPJn4AVWdDVJV+mskUNx BQmoq/Bqvk3kGjyRnb15W0je7nop6RDWxut43admzQwlU71jC6JXSi3Sr+zBg1LSj+Wz 6Z/UVs7sD6o2b9rfXk6eBwYOPfREhW/ONEtS/dDmMp+NteP2CNFtsAdBn9GZyr1lY1iX DXN5qdJn61e6FOfWER7rtx2f97Nb2k8CaFT9Z9DciBPQY368ut5dwQv5DpKEYKrPW+B4 DIzA== X-Gm-Message-State: AGi0PuZ1ebkuvS/GqjVNJMZLxJwmrdqS0LEWSkQjfb98HL0EVXlAatO5 0IHjUi7kxWFSpdwszVbTVkwb1hTyfrDigQmwcKI= X-Received: by 2002:a25:5081:: with SMTP id e123mr24424676ybb.48.1588875732811; Thu, 07 May 2020 11:22:12 -0700 (PDT) MIME-Version: 1.0 Received: by 2002:a05:6900:26b:0:0:0:0 with HTTP; Thu, 7 May 2020 11:22:12 -0700 (PDT) In-Reply-To: References: <1588795317-20879-1-git-send-email-vjitta@codeaurora.org> From: Ajay kumar Date: Thu, 7 May 2020 23:52:12 +0530 Message-ID: Subject: Re: [PATCH] iommu/iova: Retry from last rb tree node if iova search fails To: Robin Murphy Cc: vjitta@codeaurora.org, joro@8bytes.org, iommu@lists.linux-foundation.org, linux-kernel@vger.kernel.org, vinmenon@codeaurora.org, kernel-team@android.com Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 5/7/20, Robin Murphy wrote: > On 2020-05-06 9:01 pm, vjitta@codeaurora.org wrote: >> From: Vijayanand Jitta >> >> When ever a new iova alloc request comes iova is always searched >> from the cached node and the nodes which are previous to cached >> node. So, even if there is free iova space available in the nodes >> which are next to the cached node iova allocation can still fail >> because of this approach. >> >> Consider the following sequence of iova alloc and frees on >> 1GB of iova space >> >> 1) alloc - 500MB >> 2) alloc - 12MB >> 3) alloc - 499MB >> 4) free - 12MB which was allocated in step 2 >> 5) alloc - 13MB >> >> After the above sequence we will have 12MB of free iova space and >> cached node will be pointing to the iova pfn of last alloc of 13MB >> which will be the lowest iova pfn of that iova space. Now if we get an >> alloc request of 2MB we just search from cached node and then look >> for lower iova pfn's for free iova and as they aren't any, iova alloc >> fails though there is 12MB of free iova space. > > Yup, this could definitely do with improving. Unfortunately I think this > particular implementation is slightly flawed... > >> To avoid such iova search failures do a retry from the last rb tree node >> when iova search fails, this will search the entire tree and get an iova >> if its available >> >> Signed-off-by: Vijayanand Jitta >> --- >> drivers/iommu/iova.c | 11 +++++++++++ >> 1 file changed, 11 insertions(+) >> >> diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c >> index 0e6a953..2985222 100644 >> --- a/drivers/iommu/iova.c >> +++ b/drivers/iommu/iova.c >> @@ -186,6 +186,7 @@ static int __alloc_and_insert_iova_range(struct >> iova_domain *iovad, >> unsigned long flags; >> unsigned long new_pfn; >> unsigned long align_mask = ~0UL; >> + bool retry = false; >> >> if (size_aligned) >> align_mask <<= fls_long(size - 1); >> @@ -198,6 +199,8 @@ static int __alloc_and_insert_iova_range(struct >> iova_domain *iovad, >> >> curr = __get_cached_rbnode(iovad, limit_pfn); >> curr_iova = rb_entry(curr, struct iova, node); >> + >> +retry_search: >> do { >> limit_pfn = min(limit_pfn, curr_iova->pfn_lo); >> new_pfn = (limit_pfn - size) & align_mask; >> @@ -207,6 +210,14 @@ static int __alloc_and_insert_iova_range(struct >> iova_domain *iovad, >> } while (curr && new_pfn <= curr_iova->pfn_hi); >> >> if (limit_pfn < size || new_pfn < iovad->start_pfn) { >> + if (!retry) { >> + curr = rb_last(&iovad->rbroot); > > Why walk when there's an anchor node there already? However... +1 > >> + curr_iova = rb_entry(curr, struct iova, node); >> + limit_pfn = curr_iova->pfn_lo; > > ...this doesn't look right, as by now we've lost the original limit_pfn > supplied by the caller, so are highly likely to allocate beyond the > range our caller asked for. In fact AFAICS we'd start allocating from > directly directly below the anchor node, beyond the end of the entire > address space. +1 > > The logic I was imagining we want here was something like the rapidly > hacked up (and untested) diff below. > > Thanks, > Robin. > > ----->8----- > diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c > index 0e6a9536eca6..3574c19272d6 100644 > --- a/drivers/iommu/iova.c > +++ b/drivers/iommu/iova.c > @@ -186,6 +186,7 @@ static int __alloc_and_insert_iova_range(struct > iova_domain *iovad, > unsigned long flags; > unsigned long new_pfn; > unsigned long align_mask = ~0UL; > + unsigned long alloc_hi, alloc_lo; > > if (size_aligned) > align_mask <<= fls_long(size - 1); > @@ -196,17 +197,27 @@ static int __alloc_and_insert_iova_range(struct > iova_domain *iovad, > size >= iovad->max32_alloc_size) > goto iova32_full; > > + alloc_hi = IOVA_ANCHOR; > + alloc_lo = iovad->start_pfn; > +retry: > curr = __get_cached_rbnode(iovad, limit_pfn); > curr_iova = rb_entry(curr, struct iova, node); > + if (alloc_hi < curr_iova->pfn_hi) { > + alloc_lo = curr_iova->pfn_hi; > + alloc_hi = limit_pfn; > + } > + > do { > - limit_pfn = min(limit_pfn, curr_iova->pfn_lo); > - new_pfn = (limit_pfn - size) & align_mask; > + alloc_hi = min(alloc_hi, curr_iova->pfn_lo); During retry case, the curr and curr_iova is not updated. Kindly check it. Ajay > + new_pfn = (alloc_hi - size) & align_mask; > prev = curr; > curr = rb_prev(curr); > curr_iova = rb_entry(curr, struct iova, node); > } while (curr && new_pfn <= curr_iova->pfn_hi); > > - if (limit_pfn < size || new_pfn < iovad->start_pfn) { > + if (limit_pfn < size || new_pfn < alloc_lo) { > + if (alloc_lo == iovad->start_pfn) > + goto retry; > iovad->max32_alloc_size = size; > goto iova32_full; > } > _______________________________________________ > iommu mailing list > iommu@lists.linux-foundation.org > https://lists.linuxfoundation.org/mailman/listinfo/iommu >