Received: by 2002:a05:6358:9144:b0:117:f937:c515 with SMTP id r4csp3884470rwr; Sun, 7 May 2023 22:37:23 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ5aK9b9QYYn28F4xLHm5JyIlv0EgIU3E702TgbpyT0Ake9CCgKBjLEotuuOSp31u30eYeV2 X-Received: by 2002:a05:6a20:8f06:b0:ff:e397:a87a with SMTP id b6-20020a056a208f0600b000ffe397a87amr6129551pzk.33.1683524243557; Sun, 07 May 2023 22:37:23 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1683524243; cv=none; d=google.com; s=arc-20160816; b=FQNHMLz37N+7oNe0OsNadG+84gCKRhVEgzhsh6tXZWbdpdJetmw32qRdxGQsFOHz6d CEKHOrgbboIEP2wYw2kcKe4ltSSB0jOACF+dy3FnC5gCr4vAUtrIC1xirms/kkkteWxC rkZMBTFrU9RoFKNuV5e0vH/U5oIz3gzdMNpmxubHXmGfc99QvbRUCEEv+nWgXhYEzmF7 xGh3tRCQB8ILEuulobFsbopP54MgmsP2q7lBMTtjhy0zk2/q1eOpmpNaRTWmUtB7glvY anHtpu4U5auR4+KwEiraeNzdqPLEwtnqyQ/zCaYqcUfkkmZE5mrWq07E94e5sGZAgap1 0tAg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:cc:to:subject :message-id:date:from:in-reply-to:references:mime-version :dkim-signature; bh=cqwukbyFeCNx0bbIvSFkt/k69jRSOprVKk38W8Cr1Bo=; b=Je3mTiQDwFDx5fRrc0e++KX7f16Wi43CoBJrrw2DAwSGtAlal9ZgF2WFFx1Mgs2IhG MWC1yGfkKbecfqMrtYX3XWIoY9r5WaSAaEzTXLUyNynXPyGXD23Tp6TfPf9pYAlrkx8h fXei9U10zwVCdWubQCmd6nWd6/a8p4MgI2xhEwxgD6H8hgJRB5VIZfITkL4yDISWOKD4 QfKdiu5SH77dxzGqQ+GKAH3kmdymvmMX5Y3CbDVLpfRYCBMe+PKapZhan0ujcGoL1yZ4 zPzAWkAgdOpq6BTElOwmiDDUvO8Ow+iEBv5wQRPOlLELQ24mF+Fvq7xMsEzgojhWVeiN hQRA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=quTp4lfd; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id b6-20020a17090ae38600b002469c3f344asi11975916pjz.93.2023.05.07.22.37.09; Sun, 07 May 2023 22:37:23 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=quTp4lfd; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231641AbjEHFd6 (ORCPT + 99 others); Mon, 8 May 2023 01:33:58 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:50594 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232201AbjEHFdz (ORCPT ); Mon, 8 May 2023 01:33:55 -0400 Received: from dfw.source.kernel.org (dfw.source.kernel.org [IPv6:2604:1380:4641:c500::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D4A3D18DC8 for ; Sun, 7 May 2023 22:33:31 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id ECFBA61E90 for ; Mon, 8 May 2023 05:33:30 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 52E01C4339E for ; Mon, 8 May 2023 05:33:30 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1683524010; bh=6AxtpopKupEMaRzxkcHkF8s4QgFRLNFkXOduFZdPEM4=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=quTp4lfd9Uro2Gz13kSnGV4z/XPaShGsVEF+pfP/g9S7gsFFw/SiGKVoi62AWzdZ3 +rLU2qPBSQ8BV1VWgdlUTT13ROuP3SAouIghdRO4d+8rX/BxS4I3z8vKT7U/zfL1hK xmRoL5vMM7w04h9ObuHDlJaDMk1EuZdWuzUcAr8Kq6Pyu/IC61t9A2K8kw25C734IK efRowweZ+10SLHBqRq0EG8i+Jd10/m+mtiKcAVNhHTGFbASDJCrgIKTy8fNhKtVzfN toX5hs/N+X1ctC/3YyNzAAQXC0AFTt8t66lrH7aOo/VE7VzfTFJSmeApJq0Jej9iZ2 esTuLla4AluDA== Received: by mail-yw1-f181.google.com with SMTP id 00721157ae682-55a00da4e53so74060387b3.0 for ; Sun, 07 May 2023 22:33:30 -0700 (PDT) X-Gm-Message-State: AC+VfDzUs1vsE4ZQuoLlaS8+3yZwGR42XSL58Cbi5rGAqJvNtz2fYsXg WQ6c06buH1hEXC3FJp5oYr9enbTt5ILuI+rTfkM= X-Received: by 2002:a81:60c6:0:b0:54d:ea34:c31 with SMTP id u189-20020a8160c6000000b0054dea340c31mr9358431ywb.29.1683524009143; Sun, 07 May 2023 22:33:29 -0700 (PDT) MIME-Version: 1.0 References: <20230428144903.26048-1-cai.huoqing@linux.dev> In-Reply-To: From: Oded Gabbay Date: Mon, 8 May 2023 08:33:02 +0300 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [PATCH v2] accel/habanalabs: Make use of rhashtable To: Cai Huoqing Cc: Ohad Sharabi , Greg Kroah-Hartman , dri-devel@lists.freedesktop.org, linux-kernel@vger.kernel.org Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Spam-Status: No, score=-7.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, May 8, 2023 at 8:28=E2=80=AFAM Cai Huoqing = wrote: > > On 07 5=E6=9C=88 23 16:17:55, Oded Gabbay wrote: > > On Sat, May 6, 2023 at 12:25=E2=80=AFPM Cai Huoqing wrote: > > > > > > On 04 5=E6=9C=88 23 09:12:40, Oded Gabbay wrote: > > > > On Thu, May 4, 2023 at 6:00=E2=80=AFAM Cai Huoqing wrote: > > > > > > > > > > On 30 4=E6=9C=88 23 09:36:29, Oded Gabbay wrote: > > > > > > On Fri, Apr 28, 2023 at 5:49=E2=80=AFPM Cai Huoqing wrote: > > > > > > > > > > > > > > Using rhashtable to accelerate the search for userptr by addr= ess, > > > > > > > instead of using a list. > > > > > > > > > > > > > > Preferably, the lookup complexity of a hash table is O(1). > > > > > > > > > > > > > > This patch will speedup the method > > > > > > > hl_userptr_is_pinned by rhashtable_lookup_fast. > > > > > > > > > > > > > > Signed-off-by: Cai Huoqing > > > > > > > > > > > > Thanks for the patch, but the reason we never optimized this pa= th is > > > > > > because this code path is only relevant for Goya, which we don'= t want > > > > > > to change the code for anymore. > > > > > > For Gaudi we don't pin the memory in the host during submission= . It is > > > > > > done much earlier, when the user maps the memory to the device.= The > > > > > > code path in Gaudi is only in case the device is loaded with it= s MMU > > > > > > disabled. This mode was used only for debug/bring-up of the ASI= C many > > > > > > years ago. As you can see in Gaudi2, that mode was dropped even= for > > > > > > > > > > Do you mean that the userspace directly call HL_MEM_OP_MAP/HL_MEM= _OP_UNMAP > > > > > with a flag HL_MEM_USERPTR instead of pin host mem in submissmion= ? > > > > Yes, correct. > > > > > > BTW, > > > > > > Another question, > > > > > > I found the PMMU in the driver code, does it support for like CudaHos= tAlloc > > > and hostMap? > > In Habana, you allocate host memory by yourself, not through the > > driver. Just do malloc in your application. > > Then you need to map it to the device to get a virtual address that > > the device is familiar with. > > This is done by calling the MEMORY ioctl with the MAP opcode. > > > > > > > > gaudi seems can access host memory by pcie, some kernel can compute d= ata > > > in host directly, right? > > In theory, yes. However, we don't do that. We move data into the HBM > > and then the kernels work on the data. > > Oded > Agree, but What is the purpose of designing pmmu, > do you think using the pagetable in PMMU can speed up memcpy > H2D/D2H or some memory notity between CPU and device ? Without pmmu, you will have to allocate dmable memory in the host, which will have to be physical contiguous memory. That means the user can't use malloc to do it, but he must call the driver (so it can call dma_alloc_coherent or similar), which is bad. And the amount of that memory is limited, which is also bad. With PMMU, you can map any host virtual memory space, which is backed by non-contiguous physical pages, to a device virtual address. Oded > > Thanks > Cai- > > > > > > > But accessing via PCIE(64GB/s), compared to accessing HBM(3T/s), > > > has too low bandwidth. > > > > > > for nvidia, it has NVLINK. > > > > > > Thank, > > > Cai- > > > > > > > > > > > > > > > > > > debug/bring-up. > > > > > > > > > > > > Therefore, I prefer not to take this patch as validation for bo= th > > > > > > functionality and performance will take time which will be bett= er > > > > > > spent elsewhere. > > > > > > > > > > > > Thanks, > > > > > > Oded > > > > > > > > > > > > > --- > > > > > > > v1->v2: > > > > > > > Use rhashtable_free_and_destroy in hl_userptr_delete_list. > > > > > > > > > > > > > > .../habanalabs/common/command_submission.c | 16 ++++++-- > > > > > > > drivers/accel/habanalabs/common/habanalabs.h | 19 +++++---- > > > > > > > drivers/accel/habanalabs/common/memory.c | 39 +++++++++= ++-------- > > > > > > > drivers/accel/habanalabs/gaudi/gaudi.c | 16 +++++--- > > > > > > > drivers/accel/habanalabs/goya/goya.c | 14 ++++--- > > > > > > > 5 files changed, 65 insertions(+), 39 deletions(-) > > > > > > > > > > > > > > diff --git a/drivers/accel/habanalabs/common/command_submissi= on.c b/drivers/accel/habanalabs/common/command_submission.c > > > > > > > index af9d2e22c6e7..35c2ab934396 100644 > > > > > > > --- a/drivers/accel/habanalabs/common/command_submission.c > > > > > > > +++ b/drivers/accel/habanalabs/common/command_submission.c > > > > > > > @@ -312,7 +312,7 @@ static int cs_parser(struct hl_fpriv *hpr= iv, struct hl_cs_job *job) > > > > > > > parser.job_id =3D job->id; > > > > > > > > > > > > > > parser.hw_queue_id =3D job->hw_queue_id; > > > > > > > - parser.job_userptr_list =3D &job->userptr_list; > > > > > > > + parser.job_userptr_ht =3D &job->userptr_ht; > > > > > > > parser.patched_cb =3D NULL; > > > > > > > parser.user_cb =3D job->user_cb; > > > > > > > parser.user_cb_size =3D job->user_cb_size; > > > > > > > @@ -351,7 +351,7 @@ static void hl_complete_job(struct hl_dev= ice *hdev, struct hl_cs_job *job) > > > > > > > struct hl_cs *cs =3D job->cs; > > > > > > > > > > > > > > if (is_cb_patched(hdev, job)) { > > > > > > > - hl_userptr_delete_list(hdev, &job->userptr_li= st); > > > > > > > + hl_userptr_delete_list(hdev, &job->userptr_ht= ); > > > > > > > > > > > > > > /* > > > > > > > * We might arrive here from rollback and pat= ched CB wasn't > > > > > > > @@ -1284,6 +1284,7 @@ struct hl_cs_job *hl_cs_allocate_job(st= ruct hl_device *hdev, > > > > > > > enum hl_queue_type queue_type, bool is_kernel= _allocated_cb) > > > > > > > { > > > > > > > struct hl_cs_job *job; > > > > > > > + int rc; > > > > > > > > > > > > > > job =3D kzalloc(sizeof(*job), GFP_ATOMIC); > > > > > > > if (!job) > > > > > > > @@ -1296,13 +1297,20 @@ struct hl_cs_job *hl_cs_allocate_job(= struct hl_device *hdev, > > > > > > > job->queue_type =3D queue_type; > > > > > > > job->is_kernel_allocated_cb =3D is_kernel_allocated_c= b; > > > > > > > > > > > > > > - if (is_cb_patched(hdev, job)) > > > > > > > - INIT_LIST_HEAD(&job->userptr_list); > > > > > > > + if (is_cb_patched(hdev, job)) { > > > > > > > + rc =3D rhashtable_init(&job->userptr_ht, &hl_= userptr_rht_params); > > > > > > > + if (rc) > > > > > > > + goto free_job; > > > > > > > + } > > > > > > > > > > > > > > if (job->queue_type =3D=3D QUEUE_TYPE_EXT) > > > > > > > INIT_WORK(&job->finish_work, job_wq_completio= n); > > > > > > > > > > > > > > return job; > > > > > > > + > > > > > > > +free_job: > > > > > > > + kfree(job); > > > > > > > + return NULL; > > > > > > > } > > > > > > > > > > > > > > static enum hl_cs_type hl_cs_get_cs_type(u32 cs_type_flags) > > > > > > > diff --git a/drivers/accel/habanalabs/common/habanalabs.h b/d= rivers/accel/habanalabs/common/habanalabs.h > > > > > > > index eaae69a9f817..9c876d1480d2 100644 > > > > > > > --- a/drivers/accel/habanalabs/common/habanalabs.h > > > > > > > +++ b/drivers/accel/habanalabs/common/habanalabs.h > > > > > > > @@ -19,6 +19,7 @@ > > > > > > > #include > > > > > > > #include > > > > > > > #include > > > > > > > +#include > > > > > > > #include > > > > > > > #include > > > > > > > #include > > > > > > > @@ -540,6 +541,8 @@ struct hl_hints_range { > > > > > > > u64 end_addr; > > > > > > > }; > > > > > > > > > > > > > > +extern const struct rhashtable_params hl_userptr_rht_params; > > > > > > > + > > > > > > > /** > > > > > > > * struct asic_fixed_properties - ASIC specific immutable pr= operties. > > > > > > > * @hw_queues_props: H/W queues properties. > > > > > > > @@ -1915,7 +1918,7 @@ struct hl_ctx_mgr { > > > > > > > /** > > > > > > > * struct hl_userptr - memory mapping chunk information > > > > > > > * @vm_type: type of the VM. > > > > > > > - * @job_node: linked-list node for hanging the object on the= Job's list. > > > > > > > + * @job_node: hashtable node for hanging the object on the J= ob's list. > > > > > > > * @pages: pointer to struct page array > > > > > > > * @npages: size of @pages array > > > > > > > * @sgt: pointer to the scatter-gather table that holds the = pages. > > > > > > > @@ -1928,7 +1931,7 @@ struct hl_ctx_mgr { > > > > > > > */ > > > > > > > struct hl_userptr { > > > > > > > enum vm_type vm_type; /* must be first */ > > > > > > > - struct list_head job_node; > > > > > > > + struct rhash_head job_node; > > > > > > > struct page **pages; > > > > > > > unsigned int npages; > > > > > > > struct sg_table *sgt; > > > > > > > @@ -2028,7 +2031,7 @@ struct hl_cs { > > > > > > > * @patched_cb: in case of patching, this is internal CB whi= ch is submitted on > > > > > > > * the queue instead of the CB we got from the I= OCTL. > > > > > > > * @finish_work: workqueue object to run when job is complet= ed. > > > > > > > - * @userptr_list: linked-list of userptr mappings that belon= g to this job and > > > > > > > + * @userptr_ht: hashtable of userptr mappings that belong to= this job and > > > > > > > * wait for completion. > > > > > > > * @debugfs_list: node in debugfs list of command submission= jobs. > > > > > > > * @refcount: reference counter for usage of the CS job. > > > > > > > @@ -2056,7 +2059,7 @@ struct hl_cs_job { > > > > > > > struct hl_cb *user_cb; > > > > > > > struct hl_cb *patched_cb; > > > > > > > struct work_struct finish_work; > > > > > > > - struct list_head userptr_list; > > > > > > > + struct rhashtable userptr_ht; > > > > > > > struct list_head debugfs_list; > > > > > > > struct kref refcount; > > > > > > > enum hl_queue_type queue_type; > > > > > > > @@ -2075,7 +2078,7 @@ struct hl_cs_job { > > > > > > > * @user_cb: the CB we got from the user. > > > > > > > * @patched_cb: in case of patching, this is internal CB whi= ch is submitted on > > > > > > > * the queue instead of the CB we got from the I= OCTL. > > > > > > > - * @job_userptr_list: linked-list of userptr mappings that b= elong to the related > > > > > > > + * @job_userptr_ht: hashtable of userptr mappings that belon= g to the related > > > > > > > * job and wait for completion. > > > > > > > * @cs_sequence: the sequence number of the related CS. > > > > > > > * @queue_type: the type of the H/W queue this job is submit= ted to. > > > > > > > @@ -2098,7 +2101,7 @@ struct hl_cs_job { > > > > > > > struct hl_cs_parser { > > > > > > > struct hl_cb *user_cb; > > > > > > > struct hl_cb *patched_cb; > > > > > > > - struct list_head *job_userptr_list; > > > > > > > + struct rhashtable *job_userptr_ht; > > > > > > > u64 cs_sequence; > > > > > > > enum hl_queue_type queue_type; > > > > > > > u32 ctx_id; > > > > > > > @@ -3760,9 +3763,9 @@ int hl_pin_host_memory(struct hl_device= *hdev, u64 addr, u64 size, > > > > > > > struct hl_userptr *userptr); > > > > > > > void hl_unpin_host_memory(struct hl_device *hdev, struct hl_= userptr *userptr); > > > > > > > void hl_userptr_delete_list(struct hl_device *hdev, > > > > > > > - struct list_head *userptr_lis= t); > > > > > > > + struct rhashtable *userptr_ht= ); > > > > > > > bool hl_userptr_is_pinned(struct hl_device *hdev, u64 addr, = u32 size, > > > > > > > - struct list_head *userptr_lis= t, > > > > > > > + struct rhashtable *userptr_ht= , > > > > > > > struct hl_userptr **userptr); > > > > > > > > > > > > > > int hl_mmu_init(struct hl_device *hdev); > > > > > > > diff --git a/drivers/accel/habanalabs/common/memory.c b/drive= rs/accel/habanalabs/common/memory.c > > > > > > > index a7b6a273ce21..fa2104e33639 100644 > > > > > > > --- a/drivers/accel/habanalabs/common/memory.c > > > > > > > +++ b/drivers/accel/habanalabs/common/memory.c > > > > > > > @@ -23,6 +23,13 @@ MODULE_IMPORT_NS(DMA_BUF); > > > > > > > > > > > > > > #define MEM_HANDLE_INVALID ULONG_MAX > > > > > > > > > > > > > > +const struct rhashtable_params hl_userptr_rht_params =3D { > > > > > > > + .head_offset =3D offsetof(struct hl_userptr, job_node= ), > > > > > > > + .key_offset =3D offsetof(struct hl_userptr, addr), > > > > > > > + .key_len =3D sizeof(u64), > > > > > > > + .automatic_shrinking =3D true, > > > > > > > +}; > > > > > > > + > > > > > > > static int allocate_timestamps_buffers(struct hl_fpriv *hpri= v, > > > > > > > struct hl_mem_in *args, u64 *handle); > > > > > > > > > > > > > > @@ -2483,7 +2490,6 @@ int hl_pin_host_memory(struct hl_device= *hdev, u64 addr, u64 size, > > > > > > > userptr->size =3D size; > > > > > > > userptr->addr =3D addr; > > > > > > > userptr->dma_mapped =3D false; > > > > > > > - INIT_LIST_HEAD(&userptr->job_node); > > > > > > > > > > > > > > rc =3D get_user_memory(hdev, addr, size, npages, star= t, offset, > > > > > > > userptr); > > > > > > > @@ -2522,32 +2528,32 @@ void hl_unpin_host_memory(struct hl_d= evice *hdev, struct hl_userptr *userptr) > > > > > > > unpin_user_pages_dirty_lock(userptr->pages, userptr->= npages, true); > > > > > > > kvfree(userptr->pages); > > > > > > > > > > > > > > - list_del(&userptr->job_node); > > > > > > > - > > > > > > > sg_free_table(userptr->sgt); > > > > > > > kfree(userptr->sgt); > > > > > > > } > > > > > > > > > > > > > > +static void hl_userptr_free_cb(void *ptr, void *arg) > > > > > > > +{ > > > > > > > + struct hl_userptr *userptr =3D ptr; > > > > > > > + struct hl_device *hdev =3D (struct hl_device *)arg; > > > > > > > + > > > > > > > + hl_unpin_host_memory(hdev, userptr); > > > > > > > + kfree(userptr); > > > > > > > +} > > > > > > > + > > > > > > > /** > > > > > > > * hl_userptr_delete_list() - clear userptr list. > > > > > > > * @hdev: pointer to the habanalabs device structure. > > > > > > > - * @userptr_list: pointer to the list to clear. > > > > > > > + * @userptr_ht: pointer to the hashtable to clear. > > > > > > > * > > > > > > > * This function does the following: > > > > > > > * - Iterates over the list and unpins the host memory and f= rees the userptr > > > > > > > * structure. > > > > > > > */ > > > > > > > void hl_userptr_delete_list(struct hl_device *hdev, > > > > > > > - struct list_head *userptr_lis= t) > > > > > > > + struct rhashtable *userptr_ht= ) > > > > > > > { > > > > > > > - struct hl_userptr *userptr, *tmp; > > > > > > > - > > > > > > > - list_for_each_entry_safe(userptr, tmp, userptr_list, = job_node) { > > > > > > > - hl_unpin_host_memory(hdev, userptr); > > > > > > > - kfree(userptr); > > > > > > > - } > > > > > > > - > > > > > > > - INIT_LIST_HEAD(userptr_list); > > > > > > > + rhashtable_free_and_destroy(userptr_ht, hl_userptr_fr= ee_cb, hdev); > > > > > > > } > > > > > > > > > > > > > > /** > > > > > > > @@ -2555,7 +2561,7 @@ void hl_userptr_delete_list(struct hl_d= evice *hdev, > > > > > > > * @hdev: pointer to the habanalabs device structure. > > > > > > > * @addr: user address to check. > > > > > > > * @size: user block size to check. > > > > > > > - * @userptr_list: pointer to the list to clear. > > > > > > > + * @userptr_ht: pointer to the hashtable to clear. > > > > > > > * @userptr: pointer to userptr to check. > > > > > > > * > > > > > > > * This function does the following: > > > > > > > @@ -2563,10 +2569,11 @@ void hl_userptr_delete_list(struct hl= _device *hdev, > > > > > > > * pinned. If so, returns true, otherwise returns false. > > > > > > > */ > > > > > > > bool hl_userptr_is_pinned(struct hl_device *hdev, u64 addr, > > > > > > > - u32 size, struct list_head *u= serptr_list, > > > > > > > + u32 size, struct rhashtable *= userptr_ht, > > > > > > > struct hl_userptr **userptr) > > > > > > > { > > > > > > > - list_for_each_entry((*userptr), userptr_list, job_nod= e) { > > > > > > > + (*userptr) =3D rhashtable_lookup_fast(userptr_ht, &ad= dr, hl_userptr_rht_params); > > > > > > > + if (*userptr) { > > > > > > > if ((addr =3D=3D (*userptr)->addr) && (size = =3D=3D (*userptr)->size)) > > > > > > > return true; > > > > > > > } > > > > > > > diff --git a/drivers/accel/habanalabs/gaudi/gaudi.c b/drivers= /accel/habanalabs/gaudi/gaudi.c > > > > > > > index a29aa8f7b6f3..1e1433042413 100644 > > > > > > > --- a/drivers/accel/habanalabs/gaudi/gaudi.c > > > > > > > +++ b/drivers/accel/habanalabs/gaudi/gaudi.c > > > > > > > @@ -1031,7 +1031,7 @@ static int _gaudi_init_tpc_mem(struct h= l_device *hdev, > > > > > > > } > > > > > > > > > > > > > > free_job: > > > > > > > - hl_userptr_delete_list(hdev, &job->userptr_list); > > > > > > > + hl_userptr_delete_list(hdev, &job->userptr_ht); > > > > > > > hl_debugfs_remove_job(hdev, job); > > > > > > > kfree(job); > > > > > > > atomic_dec(&cb->cs_cnt); > > > > > > > @@ -4901,7 +4901,7 @@ static int gaudi_pin_memory_before_cs(s= truct hl_device *hdev, > > > > > > > int rc; > > > > > > > > > > > > > > if (hl_userptr_is_pinned(hdev, addr, le32_to_cpu(user= _dma_pkt->tsize), > > > > > > > - parser->job_userptr_list, &userptr)) > > > > > > > + parser->job_userptr_ht, &userptr)) > > > > > > > goto already_pinned; > > > > > > > > > > > > > > userptr =3D kzalloc(sizeof(*userptr), GFP_KERNEL); > > > > > > > @@ -4913,7 +4913,10 @@ static int gaudi_pin_memory_before_cs(= struct hl_device *hdev, > > > > > > > if (rc) > > > > > > > goto free_userptr; > > > > > > > > > > > > > > - list_add_tail(&userptr->job_node, parser->job_userptr= _list); > > > > > > > + rc =3D rhashtable_insert_fast(parser->job_userptr_ht, > > > > > > > + &userptr->job_node, hl_us= erptr_rht_params); > > > > > > > + if (rc) > > > > > > > + goto unpin_memory; > > > > > > > > > > > > > > rc =3D hdev->asic_funcs->asic_dma_map_sgtable(hdev, u= serptr->sgt, dir); > > > > > > > if (rc) { > > > > > > > @@ -4931,7 +4934,8 @@ static int gaudi_pin_memory_before_cs(s= truct hl_device *hdev, > > > > > > > return 0; > > > > > > > > > > > > > > unpin_memory: > > > > > > > - list_del(&userptr->job_node); > > > > > > > + rhashtable_remove_fast(parser->job_userptr_ht, > > > > > > > + &userptr->job_node, hl_userptr= _rht_params); > > > > > > > hl_unpin_host_memory(hdev, userptr); > > > > > > > free_userptr: > > > > > > > kfree(userptr); > > > > > > > @@ -5175,7 +5179,7 @@ static int gaudi_patch_dma_packet(struc= t hl_device *hdev, > > > > > > > if ((!skip_host_mem_pin) && > > > > > > > (!hl_userptr_is_pinned(hdev, addr, > > > > > > > le32_to_cpu(user_dma_= pkt->tsize), > > > > > > > - parser->job_userptr_l= ist, &userptr))) { > > > > > > > + parser->job_userptr_h= t, &userptr))) { > > > > > > > dev_err(hdev->dev, "Userptr 0x%llx + 0x%x NOT= mapped\n", > > > > > > > addr, user_dma_pkt->tsize); > > > > > > > return -EFAULT; > > > > > > > @@ -5472,7 +5476,7 @@ static int gaudi_parse_cb_no_mmu(struct= hl_device *hdev, > > > > > > > > > > > > > > free_userptr: > > > > > > > if (rc) > > > > > > > - hl_userptr_delete_list(hdev, parser->job_user= ptr_list); > > > > > > > + hl_userptr_delete_list(hdev, parser->job_user= ptr_ht); > > > > > > > return rc; > > > > > > > } > > > > > > > > > > > > > > diff --git a/drivers/accel/habanalabs/goya/goya.c b/drivers/a= ccel/habanalabs/goya/goya.c > > > > > > > index fb0ac9df841a..bfcbb9e8b126 100644 > > > > > > > --- a/drivers/accel/habanalabs/goya/goya.c > > > > > > > +++ b/drivers/accel/habanalabs/goya/goya.c > > > > > > > @@ -3347,7 +3347,7 @@ static int goya_pin_memory_before_cs(st= ruct hl_device *hdev, > > > > > > > int rc; > > > > > > > > > > > > > > if (hl_userptr_is_pinned(hdev, addr, le32_to_cpu(user= _dma_pkt->tsize), > > > > > > > - parser->job_userptr_list, &userptr)) > > > > > > > + parser->job_userptr_ht, &userptr)) > > > > > > > goto already_pinned; > > > > > > > > > > > > > > userptr =3D kzalloc(sizeof(*userptr), GFP_KERNEL); > > > > > > > @@ -3359,7 +3359,10 @@ static int goya_pin_memory_before_cs(s= truct hl_device *hdev, > > > > > > > if (rc) > > > > > > > goto free_userptr; > > > > > > > > > > > > > > - list_add_tail(&userptr->job_node, parser->job_userptr= _list); > > > > > > > + rc =3D rhashtable_insert_fast(parser->job_userptr_ht, > > > > > > > + &userptr->job_node, hl_us= erptr_rht_params); > > > > > > > + if (rc) > > > > > > > + goto unpin_memory; > > > > > > > > > > > > > > rc =3D hdev->asic_funcs->asic_dma_map_sgtable(hdev, u= serptr->sgt, dir); > > > > > > > if (rc) { > > > > > > > @@ -3377,7 +3380,8 @@ static int goya_pin_memory_before_cs(st= ruct hl_device *hdev, > > > > > > > return 0; > > > > > > > > > > > > > > unpin_memory: > > > > > > > - list_del(&userptr->job_node); > > > > > > > + rhashtable_remove_fast(parser->job_userptr_ht, > > > > > > > + &userptr->job_node, hl_userptr= _rht_params); > > > > > > > hl_unpin_host_memory(hdev, userptr); > > > > > > > free_userptr: > > > > > > > kfree(userptr); > > > > > > > @@ -3806,7 +3810,7 @@ static int goya_patch_dma_packet(struct= hl_device *hdev, > > > > > > > if ((!skip_host_mem_pin) && > > > > > > > (hl_userptr_is_pinned(hdev, addr, > > > > > > > le32_to_cpu(user_dma_pkt->tsize), > > > > > > > - parser->job_userptr_list, &userptr) = =3D=3D false)) { > > > > > > > + parser->job_userptr_ht, &userptr) =3D= =3D false)) { > > > > > > > dev_err(hdev->dev, "Userptr 0x%llx + 0x%x NOT= mapped\n", > > > > > > > addr, user_dma_pkt->tsize); > > > > > > > return -EFAULT; > > > > > > > @@ -4104,7 +4108,7 @@ static int goya_parse_cb_no_mmu(struct = hl_device *hdev, > > > > > > > > > > > > > > free_userptr: > > > > > > > if (rc) > > > > > > > - hl_userptr_delete_list(hdev, parser->job_user= ptr_list); > > > > > > > + hl_userptr_delete_list(hdev, parser->job_user= ptr_ht); > > > > > > > return rc; > > > > > > > } > > > > > > > > > > > > > > -- > > > > > > > 2.34.1 > > > > > > >