Received: by 2002:a05:6a10:f3d0:0:0:0:0 with SMTP id a16csp4545530pxv; Tue, 6 Jul 2021 03:38:28 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxFIPLfb47mlMFJNOfiEUd/uK35qWbpDTiB/YVKJDkKkn63K33d3srOMAjasapM/1GlkGBh X-Received: by 2002:a17:907:1ca4:: with SMTP id nb36mr17669329ejc.33.1625567907728; Tue, 06 Jul 2021 03:38:27 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1625567907; cv=none; d=google.com; s=arc-20160816; b=Vm712+Yfl63hlqvXpzzOMHD6l3cBZL08IScVrcIVFrTxn2vclvmX3Mvy8UKzOGFdux bpgvZmT0XvuKcafVzXtXOv/R4EDhW56YNRhfQk/ETJAv2DHVs33zQNZ/iF3oVLC1M/tD i0W4pSoMYypNqZhTHVIDnIOivbFvX6qwWVfG71H22IXDEYDCoO9gLO2/GtmHB9DJNbGp zSSi8GPgjLNONCJoSqDBmmbxH/IZssnbhlRZ7NENodX+3ShZsOkIxxlLbAKgLDpyO3yN MHW4GHA+h/aYkRyuB7SvEVZdYmnTF7WCI+WmTS0X9dbc5tj9Ze0WoMCwkRnGqqhEHyEJ R8CQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=SAa3xS01Ip/WPvwntU0vFLMJjul0CGOxjhmVkJHt/rw=; b=FRPycuFiKFmAkjedXIM1jFGEvpiIZdrxd8kWezo2ix6bvAK+PCWjHhtzhKX2LjfmvR DhgNGAcr8pSyi9tQhaHsALjvZ5hu9LEIWRHaA9vaMrlfW99DFFrcpiQfpJ2KwcwyRn0n IdfmAadFI4HYKmMogrKmg01ey9cGIwHCpjNaeexfFsV0Q9JGGlJykCDAqcfHnWrSNGl6 VIrYznn2j12IHm0dpppAlF0ng+MOckNJr4zH980NM9Zs0gjcxlj5hoN2xuat5A2/VSj3 k5kCZrB57xHwS7Oj9uTDV0QoQLd01Rm5F6QIHgi3QbHEzdTlw863aMbq/g2/F5RIEW7/ pTiw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@ffwll.ch header.s=google header.b=RO7iacH4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id e12si14654513ejy.258.2021.07.06.03.38.04; Tue, 06 Jul 2021 03:38:27 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@ffwll.ch header.s=google header.b=RO7iacH4; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231405AbhGFKjl (ORCPT + 99 others); Tue, 6 Jul 2021 06:39:41 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:34432 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231327AbhGFKjl (ORCPT ); Tue, 6 Jul 2021 06:39:41 -0400 Received: from mail-oi1-x235.google.com (mail-oi1-x235.google.com [IPv6:2607:f8b0:4864:20::235]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 34BE3C061574 for ; Tue, 6 Jul 2021 03:37:03 -0700 (PDT) Received: by mail-oi1-x235.google.com with SMTP id t3so24032948oic.5 for ; Tue, 06 Jul 2021 03:37:03 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ffwll.ch; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=SAa3xS01Ip/WPvwntU0vFLMJjul0CGOxjhmVkJHt/rw=; b=RO7iacH4Ti46dB63sJGmihRm41QJsDCce6RQmPELZZwX88PmkhaUm6ORroBBGsrxJ5 JliO2V4qJ84GuHOn5P2qmo+8BgOvGkLh6zShltaLsTIGyv2AR8Qo4CA+azevq8heOmF0 5nQVM1kQfFFf11BNsFyJVSgOMslhlh08ECCKU= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=SAa3xS01Ip/WPvwntU0vFLMJjul0CGOxjhmVkJHt/rw=; b=cJ7ZMjkWcxWQ1yN2O3IPfYyza2I9wI0AW/TiM5IBxJx9zRbnG8K4SCrZr5AcTlWUpo aXvrXkjwdo/iy14oMeTyqoPDVFQmsShqRceupSeEB2etJesDqZ0CfViOwxQ+3TgUAonG Z14hVImWWqCFucWRpbyP8xhQh9N6gL7LEGqGhM6RYrxiLNDbAi7yg7PU6To1rxymcjGh W0j02jYb+PtLk+P5OSYCwuzstm4m7mAXWPH3oz7VKbBrmqAcNVxXHXL4XVk+2QN9v44X Ix62yeulpceaXjNBSDWOi54BU0/TXEipwgKYv7KfCgAM83SHqWmwkeXU3VLxOFFdOe08 m/ag== X-Gm-Message-State: AOAM532Y0EW1m3u56xVw6sQVXJ8AhZekD2pPc14YOp5Yeitoo38WLJn1 Oj4RBqj4GkAt+/Vpf8Vix8nrVcTGmVm2UOUiFs6A9A== X-Received: by 2002:aca:f491:: with SMTP id s139mr7925290oih.128.1625567822322; Tue, 06 Jul 2021 03:37:02 -0700 (PDT) MIME-Version: 1.0 References: <20210705130314.11519-1-ogabbay@kernel.org> In-Reply-To: From: Daniel Vetter Date: Tue, 6 Jul 2021 12:36:51 +0200 Message-ID: Subject: Re: [PATCH v4 0/2] Add p2p via dmabuf to habanalabs To: Oded Gabbay Cc: Oded Gabbay , "Linux-Kernel@Vger. Kernel. Org" , Greg Kroah-Hartman , Sumit Semwal , =?UTF-8?Q?Christian_K=C3=B6nig?= , Gal Pressman , sleybo@amazon.com, Maling list - DRI developers , Jason Gunthorpe , linux-rdma , Linux Media Mailing List , Doug Ledford , Dave Airlie , Alex Deucher , Leon Romanovsky , Christoph Hellwig , amd-gfx list , "moderated list:DMA BUFFER SHARING FRAMEWORK" Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Jul 6, 2021 at 12:03 PM Oded Gabbay wrote: > > On Tue, Jul 6, 2021 at 11:40 AM Daniel Vetter wrote: > > > > On Mon, Jul 05, 2021 at 04:03:12PM +0300, Oded Gabbay wrote: > > > Hi, > > > I'm sending v4 of this patch-set following the long email thread. > > > I want to thank Jason for reviewing v3 and pointing out the errors, saving > > > us time later to debug it :) > > > > > > I consulted with Christian on how to fix patch 2 (the implementation) and > > > at the end of the day I shamelessly copied the relevant content from > > > amdgpu_vram_mgr_alloc_sgt() and amdgpu_dma_buf_attach(), regarding the > > > usage of dma_map_resource() and pci_p2pdma_distance_many(), respectively. > > > > > > I also made a few improvements after looking at the relevant code in amdgpu. > > > The details are in the changelog of patch 2. > > > > > > I took the time to write an import code into the driver, allowing me to > > > check real P2P with two Gaudi devices, one as exporter and the other as > > > importer. I'm not going to include the import code in the product, it was > > > just for testing purposes (although I can share it if anyone wants). > > > > > > I run it on a bare-metal environment with IOMMU enabled, on a sky-lake CPU > > > with a white-listed PCIe bridge (to make the pci_p2pdma_distance_many happy). > > > > > > Greg, I hope this will be good enough for you to merge this code. > > > > So we're officially going to use dri-devel for technical details review > > and then Greg for merging so we don't have to deal with other merge > > criteria dri-devel folks have? > I'm glad to receive any help or review, regardless of the subsystem > the person giving that help belongs to. > > > > > I don't expect anything less by now, but it does make the original claim > > that drivers/misc will not step all over accelerators folks a complete > > farce under the totally-not-a-gpu banner. > > > > This essentially means that for any other accelerator stack that doesn't > > fit the dri-devel merge criteria, even if it's acting like a gpu and uses > > other gpu driver stuff, you can just send it to Greg and it's good to go. > > What's wrong with Greg ??? ;) > > On a more serious note, yes, I do think the dri-devel merge criteria > is very extreme, and effectively drives-out many AI accelerator > companies that want to contribute to the kernel but can't/won't open > their software IP and patents. > > I think the expectation from AI startups (who are 90% of the deep > learning field) to cooperate outside of company boundaries is not > realistic, especially on the user-side, where the real IP of the > company resides. > > Personally I don't think there is a real justification for that at > this point of time, but if it will make you (and other people here) > happy I really don't mind creating a non-gpu accelerator subsystem > that will contain all the totally-not-a-gpu accelerators, and will > have a more relaxed criteria for upstreaming. Something along an > "rdma-core" style library looks like the correct amount of user-level > open source that should be enough. > > The question is, what will happen later ? Will it be sufficient to > "allow" us to use dmabuf and maybe other gpu stuff in the future (e.g. > hmm) ? > > If the community and dri-devel maintainers (and you among them) will > assure me it is good enough, then I'll happily contribute my work and > personal time to organize this effort and implement it. I think dri-devel stance is pretty clear and well known: We want the userspace to be open, because that's where most of the driver stack is. Without an open driver stack there's no way to ever have anything cross-vendor. And that includes the compiler and anything else you need to drive the hardware. Afaik linux cpu arch ports are also not accepted if there's no open gcc or llvm port around, because without that the overall stack just becomes useless. If that means AI companies don't want to open our their hw specs enough to allow that, so be it - all you get in that case is offloading the kernel side of the stack for convenience, with zero long term prospects to ever make this into a cross vendor subsystem stack that does something useful. If the business case says you can't open up your hw enough for that, I really don't see the point in merging such a driver, it'll be an unmaintainable stack by anyone else who's not having access to those NDA covered specs and patents and everything. If the stack is actually cross vendor to begin with that's just bonus, but generally that doesn't happen voluntarily and needs a few years to decades to get there. So that's not really something we require. tldr; just a runtime isn't enough for dri-devel. Now Greg seems to be happy to merge kernel drivers that aren't useful with the open bits provided, so *shrug*. Cheers, Daniel PS: If requiring an actually useful open driver stack is somehow *extreme* I have no idea why we even bother with merging device drivers to upstream. Just make a stable driver api and done, vendors can then do whatever they feel like and protect their "valuable IP and patents" or whatever it is. > Thanks, > oded > > > > > There's quite a lot of these floating around actually (and many do have > > semi-open runtimes, like habanalabs have now too, just not open enough to > > be actually useful). It's going to be absolutely lovely having to explain > > to these companies in background chats why habanalabs gets away with their > > stack and they don't. > > > > Or maybe we should just merge them all and give up on the idea of having > > open cross-vendor driver stacks for these accelerators. > > > > Thanks, Daniel > > > > > > > > Thanks, > > > Oded > > > > > > Oded Gabbay (1): > > > habanalabs: define uAPI to export FD for DMA-BUF > > > > > > Tomer Tayar (1): > > > habanalabs: add support for dma-buf exporter > > > > > > drivers/misc/habanalabs/Kconfig | 1 + > > > drivers/misc/habanalabs/common/habanalabs.h | 26 ++ > > > drivers/misc/habanalabs/common/memory.c | 480 +++++++++++++++++++- > > > drivers/misc/habanalabs/gaudi/gaudi.c | 1 + > > > drivers/misc/habanalabs/goya/goya.c | 1 + > > > include/uapi/misc/habanalabs.h | 28 +- > > > 6 files changed, 532 insertions(+), 5 deletions(-) > > > > > > -- > > > 2.25.1 > > > > > > > -- > > Daniel Vetter > > Software Engineer, Intel Corporation > > http://blog.ffwll.ch -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch