Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp7399013rwd; Tue, 20 Jun 2023 00:15:47 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ7WaJPEKpeHJd12DVPOD1b9uP7BEXj2vRh6yATD7nbHqc/4jV7kDLNY/oF+DeArqp7v/rKF X-Received: by 2002:a05:6358:ce03:b0:130:ea97:a225 with SMTP id gt3-20020a056358ce0300b00130ea97a225mr3695011rwb.11.1687245347610; Tue, 20 Jun 2023 00:15:47 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1687245347; cv=none; d=google.com; s=arc-20160816; b=YGqC70Gfl8BYxvtZJhKSKJdODUiUOO1zKM/qUHPzlCyDmCOM1LPXs8aGZB0qMUf71L SkZymeX+wRUXQsJJTRYPyjeGp/zHsSoISGmxPvsq1uzFMNzwVOVPbcTz8Xs5XxRSk8+X a//pkHuI/7bLoubBjCDFBWLmhUBfVgTLyNs/Kf0HQkYCxLlt6DbM7NFJC/honskNNJ6h 0v7TjEWXHNgETvN8cPjEG2DNFU+LgVBBsqDWFSrh9OcYyswQ3AIR6CuGtwYcqDUS+j+1 +ijcvFPpJG1gyEjYMGFBnyQ4fmDS/RYs8/pQZIBEDPoixznymjZw/Chj/W4JPdVsS0Rv 32Qw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:cc:to:subject :message-id:date:from:in-reply-to:references:mime-version :dkim-signature; bh=1vO0UNC8w8ZpBBQjuvFlINq9YpFy6EiHM213PIbi1oI=; b=bk2U7fOsqAYiXHWrupiBvWMr2e1NtwrZbV/hGwZVYg7VsTsRHO41YVSdSuWkEOiydv hhVgPJstZ/zqvpNCxRfbWafqdNyVvPxRMseMChzY/VJ2MmhYJBuyB9zJZrJ33lxwtwVs +ckTyWJlBWHm7/9G4WZFn4+jYHd9qkd5YQneR4G9/IJGIpf6dj5Tq3xR/or3b1UEKtEa LVDR3oS/AoczVCJr8TRz3YqvAYB6+tpF9Auz0l+lICLXKdrBd7v9QTeGTOmwz4FP3Tg9 3dIDvQ4rtUh2Ln6Om+RhcCekNeQ6e7spLy7TuiJvprV/FipK7muKMbLMO8Rpzb5Lv/QH z6Dg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=rsWYiAT6; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id i30-20020a63585e000000b0054fe2dfaab0si1096453pgm.893.2023.06.20.00.15.35; Tue, 20 Jun 2023 00:15:47 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=rsWYiAT6; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230448AbjFTHGi (ORCPT + 99 others); Tue, 20 Jun 2023 03:06:38 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54674 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229927AbjFTHGh (ORCPT ); Tue, 20 Jun 2023 03:06:37 -0400 Received: from dfw.source.kernel.org (dfw.source.kernel.org [139.178.84.217]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id CFD7CCC; Tue, 20 Jun 2023 00:06:34 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id 553BE61038; Tue, 20 Jun 2023 07:06:34 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id B4396C43397; Tue, 20 Jun 2023 07:06:33 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1687244793; bh=bSNSgEaGjfvi+rMEgSHVAqDDtfr9nRQYjTdYbwOu8zU=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=rsWYiAT6Liy3QwrdbqoSbJw9T8Zup/p/l3gyE0lIkCqgDNqcJPYwRu2780f1Xh0wk OLCJYeQ3q97lCtZVdxoFeDc6Y0F1JvvLGzXqEO5tDkaquwRH0UXKnErSSj+gA4MB35 TRQDUAfFKodCf9+sSsCRCNN5znndYB/t1PIlHshLCfVhyYRc7qTXup5mQs2WSeqJxw dMsztSS7quhAVJouCNGJUp2cU7rHgB+Nb6Awqt7y0X6bjhxBgmX+w0ntiNEJnl8yLo n6qi32AuAhGNMfwS3NNQz/hkAKcno+w2jEAQYYrd642nUzJY4LbI+DuNy7r/pa1MsO d73zXCKHKsLpA== Received: by mail-qt1-f178.google.com with SMTP id d75a77b69052e-3f9b1f43bd0so45109841cf.0; Tue, 20 Jun 2023 00:06:33 -0700 (PDT) X-Gm-Message-State: AC+VfDyTYBPlIub22AWEALLGA8kcxQscrS42KirolCBoHJ7Up5vHd7yK 5rmDqKy0KO4BIFDt9VZ6xkBR4bSsCyxcDh3DtZs= X-Received: by 2002:a05:622a:1710:b0:3ff:2164:a5c8 with SMTP id h16-20020a05622a171000b003ff2164a5c8mr4055386qtk.22.1687244792472; Tue, 20 Jun 2023 00:06:32 -0700 (PDT) MIME-Version: 1.0 References: <20230620004217.4700-1-dakr@redhat.com> <2c92bae3-0003-3c53-8ef1-6e12e5413995@redhat.com> In-Reply-To: From: Oded Gabbay Date: Tue, 20 Jun 2023 10:06:06 +0300 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [PATCH drm-next v5 00/14] [RFC] DRM GPUVA Manager & Nouveau VM_BIND UAPI To: Dave Airlie Cc: Danilo Krummrich , Donald Robson , daniel@ffwll.ch, tzimmermann@suse.de, mripard@kernel.org, corbet@lwn.net, christian.koenig@amd.com, bskeggs@redhat.com, Liam.Howlett@oracle.com, matthew.brost@intel.com, boris.brezillon@collabora.com, alexdeucher@gmail.com, bagasdotme@gmail.com, willy@infradead.org, jason@jlekstrand.net, dri-devel@lists.freedesktop.org, nouveau@lists.freedesktop.org, linux-doc@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Spam-Status: No, score=-7.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Jun 20, 2023 at 7:05=E2=80=AFAM Dave Airlie wro= te: > > Since this is feature is nouveau only currently and doesn't disturb > the current nouveau code paths, I'd like to try and get this work in > tree so other drivers can work from it. > > If there are any major objections to this, I'm happy to pull it back > out again, but I'd like to get some acks/rb in the next couple of days > in order to land some of it. > > Dave. > > > > > > forgot to add your email address to the patch series - sorry about that= . > > > > This series (v5) contains the Documentation changes you requested. > > > > - Danilo > > > > On 6/20/23 02:42, Danilo Krummrich wrote: > > > This patch series provides a new UAPI for the Nouveau driver in order= to > > > support Vulkan features, such as sparse bindings and sparse residency= . > > > > > > Furthermore, with the DRM GPUVA manager it provides a new DRM core fe= ature to > > > keep track of GPU virtual address (VA) mappings in a more generic way= . > > > > > > The DRM GPUVA manager is indented to help drivers implement userspace= -manageable > > > GPU VA spaces in reference to the Vulkan API. In order to achieve thi= s goal it > > > serves the following purposes in this context. > > > > > > 1) Provide infrastructure to track GPU VA allocations and mappin= gs, > > > making use of the maple_tree. > > > > > > 2) Generically connect GPU VA mappings to their backing buffers,= in > > > particular DRM GEM objects. Will this manager be able to connect GPU VA mappings to host memory allocations (aka user pointers) ? I only skimmed over the uapi definitions, but from that quick glance I saw you can only pass a (gem) handle to the vm bind uapi. I think it is an important feature because you don't want to have two GPU VA managers running in your driver (if that's even possible). Maybe we should at least try to make sure the uapi is/will be compatible with such an extension. Oded Oded > > > > > > 3) Provide a common implementation to perform more complex mappi= ng > > > operations on the GPU VA space. In particular splitting and m= erging > > > of GPU VA mappings, e.g. for intersecting mapping requests or= partial > > > unmap requests. > > > > > > The new VM_BIND Nouveau UAPI build on top of the DRM GPUVA manager, i= tself > > > providing the following new interfaces. > > > > > > 1) Initialize a GPU VA space via the new DRM_IOCTL_NOUVEAU_VM_IN= IT ioctl > > > for UMDs to specify the portion of VA space managed by the ke= rnel and > > > userspace, respectively. > > > > > > 2) Allocate and free a VA space region as well as bind and unbin= d memory > > > to the GPUs VA space via the new DRM_IOCTL_NOUVEAU_VM_BIND io= ctl. > > > > > > 3) Execute push buffers with the new DRM_IOCTL_NOUVEAU_EXEC ioct= l. > > > > > > Both, DRM_IOCTL_NOUVEAU_VM_BIND and DRM_IOCTL_NOUVEAU_EXEC, make use = of the DRM > > > scheduler to queue jobs and support asynchronous processing with DRM = syncobjs > > > as synchronization mechanism. > > > > > > By default DRM_IOCTL_NOUVEAU_VM_BIND does synchronous processing, > > > DRM_IOCTL_NOUVEAU_EXEC supports asynchronous processing only. > > > > > > The new VM_BIND UAPI for Nouveau makes also use of drm_exec (executio= n context > > > for GEM buffers) by Christian K=C3=B6nig. Since the patch implementin= g drm_exec was > > > not yet merged into drm-next it is part of this series, as well as a = small fix > > > for this patch, which was found while testing this series. > > > > > > This patch series is also available at [1]. > > > > > > There is a Mesa NVK merge request by Dave Airlie [2] implementing the > > > corresponding userspace parts for this series. > > > > > > The Vulkan CTS test suite passes the sparse binding and sparse reside= ncy test > > > cases for the new UAPI together with Dave's Mesa work. > > > > > > There are also some test cases in the igt-gpu-tools project [3] for t= he new UAPI > > > and hence the DRM GPU VA manager. However, most of them are testing t= he DRM GPU > > > VA manager's logic through Nouveau's new UAPI and should be considere= d just as > > > helper for implementation. > > > > > > However, I absolutely intend to change those test cases to proper kun= it test > > > cases for the DRM GPUVA manager, once and if we agree on it's usefuln= ess and > > > design. > > > > > > [1] https://gitlab.freedesktop.org/nouvelles/kernel/-/tree/new-uapi-d= rm-next / > > > https://gitlab.freedesktop.org/nouvelles/kernel/-/merge_requests= /1 > > > [2] https://gitlab.freedesktop.org/nouveau/mesa/-/merge_requests/150/ > > > [3] https://gitlab.freedesktop.org/dakr/igt-gpu-tools/-/tree/wip_nouv= eau_vm_bind > > > > > > Changes in V2: > > > =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D > > > Nouveau: > > > - Reworked the Nouveau VM_BIND UAPI to avoid memory allocations = in fence > > > signalling critical sections. Updates to the VA space are spli= t up in three > > > separate stages, where only the 2. stage executes in a fence s= ignalling > > > critical section: > > > > > > 1. update the VA space, allocate new structures and page tab= les > > > 2. (un-)map the requested memory bindings > > > 3. free structures and page tables > > > > > > - Separated generic job scheduler code from specific job impleme= ntations. > > > - Separated the EXEC and VM_BIND implementation of the UAPI. > > > - Reworked the locking parts of the nvkm/vmm RAW interface, such= that > > > (un-)map operations can be executed in fence signalling critic= al sections. > > > > > > GPUVA Manager: > > > - made drm_gpuva_regions optional for users of the GPUVA manager > > > - allow NULL GEMs for drm_gpuva entries > > > - swichted from drm_mm to maple_tree for track drm_gpuva / drm_g= puva_region > > > entries > > > - provide callbacks for users to allocate custom drm_gpuva_op st= ructures to > > > allow inheritance > > > - added user bits to drm_gpuva_flags > > > - added a prefetch operation type in order to support generating= prefetch > > > operations in the same way other operations generated > > > - hand the responsibility for mutual exclusion for a GEM's > > > drm_gpuva list to the user; simplified corresponding (un-)link= functions > > > > > > Maple Tree: > > > - I added two maple tree patches to the series, one to support c= ustom tree > > > walk macros and one to hand the locking responsibility to the = user of the > > > GPUVA manager without pre-defined lockdep checks. > > > > > > Changes in V3: > > > =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D > > > Nouveau: > > > - Reworked the Nouveau VM_BIND UAPI to do the job cleanup (inclu= ding page > > > table cleanup) within a workqueue rather than the job_free() c= allback of > > > the scheduler itself. A job_free() callback can stall the exec= ution (run() > > > callback) of the next job in the queue. Since the page table c= leanup > > > requires to take the same locks as need to be taken for page t= able > > > allocation, doing it directly in the job_free() callback would= still > > > violate the fence signalling critical path. > > > - Separated Nouveau fence allocation and emit, such that we do n= ot violate > > > the fence signalling critical path in EXEC jobs. > > > - Implement "regions" (for handling sparse mappings through PDEs= and dual > > > page tables) within Nouveau. > > > - Drop the requirement for every mapping to be contained within = a region. > > > - Add necassary synchronization of VM_BIND job operation sequenc= es in order > > > to work around limitations in page table handling. This will b= e addressed > > > in a future re-work of Nouveau's page table handling. > > > - Fixed a couple of race conditions found through more testing. = Thanks to > > > Dave for consitently trying to break it. :-) > > > > > > GPUVA Manager: > > > - Implement pre-allocation capabilities for tree modifications w= ithin fence > > > signalling critical sections. > > > - Implement accessors to to apply tree modification while walkin= g the GPUVA > > > tree in order to actually support processing of drm_gpuva_ops = through > > > callbacks in fence signalling critical sections rather than th= rough > > > pre-allocated operation lists. > > > - Remove merging of GPUVAs; the kernel has limited to none knowl= ege about > > > the semantics of mapping sequences. Hence, merging is purely s= peculative. > > > It seems that gaining a significant (or at least a measurable)= performance > > > increase through merging is way more likely to happen when use= rspace is > > > responsible for merging mappings up to the next larger page si= ze if > > > possible. > > > - Since merging was removed, regions pretty much loose their rig= ht to exist. > > > They might still be useful for handling dual page tables or si= milar > > > mechanisms, but since Nouveau seems to be the only driver havi= ng a need > > > for this for now, regions were removed from the GPUVA manager. > > > - Fixed a couple of maple_tree related issues; thanks to Liam fo= r helping me > > > out. > > > > > > Changes in V4: > > > =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D > > > Nouveau: > > > - Refactored how specific VM_BIND and EXEC jobs are created and = how their > > > arguments are passed to the generic job implementation. > > > - Fixed a UAF race condition where bind job ops could have been = freed > > > already while still waiting for a job cleanup to finish. This = is due to > > > in certain cases we need to wait for mappings actually being u= nmapped > > > before creating sparse regions in the same area. > > > - Re-based the code onto drm_exec v4 patch. > > > > > > GPUVA Manager: > > > - Fixed a maple tree related bug when pre-allocating MA states. > > > (Boris Brezillion) > > > - Made struct drm_gpuva_fn_ops a const object in all occurrences= . > > > (Boris Brezillion) > > > > > > Changes in V5: > > > =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D > > > Nouveau: > > > - Link and unlink GPUVAs outside the fence signalling critical p= ath in > > > nouveau_uvmm_bind_job_submit() holding the dma-resv lock. Mutu= al exclusion > > > of BO evicts causing mapping invalidation and regular mapping = operations > > > is ensured with dma-fences. > > > > > > GPUVA Manager: > > > - Removed the separate GEMs GPUVA list lock. Link and unlink as = well as > > > iterating the GEM's GPUVA list should be protected with the GE= M's dma-resv > > > lock instead. > > > - Renamed DRM_GPUVA_EVICTED flag to DRM_GPUVA_INVALIDATED. Mappi= ngs do not > > > get eviced, they might get invalidated due to eviction. > > > - Maple tree uses the 'unsinged long' type for node entries. Whi= le this > > > works for GPU VA spaces larger than 32-bit on 64-bit kernel, t= he GPU VA > > > space is limited to 32-bit on 32-bit kernels as well. > > > As long as we do not have a 64-bit capable maple tree for 32-b= it kernels, > > > the GPU VA manager contains checks to throw warnings when GPU = VA entries > > > exceed the maple tree's storage capabilities. > > > - Extended the Documentation and added example code as requested= by Donald > > > Robson. > > > > > > Christian K=C3=B6nig (1): > > > drm: execution context for GEM buffers v4 > > > > > > Danilo Krummrich (13): > > > maple_tree: split up MA_STATE() macro > > > drm: manager to keep track of GPUs VA mappings > > > drm: debugfs: provide infrastructure to dump a DRM GPU VA space > > > drm/nouveau: new VM_BIND uapi interfaces > > > drm/nouveau: get vmm via nouveau_cli_vmm() > > > drm/nouveau: bo: initialize GEM GPU VA interface > > > drm/nouveau: move usercopy helpers to nouveau_drv.h > > > drm/nouveau: fence: separate fence alloc and emit > > > drm/nouveau: fence: fail to emit when fence context is killed > > > drm/nouveau: chan: provide nouveau_channel_kill() > > > drm/nouveau: nvkm/vmm: implement raw ops to manage uvmm > > > drm/nouveau: implement new VM_BIND uAPI > > > drm/nouveau: debugfs: implement DRM GPU VA debugfs > > > > > > Documentation/gpu/driver-uapi.rst | 11 + > > > Documentation/gpu/drm-mm.rst | 54 + > > > drivers/gpu/drm/Kconfig | 6 + > > > drivers/gpu/drm/Makefile | 3 + > > > drivers/gpu/drm/drm_debugfs.c | 41 + > > > drivers/gpu/drm/drm_exec.c | 278 +++ > > > drivers/gpu/drm/drm_gem.c | 3 + > > > drivers/gpu/drm/drm_gpuva_mgr.c | 1971 ++++++++++++++= ++ > > > drivers/gpu/drm/nouveau/Kbuild | 3 + > > > drivers/gpu/drm/nouveau/Kconfig | 2 + > > > drivers/gpu/drm/nouveau/dispnv04/crtc.c | 9 +- > > > drivers/gpu/drm/nouveau/include/nvif/if000c.h | 26 +- > > > drivers/gpu/drm/nouveau/include/nvif/vmm.h | 19 +- > > > .../gpu/drm/nouveau/include/nvkm/subdev/mmu.h | 20 +- > > > drivers/gpu/drm/nouveau/nouveau_abi16.c | 24 + > > > drivers/gpu/drm/nouveau/nouveau_abi16.h | 1 + > > > drivers/gpu/drm/nouveau/nouveau_bo.c | 204 +- > > > drivers/gpu/drm/nouveau/nouveau_bo.h | 2 +- > > > drivers/gpu/drm/nouveau/nouveau_chan.c | 22 +- > > > drivers/gpu/drm/nouveau/nouveau_chan.h | 1 + > > > drivers/gpu/drm/nouveau/nouveau_debugfs.c | 39 + > > > drivers/gpu/drm/nouveau/nouveau_dmem.c | 9 +- > > > drivers/gpu/drm/nouveau/nouveau_drm.c | 27 +- > > > drivers/gpu/drm/nouveau/nouveau_drv.h | 94 +- > > > drivers/gpu/drm/nouveau/nouveau_exec.c | 418 ++++ > > > drivers/gpu/drm/nouveau/nouveau_exec.h | 54 + > > > drivers/gpu/drm/nouveau/nouveau_fence.c | 23 +- > > > drivers/gpu/drm/nouveau/nouveau_fence.h | 5 +- > > > drivers/gpu/drm/nouveau/nouveau_gem.c | 62 +- > > > drivers/gpu/drm/nouveau/nouveau_mem.h | 5 + > > > drivers/gpu/drm/nouveau/nouveau_prime.c | 2 +- > > > drivers/gpu/drm/nouveau/nouveau_sched.c | 461 ++++ > > > drivers/gpu/drm/nouveau/nouveau_sched.h | 123 + > > > drivers/gpu/drm/nouveau/nouveau_svm.c | 2 +- > > > drivers/gpu/drm/nouveau/nouveau_uvmm.c | 1979 ++++++++++++++= +++ > > > drivers/gpu/drm/nouveau/nouveau_uvmm.h | 107 + > > > drivers/gpu/drm/nouveau/nouveau_vmm.c | 4 +- > > > drivers/gpu/drm/nouveau/nvif/vmm.c | 100 +- > > > .../gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c | 213 +- > > > drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c | 197 +- > > > drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h | 25 + > > > .../drm/nouveau/nvkm/subdev/mmu/vmmgf100.c | 16 +- > > > .../drm/nouveau/nvkm/subdev/mmu/vmmgp100.c | 16 +- > > > .../gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c | 27 +- > > > include/drm/drm_debugfs.h | 25 + > > > include/drm/drm_drv.h | 6 + > > > include/drm/drm_exec.h | 119 + > > > include/drm/drm_gem.h | 52 + > > > include/drm/drm_gpuva_mgr.h | 682 ++++++ > > > include/linux/maple_tree.h | 7 +- > > > include/uapi/drm/nouveau_drm.h | 209 ++ > > > 51 files changed, 7566 insertions(+), 242 deletions(-) > > > create mode 100644 drivers/gpu/drm/drm_exec.c > > > create mode 100644 drivers/gpu/drm/drm_gpuva_mgr.c > > > create mode 100644 drivers/gpu/drm/nouveau/nouveau_exec.c > > > create mode 100644 drivers/gpu/drm/nouveau/nouveau_exec.h > > > create mode 100644 drivers/gpu/drm/nouveau/nouveau_sched.c > > > create mode 100644 drivers/gpu/drm/nouveau/nouveau_sched.h > > > create mode 100644 drivers/gpu/drm/nouveau/nouveau_uvmm.c > > > create mode 100644 drivers/gpu/drm/nouveau/nouveau_uvmm.h > > > create mode 100644 include/drm/drm_exec.h > > > create mode 100644 include/drm/drm_gpuva_mgr.h > > > > > > > > > base-commit: 2222dcb0775d36de28992f56455ab3967b30d380 > >