Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751371AbdGQMtg (ORCPT ); Mon, 17 Jul 2017 08:49:36 -0400 Received: from foss.arm.com ([217.140.101.70]:38806 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751290AbdGQMtf (ORCPT ); Mon, 17 Jul 2017 08:49:35 -0400 Subject: Re: What differences and relations between SVM, HSA, HMM and Unified Memory? To: Yisheng Xie , "Wuzongyong (Cordius Wu, Euler Dept)" , "iommu@lists.linux-foundation.org" , "linux-kernel@vger.kernel.org" Cc: "Wanzongshun (Vincent)" , "oded.gabbay@amd.com" , liubo95@huawei.com References: <9BD73EA91F8E404F851CF3F519B14AA8CE753F@SZXEMI503-MBS.china.huawei.com> <20c9cdd5-5118-f916-d8ad-70b7c1434d73@arm.com> <1c4f4fb0-7201-ed4c-aa88-4d7e2369238e@huawei.com> From: Jean-Philippe Brucker Message-ID: <84d5df57-f89c-6c5e-d6b2-26d6943c2217@arm.com> Date: Mon, 17 Jul 2017 13:52:46 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:52.0) Gecko/20100101 Thunderbird/52.1.1 MIME-Version: 1.0 In-Reply-To: <1c4f4fb0-7201-ed4c-aa88-4d7e2369238e@huawei.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3491 Lines: 81 On 17/07/17 12:57, Yisheng Xie wrote: > Hi Jean-Philippe, > > On 2017/6/12 19:37, Jean-Philippe Brucker wrote: >> Hello, >> >> On 10/06/17 05:06, Wuzongyong (Cordius Wu, Euler Dept) wrote: >>> Hi, >>> >>> Could someone explain differences and relations between the SVM(Shared >>> Virtual Memory, by Intel), HSA(Heterogeneous System Architecture, by AMD), >>> HMM(Heterogeneous Memory Management, by Glisse) and UM(Unified Memory, by >>> NVIDIA) ? Are these in the substitutional relation? >>> >>> As I understand it, these aim to solve the same thing, sharing pointers >>> between CPU and GPU(implement with ATS/PASID/PRI/IOMMU support). So far, >>> SVM and HSA can only be used by integrated gpu. And, Intel declare that >>> the root ports doesn’t not have the required TLP prefix support, resulting >>> that SVM can’t be used by discrete devices. So could someone tell me the >>> required TLP prefix means what specifically?> >>> With HMM, we can use allocator like malloc to manage host and device >>> memory. Does this mean that there is no need to use SVM and HSA with HMM, >>> or HMM is the basis of SVM and HAS to implement Fine-Grained system SVM >>> defined in the opencl spec? >> >> I can't provide an exhaustive answer, but I have done some work on SVM. >> Take it with a grain of salt though, I am not an expert. >> >> * HSA is an architecture that provides a common programming model for CPUs >> and accelerators (GPGPUs etc). It does have SVM requirement (I/O page >> faults, PASID and compatible address spaces), though it's only a small >> part of it. >> >> * Similarly, OpenCL provides an API for dealing with accelerators. OpenCL >> 2.0 introduced the concept of Fine-Grained System SVM, which allows to >> pass userspace pointers to devices. It is just one flavor of SVM, they >> also have coarse-grained and non-system. But they might have coined the >> name, and I believe that in the context of Linux IOMMU, when we talk about >> "SVM" it is OpenCL's fine-grained system SVM. >> [...] >> >> While SVM is only about virtual address space, > As you mentioned, SVM is only about virtual address space, I'd like to know how to > manage the physical address especially about device's RAM, before HMM? > > When OpenCL alloc a SVM pointer like: > void* p = clSVMAlloc ( > context, // an OpenCL context where this buffer is available > CL_MEM_READ_WRITE | CL_MEM_SVM_FINE_GRAIN_BUFFER, > size, // amount of memory to allocate (in bytes) > 0 // alignment in bytes (0 means default) > ); > > where this RAM come from, device RAM or host RAM? Sorry, I'm not familiar with OpenCL/GPU drivers. It is up to them to decide where to allocate memory for clSVMAlloc. My SMMU work would deal with fine-grained *system* SVM, the kind that can be obtained from malloc and doesn't require a call to clSVMAlloc. Hopefully others on this list or linux-mm might be able to help you. Thanks, Jean > Thanks > Yisheng Xie > >> HMM deals with physical >> storage. If I understand correctly, HMM allows to transparently use device >> RAM from userspace applications. So upon an I/O page fault, the mm >> subsystem will migrate data from system memory into device RAM. It would >> differ from "pure" SVM in that you would use different page directories on >> IOMMU and MMU sides, and synchronize them using MMU notifiers. But please >> don't take this at face value, I haven't had time to look into HMM yet. >> >> Thanks, >> Jean >> >> . >> >