Received: by 2002:a05:6a10:16a7:0:0:0:0 with SMTP id gp39csp2073367pxb; Sat, 14 Nov 2020 12:12:31 -0800 (PST) X-Google-Smtp-Source: ABdhPJz4PX5lsjaxw4s1yQK7CW6iP2EgSEHsn17xhJNUJ5lcAsdnAcvS0WYrYxdzuFbug0EFLtuN X-Received: by 2002:aa7:d801:: with SMTP id v1mr8668243edq.250.1605384751257; Sat, 14 Nov 2020 12:12:31 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1605384751; cv=none; d=google.com; s=arc-20160816; b=U8HYqPul/t0zMjhgHuqky9xFOrgoEHkwrks2ymL0sUiEfYBnDFMNOFWaYFU9YzWgOX Hzktyz3u8mZpVi0CZnmsaKCoX1Yt0lIZIuNe8pDaLffhGjY9iYDNFLasldsqBZHIHqFk ashojru3/RA4M+paD7sW1S8YO0MuwerxA6+mtzLw4rh8mgUPQfvmSiKxKRvBzX5jO6sa Z34wbtoTUL2NVTNBBVJs0+GiK+2DusKWXFy8ZzEVCOj+4Pp7VnQ/lSTqdR9+Fys4Iv0+ N0GJDE8uun1Yhhr3vf0YrTPG335Frf/5QXXd11avEiqFZ9t4EBo/AbzbUMYs5f9+iFBY 6sOw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject:dkim-signature; bh=67JIrhUexBLk+euYcRPybCUujGZ256Wpe49XxF/Od8o=; b=0Y39nGxijTg1iHq8W5mVbgaqOYabGnHsgIKkBQO+5Jq8t7hUZpbw44pGWZjdkCVaba aAAaMPu7nLoTb5WPp+TXO2BLU+tXprSJL7e52MoEisZOQJJq997qVSQ55Il17v14BCc7 fhgcTwj2HmVCtJHxNes2QshY38cT1NwILUFkqZNhNLv0bObTVtpNSqTENYU6MM5kGsCO gvrUSV9p+2vACKEa1JDOMZ9ypNRG2TC5a6vZkmfvdyCdV9LnxTo3VDmS8UEqn/irrjEy t+fzO7RB6Bx8bvJQc3a1s0TjYUr98hmzmpFfSUIcHw0alHD+f958esJbjJp6sGtGAPrq EFug== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@marek-ca.20150623.gappssmtp.com header.s=20150623 header.b=itpMD3co; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id v14si8767701edb.336.2020.11.14.12.12.08; Sat, 14 Nov 2020 12:12:31 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@marek-ca.20150623.gappssmtp.com header.s=20150623 header.b=itpMD3co; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726231AbgKNUKh (ORCPT + 99 others); Sat, 14 Nov 2020 15:10:37 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:51676 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726136AbgKNUKh (ORCPT ); Sat, 14 Nov 2020 15:10:37 -0500 Received: from mail-qv1-xf44.google.com (mail-qv1-xf44.google.com [IPv6:2607:f8b0:4864:20::f44]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 43AB8C0613D2 for ; Sat, 14 Nov 2020 12:10:37 -0800 (PST) Received: by mail-qv1-xf44.google.com with SMTP id y11so6731793qvu.10 for ; Sat, 14 Nov 2020 12:10:37 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marek-ca.20150623.gappssmtp.com; s=20150623; h=subject:to:cc:references:from:message-id:date:user-agent :mime-version:in-reply-to:content-language:content-transfer-encoding; bh=67JIrhUexBLk+euYcRPybCUujGZ256Wpe49XxF/Od8o=; b=itpMD3coD5rHj6pDvemIHpnPf4BB39MY7jtbuBua/ljNHEQhqlwauoK9p3NxrAmPlH tM1OmSMW5cPdEWV8j5GRMzgTXEvw7OF8+koGttFDhDJAEvy/nEVWtPGba1Pfxkf6rqQA xoh67VBTeWFhcjmWz6O70tbT6C91O+JHHO1RwL0HN1pYW2tm265UH57Gr2xCb7jwZ/MW 91omkj8gLtx3SEJVzCYmCQ3uR6lAX5pD6eGbrv1MVSQZBhvUKuGwljJ4D0JvsGXAMN/I 1Ogm4wRS+XrQ8Riey+hd9ziYiYF1dr7cWHeYsI40B5poW8sQE5wNRX4wp+ma3w9yVACI npZQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=67JIrhUexBLk+euYcRPybCUujGZ256Wpe49XxF/Od8o=; b=XK/5JIr3KzXkiGWDpO4YyCR1qluieQ6pbqTxyW29YKoMSyTySTiP4favxVTaV22AJb 8rGC+Nwr9NzoCztGPFfujFApwlReF7Y1NE6DEyndbITwYL+MhCCJqpIoaUGNFXTnV3yK KmgIEfGsLn4VeE0hfsEShER8DAMQSKKG3lFW9wK+cUg9cFkd0XJiZCl/LDIWBpaKhh3a cdQkCphkxMb5CrZdTdXq8jMROSFYYJIxahU3eMB21n+lRHLJOcyut8A960cVKiLIZ6jc gwUz6alZzPyyZ7rBk1szXeGiCFF7tJ/2NvR1iWTSHfcB3stvJoLoaF6R/fLWxgnbHwgu Q6Ow== X-Gm-Message-State: AOAM530nvS3eb2sxTSszFZKG+hDc7U5VPiuGMyNLimqsOWCiaIFKU7BW G/7pHT1RENSxdH8n3Dt8dMlldFiEr39kZ2E1yYY= X-Received: by 2002:ad4:4bb2:: with SMTP id i18mr8530477qvw.52.1605384635788; Sat, 14 Nov 2020 12:10:35 -0800 (PST) Received: from [192.168.0.189] (modemcable068.184-131-66.mc.videotron.ca. [66.131.184.68]) by smtp.gmail.com with ESMTPSA id k31sm8923935qtd.40.2020.11.14.12.10.34 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Sat, 14 Nov 2020 12:10:35 -0800 (PST) Subject: Re: [RESEND PATCH v2 4/5] drm/msm: add DRM_MSM_GEM_SYNC_CACHE for non-coherent cache maintenance To: Rob Clark Cc: Christoph Hellwig , freedreno , Sean Paul , David Airlie , Daniel Vetter , "open list:DRM DRIVER FOR MSM ADRENO GPU" , "open list:DRM DRIVER FOR MSM ADRENO GPU" , open list References: <20201114151717.5369-1-jonathan@marek.ca> <20201114151717.5369-5-jonathan@marek.ca> <20201114162406.GC24411@lst.de> <50ddcadb-c630-2ef6-cdc4-724d9823fba7@marek.ca> From: Jonathan Marek Message-ID: Date: Sat, 14 Nov 2020 15:07:20 -0500 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.2.2 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 11/14/20 2:39 PM, Rob Clark wrote: > On Sat, Nov 14, 2020 at 10:58 AM Jonathan Marek wrote: >> >> On 11/14/20 1:46 PM, Rob Clark wrote: >>> On Sat, Nov 14, 2020 at 8:24 AM Christoph Hellwig wrote: >>>> >>>> On Sat, Nov 14, 2020 at 10:17:12AM -0500, Jonathan Marek wrote: >>>>> +void msm_gem_sync_cache(struct drm_gem_object *obj, uint32_t flags, >>>>> + size_t range_start, size_t range_end) >>>>> +{ >>>>> + struct msm_gem_object *msm_obj = to_msm_bo(obj); >>>>> + struct device *dev = msm_obj->base.dev->dev; >>>>> + >>>>> + /* exit early if get_pages() hasn't been called yet */ >>>>> + if (!msm_obj->pages) >>>>> + return; >>>>> + >>>>> + /* TODO: sync only the specified range */ >>>>> + >>>>> + if (flags & MSM_GEM_SYNC_FOR_DEVICE) { >>>>> + dma_sync_sg_for_device(dev, msm_obj->sgt->sgl, >>>>> + msm_obj->sgt->nents, DMA_TO_DEVICE); >>>>> + } >>>>> + >>>>> + if (flags & MSM_GEM_SYNC_FOR_CPU) { >>>>> + dma_sync_sg_for_cpu(dev, msm_obj->sgt->sgl, >>>>> + msm_obj->sgt->nents, DMA_FROM_DEVICE); >>>>> + } >>>> >>>> Splitting this helper from the only caller is rather strange, epecially >>>> with the two unused arguments. And I think the way this is specified >>>> to take a range, but ignoring it is actively dangerous. User space will >>>> rely on it syncing everything sooner or later and then you are stuck. >>>> So just define a sync all primitive for now, and if you really need a >>>> range sync and have actually implemented it add a new ioctl for that. >>> >>> We do already have a split of ioctl "layer" which enforces valid ioctl >>> params, etc, and gem (or other) module code which is called by the >>> ioctl func. So I think it is fine to keep this split here. (Also, I >>> think at some point there will be a uring type of ioctl alternative >>> which would re-use the same gem func.) >>> >>> But I do agree that the range should be respected or added later.. >>> drm_ioctl() dispatch is well prepared for extending ioctls. >>> >>> And I assume there should be some validation that the range is aligned >>> to cache-line? Or can we flush a partial cache line? >>> >> >> The range is intended to be "sync at least this range", so that >> userspace doesn't have to worry about details like that. >> > > I don't think userspace can *not* worry about details like that. > Consider a case where the cpu and gpu are simultaneously accessing > different parts of a buffer (for ex, sub-allocation). There needs to > be cache-line separation between the two. > Right.. and it also seems like we can't get away with just flushing/invalidating the whole thing. qcom's vulkan driver has nonCoherentAtomSize=1, and it looks like dma_sync_single_for_cpu() does deal in some way with the partial cache line case, although I'm not sure that means we can have a nonCoherentAtomSize=1. > BR, > -R >