Received: by 2002:a25:4158:0:0:0:0:0 with SMTP id o85csp6719848yba; Wed, 1 May 2019 18:57:38 -0700 (PDT) X-Google-Smtp-Source: APXvYqy2WHwFqLddqpwAIDlR+Z5J465dDa+1y7zbWDYXrwrF4ndGP5XoNz9hIC7UR2dHD2kO9xOj X-Received: by 2002:aa7:8b8b:: with SMTP id r11mr1217117pfd.130.1556762258440; Wed, 01 May 2019 18:57:38 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1556762258; cv=none; d=google.com; s=arc-20160816; b=XnCdsWjcIUNFNnQoW3lOLq6qnkfkVlalc8xt5b8RC7N08eRGzhgzUmnGK85aydhK3q xhphQO7WwRAojzLL5jMpa5gvjVo4ixnHn0VE5c6ac+FYbmi44NInZ4x0Wg8oSCDJt/29 ZRPVXa8NL+X1N5t3C0ipF+5mLGpdwAF3sxNupy0mW+CGWrhts/15KLZngWPuQzrMSkZ+ eLM6VGhj/eryEOC7yqOmAtu2xo0sJ1rpAgm90j9nfG9qR5mqq2Ph3RfUQW2Xp2de77Ww zPPlMxcNpGy7U9MfW8yzCxCQ+ZzBA2JchKtCgpC3N3M93U50WIx0mCAazfxBS9ZfnFZQ X7yA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:to:subject:cc; bh=z9obC3SNT1GRoTjBLphBeumY32KtfiPG2L8Iskr4IlQ=; b=XHDbQXrc9f/+58w74v4T+xIuBXl4UYBD1oQkGPjimkgEjpXjQwW9cKJp15pnPBr9qf maIjmCXjGafy9myYYo1tteMn1IupycbAl4FcrktvX18ezdYmcGY8FNDK9cqm5X7tpHym 32r8L5Wk+uS5NEX3w/Hn3tV7lx+ShxNKuLY87MVPwMPuuIMeQmiuOtKDTh35Rc/Pagni aMhAqAz9kWm/jBQgDq0zxms5TtVGcNKG5cgj7hBGUYdcMMYtY88wGQJ7j/a4nR3ThKt/ Fnt+EEv0NaOjWKdmb0Dd72dSTCCoO+PRXlFI9gvm+KZ8lTXMVd3OvQvpmKypO+yPterh 2obA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id u25si868715pfm.173.2019.05.01.18.57.23; Wed, 01 May 2019 18:57:38 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726327AbfEBByQ (ORCPT + 99 others); Wed, 1 May 2019 21:54:16 -0400 Received: from mga18.intel.com ([134.134.136.126]:61668 "EHLO mga18.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726152AbfEBByQ (ORCPT ); Wed, 1 May 2019 21:54:16 -0400 X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga008.fm.intel.com ([10.253.24.58]) by orsmga106.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 01 May 2019 18:54:15 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.60,419,1549958400"; d="scan'208";a="145299326" Received: from allen-box.sh.intel.com (HELO [10.239.159.136]) ([10.239.159.136]) by fmsmga008.fm.intel.com with ESMTP; 01 May 2019 18:54:12 -0700 Cc: baolu.lu@linux.intel.com, David Woodhouse , Joerg Roedel , ashok.raj@intel.com, jacob.jun.pan@intel.com, alan.cox@intel.com, kevin.tian@intel.com, mika.westerberg@linux.intel.com, pengfei.xu@intel.com, Konrad Rzeszutek Wilk , Marek Szyprowski , iommu@lists.linux-foundation.org, linux-kernel@vger.kernel.org Subject: Re: [PATCH v3 02/10] swiotlb: Factor out slot allocation and free To: Robin Murphy , Christoph Hellwig References: <20190421011719.14909-1-baolu.lu@linux.intel.com> <20190421011719.14909-3-baolu.lu@linux.intel.com> <20190422164555.GA31181@lst.de> <0c6e5983-312b-0d6b-92f5-64861cd6804d@linux.intel.com> <20190423061232.GB12762@lst.de> <20190424144532.GA21480@lst.de> <20190426150433.GA19930@lst.de> <93b3d627-782d-cae0-2175-77a5a8b3fe6e@linux.intel.com> <90182d27-5764-7676-8ca6-b2773a40cfe1@arm.com> <1361b6ab-c3cf-d8ab-5f6b-9d9b7797bf02@linux.intel.com> From: Lu Baolu Message-ID: <998eadf0-0435-1a6b-7234-71554d95bb70@linux.intel.com> Date: Thu, 2 May 2019 09:47:53 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Robin, On 4/30/19 5:53 PM, Robin Murphy wrote: > On 30/04/2019 03:02, Lu Baolu wrote: >> Hi Robin, >> >> On 4/29/19 7:06 PM, Robin Murphy wrote: >>> On 29/04/2019 06:10, Lu Baolu wrote: >>>> Hi Christoph, >>>> >>>> On 4/26/19 11:04 PM, Christoph Hellwig wrote: >>>>> On Thu, Apr 25, 2019 at 10:07:19AM +0800, Lu Baolu wrote: >>>>>> This is not VT-d specific. It's just how generic IOMMU works. >>>>>> >>>>>> Normally, IOMMU works in paging mode. So if a driver issues DMA with >>>>>> IOVA  0xAAAA0123, IOMMU can remap it with a physical address >>>>>> 0xBBBB0123. >>>>>> But we should never expect IOMMU to remap 0xAAAA0123 with physical >>>>>> address of 0xBBBB0000. That's the reason why I said that IOMMU >>>>>> will not >>>>>> work there. >>>>> >>>>> Well, with the iommu it doesn't happen.  With swiotlb it obviosuly >>>>> can happen, so drivers are fine with it.  Why would that suddenly >>>>> become an issue when swiotlb is called from the iommu code? >>>>> >>>> >>>> I would say IOMMU is DMA remapping, not DMA engine. :-) >>> >>> I'm not sure I really follow the issue here - if we're copying the >>> buffer to the bounce page(s) there's no conceptual difference from >>> copying it to SWIOTLB slot(s), so there should be no need to worry >>> about the original in-page offset. >>> >>>  From the reply up-thread I guess you're trying to include an >>> optimisation to only copy the head and tail of the buffer if it spans >>> multiple pages, and directly map the ones in the middle, but AFAICS >>> that's going to tie you to also using strict mode for TLB >>> maintenance, which may not be a win overall depending on the balance >>> between invalidation bandwidth vs. memcpy bandwidth. At least if we >>> use standard SWIOTLB logic to always copy the whole thing, we should >>> be able to release the bounce pages via the flush queue to allow >>> 'safe' lazy unmaps. >>> >> >> With respect, even we use the standard SWIOTLB logic, we need to use >> the strict mode for TLB maintenance. >> >> Say, some swiotbl slots are used by untrusted device for bounce page >> purpose. When the device driver unmaps the IOVA, the slots are freed but >> the mapping is still cached in IOTLB, hence the untrusted device is >> still able to access the slots. Then the slots are allocated to other >> devices. This makes it possible for the untrusted device to access >> the data buffer of other devices. > > Sure, that's indeed how it would work right now - however since the > bounce pages will be freed and reused by the DMA API layer itself (at > the same level as the IOVAs) I see no technical reason why we couldn't > investigate deferred freeing as a future optimisation. Yes, agreed. Best regards, Lu Baolu