Received: by 2002:a05:6a10:8c0a:0:0:0:0 with SMTP id go10csp427815pxb; Tue, 9 Feb 2021 04:08:50 -0800 (PST) X-Google-Smtp-Source: ABdhPJwvAOH5oeUsFKdohoagJwuUJ47Q8DK9gdC4hsm5vRKpcM7uw4l+ES7j6/IqT/2IGl5B166K X-Received: by 2002:a50:fe02:: with SMTP id f2mr21856435edt.195.1612872530639; Tue, 09 Feb 2021 04:08:50 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1612872530; cv=none; d=google.com; s=arc-20160816; b=wQ+e1RSVUInrcXtkXzH0DJvJlU6RAlwtZH3db4aDRuNUrWn0yvTWYa3zZLYuCaIN5k AFwtLYNAAsyPk1TRRc4bb548bLOyYbuyNAGEhu3FA/1+g+5bEa2QVpmLFQJAKFuyA2zY AgBuo6jdHM28IcL5RXi4JOlzuKfAI4ovydqR9+bIIeUdDMjrqS8ia9pE5VdiU3svygzP QVsoAAljTjQuCCAQIJOqkK7KB6yG+wrOsshBgd4aeXJjgV/6DRgi20fVbmvN5Hq3S6WN rLWXgdBbWNuxPDDyEQoKfMhlKV+XIJwdwl28d6rd13To+xZMOHq86AsyQGild5Nu/TCr B/BA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :ironport-sdr:ironport-sdr; bh=BwFUD54ooR/qEeFWNOQUvky/YO+6MNMojDvBpKud6d4=; b=MgVeRdwRDHQK33wBnA0RSWP9oJ4939tlmXYy0XReiU7kmXiPo3hdkdxgG19MfTKGBI eRuB/j53kMPOCCPGUBJpCdy4nu0gnuOESjxNV7NRyn+0FP80IcVwYD4/cD6X01JcYvie P5Nk9V3r87Y8iBpf+VhSHAjgKdJRUMj/lHkMVG6iKbaIJf9WN6md+3Ph61rw776cGNlU iQSefVnIwII1+gx0fNciPFmsMPVDhZhd43XrWTGV6JXfpYaLTGjS16bd0g8G+so9/7lX rTKiGg3cE/8qaTqM7SKWXuVrMeRCTClxSFUcrKuqzXlhWrxciYW4ipAJ7GQM6m1PutAI 6klQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id y1si14805855edp.72.2021.02.09.04.08.25; Tue, 09 Feb 2021 04:08:50 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229766AbhBIMGx (ORCPT + 99 others); Tue, 9 Feb 2021 07:06:53 -0500 Received: from mga09.intel.com ([134.134.136.24]:42985 "EHLO mga09.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229558AbhBIMFU (ORCPT ); Tue, 9 Feb 2021 07:05:20 -0500 IronPort-SDR: Zkke3K6ZRBiGI6nMCMcIQ9TCibpcKJN8455RqHcBwmzO5Gc2SSgp8aMT1QIBXZUm1j2PFVQeuo BfTbjNLQ0hmw== X-IronPort-AV: E=McAfee;i="6000,8403,9889"; a="182012859" X-IronPort-AV: E=Sophos;i="5.81,164,1610438400"; d="scan'208";a="182012859" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga102.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 09 Feb 2021 04:03:16 -0800 IronPort-SDR: 0ieoaeAKU7d3c11pykydksMQ7vj3nDDcGmtrF0t/KBBgWNh9lq8KKHEiw0mCbglS6xKJI2v75a mLYjV2Vco98Q== X-IronPort-AV: E=Sophos;i="5.81,164,1610438400"; d="scan'208";a="396093725" Received: from yisun1-ubuntu.bj.intel.com (HELO yi.y.sun) ([10.238.156.116]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-SHA256; 09 Feb 2021 04:03:10 -0800 Date: Tue, 9 Feb 2021 19:57:44 +0800 From: Yi Sun To: Keqian Zhu Cc: linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org, kvm@vger.kernel.org, kvmarm@lists.cs.columbia.edu, iommu@lists.linux-foundation.org, Will Deacon , Alex Williamson , Marc Zyngier , Catalin Marinas , Kirti Wankhede , Cornelia Huck , Mark Rutland , James Morse , Robin Murphy , Suzuki K Poulose , wanghaibin.wang@huawei.com, jiangkunkun@huawei.com, yuzenghui@huawei.com, lushenming@huawei.com, kevin.tian@intel.com, yan.y.zhao@intel.com, baolu.lu@linux.intel.com Subject: Re: [RFC PATCH 10/11] vfio/iommu_type1: Optimize dirty bitmap population based on iommu HWDBM Message-ID: <20210209115744.GB28580@yi.y.sun> References: <20210128151742.18840-1-zhukeqian1@huawei.com> <20210128151742.18840-11-zhukeqian1@huawei.com> <20210207095630.GA28580@yi.y.sun> <407d28db-1f86-8d4f-ab15-3c3ac56bbe7f@huawei.com> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline In-Reply-To: <407d28db-1f86-8d4f-ab15-3c3ac56bbe7f@huawei.com> User-Agent: Mutt/1.5.21 (2010-09-15) Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 21-02-07 18:40:36, Keqian Zhu wrote: > Hi Yi, > > On 2021/2/7 17:56, Yi Sun wrote: > > Hi, > > > > On 21-01-28 23:17:41, Keqian Zhu wrote: > > > > [...] > > > >> +static void vfio_dma_dirty_log_start(struct vfio_iommu *iommu, > >> + struct vfio_dma *dma) > >> +{ > >> + struct vfio_domain *d; > >> + > >> + list_for_each_entry(d, &iommu->domain_list, next) { > >> + /* Go through all domain anyway even if we fail */ > >> + iommu_split_block(d->domain, dma->iova, dma->size); > >> + } > >> +} > > > > This should be a switch to prepare for dirty log start. Per Intel > > Vtd spec, there is SLADE defined in Scalable-Mode PASID Table Entry. > > It enables Accessed/Dirty Flags in second-level paging entries. > > So, a generic iommu interface here is better. For Intel iommu, it > > enables SLADE. For ARM, it splits block. > Indeed, a generic interface name is better. > > The vendor iommu driver plays vendor's specific actions to start dirty log, and Intel iommu and ARM smmu may differ. Besides, we may add more actions in ARM smmu driver in future. > > One question: Though I am not familiar with Intel iommu, I think it also should split block mapping besides enable SLADE. Right? > I am not familiar with ARM smmu. :) So I want to clarify if the block in smmu is big page, e.g. 2M page? Intel Vtd manages the memory per page, 4KB/2MB/1GB. There are two ways to manage dirty pages. 1. Keep default granularity. Just set SLADE to enable the dirty track. 2. Split big page to 4KB to get finer granularity. But question about the second solution is if it can benefit the user space, e.g. live migration. If my understanding about smmu block (i.e. the big page) is correct, have you collected some performance data to prove that the split can improve performance? Thanks! > Thanks, > Keqian