Received: by 2002:a05:6902:102b:0:0:0:0 with SMTP id x11csp1003134ybt; Wed, 1 Jul 2020 15:56:53 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyhCPctfp8UF2dfnMalLNKXq3ghvD2Ugy0vAXE9ZqYe/QvnD+E3CajUr0igTXz8SqvMwosF X-Received: by 2002:a17:906:6558:: with SMTP id u24mr12436493ejn.364.1593644213544; Wed, 01 Jul 2020 15:56:53 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1593644213; cv=none; d=google.com; s=arc-20160816; b=EjQhYWVeQkSud9E6ZXVbuZFOXcIhoWbNAXLraNwunqpIJZYnjtXPoUU9OTYnXoknj6 M8q+WQVf4ByfVuwC+B3SUU31Y59teBMXvnTLZJRDPq1cpXMAYiEJWIZJEXHIdNW+GK+r r9i1Qy90PwbxrtrPKVb9aHEnrpQx+RA1iW8Hfl7qIuwmZ5SqmVJipCfhFlXRlarcn4Wg ut0JiOqIt57onALNg5qshzsD6+Cx4LYj6HIo0euDt8mYWvHpqpzldCr+SAzmC/Z9SGKZ YGwiL34tLXbl2xdxIqX+QBu2xnbHmoVondO+JSCVR2yWrRNP0FR7XPyTVGe/R9aepyvz jjGw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:dkim-signature:content-transfer-encoding :mime-version:references:in-reply-to:message-id:date:subject:cc:to :from; bh=UL+bKmwqRdPEbqM2DAzkFsHMUp8NkFT28XTM8su6HPs=; b=AtPDO5ElZvWc2PKStCpngBZ1Cmy9HTna8rC2o4Z88mwi6u9za46xI/jtkeBf3gt+Gv 2qMuEDyYWK1R6XuG1u1n7bBON02VwC52f17zcKzq+Za3yE+NGaMlCgguEqwVMiZTWnCd uIU6oubo5EZmPZ3lw/SzKmUv31iCNUEb32bwSxzcu4kOdq/fniAWHuRI5JFal63ZeU9p LOK6frUw5Dx32kvoidz1xnjpkliM9QAbm0AUny1KMNM6ar/ZoAyRNKaRaz8EXmA3MQUZ Fb+v9SgZwE0O+cjR1SdvJc6LRkLPjTCxWgbvJW/sUIKkSQcjiCI58tjoUn+BwGerHe+a PfZA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=iCjwNcVg; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id ob20si4650541ejb.646.2020.07.01.15.56.30; Wed, 01 Jul 2020 15:56:53 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=iCjwNcVg; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727895AbgGAWyS (ORCPT + 99 others); Wed, 1 Jul 2020 18:54:18 -0400 Received: from hqnvemgate24.nvidia.com ([216.228.121.143]:11162 "EHLO hqnvemgate24.nvidia.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727817AbgGAWyQ (ORCPT ); Wed, 1 Jul 2020 18:54:16 -0400 Received: from hqpgpgate102.nvidia.com (Not Verified[216.228.121.13]) by hqnvemgate24.nvidia.com (using TLS: TLSv1.2, DES-CBC3-SHA) id ; Wed, 01 Jul 2020 15:52:36 -0700 Received: from hqmail.nvidia.com ([172.20.161.6]) by hqpgpgate102.nvidia.com (PGP Universal service); Wed, 01 Jul 2020 15:54:16 -0700 X-PGP-Universal: processed; by hqpgpgate102.nvidia.com on Wed, 01 Jul 2020 15:54:16 -0700 Received: from HQMAIL105.nvidia.com (172.20.187.12) by HQMAIL111.nvidia.com (172.20.187.18) with Microsoft SMTP Server (TLS) id 15.0.1473.3; Wed, 1 Jul 2020 22:54:06 +0000 Received: from rnnvemgw01.nvidia.com (10.128.109.123) by HQMAIL105.nvidia.com (172.20.187.12) with Microsoft SMTP Server (TLS) id 15.0.1473.3 via Frontend Transport; Wed, 1 Jul 2020 22:54:06 +0000 Received: from rcampbell-dev.nvidia.com (Not Verified[10.110.48.66]) by rnnvemgw01.nvidia.com with Trustwave SEG (v7,5,8,10121) id ; Wed, 01 Jul 2020 15:54:06 -0700 From: Ralph Campbell To: , , , , CC: Jerome Glisse , John Hubbard , Christoph Hellwig , Jason Gunthorpe , "Andrew Morton" , Shuah Khan , "Ben Skeggs" , Ralph Campbell Subject: [PATCH v3 5/5] hmm: add tests for hmm_pfn_to_map_order() Date: Wed, 1 Jul 2020 15:53:52 -0700 Message-ID: <20200701225352.9649-6-rcampbell@nvidia.com> X-Mailer: git-send-email 2.20.1 In-Reply-To: <20200701225352.9649-1-rcampbell@nvidia.com> References: <20200701225352.9649-1-rcampbell@nvidia.com> MIME-Version: 1.0 X-NVConfidentiality: public Content-Transfer-Encoding: quoted-printable Content-Type: text/plain DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1593643956; bh=UL+bKmwqRdPEbqM2DAzkFsHMUp8NkFT28XTM8su6HPs=; h=X-PGP-Universal:From:To:CC:Subject:Date:Message-ID:X-Mailer: In-Reply-To:References:MIME-Version:X-NVConfidentiality: Content-Transfer-Encoding:Content-Type; b=iCjwNcVg7bkn/my+AxsLEl8MhQESiDTQPBs3PLFh2F40Nc/cXrK8c3tmadkRomcIE D7ltu2olJDS9P8HASUKewPs5SBsFA3PYTp+DT616AuomF7Tx//09pTgE8yZzFdryYk aqdkso4Rl3YUGTzj4Ndf4FgBDAQKqsABk5Knpg6FBb0YYhNcucpJ16wHePWDwEotOo ZIW0XTifi4pkKuin//VSry09hlWEmzt/cLQoUXFAoOO8k9RvvxHz6pMFNaJPe9jQxP biTGEKAZPI7nrxeCVsnNCO09M5GT+FLtSrHKxx0dGNGIKjuISD7RNk0P0952GHn5nM dxedWuEOm5DbQ== Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Add a sanity test for hmm_range_fault() returning the page mapping size order. Signed-off-by: Ralph Campbell --- lib/test_hmm.c | 4 ++ lib/test_hmm_uapi.h | 4 ++ tools/testing/selftests/vm/hmm-tests.c | 76 ++++++++++++++++++++++++++ 3 files changed, 84 insertions(+) diff --git a/lib/test_hmm.c b/lib/test_hmm.c index a2a82262b97b..9aa577afc269 100644 --- a/lib/test_hmm.c +++ b/lib/test_hmm.c @@ -766,6 +766,10 @@ static void dmirror_mkentry(struct dmirror *dmirror, s= truct hmm_range *range, *perm |=3D HMM_DMIRROR_PROT_WRITE; else *perm |=3D HMM_DMIRROR_PROT_READ; + if (hmm_pfn_to_map_order(entry) + PAGE_SHIFT =3D=3D PMD_SHIFT) + *perm |=3D HMM_DMIRROR_PROT_PMD; + else if (hmm_pfn_to_map_order(entry) + PAGE_SHIFT =3D=3D PUD_SHIFT) + *perm |=3D HMM_DMIRROR_PROT_PUD; } =20 static bool dmirror_snapshot_invalidate(struct mmu_interval_notifier *mni, diff --git a/lib/test_hmm_uapi.h b/lib/test_hmm_uapi.h index 67b3b2e6ff5d..670b4ef2a5b6 100644 --- a/lib/test_hmm_uapi.h +++ b/lib/test_hmm_uapi.h @@ -40,6 +40,8 @@ struct hmm_dmirror_cmd { * HMM_DMIRROR_PROT_NONE: unpopulated PTE or PTE with no access * HMM_DMIRROR_PROT_READ: read-only PTE * HMM_DMIRROR_PROT_WRITE: read/write PTE + * HMM_DMIRROR_PROT_PMD: PMD sized page is fully mapped by same permission= s + * HMM_DMIRROR_PROT_PUD: PUD sized page is fully mapped by same permission= s * HMM_DMIRROR_PROT_ZERO: special read-only zero page * HMM_DMIRROR_PROT_DEV_PRIVATE_LOCAL: Migrated device private page on the * device the ioctl() is made @@ -51,6 +53,8 @@ enum { HMM_DMIRROR_PROT_NONE =3D 0x00, HMM_DMIRROR_PROT_READ =3D 0x01, HMM_DMIRROR_PROT_WRITE =3D 0x02, + HMM_DMIRROR_PROT_PMD =3D 0x04, + HMM_DMIRROR_PROT_PUD =3D 0x08, HMM_DMIRROR_PROT_ZERO =3D 0x10, HMM_DMIRROR_PROT_DEV_PRIVATE_LOCAL =3D 0x20, HMM_DMIRROR_PROT_DEV_PRIVATE_REMOTE =3D 0x30, diff --git a/tools/testing/selftests/vm/hmm-tests.c b/tools/testing/selftes= ts/vm/hmm-tests.c index 79db22604019..b533dd08da1d 100644 --- a/tools/testing/selftests/vm/hmm-tests.c +++ b/tools/testing/selftests/vm/hmm-tests.c @@ -1291,6 +1291,82 @@ TEST_F(hmm2, snapshot) hmm_buffer_free(buffer); } =20 +/* + * Test the hmm_range_fault() HMM_PFN_PMD flag for large pages that + * should be mapped by a large page table entry. + */ +TEST_F(hmm, compound) +{ + struct hmm_buffer *buffer; + unsigned long npages; + unsigned long size; + int *ptr; + unsigned char *m; + int ret; + long pagesizes[4]; + int n, idx; + unsigned long i; + + /* Skip test if we can't allocate a hugetlbfs page. */ + + n =3D gethugepagesizes(pagesizes, 4); + if (n <=3D 0) + return; + for (idx =3D 0; --n > 0; ) { + if (pagesizes[n] < pagesizes[idx]) + idx =3D n; + } + size =3D ALIGN(TWOMEG, pagesizes[idx]); + npages =3D size >> self->page_shift; + + buffer =3D malloc(sizeof(*buffer)); + ASSERT_NE(buffer, NULL); + + buffer->ptr =3D get_hugepage_region(size, GHR_STRICT); + if (buffer->ptr =3D=3D NULL) { + free(buffer); + return; + } + + buffer->size =3D size; + buffer->mirror =3D malloc(npages); + ASSERT_NE(buffer->mirror, NULL); + + /* Initialize the pages the device will snapshot in buffer->ptr. */ + for (i =3D 0, ptr =3D buffer->ptr; i < size / sizeof(*ptr); ++i) + ptr[i] =3D i; + + /* Simulate a device snapshotting CPU pagetables. */ + ret =3D hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages); + ASSERT_EQ(ret, 0); + ASSERT_EQ(buffer->cpages, npages); + + /* Check what the device saw. */ + m =3D buffer->mirror; + for (i =3D 0; i < npages; ++i) + ASSERT_EQ(m[i], HMM_DMIRROR_PROT_WRITE | + HMM_DMIRROR_PROT_PMD); + + /* Make the region read-only. */ + ret =3D mprotect(buffer->ptr, size, PROT_READ); + ASSERT_EQ(ret, 0); + + /* Simulate a device snapshotting CPU pagetables. */ + ret =3D hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages); + ASSERT_EQ(ret, 0); + ASSERT_EQ(buffer->cpages, npages); + + /* Check what the device saw. */ + m =3D buffer->mirror; + for (i =3D 0; i < npages; ++i) + ASSERT_EQ(m[i], HMM_DMIRROR_PROT_READ | + HMM_DMIRROR_PROT_PMD); + + free_hugepage_region(buffer->ptr); + buffer->ptr =3D NULL; + hmm_buffer_free(buffer); +} + /* * Test two devices reading the same memory (double mapped). */ --=20 2.20.1