Received: by 2002:a05:6a10:a852:0:0:0:0 with SMTP id d18csp2930201pxy; Mon, 3 May 2021 11:07:38 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyYK830tEkUbPKEC/BR9T8fdemGpXzLe8TGIMS8YAqwXUL2G43pdK3gEACC1n2pPxyekuZ/ X-Received: by 2002:a50:f41a:: with SMTP id r26mr18434833edm.339.1620065258082; Mon, 03 May 2021 11:07:38 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1620065258; cv=none; d=google.com; s=arc-20160816; b=UmGZ3z8Zr+P2iTvbS+MeHYHpGPwcdZLXVbVAL8CpTWhqYIymYQYD1YBMg0KqsVxxzO /s7L33LIxR1WBzU93dmxXxtuC+lwYqr7sj5KN+2uvXRJPDIGHApTj99EVZeojOMfna5R 232XzvE382rwlR8QJNcuNX3DU7FbS+Ifh/qNOF8tbqmBTAVLBuhWzALJQIqltFEtjHTa e5L9sr7zxd0T6JQ/cSxhyKyB15jF/gxHVKaURfF/2kIHmd4UPBe5L/KrkPCNxCweNJYP h8eQEvC5Ei4SmY6t8e3evuJbH/o6RC2eECjYkiyKHJT745MoqtTpDVbXLBNQ54gDYtBy OWrg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:subject:from :references:cc:to:dkim-signature; bh=e5ncV0IuRV2X8acsw7XWidCK/BBCiCt6KbPi7m/CD8g=; b=uKyXeW9adObThROAGwxlWiJHpTIIvgQO4+/mG/gnFJ2DjetdSSveT7/kA2PaXOTyRs NIW4n6KAZaI1Xl+xioA2qKgWwLN+jKEoYh3hry2bdHEFsS11mpd6EFfjXbP9OouyRgDT r+UMm39duUMDCxXqJXaSHYLD2Z7auAx8GgYPaGDSTcgkl9evVebn/TzJ+9mu2byEp+SS SlVcC2YAX3/Lio5Wdp3CvOgz7ldiKBLHeu0HiBVxUx7WwSR+7DW9GIO/36IprgwQ/jv3 L03FGNHEElCoKavQRW7e5Krk548yKl+qOnokRuUCHNpySFifMm3rEsGc3GCEqfMMSXwb EN3w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=iQrD531w; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id bq19si1167929edb.311.2021.05.03.11.07.12; Mon, 03 May 2021 11:07:38 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=iQrD531w; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233773AbhECNaO (ORCPT + 99 others); Mon, 3 May 2021 09:30:14 -0400 Received: from us-smtp-delivery-124.mimecast.com ([170.10.133.124]:35768 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233983AbhECNaM (ORCPT ); Mon, 3 May 2021 09:30:12 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1620048559; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=e5ncV0IuRV2X8acsw7XWidCK/BBCiCt6KbPi7m/CD8g=; b=iQrD531wA7j91RBb9LZjgWcqgjZQJvTBbZdJD6QTpgi0kzkPYuT0QOvUZSx0A4f/hecTEf FDNONHnHm52Ui++UuegmTS/+5IFGAwEWW0siWLYvwPCZB8S4LJOOvsaHg6en6WsWvh5lWp FWCLILNEzS/qsZajNXf5w1t2kvZSaHA= Received: from mail-ed1-f70.google.com (mail-ed1-f70.google.com [209.85.208.70]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-472-cGuJiIdhN1yEUP3P8j4GPg-1; Mon, 03 May 2021 09:29:17 -0400 X-MC-Unique: cGuJiIdhN1yEUP3P8j4GPg-1 Received: by mail-ed1-f70.google.com with SMTP id g19-20020a0564021813b029038811907178so4514432edy.14 for ; Mon, 03 May 2021 06:29:17 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:to:cc:references:from:subject:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=e5ncV0IuRV2X8acsw7XWidCK/BBCiCt6KbPi7m/CD8g=; b=kn4yHSRc7qtkzVm8jJPHDUEagfuFDad8V/wQMSYD/zcABCrPr4RXSWBuk7bqAuHqQP NUyDj9BiC+Phciffpw0wzs2oid0Y9yG3MWLTCJiNxxgrW1BCSMqbSqomce6ZrsyrQuaX E7vTfte1W9Wb0BBBS/tgg7KIApzqC/J5VFQlnzUB8DJdIHU+8r/l/322ekHhn92OSiha VUKJfs28xMR7g3Nb7EIu7biXeOh9TewOkxV93ge+grZr4BVknOpT8tq50vqlL2MpBLdP mJqVYzmmn8S8HIMbQJIy7XciVY3vRJrlrh2u/O9Fh+BbZHYsRXvHUCtGuXIZCrvMBFm3 cCQg== X-Gm-Message-State: AOAM533qn1HIoHco9S0M5Kt4hcoivROTV+XMbLOLeu/XX6lPvUQJBOho zRAPnwiOCNfmA8Ta7D6qSovWp/xgQmSs/XhN5wg4CBUMVpNELrqAjMY620seWrcLPsUG0ROfWEa 997FfuK3RjnNrbw/ihtEgeclx X-Received: by 2002:a17:906:57c3:: with SMTP id u3mr6342365ejr.162.1620048556669; Mon, 03 May 2021 06:29:16 -0700 (PDT) X-Received: by 2002:a17:906:57c3:: with SMTP id u3mr6342344ejr.162.1620048556461; Mon, 03 May 2021 06:29:16 -0700 (PDT) Received: from ?IPv6:2001:b07:6468:f312:c8dd:75d4:99ab:290a? ([2001:b07:6468:f312:c8dd:75d4:99ab:290a]) by smtp.gmail.com with ESMTPSA id r16sm4130883edq.87.2021.05.03.06.29.15 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Mon, 03 May 2021 06:29:15 -0700 (PDT) To: Ben Gardon , linux-kernel@vger.kernel.org, kvm@vger.kernel.org Cc: Peter Xu , Sean Christopherson , Peter Shier , Junaid Shahid , Jim Mattson , Yulei Zhang , Wanpeng Li , Vitaly Kuznetsov , Xiao Guangrong References: <20210429211833.3361994-1-bgardon@google.com> <20210429211833.3361994-7-bgardon@google.com> From: Paolo Bonzini Subject: Re: [PATCH v2 6/7] KVM: mmu: Add slots_arch_lock for memslot arch fields Message-ID: <1e9c77a9-adec-0a2d-5483-70cb2332d529@redhat.com> Date: Mon, 3 May 2021 15:29:14 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.8.1 MIME-Version: 1.0 In-Reply-To: <20210429211833.3361994-7-bgardon@google.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 29/04/21 23:18, Ben Gardon wrote: > Add a new lock to protect the arch-specific fields of memslots if they > need to be modified in a kvm->srcu read critical section. A future > commit will use this lock to lazily allocate memslot rmaps for x86. Here there should be a blurb about the possible races that can happen and why we decided for the slots_arch_lock. > Signed-off-by: Ben Gardon > --- > include/linux/kvm_host.h | 9 +++++++++ > virt/kvm/kvm_main.c | 31 ++++++++++++++++++++++++++----- > 2 files changed, 35 insertions(+), 5 deletions(-) > > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index 8895b95b6a22..2d5e797fbb08 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -472,6 +472,15 @@ struct kvm { > #endif /* KVM_HAVE_MMU_RWLOCK */ > > struct mutex slots_lock; > + > + /* > + * Protects the arch-specific fields of struct kvm_memory_slots in > + * use by the VM. To be used under the slots_lock (above) or in a > + * kvm->srcu read cirtical section where acquiring the slots_lock > + * would lead to deadlock with the synchronize_srcu in > + * install_new_memslots. > + */ I think usage under slots_lock need not be mentioned here. More like this: /* * Protects the arch-specific fields of struct kvm_memory_slots * in use by the VM. Usually these are initialized by * kvm_arch_prepare_memory_region and then protected by * kvm->srcu; however, if they need to be initialized outside * kvm_arch_prepare_memory_region, slots_arch_lock can * be used instead as it is also held when calling * kvm_arch_prepare_memory_region itself. Note that using * slots_lock would lead to deadlock with install_new_memslots, * because it is held during synchronize_srcu: * * idx = srcu_read_lock(&kvm->srcu); * mutex_lock(&kvm->slots_lock); * mutex_lock(&kvm->slots_lock); * synchronize_srcu(&kvm->srcu); */ (Though a better place for this is in Documentation/virtual/kvm/locking.rst). Paolo > + struct mutex slots_arch_lock; > struct mm_struct *mm; /* userspace tied to this vm */ > struct kvm_memslots __rcu *memslots[KVM_ADDRESS_SPACE_NUM]; > struct kvm_vcpu *vcpus[KVM_MAX_VCPUS]; > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index c8010f55e368..97b03fa2d0c8 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -908,6 +908,7 @@ static struct kvm *kvm_create_vm(unsigned long type) > mutex_init(&kvm->lock); > mutex_init(&kvm->irq_lock); > mutex_init(&kvm->slots_lock); > + mutex_init(&kvm->slots_arch_lock); > INIT_LIST_HEAD(&kvm->devices); > > BUILD_BUG_ON(KVM_MEM_SLOTS_NUM > SHRT_MAX); > @@ -1280,6 +1281,10 @@ static struct kvm_memslots *install_new_memslots(struct kvm *kvm, > slots->generation = gen | KVM_MEMSLOT_GEN_UPDATE_IN_PROGRESS; > > rcu_assign_pointer(kvm->memslots[as_id], slots); > + > + /* Acquired in kvm_set_memslot. */ > + mutex_unlock(&kvm->slots_arch_lock); > + > synchronize_srcu_expedited(&kvm->srcu); > > /* > @@ -1351,6 +1356,9 @@ static int kvm_set_memslot(struct kvm *kvm, > struct kvm_memslots *slots; > int r; > > + /* Released in install_new_memslots. */ > + mutex_lock(&kvm->slots_arch_lock); > + > slots = kvm_dup_memslots(__kvm_memslots(kvm, as_id), change); > if (!slots) > return -ENOMEM; > @@ -1364,10 +1372,9 @@ static int kvm_set_memslot(struct kvm *kvm, > slot->flags |= KVM_MEMSLOT_INVALID; > > /* > - * We can re-use the old memslots, the only difference from the > - * newly installed memslots is the invalid flag, which will get > - * dropped by update_memslots anyway. We'll also revert to the > - * old memslots if preparing the new memory region fails. > + * We can re-use the memory from the old memslots. > + * It will be overwritten with a copy of the new memslots > + * after reacquiring the slots_arch_lock below. > */ > slots = install_new_memslots(kvm, as_id, slots); > > @@ -1379,6 +1386,17 @@ static int kvm_set_memslot(struct kvm *kvm, > * - kvm_is_visible_gfn (mmu_check_root) > */ > kvm_arch_flush_shadow_memslot(kvm, slot); > + > + /* Released in install_new_memslots. */ > + mutex_lock(&kvm->slots_arch_lock); > + > + /* > + * The arch-specific fields of the memslots could have changed > + * between releasing the slots_arch_lock in > + * install_new_memslots and here, so get a fresh copy of the > + * slots. > + */ > + kvm_copy_memslots(__kvm_memslots(kvm, as_id), slots); > } > > r = kvm_arch_prepare_memory_region(kvm, new, mem, change); > @@ -1394,8 +1412,11 @@ static int kvm_set_memslot(struct kvm *kvm, > return 0; > > out_slots: > - if (change == KVM_MR_DELETE || change == KVM_MR_MOVE) > + if (change == KVM_MR_DELETE || change == KVM_MR_MOVE) { > + slot = id_to_memslot(slots, old->id); > + slot->flags &= ~KVM_MEMSLOT_INVALID; > slots = install_new_memslots(kvm, as_id, slots); > + } > kvfree(slots); > return r; > } >