Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752932AbdHQOyV (ORCPT ); Thu, 17 Aug 2017 10:54:21 -0400 Received: from mx1.redhat.com ([209.132.183.28]:60314 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752751AbdHQOyT (ORCPT ); Thu, 17 Aug 2017 10:54:19 -0400 DMARC-Filter: OpenDMARC Filter v1.3.2 mx1.redhat.com 7F169356DA Authentication-Results: ext-mx06.extmail.prod.ext.phx2.redhat.com; dmarc=none (p=none dis=none) header.from=redhat.com Authentication-Results: ext-mx06.extmail.prod.ext.phx2.redhat.com; spf=fail smtp.mailfrom=rkrcmar@redhat.com Date: Thu, 17 Aug 2017 16:54:12 +0200 From: Radim =?utf-8?B?S3LEjW3DocWZ?= To: Alexander Graf Cc: linux-mips@linux-mips.org, linux-arm-kernel@lists.infradead.org, kvm@vger.kernel.org, kvm-ppc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-s390@vger.kernel.org, Marc Zyngier , Christian Borntraeger , James Hogan , Christoffer Dall , Paul Mackerras , Cornelia Huck , David Hildenbrand , Paolo Bonzini Subject: Re: [PATCH RFC 0/2] KVM: use RCU to allow dynamic kvm->vcpus array Message-ID: <20170817145411.GE2566@flask> References: <20170816194037.9460-1-rkrcmar@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.30]); Thu, 17 Aug 2017 14:54:19 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2551 Lines: 57 2017-08-17 09:04+0200, Alexander Graf: > On 16.08.17 21:40, Radim Krčmář wrote: > > The goal is to increase KVM_MAX_VCPUS without worrying about memory > > impact of many small guests. > > > > This is a second out of three major "dynamic" options: > > 1) size vcpu array at VM creation time > > 2) resize vcpu array when new VCPUs are created > > 3) use a lockless list/tree for VCPUs > > > > The disadvantage of (1) is its requirement on userspace changes and > > limited flexibility because userspace must provide the maximal count on > > start. The main advantage is that kvm->vcpus will work like it does > > now. It has been posted as "[PATCH 0/4] KVM: add KVM_CREATE_VM2 to > > allow dynamic kvm->vcpus array", > > http://www.mail-archive.com/linux-kernel@vger.kernel.org/msg1377285.html > > > > The main problem of (2), this series, is that we cannot extend the array > > in place and therefore require some kind of protection when moving it. > > RCU seems best, but it makes the code slower and harder to deal with. > > The main advantage is that we do not need userspace changes. > > Creating/Destroying vcpus is not something I consider a fast path, so why > should we optimize for it? The case that needs to be fast is execution. Right, the creation is not important. I was concerned about the use of lock() and unlock() needed for every access -- both in performance and code, because the common case where hotplug doesn't happen and all VCPUs are created upfront doesn't even need any runtime protection. > What if we just sent a "vcpu move" request to all vcpus with the new pointer > after it moved? That way the vcpu thread itself would be responsible for the > migration to the new memory region. Only if all vcpus successfully moved, > keep rolling (and allow foreign get_vcpu again). I'm not sure if I understood this. You propose to cache kvm->vcpus in vcpu->vcpus and do an extensions of this, int vcpu_create(...) { if (resize_needed(kvm->vcpus)) { old_vcpus = kvm->vcpus kvm->vcpus = make_bigger(kvm->vcpus) kvm_make_all_cpus_request(kvm, KVM_REQ_UPDATE_VCPUS) free(old_vcpus) } vcpu->vcpus = kvm->vcpus } with added extra locking, (S)RCU, on accesses that do not come from VCPUs (irqfd and VM ioctl)? > That way we should be basically lock-less and scale well. For additional > icing, feel free to increase the vcpu array x2 every time it grows to not > run into the slow path too often. Yeah, I skipped the growing as it was not necessary for the illustration.