Received: by 2002:a05:6a10:5bc5:0:0:0:0 with SMTP id os5csp1787901pxb; Wed, 20 Oct 2021 11:44:53 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyZyV+ijvgESC5BnpFLgyQUwp8IUnCUZjT6i9aUphtPEtEi41SgkEWFKgQ1pVQZ7hoKrucz X-Received: by 2002:a17:906:131a:: with SMTP id w26mr902118ejb.548.1634755492899; Wed, 20 Oct 2021 11:44:52 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1634755492; cv=none; d=google.com; s=arc-20160816; b=YeduTfis0HZjWKA+EVg3JmpyGvanOBaYEfqRdSRYAny1MaMvpnlR8rgPo1a1j+dI0n VhxV1uNAFFUNU7bpzzrBK7lyJM4Oi0148qvxwQv1Gwv7CwFcLeZXXQJNV0u0pVVQMT0p W9kaz3MIoe3uM0Tzmc09AutcfLEXEn8PcePPE5Nshs6waihPNGbNwuDttsg6f98WS2TK VcrrIYGxExYxZJKjNi32rB+iyjoUVRREPk72lsb/x6WN1vEWmna5D68l+El9DQBsQD/E 2ux1HEA3ML336aFHaaPGfhlgagJCs4fSRs4Q141hUO52NIyczZxJS7E7h4dDY50zUWsO TWEA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:subject:from :references:cc:to; bh=H97caubY5oiBUQmSY5oR6T0cjGcf4i8VHNgYne9Ee/g=; b=BsB75he9k0T4n9BPG06rCYST8h50qLhz6lSSXwmYkiWXcwgrhfNRVWK3wKd7dTTl2e 9cVjpOu+WxKivX6VbOIwUc4Gy5vrGJFWr8WU9lYv8GhMDH4WLY8u/8g8vJhuK2zLihnj LtqwAVHMnse1aDwAChjB7GMmSFsGnEJzCgptEmGHGtNgeHRNAv9nnGDkJuiOm7sZhaSC vAKE2wZYma/U+cAkx4oGGI566bgLhFubuBQyH2AYzDtwd+wmQEsg5friq5mlTwIOhDZ8 1uEi2/mkypWkDZz7pKo+8gM6kV+V/zxtzANysCHDui57TNERqIo0IZYreKz/+qwfmyEG Qw2Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id oz30si433159ejc.667.2021.10.20.11.44.22; Wed, 20 Oct 2021 11:44:52 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231331AbhJTSor (ORCPT + 99 others); Wed, 20 Oct 2021 14:44:47 -0400 Received: from vps-vb.mhejs.net ([37.28.154.113]:40950 "EHLO vps-vb.mhejs.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231271AbhJTSoq (ORCPT ); Wed, 20 Oct 2021 14:44:46 -0400 Received: from MUA by vps-vb.mhejs.net with esmtps (TLS1.2) tls TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 (Exim 4.94.2) (envelope-from ) id 1mdGXl-0002gu-8u; Wed, 20 Oct 2021 20:42:29 +0200 To: Sean Christopherson Cc: Paolo Bonzini , Vitaly Kuznetsov , Wanpeng Li , Jim Mattson , Igor Mammedov , Marc Zyngier , James Morse , Julien Thierry , Suzuki K Poulose , Huacai Chen , Aleksandar Markovic , Paul Mackerras , Christian Borntraeger , Janosch Frank , David Hildenbrand , Cornelia Huck , Claudio Imbrenda , Joerg Roedel , kvm@vger.kernel.org, linux-kernel@vger.kernel.org References: <555f58fdaec120aa7a6f6fbad06cca796a8c9168.1632171479.git.maciej.szmigiero@oracle.com> From: "Maciej S. Szmigiero" Subject: Re: [PATCH v5 08/13] KVM: Resolve memslot ID via a hash table instead of via a static array Message-ID: <1729cda2-83f0-ed03-c6b4-4418de80f933@maciej.szmigiero.name> Date: Wed, 20 Oct 2021 20:42:23 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.13.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 20.10.2021 02:43, Sean Christopherson wrote: > On Mon, Sep 20, 2021, Maciej S. Szmigiero wrote: >> --- >> include/linux/kvm_host.h | 16 +++++------ >> virt/kvm/kvm_main.c | 61 +++++++++++++++++++++++++++++++--------- >> 2 files changed, 55 insertions(+), 22 deletions(-) >> >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h >> index 8fd9644f40b2..d2acc00a6472 100644 >> --- a/include/linux/kvm_host.h >> +++ b/include/linux/kvm_host.h >> @@ -29,6 +29,7 @@ >> #include >> #include >> #include >> +#include >> #include >> >> #include >> @@ -426,6 +427,7 @@ static inline int kvm_vcpu_exiting_guest_mode(struct kvm_vcpu *vcpu) >> #define KVM_MEM_MAX_NR_PAGES ((1UL << 31) - 1) >> >> struct kvm_memory_slot { >> + struct hlist_node id_node; >> gfn_t base_gfn; >> unsigned long npages; >> unsigned long *dirty_bitmap; >> @@ -528,7 +530,7 @@ static inline int kvm_arch_vcpu_memslots_id(struct kvm_vcpu *vcpu) >> struct kvm_memslots { >> u64 generation; >> /* The mapping table from slot id to the index in memslots[]. */ >> - short id_to_index[KVM_MEM_SLOTS_NUM]; >> + DECLARE_HASHTABLE(id_hash, 7); > > Can you add a comment explaining the rationale for size "7"? Not necessarily the > justification in choosing "7", more so the tradeoffs between performance, memory, > etc... so that all your work/investigation isn't lost and doesn't have to be repeated > if someone wants to tweak this in the future. Will add such comment. >> atomic_t last_used_slot; >> int used_slots; >> struct kvm_memory_slot memslots[]; >> @@ -795,16 +797,14 @@ static inline struct kvm_memslots *kvm_vcpu_memslots(struct kvm_vcpu *vcpu) >> static inline >> struct kvm_memory_slot *id_to_memslot(struct kvm_memslots *slots, int id) >> { >> - int index = slots->id_to_index[id]; >> struct kvm_memory_slot *slot; >> >> - if (index < 0) >> - return NULL; >> - >> - slot = &slots->memslots[index]; >> + hash_for_each_possible(slots->id_hash, slot, id_node, id) { >> + if (slot->id == id) >> + return slot; > > Hmm, related to the hash, it might be worth adding a stat here to count collisions. > Might be more pain than it's worth though since we don't have @kvm. It's a good idea if it turns out that it's worth optimizing the code further (by, for example, introducing a self-resizing hash table, which would bring a significant increase in complexity for rather uncertain gains). >> @@ -1274,30 +1275,46 @@ static inline int kvm_memslot_insert_back(struct kvm_memslots *slots) >> * itself is not preserved in the array, i.e. not swapped at this time, only >> * its new index into the array is tracked. Returns the changed memslot's >> * current index into the memslots array. >> + * The memslot at the returned index will not be in @slots->id_hash by then. >> + * @memslot is a detached struct with desired final data of the changed slot. >> */ >> static inline int kvm_memslot_move_backward(struct kvm_memslots *slots, >> struct kvm_memory_slot *memslot) >> { >> struct kvm_memory_slot *mslots = slots->memslots; >> + struct kvm_memory_slot *mmemslot = id_to_memslot(slots, memslot->id); > > My comment from v3 about the danger of "mmemslot" still stands. FWIW, I dislike > "mslots" as well, but that predates me, and all of this will go away in the end :-) > > On Wed, May 19, 2021 at 3:31 PM Sean Christopherson wrote: >> On Sun, May 16, 2021, Maciej S. Szmigiero wrote: >>> struct kvm_memory_slot *mslots = slots->memslots; >>> + struct kvm_memory_slot *dmemslot = id_to_memslot(slots, memslot->id); >> >> I vote to call these local vars "old", or something along those lines. dmemslot >> isn't too bad, but mmemslot in the helpers below is far too similar to memslot, >> and using the wrong will cause nasty explosions. > Will rename "mmemslot" to "oldslot" in kvm_memslot_move_backward(), too. >> int i; >> >> - if (slots->id_to_index[memslot->id] == -1 || !slots->used_slots) >> + if (!mmemslot || !slots->used_slots) >> return -1; >> >> + /* >> + * The loop below will (possibly) overwrite the target memslot with >> + * data of the next memslot, or a similar loop in >> + * kvm_memslot_move_forward() will overwrite it with data of the >> + * previous memslot. >> + * Then update_memslots() will unconditionally overwrite and re-add >> + * it to the hash table. >> + * That's why the memslot has to be first removed from the hash table >> + * here. >> + */ > > Is this reword accurate? > > /* > * Delete the slot from the hash table before sorting the remaining > * slots, the slot's data may be overwritten when copying slots as part > * of the sorting proccess. update_memslots() will unconditionally > * rewrite the entire slot and re-add it to the hash table. > */ It's accurate, will replace the comment with the proposed one. >> @@ -1369,6 +1391,9 @@ static inline int kvm_memslot_move_forward(struct kvm_memslots *slots, >> * most likely to be referenced, sorting it to the front of the array was >> * advantageous. The current binary search starts from the middle of the array >> * and uses an LRU pointer to improve performance for all memslots and GFNs. >> + * >> + * @memslot is a detached struct, not a part of the current or new memslot >> + * array. >> */ >> static void update_memslots(struct kvm_memslots *slots, >> struct kvm_memory_slot *memslot, >> @@ -1393,7 +1418,8 @@ static void update_memslots(struct kvm_memslots *slots, >> * its index accordingly. >> */ >> slots->memslots[i] = *memslot; >> - slots->id_to_index[memslot->id] = i; >> + hash_add(slots->id_hash, &slots->memslots[i].id_node, >> + memslot->id); > > Let this poke out past 80 chars, i.e. drop the newline. Will do. Thanks, Maciej