Received: by 2002:a25:23cc:0:0:0:0:0 with SMTP id j195csp632993ybj; Thu, 7 May 2020 04:08:11 -0700 (PDT) X-Google-Smtp-Source: APiQypIUbsvwNkjpw8Ymr8L68PT9DPd94CuVyjcg+MNkCY0lzfXTiz1Qf5hmQNn97ygJqgDDVsIk X-Received: by 2002:a17:906:4c8e:: with SMTP id q14mr11959162eju.208.1588849690941; Thu, 07 May 2020 04:08:10 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588849690; cv=none; d=google.com; s=arc-20160816; b=1AS6xgL4PfK9+q5s37WQd3Xs6mAzKMn69evrzBTYg9f7vWHtT+qf87MRP/y1N/TsHv qUt+qL89b8GRaoKjO6yBnzWFU8JOxCOqgWvTYOZAaxusu8TGG7TpQ0h+AmoKYjcbpejW UkOL8vKYrK2BbX19FPiVXSt0kcPVHQ08n3a05OOZo6tp/F5GJzZbiSrxPAq3L03BN+C5 U1JccYGNOlEnYZl1q8a1y8uyMX5Zp22iYgo/VN4hAd84b98erwJ9sn94fKuNNFKUPIoV Ra2V9l6dhj1l4nsvpvM5r0PZv3QKwB+XLbFd+6TGcCCtTsrVFJRzpe5JPij7X6iNra3r toPg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature; bh=ETQ6uKGHMQumDeMEnt+lmo6qtl8Y1L8zkXdal0m3TTw=; b=qxbc9djqxMUxySY9KOS5SnPK02PtCLXNnR+i/NjIJIrOb5FcWTkAGofWJA4hEQza8E BhyAqIdtlA4K5zSITkyzeua/8MgUNNz+82rnTPbUuR6LkVxtHUweVEDUG5sxgAEfCBZc EE8L7Zi2uc+yY0sJB6ngKjzbDCqZPZl/2jItCh1Cs60/GlKh8GtJx61+C5Hi5ZBJ+MiS 5ttSmsYGYn6MF+iDnF9iaovfKxAFJGkVOj3r9el5Au+jKx/PF2UaiLUpduCfHfwNSp2Z 1JywWpYnZPvu5hnhDOu2wZaZPPe2BcwS3A+VvVxXLeCbyXxJiRw27UwAtjMwmERKuGHI etGg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b="Xi83Jr/j"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id i5si3228752ejr.71.2020.05.07.04.07.42; Thu, 07 May 2020 04:08:10 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b="Xi83Jr/j"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727890AbgEGKdm (ORCPT + 99 others); Thu, 7 May 2020 06:33:42 -0400 Received: from us-smtp-delivery-1.mimecast.com ([205.139.110.120]:57176 "EHLO us-smtp-1.mimecast.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1727878AbgEGKdj (ORCPT ); Thu, 7 May 2020 06:33:39 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1588847617; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=ETQ6uKGHMQumDeMEnt+lmo6qtl8Y1L8zkXdal0m3TTw=; b=Xi83Jr/jloetrwDLBEruUm4uTUXniiwsL3YZRtmr+nBRK5MeFboQlLhqaoGfNOxpf43Lae tVu1cAtuf/n9UR7H/Utueul25ReLVbrr8HPuDLYRrNwx345FLb1lKrqYbC9idwhzicxhFs BFtvmB8OGP7E+z4ikP/mq4E33ElK8p0= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-346-zu1RLbGwPVy20Z2PfAoPog-1; Thu, 07 May 2020 06:33:35 -0400 X-MC-Unique: zu1RLbGwPVy20Z2PfAoPog-1 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.phx2.redhat.com [10.5.11.14]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 7B53F19200C1; Thu, 7 May 2020 10:33:34 +0000 (UTC) Received: from t480s.redhat.com (ovpn-113-245.ams2.redhat.com [10.36.113.245]) by smtp.corp.redhat.com (Postfix) with ESMTP id 7BF5D5D9C5; Thu, 7 May 2020 10:33:32 +0000 (UTC) From: David Hildenbrand To: linux-kernel@vger.kernel.org Cc: linux-mm@kvack.org, virtio-dev@lists.oasis-open.org, virtualization@lists.linux-foundation.org, kvm@vger.kernel.org, Michal Hocko , Andrew Morton , "Michael S . Tsirkin" , David Hildenbrand , Pankaj Gupta Subject: [PATCH v3 15/15] virtio-mem: Try to unplug the complete online memory block first Date: Thu, 7 May 2020 12:31:19 +0200 Message-Id: <20200507103119.11219-16-david@redhat.com> In-Reply-To: <20200507103119.11219-1-david@redhat.com> References: <20200507103119.11219-1-david@redhat.com> MIME-Version: 1.0 X-Scanned-By: MIMEDefang 2.79 on 10.5.11.14 Content-Transfer-Encoding: quoted-printable Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Right now, we always try to unplug single subblocks when processing an online memory block. Let's try to unplug the complete online memory block first, in case it is fully plugged and the unplug request is large enough. Fallback to single subblocks in case the memory block cannot get unplugged as a whole. Cc: "Michael S. Tsirkin" Cc: Pankaj Gupta Signed-off-by: David Hildenbrand --- drivers/virtio/virtio_mem.c | 88 ++++++++++++++++++++++++------------- 1 file changed, 57 insertions(+), 31 deletions(-) diff --git a/drivers/virtio/virtio_mem.c b/drivers/virtio/virtio_mem.c index abd93b778a26..9e523db3bee1 100644 --- a/drivers/virtio/virtio_mem.c +++ b/drivers/virtio/virtio_mem.c @@ -1307,6 +1307,46 @@ static int virtio_mem_mb_unplug_any_sb_offline(str= uct virtio_mem *vm, return 0; } =20 +/* + * Unplug the given plugged subblocks of an online memory block. + * + * Will modify the state of the memory block. + */ +static int virtio_mem_mb_unplug_sb_online(struct virtio_mem *vm, + unsigned long mb_id, int sb_id, + int count) +{ + const unsigned long nr_pages =3D PFN_DOWN(vm->subblock_size) * count; + unsigned long start_pfn; + int rc; + + start_pfn =3D PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) + + sb_id * vm->subblock_size); + rc =3D alloc_contig_range(start_pfn, start_pfn + nr_pages, + MIGRATE_MOVABLE, GFP_KERNEL); + if (rc =3D=3D -ENOMEM) + /* whoops, out of memory */ + return rc; + if (rc) + return -EBUSY; + + /* Mark it as fake-offline before unplugging it */ + virtio_mem_set_fake_offline(start_pfn, nr_pages, true); + adjust_managed_page_count(pfn_to_page(start_pfn), -nr_pages); + + /* Try to unplug the allocated memory */ + rc =3D virtio_mem_mb_unplug_sb(vm, mb_id, sb_id, count); + if (rc) { + /* Return the memory to the buddy. */ + virtio_mem_fake_online(start_pfn, nr_pages); + return rc; + } + + virtio_mem_mb_set_state(vm, mb_id, + VIRTIO_MEM_MB_STATE_ONLINE_PARTIAL); + return 0; +} + /* * Unplug the desired number of plugged subblocks of an online memory bl= ock. * Will skip subblock that are busy. @@ -1321,16 +1361,21 @@ static int virtio_mem_mb_unplug_any_sb_online(str= uct virtio_mem *vm, unsigned long mb_id, uint64_t *nb_sb) { - const unsigned long nr_pages =3D PFN_DOWN(vm->subblock_size); - unsigned long start_pfn; int rc, sb_id; =20 - /* - * TODO: To increase the performance we want to try bigger, consecutive - * subblocks first before falling back to single subblocks. Also, - * we should sense via something like is_mem_section_removable() - * first if it makes sense to go ahead any try to allocate. - */ + /* If possible, try to unplug the complete block in one shot. */ + if (*nb_sb >=3D vm->nb_sb_per_mb && + virtio_mem_mb_test_sb_plugged(vm, mb_id, 0, vm->nb_sb_per_mb)) { + rc =3D virtio_mem_mb_unplug_sb_online(vm, mb_id, 0, + vm->nb_sb_per_mb); + if (!rc) { + *nb_sb -=3D vm->nb_sb_per_mb; + goto unplugged; + } else if (rc !=3D -EBUSY) + return rc; + } + + /* Fallback to single subblocks. */ for (sb_id =3D vm->nb_sb_per_mb - 1; sb_id >=3D 0 && *nb_sb; sb_id--) { /* Find the next candidate subblock */ while (sb_id >=3D 0 && @@ -1339,34 +1384,15 @@ static int virtio_mem_mb_unplug_any_sb_online(str= uct virtio_mem *vm, if (sb_id < 0) break; =20 - start_pfn =3D PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) + - sb_id * vm->subblock_size); - rc =3D alloc_contig_range(start_pfn, start_pfn + nr_pages, - MIGRATE_MOVABLE, GFP_KERNEL); - if (rc =3D=3D -ENOMEM) - /* whoops, out of memory */ - return rc; - if (rc) - /* memory busy, we can't unplug this chunk */ + rc =3D virtio_mem_mb_unplug_sb_online(vm, mb_id, sb_id, 1); + if (rc =3D=3D -EBUSY) continue; - - /* Mark it as fake-offline before unplugging it */ - virtio_mem_set_fake_offline(start_pfn, nr_pages, true); - adjust_managed_page_count(pfn_to_page(start_pfn), -nr_pages); - - /* Try to unplug the allocated memory */ - rc =3D virtio_mem_mb_unplug_sb(vm, mb_id, sb_id, 1); - if (rc) { - /* Return the memory to the buddy. */ - virtio_mem_fake_online(start_pfn, nr_pages); + else if (rc) return rc; - } - - virtio_mem_mb_set_state(vm, mb_id, - VIRTIO_MEM_MB_STATE_ONLINE_PARTIAL); *nb_sb -=3D 1; } =20 +unplugged: /* * Once all subblocks of a memory block were unplugged, offline and * remove it. This will usually not fail, as no memory is in use --=20 2.25.3