Received: by 2002:a25:ca44:0:0:0:0:0 with SMTP id a65csp1028238ybg; Mon, 27 Jul 2020 06:11:15 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyRsRH+WS9MKxGTMT5EmwwLU2jQMw5TlfRhVC0lZho2Qxxa9hrHkVOMy1r+3knovWjBE0OO X-Received: by 2002:a05:6402:1b1c:: with SMTP id by28mr21243298edb.270.1595855475342; Mon, 27 Jul 2020 06:11:15 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1595855475; cv=none; d=google.com; s=arc-20160816; b=Ylo+HQlP5u+I96X9UgccWVOdh3y0swKgO8qUYZuJm3nIvqrcR3h3O9Y5MRpWmZtmu8 h6peySKaXSE1cZKnYdSUkesIakdiQO1UK67UAygrfX944AC1ygsOjWq6RuYHEBGKs35V SL4ExqHy0DTt0tXQAT/uxRM6wRa83JkAHigsypRGSubV6gEDmMYv8I8VK0nlVCfwTdCW GwuwAFx0wSJq5gU6GR1/rC4RAsUwNP7KJgkwX2SwVgNS1E5kBaQLr+DaU/Ou4wfu906X Bhlc1Omtexwmr6hkbutKxPUaLHbgR1aH9SOKMmlBketChJuJaeqpchrTkNEYsTxstH8H IaPA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:references:cc:to:from:subject; bh=KIuzdkxSLlPlvlr+ruWaoEHAMHw/XqaK2ccOphHR4jo=; b=Zf/FZBy5iST/4XT14UQOQHe2SlCZ4of0dFFcaM0xAQQ1+Kjbgj4uENFXFbZhL0Eaio kvPt4KaNTTPAMg67dC8YAXK6DKXuCSy0oq8bz5gd6D65H4a1O76klTSHqM4mJSUuQWvt D8n0iOwhJ5bg8CWcqpeWgiEELygkSHx1Cp2YeNYLulhMuwhAttMpJbFGN4+r6qJThSH4 5uSe+1GAP5dpvfG3UDCke4zH20cY/tCvrEWjeq2qjyrcNgIRWq5rsh+QCKo8GGrujOLk i0kC0i5rmVTx5zZXdp5zWQgt6aij4KcdTSeaKWyBjnG+PT8g4tnNXPcLLNWMlmyHG8dS rKgw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id i22si5626613edg.217.2020.07.27.06.10.52; Mon, 27 Jul 2020 06:11:15 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728602AbgG0NKP (ORCPT + 99 others); Mon, 27 Jul 2020 09:10:15 -0400 Received: from out30-131.freemail.mail.aliyun.com ([115.124.30.131]:56100 "EHLO out30-131.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728141AbgG0NKO (ORCPT ); Mon, 27 Jul 2020 09:10:14 -0400 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R111e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e01422;MF=shile.zhang@linux.alibaba.com;NM=1;PH=DS;RN=8;SR=0;TI=SMTPD_---0U3zRBsO_1595855409; Received: from B-J2UMLVDL-1650.local(mailfrom:shile.zhang@linux.alibaba.com fp:SMTPD_---0U3zRBsO_1595855409) by smtp.aliyun-inc.com(127.0.0.1); Mon, 27 Jul 2020 21:10:10 +0800 Subject: Re: [PATCH v2] virtio_ring: use alloc_pages_node for NUMA-aware allocation From: Shile Zhang To: "Michael S. Tsirkin" Cc: Jason Wang , virtualization@lists.linux-foundation.org, linux-kernel@vger.kernel.org, kernel test robot , Jiang Liu , linux-pci@vger.kernel.org, bhelgaas@google.com References: <20200721070013.62894-1-shile.zhang@linux.alibaba.com> <20200721041550-mutt-send-email-mst@kernel.org> Message-ID: <222b40f1-a36c-0375-e965-cd949e8b9eeb@linux.alibaba.com> Date: Mon, 27 Jul 2020 21:10:09 +0800 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2020/7/21 19:28, Shile Zhang wrote: > > > On 2020/7/21 16:18, Michael S. Tsirkin wrote: >> On Tue, Jul 21, 2020 at 03:00:13PM +0800, Shile Zhang wrote: >>> Use alloc_pages_node() allocate memory for vring queue with proper >>> NUMA affinity. >>> >>> Reported-by: kernel test robot >>> Suggested-by: Jiang Liu >>> Signed-off-by: Shile Zhang >> >> Do you observe any performance gains from this patch? > > Thanks for your comments! > Yes, the bandwidth can boost more than doubled (from 30Gbps to 80GBps) > with this changes in my test env (8 numa nodes), with netperf test. > >> >> I also wonder why isn't the probe code run on the correct numa node? >> That would fix a wide class of issues like this without need to tweak >> drivers. > > Good point, I'll check this, thanks! Sorry, I have no idea about how the probe code to grab the appropriate NUMA node. > >> >> Bjorn, what do you think? Was this considered? Hi Bjorn, Could you please give any comments about this issue? Thanks! >> >>> --- >>> Changelog >>> v1 -> v2: >>> - fixed compile warning reported by LKP. >>> --- >>>   drivers/virtio/virtio_ring.c | 10 ++++++---- >>>   1 file changed, 6 insertions(+), 4 deletions(-) >>> >>> diff --git a/drivers/virtio/virtio_ring.c b/drivers/virtio/virtio_ring.c >>> index 58b96baa8d48..d38fd6872c8c 100644 >>> --- a/drivers/virtio/virtio_ring.c >>> +++ b/drivers/virtio/virtio_ring.c >>> @@ -276,9 +276,11 @@ static void *vring_alloc_queue(struct >>> virtio_device *vdev, size_t size, >>>           return dma_alloc_coherent(vdev->dev.parent, size, >>>                         dma_handle, flag); >>>       } else { >>> -        void *queue = alloc_pages_exact(PAGE_ALIGN(size), flag); >>> - >>> -        if (queue) { >>> +        void *queue = NULL; >>> +        struct page *page = >>> alloc_pages_node(dev_to_node(vdev->dev.parent), >>> +                             flag, get_order(size)); >>> +        if (page) { >>> +            queue = page_address(page); >>>               phys_addr_t phys_addr = virt_to_phys(queue); >>>               *dma_handle = (dma_addr_t)phys_addr; >>> @@ -308,7 +310,7 @@ static void vring_free_queue(struct virtio_device >>> *vdev, size_t size, >>>       if (vring_use_dma_api(vdev)) >>>           dma_free_coherent(vdev->dev.parent, size, queue, dma_handle); >>>       else >>> -        free_pages_exact(queue, PAGE_ALIGN(size)); >>> +        free_pages((unsigned long)queue, get_order(size)); >>>   } >>>   /* >>> -- >>> 2.24.0.rc2