Received: by 2002:ac0:a5a6:0:0:0:0:0 with SMTP id m35-v6csp3630220imm; Mon, 10 Sep 2018 22:39:42 -0700 (PDT) X-Google-Smtp-Source: ANB0VdY92YNd0AVqSXF5g2mgkvBjmyYiICsLLjfPyOB+cZk6Q7CvyO2zPQLyCEp06TUwu4WHvubq X-Received: by 2002:a63:8442:: with SMTP id k63-v6mr26165679pgd.388.1536644382473; Mon, 10 Sep 2018 22:39:42 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1536644382; cv=none; d=google.com; s=arc-20160816; b=luql770xuA2pkIJY0sxNp+awdPeKemHZYxNNpwA1bRU5OTKcIYjku/DGKtDy7DDV30 nHijzitFNYnWDeGwyfybFLvg4oJg4nfN1d9SChSUQv8HDTzDNKViqwqtmcoPrj1QSZBy ou95lgVDgUx1q8avVwM6YOvyO20/PlqiAvM2h645CF/y2RpqC890+UYtJ50zG6/UM1db KaCYdA/jdKFEajdg03Nglpe8uJ0SDtJhV8rw5MA36S/gwtsM7CJx1Gc7et+EaKdYTjDl vaepfsFlMrHeh9vaQW7FCEv+hSt2B7znj6NaSTs37ER011wFnccSKzfxEQCeGmCg1Qvt Bxtw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-transfer-encoding:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=2gPLJcFW7Q9s48qEQXk7fWeXLM53940Dsi2Gub4QTJo=; b=D7z7y8ORAVjTSpTtkMPwb52WIBgDA3LP95pZaryRl9MmMuPyFWNDG95w82HSSSh1Dz Z6kRcpkFemW2ZkJrp9PdHe9QrPMPVOQLT3id66MIshgPAlUw1kk78Ch3OpOxW5xvIQcz fR7JohqReChq77yAeX6/ZcY/FdGet4pvAtmYAK73FnWXorhaSAf9clf20U+9WeFSQ4Hy 9dFJTJDqniWr0nQJd09PqUHxP74sMcWCMiHFjhWV8EINo4yuzkHA4ymgcgzFg5qDppf1 lKtPXNzDdze87ZAmaDl43zrdOYMbsP3mFbwTKuR5HoHfmo2gdSZryLmJQjo4t8yn/KPn 5VFQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id d7-v6si16427117pls.136.2018.09.10.22.39.26; Mon, 10 Sep 2018 22:39:42 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726511AbeIKKgD (ORCPT + 99 others); Tue, 11 Sep 2018 06:36:03 -0400 Received: from mga09.intel.com ([134.134.136.24]:11489 "EHLO mga09.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726301AbeIKKgC (ORCPT ); Tue, 11 Sep 2018 06:36:02 -0400 X-Amp-Result: UNSCANNABLE X-Amp-File-Uploaded: False Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga102.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 10 Sep 2018 22:38:27 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.53,359,1531810800"; d="scan'208";a="87696576" Received: from btwcube1.sh.intel.com (HELO debian) ([10.67.104.194]) by fmsmga004.fm.intel.com with ESMTP; 10 Sep 2018 22:38:25 -0700 Date: Tue, 11 Sep 2018 13:37:26 +0800 From: Tiwei Bie To: Jason Wang Cc: "Michael S. Tsirkin" , virtualization@lists.linux-foundation.org, linux-kernel@vger.kernel.org, netdev@vger.kernel.org, virtio-dev@lists.oasis-open.org, wexu@redhat.com, jfreimann@redhat.com Subject: Re: [virtio-dev] Re: [PATCH net-next v2 0/5] virtio: support packed ring Message-ID: <20180911053726.GA7472@debian> References: <20180711022711.7090-1-tiwei.bie@intel.com> <20180827170005-mutt-send-email-mst@kernel.org> <20180907012225.GA32677@debian> <20180907084509-mutt-send-email-mst@kernel.org> <20180910030053.GA15645@debian> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Sep 10, 2018 at 11:33:17AM +0800, Jason Wang wrote: > On 2018年09月10日 11:00, Tiwei Bie wrote: > > On Fri, Sep 07, 2018 at 09:00:49AM -0400, Michael S. Tsirkin wrote: > > > On Fri, Sep 07, 2018 at 09:22:25AM +0800, Tiwei Bie wrote: > > > > On Mon, Aug 27, 2018 at 05:00:40PM +0300, Michael S. Tsirkin wrote: > > > > > Are there still plans to test the performance with vost pmd? > > > > > vhost doesn't seem to show a performance gain ... > > > > > > > > > I tried some performance tests with vhost PMD. In guest, the > > > > XDP program will return XDP_DROP directly. And in host, testpmd > > > > will do txonly fwd. > > > > > > > > When burst size is 1 and packet size is 64 in testpmd and > > > > testpmd needs to iterate 5 Tx queues (but only the first two > > > > queues are enabled) to prepare and inject packets, I got ~12% > > > > performance boost (5.7Mpps -> 6.4Mpps). And if the vhost PMD > > > > is faster (e.g. just need to iterate the first two queues to > > > > prepare and inject packets), then I got similar performance > > > > for both rings (~9.9Mpps) (packed ring's performance can be > > > > lower, because it's more complicated in driver.) > > > > > > > > I think packed ring makes vhost PMD faster, but it doesn't make > > > > the driver faster. In packed ring, the ring is simplified, and > > > > the handling of the ring in vhost (device) is also simplified, > > > > but things are not simplified in driver, e.g. although there is > > > > no desc table in the virtqueue anymore, driver still needs to > > > > maintain a private desc state table (which is still managed as > > > > a list in this patch set) to support the out-of-order desc > > > > processing in vhost (device). > > > > > > > > I think this patch set is mainly to make the driver have a full > > > > functional support for the packed ring, which makes it possible > > > > to leverage the packed ring feature in vhost (device). But I'm > > > > not sure whether there is any other better idea, I'd like to > > > > hear your thoughts. Thanks! > > > Just this: Jens seems to report a nice gain with virtio and > > > vhost pmd across the board. Try to compare virtio and > > > virtio pmd to see what does pmd do better? > > The virtio PMD (drivers/net/virtio) in DPDK doesn't need to share > > the virtio ring operation code with other drivers and is highly > > optimized for network. E.g. in Rx, the Rx burst function won't > > chain descs. So the ID management for the Rx ring can be quite > > simple and straightforward, we just need to initialize these IDs > > when initializing the ring and don't need to change these IDs > > in data path anymore (the mergable Rx code in that patch set > > assumes the descs will be written back in order, which should be > > fixed. I.e., the ID in the desc should be used to index vq->descx[]). > > The Tx code in that patch set also assumes the descs will be > > written back by device in order, which should be fixed. > > Yes it is. I think I've pointed it out in some early version of pmd patch. > So I suspect part (or all) of the boost may come from in order feature. > > > > > But in kernel virtio driver, the virtio_ring.c is very generic. > > The enqueue (virtqueue_add()) and dequeue (virtqueue_get_buf_ctx()) > > functions need to support all the virtio devices and should be > > able to handle all the possible cases that may happen. So although > > the packed ring can be very efficient in some cases, currently > > the room to optimize the performance in kernel's virtio_ring.c > > isn't that much. If we want to take the fully advantage of the > > packed ring's efficiency, we need some further e.g. API changes > > in virtio_ring.c, which shouldn't be part of this patch set. > > Could you please share more thoughts on this e.g how to improve the API? > Notice since the API is shared by both split ring and packed ring, it may > improve the performance of split ring as well. One can easily imagine a > batching API, but it does not have many real users now, the only case is the > XDP transmission which can accept an array of XDP frames. I don't have detailed thoughts on this yet. But kernel's virtio_ring.c is quite generic compared with what we did in virtio PMD. > > > So > > I still think this patch set is mainly to make the kernel virtio > > driver to have a full functional support of the packed ring, and > > we can't expect impressive performance boost with it. > > We can only gain when virtio ring layout is the bottleneck. If there're > bottlenecks elsewhere, we probably won't see any increasing in the numbers. > Vhost-net is an example, and lots of optimizations have proved that virtio > ring is not the main bottleneck for the current codes. I suspect it also the > case of virtio driver. Did perf tell us any interesting things in virtio > driver? > > Thanks > > > > > > > > > > > On Wed, Jul 11, 2018 at 10:27:06AM +0800, Tiwei Bie wrote: > > > > > > Hello everyone, > > > > > > > > > > > > This patch set implements packed ring support in virtio driver. > > > > > > > > > > > > Some functional tests have been done with Jason's > > > > > > packed ring implementation in vhost: > > > > > > > > > > > > https://lkml.org/lkml/2018/7/3/33 > > > > > > > > > > > > Both of ping and netperf worked as expected. > > > > > > > > > > > > v1 -> v2: > > > > > > - Use READ_ONCE() to read event off_wrap and flags together (Jason); > > > > > > - Add comments related to ccw (Jason); > > > > > > > > > > > > RFC (v6) -> v1: > > > > > > - Avoid extra virtio_wmb() in virtqueue_enable_cb_delayed_packed() > > > > > > when event idx is off (Jason); > > > > > > - Fix bufs calculation in virtqueue_enable_cb_delayed_packed() (Jason); > > > > > > - Test the state of the desc at used_idx instead of last_used_idx > > > > > > in virtqueue_enable_cb_delayed_packed() (Jason); > > > > > > - Save wrap counter (as part of queue state) in the return value > > > > > > of virtqueue_enable_cb_prepare_packed(); > > > > > > - Refine the packed ring definitions in uapi; > > > > > > - Rebase on the net-next tree; > > > > > > > > > > > > RFC v5 -> RFC v6: > > > > > > - Avoid tracking addr/len/flags when DMA API isn't used (MST/Jason); > > > > > > - Define wrap counter as bool (Jason); > > > > > > - Use ALIGN() in vring_init_packed() (Jason); > > > > > > - Avoid using pointer to track `next` in detach_buf_packed() (Jason); > > > > > > - Add comments for barriers (Jason); > > > > > > - Don't enable RING_PACKED on ccw for now (noticed by Jason); > > > > > > - Refine the memory barrier in virtqueue_poll(); > > > > > > - Add a missing memory barrier in virtqueue_enable_cb_delayed_packed(); > > > > > > - Remove the hacks in virtqueue_enable_cb_prepare_packed(); > > > > > > > > > > > > RFC v4 -> RFC v5: > > > > > > - Save DMA addr, etc in desc state (Jason); > > > > > > - Track used wrap counter; > > > > > > > > > > > > RFC v3 -> RFC v4: > > > > > > - Make ID allocation support out-of-order (Jason); > > > > > > - Various fixes for EVENT_IDX support; > > > > > > > > > > > > RFC v2 -> RFC v3: > > > > > > - Split into small patches (Jason); > > > > > > - Add helper virtqueue_use_indirect() (Jason); > > > > > > - Just set id for the last descriptor of a list (Jason); > > > > > > - Calculate the prev in virtqueue_add_packed() (Jason); > > > > > > - Fix/improve desc suppression code (Jason/MST); > > > > > > - Refine the code layout for XXX_split/packed and wrappers (MST); > > > > > > - Fix the comments and API in uapi (MST); > > > > > > - Remove the BUG_ON() for indirect (Jason); > > > > > > - Some other refinements and bug fixes; > > > > > > > > > > > > RFC v1 -> RFC v2: > > > > > > - Add indirect descriptor support - compile test only; > > > > > > - Add event suppression supprt - compile test only; > > > > > > - Move vring_packed_init() out of uapi (Jason, MST); > > > > > > - Merge two loops into one in virtqueue_add_packed() (Jason); > > > > > > - Split vring_unmap_one() for packed ring and split ring (Jason); > > > > > > - Avoid using '%' operator (Jason); > > > > > > - Rename free_head -> next_avail_idx (Jason); > > > > > > - Add comments for virtio_wmb() in virtqueue_add_packed() (Jason); > > > > > > - Some other refinements and bug fixes; > > > > > > > > > > > > Thanks! > > > > > > > > > > > > Tiwei Bie (5): > > > > > > virtio: add packed ring definitions > > > > > > virtio_ring: support creating packed ring > > > > > > virtio_ring: add packed ring support > > > > > > virtio_ring: add event idx support in packed ring > > > > > > virtio_ring: enable packed ring > > > > > > > > > > > > drivers/s390/virtio/virtio_ccw.c | 14 + > > > > > > drivers/virtio/virtio_ring.c | 1365 ++++++++++++++++++++++------ > > > > > > include/linux/virtio_ring.h | 8 +- > > > > > > include/uapi/linux/virtio_config.h | 3 + > > > > > > include/uapi/linux/virtio_ring.h | 43 + > > > > > > 5 files changed, 1157 insertions(+), 276 deletions(-) > > > > > > > > > > > > -- > > > > > > 2.18.0 > > > > > --------------------------------------------------------------------- > > > > > To unsubscribe, e-mail: virtio-dev-unsubscribe@lists.oasis-open.org > > > > > For additional commands, e-mail: virtio-dev-help@lists.oasis-open.org > > > > > >