Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp367461pxj; Thu, 3 Jun 2021 08:36:58 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwllrjlIBiwRhWXSb7D+YPiNM79/S4KMqKbXWooPwKDb2vA86AF8XDejHUjjwWzz6MFuuEL X-Received: by 2002:a05:6402:50d2:: with SMTP id h18mr149742edb.10.1622734617922; Thu, 03 Jun 2021 08:36:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1622734617; cv=none; d=google.com; s=arc-20160816; b=wUZGKt1q/TfcaNNVn505DiMgoS2WuyR3iSxICliE0YxWGK+a7rtlCVyNm/e3UQCUqu R5qCDuF1XLWBp9KqQ4wE4B087Z1/7VGuEGdLjTQIWmX+fJqTc8zTb3xlAePIh4fsBDLl bueXRm0lYO5Z5SaZOoqBbSy68T2VbwL2p6G2JAezCUjP8b/PwkTyi7+WaZ/rVzwuCJG5 GBn4pKSoTbEL7TgMejrRzlEYq7fLkweZORTOVipFCIlUuH/P4kbVM5tuIifkqTV5gGhA 7AyFplo1E65f52D3pNxL8C05tyGgLoxcGDpwX7uTUgyTMwS70oVHb36bLb97OYZfCAX7 k+Ww== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=HHcM+IgNoZ1DlGmkHUu2OKnC00B3Yz63C0t0xp+6d/4=; b=SPyfXlNsdC82WVGzRuq9vbsCqN7SYUJ2vGAVkctl4+pr8ccTD7XMlMHgNLEN3u/q// oLwhkBxSzeY1NcNzKpY0jlsEIbo2Slq0oM/f3oiysX52DO+Mu8QbbMn31o3o2M21YTDO ktAmvHYNMhGlDpC44gqLF0ZtcdviLdPUopNMgEWMdlRo0ol4JvF+K3cg2Z7ouNgMsyCy AuH17vHbbReR0AHYqOXCQFl8+eJnndsHXDRxRG6ZVNZloNU5Sgig/d7BeZKm93SlyhEc nn2/WiRa4QJlp+EMR/Z4I9EevaFdfAg82iB1W1uBeARSN7L/HNeTmbvvlqVtT2pdh0WH hkag== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=DdF0vWcG; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id x17si2555332edd.342.2021.06.03.08.36.35; Thu, 03 Jun 2021 08:36:57 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=DdF0vWcG; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229966AbhFCPgw (ORCPT + 99 others); Thu, 3 Jun 2021 11:36:52 -0400 Received: from us-smtp-delivery-124.mimecast.com ([216.205.24.124]:57760 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229791AbhFCPgw (ORCPT ); Thu, 3 Jun 2021 11:36:52 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1622734507; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=HHcM+IgNoZ1DlGmkHUu2OKnC00B3Yz63C0t0xp+6d/4=; b=DdF0vWcGiUdmDTPoL9Zv/FvYlpMwaon6lGPgAaQJrkesNStC0gYajfaPOZfqbfw98/BfNr qhlWtigtt/nVefUg1jHbAG8zY+mCkQh++wUSLu6WOW0B97mP+DK+c5T9bzwvbj1dCt2x+G 5hix03+/ZXUT77C2otGnXEXqGfNOAeI= Received: from mail-ej1-f72.google.com (mail-ej1-f72.google.com [209.85.218.72]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-243-lGwXjWJKMpaGXbnpaMPz5A-1; Thu, 03 Jun 2021 11:35:05 -0400 X-MC-Unique: lGwXjWJKMpaGXbnpaMPz5A-1 Received: by mail-ej1-f72.google.com with SMTP id p18-20020a1709067852b02903dab2a3e1easo2086801ejm.17 for ; Thu, 03 Jun 2021 08:35:05 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=HHcM+IgNoZ1DlGmkHUu2OKnC00B3Yz63C0t0xp+6d/4=; b=rNJjz3AWBHCYoyemddFr2y9mFgsit5803lcHdGGHmV6gBsFWWqZF7GZudy83uPAozN dU5DZYADhdK7DX8Zw2BPJGS01Iswju3TGgUFK9ThYC3tHDCk5I82ci9ESCNVR3L3QRVR 747/lSgd+BiJIAi5PGIR0JdxO8ZQWUXoTrRTFhBFWgb8aov342OE4JPailWKiTbOoHoQ gAZzqIKkU6Cjw34EUbB3iN6THnIXZctfsOtu3fSeGVztRd2/ewm8N5Oc6Jnu7XflGNcs dVz7xiOWbhzt+ZbgAV8Vwi6YxU6ixVsRL3yTy7iF+9rz2yv1VPgGoV5VXc32MuzhT/T9 C2xg== X-Gm-Message-State: AOAM532AbCEg/uqCi4/pVu9Bjf2/JpafodMa4bXdhL8TCcBoHKDwKN8q 1/+JDzSlgj0h8dhFw+Bw+A8oJDFJAMEFRi6E7qtJL16jv2u2wockHhN8IOpNefWIgcxVjfxtVNP 4ldbvlUbrpkwBmXpiwdTvU6zo X-Received: by 2002:a05:6402:702:: with SMTP id w2mr66159edx.189.1622734504256; Thu, 03 Jun 2021 08:35:04 -0700 (PDT) X-Received: by 2002:a05:6402:702:: with SMTP id w2mr66134edx.189.1622734504040; Thu, 03 Jun 2021 08:35:04 -0700 (PDT) Received: from steredhat ([5.170.129.82]) by smtp.gmail.com with ESMTPSA id p13sm1920100edq.67.2021.06.03.08.35.01 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 03 Jun 2021 08:35:03 -0700 (PDT) Date: Thu, 3 Jun 2021 17:34:59 +0200 From: Stefano Garzarella To: Arseny Krasnov Cc: Stefan Hajnoczi , "Michael S. Tsirkin" , Jason Wang , "David S. Miller" , Jakub Kicinski , Jorgen Hansen , Colin Ian King , Andra Paraschiv , Norbert Slusarek , kvm@vger.kernel.org, virtualization@lists.linux-foundation.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, oxffffaa@gmail.com Subject: Re: [PATCH v10 15/18] vhost/vsock: support SEQPACKET for transport Message-ID: <20210603153459.4qncp25nssuby4vp@steredhat> References: <20210520191357.1270473-1-arseny.krasnov@kaspersky.com> <20210520191916.1272540-1-arseny.krasnov@kaspersky.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Disposition: inline In-Reply-To: <20210520191916.1272540-1-arseny.krasnov@kaspersky.com> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, May 20, 2021 at 10:19:13PM +0300, Arseny Krasnov wrote: Please describe better the changes included in this patch in the first part of the commit message. >As vhost places data in buffers of guest's rx queue, keep SEQ_EOR >bit set only when last piece of data is copied. Otherwise we get >sequence packets for one socket in guest's rx queue with SEQ_EOR bit >set. Also remove ignore of non-stream type of packets, handle SEQPACKET >feature bit. > >Signed-off-by: Arseny Krasnov >--- > v9 -> v10: > 1) Move 'restore_flag' handling to 'payload_len' calculation > block. > > drivers/vhost/vsock.c | 44 +++++++++++++++++++++++++++++++++++++++---- > 1 file changed, 40 insertions(+), 4 deletions(-) > >diff --git a/drivers/vhost/vsock.c b/drivers/vhost/vsock.c >index 5e78fb719602..63d15beaad05 100644 >--- a/drivers/vhost/vsock.c >+++ b/drivers/vhost/vsock.c >@@ -31,7 +31,8 @@ > > enum { > VHOST_VSOCK_FEATURES = VHOST_FEATURES | >- (1ULL << VIRTIO_F_ACCESS_PLATFORM) >+ (1ULL << VIRTIO_F_ACCESS_PLATFORM) | >+ (1ULL << VIRTIO_VSOCK_F_SEQPACKET) > }; > > enum { >@@ -56,6 +57,7 @@ struct vhost_vsock { > atomic_t queued_replies; > > u32 guest_cid; >+ bool seqpacket_allow; > }; > > static u32 vhost_transport_get_local_cid(void) >@@ -112,6 +114,7 @@ vhost_transport_do_send_pkt(struct vhost_vsock *vsock, > size_t nbytes; > size_t iov_len, payload_len; > int head; >+ bool restore_flag = false; > > spin_lock_bh(&vsock->send_pkt_list_lock); > if (list_empty(&vsock->send_pkt_list)) { >@@ -168,9 +171,15 @@ vhost_transport_do_send_pkt(struct vhost_vsock *vsock, > /* If the packet is greater than the space available in the > * buffer, we split it using multiple buffers. > */ >- if (payload_len > iov_len - sizeof(pkt->hdr)) >+ if (payload_len > iov_len - sizeof(pkt->hdr)) { > payload_len = iov_len - sizeof(pkt->hdr); > Please, add a comment here to explain why we need this. >+ if (le32_to_cpu(pkt->hdr.flags) & >VIRTIO_VSOCK_SEQ_EOR) { >+ pkt->hdr.flags &= ~cpu_to_le32(VIRTIO_VSOCK_SEQ_EOR); >+ restore_flag = true; >+ } >+ } >+ > /* Set the correct length in the header */ > pkt->hdr.len = cpu_to_le32(payload_len); > >@@ -181,6 +190,9 @@ vhost_transport_do_send_pkt(struct vhost_vsock >*vsock, > break; > } > >+ if (restore_flag) >+ pkt->hdr.flags |= cpu_to_le32(VIRTIO_VSOCK_SEQ_EOR); >+ Maybe we can restore the flag only if we are queueing again the same packet, I mean in the `if (pkt->off < pkt->len) {` branch below. What do you think? > nbytes = copy_to_iter(pkt->buf + pkt->off, payload_len, > &iov_iter); > if (nbytes != payload_len) { >@@ -354,8 +366,7 @@ vhost_vsock_alloc_pkt(struct vhost_virtqueue *vq, > return NULL; > } > >- if (le16_to_cpu(pkt->hdr.type) == VIRTIO_VSOCK_TYPE_STREAM) >- pkt->len = le32_to_cpu(pkt->hdr.len); >+ pkt->len = le32_to_cpu(pkt->hdr.len); > > /* No payload */ > if (!pkt->len) >@@ -398,6 +409,8 @@ static bool vhost_vsock_more_replies(struct >vhost_vsock *vsock) > return val < vq->num; > } > >+static bool vhost_transport_seqpacket_allow(u32 remote_cid); >+ > static struct virtio_transport vhost_transport = { > .transport = { > .module = THIS_MODULE, >@@ -424,6 +437,10 @@ static struct virtio_transport vhost_transport = { > .stream_is_active = virtio_transport_stream_is_active, > .stream_allow = virtio_transport_stream_allow, > >+ .seqpacket_dequeue = virtio_transport_seqpacket_dequeue, >+ .seqpacket_enqueue = virtio_transport_seqpacket_enqueue, >+ .seqpacket_allow = vhost_transport_seqpacket_allow, >+ > .notify_poll_in = virtio_transport_notify_poll_in, > .notify_poll_out = virtio_transport_notify_poll_out, > .notify_recv_init = virtio_transport_notify_recv_init, >@@ -441,6 +458,22 @@ static struct virtio_transport vhost_transport = { > .send_pkt = vhost_transport_send_pkt, > }; > >+static bool vhost_transport_seqpacket_allow(u32 remote_cid) >+{ >+ struct vhost_vsock *vsock; >+ bool seqpacket_allow = false; >+ >+ rcu_read_lock(); >+ vsock = vhost_vsock_get(remote_cid); >+ >+ if (vsock) >+ seqpacket_allow = vsock->seqpacket_allow; >+ >+ rcu_read_unlock(); >+ >+ return seqpacket_allow; >+} >+ > static void vhost_vsock_handle_tx_kick(struct vhost_work *work) > { > struct vhost_virtqueue *vq = container_of(work, struct vhost_virtqueue, >@@ -785,6 +818,9 @@ static int vhost_vsock_set_features(struct vhost_vsock *vsock, u64 features) > goto err; > } > >+ if (features & (1ULL << VIRTIO_VSOCK_F_SEQPACKET)) >+ vsock->seqpacket_allow = true; >+ > for (i = 0; i < ARRAY_SIZE(vsock->vqs); i++) { > vq = &vsock->vqs[i]; > mutex_lock(&vq->mutex); >-- >2.25.1 >