Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp1419344pxj; Fri, 18 Jun 2021 06:46:06 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwostVwfScNwsccRxULurXniFYx8xw61Sa42U455owJjLjuiVXvyzCL8fJ4L535HPculbHf X-Received: by 2002:a05:6e02:8ee:: with SMTP id n14mr6183132ilt.205.1624023966776; Fri, 18 Jun 2021 06:46:06 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1624023966; cv=none; d=google.com; s=arc-20160816; b=zg23yiSNHHrTSYyGFYolcZO7JoNiBcfbbgyatviGpdSaYJxK+xJwCMP2WtdJdklg4Z +W/28EP3JcLXfqdQyrRu41SRybAv52MnWKZRti8IdmtoWqyD/xsWRB2Wl2mImU9y88y9 VXtuhEwOLxhnFMwJwhYF36nc8pMQp3wu71WhhOsHN8Jh3/Uo2hcey7J7lNMgYSmBm5U3 Lwxk6XZl2G9fQQRU8msRa48uJBCU6Fx7kYb4+Rq4L2dxE2pQQv2+/CfPYmAZGS0ol01a EH1rrWChitMmT9169WKqI3w7+yQz+lei+NzyTVOYPgeECTlHQzOii2HdEkg2W3Sf5cay 00fw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=TeeOPx856RlPCCZLEcV71KwWexsoQ+3mbRzQGTsGO84=; b=xUAs1R8ZbpoeHaYcgEoeAZcymSeTskeyHdW0nBNTduF/KYdvBxqYzzxBSJS4m0jT1L hCq8kexMhv//xLRNARj84EYAmNpIYmGT7NojQsWTBAuCB1UG2i/sGTI7K6cBjJnlBxKB PbbdlISadZH/1mMQAePR+p2aW0RudM9AoDaDz4l9ghJnLhWM1h9cUNsSTgs/RWZDpiZN WmATBVqLgaWBmoFBeeVV/sVgYlVgmFVsP4j5ezs9ifHlnBtKeAi7aMTlK4FmvWZTwJXQ 7vjLVILXj1ya7TDKKPseSXU+JHXbqymk+yF9FACj3lQqy8c3Rr5JOaN8TtoI5S5l9HMg Q+2w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=Dk85mKqz; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id l15si9777959jaj.27.2021.06.18.06.45.52; Fri, 18 Jun 2021 06:46:06 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=Dk85mKqz; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233967AbhFRNql (ORCPT + 99 others); Fri, 18 Jun 2021 09:46:41 -0400 Received: from us-smtp-delivery-124.mimecast.com ([170.10.133.124]:44637 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233893AbhFRNqk (ORCPT ); Fri, 18 Jun 2021 09:46:40 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1624023871; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=TeeOPx856RlPCCZLEcV71KwWexsoQ+3mbRzQGTsGO84=; b=Dk85mKqzUt1Uq7PV+32/LtUr5kqanoPDbILxtmBkbHwsyRqbfoOLOf2Y4kENbEHFfGMsdm i5T+eW0tZ/qB9E6CLztmnJE9cOJy3SnZMysMPKexAJDpKwWVi13iso4D6Y+SPVoVSvUy7x 5nk95FrETVZWdUCGDQtc7wuqqSo82AE= Received: from mail-ed1-f70.google.com (mail-ed1-f70.google.com [209.85.208.70]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-34-ChLHSrNiNeSUjo1IXijYWg-1; Fri, 18 Jun 2021 09:44:30 -0400 X-MC-Unique: ChLHSrNiNeSUjo1IXijYWg-1 Received: by mail-ed1-f70.google.com with SMTP id v8-20020a0564023488b0290393873961f6so3643733edc.17 for ; Fri, 18 Jun 2021 06:44:30 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=TeeOPx856RlPCCZLEcV71KwWexsoQ+3mbRzQGTsGO84=; b=H2Pk2b//CYtCXvQIprl9CyZTw9TQ14U5PVB1CnCxFMupRPkvCEv3tYXL2Sw3dRinO7 gLzU9KCSsc6LSYt0QUV2Wx7SaPsUxCDU2OvPdVBiUWsZiGxn7bGKu2VeDE6qm9JarsLo 6lePHXNu/TJeEEStwnJhtMnpNIGvzDI4tivROLwsh1oK/a79m6YpmiaSCJZZh51QHDLk ZWz+KEMdmaN+klKmpMDSIvN1/LSD0oaeJ/2ilOD5O5HfEPA9xHLRlnuySTahsEAoKSb4 qkj3Klr75KzkXBHg1jKJRpEXwVeybEsrZEq/fAc37UG3xEh3clX0hEuSL9TzNXUlLrA5 9kRA== X-Gm-Message-State: AOAM532mFvfyeih+8RX8S3WLhF4TmYdjZ5gidCBITQt215cZJgiQ9kEG 1WvyFQMbjyt/F97pfb+ABys8aGHeS02D7KXhCJc/GfChPwpcQTVVDccwp71zsFZLkiFWQufwvcq SGnTjiKtSvg9ox/NTSc7KgCoM X-Received: by 2002:a50:f9ca:: with SMTP id a10mr4968350edq.97.1624023869087; Fri, 18 Jun 2021 06:44:29 -0700 (PDT) X-Received: by 2002:a50:f9ca:: with SMTP id a10mr4968330edq.97.1624023868927; Fri, 18 Jun 2021 06:44:28 -0700 (PDT) Received: from steredhat.lan ([5.170.130.11]) by smtp.gmail.com with ESMTPSA id da28sm1414842edb.0.2021.06.18.06.44.26 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 18 Jun 2021 06:44:28 -0700 (PDT) Date: Fri, 18 Jun 2021 15:44:23 +0200 From: Stefano Garzarella To: Arseny Krasnov Cc: Stefan Hajnoczi , "Michael S. Tsirkin" , Jason Wang , "David S. Miller" , Jakub Kicinski , Norbert Slusarek , Andra Paraschiv , Colin Ian King , kvm@vger.kernel.org, virtualization@lists.linux-foundation.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, oxffffaa@gmail.com Subject: Re: [PATCH v11 11/18] virtio/vsock: dequeue callback for SOCK_SEQPACKET Message-ID: <20210618134423.mksgnbmchmow4sgh@steredhat.lan> References: <20210611110744.3650456-1-arseny.krasnov@kaspersky.com> <20210611111241.3652274-1-arseny.krasnov@kaspersky.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii; format=flowed Content-Disposition: inline In-Reply-To: <20210611111241.3652274-1-arseny.krasnov@kaspersky.com> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Arseny, the series looks great, I have just a question below about seqpacket_dequeue. I also sent a couple a simple fixes, it would be great if you can review them: https://lore.kernel.org/netdev/20210618133526.300347-1-sgarzare@redhat.com/ On Fri, Jun 11, 2021 at 02:12:38PM +0300, Arseny Krasnov wrote: >Callback fetches RW packets from rx queue of socket until whole record >is copied(if user's buffer is full, user is not woken up). This is done >to not stall sender, because if we wake up user and it leaves syscall, >nobody will send credit update for rest of record, and sender will wait >for next enter of read syscall at receiver's side. So if user buffer is >full, we just send credit update and drop data. > >Signed-off-by: Arseny Krasnov >--- > v10 -> v11: > 1) 'msg_count' field added to count current number of EORs. > 2) 'msg_ready' argument removed from callback. > 3) If 'memcpy_to_msg()' failed during copy loop, there will be > no next attempts to copy data, rest of record will be freed. > > include/linux/virtio_vsock.h | 5 ++ > net/vmw_vsock/virtio_transport_common.c | 84 +++++++++++++++++++++++++ > 2 files changed, 89 insertions(+) > >diff --git a/include/linux/virtio_vsock.h b/include/linux/virtio_vsock.h >index dc636b727179..1d9a302cb91d 100644 >--- a/include/linux/virtio_vsock.h >+++ b/include/linux/virtio_vsock.h >@@ -36,6 +36,7 @@ struct virtio_vsock_sock { > u32 rx_bytes; > u32 buf_alloc; > struct list_head rx_queue; >+ u32 msg_count; > }; > > struct virtio_vsock_pkt { >@@ -80,6 +81,10 @@ virtio_transport_dgram_dequeue(struct vsock_sock *vsk, > struct msghdr *msg, > size_t len, int flags); > >+ssize_t >+virtio_transport_seqpacket_dequeue(struct vsock_sock *vsk, >+ struct msghdr *msg, >+ int flags); > s64 virtio_transport_stream_has_data(struct vsock_sock *vsk); > s64 virtio_transport_stream_has_space(struct vsock_sock *vsk); > >diff --git a/net/vmw_vsock/virtio_transport_common.c b/net/vmw_vsock/virtio_transport_common.c >index ad0d34d41444..1e1df19ec164 100644 >--- a/net/vmw_vsock/virtio_transport_common.c >+++ b/net/vmw_vsock/virtio_transport_common.c >@@ -393,6 +393,78 @@ virtio_transport_stream_do_dequeue(struct vsock_sock *vsk, > return err; > } > >+static int virtio_transport_seqpacket_do_dequeue(struct vsock_sock *vsk, >+ struct msghdr *msg, >+ int flags) >+{ >+ struct virtio_vsock_sock *vvs = vsk->trans; >+ struct virtio_vsock_pkt *pkt; >+ int dequeued_len = 0; >+ size_t user_buf_len = msg_data_left(msg); >+ bool copy_failed = false; >+ bool msg_ready = false; >+ >+ spin_lock_bh(&vvs->rx_lock); >+ >+ if (vvs->msg_count == 0) { >+ spin_unlock_bh(&vvs->rx_lock); >+ return 0; >+ } >+ >+ while (!msg_ready) { >+ pkt = list_first_entry(&vvs->rx_queue, struct virtio_vsock_pkt, list); >+ >+ if (!copy_failed) { >+ size_t pkt_len; >+ size_t bytes_to_copy; >+ >+ pkt_len = (size_t)le32_to_cpu(pkt->hdr.len); >+ bytes_to_copy = min(user_buf_len, pkt_len); >+ >+ if (bytes_to_copy) { >+ int err; >+ >+ /* sk_lock is held by caller so no one else can dequeue. >+ * Unlock rx_lock since memcpy_to_msg() may sleep. >+ */ >+ spin_unlock_bh(&vvs->rx_lock); >+ >+ err = memcpy_to_msg(msg, pkt->buf, bytes_to_copy); >+ if (err) { >+ /* Copy of message failed, set flag to skip >+ * copy path for rest of fragments. Rest of >+ * fragments will be freed without copy. >+ */ >+ copy_failed = true; >+ dequeued_len = err; If we fail to copy the message we will discard the entire packet. Is it acceptable for the user point of view, or we should leave the packet in the queue and the user can retry, maybe with a different buffer? Then we can remove the packets only when we successfully copied all the fragments. I'm not sure make sense, maybe better to check also other implementations :-) Thanks, Stefano >+ } else { >+ user_buf_len -= bytes_to_copy; >+ } >+ >+ spin_lock_bh(&vvs->rx_lock); >+ } >+ >+ if (dequeued_len >= 0) >+ dequeued_len += pkt_len; >+ } >+ >+ if (le32_to_cpu(pkt->hdr.flags) & VIRTIO_VSOCK_SEQ_EOR) { >+ msg_ready = true; >+ vvs->msg_count--; >+ } >+ >+ virtio_transport_dec_rx_pkt(vvs, pkt); >+ list_del(&pkt->list); >+ virtio_transport_free_pkt(pkt); >+ } >+ >+ spin_unlock_bh(&vvs->rx_lock); >+ >+ virtio_transport_send_credit_update(vsk); >+ >+ return dequeued_len; >+} >+ > ssize_t > virtio_transport_stream_dequeue(struct vsock_sock *vsk, > struct msghdr *msg, >@@ -405,6 +477,18 @@ virtio_transport_stream_dequeue(struct vsock_sock *vsk, > } > EXPORT_SYMBOL_GPL(virtio_transport_stream_dequeue); > >+ssize_t >+virtio_transport_seqpacket_dequeue(struct vsock_sock *vsk, >+ struct msghdr *msg, >+ int flags) >+{ >+ if (flags & MSG_PEEK) >+ return -EOPNOTSUPP; >+ >+ return virtio_transport_seqpacket_do_dequeue(vsk, msg, flags); >+} >+EXPORT_SYMBOL_GPL(virtio_transport_seqpacket_dequeue); >+ > int > virtio_transport_dgram_dequeue(struct vsock_sock *vsk, > struct msghdr *msg, >-- >2.25.1 >