Received: by 10.192.165.148 with SMTP id m20csp2234465imm; Thu, 26 Apr 2018 07:53:40 -0700 (PDT) X-Google-Smtp-Source: AIpwx4/T8V8KiYeAVGXgoltk8xj4FDZZH+ZJhi7ySsMOdXActt8knGnOb5DmSIN5+vEqsKTXrG20 X-Received: by 10.99.112.82 with SMTP id a18mr18588323pgn.148.1524754420789; Thu, 26 Apr 2018 07:53:40 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1524754420; cv=none; d=google.com; s=arc-20160816; b=rKbXmFGjkSLib/wmLk1hmDKF+d7HKKEhma52y/u9vkMTtRyh55Bv1o2mbrMgll0aiN MuAGZbvybHhitjnRAiGJSjN8NmBs8c4r88VQWIivsQcOTXucTzUjy2HADcuc2W+6tpC3 +OUwW3bJIgvyFnlskXIRlalQoRWpgBkJ8pfDeNk6XTN629+7Rwn/ZHVhDuwqQMjZJocl KvNXyPzf6Fph0ZfKVsxaQQ/GqyF1rwOGXRth1e2uySLXhywK7shMPAghQMZ5f9PeIIC6 gyY8ZZMb4yQwySmLNy4Vu7xqars+QWAAnfarq0DhiiIP0UD0erCscWGi3CYNT6yrWugH 75tA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:dkim-signature:arc-authentication-results; bh=zREpjI59PPH0VglY22tCSFujyK2RZekNVWn0BYcgOEY=; b=CwiySmXYAuPcDRXWyS5RCJ7bBrSBA/Z5yZytBMzu0pTA2K6zbU9y2pNTDzZld42WC9 H2sWIb3XHwfCx+eVjhDDzWUgOQ2J9OGlTFGcX5E+8hPlJX0Qaktg5IquE8ZiOOyjwdah j0OAziCrsH+fwaaHYveVrLT1/xjX95ULsmC1CoAKdF8RMxGjxSdL5/wR0NDu25/0T/tJ FG2H9j0Zm12z84I8EeRn2yt2Kz2f/a8lLq3mT0FcsS1QR2rLVOA5mSPTcorq3JtZMkqT Gl6ZR5HOJGyx20JD6OsUL0+OjNfyKhfVewmnjzdc6xuexrq2Uop7QF4wEmN2P2WbaX9P x3yw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=llVwpi2/; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j33-v6si19009700pld.395.2018.04.26.07.53.26; Thu, 26 Apr 2018 07:53:40 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=llVwpi2/; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756644AbeDZOvo (ORCPT + 99 others); Thu, 26 Apr 2018 10:51:44 -0400 Received: from mail-pf0-f178.google.com ([209.85.192.178]:32831 "EHLO mail-pf0-f178.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756548AbeDZOvD (ORCPT ); Thu, 26 Apr 2018 10:51:03 -0400 Received: by mail-pf0-f178.google.com with SMTP id f15so18586752pfn.0 for ; Thu, 26 Apr 2018 07:51:03 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=zREpjI59PPH0VglY22tCSFujyK2RZekNVWn0BYcgOEY=; b=llVwpi2/hfQUOvNBT1AwLdUsLg1CuwrtjtZdR0kYSNyk9vagO/HQ/4tCch2GRxus0i jFXWtHW0Q21fmU1JgtGqHBa9ucpmM+lybn4oLJ7IRHchtqdf0IkvYJC8N/x/fY2QKQGR yzdA1nrWRiCnptwMOPzOg5u6MhVMx9DHg2uP2J5mHHwcBKnUV2m6Ytn+wq8W/jWvs8zU n4P44hFMBHLdR6Asx/IsvBrl6BO1/YK9bPzEg4f6TxHX8AqEWGsH4syAaxE5c6eemutu GlijPWBJgl3aEjg5FZRmSCb42bGGJOIoA7AuedJ8+m/Jg+aI05fnVP9919mLUfgWTRsv Yyzw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=zREpjI59PPH0VglY22tCSFujyK2RZekNVWn0BYcgOEY=; b=PnmeQ7VR8F0orSu1FI1pavMmMRu4UOIxtWZws0AJxNjetAPoWSXvvD5TXpr/y+/M2p faMjfUL/KfTQ5VJE0eWTMsJaZ+gY7NnW5XIswpAP0Tlf+fAknXOv4RN8ztkyzhto+Aej 2zt50bwPAldQAOH1eA9ZF58Bj8H85lrwoVzETC3YiOJm3y3zaNUAY1Dp18O0nd0nuut9 kjrMCOAUsw3SGUAnPOPVP53euTauqG+x06Vwhjiyw0EcZcWWvbqgP8CpfvHt7POnWrEw Kd5QV7lEMrbPqY3ksAeYMf/Do/11hubx07Sg6gmQ+yGS4J4QVic62kFMc/Ta0PXB4MT7 Uk6w== X-Gm-Message-State: ALQs6tCBdBCUEiMiuD6JGjQTyzfNIX6omocuLSyrDT2g7GNo65KiZNRJ L1L3aZs7ury9oVgxwNkGH1zjow== X-Received: by 2002:a17:902:b081:: with SMTP id p1-v6mr35250683plr.31.1524754262924; Thu, 26 Apr 2018 07:51:02 -0700 (PDT) Received: from localhost ([2620:15c:2c4:201:f5a:7eca:440a:3ead]) by smtp.gmail.com with ESMTPSA id v1sm37586195pfg.144.2018.04.26.07.51.01 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Thu, 26 Apr 2018 07:51:01 -0700 (PDT) From: Eric Dumazet To: "David S . Miller" Cc: netdev , Andy Lutomirski , linux-kernel , linux-mm , Ka-Cheong Poon , Eric Dumazet , Eric Dumazet , Soheil Hassas Yeganeh Subject: [PATCH v3 net-next 1/2] tcp: add TCP_ZEROCOPY_RECEIVE support for zerocopy receive Date: Thu, 26 Apr 2018 07:50:55 -0700 Message-Id: <20180426145056.220325-2-edumazet@google.com> X-Mailer: git-send-email 2.17.0.484.g0c8726318c-goog In-Reply-To: <20180426145056.220325-1-edumazet@google.com> References: <20180426145056.220325-1-edumazet@google.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org When adding tcp mmap() implementation, I forgot that socket lock had to be taken before current->mm->mmap_sem. syzbot eventually caught the bug. Since we can not lock the socket in tcp mmap() handler we have to split the operation in two phases. 1) mmap() on a tcp socket simply reserves VMA space, and nothing else. This operation does not involve any TCP locking. 2) getsockopt(fd, IPPROTO_TCP, TCP_ZEROCOPY_RECEIVE, ...) implements the transfert of pages from skbs to one VMA. This operation only uses down_read(¤t->mm->mmap_sem) after holding TCP lock, thus solving the lockdep issue. This new implementation was suggested by Andy Lutomirski with great details. Benefits are : - Better scalability, in case multiple threads reuse VMAS (without mmap()/munmap() calls) since mmap_sem wont be write locked. - Better error recovery. The previous mmap() model had to provide the expected size of the mapping. If for some reason one part could not be mapped (partial MSS), the whole operation had to be aborted. With the tcp_zerocopy_receive struct, kernel can report how many bytes were successfuly mapped, and how many bytes should be read to skip the problematic sequence. - No more memory allocation to hold an array of page pointers. 16 MB mappings needed 32 KB for this array, potentially using vmalloc() :/ - skbs are freed while mmap_sem has been released Following patch makes the change in tcp_mmap tool to demonstrate one possible use of mmap() and setsockopt(... TCP_ZEROCOPY_RECEIVE ...) Note that memcg might require additional changes. Fixes: 93ab6cc69162 ("tcp: implement mmap() for zero copy receive") Signed-off-by: Eric Dumazet Reported-by: syzbot Suggested-by: Andy Lutomirski Cc: linux-mm@kvack.org Cc: Soheil Hassas Yeganeh --- include/uapi/linux/tcp.h | 8 ++ net/ipv4/tcp.c | 192 ++++++++++++++++++++------------------- 2 files changed, 109 insertions(+), 91 deletions(-) diff --git a/include/uapi/linux/tcp.h b/include/uapi/linux/tcp.h index 379b08700a542d49bbce9b4b49b17879d00b69bb..e9e8373b34b9ddc735329341b91f455bf5c0b17c 100644 --- a/include/uapi/linux/tcp.h +++ b/include/uapi/linux/tcp.h @@ -122,6 +122,7 @@ enum { #define TCP_MD5SIG_EXT 32 /* TCP MD5 Signature with extensions */ #define TCP_FASTOPEN_KEY 33 /* Set the key for Fast Open (cookie) */ #define TCP_FASTOPEN_NO_COOKIE 34 /* Enable TFO without a TFO cookie */ +#define TCP_ZEROCOPY_RECEIVE 35 struct tcp_repair_opt { __u32 opt_code; @@ -276,4 +277,11 @@ struct tcp_diag_md5sig { __u8 tcpm_key[TCP_MD5SIG_MAXKEYLEN]; }; +/* setsockopt(fd, IPPROTO_TCP, TCP_ZEROCOPY_RECEIVE, ...) */ + +struct tcp_zerocopy_receive { + __u64 address; /* in: address of mapping */ + __u32 length; /* in/out: number of bytes to map/mapped */ + __u32 recv_skip_hint; /* out: amount of bytes to skip */ +}; #endif /* _UAPI_LINUX_TCP_H */ diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c index dfd090ea54ad47112fc23c61180b5bf8edd2c736..c10c4a41ad39d6f8ae472882b243c2b70c915546 100644 --- a/net/ipv4/tcp.c +++ b/net/ipv4/tcp.c @@ -1726,118 +1726,111 @@ int tcp_set_rcvlowat(struct sock *sk, int val) } EXPORT_SYMBOL(tcp_set_rcvlowat); -/* When user wants to mmap X pages, we first need to perform the mapping - * before freeing any skbs in receive queue, otherwise user would be unable - * to fallback to standard recvmsg(). This happens if some data in the - * requested block is not exactly fitting in a page. - * - * We only support order-0 pages for the moment. - * mmap() on TCP is very strict, there is no point - * trying to accommodate with pathological layouts. - */ +static const struct vm_operations_struct tcp_vm_ops = { +}; + int tcp_mmap(struct file *file, struct socket *sock, struct vm_area_struct *vma) { - unsigned long size = vma->vm_end - vma->vm_start; - unsigned int nr_pages = size >> PAGE_SHIFT; - struct page **pages_array = NULL; - u32 seq, len, offset, nr = 0; - struct sock *sk = sock->sk; - const skb_frag_t *frags; + if (vma->vm_flags & (VM_WRITE | VM_EXEC)) + return -EPERM; + vma->vm_flags &= ~(VM_MAYWRITE | VM_MAYEXEC); + + /* Instruct vm_insert_page() to not down_read(mmap_sem) */ + vma->vm_flags |= VM_MIXEDMAP; + + vma->vm_ops = &tcp_vm_ops; + return 0; +} +EXPORT_SYMBOL(tcp_mmap); + +static int tcp_zerocopy_receive(struct sock *sk, + struct tcp_zerocopy_receive *zc) +{ + unsigned long address = (unsigned long)zc->address; + const skb_frag_t *frags = NULL; + u32 length = 0, seq, offset; + struct vm_area_struct *vma; + struct sk_buff *skb = NULL; struct tcp_sock *tp; - struct sk_buff *skb; int ret; - if (vma->vm_pgoff || !nr_pages) + if (address & (PAGE_SIZE - 1) || address != zc->address) return -EINVAL; - if (vma->vm_flags & VM_WRITE) - return -EPERM; - /* TODO: Maybe the following is not needed if pages are COW */ - vma->vm_flags &= ~VM_MAYWRITE; - - lock_sock(sk); - - ret = -ENOTCONN; if (sk->sk_state == TCP_LISTEN) - goto out; + return -ENOTCONN; sock_rps_record_flow(sk); - if (tcp_inq(sk) < size) { - ret = sock_flag(sk, SOCK_DONE) ? -EIO : -EAGAIN; + down_read(¤t->mm->mmap_sem); + + ret = -EINVAL; + vma = find_vma(current->mm, address); + if (!vma || vma->vm_start > address || vma->vm_ops != &tcp_vm_ops) goto out; - } + zc->length = min_t(unsigned long, zc->length, vma->vm_end - address); + tp = tcp_sk(sk); seq = tp->copied_seq; - /* Abort if urgent data is in the area */ - if (unlikely(tp->urg_data)) { - u32 urg_offset = tp->urg_seq - seq; + zc->length = min_t(u32, zc->length, tcp_inq(sk)); + zc->length &= ~(PAGE_SIZE - 1); - ret = -EINVAL; - if (urg_offset < size) - goto out; - } - ret = -ENOMEM; - pages_array = kvmalloc_array(nr_pages, sizeof(struct page *), - GFP_KERNEL); - if (!pages_array) - goto out; - skb = tcp_recv_skb(sk, seq, &offset); - ret = -EINVAL; -skb_start: - /* We do not support anything not in page frags */ - offset -= skb_headlen(skb); - if ((int)offset < 0) - goto out; - if (skb_has_frag_list(skb)) - goto out; - len = skb->data_len - offset; - frags = skb_shinfo(skb)->frags; - while (offset) { - if (frags->size > offset) - goto out; - offset -= frags->size; - frags++; - } - while (nr < nr_pages) { - if (len) { - if (len < PAGE_SIZE) - goto out; - if (frags->size != PAGE_SIZE || frags->page_offset) - goto out; - pages_array[nr++] = skb_frag_page(frags); - frags++; - len -= PAGE_SIZE; - seq += PAGE_SIZE; - continue; - } - skb = skb->next; - offset = seq - TCP_SKB_CB(skb)->seq; - goto skb_start; - } - /* OK, we have a full set of pages ready to be inserted into vma */ - for (nr = 0; nr < nr_pages; nr++) { - ret = vm_insert_page(vma, vma->vm_start + (nr << PAGE_SHIFT), - pages_array[nr]); - if (ret) - goto out; - } - /* operation is complete, we can 'consume' all skbs */ - tp->copied_seq = seq; - tcp_rcv_space_adjust(sk); - - /* Clean up data we have read: This will do ACK frames. */ - tcp_recv_skb(sk, seq, &offset); - tcp_cleanup_rbuf(sk, size); + zap_page_range(vma, address, zc->length); + zc->recv_skip_hint = 0; ret = 0; + while (length + PAGE_SIZE <= zc->length) { + if (zc->recv_skip_hint < PAGE_SIZE) { + if (skb) { + skb = skb->next; + offset = seq - TCP_SKB_CB(skb)->seq; + } else { + skb = tcp_recv_skb(sk, seq, &offset); + } + + zc->recv_skip_hint = skb->len - offset; + offset -= skb_headlen(skb); + if ((int)offset < 0 || skb_has_frag_list(skb)) + break; + frags = skb_shinfo(skb)->frags; + while (offset) { + if (frags->size > offset) + goto out; + offset -= frags->size; + frags++; + } + } + if (frags->size != PAGE_SIZE || frags->page_offset) + break; + ret = vm_insert_page(vma, address + length, + skb_frag_page(frags)); + if (ret) + break; + length += PAGE_SIZE; + seq += PAGE_SIZE; + zc->recv_skip_hint -= PAGE_SIZE; + frags++; + } out: - release_sock(sk); - kvfree(pages_array); + up_read(¤t->mm->mmap_sem); + if (length) { + tp->copied_seq = seq; + tcp_rcv_space_adjust(sk); + + /* Clean up data we have read: This will do ACK frames. */ + tcp_recv_skb(sk, seq, &offset); + tcp_cleanup_rbuf(sk, length); + ret = 0; + if (length == zc->length) + zc->recv_skip_hint = 0; + } else { + if (!zc->recv_skip_hint && sock_flag(sk, SOCK_DONE)) + ret = -EIO; + } + zc->length = length; return ret; } -EXPORT_SYMBOL(tcp_mmap); static void tcp_update_recv_tstamps(struct sk_buff *skb, struct scm_timestamping *tss) @@ -3472,6 +3465,23 @@ static int do_tcp_getsockopt(struct sock *sk, int level, } return 0; } + case TCP_ZEROCOPY_RECEIVE: { + struct tcp_zerocopy_receive zc; + int err; + + if (get_user(len, optlen)) + return -EFAULT; + if (len != sizeof(zc)) + return -EINVAL; + if (copy_from_user(&zc, optval, len)) + return -EFAULT; + lock_sock(sk); + err = tcp_zerocopy_receive(sk, &zc); + release_sock(sk); + if (!err && copy_to_user(optval, &zc, len)) + err = -EFAULT; + return err; + } default: return -ENOPROTOOPT; } -- 2.17.0.484.g0c8726318c-goog