Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp133604ybi; Tue, 2 Jul 2019 17:41:52 -0700 (PDT) X-Google-Smtp-Source: APXvYqz+YEj6kFi/Ip/JmIkc8q6udph5rdZXIu3QVLBDFwOUoLiT8uOp5shLrOccG19loa+5fl6m X-Received: by 2002:a63:e1e:: with SMTP id d30mr10749410pgl.100.1562114512475; Tue, 02 Jul 2019 17:41:52 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1562114512; cv=none; d=google.com; s=arc-20160816; b=LrS2JZ2zd9kIOsgjBoxrPv66mqhWQe7fz+n+ww54ECSgIix/IgYfOOPCNgZ7bX7Clc 6+ph52CQ1d1rlojda8QPPDKL5MloTJRpYs61gk5ryeuYuw8vWgG4Q/OoLASic0jBqAA+ XKVv6OskWcLXT6c1TPBhZnWz31YVM0bV4Ad1y70qk/9b9n85emwwI+/N/NayVZE/HZD6 QHoIvWjNs+p31GS0Xwy4YmjOADP6WI7VFJdMOvO7ubn6Gjcc/l6HcWk1ktlbtkMoow42 zVX9LYZmcWa8PRqOuw3qKczVOJo4CeJkxzzIFblJg86bmaIkPiNj7NR9TIHw/kPNu8Xl lsAA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:subject:cc:to:from:date; bh=4onPMRUsq3OEkuqd5424Ft025moQXkbN5NG6TO3PwEg=; b=ElJzIgy6XuI7mAM8RR87WJl8pio30g0WM1UOsVSnGV0d8TaXoodoquXUodbgefH6Tv sxUqR/7O125sisH5a3F1yy/Bn3MNAI7TRl8Tgj2LcD3gbz9cmBfa+/ZvaVJFV3kQyceV SPdbSYvUbfqeVpzrXl5QoVaoQ6/QM0xmV/M9EWMUVFuVDArkeJ1i6rtXJoAi0d5kovdd 9CMB+wj5dP7LcLYW0r1ZQ8jChPIVihxQA+cP64n5IpRVIJq1BI6oUDjSIhpoksy721m0 /41DyBihpoWOsDT5Wli9u+CMsX0hEfbUZIO80WFG456AT2DbopMIpCZ83hUIfHKrYpH5 +tIg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q13si206836pjb.13.2019.07.02.17.41.37; Tue, 02 Jul 2019 17:41:52 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727261AbfGCAjx (ORCPT + 99 others); Tue, 2 Jul 2019 20:39:53 -0400 Received: from mx1.redhat.com ([209.132.183.28]:49796 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727040AbfGCAjx (ORCPT ); Tue, 2 Jul 2019 20:39:53 -0400 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.phx2.redhat.com [10.5.11.14]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 844B53084042; Tue, 2 Jul 2019 21:02:49 +0000 (UTC) Received: from carbon (ovpn-200-45.brq.redhat.com [10.40.200.45]) by smtp.corp.redhat.com (Postfix) with ESMTP id D13545D9C6; Tue, 2 Jul 2019 21:02:43 +0000 (UTC) Date: Tue, 2 Jul 2019 23:02:41 +0200 From: Jesper Dangaard Brouer To: Ivan Khoronzhuk Cc: netdev@vger.kernel.org, Ilias Apalodimas , grygorii.strashko@ti.com, jakub.kicinski@netronome.com, daniel@iogearbox.net, john.fastabend@gmail.com, ast@kernel.org, linux-kernel@vger.kernel.org, linux-omap@vger.kernel.org, brouer@redhat.com Subject: Re: [PATCH] net: core: page_pool: add user refcnt and reintroduce page_pool_destroy Message-ID: <20190702230241.3be6d787@carbon> In-Reply-To: <20190702185839.GH4510@khorivan> References: <20190702153902.0e42b0b2@carbon> <156207778364.29180.5111562317930943530.stgit@firesoul> <20190702144426.GD4510@khorivan> <20190702165230.6caa36e3@carbon> <20190702145612.GF4510@khorivan> <20190702171029.76c60538@carbon> <20190702152112.GG4510@khorivan> <20190702202907.15fb30ce@carbon> <20190702185839.GH4510@khorivan> MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 2.79 on 10.5.11.14 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.40]); Tue, 02 Jul 2019 21:03:00 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, 2 Jul 2019 21:58:40 +0300 Ivan Khoronzhuk wrote: > On Tue, Jul 02, 2019 at 08:29:07PM +0200, Jesper Dangaard Brouer wrote: > >On Tue, 2 Jul 2019 18:21:13 +0300 > >Ivan Khoronzhuk wrote: > > > >> On Tue, Jul 02, 2019 at 05:10:29PM +0200, Jesper Dangaard Brouer wrote: > >> >On Tue, 2 Jul 2019 17:56:13 +0300 > >> >Ivan Khoronzhuk wrote: > >> > > >> >> On Tue, Jul 02, 2019 at 04:52:30PM +0200, Jesper Dangaard Brouer wrote: > >> >> >On Tue, 2 Jul 2019 17:44:27 +0300 > >> >> >Ivan Khoronzhuk wrote: > >> >> > > >> >> >> On Tue, Jul 02, 2019 at 04:31:39PM +0200, Jesper Dangaard Brouer wrote: > >> >> >> >From: Ivan Khoronzhuk > >> >> >> > > >> >> >> >Jesper recently removed page_pool_destroy() (from driver invocation) and > >> >> >> >moved shutdown and free of page_pool into xdp_rxq_info_unreg(), in-order to > >> >> >> >handle in-flight packets/pages. This created an asymmetry in drivers > >> >> >> >create/destroy pairs. > >> >> >> > > >> >> >> >This patch add page_pool user refcnt and reintroduce page_pool_destroy. > >> >> >> >This serves two purposes, (1) simplify drivers error handling as driver now > >> >> >> >drivers always calls page_pool_destroy() and don't need to track if > >> >> >> >xdp_rxq_info_reg_mem_model() was unsuccessful. (2) allow special cases > >> >> >> >where a single RX-queue (with a single page_pool) provides packets for two > >> >> >> >net_device'es, and thus needs to register the same page_pool twice with two > >> >> >> >xdp_rxq_info structures. > >> >> >> > >> >> >> As I tend to use xdp level patch there is no more reason to mention (2) case > >> >> >> here. XDP patch serves it better and can prevent not only obj deletion but also > >> >> >> pool flush, so, this one patch I could better leave only for (1) case. > >> >> > > >> >> >I don't understand what you are saying. > >> >> > > >> >> >Do you approve this patch, or do you reject this patch? > >> >> > > >> >> It's not reject, it's proposition to use both, XDP and page pool patches, > >> >> each having its goal. > >> > > >> >Just to be clear, if you want this patch to get accepted you have to > >> >reply with your Signed-off-by (as I wrote). > >> > > >> >Maybe we should discuss it in another thread, about why you want two > >> >solutions to the same problem. > >> > >> If it solves same problem I propose to reject this one and use this: > >> https://lkml.org/lkml/2019/7/2/651 > > > >No, I propose using this one, and rejecting the other one. > > There is at least several arguments against this one (related (2) purpose) > > It allows: > - avoid changes to page_pool/mlx5/netsec > - save not only allocator obj but allocator "page/buffer flush" > - buffer flush can be present not only in page_pool but for other allocators > that can behave differently and not so simple solution. > - to not limit cpsw/(potentially others) to use "page_pool" allocator only > .... > > This patch better leave also, as it simplifies error path for page_pool and > have more error prone usage comparing with existent one. > > Please, don't limit cpsw and potentially other drivers to use only > page_pool it can be zca or etc... I don't won't to modify each allocator. > I propose to add both as by fact they solve different problems with common > solution. I'm trying to limit the scope of your changes, for your special case, because I'm afraid this more common solution is going to limit our options, painting ourselves into a corner. E.g. for correct lifetime handling, I think we actually need to do a dev_hold() on the net_device. (Changes in f71fec47c2 might not be enough, but I first need to dig into the details and ask Hellwig about some details). Adding that after your patch is more complicated (if even doable). E.g. doing dev_hold() on the net_device, can also turn into a performance advantage, when/if page_pool is extended to also "travel" into SKBs. (Allowing to elide such dev_hold() calls in netstack). I also worry about the possible performance impact these changes will have down the road. (For the RX/alloc side it should be clear by now that we gain a lot of performance with the single RX-queue binding and napi protection). On the return/free side performance *need* to be improved (it doesn't scale). I'm basically looking at different ways to bulk return pages into the ptr_ring, which requires changes in page_pool and likely in xdp_allocator structure. Which your changes are complicating. This special use-case, seems confined to your driver. And Ilias told me that XDP is not really a performance benefit for this driver as the HW PPS-limit is hit before the XDP and netstack limit. I ask, does it make sense to add XDP to this driver, if it complicates the code for everybody else? -- Best regards, Jesper Dangaard Brouer MSc.CS, Principal Kernel Engineer at Red Hat LinkedIn: http://www.linkedin.com/in/brouer