Received: by 10.213.65.68 with SMTP id h4csp499195imn; Fri, 16 Mar 2018 09:35:48 -0700 (PDT) X-Google-Smtp-Source: AG47ELtw+C1i/TrgraHoBKtyGHP4kwmzW3VoXOZ441ipIDQ3GP2vYrCtHd3EzeyRcD5TKhglYvck X-Received: by 10.101.76.13 with SMTP id u13mr1881140pgq.287.1521218147964; Fri, 16 Mar 2018 09:35:47 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1521218147; cv=none; d=google.com; s=arc-20160816; b=wSWfcKM/1O2eyF2QdgIhkog0VFki9a3231ZIOSrXQoUvPYwAFd9m//Ku3l7OO3ka1k 56TUbYXF3kmCdEUUj1vAQspD1We3h8Msy0tfIf+DgTwEUDw0yCQQrnG8gUjwvvB8Qu16 U0r5z9I2dLidyQVxx8e+xfhjv7WPIYbN6Kmy1NxRmJb+YeBpFs7+9IP5zoDa6rz0SSU4 4dvC6+4Pw+ZivbEfINC3olgNGw9PmXd1JrS88fYO2rr5vFJMr6Ljo1QOZn1DhESqh1OF nWcUBbqLZghfO1AuNOYiRoYYVUzc25lp6VtCbUD1nyjHJ8m2t7+2Dt+72MaGHWv6rRX6 sCMg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:mime-version:user-agent:references :in-reply-to:message-id:date:subject:cc:to:from :arc-authentication-results; bh=5PEtK9Uj8hiBeM64O6TtgfPSVdQdNAnRVNoLdQb1F1Y=; b=GUWPzTIK85cDcPOMH3G6v6s6cB9O8/l0E78A3hKqXbb1T9AU+JUO1IlIwm7EJONxVo 1IJI7e/zr8cRc4yC2ecHMCyPDQqGWZL1uEU6oJrxJSgQaP1EGGMaRbHCgd6XWAdiRV7O 7qrUD3sz8Ve1+fvQR2GXIC/X3yiHr5/kOjEeE08SWEUsQZpcBigcmnx+QCdXQap3VM16 nZ6Ulef3UHQ6gocuGyMCKZTxwxVHEs0+O92VoOeT6lWIwtJuVs/2V3T4Gp6FDsieDIKC 79nwTVqYGOLtAnz+Ne6xZaXaw4WDFwB9qY1QxJ5GVTpl78pb2w4d2Wvy7fMdk/ZDyXlX Hodw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id r63-v6si6457146plb.356.2018.03.16.09.35.33; Fri, 16 Mar 2018 09:35:47 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752990AbeCPQeI (ORCPT + 99 others); Fri, 16 Mar 2018 12:34:08 -0400 Received: from mail.linuxfoundation.org ([140.211.169.12]:42048 "EHLO mail.linuxfoundation.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S934475AbeCPPiD (ORCPT ); Fri, 16 Mar 2018 11:38:03 -0400 Received: from localhost (LFbn-1-12247-202.w90-92.abo.wanadoo.fr [90.92.61.202]) by mail.linuxfoundation.org (Postfix) with ESMTPSA id 71D5D125F; Fri, 16 Mar 2018 15:37:57 +0000 (UTC) From: Greg Kroah-Hartman To: linux-kernel@vger.kernel.org Cc: Greg Kroah-Hartman , stable@vger.kernel.org, John Fastabend , "David S. Miller" , Sasha Levin Subject: [PATCH 4.14 090/109] net: sched: drop qdisc_reset from dev_graft_qdisc Date: Fri, 16 Mar 2018 16:23:59 +0100 Message-Id: <20180316152334.821254067@linuxfoundation.org> X-Mailer: git-send-email 2.16.2 In-Reply-To: <20180316152329.844663293@linuxfoundation.org> References: <20180316152329.844663293@linuxfoundation.org> User-Agent: quilt/0.65 X-stable: review MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org 4.14-stable review patch. If anyone has any objections, please let me know. ------------------ From: John Fastabend [ Upstream commit 7bbde83b1860c28a1cc35516352c4e7e5172c29a ] In qdisc_graft_qdisc a "new" qdisc is attached and the 'qdisc_destroy' operation is called on the old qdisc. The destroy operation will wait a rcu grace period and call qdisc_rcu_free(). At which point gso_cpu_skb is free'd along with all stats so no need to zero stats and gso_cpu_skb from the graft operation itself. Further after dropping the qdisc locks we can not continue to call qdisc_reset before waiting an rcu grace period so that the qdisc is detached from all cpus. By removing the qdisc_reset() here we get the correct property of waiting an rcu grace period and letting the qdisc_destroy operation clean up the qdisc correctly. Note, a refcnt greater than 1 would cause the destroy operation to be aborted however if this ever happened the reference to the qdisc would be lost and we would have a memory leak. Signed-off-by: John Fastabend Signed-off-by: David S. Miller Signed-off-by: Sasha Levin Signed-off-by: Greg Kroah-Hartman --- net/sched/sch_generic.c | 28 +++++++++++++++++++--------- 1 file changed, 19 insertions(+), 9 deletions(-) --- a/net/sched/sch_generic.c +++ b/net/sched/sch_generic.c @@ -743,10 +743,6 @@ struct Qdisc *dev_graft_qdisc(struct net root_lock = qdisc_lock(oqdisc); spin_lock_bh(root_lock); - /* Prune old scheduler */ - if (oqdisc && refcount_read(&oqdisc->refcnt) <= 1) - qdisc_reset(oqdisc); - /* ... and graft new one */ if (qdisc == NULL) qdisc = &noop_qdisc; @@ -897,6 +893,16 @@ static bool some_qdisc_is_busy(struct ne return false; } +static void dev_qdisc_reset(struct net_device *dev, + struct netdev_queue *dev_queue, + void *none) +{ + struct Qdisc *qdisc = dev_queue->qdisc_sleeping; + + if (qdisc) + qdisc_reset(qdisc); +} + /** * dev_deactivate_many - deactivate transmissions on several devices * @head: list of devices to deactivate @@ -907,7 +913,6 @@ static bool some_qdisc_is_busy(struct ne void dev_deactivate_many(struct list_head *head) { struct net_device *dev; - bool sync_needed = false; list_for_each_entry(dev, head, close_list) { netdev_for_each_tx_queue(dev, dev_deactivate_queue, @@ -917,20 +922,25 @@ void dev_deactivate_many(struct list_hea &noop_qdisc); dev_watchdog_down(dev); - sync_needed |= !dev->dismantle; } /* Wait for outstanding qdisc-less dev_queue_xmit calls. * This is avoided if all devices are in dismantle phase : * Caller will call synchronize_net() for us */ - if (sync_needed) - synchronize_net(); + synchronize_net(); /* Wait for outstanding qdisc_run calls. */ - list_for_each_entry(dev, head, close_list) + list_for_each_entry(dev, head, close_list) { while (some_qdisc_is_busy(dev)) yield(); + /* The new qdisc is assigned at this point so we can safely + * unwind stale skb lists and qdisc statistics + */ + netdev_for_each_tx_queue(dev, dev_qdisc_reset, NULL); + if (dev_ingress_queue(dev)) + dev_qdisc_reset(dev, dev_ingress_queue(dev), NULL); + } } void dev_deactivate(struct net_device *dev)