Received: by 2002:a05:6358:1087:b0:cb:c9d3:cd90 with SMTP id j7csp1298841rwi; Thu, 13 Oct 2022 11:40:12 -0700 (PDT) X-Google-Smtp-Source: AMsMyM66iXOqSstlBBoxYt1hXxEs9tidtivHe6l7TJ6dMC2cfgScQSYxa+7MCi3LzNXX2J1TqhEl X-Received: by 2002:a17:907:72d1:b0:78e:16cb:31af with SMTP id du17-20020a17090772d100b0078e16cb31afmr811781ejc.231.1665686412485; Thu, 13 Oct 2022 11:40:12 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1665686412; cv=none; d=google.com; s=arc-20160816; b=p7IAl1oSTq2NXq0YXZUN39i+qjmHjAlOBcIrIIhB0G1JOeBCHNdRDue2tdizD0EQtm kKMEPOnHcZ2qokx0ryAjtR/JxdmVIH0WNTrBnG4cqx87D7NFR4L3j9EOV+hxwxnN9oON y89OrJwHlieGdtkwTy7KhtR/zDzKSDMkXfTOAZ69PKVPgnNBDkBcINNPyiFA6vFv6UgE E34P3CjjNq0ICiHpFkRI4hkkccTtboAAUKg5ecjYYi/R4VR4i9w4XdQE8zrg3G3shDvR vKUojO81zR4A0f27D2tkVlonQioiwimVATHgS0UC5/AoMnDW0ot2FWcxDyjsvmzHQyf+ BZlw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :user-agent:references:in-reply-to:message-id:date:subject:cc:to :from:dkim-signature; bh=6xUYP0N+KL+/7fLdmtilqmBfURf1xBMC9zAqWemjNHY=; b=K0UiQ9W9s+sWIRtxnXWqWGcmKQGqL7Q4fRhJtV9aUIRczYoheSVezdBGHDG/y+LBwy aWf0iEUriEU2VKYoDpohyabXMU1cIT9PysCQRMulRMg6Qq+2TlPArbie2RhwN7N/OjPd WGB0HxCvebRCXBLn1bVxNLh7oYABwEvOy4EqGVac/SVvDiq6OQl8O7QDdXO0hS+18Ker e0fu4aUbxK4g9Pg41Gp/tnWWaEBwGpXq1kukXSXtwNwPzdQDYmL+U40XSLFm0+3hxSCG ZbPhTpZMCrJz6KYIKdgjoZSMCk1Rm+J+lsWv6hcwjwxmQpuhsrPoc5KaUFMyVVWN9mPt Kutw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linuxfoundation.org header.s=korg header.b=W2oc5Owk; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linuxfoundation.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id pk21-20020a170906d7b500b0078db6f488c2si367193ejb.56.2022.10.13.11.39.45; Thu, 13 Oct 2022 11:40:12 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@linuxfoundation.org header.s=korg header.b=W2oc5Owk; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linuxfoundation.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231725AbiJMSPa (ORCPT + 99 others); Thu, 13 Oct 2022 14:15:30 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38826 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231622AbiJMSMr (ORCPT ); Thu, 13 Oct 2022 14:12:47 -0400 Received: from dfw.source.kernel.org (dfw.source.kernel.org [IPv6:2604:1380:4641:c500::1]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 66CCA2DCA; Thu, 13 Oct 2022 11:09:23 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id A269F61A3F; Thu, 13 Oct 2022 18:01:27 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id DF02BC433B5; Thu, 13 Oct 2022 18:01:26 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=linuxfoundation.org; s=korg; t=1665684087; bh=Qgu883AQmFoIRpgdwPNsgCs+MfY/HWZXNTNzgBkZpX8=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=W2oc5OwkxgIXDer7JS6JyRj3kaUPrQ4LbqvAckThNUDeHaZsj4DKWLQieyuTtEC88 2ebsCHVhHnfEDa2N3wMp/hKpjBjVLBucUViGExuSOUcYrdNLzYKS8zRArcQiggoYMd NDM6eFFx9HWOhd6NQai8t0djNXChDB6jiaenbhRg= From: Greg Kroah-Hartman To: linux-kernel@vger.kernel.org Cc: Greg Kroah-Hartman , stable@vger.kernel.org, Sherry Yang , Paul Webb , Phillip Goerl , Jack Vogel , Nicky Veitch , Colm Harrington , Ramanan Govindarajan , Sebastian Andrzej Siewior , Dominik Brodowski , Tejun Heo , Sultan Alsawaf , "Jason A. Donenfeld" Subject: [PATCH 6.0 20/34] random: use expired timer rather than wq for mixing fast pool Date: Thu, 13 Oct 2022 19:52:58 +0200 Message-Id: <20221013175147.042953145@linuxfoundation.org> X-Mailer: git-send-email 2.38.0 In-Reply-To: <20221013175146.507746257@linuxfoundation.org> References: <20221013175146.507746257@linuxfoundation.org> User-Agent: quilt/0.67 MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-7.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Jason A. Donenfeld commit 748bc4dd9e663f23448d8ad7e58c011a67ea1eca upstream. Previously, the fast pool was dumped into the main pool periodically in the fast pool's hard IRQ handler. This worked fine and there weren't problems with it, until RT came around. Since RT converts spinlocks into sleeping locks, problems cropped up. Rather than switching to raw spinlocks, the RT developers preferred we make the transformation from originally doing: do_some_stuff() spin_lock() do_some_other_stuff() spin_unlock() to doing: do_some_stuff() queue_work_on(some_other_stuff_worker) This is an ordinary pattern done all over the kernel. However, Sherry noticed a 10% performance regression in qperf TCP over a 40gbps InfiniBand card. Quoting her message: > MT27500 Family [ConnectX-3] cards: > Infiniband device 'mlx4_0' port 1 status: > default gid: fe80:0000:0000:0000:0010:e000:0178:9eb1 > base lid: 0x6 > sm lid: 0x1 > state: 4: ACTIVE > phys state: 5: LinkUp > rate: 40 Gb/sec (4X QDR) > link_layer: InfiniBand > > Cards are configured with IP addresses on private subnet for IPoIB > performance testing. > Regression identified in this bug is in TCP latency in this stack as reported > by qperf tcp_lat metric: > > We have one system listen as a qperf server: > [root@yourQperfServer ~]# qperf > > Have the other system connect to qperf server as a client (in this > case, it’s X7 server with Mellanox card): > [root@yourQperfClient ~]# numactl -m0 -N0 qperf 20.20.20.101 -v -uu -ub --time 60 --wait_server 20 -oo msg_size:4K:1024K:*2 tcp_lat Rather than incur the scheduling latency from queue_work_on, we can instead switch to running on the next timer tick, on the same core. This also batches things a bit more -- once per jiffy -- which is okay now that mix_interrupt_randomness() can credit multiple bits at once. Reported-by: Sherry Yang Tested-by: Paul Webb Cc: Sherry Yang Cc: Phillip Goerl Cc: Jack Vogel Cc: Nicky Veitch Cc: Colm Harrington Cc: Ramanan Govindarajan Cc: Sebastian Andrzej Siewior Cc: Dominik Brodowski Cc: Tejun Heo Cc: Sultan Alsawaf Cc: stable@vger.kernel.org Fixes: 58340f8e952b ("random: defer fast pool mixing to worker") Signed-off-by: Jason A. Donenfeld Signed-off-by: Greg Kroah-Hartman --- drivers/char/random.c | 18 +++++++++++------- 1 file changed, 11 insertions(+), 7 deletions(-) --- a/drivers/char/random.c +++ b/drivers/char/random.c @@ -923,17 +923,20 @@ struct fast_pool { unsigned long pool[4]; unsigned long last; unsigned int count; - struct work_struct mix; + struct timer_list mix; }; +static void mix_interrupt_randomness(struct timer_list *work); + static DEFINE_PER_CPU(struct fast_pool, irq_randomness) = { #ifdef CONFIG_64BIT #define FASTMIX_PERM SIPHASH_PERMUTATION - .pool = { SIPHASH_CONST_0, SIPHASH_CONST_1, SIPHASH_CONST_2, SIPHASH_CONST_3 } + .pool = { SIPHASH_CONST_0, SIPHASH_CONST_1, SIPHASH_CONST_2, SIPHASH_CONST_3 }, #else #define FASTMIX_PERM HSIPHASH_PERMUTATION - .pool = { HSIPHASH_CONST_0, HSIPHASH_CONST_1, HSIPHASH_CONST_2, HSIPHASH_CONST_3 } + .pool = { HSIPHASH_CONST_0, HSIPHASH_CONST_1, HSIPHASH_CONST_2, HSIPHASH_CONST_3 }, #endif + .mix = __TIMER_INITIALIZER(mix_interrupt_randomness, 0) }; /* @@ -975,7 +978,7 @@ int __cold random_online_cpu(unsigned in } #endif -static void mix_interrupt_randomness(struct work_struct *work) +static void mix_interrupt_randomness(struct timer_list *work) { struct fast_pool *fast_pool = container_of(work, struct fast_pool, mix); /* @@ -1029,10 +1032,11 @@ void add_interrupt_randomness(int irq) if (new_count < 1024 && !time_is_before_jiffies(fast_pool->last + HZ)) return; - if (unlikely(!fast_pool->mix.func)) - INIT_WORK(&fast_pool->mix, mix_interrupt_randomness); fast_pool->count |= MIX_INFLIGHT; - queue_work_on(raw_smp_processor_id(), system_highpri_wq, &fast_pool->mix); + if (!timer_pending(&fast_pool->mix)) { + fast_pool->mix.expires = jiffies; + add_timer_on(&fast_pool->mix, raw_smp_processor_id()); + } } EXPORT_SYMBOL_GPL(add_interrupt_randomness);