From: Eric Dumazet Subject: Re: [PATCH] net/sock: move memory_allocated over to percpu_counter variables Date: Fri, 7 Sep 2018 00:03:13 -0700 Message-ID: References: <20180906192034.8467-1-olof@lixom.net> <20180907033257.2nlgiqm2t4jiwhzc@gondor.apana.org.au> Mime-Version: 1.0 Content-Type: text/plain; charset="UTF-8" Cc: Herbert Xu , David Miller , Neil Horman , Marcelo Ricardo Leitner , Vladislav Yasevich , Alexey Kuznetsov , Hideaki YOSHIFUJI , linux-crypto@vger.kernel.org, LKML , linux-sctp@vger.kernel.org, netdev , linux-decnet-user@lists.sourceforge.net, kernel-team , Yuchung Cheng , Neal Cardwell To: Olof Johansson Return-path: In-Reply-To: Sender: linux-kernel-owner@vger.kernel.org List-Id: linux-crypto.vger.kernel.org On Thu, Sep 6, 2018 at 11:20 PM Olof Johansson wrote: > > Hi, > > On Thu, Sep 6, 2018 at 8:32 PM, Herbert Xu wrote: > > On Thu, Sep 06, 2018 at 12:33:58PM -0700, Eric Dumazet wrote: > >> On Thu, Sep 6, 2018 at 12:21 PM Olof Johansson wrote: > >> > > >> > Today these are all global shared variables per protocol, and in > >> > particular tcp_memory_allocated can get hot on a system with > >> > large number of CPUs and a substantial number of connections. > >> > > >> > Moving it over to a per-cpu variable makes it significantly cheaper, > >> > and the added overhead when summing up the percpu copies is still smaller > >> > than the cost of having a hot cacheline bouncing around. > >> > >> I am curious. We never noticed contention on this variable, at least for TCP. > > > > Yes these variables are heavily amortised so I'm surprised that > > they would cause much contention. > > > >> Please share some numbers with us. > > > > Indeed. > > Certainly, just had to collect them again. > > This is on a dual xeon box, with ~150-200k TCP connections. I see > about .7% CPU spent in __sk_mem_{reduce,raise}_allocated in the > inlined atomic ops, most of those in reduce. > > Call path for reduce is practically all from tcp_write_timer on softirq: > > __sk_mem_reduce_allocated > tcp_write_timer > call_timer_fn > run_timer_softirq > __do_softirq > irq_exit > smp_apic_timer_interrupt > apic_timer_interrupt > cpuidle_enter_state > > With this patch, I see about .18+.11+.07 = .36% in percpu-related > functions called from the same __sk_mem functions. > > Now, that's a halving of cycles samples on that specific setup. The > real difference though, is on another platform where atomics are more > expensive. There, this makes a significant difference. Unfortunately, > I can't share specifics but I think this change stands on its own on > the dual xeon setup as well, maybe with slightly less strong wording > on just how hot the variable/line happens to be. Problem is : we have platforms with more than 100 cpus, and sk_memory_allocated() cost will be too expensive, especially if the host is under memory pressure, since all cpus will touch their private counter. per cpu variables do not really scale, they were ok 10 years ago when no more than 16 cpus were the norm. I would prefer change TCP to not aggressively call __sk_mem_reduce_allocated() from tcp_write_timer() Ideally only tcp_retransmit_timer() should attempt to reduce forward allocations, after recurring timeout. Note that after 20c64d5cd5a2bdcdc8982a06cb05e5e1bd851a3d ("net: avoid sk_forward_alloc overflows") we have better control over sockets having huge forward allocations. Something like : diff --git a/net/ipv4/tcp_timer.c b/net/ipv4/tcp_timer.c index 7fdf222a0bdfe9775970082f6b5dcdcc82b2ae1a..7e2e17cde9b6a9be835edfac26b64f4ce9411538 100644 --- a/net/ipv4/tcp_timer.c +++ b/net/ipv4/tcp_timer.c @@ -505,6 +505,8 @@ void tcp_retransmit_timer(struct sock *sk) mib_idx = LINUX_MIB_TCPTIMEOUTS; } __NET_INC_STATS(sock_net(sk), mib_idx); + } else { + sk_mem_reclaim(sk); } tcp_enter_loss(sk); @@ -576,11 +578,11 @@ void tcp_write_timer_handler(struct sock *sk) if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) || !icsk->icsk_pending) - goto out; + return; if (time_after(icsk->icsk_timeout, jiffies)) { sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); - goto out; + return; } tcp_mstamp_refresh(tcp_sk(sk)); @@ -602,9 +604,6 @@ void tcp_write_timer_handler(struct sock *sk) tcp_probe_timer(sk); break; } - -out: - sk_mem_reclaim(sk); }