Received: by 2002:a05:7412:b995:b0:f9:9502:5bb8 with SMTP id it21csp889904rdb; Fri, 22 Dec 2023 07:59:32 -0800 (PST) X-Google-Smtp-Source: AGHT+IHnwxB9RLRYVC1wwmCxUc5AelTWMh8qZ0mMMA6LJZ9PTv9GBLZDYEs+pDuthys9OA3f5Irm X-Received: by 2002:aa7:d497:0:b0:54b:27e7:f965 with SMTP id b23-20020aa7d497000000b0054b27e7f965mr812187edr.19.1703260772309; Fri, 22 Dec 2023 07:59:32 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1703260772; cv=none; d=google.com; s=arc-20160816; b=gL6iMyAVOlSqQaaDO/sv7Joab8jiXMd2GNZFm8N/JOUlkXvmb+Lm8pW0wJS5mOKAhq IYB/c8Yzr/4lCoVyo8PXGphVEWDQ0ISOIK0II1jIhD/0RIZA3f0LOrXgTsNPLhnotx8q Fk3KdL/vLPmKa1gHbaREef3EA74jZtp386RyecuWO1owtC1f28tbma7l1LhJl1QJc/1F 1Fo+o3tqT466bnKAiImsp9B0cpxACkTpRRrI+KKTCxqfU14VZOIGrjOrK+pG2DzmJRUz VfNGCwSW5Z4PmBjEGE7ptcSHUt10pdMmEE1r1PTVvGPdmNkTUcfhCmuGWtWOGyU/fICU V8zw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:list-unsubscribe:list-subscribe:list-id:precedence :dkim-signature; bh=OAR7QlHd39RQ33GYs0BoNsJCPx6aNi6FZcLgvPLmh3E=; fh=jJBbNPnQZtZvstbxiz5RTvIPxy/uxViwSQpNAR2vU2o=; b=yFYqUk4Y5pygvlGgh7GfPU4AQ/6aeQ6tNmlEjoxzI24xOiDcBWtQAKgnG1pI2rAXH5 YMHrGdibvTN/Tx80YVkl7/BVBk0FLitxvFqGztW0Vv79koXBOKmPIQIFQlIkX1mqc8+z At8mbWLvh1omc2TPNW4rRKo3vlSquPiyy3wWqrSIWN0gzJC5wJmrQ623HTOMB3lIo+eW CW/iFJ9qsygbDWApGZ8fQWpPDEgPiLWk5SkVPvSRfM7uDUSAs7lGsiJsOq4Eti2cl2vz 21r+7zXoetl1Ila81LxK68QaKrFT6VGARsHlhWl9v8l+Gl/S4Mxapi/pLlEymiAyCtGN gvUg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=CBkoEhMF; spf=pass (google.com: domain of linux-kernel+bounces-9859-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:4601:e00::3 as permitted sender) smtp.mailfrom="linux-kernel+bounces-9859-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from am.mirrors.kernel.org (am.mirrors.kernel.org. [2604:1380:4601:e00::3]) by mx.google.com with ESMTPS id u21-20020a50d515000000b005534fb7d47asi1854530edi.538.2023.12.22.07.59.32 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 22 Dec 2023 07:59:32 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-9859-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:4601:e00::3 as permitted sender) client-ip=2604:1380:4601:e00::3; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=CBkoEhMF; spf=pass (google.com: domain of linux-kernel+bounces-9859-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:4601:e00::3 as permitted sender) smtp.mailfrom="linux-kernel+bounces-9859-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by am.mirrors.kernel.org (Postfix) with ESMTPS id DFF981F226A0 for ; Fri, 22 Dec 2023 15:59:31 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 34983241F0; Fri, 22 Dec 2023 15:59:25 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=linaro.org header.i=@linaro.org header.b="CBkoEhMF" X-Original-To: linux-kernel@vger.kernel.org Received: from mail-pj1-f47.google.com (mail-pj1-f47.google.com [209.85.216.47]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id CB990241E4 for ; Fri, 22 Dec 2023 15:59:22 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linaro.org Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linaro.org Received: by mail-pj1-f47.google.com with SMTP id 98e67ed59e1d1-28bfdf3cd12so941582a91.3 for ; Fri, 22 Dec 2023 07:59:22 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; t=1703260762; x=1703865562; darn=vger.kernel.org; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc:subject:date:message-id:reply-to; bh=OAR7QlHd39RQ33GYs0BoNsJCPx6aNi6FZcLgvPLmh3E=; b=CBkoEhMFP3jw3I2fasjhvaHbdgGV1lSlARnX5gXuYJdri2+IG0BoRSxbzxO3Ey09/i kyN4TDMwVJt9XZU+KIOeLC5gxGFKsQhuz6LWdXXGG1DWf24fpyIcCcAhP8IL1/XGSruO 6GVWKG5E1O36YgLdRd6bEU4zyVoInIV0cMLJl8tTI8YnJddrMiolsV4bcG7rVUQQyU+K hBO+DG4L4JDouDay0X/vPPfeJZhiIWjDmf4hF7IUt4TjjbjR43cKaL/rAXL7oeEyTki8 EWFK4USr51VXGpS6IF+s2DI28UbnWz3QuMXo/wzcv7yOzPZJDZ4PuCJGyucfOjW1DuiR lB2Q== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1703260762; x=1703865562; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=OAR7QlHd39RQ33GYs0BoNsJCPx6aNi6FZcLgvPLmh3E=; b=kEM1L95ouMy8KoviscDbqCfr6TOJKmevRkVwIw/wf7y0v9hLbN5Ggip7KhEMSQwt0K QHffNXVFSs2XPY7uRNwQvNZtGib6+PMmIKipnWpdUHnqtbe7tIo2+wae2DskZxiUBliU zpK6bo/mA1q+ysZd5Fv2sn3iTlyLRG8sfJWZtRhs3gy6ke0yxcIPQxzGwNjSI63RuXcU h2YWH9OXier4ZmkjW2tP2TyyDrZKilQhI1Dgt8lWGZARAXaafF/q5RpsRwbxS0e5HmgX cZg4H1vwAlpEO0lxxYzUTEwW6ZjNKmrDbdGwAYis+sdKWsk+KSlmp0gByB49KknCcpL0 zwyA== X-Gm-Message-State: AOJu0Yx+AB90jeF4Cx2wVMJBi0lQ1RxV4Z6PXyP3nXhZUpdpqnCpcQgz qa7Z+nHcnXJ7G0M0AIQ49N4ZglkTIlBK9nA+gWrDewitKCbXUA== X-Received: by 2002:a17:90a:784e:b0:28a:c616:ff1c with SMTP id y14-20020a17090a784e00b0028ac616ff1cmr814372pjl.97.1703260762031; Fri, 22 Dec 2023 07:59:22 -0800 (PST) Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 References: <20231003215159.GJ1539@noisy.programming.kicks-ass.net> <20231004120544.GA6307@noisy.programming.kicks-ass.net> <20231004174801.GE19999@noisy.programming.kicks-ass.net> <20231009102949.GC14330@noisy.programming.kicks-ass.net> <98b3df1-79b7-836f-e334-afbdd594b55@inria.fr> <93112fbe-30be-eab8-427c-5d4670a0f94e@inria.fr> <9dc451b5-9dd8-89f2-1c9c-7c358faeaad@inria.fr> In-Reply-To: <9dc451b5-9dd8-89f2-1c9c-7c358faeaad@inria.fr> From: Vincent Guittot Date: Fri, 22 Dec 2023 16:59:10 +0100 Message-ID: Subject: Re: EEVDF and NUMA balancing To: Julia Lawall Cc: Peter Zijlstra , Ingo Molnar , Dietmar Eggemann , Mel Gorman , linux-kernel@vger.kernel.org Content-Type: text/plain; charset="UTF-8" On Fri, 22 Dec 2023 at 16:00, Julia Lawall wrote: > > > > On Fri, 22 Dec 2023, Vincent Guittot wrote: > > > On Thu, 21 Dec 2023 at 19:20, Julia Lawall wrote: > > > > > > > > > > > > On Wed, 20 Dec 2023, Vincent Guittot wrote: > > > > > > > On Tue, 19 Dec 2023 at 18:51, Julia Lawall wrote: > > > > > > > > > > > > One CPU has 2 threads, and the others have one. The one with two threads > > > > > > > is returned as the busiest one. But nothing happens, because both of them > > > > > > > prefer the socket that they are on. > > > > > > > > > > > > This explains way load_balance uses migrate_util and not migrate_task. > > > > > > One CPU with 2 threads can be overloaded > > > > > > > > > > > > ok, so it seems that your 1st problem is that you have 2 threads on > > > > > > the same CPU whereas you should have an idle core in this numa node. > > > > > > All cores are sharing the same LLC, aren't they ? > > > > > > > > > > Sorry, not following this. > > > > > > > > > > Socket 1 has N-1 threads, and thus an idle CPU. > > > > > Socket 2 has N+1 threads, and thus one CPU with two threads. > > > > > > > > > > Socket 1 tries to steal from that one CPU with two threads, but that > > > > > fails, because both threads prefer being on Socket 2. > > > > > > > > > > Since most (or all?) of the threads on Socket 2 perfer being on Socket 2. > > > > > the only hope for Socket 1 to fill in its idle core is active balancing. > > > > > But active balancing is not triggered because of migrate_util and because > > > > > CPU_NEWLY_IDLE prevents the failure counter from ebing increased. > > > > > > > > CPU_NEWLY_IDLE load_balance doesn't aims to do active load balance so > > > > you should focus on the CPU_NEWLY_IDLE load_balance > > > > > > I'm still perplexed why a core that has been idle for 1 second or more is > > > considered to be newly idle. > > > > CPU_NEWLY_IDLE load balance is called when the scheduler was > > scheduling something that just migrated or went back to sleep and > > doesn't have anything to schedule so it tries to pull a task from > > somewhere else. > > > > But you should still have some CPU_IDLE load balance according to your > > description where one CPU of the socket remains idle and those will > > increase the nr_balance_failed > > This happens. But not often. > > > I'm surprised that you have mainly CPU_NEWLY_IDLE. Do you know the reason ? > > No. They come from do_idle calling the scheduler. I will look into why > this happens so often. Hmm, the CPU was idle and received a need resched which triggered the scheduler but there was nothing to schedule so it goes back to idle after running a newly_idle _load_balance. > > > > > > > > > > > > > > > > > > > > The part that I am currently missing to understand is that when I convert > > > > > CPU_NEWLY_IDLE to CPU_IDLE, it typically picks a CPU with only one thread > > > > > as busiest. I have the impression that the fbq_type intervenes to cause > > > > > > > > find_busiest_queue skips rqs which only have threads preferring being > > > > in there. So it selects another rq with a thread that doesn't prefer > > > > its current node. > > > > > > > > do you know what is the value of env->fbq_type ? > > > > > > I have seen one trace in which it is all. There are 33 tasks on one > > > socket, and they are all considered to have a preference for that socket. > > > > With env->fbq_type == all, load_balance and find_busiest_queue should > > be able to select the actual busiest queue with 2 threads. > > That's what it does. But nothing can be stolen because there is no active > balancing. My patch below should enable to pull a task from the 1st idle load balance that fails > > > > > But then I imagine that can_migrate/ migrate_degrades_locality > > prevents to detach the task > > Exactly. > > julia > > > > > > > But I have another trace in which it is regular. There are 33 tasks on > > > the socket, but only 32 have a preference. > > > > > > > > > > > need_active_balance() probably needs a new condition for the numa case > > > > where the busiest queue can't be selected and we have to trigger an > > > > active load_balance on a rq with only 1 thread but that is not running > > > > on its preferred node. Something like the untested below : > > > > > > > > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > > > > index e5da5eaab6ce..de1474191488 100644 > > > > --- a/kernel/sched/fair.c > > > > +++ b/kernel/sched/fair.c > > > > @@ -11150,6 +11150,24 @@ imbalanced_active_balance(struct lb_env *env) > > > > return 0; > > > > } > > > > > > > > +static inline bool > > > > +numa_active_balance(struct lb_env *env) > > > > +{ > > > > + struct sched_domain *sd = env->sd; > > > > + > > > > + /* > > > > + * We tried to migrate only a !numa task or a task on wrong node but > > > > + * the busiest queue with such task has only 1 running task. Previous > > > > + * attempt has failed so force the migration of such task. > > > > + */ > > > > + if ((env->fbq_type < all) && > > > > + (env->src_rq->cfs.h_nr_running == 1) && > > > > + (sd->nr_balance_failed > 0)) > > > > > > The last condition will still be a problem because of CPU_NEWLY_IDLE. The > > > nr_balance_failed counter doesn't get incremented very often. > > > > It waits for at least 1 failed CPU_IDLE load_balance > > > > > > > > julia > > > > > > > + return 1; > > > > + > > > > + return 0; > > > > +} > > > > + > > > > static int need_active_balance(struct lb_env *env) > > > > { > > > > struct sched_domain *sd = env->sd; > > > > @@ -11176,6 +11194,9 @@ static int need_active_balance(struct lb_env *env) > > > > if (env->migration_type == migrate_misfit) > > > > return 1; > > > > > > > > + if (numa_active_balance(env)) > > > > + return 1; > > > > + > > > > return 0; > > > > } > > > > > > > > > > > > > it to avoid the CPU with two threads that already prefer Socket 2. But I > > > > > don't know at the moment why that is the case. In any case, it's fine to > > > > > active balance from a CPU with only one thread, because Socket 2 will > > > > > even itself out afterwards. > > > > > > > > > > > > > > > > > You should not have more than 1 thread per CPU when there are N+1 > > > > > > threads on a node with N cores / 2N CPUs. > > > > > > > > > > Hmm, I think there is a miscommunication about cores and CPUs. The > > > > > machine has two sockets with 16 physical cores each, and thus 32 > > > > > hyperthreads. There are 64 threads running. > > > > > > > > Ok, I have been confused by what you wrote previously: > > > > " The context is that there are 2N threads running on 2N cores, one thread > > > > gets NUMA balanced to the other socket, leaving N+1 threads on one socket > > > > and N-1 threads on the other socket." > > > > > > > > I have assumed that there were N cores and 2N CPUs per socket as you > > > > mentioned Intel Xeon 6130 in the commit message . My previous emails > > > > don't apply at all with N CPUs per socket and the group_overloaded is > > > > correct. > > > > > > > > > > > > > > > > > > > > > > julia > > > > > > > > > > > This will enable the > > > > > > load_balance to try to migrate a task instead of some util(ization) > > > > > > and you should reach the active load balance. > > > > > > > > > > > > > > > > > > > > > In theory you should have the > > > > > > > > local "group_has_spare" and the busiest "group_fully_busy" (at most). > > > > > > > > This means that no group should be overloaded and load_balance should > > > > > > > > not try to migrate utli but only task > > > > > > > > > > > > > > I didn't collect information about the groups. I will look into that. > > > > > > > > > > > > > > julia > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > and changing the above test to: > > > > > > > > > > > > > > > > > > if ((env->migration_type == migrate_task || env->migration_type == migrate_util) && > > > > > > > > > (sd->nr_balance_failed > sd->cache_nice_tries+2)) > > > > > > > > > > > > > > > > > > seems to solve the problem. > > > > > > > > > > > > > > > > > > I will test this on more applications. But let me know if the above > > > > > > > > > solution seems completely inappropriate. Maybe it violates some other > > > > > > > > > constraints. > > > > > > > > > > > > > > > > > > I have no idea why this problem became more visible with EEVDF. It seems > > > > > > > > > to have to do with the time slices all turning out to be the same. I got > > > > > > > > > the same behavior in 6.5 by overwriting the timeslice calculation to > > > > > > > > > always return 1. But I don't see the connection between the timeslice and > > > > > > > > > the behavior of the idle task. > > > > > > > > > > > > > > > > > > thanks, > > > > > > > > > julia > > > > > > > > > > > > > > > > > > > >