Received: by 2002:a05:6358:16cc:b0:ea:6187:17c9 with SMTP id r12csp9911862rwl; Wed, 11 Jan 2023 11:43:04 -0800 (PST) X-Google-Smtp-Source: AMrXdXursn4sjq498OGsa5+u57AuI/3+1O059xOyCj05alqvDMIT1aehG2O0qsPT5vsV4ZHj/eR0 X-Received: by 2002:a05:6402:150b:b0:493:a6eb:874e with SMTP id f11-20020a056402150b00b00493a6eb874emr24172026edw.5.1673466184421; Wed, 11 Jan 2023 11:43:04 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1673466184; cv=none; d=google.com; s=arc-20160816; b=X+pZLMa+Dbx0sCmMmBGMj/x5uz7W7LKW6QcEbThKj5MAIlmBRuV/7z2q+Az8LmMmp9 kiE25odOmEJr7FB+X3r1wiCiUHYGXVyG127kGF8txK9oHS9Zd4eqGL55kBiVRiRvBrp4 EcOzOPFaem3u9c8fuL8ZZPP8let7Ipbpj/qERyaisT60Ozq/LHyn+qWDxyLi7c+kSsl7 cb2WyYQj3MS9L0WyqEVzvP+DY5kiBQiX4yoxwr5OxFlY5qpvcX/v8/0dB6X0SgQjjggj XqHahUeD0PGRYs5uKEfd2r1mZk8wnQ1IKQq6a2+OEi25ZjlOJeMFwAGSrkJwuiZFRVwu jQXw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to :content-language:references:cc:to:from:subject:user-agent :mime-version:date:message-id:dkim-signature; bh=lP+IyeSfqqhlvSsIkoGIIRpA2P+LRGc74DZkL7AQAlc=; b=N6bDG+D+HgjoH7KrWoleBHraEMVWe1wxBtcIcxi+4Qz64M47SOtrE3us0mAjPIqyQz jCVOibQNb8shwt2O50T7Gfg99CeROyYJtCMQB9QFsPlRBrlupoqTO65CkUkHcL2xoYNn ywvHrTBef7TudRjJAk1p8eLO+ZAaE4S8o6Y+n0ku99DVdFTpG6uumhhvZVSJ5PnGMIgW k7H3eAGh49xYwe9jUdcGLE2MKeItfOUOgLmedJHqG5KF2YT1JeQELzTwtbF/g9SD8adY iDnBrgOKWnlbGNgWCY/DzO9E4B1byydO/q940uLuZC/bHfiFNmQcHvjbFffXNb/Ay6RN er/Q== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@efficios.com header.s=smtpout1 header.b=BYdzV5aG; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=efficios.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id o6-20020a056402444600b004829afcc563si14376269edb.404.2023.01.11.11.42.51; Wed, 11 Jan 2023 11:43:04 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@efficios.com header.s=smtpout1 header.b=BYdzV5aG; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=efficios.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232487AbjAKTaz (ORCPT + 51 others); Wed, 11 Jan 2023 14:30:55 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:41460 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230197AbjAKTaw (ORCPT ); Wed, 11 Jan 2023 14:30:52 -0500 Received: from smtpout.efficios.com (unknown [IPv6:2607:5300:203:b2ee::31e5]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1C7DFD13D for ; Wed, 11 Jan 2023 11:30:48 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=efficios.com; s=smtpout1; t=1673465447; bh=P5o+X5w9OffEjvSaI+GOzWtEzWbMVtkRV9QeB10s/Uo=; h=Date:Subject:From:To:Cc:References:In-Reply-To:From; b=BYdzV5aGrS2mDWushzIiZHeGklmS+cyaMUOmgCWDDprghvolkS7hL3SxkQfQX0Hyj 9AQh1Wja5Oez71znxfEf9bgyIcJWIteRHxBFPAJ7lBAqji9a8Yt2x6UzSMaVLTCZV3 Y9rWPBK8qbWVQAfo3B3g2OjSGPRKLl1duFIfsFbRUJMgh7wlDtJXyVHPyTFi4FDUWW XxIA8aXRWf+eSeJJrsKCw4b2NL5jGDpwFVmSuUid7L4mJoXcAdJ/4gF1+2txXhZ8d/ ACiXvcvbA1PVFeyOqffnZZMaRn8hpbGitZF+u+T6HGPuR4I9NEPyOmjHIdcHLPvlrA jYclhyVaj2BYA== Received: from [172.16.0.101] (192-222-180-24.qc.cable.ebox.net [192.222.180.24]) by smtpout.efficios.com (Postfix) with ESMTPSA id 4Nsd7q5WfrzgJM; Wed, 11 Jan 2023 14:30:47 -0500 (EST) Message-ID: Date: Wed, 11 Jan 2023 14:31:21 -0500 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.6.0 Subject: Re: rseq CPU ID not correct on 6.0 kernels for pinned threads From: Mathieu Desnoyers To: Florian Weimer , linux-kernel@vger.kernel.org Cc: Ingo Molnar , Peter Zijlstra , libc-alpha@sourceware.org References: <87lem9cnxr.fsf@oldenburg.str.redhat.com> <8b52b0d6-c973-f959-b44a-1b54fb808a04@efficios.com> Content-Language: en-US In-Reply-To: <8b52b0d6-c973-f959-b44a-1b54fb808a04@efficios.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-1.3 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,NICE_REPLY_A,RDNS_NONE, SPF_HELO_NONE,SPF_PASS autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2023-01-11 09:52, Mathieu Desnoyers wrote: > On 2023-01-11 06:26, Florian Weimer wrote: >> The glibc test suite contains a test that verifies that sched_getcpu >> returns the expected CPU number for a thread that is pinned (via >> sched_setaffinity) to a specific CPU.  There are other threads running >> which attempt to de-schedule the pinned thread from its CPU.  I believe >> the test is correctly doing what it is expected to do; it is invalid >> only if one believes that it is okay for the kernel to disregard the >> affinity mask for scheduling decisions. >> >> These days, we use the cpu_id rseq field as the return value of >> sched_getcpu if the kernel has rseq support (which it has in these >> cases). >> >> This test has started failing sporadically for us, some time around >> kernel 6.0.  I see failure occasionally on a Fedora builder, it runs: >> >> Linux buildvm-x86-26.iad2.fedoraproject.org 6.0.15-300.fc37.x86_64 #1 >> SMP PREEMPT_DYNAMIC Wed Dec 21 18:33:23 UTC 2022 x86_64 x86_64 x86_64 >> GNU/Linux >> >> I think I've seen it on the x86-64 builder only, but that might just be >> an accident. >> >> The failing tests log this output: >> >> =====FAIL: nptl/tst-thread-affinity-pthread.out===== >> info: Detected CPU set size (in bits): 64 >> info: Maximum test CPU: 5 >> error: Pinned thread 1 ran on impossible cpu 0 >> error: Pinned thread 0 ran on impossible cpu 0 >> info: Main thread ran on 4 CPU(s) of 6 available CPU(s) >> info: Other threads ran on 6 CPU(s) >> =====FAIL: nptl/tst-thread-affinity-pthread2.out===== >> info: Detected CPU set size (in bits): 64 >> info: Maximum test CPU: 5 >> error: Pinned thread 1 ran on impossible cpu 1 >> error: Pinned thread 2 ran on impossible cpu 0 >> error: Pinned thread 3 ran on impossible cpu 3 >> info: Main thread ran on 5 CPU(s) of 6 available CPU(s) >> info: Other threads ran on 6 CPU(s) >> >> But I also encountered one local failure, but it is rare.  Maybe it's >> load-related.  There shouldn't be any CPU unplug or anything like that >> involved here. >> >> I am not entirely sure if something is changing CPU affinities from >> outside the process (which would be quite wrong, but not a kernel bug). >> But in the past, our glibc test has detected real rseq cpu_id >> brokenness, so I'm leaning towards that as the cause this time, too. > > It can be caused by rseq failing to update the cpu number field on > return to userspace. Tthis could be validated by printing the regular > getcpu vdso value and/or the value returned by the getcpu system call > when the error is triggered, and see whether it matches the rseq cpu id > value. > > It can also be caused by scheduler failure to take the affinity into > account. > > As you also point out, it can also be caused by some other task > modifying the affinity of your task concurrently. You could print > the result of sched_getaffinity on error to get a better idea of > the expected vs actual mask. > > Lastly, it could be caused by CPU hotplug which would set all bits > in the affinity mask as a fallback. As you mention it should not be > the cause there. > > Can you share your kernel configuration ? Also, can you provide more information about the cpufreq driver and governor used in your system ? e.g. output of cpupower frequency-info and also output of sysctl kernel.sched_energy_aware Is this on a physical machine or in a virtual machine ? Thanks, Mathieu > > Thanks, > > Mathieu > >> >> Thanks, >> Florian >> > -- Mathieu Desnoyers EfficiOS Inc. https://www.efficios.com