Received: by 2002:a05:6a10:2726:0:0:0:0 with SMTP id ib38csp215040pxb; Wed, 23 Mar 2022 16:38:41 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzujQfbw+BVeWT/iKSJNvW/M5OMBYkH9PykS4Nb30WeZYfRv0rQfV03fgekQdgIleVAWGUZ X-Received: by 2002:a17:907:1b09:b0:6d8:faa8:4a06 with SMTP id mp9-20020a1709071b0900b006d8faa84a06mr2724555ejc.701.1648078721723; Wed, 23 Mar 2022 16:38:41 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1648078721; cv=none; d=google.com; s=arc-20160816; b=KuDbsnfGRaaI15X0Be5q7+C7BWEpGBW9dDVgIVN6i683FzGRmbVrGed/JQzqzaMDxM Jtjt1MSBOd7KkPbYuSWWr7Dwr0bFXLwzOigtxtbP8LDwuFxyppLlOciqPdHH3bLdSJ2i Suq2m7MDInfxQj5yvOwl6UFVTeBbcyQwcB4isTLwbpu+WP1l/Y9jf44Up6FfQfWVg3j9 Hsfqo4yYLLKD7rEKu0LJIcp0d2GR9PElIII5Rf9olQgtTFg65FkdcZLveEnWioWGov1e +6FlRCzaTei8aVooB0HK+TaWXS6DpiFhAFMX8xpzQHe34Yxr/Akd9jT/aSeAMGndlWUw 5JzQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=Pgn+1JHw/CHLccewmVrT5fNYmRXbpY7xb/kSDqUDndM=; b=sUSxLFh3XJgAXm/HYDH8nsYok9pia5FW8I1X2SabHztZ7yckhzwLNN/JF0u5lHft93 sKa9j0Gr7blulstHqZPg0FSj81G3OXZ+pdBULtROfupJBTooPtrgXo8O3K7tRAh39eOu CR1D6IbqXWK6AOc5XlzsJn5H2rtAWJoYTyuVi7JE2t3gGnLPoTnGQFfAATWupDLsqSGN SeMeliJhbKpw3iCkiZGy8OGrLCaMfkjsFNOml85JrW3hjBBQdyTIyQ2Z/qYNPQXa2//m zL5y0CmbkT0dtqdX84ORhwJj6sdlmdk/yDkbQC7HwBZ3FAcb4IIcmB57pGPlxEvc+02P lHRQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=GVfzuF71; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id hd18-20020a170907969200b006e05946ea68si5706487ejc.1002.2022.03.23.16.38.16; Wed, 23 Mar 2022 16:38:41 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=GVfzuF71; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235830AbiCWLhd (ORCPT + 99 others); Wed, 23 Mar 2022 07:37:33 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:45122 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232743AbiCWLhb (ORCPT ); Wed, 23 Mar 2022 07:37:31 -0400 Received: from mail-wm1-x329.google.com (mail-wm1-x329.google.com [IPv6:2a00:1450:4864:20::329]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 43A617560E for ; Wed, 23 Mar 2022 04:36:00 -0700 (PDT) Received: by mail-wm1-x329.google.com with SMTP id p189so744135wmp.3 for ; Wed, 23 Mar 2022 04:36:00 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=Pgn+1JHw/CHLccewmVrT5fNYmRXbpY7xb/kSDqUDndM=; b=GVfzuF71g5LqYrBx7bHu3yG77CW1N11Q2XqfFcsNkK0dx+7Ghqw5o2k/jCv8qe8g5l R89j6Au5Zos118VJqYPz8j/l5Rc4B5MxO3qykCiQGA4qkwIYv+edsmAfEOfLI70Dkwg8 zyt/jh6a2k/C9Kb4MocEzwgLVXKjqmmDPxT7m2tFaJsa3uH9fC/gCZTBzlqlWoOUXbYy nI1nVPjKzLLCS9t2FeJQr7CMmVygkRHkoUfoyr64fhrKMDKpWDlUIyIcqYd00XtpKFrP MqWYYUQWB6wCD035qQFjbizkXNGctq9FmWqAfF+qln6f++HjMQ5SW61IgzAoAIsGupGt RKRw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=Pgn+1JHw/CHLccewmVrT5fNYmRXbpY7xb/kSDqUDndM=; b=inp+jvX+pgMOKyAg0Xm8hsd7eCDOLMSx5n/mE23lWUxIxgj9+7XKwyIloX02WncD9Z Q6JTQ5hYInXgpEndOZRYo7QzSonmR6RgAvyK/golGjaRpMRSORBF71yDretSY+ukXpp/ TkD3UjxIbU66fgwCvhCQgEjAyXVP7NenfETEWfIVB65qlBbbkpRV3LT/N1ENpJXn0xfb ajGwe2J3mN9jlMJvACu9PB11YY04UIgAfOC7SAbGsM02LlWSWnlAEhll+MUJnJnluzom 4dqjO6IepEaz35g45LnBO2lwqs2SYj4lh4atQCXCD7/Xpr4nZ/2LRfEtiymn6zKnWjHW uHOg== X-Gm-Message-State: AOAM533JTEYCaI+tB8MxsHru4CwTrxCq0ctCElvD9zfHApaiO1u9fGBr nGSNZ9TTXZK93G0nhn/79lalfQmTwgYbD8K1E72rtQ== X-Received: by 2002:a05:600c:19ca:b0:38c:b84f:421a with SMTP id u10-20020a05600c19ca00b0038cb84f421amr7997487wmq.118.1648035358616; Wed, 23 Mar 2022 04:35:58 -0700 (PDT) MIME-Version: 1.0 References: <20220308205000.27646-1-mike.leach@linaro.org> <639c902b-535e-8627-de04-9bcbbdc98ed5@arm.com> <17215ebc-dbf2-81ca-7ef4-b4ebc3cb9d28@arm.com> In-Reply-To: From: Mike Leach Date: Wed, 23 Mar 2022 11:35:47 +0000 Message-ID: Subject: Re: [PATCH 00/10] coresight: Add new API to allocate trace source ID values To: Suzuki Kuruppassery Poulose Cc: coresight@lists.linaro.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, mathieu.poirier@linaro.org, peterz@infradead.org, mingo@redhat.com, acme@kernel.org, linux-perf-users@vger.kernel.org, leo.yan@linaro.org, James Clark Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Suzuki On Wed, 23 Mar 2022 at 10:41, Suzuki Kuruppassery Poulose wrote: > > Hi Mike > > On 23/03/2022 10:07, Mike Leach wrote: > > Hi Suzuki, > > > > On Tue, 22 Mar 2022 at 18:52, Suzuki K Poulose wrote: > >> > >> Hi Mike > >> > >> On 22/03/2022 14:27, Mike Leach wrote: > >>> Hi Suzuki > >>> > >>> On Tue, 22 Mar 2022 at 12:35, Suzuki Kuruppassery Poulose > >>> wrote: > >>>> > >>>> On 22/03/2022 11:38, Mike Leach wrote: > >>>>> HI Suzuki, > >>>>> > >>>>> On Tue, 22 Mar 2022 at 10:43, Suzuki Kuruppassery Poulose > >>>>> wrote: > >>>>>> > >>>>>> + Cc: James Clark > >>>>>> > >>>>>> Hi Mike, > >>>>>> > >>>>>> On 08/03/2022 20:49, Mike Leach wrote: > >>>>>>> The current method for allocating trace source ID values to sources is > >>>>>>> to use a fixed algorithm for CPU based sources of (cpu_num * 2 + 0x10). > >>>>>>> The STM is allocated ID 0x1. > >>>>>>> > >>>>>>> This fixed algorithm is used in both the CoreSight driver code, and by > >>>>>>> perf when writing the trace metadata in the AUXTRACE_INFO record. > >>>>>>> > >>>>>>> The method needs replacing as currently:- > >>>>>>> 1. It is inefficient in using available IDs. > >>>>>>> 2. Does not scale to larger systems with many cores and the algorithm > >>>>>>> has no limits so will generate invalid trace IDs for cpu number > 44. > >>>>>> > >>>>>> Thanks for addressing this issue. > >>>>>> > >>>>>>> > >>>>>>> Additionally requirements to allocate additional system IDs on some > >>>>>>> systems have been seen. > >>>>>>> > >>>>>>> This patch set introduces an API that allows the allocation of trace IDs > >>>>>>> in a dynamic manner. > >>>>>>> > >>>>>>> Architecturally reserved IDs are never allocated, and the system is > >>>>>>> limited to allocating only valid IDs. > >>>>>>> > >>>>>>> Each of the current trace sources ETM3.x, ETM4.x and STM is updated to use > >>>>>>> the new API. > >>>>>>> > >>>>>>> perf handling is changed so that the ID associated with the CPU is read > >>>>>>> from sysfs. The ID allocator is notified when perf events start and stop > >>>>>>> so CPU based IDs are kept constant throughout any perf session. > >>>>>>> > >>>>>>> For the ETMx.x devices IDs are allocated on certain events > >>>>>>> a) When using sysfs, an ID will be allocated on hardware enable, and freed > >>>>>>> when the sysfs reset is written. > >>>>>>> b) When using perf, ID is allocated on hardware enable, and freed on > >>>>>>> hardware disable. > >>>>>>> > >>>>>>> For both cases the ID is allocated when sysfs is read to get the current > >>>>>>> trace ID. This ensures that consistent decode metadata can be extracted > >>>>>>> from the system where this read occurs before device enable. > >>>>>> > >>>>>> > >>>>>>> > >>>>>>> Note: This patchset breaks backward compatibility for perf record. > >>>>>>> Because the method for generating the AUXTRACE_INFO meta data has > >>>>>>> changed, using an older perf record will result in metadata that > >>>>>>> does not match the trace IDs used in the recorded trace data. > >>>>>>> This mismatch will cause subsequent decode to fail. Older versions of > >>>>>>> perf will still be able to decode data generated by the updated system. > >>>>>> > >>>>>> I have some concerns over this and the future plans for the dynamic > >>>>>> allocation per sink. i.e., we are breaking/modifying the perf now to > >>>>>> accommodate the dynamic nature of the trace id of a given CPU/ETM. > >>>>> > >>>>> I don't beleive we have a choice for this - we cannot retain what is > >>>>> an essentially broken allocation mechanism. > >>>>> > >>>> > >>>> I completely agree and I am happy with the current step by step approach > >>>> of moving to a dynamic allocation scheme. Apologies, this wasn't > >>>> conveyed appropriately. > >>>> > >>>>>> The proposed approach of exposing this via sysfs may (am not sure if > >>>>>> this would be the case) break for the trace-id per sink change, as a > >>>>>> sink could assign different trace-id for a CPU depending. > >>>>>> > >>>>> > >>>>> If a path exists between a CPU and a sink - the current framework as > >>>>> far as I can tell would not allow for a new path to be set up between > >>>>> the cpu and another sink. > >>>> > >>>> e.g, if we have concurrent perf sessions, in the future with sink based > >>>> allocation : > >>>> > >>>> perf record -e cs_etm/@sink1/... payload1 > >>>> perf record -e cs_etm/@sink2/... payload2 > >>>> perf record -e cs_etm// ... payload3 > >>>> > >>>> The trace id allocated for first session for CPU0 *could* be different > >>>> from that of the second or the third. > >>> > >>> If these sessions run concurrently then the same Trace ID will be used > >>> for CPU0 for all the sessions. > >>> We ensure this by notifications that a cs_etm session is starting and > >>> stopping - and keep a refcount. > >> > >> The scheme is fine now, with a global trace-id map. But with per-sink > >> allocation, this could cause problems. > >> > >> e.g., there could be a situation where: > >> > >> trace_id[CPU0][sink0] == trace_id[CPU1][sink1] > >> > >> So if we have a session where both CPU0 and CPU1 trace to a common sink, > >> we get the trace mixed with no way of splitting them. As the perf will > >> read the trace-id for CPU0 from that of sink0 and CPU1 from sink1. > > > > I think we need to consider the CoreSight hardware topology here. > > > > Any CPUx that can trace to a sink reachable by another CPUy must > > always get trace IDs from the same pool as CPUy. > > > > Consider the options for multi sink topologies:- > > > > CPU0->funnel0->ETF->ETR > > CPU1--+^ > > > > Clearly - in-line sinks can never have simultaneous independent > > sessions - the session into ETR traces through ETF > > > > Now we could have replicators / programmable replicators - > > > > ATB->Prog replicator->ETR0 > > +->ETR1 > > > > however programmable replicators use trace ID for filtering - this is > > effectively a single sink on the input, so once again the Trace IDs > > must come from the same pool. > > > > Now, we could have independent per cluster / socket topology > > Cluster 0 > > CPU0->funnel0->ETF0->ETR0 > > CPU1--+^ > > > > Cluster 1 > > CPU2->funnel1->ETF1->ETR1 > > CPU3--+^ > > > > What if the ETR was a common one ? i.e. > > Cluster0 > CPU0 -> ETF0 ..... > \ > Cluster1 -- ETR0 > CPU1 -> ETF1 ..... / > > And lets there are 3 perf sessions in parallel, started in the > order below > > perf record -e cs_etm/@etf0/ app1 # CPU0 gets a trace-id[etf0] -> 0x50 > perf record -e cs_etm/@etf1/ app2 # CPU1 gets a trace-id[etf1] -> 0x50 > perf record -e cs_etm/@etr/ app3 # CPU0 and CPU1 both use the existing > trace ids from the allocations. > This could be treated as a single combined topology - as soon as any sink is reachable by CPU0 and CPU1 then we have to treat this as having a single pool of trace IDs and so CPU0 and CPU1 cannot have the same ID. Alternatively, once the allocation metadata is expanded to recognize trace ID pools - it is entirely possible to ensure that the CPU / ID fixing is done on a per pool basis. One of the reasons we need to ensure that the CPU / ID allocation remains constant is that an event can be scheduled on a CPU multiple times for a single aux buffer - resulting in multiple trace blocks in the buffer / multiple buffers in the data file, so we cannot have the ID change mid buffer / file without significant changes to the decode process and tracking of CPU / ID changes on a intra buffer basis. Mike > So, when app3 threads are scheduled on CPU0 & CPU1, we get the trace in > ETR with the same trace-id of 0x50. > > Suzuki -- Mike Leach Principal Engineer, ARM Ltd. Manchester Design Centre. UK