Received: by 2002:a05:6a10:2726:0:0:0:0 with SMTP id ib38csp689471pxb; Thu, 24 Mar 2022 05:21:00 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzVq8hyZspheHa2O6S+ASvXHbJ/roSjEXZnrp7wjzcUEd3r9W4xc7iVGAryzGXESCRQhJJf X-Received: by 2002:a63:ec43:0:b0:382:11ef:b94a with SMTP id r3-20020a63ec43000000b0038211efb94amr3924915pgj.409.1648124460564; Thu, 24 Mar 2022 05:21:00 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1648124460; cv=none; d=google.com; s=arc-20160816; b=M1Sw5CbAdUl0Sv0QcLM7yuXJAMPIlsGh67ZhEjAFvn8hZM5P9vD36ZnQAmEDylSC9a SQBRgyb9I1ZTc4hbCmb6THoMPdEYVUWjrxnjmll2Zr4VD7noVrH9bTdVCyMt0CPlybqL WMgyUssRff+/mFNHlIbW5emIWDSgKV4yWZNgkE62/khbEI4dI1w/WTL3f7T5IlzMMV0P D280I5Tb+Z6wyFLkkJnKmCDRBMh9o4Oo7geg+KT01OcaBt5dFYvwi/bW1n19aMIH1Rxz 188q8/L+k/M0G54286+32QAxSVm04e7ojKfwsZh5TS8LIkVhDvBPKy92OYsguftkrZjU 2JhA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=ky26SBCtUapKPBOlXmRYotWknEZHsIyCEymHbAAXUHA=; b=onoVPiOJwAPai9xsBMbbJSBrjiSn2JCg3vpMoLb6ApmmENoLqgclf62zY18pcGxqFt 3lh1SOTfa8DqoeB5MsZi3CYuymOWnr/TRWz6WGjoh3IbGCoVnX8tg8whUFdlVr7PxuOu T4Tn6zSgaWOVotM80K/XlZKkmMUqAXMpmHb7Xqo+SxTWyOweOi3qudAlWVHTiS6Bgp/w 5+dazyRQg9CSPQ2wykCcKs5EYHiP726+wO9D1HOF0whufKKr5Ok9sIBc62fVNPpW9txO qWQqyL4yihqn0WCytpWZe1sOM5HYzoYfjUqQUJvs0ouyly0uWJ5KHVfWWE/IrK3fGiOX liyQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=scrEo17J; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id s21-20020a17090a5d1500b001bf6715212dsi8046151pji.104.2022.03.24.05.20.44; Thu, 24 Mar 2022 05:21:00 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=scrEo17J; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235497AbiCWLGq (ORCPT + 99 others); Wed, 23 Mar 2022 07:06:46 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:44874 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235234AbiCWLGp (ORCPT ); Wed, 23 Mar 2022 07:06:45 -0400 Received: from mail-wr1-x42b.google.com (mail-wr1-x42b.google.com [IPv6:2a00:1450:4864:20::42b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 652274A917 for ; Wed, 23 Mar 2022 04:05:14 -0700 (PDT) Received: by mail-wr1-x42b.google.com with SMTP id u16so1579131wru.4 for ; Wed, 23 Mar 2022 04:05:14 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=ky26SBCtUapKPBOlXmRYotWknEZHsIyCEymHbAAXUHA=; b=scrEo17J6Q8JXfoIHKn+qES+ujya4Mi1cDoSDQ0sOkVhWpg5r+D2pgkOT+SMlVt2CR mzGJo0itskQ/WFjsJw61drGodykkhhLpruvdob+lAfaSb09+b8Pg0D/zWT1vGTIwnHYc QJg1R/13UmeYphpXUG1RTzbI46BGHVGXQG36+H61REizUjggA4Kot+iEf48T2+iaQZ+d dgNv13ieJjvGP5ox+LbUAteIF1K1Xtmtq57qbghuKn+JsA+fUKbsftBorjkdgczLqi/H 8yyE43q4YwN4eZkI2S8x+FCjC0kTHNn54FSdITQ/DzEuLTDMqbwxKGKxo2n/uhE7qlkN 79Gw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=ky26SBCtUapKPBOlXmRYotWknEZHsIyCEymHbAAXUHA=; b=EHBjW2tzQ8i6SxifH7xoV/8LrtKcTNLZ2RKJ7ZqGsfm11aNuUtztbgbIMziOa8renz FHVjAP96EDwxp2I6NJVYZpJZ+riOSvZO847NbYM24pn4uyZb6la11qjVyylSD3uVUYBd PqkO9UfzJQ5D3EXr/fCrJoHWM1/aPWr7lHTKHvfT8ddURH6Z40VnJfmrjO1Nxs6tLMLp y2PKuZg8AwuFL8r2TE04u+VwEOVGEpaekeuLmqu+upmdnit8VdAvhtgIMBLRLEpnxHf3 pXOBwboRksShocaqWfpVAhCI8g7i3nPJav6jhoGAy8CSm9Bcl6Xq0YTjG7IYIeVzCJAJ AmUw== X-Gm-Message-State: AOAM531bxQd0Rd3azgfb2Ms9t/yuUJ1ydGJ02T7IE1qk8mk6aUoRN185 n5rZ0+eug7tylIuNCPekhXmdchF444w+078kQTQIVg== X-Received: by 2002:adf:f5c4:0:b0:205:2836:f1e5 with SMTP id k4-20020adff5c4000000b002052836f1e5mr6559414wrp.519.1648033512911; Wed, 23 Mar 2022 04:05:12 -0700 (PDT) MIME-Version: 1.0 References: <20220308205000.27646-1-mike.leach@linaro.org> <639c902b-535e-8627-de04-9bcbbdc98ed5@arm.com> <17215ebc-dbf2-81ca-7ef4-b4ebc3cb9d28@arm.com> In-Reply-To: From: Mike Leach Date: Wed, 23 Mar 2022 11:05:02 +0000 Message-ID: Subject: Re: [PATCH 00/10] coresight: Add new API to allocate trace source ID values To: Al Grant Cc: Suzuki Poulose , "coresight@lists.linaro.org" , "linux-arm-kernel@lists.infradead.org" , "linux-kernel@vger.kernel.org" , "peterz@infradead.org" , "mingo@redhat.com" , "acme@kernel.org" , "linux-perf-users@vger.kernel.org" , James Clark Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Al, On Wed, 23 Mar 2022 at 10:36, Al Grant wrote: > > > -----Original Message----- > > From: Mike Leach > > Sent: 23 March 2022 10:08 > > To: Suzuki Poulose > > Cc: coresight@lists.linaro.org; linux-arm-kernel@lists.infradead.org; linux- > > kernel@vger.kernel.org; peterz@infradead.org; mingo@redhat.com; > > acme@kernel.org; linux-perf-users@vger.kernel.org; James Clark > > > > Subject: Re: [PATCH 00/10] coresight: Add new API to allocate trace source ID > > values > > > > Hi Suzuki, > > > > On Tue, 22 Mar 2022 at 18:52, Suzuki K Poulose > > wrote: > > > > > > Hi Mike > > > > > > On 22/03/2022 14:27, Mike Leach wrote: > > > > Hi Suzuki > > > > > > > > On Tue, 22 Mar 2022 at 12:35, Suzuki Kuruppassery Poulose > > > > wrote: > > > >> > > > >> On 22/03/2022 11:38, Mike Leach wrote: > > > >>> HI Suzuki, > > > >>> > > > >>> On Tue, 22 Mar 2022 at 10:43, Suzuki Kuruppassery Poulose > > > >>> wrote: > > > >>>> > > > >>>> + Cc: James Clark > > > >>>> > > > >>>> Hi Mike, > > > >>>> > > > >>>> On 08/03/2022 20:49, Mike Leach wrote: > > > >>>>> The current method for allocating trace source ID values to > > > >>>>> sources is to use a fixed algorithm for CPU based sources of > > (cpu_num * 2 + 0x10). > > > >>>>> The STM is allocated ID 0x1. > > > >>>>> > > > >>>>> This fixed algorithm is used in both the CoreSight driver code, > > > >>>>> and by perf when writing the trace metadata in the AUXTRACE_INFO > > record. > > > >>>>> > > > >>>>> The method needs replacing as currently:- 1. It is inefficient > > > >>>>> in using available IDs. > > > >>>>> 2. Does not scale to larger systems with many cores and the > > > >>>>> algorithm has no limits so will generate invalid trace IDs for cpu > > number > 44. > > > >>>> > > > >>>> Thanks for addressing this issue. > > > >>>> > > > >>>>> > > > >>>>> Additionally requirements to allocate additional system IDs on > > > >>>>> some systems have been seen. > > > >>>>> > > > >>>>> This patch set introduces an API that allows the allocation of > > > >>>>> trace IDs in a dynamic manner. > > > >>>>> > > > >>>>> Architecturally reserved IDs are never allocated, and the system > > > >>>>> is limited to allocating only valid IDs. > > > >>>>> > > > >>>>> Each of the current trace sources ETM3.x, ETM4.x and STM is > > > >>>>> updated to use the new API. > > > >>>>> > > > >>>>> perf handling is changed so that the ID associated with the CPU > > > >>>>> is read from sysfs. The ID allocator is notified when perf > > > >>>>> events start and stop so CPU based IDs are kept constant > > throughout any perf session. > > > >>>>> > > > >>>>> For the ETMx.x devices IDs are allocated on certain events > > > >>>>> a) When using sysfs, an ID will be allocated on hardware enable, > > > >>>>> and freed when the sysfs reset is written. > > > >>>>> b) When using perf, ID is allocated on hardware enable, and > > > >>>>> freed on hardware disable. > > > >>>>> > > > >>>>> For both cases the ID is allocated when sysfs is read to get the > > > >>>>> current trace ID. This ensures that consistent decode metadata > > > >>>>> can be extracted from the system where this read occurs before > > device enable. > > > >>>> > > > >>>> > > > >>>>> > > > >>>>> Note: This patchset breaks backward compatibility for perf record. > > > >>>>> Because the method for generating the AUXTRACE_INFO meta data > > > >>>>> has changed, using an older perf record will result in metadata > > > >>>>> that does not match the trace IDs used in the recorded trace data. > > > >>>>> This mismatch will cause subsequent decode to fail. Older > > > >>>>> versions of perf will still be able to decode data generated by the > > updated system. > > > >>>> > > > >>>> I have some concerns over this and the future plans for the > > > >>>> dynamic allocation per sink. i.e., we are breaking/modifying the > > > >>>> perf now to accommodate the dynamic nature of the trace id of a > > given CPU/ETM. > > > >>> > > > >>> I don't beleive we have a choice for this - we cannot retain what > > > >>> is an essentially broken allocation mechanism. > > > >>> > > > >> > > > >> I completely agree and I am happy with the current step by step > > > >> approach of moving to a dynamic allocation scheme. Apologies, this > > > >> wasn't conveyed appropriately. > > > >> > > > >>>> The proposed approach of exposing this via sysfs may (am not sure > > > >>>> if this would be the case) break for the trace-id per sink > > > >>>> change, as a sink could assign different trace-id for a CPU depending. > > > >>>> > > > >>> > > > >>> If a path exists between a CPU and a sink - the current framework > > > >>> as far as I can tell would not allow for a new path to be set up > > > >>> between the cpu and another sink. > > > >> > > > >> e.g, if we have concurrent perf sessions, in the future with sink > > > >> based allocation : > > > >> > > > >> perf record -e cs_etm/@sink1/... payload1 perf record -e > > > >> cs_etm/@sink2/... payload2 > > > >> perf record -e cs_etm// ... payload3 > > > >> > > > >> The trace id allocated for first session for CPU0 *could* be > > > >> different from that of the second or the third. > > > > > > > > If these sessions run concurrently then the same Trace ID will be > > > > used for CPU0 for all the sessions. > > > > We ensure this by notifications that a cs_etm session is starting > > > > and stopping - and keep a refcount. > > > > > > The scheme is fine now, with a global trace-id map. But with per-sink > > > allocation, this could cause problems. > > > > > > e.g., there could be a situation where: > > > > > > trace_id[CPU0][sink0] == trace_id[CPU1][sink1] > > > > > > So if we have a session where both CPU0 and CPU1 trace to a common > > > sink, we get the trace mixed with no way of splitting them. As the > > > perf will read the trace-id for CPU0 from that of sink0 and CPU1 from sink1. > > > > I think we need to consider the CoreSight hardware topology here. > > > > Any CPUx that can trace to a sink reachable by another CPUy must always > > get trace IDs from the same pool as CPUy. > > > > Consider the options for multi sink topologies:- > > > > CPU0->funnel0->ETF->ETR > > CPU1--+^ > > > > Clearly - in-line sinks can never have simultaneous independent sessions - > > the session into ETR traces through ETF > > > > Now we could have replicators / programmable replicators - > > > > ATB->Prog replicator->ETR0 > > +->ETR1 > > > > however programmable replicators use trace ID for filtering - this is > > effectively a single sink on the input, so once again the Trace IDs must come > > from the same pool. > > > > Now, we could have independent per cluster / socket topology Cluster 0 > > CPU0->funnel0->ETF0->ETR0 > > CPU1--+^ > > > > Cluster 1 > > CPU2->funnel1->ETF1->ETR1 > > CPU3--+^ > > > > Here cluster 0 & 1 can have independent sets of trace IDs as their respective > > cores can never trace to the same sink. > > > > Finally we have the ETE+TRBE 1:1 type topologies. These could actually not > > bother allocating any trace ID when in 1:1 mode, which should probably be a > > follow on incremental addition to this initial set. > > > > So, my conclusion when I was considering all this is that "per-sink" > > trace Id allocation is in fact "per unique trace path set" allocation. > > The id pools can't always be defined statically in all situations. > E.g. if you have 128 CPUs each with their own ETR, and also > replicated into a funnel network leading to a common ETR. > Agreed - and this will be an issue that is needed to be considered when implementing multiple ID pools. This is a possibility when ETE traces with TRBE switched off. > This topology supports (at least) two distinct modes: (a) all CPUs > enabled for tracing to their own ETRs (b) a selection of CPUs > (up to some limit), combined together. Both are valid dynamic > configurations. Perf might have a preference on which one to use, > but both are valid. > > But there's no static id pool that works for both. The pool for > (b) has to allocate to some random selection of 128 CPUs, > from only around 120 numbers. The pool has to take account of > which CPUs are selected. > Also agreed - it is entirely possible to run out of IDs. The choice then is to not trace on any, or simply trace on the first N that get valid IDs. Which has to be a function of the perf handing code. IDs can be allocated when perf is running through the code to set up the trace paths from the selected cores. The present system "allows" trace to go from all cores by allocating invalid trace IDs once we get past too many cores (46 at present). > So your comment "any CPU that can trace..." has to be interpreted as > "any CPU that can trace in the currently configured dynamic topology" > rather than "any CPU that can be dynamically configured to trace..."... > is that what you meant? Pretty much. The real problems with "per-sink" / Trace ID pools is recognizing the current set up topology when this can be dynamic. I anticipated that the metadata attached to allocating trace IDs will have to expand to recognise this one we get to advanced topologies - which is why the API was designed to always pass in the metadata. > > Alternatively, we could just declare that such systems are too > complicated to support, and say that we wouldn't support the > use of a global sink that (statically) was reachable by 128 CPUs. > I think that we can support these systems - it is just that the user / perf will have to accept that there is a limit to the number of trace IDs that can be allocated & hence the number of CPUs that will actually trace the event. That will mean selecting a set of CPUS to trace on, or not scheduling the event on a core that cannot trace due to the lack of a trace ID. The infrastructure will not enable a trace path if that path cannot allocate a unique ID for the current trace ID set. (this has always been the case - the only change is that it is handled by the dynamic allocator now) Alternatively we could simply say that Trace IDs are a limited resource - we will only support using a single pool at once (111 possible IDs when you take into account the reserved values) - which is the situation with this patchset. Mike > Al > > > > > > > > > > > > > > So my point is, we are changing the ABI for perf to grab the TraceID > > > with your patches. And clearly this approach could break easily when > > > we extend to sink-based idmap. So, lets make the ABI change for perf > > > scalable and bullet proof (as far as we can) by exposing this > > > information via the perf RECORD. That way any future changes in the > > > scheme won't affect the perf as long as it has a reliable information > > > within each "record". > > > > > > > > > My point is, let us fix this once and for all, so that we don't need > > > to change this again. I understand this involves more work in the perf > > > tool. I believe that is for better > > > > > > Thoughts ? > > > > > > > My preference is the incremental approach. > > Fix the trace ID allocation issues that partners are having now, then update > > to the perf record approach in a separate follow up patchset. > > Then when we start to see systems that require it - update to using the per- > > unique-path trace ID pools. > > > > Regards > > > > Mike > > > > > Suzuki > > > > > > > > -- > > Mike Leach > > Principal Engineer, ARM Ltd. > > Manchester Design Centre. UK > > _______________________________________________ > > CoreSight mailing list -- coresight@lists.linaro.org To unsubscribe send an > > email to coresight-leave@lists.linaro.org -- Mike Leach Principal Engineer, ARM Ltd. Manchester Design Centre. UK