Received: by 2002:a05:6a10:6d10:0:0:0:0 with SMTP id gq16csp952633pxb; Fri, 15 Apr 2022 16:26:22 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz5LsfCOq0HnqddI9hhows6wRJZudMwwR2hJ9/nWwQFLQP8pHywO0v30BsKBm/uheJybRmK X-Received: by 2002:a17:90b:3ec5:b0:1c7:77ab:3854 with SMTP id rm5-20020a17090b3ec500b001c777ab3854mr1217274pjb.156.1650065182698; Fri, 15 Apr 2022 16:26:22 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1650065182; cv=none; d=google.com; s=arc-20160816; b=jj/ZTvbaTmEPswR3VHo5+LAAaY23Ry0bdY/JkAhzMocMqUFQ20McH/3QnRVIF9Xcm6 8rMAETY5dTZeAOiFIr7Wyaakj0dgropni/Ye2LaZOLsipdnBYm/88xI2lHrSjXd36SVh 65Cd0zQCqAA+buLCx3H5P5cy26tpPISmIw9BP1UxGlQQZRRHBj09pRonqUTAGva2+Lti xmd1cFtPCPqnaK2ZJCABppipjLbdC/1cml3jhSNRuvo65ZV/MS5RrmGeVVS9jPF5S1NC nUJ/aFwTiMgCb/1D406b9mQFTz49USZxEQiB2j0TAWMsYY04ycy5VCg96AjhC/B3UKvN 0hjw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=26C6jzBDYX2KyIOZi1XWpKUSUT0DJxPrIpF38ct4maw=; b=vqZUChGE7CSaEdDWLOgjdoLk40Lu7JlHRmMb+NEiDLxuSzpyNduf1VMxL/bbp14Xp0 Zi5ZJKyigc6vgaKMFDyonxTFFHOG4UNhtU7n1fTgVkaamEqu3p0r9kDmnSBAf8nX6C2E iZmAzZgP0eRj2sPKmNZiu9qKkEbjA7BPsXYrn3qDPYpDsu2S6Zu+Ge/iCjxBt3aRrK/E c2EJ6R0hLGKR6rCl85xKaLqmMKD3dGNpywlP2fyWjFo2QpsvtcbRhKB4FSIJKM+sY8TI T05R6Bi3hQvw6RxXLLFfmeg6PApwWcZ51LYbn5gLxUyRsWDrxo/pLbHqv10pKslS1uq/ 3KcQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@ventanamicro.com header.s=google header.b=jx2lf6rq; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id t21-20020a63eb15000000b003821f536216si2605618pgh.578.2022.04.15.16.26.08; Fri, 15 Apr 2022 16:26:22 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@ventanamicro.com header.s=google header.b=jx2lf6rq; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S242693AbiDNLcE (ORCPT + 99 others); Thu, 14 Apr 2022 07:32:04 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:49564 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231741AbiDNLcC (ORCPT ); Thu, 14 Apr 2022 07:32:02 -0400 Received: from mail-lf1-x134.google.com (mail-lf1-x134.google.com [IPv6:2a00:1450:4864:20::134]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 3F7E33616B for ; Thu, 14 Apr 2022 04:29:37 -0700 (PDT) Received: by mail-lf1-x134.google.com with SMTP id p10so8492422lfa.12 for ; Thu, 14 Apr 2022 04:29:37 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ventanamicro.com; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=26C6jzBDYX2KyIOZi1XWpKUSUT0DJxPrIpF38ct4maw=; b=jx2lf6rqh2orVVm6dXJqh1Fy07Tj7XpxI5Q8zXThz/AOcwEUFDIkS2sdUN/0KOyYyx vZAMxVXGbnunwf2cV4FLFjo9wwfwxqceyV1dNega39scX2rgA8V4CRi9v2FqCuiScLw1 Sa+K/XAGguCNRLCI9Q9duMIJ8zIzUW7bHjDarvo7z3vZiNyWlVUjMOo0ciEbtr/vKHRP UH00UXWhLgjTwx55Jdjsw+r4DwERFKAm4uIAWUKH0j1lVvgepsmDGGniYU1JLVqogBXq RlzgxzXN3rqM1E+rK5ouCOZJzgUb6jtIxCoMX7Jy3BYWLZXRBkNoZk+CTfHK0pumtOAy 02EQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=26C6jzBDYX2KyIOZi1XWpKUSUT0DJxPrIpF38ct4maw=; b=COX4LyBVw5rohG+IIxdU+3ZLw1NN4hj05liw6rPMVUQPOM9xxdiVYS2CsBK8n/E55i PiZfQdNkMWtyF2UFIj8Is4FoFR6OunnrfCc+N80gw9iImfvDgTdXhYosZBsnfEUDuXqq x5qI0oxFHiGcIm9/yYfxWO1fYoYuZBcLnlJztp3CJxH+iseEFUneQZFEmGQAOM4Y0Dup BgSBj1ZBmsi36nXln1X3eYDhxYyyPNgUN7EXxcUt/RTrXFqSykXdhhlu4xHqHGFcOPff irocuJosqWZOJytLHhbfkyKtB1pA2wg6kyXP1Qf8kiE/IgP0H5wO3AXwTJsB9WaritdM F9qw== X-Gm-Message-State: AOAM5336xFX8nMeMAfUp2ZsJYKaz5Bf/zTuoYbUnCK+9edoRmTof4Mm/ 7UKnlrXwSli5o2391HfUwsViFxho28yMfuUDFpv45XhVVSpY88wK X-Received: by 2002:a19:6b13:0:b0:46b:b970:5792 with SMTP id d19-20020a196b13000000b0046bb9705792mr1706263lfa.100.1649935775390; Thu, 14 Apr 2022 04:29:35 -0700 (PDT) MIME-Version: 1.0 References: <20220324151258.943896-1-apatel@ventanamicro.com> <20220324151258.943896-4-apatel@ventanamicro.com> <87a6cshf6y.ffs@tglx> In-Reply-To: <87a6cshf6y.ffs@tglx> From: Anup Patel Date: Thu, 14 Apr 2022 16:59:23 +0530 Message-ID: Subject: Re: [PATCH v5 3/7] genirq: Add mechanism to multiplex a single HW IPI To: Thomas Gleixner Cc: Palmer Dabbelt , Paul Walmsley , Marc Zyngier , Daniel Lezcano , Atish Patra , Alistair Francis , Anup Patel , linux-riscv , "linux-kernel@vger.kernel.org List" Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Apr 11, 2022 at 1:41 AM Thomas Gleixner wrote: > > Anup, > > On Thu, Mar 24 2022 at 20:42, Anup Patel wrote: > > All RISC-V platforms have a single HW IPI provided by the INTC local > > interrupt controller. The HW method to trigger INTC IPI can be through > > external irqchip (e.g. RISC-V AIA), through platform specific device > > (e.g. SiFive CLINT timer), or through firmware (e.g. SBI IPI call). > > > > To support multiple IPIs on RISC-V, we need a generic mechanism to > > create multiple per-CPU vIRQs using a single HW IPI hence this patch. > > git grep 'This patch' Documentation/process Okay, I will update the commit description as-per Documentation. > > > The generic IPI multiplex mechanism added by this patch can also be > > useful to other architectures. > > Which ones? Sane architectures have more than one IPI. Currently, the IPI muxing is shared code for various RISC-V drivers (such as CLINT driver, SBI IPI irqchip driver, and AIA (coming soon)). Overall, the IPI muxing seems independent of RISC-V so maybe it is useful to have it as common selectable API. > > > diff --git a/include/linux/irq.h b/include/linux/irq.h > > index 848e1e12c5c6..cdce7eae2f37 100644 > > --- a/include/linux/irq.h > > +++ b/include/linux/irq.h > > @@ -1248,6 +1248,34 @@ int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest); > > int ipi_send_single(unsigned int virq, unsigned int cpu); > > int ipi_send_mask(unsigned int virq, const struct cpumask *dest); > > > > +#define IPI_MUX_NR_IRQS BITS_PER_LONG > > +struct ipi_mux_ops { > > This is unreadable. Newlines exist for a reason. Okay, I will add a newline above the struct. > > > + void (*ipi_mux_clear)(unsigned int parent_virq); > > + void (*ipi_mux_send)(unsigned int parent_virq, > > + const struct cpumask *mask); > > +}; > > + > > +/* Process multiplexed IPIs */ > > +void ipi_mux_process(void); > > + > > +/* > > + * Create multiple IPIs (total IPI_MUX_NR_IRQS) multiplexed on top of a > > + * single parent IPI. > > + * > > + * If the parent IPI > 0 then ipi_mux_process() will be automatically > > + * called via chained handler. > > + * > > + * If the parent IPI <= 0 then it is responsiblity of irqchip drivers > > + * to explicitly call ipi_mux_process() for processing muxed > > + * IPIs. > > + * > > + * Returns first virq of the muxed IPIs upon success or <=0 upon failure > > + */ > > +int ipi_mux_create(unsigned int parent_virq, const struct ipi_mux_ops *ops); > > While it is kinda sensible to have the documentation near the > declaration, I prefer it to be near the code because thats where it > matters and also has a higher chance to be updated when the code > changes. Okay, I will move documentation near the code. > > Please use proper kernel doc while at it. Sure, I will update. > > > +static unsigned int ipi_mux_parent_virq; > > +static struct irq_domain *ipi_mux_domain; > > +static const struct ipi_mux_ops *ipi_mux_ops; > > +static DEFINE_PER_CPU(unsigned long, ipi_mux_bits); > > + > > +static void ipi_mux_dummy(struct irq_data *d) > > +{ > > +} > > + > > +static void ipi_mux_send_mask(struct irq_data *d, const struct cpumask *mask) > > +{ > > + int cpu; > > + > > + /* Barrier before doing atomic bit update to IPI bits */ > > + smp_mb__before_atomic(); > > + > > + for_each_cpu(cpu, mask) > > + set_bit(d->hwirq, per_cpu_ptr(&ipi_mux_bits, cpu)); > > + > > + /* Barrier after doing atomic bit update to IPI bits */ > > + smp_mb__after_atomic(); > > + > > + /* Trigger the parent IPI */ > > + ipi_mux_ops->ipi_mux_send(ipi_mux_parent_virq, mask); > > +} > > + > > +static struct irq_chip ipi_mux_chip = { > > + .name = "RISC-V IPI Mux", > > RISC-V IPI Mux is a truly generic name :) Aargh, I forgot to remove "RISC-V" from the name here. I will update. > > > + .irq_mask = ipi_mux_dummy, > > + .irq_unmask = ipi_mux_dummy, > > + .ipi_send_mask = ipi_mux_send_mask, > > +}; > > + > > +static int ipi_mux_domain_map(struct irq_domain *d, unsigned int irq, > > + irq_hw_number_t hwirq) > > +{ > > + irq_set_percpu_devid(irq); > > + irq_domain_set_info(d, irq, hwirq, &ipi_mux_chip, d->host_data, > > + handle_percpu_devid_irq, NULL, NULL); > > + > > + return 0; > > +} > > + > > +static int ipi_mux_domain_alloc(struct irq_domain *d, unsigned int virq, > > + unsigned int nr_irqs, void *arg) > > +{ > > + int i, ret; > > + irq_hw_number_t hwirq; > > + unsigned int type = IRQ_TYPE_NONE; > > + struct irq_fwspec *fwspec = arg; > > Documentation/process/maintainer-tip.rst #coding-style-notes Okay, I will refer and update. > > > + ret = irq_domain_translate_onecell(d, fwspec, &hwirq, &type); > > + if (ret) > > + return ret; > > + > > + for (i = 0; i < nr_irqs; i++) { > > + ret = ipi_mux_domain_map(d, virq + i, hwirq + i); > > + if (ret) > > + return ret; > > + } > > + > > + return 0; > > +} > > + > > +static const struct irq_domain_ops ipi_mux_domain_ops = { > > + .translate = irq_domain_translate_onecell, > > + .alloc = ipi_mux_domain_alloc, > > + .free = irq_domain_free_irqs_top, > > +}; > > + > > +void ipi_mux_process(void) > > +{ > > + int err; > > + unsigned long irqs, *bits = this_cpu_ptr(&ipi_mux_bits); > > + irq_hw_number_t hwirq; > > + > > + while (true) { > > + /* Clear the parent IPI */ > > + ipi_mux_ops->ipi_mux_clear(ipi_mux_parent_virq); > > This being in a loop smells fishy at least without a comment. And the > more I read all of this the less I'm convinced that this code can be > used by anything else than RISCV. The original IPI muxing code in RISC-V had this loop so I did not remove it. Actually, the loop is redundant because if a CPU gets another IPI while it was in ipi_mux_process() then another interrupt will be taken and ipi_mux_process() will be called again. I test more and remove this loop. > > > + /* Order bit clearing and data access. */ > > + mb(); > > This mb() pairs with what? Memory barriers have a counterpart and it's > mandatory to document that in the comment. It pairs with barriers in ipi_mux_send_mask(). I will update the comment. > > > + irqs = xchg(bits, 0); > > + if (!irqs) > > + break; > > + > > + for_each_set_bit(hwirq, &irqs, IPI_MUX_NR_IRQS) { > > + err = generic_handle_domain_irq(ipi_mux_domain, > > + hwirq); > > + if (unlikely(err)) > > + pr_warn_ratelimited( > > + "can't find mapping for hwirq %lu\n", > > + hwirq); > > + } > > + } > > +} > > + > > + > > +void ipi_mux_destroy(void) > > Seriously? You provide a function to rip the IPI mechanism out in a > running system? What's that for? > > > +{ > > + if (!ipi_mux_domain) > > + return; > > + > > + irq_domain_remove(ipi_mux_domain); > > + ipi_mux_domain = NULL; > > + ipi_mux_parent_virq = 0; > > If it would be useful, then this would leak the hotplug callbacks, but > the good news is that after tearing down the IPI domain hotplug does not > work anymore :) The only use of this function was to clean up in-case the irqchip driver failed after creating mux. I will certainly remove this function in the next patch revision. > > Thanks, > > tglx Regards, Anup