Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp967082ybi; Tue, 16 Jul 2019 07:45:50 -0700 (PDT) X-Google-Smtp-Source: APXvYqzcpAsLZQ0dgd7WY1SE6gpdx/aWBWrBYRVb4ZKSpfvikWLHJAPJS0Oc1rvfRK0dYtPZwNuZ X-Received: by 2002:a63:608c:: with SMTP id u134mr32782428pgb.274.1563288350111; Tue, 16 Jul 2019 07:45:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1563288350; cv=none; d=google.com; s=arc-20160816; b=GBPhnQxUgv5WW1hm93qyKqj8FFd4AGwC8mQ+T80P9G6B2W70649emDa0aJA78zgGOo C4+hGodBS0isfM1/NFuV+P4Sojbx1xsZuEo5S+q7BIZrZNaxd7U0d/5uAOh/PPlZKbbd ijopwCDDRCDVz9WOdt9y53UtTwGngGQ6zfJXxpyiO566pKFxMV75bgDJgOlKQQcdzamX ZU9iJTU+rRi7Crnk7JqiTSsFcCCRLAZueaM8lwnFZA7a6Nu7RvcYcwS9iFgmATlPuPOE T0Rq2OUvhbdn7ecoIvr/1cZYQOSEZGC/b2LohApOh1o365qjGZFnGsZ22l3thcerf81V hbvQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-language :content-transfer-encoding:in-reply-to:mime-version:user-agent:date :message-id:organization:from:references:cc:to:subject; bh=+Rr9BnABRrXCFBqBC393LyiNovDjBr8GebC8tWdgpw0=; b=wvT3XyxRvidbLtwoYg6Fr8iGVM46qu6m02H/bv5yZ0gxkeQFLBpn/DL7DvSiRcH/5w NV8u8z3otU1fDLGKC24G9x4f0WzF/OyUvt5xhKQ8ClkKGdrnGNrc1ZwCqVhBrK9wKkD/ yaBUPJfouH87gCJdlgLBTQ0GziUCIMgwXFOOlTpdTqCIU3iu55tza+wIilYMR050T8Oh d9FghFwAm0u5Zp6xIslRAvFPyVUlJ4+q3zvkMPeMGfz5QAIx6V5KMtzkeSQD5cYmigwU jXMUVgfWA1oseIvTzBqs+DiFzu5EJ+c4YdwS6b5jmP+Nh07wZNswttmb8PQF1Etm1MZI a0FQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id z19si20670380pfa.19.2019.07.16.07.45.33; Tue, 16 Jul 2019 07:45:50 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2387838AbfGPOoL (ORCPT + 99 others); Tue, 16 Jul 2019 10:44:11 -0400 Received: from mx1.redhat.com ([209.132.183.28]:58884 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726997AbfGPOoK (ORCPT ); Tue, 16 Jul 2019 10:44:10 -0400 Received: from smtp.corp.redhat.com (int-mx01.intmail.prod.int.phx2.redhat.com [10.5.11.11]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 14A2E85546; Tue, 16 Jul 2019 14:44:10 +0000 (UTC) Received: from llong.remote.csb (ovpn-122-180.rdu2.redhat.com [10.10.122.180]) by smtp.corp.redhat.com (Postfix) with ESMTP id 457726013A; Tue, 16 Jul 2019 14:44:07 +0000 (UTC) Subject: Re: [PATCH v3 3/5] locking/qspinlock: Introduce CNA into the slow path of qspinlock To: Alex Kogan Cc: linux@armlinux.org.uk, peterz@infradead.org, mingo@redhat.com, will.deacon@arm.com, arnd@arndb.de, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, tglx@linutronix.de, bp@alien8.de, hpa@zytor.com, x86@kernel.org, guohanjun@huawei.com, jglauber@marvell.com, steven.sistare@oracle.com, daniel.m.jordan@oracle.com, dave.dice@oracle.com, rahul.x.yadav@oracle.com References: <20190715192536.104548-1-alex.kogan@oracle.com> <20190715192536.104548-4-alex.kogan@oracle.com> <77bba626-f3e6-45a8-aae8-43b945d0fab9@redhat.com> <32DD898E-0F5E-4A63-9795-F78411B77A98@oracle.com> From: Waiman Long Organization: Red Hat Message-ID: <8ceec931-2921-6ee6-2642-476b4a12281e@redhat.com> Date: Tue, 16 Jul 2019 10:44:06 -0400 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.7.0 MIME-Version: 1.0 In-Reply-To: <32DD898E-0F5E-4A63-9795-F78411B77A98@oracle.com> Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 8bit Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.11 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.28]); Tue, 16 Jul 2019 14:44:10 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 7/16/19 10:26 AM, Alex Kogan wrote: >> On Jul 15, 2019, at 5:30 PM, Waiman Long wrote: >> >> On 7/15/19 3:25 PM, Alex Kogan wrote: >> >>> /* >>> - * On 64-bit architectures, the mcs_spinlock structure will be 16 bytes in >>> - * size and four of them will fit nicely in one 64-byte cacheline. For >>> - * pvqspinlock, however, we need more space for extra data. To accommodate >>> - * that, we insert two more long words to pad it up to 32 bytes. IOW, only >>> - * two of them can fit in a cacheline in this case. That is OK as it is rare >>> - * to have more than 2 levels of slowpath nesting in actual use. We don't >>> - * want to penalize pvqspinlocks to optimize for a rare case in native >>> - * qspinlocks. >>> + * On 64-bit architectures, the mcs_spinlock structure will be 20 bytes in >>> + * size. For pvqspinlock or the NUMA-aware variant, however, we need more >>> + * space for extra data. To accommodate that, we insert two more long words >>> + * to pad it up to 36 bytes. >>> */ >> The 20 bytes figure is wrong. It is actually 24 bytes for 64-bit as the >> mcs_spinlock structure is 8-byte aligned. For better cacheline >> alignment, I will like to keep mcs_spinlock to 16 bytes as before. >> Instead, you can use encode_tail() to store the CNA node pointer in >> "locked". For instance, use (encode_tail() << 1) in locked to >> distinguish it from the regular locked=1 value. > I think this can work. > decode_tail() will get the actual node pointer from the encoded value. > And that would keep the size of mcs_spinlock intact. > Good idea, thanks! > > BTW, maybe better change those function names to encode_node() / decode_node() then? The names look good to me. > >>> s >>> + >>> +static void cna_init_node(struct mcs_spinlock *node) >>> +{ >>> + struct cna_node *cn = CNA_NODE(node); >>> + struct mcs_spinlock *base_node; >>> + int cpuid; >>> + >>> + BUILD_BUG_ON(sizeof(struct cna_node) > sizeof(struct qnode)); >>> + /* we store a pointer in the node's @locked field */ >>> + BUILD_BUG_ON(sizeof(uintptr_t) > sizeof_field(struct mcs_spinlock, locked)); >>> + >>> + cpuid = smp_processor_id(); >>> + cn->numa_node = cpu_to_node(cpuid); >>> + >>> + base_node = this_cpu_ptr(&qnodes[0].mcs); >>> + cn->encoded_tail = encode_tail(cpuid, base_node->count - 1); >>> +} >> >> I think you can use an early_init call to initialize the numa_node and >> encoded_tail values for all the per-cpu CNA nodes instead of doing it >> every time a node is used. If it turns out that pv_qspinlock is used, >> the pv_node_init() will properly re-initialize it. > Yes, this should work. Thanks. > > BTW, should not we initialize `cpu` in pv_init_node() that same way? We would also initialize cpu this way in pv_init_node. The smp_processor_id() call is relatively cheap, but the initialization done here is more expensive. >> The only thing left >> to do here is perhaps setting tail to NULL. > There is no need to initialize cna_node.tail — we never access it unless > the node is at the head of the secondary queue, and in that case we > initialize it before placing the node at the head of that queue > (see find_successor()). OK. -Longman