Received: by 2002:a25:ad19:0:0:0:0:0 with SMTP id y25csp4654645ybi; Mon, 15 Jul 2019 12:27:54 -0700 (PDT) X-Google-Smtp-Source: APXvYqxvrCbcOnViMLklJjLKG27g5OXql3SAwDMJlSNGt+nVJfTqw0LPEf2F06E/WVMKb/uG3He/ X-Received: by 2002:a65:55c9:: with SMTP id k9mr29121401pgs.142.1563218873989; Mon, 15 Jul 2019 12:27:53 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1563218873; cv=none; d=google.com; s=arc-20160816; b=I5x0Ql+kJ0GUrDHozWOyHNf74EDnaVwvD82UQl1kVHPleJPouun0zyGnK+iKk7yil5 oSNrWXLvr8imI0jYPyjDuddA6pK2uuOU6NfZzvR7I0WnmBrOmVtQK0ExNjjW1cVNYjPC ZOuCDV3KzcpYb41/X0thN0+xRUchLoa46q+49gmdSn8Yu7ZHBBnvhjfMirRckTP62os1 snCxjmJltEysg+hm7PFdaHRD1QX5JowRzcovbkuJd0tUncnhuxCqv9uGsVxknVjfegQb m3urZzqN8yGm4tGF+qiGs/Zke62o+n1L8Un5rohKQmKDNKGv5YYPpzbzf9fVzTJ+EizP covQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:dkim-signature; bh=wRf1OAZyxbFwGPam2s6jhgORpRyKZydMuCtiU+UituY=; b=KBw6nhXhu+kf76s1PxJjdMMXKGtM8XtMBC8hpyUr1VXWnMcZJsgLzbL3sajSEll/Bv vDMyq9SFIbQT4IdxiMdXQauYbvGdh6+yMkSFE4dVEHSnXtlFayVrpAiWdaJ6xysAPZK2 TrMjNWWTqShCejVW26dGtji3yEbk+2ScAueJZF0JLBpaQ08XeEk4tN0J4nxvZ4YbOUYj 6a9wi0h6wD7G2SwFmmj8K4MNek9tjwDiMhePNUEhorvTpiY2px8H/ry4oLqBSt6ZauXz +QsXYhEwx6rriZj4XJLS4/Fw7SkCfgtI6d5J8iKZXqqvELFc7UlUcoYZjz4iuaaNeMUh anZA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2018-07-02 header.b=NGGwCIz0; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id r3si16391454pls.176.2019.07.15.12.27.37; Mon, 15 Jul 2019 12:27:53 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2018-07-02 header.b=NGGwCIz0; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730379AbfGOT1G (ORCPT + 99 others); Mon, 15 Jul 2019 15:27:06 -0400 Received: from userp2130.oracle.com ([156.151.31.86]:38464 "EHLO userp2130.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1730002AbfGOT1G (ORCPT ); Mon, 15 Jul 2019 15:27:06 -0400 Received: from pps.filterd (userp2130.oracle.com [127.0.0.1]) by userp2130.oracle.com (8.16.0.27/8.16.0.27) with SMTP id x6FJOdHx018491; Mon, 15 Jul 2019 19:26:01 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=oracle.com; h=from : to : cc : subject : date : message-id : in-reply-to : references; s=corp-2018-07-02; bh=wRf1OAZyxbFwGPam2s6jhgORpRyKZydMuCtiU+UituY=; b=NGGwCIz0YycmENfUgJpt7Vz/BXNTXpEIuK8+y8A5vU+yrp83xL2kwk35cnfQRDYJ3U6Q qZgiYIDJCcUSey+G6FRSY/i1AZmcqDiWYC2GWf2ecYQT/Tgupobfgm7WZbyxQDZSspwa vi1QjBoBsRHnAjKVsLtZ297mQKv44sI1JARrlzYD+V6XLzmwY831JJTRU9vbZ96j5L12 kAfPMMgJ5ZQWSmZoN/gUdkFMvc1dPksXLJpIrDNiQbaCdfXekY05Sv+5k/yaNtyaO0PX zi7/7FVhEGmOoGLCJ+JAcMhW49J4z4V3l9p0Og2tIBnD3pdvT8VqRJ2uXvnfNVxHAbnb nA== Received: from userp3030.oracle.com (userp3030.oracle.com [156.151.31.80]) by userp2130.oracle.com with ESMTP id 2tq6qtghyj-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Mon, 15 Jul 2019 19:26:00 +0000 Received: from pps.filterd (userp3030.oracle.com [127.0.0.1]) by userp3030.oracle.com (8.16.0.27/8.16.0.27) with SMTP id x6FJMeim185347; Mon, 15 Jul 2019 19:26:00 GMT Received: from userv0122.oracle.com (userv0122.oracle.com [156.151.31.75]) by userp3030.oracle.com with ESMTP id 2tq4dtg9hv-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Mon, 15 Jul 2019 19:26:00 +0000 Received: from abhmp0013.oracle.com (abhmp0013.oracle.com [141.146.116.19]) by userv0122.oracle.com (8.14.4/8.14.4) with ESMTP id x6FJPsdA007776; Mon, 15 Jul 2019 19:25:54 GMT Received: from ol-bur-x5-4.us.oracle.com (/10.152.128.37) by default (Oracle Beehive Gateway v4.0) with ESMTP ; Mon, 15 Jul 2019 12:25:54 -0700 From: Alex Kogan To: linux@armlinux.org.uk, peterz@infradead.org, mingo@redhat.com, will.deacon@arm.com, arnd@arndb.de, longman@redhat.com, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, tglx@linutronix.de, bp@alien8.de, hpa@zytor.com, x86@kernel.org, guohanjun@huawei.com, jglauber@marvell.com Cc: steven.sistare@oracle.com, daniel.m.jordan@oracle.com, alex.kogan@oracle.com, dave.dice@oracle.com, rahul.x.yadav@oracle.com Subject: [PATCH v3 2/5] locking/qspinlock: Refactor the qspinlock slow path Date: Mon, 15 Jul 2019 15:25:33 -0400 Message-Id: <20190715192536.104548-3-alex.kogan@oracle.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20190715192536.104548-1-alex.kogan@oracle.com> References: <20190715192536.104548-1-alex.kogan@oracle.com> X-Proofpoint-Virus-Version: vendor=nai engine=6000 definitions=9319 signatures=668688 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 suspectscore=0 malwarescore=0 phishscore=0 bulkscore=0 spamscore=0 mlxscore=0 mlxlogscore=999 adultscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.0.1-1810050000 definitions=main-1907150221 X-Proofpoint-Virus-Version: vendor=nai engine=6000 definitions=9319 signatures=668688 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 priorityscore=1501 malwarescore=0 suspectscore=0 phishscore=0 bulkscore=0 spamscore=0 clxscore=1015 lowpriorityscore=0 mlxscore=0 impostorscore=0 mlxlogscore=999 adultscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.0.1-1810050000 definitions=main-1907150222 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Move some of the code manipulating the spin lock into separate functions. This would allow easier integration of alternative ways to manipulate that lock. Signed-off-by: Alex Kogan Reviewed-by: Steve Sistare --- kernel/locking/qspinlock.c | 40 ++++++++++++++++++++++++++++++++++++++-- 1 file changed, 38 insertions(+), 2 deletions(-) diff --git a/kernel/locking/qspinlock.c b/kernel/locking/qspinlock.c index 961781624638..5668466b3006 100644 --- a/kernel/locking/qspinlock.c +++ b/kernel/locking/qspinlock.c @@ -297,6 +297,36 @@ static __always_inline u32 __pv_wait_head_or_lock(struct qspinlock *lock, #define queued_spin_lock_slowpath native_queued_spin_lock_slowpath #endif +/* + * set_locked_empty_mcs - Try to set the spinlock value to _Q_LOCKED_VAL, + * and by doing that unlock the MCS lock when its waiting queue is empty + * @lock: Pointer to queued spinlock structure + * @val: Current value of the lock + * @node: Pointer to the MCS node of the lock holder + * + * *,*,* -> 0,0,1 + */ +static __always_inline bool __set_locked_empty_mcs(struct qspinlock *lock, + u32 val, + struct mcs_spinlock *node) +{ + return atomic_try_cmpxchg_relaxed(&lock->val, &val, _Q_LOCKED_VAL); +} + +/* + * pass_mcs_lock - pass the MCS lock to the next waiter + * @node: Pointer to the MCS node of the lock holder + * @next: Pointer to the MCS node of the first waiter in the MCS queue + */ +static __always_inline void __pass_mcs_lock(struct mcs_spinlock *node, + struct mcs_spinlock *next) +{ + arch_mcs_spin_unlock_contended(&next->locked, 1); +} + +#define set_locked_empty_mcs __set_locked_empty_mcs +#define pass_mcs_lock __pass_mcs_lock + #endif /* _GEN_PV_LOCK_SLOWPATH */ /** @@ -541,7 +571,7 @@ void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) * PENDING will make the uncontended transition fail. */ if ((val & _Q_TAIL_MASK) == tail) { - if (atomic_try_cmpxchg_relaxed(&lock->val, &val, _Q_LOCKED_VAL)) + if (set_locked_empty_mcs(lock, val, node)) goto release; /* No contention */ } @@ -558,7 +588,7 @@ void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) if (!next) next = smp_cond_load_relaxed(&node->next, (VAL)); - arch_mcs_spin_unlock_contended(&next->locked, 1); + pass_mcs_lock(node, next); pv_kick_node(lock, next); release: @@ -583,6 +613,12 @@ EXPORT_SYMBOL(queued_spin_lock_slowpath); #undef pv_kick_node #undef pv_wait_head_or_lock +#undef set_locked_empty_mcs +#define set_locked_empty_mcs __set_locked_empty_mcs + +#undef pass_mcs_lock +#define pass_mcs_lock __pass_mcs_lock + #undef queued_spin_lock_slowpath #define queued_spin_lock_slowpath __pv_queued_spin_lock_slowpath -- 2.11.0 (Apple Git-81)