Received: by 2002:a25:4158:0:0:0:0:0 with SMTP id o85csp1102955yba; Wed, 24 Apr 2019 15:11:07 -0700 (PDT) X-Google-Smtp-Source: APXvYqwNTBm/JYtDwamLXQN0qGMTl4X+esXwu0gSYx9y6jBj1mgisGDlDCDDZv2Ux6vVQMqRkJu1 X-Received: by 2002:aa7:991b:: with SMTP id z27mr8459248pff.168.1556143867871; Wed, 24 Apr 2019 15:11:07 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1556143867; cv=none; d=google.com; s=arc-20160816; b=SMZcX1UyA6aC4BHa8uA2WK0phl+dEMsu0xYbOCPbm70usYt5rFZiD/Hw1a/yysHJ5g hR8htNi+rvexk3WHBnfW2bU5DsPziY7UZd7GpWU3Afey+yrQ+wjTlIvDk5gctEibV/if QusKdGQGw0e8iX38Eb57vmxv39+LbsZXp84/0uqPw0gAU+5ShmOneUj7jrymwL2oAEuQ VoRCyvzLpUlZ8TuqWO/jLmMM18/MuZ9GJ043uel6T3BJodEWA2inp+RASz51UhX/Ejd/ dkldzYAvop83eRLbeQ0ZVQOJ2D+UTvn/fiTs/vj8YJKASGIJLQLVGJQSsAnEeTIbITDZ MCZQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-language :content-transfer-encoding:in-reply-to:mime-version:user-agent:date :message-id:organization:from:references:cc:to:subject; bh=I/VeNK4tZKZddQiSDu5iryPnTBF0ZzaQEIRwZaoSWTk=; b=oCpzejgUwZP1xb9eN4SzKagp1CDIXjszAcbu51SnSFIEB9QN4UEow455EO20BAldsN qPtuaWcx7FDPimH93/b5mzq73NNe6wKikHVBUDt2giW7ebTbnZJkqSNQFspACyk8F4FE a10j44wSE8+Otc82ji6pmuraE/Z4y5K8+bgI496xtXbCu+OepPRPlv2spTRjxUu/LoIj T3iCEmmpfFFaGmJ/kPj88sxR1Ofaz7zynkAOTwL1tx7n9kHTtEQbNVj2J2wGGBOKF7GH ucjsLnwMbWi9Fwk8+EP9J/ykXKWSWVzbzKFcoymujDdDG/Wuz4Cmjz4qM8ugtrOZgb/n adfQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 98si20416885plb.84.2019.04.24.15.10.52; Wed, 24 Apr 2019 15:11:07 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1732740AbfDXQtI (ORCPT + 99 others); Wed, 24 Apr 2019 12:49:08 -0400 Received: from mx1.redhat.com ([209.132.183.28]:55106 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1732574AbfDXQtG (ORCPT ); Wed, 24 Apr 2019 12:49:06 -0400 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 89640307D861; Wed, 24 Apr 2019 16:49:06 +0000 (UTC) Received: from llong.remote.csb (dhcp-17-85.bos.redhat.com [10.18.17.85]) by smtp.corp.redhat.com (Postfix) with ESMTP id 8FAD660C9C; Wed, 24 Apr 2019 16:49:05 +0000 (UTC) Subject: Re: [PATCH v4 14/16] locking/rwsem: Guard against making count negative To: Peter Zijlstra Cc: Linus Torvalds , Ingo Molnar , Will Deacon , Thomas Gleixner , Linux List Kernel Mailing , the arch/x86 maintainers , Davidlohr Bueso , Tim Chen , huang ying References: <4cbd3c18-c9c0-56eb-4e01-ee355a69057a@redhat.com> <20190419102647.GP7905@worktop.programming.kicks-ass.net> <20190419120207.GO4038@hirez.programming.kicks-ass.net> <20190419130304.GV14281@hirez.programming.kicks-ass.net> <20190419131522.GW14281@hirez.programming.kicks-ass.net> <57620139-92a3-4a21-56bd-5d6fff23214f@redhat.com> <7b1bfc26-6e90-bd65-ab46-08413acd80e9@redhat.com> <20190423141714.GO11158@hirez.programming.kicks-ass.net> <4f62d7f2-e5f6-500e-3e70-b1d1978f7140@redhat.com> <20190424070959.GE4038@hirez.programming.kicks-ass.net> From: Waiman Long Organization: Red Hat Message-ID: <51589ac0-3e1f-040e-02bf-b6de77cbda1d@redhat.com> Date: Wed, 24 Apr 2019 12:49:05 -0400 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: <20190424070959.GE4038@hirez.programming.kicks-ass.net> Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.48]); Wed, 24 Apr 2019 16:49:06 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 4/24/19 3:09 AM, Peter Zijlstra wrote: > On Tue, Apr 23, 2019 at 03:12:16PM -0400, Waiman Long wrote: >> That is true in general, but doing preempt_disable/enable across >> function boundary is ugly and prone to further problems down the road. > We do worse things in this code, and the thing Linus proposes is > actually quite simple, something like so: > > --- > --- a/kernel/locking/rwsem.c > +++ b/kernel/locking/rwsem.c > @@ -912,7 +904,7 @@ rwsem_down_read_slowpath(struct rw_semap > raw_spin_unlock_irq(&sem->wait_lock); > break; > } > - schedule(); > + schedule_preempt_disabled(); > lockevent_inc(rwsem_sleep_reader); > } > > @@ -1121,6 +1113,7 @@ static struct rw_semaphore *rwsem_downgr > */ > inline void __down_read(struct rw_semaphore *sem) > { > + preempt_disable(); > if (unlikely(atomic_long_fetch_add_acquire(RWSEM_READER_BIAS, > &sem->count) & RWSEM_READ_FAILED_MASK)) { > rwsem_down_read_slowpath(sem, TASK_UNINTERRUPTIBLE); > @@ -1129,10 +1122,12 @@ inline void __down_read(struct rw_semaph > } else { > rwsem_set_reader_owned(sem); > } > + preempt_enable(); > } > > static inline int __down_read_killable(struct rw_semaphore *sem) > { > + preempt_disable(); > if (unlikely(atomic_long_fetch_add_acquire(RWSEM_READER_BIAS, > &sem->count) & RWSEM_READ_FAILED_MASK)) { > if (IS_ERR(rwsem_down_read_slowpath(sem, TASK_KILLABLE))) > @@ -1142,6 +1137,7 @@ static inline int __down_read_killable(s > } else { > rwsem_set_reader_owned(sem); > } > + preempt_enable(); > return 0; > } > Making that change will help the slowpath to has less preemption points. For an uncontended rwsem, this offers no real benefit. Adding preempt_disable() is more complicated than I originally thought. Maybe we are too paranoid about the possibility of a large number of preemptions happening just at the right moment. If p is the probably of a preemption in the middle of the inc-check-dec sequence, which I have already moved as close to each other as possible. We are talking a probability of p^32768. Since p will be really small, the compound probability will be infinitesimally small. So I would like to not do preemption now for the current patchset. We can restart the discussion later on if there is a real concern that it may actually happen. Please let me know if you still want to add preempt_disable() for the read lock. Cheers, Longman