Received: by 2002:a05:6a10:a852:0:0:0:0 with SMTP id d18csp1626301pxy; Thu, 6 May 2021 11:45:55 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz+B3NVVAqwOn4xpy2Yd2e9HRAmgSzZq3svhkShDYcggox9Ehzsf+eHJl6ANbm6b36xO4kW X-Received: by 2002:a63:f056:: with SMTP id s22mr5653686pgj.369.1620326755273; Thu, 06 May 2021 11:45:55 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1620326755; cv=none; d=google.com; s=arc-20160816; b=NQclTqzQYCuwj9NrrtCwF8cmtRMvr2OuNSnwjg1YuyCu2BgIM3nNYbE/U0BUFHBti1 mdM6+dgtAMgsoGRLLEf2CV06Z3MTk8lRnGAnsG9N21S87hD3m3nJPpuOU8QF7B1GVB5A 1m9GNtA6jDn0CjPn0X1OsFNt5GhoTQt796rfEX/Hr9fWtJ6awzeT4Mbep5lQg2HgHBIu isKudwvEUgpiUDqouwiiZRy5p6AGovyGG3n9HII4v3nX+Yozeu+BkWHG0U0fUIgvzMZB HbfqNCRnIXmIlPrFSL5DOUxI3cR6S40TaQ7EX8KlRA8mWpY23n3zAtDqxK2iii3fea1h h4+Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject:dkim-signature; bh=CQ9Mos2syCik0TQ8OY7wSpK2F3tBlzqwRXD+sm7DThY=; b=Z60VmvjoZLzkkx+R1Uw8iOc+aZnUDixps1AtKObQ60xhGOTYQ9aDRMmDzz3LU2s1Pp 4zh0YPFdyl2Thse1s98dYLEMBbXbLECzGNdFskPnNsD24nrmxmnUyTuS18G9o/pdYdTG Mxd5WC9ebsRHz8+QQatfOFd0Pb9lyDMOZPlWLH+Xzyopf+VL2kWjvWEdGU/oYAdQB+5O 4Z8iWi6uckfvlHgtgyiumrId9FaTFoTfR6NnIXZTexx5bKBeO79MveLcI2verflzwIpU v+QN3R+zpCvDSc1A+whRDy17pkfTLsPBEZM3+v+bELQeogibQEF9ldn3pHdL0UUiL810 G7AA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=P9jLJo6A; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id b4si4214797plz.90.2021.05.06.11.45.42; Thu, 06 May 2021 11:45:55 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=P9jLJo6A; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234951AbhEFOwL (ORCPT + 99 others); Thu, 6 May 2021 10:52:11 -0400 Received: from us-smtp-delivery-124.mimecast.com ([216.205.24.124]:24375 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234759AbhEFOwK (ORCPT ); Thu, 6 May 2021 10:52:10 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1620312672; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=CQ9Mos2syCik0TQ8OY7wSpK2F3tBlzqwRXD+sm7DThY=; b=P9jLJo6A1wzPrv4cAu32yWwpz1flbfGXDSxXvrctPZkXY7Dty2q7DAMqNLUAQ2QhmYS4sL AxgVOP9kZ1YRaQcjEyedymf4p+Y4gtfNlMQJH78VRN0/yJ7lNtlgMfJEe0KDimmc8BuyyD bmZTAr+abipgqCmes8Nwbg6js1TSGbo= Received: from mail-wr1-f69.google.com (mail-wr1-f69.google.com [209.85.221.69]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-165-O1nU1hnoPYai3RVCOp0h2A-1; Thu, 06 May 2021 10:51:08 -0400 X-MC-Unique: O1nU1hnoPYai3RVCOp0h2A-1 Received: by mail-wr1-f69.google.com with SMTP id a12-20020a5d6cac0000b0290109c3c8d66fso2295417wra.15 for ; Thu, 06 May 2021 07:51:08 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=CQ9Mos2syCik0TQ8OY7wSpK2F3tBlzqwRXD+sm7DThY=; b=n+o8QGRH08GYtEe34i+Ebxjer3TrF4j7AbWCz24Of373LlNO+ms0mN+6EXCCDAJ8bG 1uU6W9Ksd32VWbyymbadIGs2RVsrPwnm+qye1oSAWHCeSsUjlQpRZ8F4qNjkzsBAdpY1 ttOOvvTWwNBZiQoFZpFPfreV7KtrX7r09Kzxl1vXZw2bWqXwAJuKSAL32apaYEuW2tBp w/ioUZ69g1GWeKTNYjoi32lXBVCiXjXJbtlTHiS1Cb5FBnrxZM2N9+CCZv2HNHw7ak1T DIYCqcB0bAirXxJnoG0nXFiizXYVvvzAtk7MEm5/HALUih12wsnAexMHGcmji8tj1ZVM L3eQ== X-Gm-Message-State: AOAM532psaNcEFrdLim+AkZ6NrbnSvmUYMOH0XMP4ZHPTXpHojTUlkPR gA0eBTY1Wymf2arEr8AMBNUzyJTwesDE/VxMmoexjUbum6FAoNhCYbY+2vJGYgnO3ZstitXMT3+ nCsk2d0q+IMb141as//gmYOa8mKhNhw3+M9EG/20/2M90dZPKz/Tfplo7GY18Z9F2bRx6isPv/C /+ X-Received: by 2002:a1c:ed05:: with SMTP id l5mr15782692wmh.154.1620312667347; Thu, 06 May 2021 07:51:07 -0700 (PDT) X-Received: by 2002:a1c:ed05:: with SMTP id l5mr15782658wmh.154.1620312667063; Thu, 06 May 2021 07:51:07 -0700 (PDT) Received: from ?IPv6:2001:b07:6468:f312:c8dd:75d4:99ab:290a? ([2001:b07:6468:f312:c8dd:75d4:99ab:290a]) by smtp.gmail.com with ESMTPSA id z18sm4339616wrh.16.2021.05.06.07.51.05 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Thu, 06 May 2021 07:51:06 -0700 (PDT) Subject: Re: KVM: x86: Prevent deadlock against tk_core.seq To: Thomas Gleixner , kvm@vger.kernel.org Cc: Sean Christopherson , x86@kernel.org, LKML References: <87h7jgm1zy.ffs@nanos.tec.linutronix.de> From: Paolo Bonzini Message-ID: Date: Thu, 6 May 2021 16:51:05 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.8.1 MIME-Version: 1.0 In-Reply-To: <87h7jgm1zy.ffs@nanos.tec.linutronix.de> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 06/05/21 15:21, Thomas Gleixner wrote: > syzbot reported a possible deadlock in pvclock_gtod_notify(): > > CPU 0 CPU 1 > write_seqcount_begin(&tk_core.seq); > pvclock_gtod_notify() spin_lock(&pool->lock); > queue_work(..., &pvclock_gtod_work) ktime_get() > spin_lock(&pool->lock); do { > seq = read_seqcount_begin(tk_core.seq) > ... > } while (read_seqcount_retry(&tk_core.seq, seq); > > While this is unlikely to happen, it's possible. > > Delegate queue_work() to irq_work() which postpones it until the > tk_core.seq write held region is left and interrupts are reenabled. > > Fixes: 16e8d74d2da9 ("KVM: x86: notifier for clocksource changes") > Reported-by: syzbot+6beae4000559d41d80f8@syzkaller.appspotmail.com > Signed-off-by: Thomas Gleixner > --- > Link: https://lore.kernel.org/r/0000000000001d43ac05c0f5c6a0@google.com > --- > arch/x86/kvm/x86.c | 22 ++++++++++++++++++---- > 1 file changed, 18 insertions(+), 4 deletions(-) > > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -8040,6 +8040,18 @@ static void pvclock_gtod_update_fn(struc > static DECLARE_WORK(pvclock_gtod_work, pvclock_gtod_update_fn); > > /* > + * Indirection to move queue_work() out of the tk_core.seq write held > + * region to prevent possible deadlocks against time accessors which > + * are invoked with work related locks held. > + */ > +static void pvclock_irq_work_fn(struct irq_work *w) > +{ > + queue_work(system_long_wq, &pvclock_gtod_work); > +} > + > +static DEFINE_IRQ_WORK(pvclock_irq_work, pvclock_irq_work_fn); > + > +/* > * Notification about pvclock gtod data update. > */ > static int pvclock_gtod_notify(struct notifier_block *nb, unsigned long unused, > @@ -8050,13 +8062,14 @@ static int pvclock_gtod_notify(struct no > > update_pvclock_gtod(tk); > > - /* disable master clock if host does not trust, or does not > - * use, TSC based clocksource. > + /* > + * Disable master clock if host does not trust, or does not use, > + * TSC based clocksource. Delegate queue_work() to irq_work as > + * this is invoked with tk_core.seq write held. > */ > if (!gtod_is_based_on_tsc(gtod->clock.vclock_mode) && > atomic_read(&kvm_guest_has_master_clock) != 0) > - queue_work(system_long_wq, &pvclock_gtod_work); > - > + irq_work_queue(&pvclock_irq_work); > return 0; > } > > @@ -8168,6 +8181,7 @@ void kvm_arch_exit(void) > cpuhp_remove_state_nocalls(CPUHP_AP_X86_KVM_CLK_ONLINE); > #ifdef CONFIG_X86_64 > pvclock_gtod_unregister_notifier(&pvclock_gtod_notifier); > + irq_work_sync(&pvclock_irq_work); > cancel_work_sync(&pvclock_gtod_work); > #endif > kvm_x86_ops.hardware_enable = NULL; > Queued, thanks. Paolo