Received: by 2002:a25:824b:0:0:0:0:0 with SMTP id d11csp3264145ybn; Fri, 27 Sep 2019 03:49:14 -0700 (PDT) X-Google-Smtp-Source: APXvYqx02v87mMiLAu7z7IEqhe3LH/0ImwhIwqIiIBII++d63irYO8nwffia9meGXCiSkna4bWRr X-Received: by 2002:aa7:dad9:: with SMTP id x25mr3752653eds.7.1569581354827; Fri, 27 Sep 2019 03:49:14 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1569581354; cv=none; d=google.com; s=arc-20160816; b=FdXg2SsLkEHoru7yXOB3xxiS2bH8GeE1KPewYt1PSUwaHVhu3h7P2ajNzr5EMlXCKF GNBRT1Vx5vPTsKXKAmx1QiqY0LN+j3RmcdVhj0fj63cRIfQ+tsu1L+pjIhQzP38lbRp0 ugZn58LLILH2f5EH6AsvhEokTTGKdVFNUjfSvVYZNb8xstQuq0T41wBsjYsJeaJSY7JF SDB46vGCHkj5wKyU80x1XLdztMoa0WVMTAPTS5TlUCRqsbiXBVj9OV2KGhx4G958EDnA ZGbZap0o92Xc3eheobNAsWYAmDxUAVbPbMEIC7aj0DcS2nQmLJfv9nDDjhCHEMGiHuTY URIw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:references:cc:to:subject:from; bh=xKTZL2zdjAX6XbZKvqdB/Yc9cJPTZsmfBJknrXQt+xQ=; b=p2TPaaiNa8t757oBXqlBAxvxv6NYpSrX+394kPo/GdnopWphiyVuUQZ4fUtFTJVhhR o237rggkMTzmz8A98VM/WwitY9ZPFpmpTO6aBhdp8FD66PDulNs4p3bJyhBveBwuhP5G dvVYQ1146eI2KQClRlRExRhlIoaHYQnIoCReLlq7Jn5JxNWs3jUrU2iosZSh1Z8FF+oq NJoDQWaKcK1KTQiN62ebNsGLEiW1jTkRtkiPis4ehVoRUbQcLpF1clpUj5S4jwJoPMEx owjGKY8JaWSWvfzLqfsjPlffMI+aT74mHSymHfOpmft2HEBBV5s160Trki7yMLnkd7su YcIQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q1si1355109edj.354.2019.09.27.03.48.49; Fri, 27 Sep 2019 03:49:14 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726483AbfI0Ksd (ORCPT + 99 others); Fri, 27 Sep 2019 06:48:33 -0400 Received: from foss.arm.com ([217.140.110.172]:48924 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725890AbfI0Ksc (ORCPT ); Fri, 27 Sep 2019 06:48:32 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id E949628; Fri, 27 Sep 2019 03:48:31 -0700 (PDT) Received: from [10.1.196.133] (e112269-lin.cambridge.arm.com [10.1.196.133]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 8EE103F67D; Fri, 27 Sep 2019 03:48:30 -0700 (PDT) From: Steven Price Subject: Re: drm_sched with panfrost crash on T820 To: Neil Armstrong , daniel@ffwll.ch, airlied@linux.ie, =?UTF-8?Q?Christian_K=c3=b6nig?= Cc: Tomeu Vizoso , linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, "open list:ARM/Amlogic Meson..." , Erico Nunes , Rob Herring References: <3fb178d8-f069-0ae2-1ed3-4ded84a71951@arm.com> Message-ID: <26ae2a4d-8df1-e8db-3060-41638ed63e2a@arm.com> Date: Fri, 27 Sep 2019 11:48:29 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.9.0 MIME-Version: 1.0 In-Reply-To: <3fb178d8-f069-0ae2-1ed3-4ded84a71951@arm.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-GB Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 27/09/2019 10:55, Steven Price wrote: [...] > One obvious issue with the DRM scheduler is that there is a call to > cancel_delayed_work() in drm_sched_stop() which to me looks like it > should be cancel_delayed_work_sync() to ensure that the timeout handling > has completed. > > However in the above scenario a _sync() variety would then cause a > deadlock (one thread has pfdev->reset_lock and is waiting for the other > thread which is trying to take the lock). > > So we need to update Panfrost so that it can coordinate the reset > between schedulers. Can you try something like the following (untested): And actually testing it I of course discover it doesn't quite work. We do need the cancel_delayed_work_sync() in the DRM scheduler (when stopping a different scheduler) and we need to avoid holding the reset_lock during the drm_sched_stop() call to prevent deadlocking with another thread handling a timeout. Can you give the following patch a spin? I don't have a great reproduction case, so it would be good to get some confidence it fixes the problem. ----8<---- From 521a286789260197ae94f698932ebf369efc45ad Mon Sep 17 00:00:00 2001 From: Steven Price Date: Fri, 27 Sep 2019 11:42:40 +0100 Subject: [PATCH] drm/panfrost: Handle resetting on timeout better Panfrost uses multiple schedulers (one for each slot, so 2 in reality), and on a timeout has to stop all the schedulers to safely perform a reset. However more than one scheduler can trigger a timeout at the same time. This race condition results in jobs being freed while they are still in use. Modify drm_sched_stop() to call cancel_delayed_work_sync() when stopping a different scheduler to the one belonging to the passed in job. panfrost_job_timedout() is also modified to only allow one thread at a time to handle the reset. Any subsequent threads simply return assuming that the first thread will handle the situation. Signed-off-by: Steven Price --- drivers/gpu/drm/panfrost/panfrost_device.h | 2 ++ drivers/gpu/drm/panfrost/panfrost_job.c | 11 ++++++++++- drivers/gpu/drm/scheduler/sched_main.c | 5 ++++- 3 files changed, 16 insertions(+), 2 deletions(-) diff --git a/drivers/gpu/drm/panfrost/panfrost_device.h b/drivers/gpu/drm/panfrost/panfrost_device.h index f503c566e99f..6441c7fba6c4 100644 --- a/drivers/gpu/drm/panfrost/panfrost_device.h +++ b/drivers/gpu/drm/panfrost/panfrost_device.h @@ -99,6 +99,8 @@ struct panfrost_device { unsigned long cur_volt; struct panfrost_devfreq_slot slot[NUM_JOB_SLOTS]; } devfreq; + + bool is_resetting; }; struct panfrost_mmu { diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c index 05c85f45a0de..1b2019e08b43 100644 --- a/drivers/gpu/drm/panfrost/panfrost_job.c +++ b/drivers/gpu/drm/panfrost/panfrost_job.c @@ -388,13 +388,21 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job) mutex_lock(&pfdev->reset_lock); + if (pfdev->is_resetting) { + mutex_unlock(&pfdev->reset_lock); + return; + } + pfdev->is_resetting = true; + + mutex_unlock(&pfdev->reset_lock); + for (i = 0; i < NUM_JOB_SLOTS; i++) drm_sched_stop(&pfdev->js->queue[i].sched, sched_job); if (sched_job) drm_sched_increase_karma(sched_job); - /* panfrost_core_dump(pfdev); */ + mutex_lock(&pfdev->reset_lock); panfrost_devfreq_record_transition(pfdev, js); panfrost_device_reset(pfdev); @@ -406,6 +414,7 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job) for (i = 0; i < NUM_JOB_SLOTS; i++) drm_sched_start(&pfdev->js->queue[i].sched, true); + pfdev->is_resetting = false; mutex_unlock(&pfdev->reset_lock); } diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c index 148468447ba9..bc6d1862ec8a 100644 --- a/drivers/gpu/drm/scheduler/sched_main.c +++ b/drivers/gpu/drm/scheduler/sched_main.c @@ -415,7 +415,10 @@ void drm_sched_stop(struct drm_gpu_scheduler *sched, struct drm_sched_job *bad) * this TDR finished and before the newly restarted jobs had a * chance to complete. */ - cancel_delayed_work(&sched->work_tdr); + if (bad->sched != sched) + cancel_delayed_work_sync(&sched->work_tdr); + else + cancel_delayed_work(&sched->work_tdr); } EXPORT_SYMBOL(drm_sched_stop); -- 2.20.1