Received: by 2002:a25:824b:0:0:0:0:0 with SMTP id d11csp3555184ybn; Fri, 27 Sep 2019 08:03:23 -0700 (PDT) X-Google-Smtp-Source: APXvYqxWgm2Ay4VQAo6JBu042QI80QZ4mv+3I9gJ3r6tQVPsIcLQfRv9Tr21EoN8mfgFrJ/foYD9 X-Received: by 2002:a05:6000:45:: with SMTP id k5mr3001758wrx.259.1569596603715; Fri, 27 Sep 2019 08:03:23 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1569596603; cv=none; d=google.com; s=arc-20160816; b=pX3aeiXilFMQNEyzz53rGFm/dDx6oOPdEpA6V3mu2lGmhE3xRyfMJPZXqmSJTcS5fP 3CfX7G9I6apVv45v2vdVBIYCvJLJXq0FDMCk2mApXf/QHHvVXSBcFLCrZEH18kWieB5h xAAP2ZtjX66DhQqhoVB7+Pq4ESqHxjsicmYsX/5bMX5Qq190i4z6sDzzknm98WyiSAuU dT4iS4awaA2SQurtZMPzXuexdxUOa2FTD54qFJOPHJfI916/nwXspdLS7HHiuEq15m5b +0PkbMnhGzEFI9m7DR9NWLBa+zYTHxzH6owBGZJhA690zbKsEGQoHQIK/fi0O/3kFUmw HFrg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=MJGVbj0TWtO2jge7C0Y1rWidEjNivfcHvYogcRGmefI=; b=zymy4Yoytwwub4qPkyMDHeP5QP3Ng0tq8fJmqsWvNJUkqKsgCex046IG5sn8qJOtnb PrgY5sUdQOA0QmWRBIPwsaMo7jQPrnafMSOfeLXAw8Q1TZnq68lvmY3Zw3amRn1WaTai 5tMjkEo+Lf/xP3Ew808kdWpBrC/58FhOpyPEaexG9Aojma13GreTTYJr6U0LhT41YZ9C 665wNW7EDq1a/lxxhGszjZqlWRBUszQt3oeWM5MToxvhtOZ/Yar85I7pQGWC77FzHR9S Zi6pUMFlQbcCqWa3GJX6ZIF7IL+Vc5briCQjIbZsgYKpCdqSk806mED/p2yInastscu6 tskw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id h50si1694009ede.423.2019.09.27.08.02.53; Fri, 27 Sep 2019 08:03:23 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727542AbfI0PAV (ORCPT + 99 others); Fri, 27 Sep 2019 11:00:21 -0400 Received: from foss.arm.com ([217.140.110.172]:54378 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726926AbfI0PAV (ORCPT ); Fri, 27 Sep 2019 11:00:21 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 4A5AE28; Fri, 27 Sep 2019 08:00:20 -0700 (PDT) Received: from [10.1.196.133] (e112269-lin.cambridge.arm.com [10.1.196.133]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id F28703F534; Fri, 27 Sep 2019 08:00:18 -0700 (PDT) Subject: Re: drm_sched with panfrost crash on T820 To: Neil Armstrong , daniel@ffwll.ch, airlied@linux.ie, =?UTF-8?Q?Christian_K=c3=b6nig?= Cc: Tomeu Vizoso , linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, "open list:ARM/Amlogic Meson..." , Erico Nunes References: <3fb178d8-f069-0ae2-1ed3-4ded84a71951@arm.com> <26ae2a4d-8df1-e8db-3060-41638ed63e2a@arm.com> <1eec2f1b-0467-cd4d-aa22-23c70388ac0f@baylibre.com> <645fd795-7d22-268d-a8af-5ff090379505@baylibre.com> From: Steven Price Message-ID: <5e4592e8-e1ef-864d-5ed4-3b82f54fb43d@arm.com> Date: Fri, 27 Sep 2019 16:00:17 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.9.0 MIME-Version: 1.0 In-Reply-To: <645fd795-7d22-268d-a8af-5ff090379505@baylibre.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-GB Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 27/09/2019 12:48, Neil Armstrong wrote: > Hi, > > On 27/09/2019 13:27, Neil Armstrong wrote: >> Hi Steven, >> >> Thanks for your prompt reaction ! >> >> On 27/09/2019 12:48, Steven Price wrote: >>> On 27/09/2019 10:55, Steven Price wrote: >>> [...] >>>> One obvious issue with the DRM scheduler is that there is a call to >>>> cancel_delayed_work() in drm_sched_stop() which to me looks like it >>>> should be cancel_delayed_work_sync() to ensure that the timeout handling >>>> has completed. >>>> >>>> However in the above scenario a _sync() variety would then cause a >>>> deadlock (one thread has pfdev->reset_lock and is waiting for the other >>>> thread which is trying to take the lock). >>>> >>>> So we need to update Panfrost so that it can coordinate the reset >>>> between schedulers. Can you try something like the following (untested): >>> >>> And actually testing it I of course discover it doesn't quite work. We >>> do need the cancel_delayed_work_sync() in the DRM scheduler (when >>> stopping a different scheduler) and we need to avoid holding the >>> reset_lock during the drm_sched_stop() call to prevent deadlocking with >>> another thread handling a timeout. >> >> Yep the first patch wasn't fixing the issue all the time. >> >>> >>> Can you give the following patch a spin? I don't have a great >>> reproduction case, so it would be good to get some confidence it fixes >>> the problem. >> >> Running it right now. > > First run gave me (while applying on v5.3): > [ 307.969230] Unable to handle kernel NULL pointer dereference at virtual address 00000000000000c0 > [...] > [ 308.029358] Hardware name: Khadas VIM2 (DT) > [ 308.033510] Workqueue: events drm_sched_job_timedout > [ 308.038416] pstate: 80000005 (Nzcv daif -PAN -UAO) > [ 308.043160] pc : drm_sched_start+0x88/0x138 > [ 308.047300] lr : drm_sched_start+0xb0/0x138 > [...] > [ 308.133635] Call trace: > [ 308.136052] drm_sched_start+0x88/0x138 > [ 308.139847] panfrost_job_timedout+0x1cc/0x208 > [ 308.144245] drm_sched_job_timedout+0x44/0xa8 > [ 308.148560] process_one_work+0x1e0/0x320 > [ 308.152524] worker_thread+0x40/0x450 > [ 308.156149] kthread+0x124/0x128 > [ 308.159342] ret_from_fork+0x10/0x18 > [ 308.162879] Code: 54000280 f9400862 52800020 aa1a03e1 (f940605b) > [ 308.168914] ---[ end trace 256b7f5faec101d2 ]--- > > Bad pointer seems to be struct dma_fence *fence = s_job->s_fence->parent that > could be a NULL return from panfrost_job_run(). I haven't managed reproduce this locally, admittedly the test case I was using before relies on changes in drm-misc-next (HEAPs specifically), so it might be a bug present in v5.3 which isn't present in drm-misc-next. From the code dumped we have: 0: 54000280 b.eq 50 <.text+0x50> // b.none 4: f9400862 ldr x2, [x3, #16] 8: 52800020 mov w0, #0x1 // #1 c: aa1a03e1 mov x1, x26 10: f940605b ldr x27, [x2, #192] Which looks like the expression s_job->s_fence->parent, so it looks like s_job->s_fence == NULL. Which looks to me like drm_sched_job_cleanup() has been called on the job. But I can't work out why. Steve > Neil >> >> Thanks, >> Neil >> >>> >>> ----8<---- >>> From 521a286789260197ae94f698932ebf369efc45ad Mon Sep 17 00:00:00 2001 >>> From: Steven Price >>> Date: Fri, 27 Sep 2019 11:42:40 +0100 >>> Subject: [PATCH] drm/panfrost: Handle resetting on timeout better >>> >>> Panfrost uses multiple schedulers (one for each slot, so 2 in reality), >>> and on a timeout has to stop all the schedulers to safely perform a >>> reset. However more than one scheduler can trigger a timeout at the same >>> time. This race condition results in jobs being freed while they are >>> still in use. >>> >>> Modify drm_sched_stop() to call cancel_delayed_work_sync() when stopping >>> a different scheduler to the one belonging to the passed in job. >>> panfrost_job_timedout() is also modified to only allow one thread at a >>> time to handle the reset. Any subsequent threads simply return assuming >>> that the first thread will handle the situation. >>> >>> Signed-off-by: Steven Price >>> --- >>> drivers/gpu/drm/panfrost/panfrost_device.h | 2 ++ >>> drivers/gpu/drm/panfrost/panfrost_job.c | 11 ++++++++++- >>> drivers/gpu/drm/scheduler/sched_main.c | 5 ++++- >>> 3 files changed, 16 insertions(+), 2 deletions(-) >>> >>> diff --git a/drivers/gpu/drm/panfrost/panfrost_device.h b/drivers/gpu/drm/panfrost/panfrost_device.h >>> index f503c566e99f..6441c7fba6c4 100644 >>> --- a/drivers/gpu/drm/panfrost/panfrost_device.h >>> +++ b/drivers/gpu/drm/panfrost/panfrost_device.h >>> @@ -99,6 +99,8 @@ struct panfrost_device { >>> unsigned long cur_volt; >>> struct panfrost_devfreq_slot slot[NUM_JOB_SLOTS]; >>> } devfreq; >>> + >>> + bool is_resetting; >>> }; >>> >>> struct panfrost_mmu { >>> diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c >>> index 05c85f45a0de..1b2019e08b43 100644 >>> --- a/drivers/gpu/drm/panfrost/panfrost_job.c >>> +++ b/drivers/gpu/drm/panfrost/panfrost_job.c >>> @@ -388,13 +388,21 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job) >>> >>> mutex_lock(&pfdev->reset_lock); >>> >>> + if (pfdev->is_resetting) { >>> + mutex_unlock(&pfdev->reset_lock); >>> + return; >>> + } >>> + pfdev->is_resetting = true; >>> + >>> + mutex_unlock(&pfdev->reset_lock); >>> + >>> for (i = 0; i < NUM_JOB_SLOTS; i++) >>> drm_sched_stop(&pfdev->js->queue[i].sched, sched_job); >>> >>> if (sched_job) >>> drm_sched_increase_karma(sched_job); >>> >>> - /* panfrost_core_dump(pfdev); */ >>> + mutex_lock(&pfdev->reset_lock); >>> >>> panfrost_devfreq_record_transition(pfdev, js); >>> panfrost_device_reset(pfdev); >>> @@ -406,6 +414,7 @@ static void panfrost_job_timedout(struct drm_sched_job *sched_job) >>> for (i = 0; i < NUM_JOB_SLOTS; i++) >>> drm_sched_start(&pfdev->js->queue[i].sched, true); >>> >>> + pfdev->is_resetting = false; >>> mutex_unlock(&pfdev->reset_lock); >>> } >>> >>> diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c >>> index 148468447ba9..bc6d1862ec8a 100644 >>> --- a/drivers/gpu/drm/scheduler/sched_main.c >>> +++ b/drivers/gpu/drm/scheduler/sched_main.c >>> @@ -415,7 +415,10 @@ void drm_sched_stop(struct drm_gpu_scheduler *sched, struct drm_sched_job *bad) >>> * this TDR finished and before the newly restarted jobs had a >>> * chance to complete. >>> */ >>> - cancel_delayed_work(&sched->work_tdr); >>> + if (bad->sched != sched) >>> + cancel_delayed_work_sync(&sched->work_tdr); >>> + else >>> + cancel_delayed_work(&sched->work_tdr); >>> } >>> >>> EXPORT_SYMBOL(drm_sched_stop); >>> >> > > _______________________________________________ > dri-devel mailing list > dri-devel@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/dri-devel >