Received: by 2002:a25:ab43:0:0:0:0:0 with SMTP id u61csp960919ybi; Fri, 14 Jun 2019 06:20:38 -0700 (PDT) X-Google-Smtp-Source: APXvYqx6mdxrc2cdWVtjZbDihnV0OGlbW2BJbBs2IqYNNK14CCAIze/aIUEyJ7Ep6FJrdBFvw1zb X-Received: by 2002:a62:386:: with SMTP id 128mr101448447pfd.10.1560518438516; Fri, 14 Jun 2019 06:20:38 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1560518438; cv=none; d=google.com; s=arc-20160816; b=Pf2T0CdpyZuiYgDCA3oizuC9dTIMjdH2FNT3aQKM0+TgmdAmawnOpp1U8Iy6cVhDY2 x0qTB+ccGoPVcl7FximyqzIMNBg3Ce2QxvTyGPk7R9YWKvgJE9q31FiZp36QzyxKpt5w gepc6K9bJC29WKyYswLIzuz8VXsYbCBF3CyYXJBA0lHMQ1B2CoF9++H8Oynzg/qHDB2U MBHw7EvOmSVRJqS8J+Kn+9BBT1WFDU3l5/mXGgiTYyItfjhJEygdfOuunVHjm+1sh51t HFVE9fWL5+Reg4ur8PplOLE4trq7UhcmSLG1DWcxECLESAYWrERNNiJtl8sDlCsYObpk 4vxA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-transfer-encoding:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=J27b+HrlctMzcBwFIcA+Fda937PR3Aw4zcivd7d9+UI=; b=ScqwNQa+Lh8ViVuMmT6oJQwnK3SFfH0mLztOvJljxIQeBecJy3w2HkSCpVE//eyKZk kMFOIPzjmNsKSarWx2ldh6WkjbARvvbbD1ek3WXwniSi3k+n5JAtkBLP6OiZQx8ncxt8 lizfujatRXk7IvPzq/T1MZfPVxpDXV/QO7jkXtVwD4eXaLuPQjfFzf864a7ACLDCG4h/ LMMaF7CMrbBljrWq8KpU6fwkZz6JZ+E221cE94wQSy3aW/B1zbv/D/Umiws6w5QNcoe2 fM0s7gpZNJy9snKryVGuxwZdkkB6tvbregf6OUAC0lqFuMwa/r+gRkVCTdqb+ITjsQ/o mceg== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@infradead.org header.s=merlin.20170209 header.b=bLymlhl3; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j10si2614020pgh.273.2019.06.14.06.20.21; Fri, 14 Jun 2019 06:20:38 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=fail header.i=@infradead.org header.s=merlin.20170209 header.b=bLymlhl3; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728049AbfFNNTm (ORCPT + 99 others); Fri, 14 Jun 2019 09:19:42 -0400 Received: from merlin.infradead.org ([205.233.59.134]:38772 "EHLO merlin.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727913AbfFNNTm (ORCPT ); Fri, 14 Jun 2019 09:19:42 -0400 DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=merlin.20170209; h=In-Reply-To:Content-Transfer-Encoding: Content-Type:MIME-Version:References:Message-ID:Subject:Cc:To:From:Date: Sender:Reply-To:Content-ID:Content-Description:Resent-Date:Resent-From: Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Id:List-Help: List-Unsubscribe:List-Subscribe:List-Post:List-Owner:List-Archive; bh=J27b+HrlctMzcBwFIcA+Fda937PR3Aw4zcivd7d9+UI=; b=bLymlhl3W46V0GdF77FYVHhqZb fOoDaDU3o7LR/EN8DoFEjda8u2W8HILBZOqJYlwb4K8+AdK9uaRK99WklFqCR6GUw5IfBrAhSjMzR jNk8l5LnEFd7+3hofHl/+8WfRX9oDmsl0A/EQoEROP7mlGqR6ZGtPCLXzSLUJMLoaGvghudrjaDrJ dp6N5LQNtMFHZoa/i8nqd3FEgbKQmBMxWr2r2kRYDM+tgpza25AGOjptXaeChq3WW8vidR/PZO/Ku BghMnb8Wx/tXIGSY5cTuS48kQ3BV5cXPxWi0gC7HRyTVABcdmgeUtz9LSL/iQmr9w57qMMBZqGwo8 /MXfEiNA==; Received: from j217100.upc-j.chello.nl ([24.132.217.100] helo=hirez.programming.kicks-ass.net) by merlin.infradead.org with esmtpsa (Exim 4.92 #3 (Red Hat Linux)) id 1hbm6x-0007w5-1o; Fri, 14 Jun 2019 13:19:19 +0000 Received: by hirez.programming.kicks-ass.net (Postfix, from userid 1000) id F03D820245B51; Fri, 14 Jun 2019 15:19:16 +0200 (CEST) Date: Fri, 14 Jun 2019 15:19:16 +0200 From: Peter Zijlstra To: Christian =?iso-8859-1?Q?K=F6nig?= Cc: daniel@ffwll.ch, l.stach@pengutronix.de, linux+etnaviv@armlinux.org.uk, christian.gmeiner@gmail.com, yuq825@gmail.com, eric@anholt.net, thellstrom@vmware.com, dri-devel@lists.freedesktop.org, linux-kernel@vger.kernel.org, etnaviv@lists.freedesktop.org, lima@lists.freedesktop.org Subject: Re: [PATCH 3/6] drm/gem: use new ww_mutex_(un)lock_for_each macros Message-ID: <20190614131916.GQ3436@hirez.programming.kicks-ass.net> References: <20190614124125.124181-1-christian.koenig@amd.com> <20190614124125.124181-4-christian.koenig@amd.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <20190614124125.124181-4-christian.koenig@amd.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Jun 14, 2019 at 02:41:22PM +0200, Christian K?nig wrote: > Use the provided macros instead of implementing deadlock handling on our own. > > Signed-off-by: Christian K?nig > --- > drivers/gpu/drm/drm_gem.c | 49 ++++++++++----------------------------- > 1 file changed, 12 insertions(+), 37 deletions(-) > > diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c > index 50de138c89e0..6e4623d3bee2 100644 > --- a/drivers/gpu/drm/drm_gem.c > +++ b/drivers/gpu/drm/drm_gem.c > @@ -1307,51 +1307,26 @@ int > drm_gem_lock_reservations(struct drm_gem_object **objs, int count, > struct ww_acquire_ctx *acquire_ctx) > { > - int contended = -1; > + struct ww_mutex *contended; > int i, ret; > > ww_acquire_init(acquire_ctx, &reservation_ww_class); > > -retry: > - if (contended != -1) { > - struct drm_gem_object *obj = objs[contended]; > - > - ret = ww_mutex_lock_slow_interruptible(&obj->resv->lock, > - acquire_ctx); > - if (ret) { > - ww_acquire_done(acquire_ctx); > - return ret; > - } > - } > - > - for (i = 0; i < count; i++) { > - if (i == contended) > - continue; > - > - ret = ww_mutex_lock_interruptible(&objs[i]->resv->lock, > - acquire_ctx); > - if (ret) { > - int j; > - > - for (j = 0; j < i; j++) > - ww_mutex_unlock(&objs[j]->resv->lock); > - > - if (contended != -1 && contended >= i) > - ww_mutex_unlock(&objs[contended]->resv->lock); > - > - if (ret == -EDEADLK) { > - contended = i; > - goto retry; > - } > - > - ww_acquire_done(acquire_ctx); > - return ret; > - } > - } I note all the sites you use this on are simple idx iterators; so how about something like so: int ww_mutex_unlock_all(int count, void *data, struct ww_mutex *(*func)(int, void *)) { int i; for (i = 0; i < count; i++) { lock = func(i, data); ww_mutex_unlock(lock); } } int ww_mutex_lock_all(int count, struct ww_acquire_context *acquire_ctx, bool intr, void *data, struct ww_mutex *(*func)(int, void *)) { int i, ret, contended = -1; struct ww_mutex *lock; retry: if (contended != -1) { lock = func(contended, data); if (intr) ret = ww_mutex_lock_slow_interruptible(lock, acquire_ctx); else ret = ww_mutex_lock_slow(lock, acquire_ctx), 0; if (ret) { ww_acquire_done(acquire_ctx); return ret; } } for (i = 0; i < count; i++) { if (i == contended) continue; lock = func(i, data); if (intr) ret = ww_mutex_lock_interruptible(lock, acquire_ctx); else ret = ww_mutex_lock(lock, acquire_ctx), 0; if (ret) { ww_mutex_unlock_all(i, data, func); if (contended > i) { lock = func(contended, data); ww_mutex_unlock(lock); } if (ret == -EDEADLK) { contended = i; goto retry; } ww_acquire_done(acquire_ctx); return ret; } } ww_acquire_done(acquire_ctx); return 0; } > + ww_mutex_lock_for_each(for (i = 0; i < count; i++), > + &objs[i]->resv->lock, contended, ret, true, > + acquire_ctx) > + if (ret) > + goto error; which then becomes: struct ww_mutex *gem_ww_mutex_func(int i, void *data) { struct drm_gem_object **objs = data; return &objs[i]->resv->lock; } ret = ww_mutex_lock_all(count, acquire_ctx, true, objs, gem_ww_mutex_func); > ww_acquire_done(acquire_ctx); > > return 0; > + > +error: > + ww_mutex_unlock_for_each(for (i = 0; i < count; i++), > + &objs[i]->resv->lock, contended); > + ww_acquire_done(acquire_ctx); > + return ret; > } > EXPORT_SYMBOL(drm_gem_lock_reservations); > > -- > 2.17.1 >