Received: by 10.223.185.116 with SMTP id b49csp2375230wrg; Mon, 5 Mar 2018 01:34:35 -0800 (PST) X-Google-Smtp-Source: AG47ELsLdrY+E1BlZlLXunwT5yi8tc+7Rqsz7MqtL6gIsZpAs+5MVaiUH0x0a84cnqCMPvSv7NcQ X-Received: by 2002:a17:902:bcc5:: with SMTP id o5-v6mr12074452pls.86.1520242474961; Mon, 05 Mar 2018 01:34:34 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1520242474; cv=none; d=google.com; s=arc-20160816; b=gX4dNFQ4cqihYI0MC6j3LIqg9AodGu3anl59auvrYn0DRQx4cOKGvmPK6qZZ84LYed B1AWM1iSAd/el4sV5263vGdB+L5N7xV1DuMO+76T0bk3PA6BzV2HLEt7Ss/KmZRQfnl3 Jhke09mqhAL+6MYPHF9KcFXH0QX5gA/D31os0nDCZl2yQVCbMd8qnH0TAxcOJ7x5pOBt 70SiDoOAmtIfFxmSEI287IPt8yxplnQfAsmlR+FBEYOqjfZ3ilZCJ+pBgquSQC3SXpa5 wduYgO9dUBo69hb/Bn0CC71wjc8E8QTALByJTZpYPC5ts3QG1ekWfHjBT8m1eTNo+E8L ITiA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:mail-followup-to :message-id:subject:cc:to:from:date:dkim-signature :arc-authentication-results; bh=h/uQo5uLmFYC/JW6qgA/92M9wCDtpuhAWSnFCNCbdGA=; b=DwbvTdJSqLbvAUmTQxVG9IoCBz9DVTvYYe0V8PPPlrezCtMXa6aUxkjkIJ+VC0MMur QEmw0rnexdkPrtmVNGlFeCjTkgEUijxdlj9WwZzIpHLAkFlBtmZxa6iFfNB978GkncT0 fDEsbGT33r/qcwQzmgsqtk7Tj7XNJYBhaL3RjrlNlpc6zjTn3GbpD6Vti/FjpQV5/xz8 dlS7T8/FrNvGi5Y/0EYDf0WTKpJUIg/362Uq3Sbm4aPdiWuF1FH5NSDsdkSog/grq3ls dSsjxYgiMcg0Y0pBwcv4Al2JIqbzfedXnQhTj6bhBOO+5ndnIxLoDda1b3zsCDpJQnGX r/ow== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@ffwll.ch header.s=google header.b=knRz6MX7; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q11si8164316pgf.90.2018.03.05.01.34.20; Mon, 05 Mar 2018 01:34:34 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=fail header.i=@ffwll.ch header.s=google header.b=knRz6MX7; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S933221AbeCEJZY (ORCPT + 99 others); Mon, 5 Mar 2018 04:25:24 -0500 Received: from mail-wm0-f66.google.com ([74.125.82.66]:54052 "EHLO mail-wm0-f66.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751776AbeCEJZT (ORCPT ); Mon, 5 Mar 2018 04:25:19 -0500 Received: by mail-wm0-f66.google.com with SMTP id t74so14408264wme.3 for ; Mon, 05 Mar 2018 01:25:19 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ffwll.ch; s=google; h=sender:date:from:to:cc:subject:message-id:mail-followup-to :references:mime-version:content-disposition:in-reply-to:user-agent; bh=h/uQo5uLmFYC/JW6qgA/92M9wCDtpuhAWSnFCNCbdGA=; b=knRz6MX7dU89mHv71FcNQZp1hh2nakmw3wH9lDRajBKDMv2p7O00V2MwmoEpNokhze kMTIXAdM52lYBpyspxabIsv5ttJaCI+M99BWxdBNoI3Wd7ZfDJEAcJ5+/Lxvvw9Y4GOA dffxxadNe5KfS8EzGbmPtam1U+OKAmuwDA4Vg= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:sender:date:from:to:cc:subject:message-id :mail-followup-to:references:mime-version:content-disposition :in-reply-to:user-agent; bh=h/uQo5uLmFYC/JW6qgA/92M9wCDtpuhAWSnFCNCbdGA=; b=aiyHEyL81BPQH5MSSZHdszTcDTHjeTGz0bIHjyHCp41t//N6K96gihLA5hZs3ueJBR rkJfgN+4bfZzDAwYiSO+Ijf3OcwCy20mW4wk7FZMgtmWvFzYdz6TsvdcPQjWBNRmyxQx XEqA+4/hcr/UG2uSbLS1f9kdiAbBRCzY9RhaBJ5fkjqymAaG1K0C7LWhv8ache4eWkeK GHKhU1ZQkz5fylcWtQO3pp+51hlXGhzI7/tzcCrM8tyrHWkpW/Mv0DTRHBvsd2iQz+PT Bq3bewCO8I1XMgIjVcIdVp+WbJdfsHJzm0MH71Kw+ks1tkhn3QRmBQe+CN3NDuybP6aO t27g== X-Gm-Message-State: APf1xPDD8zFc7VrFE7rwGCQ/FYoft1AI2aTHEqqhZAdHevRMOQhmGOSs Jk5z6VeEWnCN2iUxAfebPzrfqw== X-Received: by 10.80.224.9 with SMTP id e9mr17766401edl.218.1520241918367; Mon, 05 Mar 2018 01:25:18 -0800 (PST) Received: from phenom.ffwll.local ([2a02:168:5635:0:39d2:f87e:2033:9f6]) by smtp.gmail.com with ESMTPSA id k5sm12011498edc.3.2018.03.05.01.25.17 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Mon, 05 Mar 2018 01:25:17 -0800 (PST) Date: Mon, 5 Mar 2018 10:25:15 +0100 From: Daniel Vetter To: Oleksandr Andrushchenko Cc: xen-devel@lists.xenproject.org, linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, airlied@linux.ie, daniel.vetter@intel.com, seanpaul@chromium.org, gustavo@padovan.org, jgross@suse.com, boris.ostrovsky@oracle.com, konrad.wilk@oracle.com, Oleksandr Andrushchenko Subject: Re: [PATCH 9/9] drm/xen-front: Implement communication with backend Message-ID: <20180305092515.GJ22212@phenom.ffwll.local> Mail-Followup-To: Oleksandr Andrushchenko , xen-devel@lists.xenproject.org, linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, airlied@linux.ie, daniel.vetter@intel.com, seanpaul@chromium.org, gustavo@padovan.org, jgross@suse.com, boris.ostrovsky@oracle.com, konrad.wilk@oracle.com, Oleksandr Andrushchenko References: <1519200222-20623-1-git-send-email-andr2000@gmail.com> <1519200222-20623-10-git-send-email-andr2000@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1519200222-20623-10-git-send-email-andr2000@gmail.com> X-Operating-System: Linux phenom 4.14.0-3-amd64 User-Agent: Mutt/1.9.3 (2018-01-21) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Feb 21, 2018 at 10:03:42AM +0200, Oleksandr Andrushchenko wrote: > From: Oleksandr Andrushchenko > > Handle communication with the backend: > - send requests and wait for the responses according > to the displif protocol > - serialize access to the communication channel > - time-out used for backend communication is set to 3000 ms > - manage display buffers shared with the backend > > Signed-off-by: Oleksandr Andrushchenko After the demidlayering it probably makes sense to merge this with the overall kms/basic-drm-driver patch. Up to you really. -Daniel > --- > drivers/gpu/drm/xen/xen_drm_front.c | 327 +++++++++++++++++++++++++++++++++++- > drivers/gpu/drm/xen/xen_drm_front.h | 5 + > 2 files changed, 327 insertions(+), 5 deletions(-) > > diff --git a/drivers/gpu/drm/xen/xen_drm_front.c b/drivers/gpu/drm/xen/xen_drm_front.c > index 8de88e359d5e..5ad546231d30 100644 > --- a/drivers/gpu/drm/xen/xen_drm_front.c > +++ b/drivers/gpu/drm/xen/xen_drm_front.c > @@ -31,12 +31,146 @@ > #include "xen_drm_front_evtchnl.h" > #include "xen_drm_front_shbuf.h" > > +/* timeout in ms to wait for backend to respond */ > +#define VDRM_WAIT_BACK_MS 3000 > + > +struct xen_drm_front_dbuf { > + struct list_head list; > + uint64_t dbuf_cookie; > + uint64_t fb_cookie; > + struct xen_drm_front_shbuf *shbuf; > +}; > + > +static int dbuf_add_to_list(struct xen_drm_front_info *front_info, > + struct xen_drm_front_shbuf *shbuf, uint64_t dbuf_cookie) > +{ > + struct xen_drm_front_dbuf *dbuf; > + > + dbuf = kzalloc(sizeof(*dbuf), GFP_KERNEL); > + if (!dbuf) > + return -ENOMEM; > + > + dbuf->dbuf_cookie = dbuf_cookie; > + dbuf->shbuf = shbuf; > + list_add(&dbuf->list, &front_info->dbuf_list); > + return 0; > +} > + > +static struct xen_drm_front_dbuf *dbuf_get(struct list_head *dbuf_list, > + uint64_t dbuf_cookie) > +{ > + struct xen_drm_front_dbuf *buf, *q; > + > + list_for_each_entry_safe(buf, q, dbuf_list, list) > + if (buf->dbuf_cookie == dbuf_cookie) > + return buf; > + > + return NULL; > +} > + > +static void dbuf_flush_fb(struct list_head *dbuf_list, uint64_t fb_cookie) > +{ > + struct xen_drm_front_dbuf *buf, *q; > + > + list_for_each_entry_safe(buf, q, dbuf_list, list) > + if (buf->fb_cookie == fb_cookie) > + xen_drm_front_shbuf_flush(buf->shbuf); > +} > + > +static void dbuf_free(struct list_head *dbuf_list, uint64_t dbuf_cookie) > +{ > + struct xen_drm_front_dbuf *buf, *q; > + > + list_for_each_entry_safe(buf, q, dbuf_list, list) > + if (buf->dbuf_cookie == dbuf_cookie) { > + list_del(&buf->list); > + xen_drm_front_shbuf_unmap(buf->shbuf); > + xen_drm_front_shbuf_free(buf->shbuf); > + kfree(buf); > + break; > + } > +} > + > +static void dbuf_free_all(struct list_head *dbuf_list) > +{ > + struct xen_drm_front_dbuf *buf, *q; > + > + list_for_each_entry_safe(buf, q, dbuf_list, list) { > + list_del(&buf->list); > + xen_drm_front_shbuf_unmap(buf->shbuf); > + xen_drm_front_shbuf_free(buf->shbuf); > + kfree(buf); > + } > +} > + > +static struct xendispl_req *be_prepare_req( > + struct xen_drm_front_evtchnl *evtchnl, uint8_t operation) > +{ > + struct xendispl_req *req; > + > + req = RING_GET_REQUEST(&evtchnl->u.req.ring, > + evtchnl->u.req.ring.req_prod_pvt); > + req->operation = operation; > + req->id = evtchnl->evt_next_id++; > + evtchnl->evt_id = req->id; > + return req; > +} > + > +static int be_stream_do_io(struct xen_drm_front_evtchnl *evtchnl, > + struct xendispl_req *req) > +{ > + reinit_completion(&evtchnl->u.req.completion); > + if (unlikely(evtchnl->state != EVTCHNL_STATE_CONNECTED)) > + return -EIO; > + > + xen_drm_front_evtchnl_flush(evtchnl); > + return 0; > +} > + > +static int be_stream_wait_io(struct xen_drm_front_evtchnl *evtchnl) > +{ > + if (wait_for_completion_timeout(&evtchnl->u.req.completion, > + msecs_to_jiffies(VDRM_WAIT_BACK_MS)) <= 0) > + return -ETIMEDOUT; > + > + return evtchnl->u.req.resp_status; > +} > + > static int be_mode_set(struct xen_drm_front_drm_pipeline *pipeline, uint32_t x, > uint32_t y, uint32_t width, uint32_t height, uint32_t bpp, > uint64_t fb_cookie) > > { > - return 0; > + struct xen_drm_front_evtchnl *evtchnl; > + struct xen_drm_front_info *front_info; > + struct xendispl_req *req; > + unsigned long flags; > + int ret; > + > + front_info = pipeline->drm_info->front_info; > + evtchnl = &front_info->evt_pairs[pipeline->index].req; > + if (unlikely(!evtchnl)) > + return -EIO; > + > + mutex_lock(&front_info->req_io_lock); > + > + spin_lock_irqsave(&front_info->io_lock, flags); > + req = be_prepare_req(evtchnl, XENDISPL_OP_SET_CONFIG); > + req->op.set_config.x = x; > + req->op.set_config.y = y; > + req->op.set_config.width = width; > + req->op.set_config.height = height; > + req->op.set_config.bpp = bpp; > + req->op.set_config.fb_cookie = fb_cookie; > + > + ret = be_stream_do_io(evtchnl, req); > + spin_unlock_irqrestore(&front_info->io_lock, flags); > + > + if (ret == 0) > + ret = be_stream_wait_io(evtchnl); > + > + mutex_unlock(&front_info->req_io_lock); > + return ret; > } > > static int be_dbuf_create_int(struct xen_drm_front_info *front_info, > @@ -44,7 +178,69 @@ static int be_dbuf_create_int(struct xen_drm_front_info *front_info, > uint32_t bpp, uint64_t size, struct page **pages, > struct sg_table *sgt) > { > + struct xen_drm_front_evtchnl *evtchnl; > + struct xen_drm_front_shbuf *shbuf; > + struct xendispl_req *req; > + struct xen_drm_front_shbuf_cfg buf_cfg; > + unsigned long flags; > + int ret; > + > + evtchnl = &front_info->evt_pairs[GENERIC_OP_EVT_CHNL].req; > + if (unlikely(!evtchnl)) > + return -EIO; > + > + memset(&buf_cfg, 0, sizeof(buf_cfg)); > + buf_cfg.xb_dev = front_info->xb_dev; > + buf_cfg.pages = pages; > + buf_cfg.size = size; > + buf_cfg.sgt = sgt; > + buf_cfg.be_alloc = front_info->cfg.be_alloc; > + > + shbuf = xen_drm_front_shbuf_alloc(&buf_cfg); > + if (!shbuf) > + return -ENOMEM; > + > + ret = dbuf_add_to_list(front_info, shbuf, dbuf_cookie); > + if (ret < 0) { > + xen_drm_front_shbuf_free(shbuf); > + return ret; > + } > + > + mutex_lock(&front_info->req_io_lock); > + > + spin_lock_irqsave(&front_info->io_lock, flags); > + req = be_prepare_req(evtchnl, XENDISPL_OP_DBUF_CREATE); > + req->op.dbuf_create.gref_directory = > + xen_drm_front_shbuf_get_dir_start(shbuf); > + req->op.dbuf_create.buffer_sz = size; > + req->op.dbuf_create.dbuf_cookie = dbuf_cookie; > + req->op.dbuf_create.width = width; > + req->op.dbuf_create.height = height; > + req->op.dbuf_create.bpp = bpp; > + if (buf_cfg.be_alloc) > + req->op.dbuf_create.flags |= XENDISPL_DBUF_FLG_REQ_ALLOC; > + > + ret = be_stream_do_io(evtchnl, req); > + spin_unlock_irqrestore(&front_info->io_lock, flags); > + > + if (ret < 0) > + goto fail; > + > + ret = be_stream_wait_io(evtchnl); > + if (ret < 0) > + goto fail; > + > + ret = xen_drm_front_shbuf_map(shbuf); > + if (ret < 0) > + goto fail; > + > + mutex_unlock(&front_info->req_io_lock); > return 0; > + > +fail: > + mutex_unlock(&front_info->req_io_lock); > + dbuf_free(&front_info->dbuf_list, dbuf_cookie); > + return ret; > } > > static int be_dbuf_create_from_sgt(struct xen_drm_front_info *front_info, > @@ -66,26 +262,144 @@ static int be_dbuf_create_from_pages(struct xen_drm_front_info *front_info, > static int be_dbuf_destroy(struct xen_drm_front_info *front_info, > uint64_t dbuf_cookie) > { > - return 0; > + struct xen_drm_front_evtchnl *evtchnl; > + struct xendispl_req *req; > + unsigned long flags; > + bool be_alloc; > + int ret; > + > + evtchnl = &front_info->evt_pairs[GENERIC_OP_EVT_CHNL].req; > + if (unlikely(!evtchnl)) > + return -EIO; > + > + be_alloc = front_info->cfg.be_alloc; > + > + /* > + * for the backend allocated buffer release references now, so backend > + * can free the buffer > + */ > + if (be_alloc) > + dbuf_free(&front_info->dbuf_list, dbuf_cookie); > + > + mutex_lock(&front_info->req_io_lock); > + > + spin_lock_irqsave(&front_info->io_lock, flags); > + req = be_prepare_req(evtchnl, XENDISPL_OP_DBUF_DESTROY); > + req->op.dbuf_destroy.dbuf_cookie = dbuf_cookie; > + > + ret = be_stream_do_io(evtchnl, req); > + spin_unlock_irqrestore(&front_info->io_lock, flags); > + > + if (ret == 0) > + ret = be_stream_wait_io(evtchnl); > + > + /* > + * do this regardless of communication status with the backend: > + * if we cannot remove remote resources remove what we can locally > + */ > + if (!be_alloc) > + dbuf_free(&front_info->dbuf_list, dbuf_cookie); > + > + mutex_unlock(&front_info->req_io_lock); > + return ret; > } > > static int be_fb_attach(struct xen_drm_front_info *front_info, > uint64_t dbuf_cookie, uint64_t fb_cookie, uint32_t width, > uint32_t height, uint32_t pixel_format) > { > - return 0; > + struct xen_drm_front_evtchnl *evtchnl; > + struct xen_drm_front_dbuf *buf; > + struct xendispl_req *req; > + unsigned long flags; > + int ret; > + > + evtchnl = &front_info->evt_pairs[GENERIC_OP_EVT_CHNL].req; > + if (unlikely(!evtchnl)) > + return -EIO; > + > + buf = dbuf_get(&front_info->dbuf_list, dbuf_cookie); > + if (!buf) > + return -EINVAL; > + > + buf->fb_cookie = fb_cookie; > + > + mutex_lock(&front_info->req_io_lock); > + > + spin_lock_irqsave(&front_info->io_lock, flags); > + req = be_prepare_req(evtchnl, XENDISPL_OP_FB_ATTACH); > + req->op.fb_attach.dbuf_cookie = dbuf_cookie; > + req->op.fb_attach.fb_cookie = fb_cookie; > + req->op.fb_attach.width = width; > + req->op.fb_attach.height = height; > + req->op.fb_attach.pixel_format = pixel_format; > + > + ret = be_stream_do_io(evtchnl, req); > + spin_unlock_irqrestore(&front_info->io_lock, flags); > + > + if (ret == 0) > + ret = be_stream_wait_io(evtchnl); > + > + mutex_unlock(&front_info->req_io_lock); > + return ret; > } > > static int be_fb_detach(struct xen_drm_front_info *front_info, > uint64_t fb_cookie) > { > - return 0; > + struct xen_drm_front_evtchnl *evtchnl; > + struct xendispl_req *req; > + unsigned long flags; > + int ret; > + > + evtchnl = &front_info->evt_pairs[GENERIC_OP_EVT_CHNL].req; > + if (unlikely(!evtchnl)) > + return -EIO; > + > + mutex_lock(&front_info->req_io_lock); > + > + spin_lock_irqsave(&front_info->io_lock, flags); > + req = be_prepare_req(evtchnl, XENDISPL_OP_FB_DETACH); > + req->op.fb_detach.fb_cookie = fb_cookie; > + > + ret = be_stream_do_io(evtchnl, req); > + spin_unlock_irqrestore(&front_info->io_lock, flags); > + > + if (ret == 0) > + ret = be_stream_wait_io(evtchnl); > + > + mutex_unlock(&front_info->req_io_lock); > + return ret; > } > > static int be_page_flip(struct xen_drm_front_info *front_info, int conn_idx, > uint64_t fb_cookie) > { > - return 0; > + struct xen_drm_front_evtchnl *evtchnl; > + struct xendispl_req *req; > + unsigned long flags; > + int ret; > + > + if (unlikely(conn_idx >= front_info->num_evt_pairs)) > + return -EINVAL; > + > + dbuf_flush_fb(&front_info->dbuf_list, fb_cookie); > + evtchnl = &front_info->evt_pairs[conn_idx].req; > + > + mutex_lock(&front_info->req_io_lock); > + > + spin_lock_irqsave(&front_info->io_lock, flags); > + req = be_prepare_req(evtchnl, XENDISPL_OP_PG_FLIP); > + req->op.pg_flip.fb_cookie = fb_cookie; > + > + ret = be_stream_do_io(evtchnl, req); > + spin_unlock_irqrestore(&front_info->io_lock, flags); > + > + if (ret == 0) > + ret = be_stream_wait_io(evtchnl); > + > + mutex_unlock(&front_info->req_io_lock); > + return ret; > } > > static void xen_drm_drv_unload(struct xen_drm_front_info *front_info) > @@ -183,6 +497,7 @@ static void xen_drv_remove_internal(struct xen_drm_front_info *front_info) > { > xen_drm_drv_deinit(front_info); > xen_drm_front_evtchnl_free_all(front_info); > + dbuf_free_all(&front_info->dbuf_list); > } > > static int backend_on_initwait(struct xen_drm_front_info *front_info) > @@ -310,6 +625,8 @@ static int xen_drv_probe(struct xenbus_device *xb_dev, > > front_info->xb_dev = xb_dev; > spin_lock_init(&front_info->io_lock); > + mutex_init(&front_info->req_io_lock); > + INIT_LIST_HEAD(&front_info->dbuf_list); > front_info->drm_pdrv_registered = false; > dev_set_drvdata(&xb_dev->dev, front_info); > return xenbus_switch_state(xb_dev, XenbusStateInitialising); > diff --git a/drivers/gpu/drm/xen/xen_drm_front.h b/drivers/gpu/drm/xen/xen_drm_front.h > index c6f52c892434..db32d00145d1 100644 > --- a/drivers/gpu/drm/xen/xen_drm_front.h > +++ b/drivers/gpu/drm/xen/xen_drm_front.h > @@ -137,6 +137,8 @@ struct xen_drm_front_info { > struct xenbus_device *xb_dev; > /* to protect data between backend IO code and interrupt handler */ > spinlock_t io_lock; > + /* serializer for backend IO: request/response */ > + struct mutex req_io_lock; > bool drm_pdrv_registered; > /* virtual DRM platform device */ > struct platform_device *drm_pdev; > @@ -144,6 +146,9 @@ struct xen_drm_front_info { > int num_evt_pairs; > struct xen_drm_front_evtchnl_pair *evt_pairs; > struct xen_drm_front_cfg cfg; > + > + /* display buffers */ > + struct list_head dbuf_list; > }; > > #endif /* __XEN_DRM_FRONT_H_ */ > -- > 2.7.4 > > _______________________________________________ > dri-devel mailing list > dri-devel@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/dri-devel -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch