Received: by 2002:a25:c205:0:0:0:0:0 with SMTP id s5csp1522012ybf; Thu, 27 Feb 2020 12:38:12 -0800 (PST) X-Google-Smtp-Source: APXvYqyAaFMLsAjcKdHDkVjtMG1xd0Zb1Z9N+HsQSsQdxSLcFfwmKHf+e3vhtoqwnNp1y+CHlsn0 X-Received: by 2002:a05:6830:1098:: with SMTP id y24mr575350oto.197.1582835891717; Thu, 27 Feb 2020 12:38:11 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1582835891; cv=none; d=google.com; s=arc-20160816; b=RCdoLv6hd8n+QXrnXOxvGIsOA7Dea42853EiYHu2A42dRhxzGuorgnR1wlrFIhKYxo XNwt9Wsr2//a2zYf0FFSCrZjlsKFaimLxksRkAXJOseu3CjarjBN/duN5Nh7MKhqzpUe 7lhTFg3ApqyFArrB9vi1MBAZWQxdAkQiPuaCBg1NdzKF2mB79Jg+YP39VHYbTnSVu1OH zLMTCWtiLh5yh2yBAE9CG724nWsk7OU7skQzWKMOfnqA9sJ+3RnP/UhHf1RpME30STqQ bDtG/lxbngMDoL1QBLyKNDyHp/1dDlushvOsaQpf+BWZovLkdw22b+V2JXUwIuluE1OW NykA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=S52F+Z8loCBC9QLJBM7st+OZaBKLb/smZzBpNMziyuk=; b=q8zHSLxVSC358M0Tt6+zHHP9wFTk5zxGZW/rK3EwrI8DAiHJqEA0RWvN/ZS3HUfbEv iTyj69rchYsgC6zfVQwZCsXC5VuBLjaWaylCakhHOnYRxR+jR0Q0JEmX/39mNCb6hyYq xvVCWeN7MY++bfFQaTzF+HSSsRlqVHhca12LEq0ZAA0TphlMLN3Fszl3Q1tBpHxUwbTd KY9gz47IQOB+ap4n/qzghIBKH91uMx7NctgCU2mHAogDJNZnW0PjkK75oTYNqu8Hemdl nDZ2dH6ZlaMdaKUEGxC7ONzMcNEDbeKzXer9OjzpjakEuvWk85OGOr/ZYmNnxWq6OD7e 80HA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=VUWplpBz; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j25si501559oij.242.2020.02.27.12.37.58; Thu, 27 Feb 2020 12:38:11 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=VUWplpBz; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729555AbgB0Uhm (ORCPT + 99 others); Thu, 27 Feb 2020 15:37:42 -0500 Received: from mail.kernel.org ([198.145.29.99]:37006 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726793AbgB0Uhl (ORCPT ); Thu, 27 Feb 2020 15:37:41 -0500 Received: from mail-qt1-f182.google.com (mail-qt1-f182.google.com [209.85.160.182]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 8CCA8246A3 for ; Thu, 27 Feb 2020 20:37:38 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1582835858; bh=l9nS99EYzLaPL4XHnqtGfQeTw+aMLJUOcFkui2aBGso=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=VUWplpBzQY86XrlMe+2sLdHtD5pR8UYl66QbxDYdqWU7NWhrv856kGBI7A5wm4Omp JGQ84ReOIXTobBRXRW45izEKII7ygHYs3Dnw/jADlmY2lMgyMabaSXc2efzqspEvz4 9l+V6B+jPUYKY4goOVTjrnCuw0QX/HsusauE+1ns= Received: by mail-qt1-f182.google.com with SMTP id p34so342744qtb.6 for ; Thu, 27 Feb 2020 12:37:38 -0800 (PST) X-Gm-Message-State: APjAAAVjeME81hREgFELj/tR9d50YnvaLhNtMc7MNLqhvSYYJeKDoVFF 2KLziFan/P4j7EpsdMlg4HcopTwn8ir/ADuW1w== X-Received: by 2002:ac8:59:: with SMTP id i25mr1223027qtg.110.1582835857455; Thu, 27 Feb 2020 12:37:37 -0800 (PST) MIME-Version: 1.0 References: <1582710377-15489-1-git-send-email-kevin3.tang@gmail.com> <1582710377-15489-5-git-send-email-kevin3.tang@gmail.com> In-Reply-To: <1582710377-15489-5-git-send-email-kevin3.tang@gmail.com> From: Rob Herring Date: Thu, 27 Feb 2020 14:37:25 -0600 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [PATCH RFC v4 4/6] drm/sprd: add Unisoc's drm display controller driver To: Kevin Tang Cc: Maarten Lankhorst , Maxime Ripard , Sean Paul , David Airlie , Daniel Vetter , Mark Rutland , Orson Zhai , Baolin Wang , Lyra Zhang , "linux-kernel@vger.kernel.org" , dri-devel Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Feb 26, 2020 at 3:46 AM Kevin Tang wrote: > > Adds DPU(Display Processor Unit) support for the Unisoc's display subsystem. > It's support multi planes, scaler, rotation, PQ(Picture Quality) and more. > > Cc: Orson Zhai > Cc: Baolin Wang > Cc: Chunyan Zhang > Signed-off-by: Kevin Tang > --- > drivers/gpu/drm/sprd/Makefile | 5 +- > drivers/gpu/drm/sprd/dpu/Makefile | 7 + > drivers/gpu/drm/sprd/dpu/dpu_r2p0.c | 770 ++++++++++++++++++++++++++++++++++++ > drivers/gpu/drm/sprd/sprd_dpu.c | 586 +++++++++++++++++++++++++++ > drivers/gpu/drm/sprd/sprd_dpu.h | 127 ++++++ > drivers/gpu/drm/sprd/sprd_drm.c | 1 + > drivers/gpu/drm/sprd/sprd_drm.h | 2 + > 7 files changed, 1497 insertions(+), 1 deletion(-) > create mode 100644 drivers/gpu/drm/sprd/dpu/Makefile > create mode 100644 drivers/gpu/drm/sprd/dpu/dpu_r2p0.c > create mode 100644 drivers/gpu/drm/sprd/sprd_dpu.c > create mode 100644 drivers/gpu/drm/sprd/sprd_dpu.h > > diff --git a/drivers/gpu/drm/sprd/Makefile b/drivers/gpu/drm/sprd/Makefile > index 86d95d9..88ab32a 100644 > --- a/drivers/gpu/drm/sprd/Makefile > +++ b/drivers/gpu/drm/sprd/Makefile > @@ -2,4 +2,7 @@ > > subdir-ccflags-y += -I$(srctree)/$(src) > > -obj-y := sprd_drm.o > +obj-y := sprd_drm.o \ > + sprd_dpu.o > + > +obj-y += dpu/ > diff --git a/drivers/gpu/drm/sprd/dpu/Makefile b/drivers/gpu/drm/sprd/dpu/Makefile > new file mode 100644 > index 0000000..73bd497 > --- /dev/null > +++ b/drivers/gpu/drm/sprd/dpu/Makefile > @@ -0,0 +1,7 @@ > +# SPDX-License-Identifier: GPL-2.0 > + > +ifdef CONFIG_ARM64 > +KBUILD_CFLAGS += -mstrict-align > +endif > + > +obj-y += dpu_r2p0.o > diff --git a/drivers/gpu/drm/sprd/dpu/dpu_r2p0.c b/drivers/gpu/drm/sprd/dpu/dpu_r2p0.c > new file mode 100644 > index 0000000..984fa9b > --- /dev/null > +++ b/drivers/gpu/drm/sprd/dpu/dpu_r2p0.c > @@ -0,0 +1,770 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (C) 2020 Unisoc Inc. > + */ > + > +#include > +#include > +#include > +#include "sprd_dpu.h" > + > +#define DISPC_INT_FBC_PLD_ERR_MASK BIT(8) > +#define DISPC_INT_FBC_HDR_ERR_MASK BIT(9) > + > +#define DISPC_INT_MMU_INV_WR_MASK BIT(19) > +#define DISPC_INT_MMU_INV_RD_MASK BIT(18) > +#define DISPC_INT_MMU_VAOR_WR_MASK BIT(17) > +#define DISPC_INT_MMU_VAOR_RD_MASK BIT(16) > + > +struct layer_reg { > + u32 addr[4]; > + u32 ctrl; > + u32 size; > + u32 pitch; > + u32 pos; > + u32 alpha; > + u32 ck; > + u32 pallete; > + u32 crop_start; > +}; > + > +struct wb_region_reg { > + u32 pos; > + u32 size; > +}; > + > +/* dpu controller register */ > +struct dpu_reg { structs for registers is not normal coding style. #defines of offsets is. More below. > + u32 dpu_version; > + u32 dpu_ctrl; > + u32 dpu_cfg0; > + u32 dpu_cfg1; > + u32 dpu_cfg2; > + u32 dpu_secure; > + u32 reserved_0x0018_0x001C[2]; > + u32 panel_size; > + u32 blend_size; > + u32 reserved_0x0028; > + u32 bg_color; > + struct layer_reg layers[8]; > + u32 wb_base_addr; > + u32 wb_ctrl; > + u32 wb_cfg; > + u32 wb_pitch; > + struct wb_region_reg region[3]; > + u32 reserved_0x01D8_0x01DC[2]; > + u32 dpu_int_en; > + u32 dpu_int_clr; > + u32 dpu_int_sts; > + u32 dpu_int_raw; > + u32 dpi_ctrl; > + u32 dpi_h_timing; > + u32 dpi_v_timing; > + u32 reserved_0x01FC; > + u32 dpu_enhance_cfg; > + u32 reserved_0x0204_0x020C[3]; > + u32 epf_epsilon; > + u32 epf_gain0_3; > + u32 epf_gain4_7; > + u32 epf_diff; > + u32 reserved_0x0220_0x023C[8]; > + u32 hsv_lut_addr; > + u32 hsv_lut_wdata; > + u32 hsv_lut_rdata; > + u32 reserved_0x024C_0x027C[13]; > + u32 cm_coef01_00; > + u32 cm_coef03_02; > + u32 cm_coef11_10; > + u32 cm_coef13_12; > + u32 cm_coef21_20; > + u32 cm_coef23_22; > + u32 reserved_0x0298_0x02BC[10]; > + u32 slp_cfg0; > + u32 slp_cfg1; > + u32 reserved_0x02C8_0x02FC[14]; > + u32 gamma_lut_addr; > + u32 gamma_lut_wdata; > + u32 gamma_lut_rdata; > + u32 reserved_0x030C_0x033C[13]; > + u32 checksum_en; > + u32 checksum0_start_pos; > + u32 checksum0_end_pos; > + u32 checksum1_start_pos; > + u32 checksum1_end_pos; > + u32 checksum0_result; > + u32 checksum1_result; > + u32 reserved_0x035C; > + u32 dpu_sts[18]; > + u32 reserved_0x03A8_0x03AC[2]; > + u32 dpu_fbc_cfg0; > + u32 dpu_fbc_cfg1; > + u32 reserved_0x03B8_0x03EC[14]; > + u32 rf_ram_addr; > + u32 rf_ram_rdata_low; > + u32 rf_ram_rdata_high; > + u32 reserved_0x03FC_0x07FC[257]; > + u32 mmu_en; > + u32 mmu_update; > + u32 mmu_min_vpn; > + u32 mmu_vpn_range; > + u32 mmu_pt_addr; > + u32 mmu_default_page; > + u32 mmu_vaor_addr_rd; > + u32 mmu_vaor_addr_wr; > + u32 mmu_inv_addr_rd; > + u32 mmu_inv_addr_wr; > + u32 mmu_uns_addr_rd; > + u32 mmu_uns_addr_wr; > + u32 mmu_miss_cnt; > + u32 mmu_pt_update_qos; > + u32 mmu_version; > + u32 mmu_min_ppn1; > + u32 mmu_ppn_range1; > + u32 mmu_min_ppn2; > + u32 mmu_ppn_range2; > + u32 mmu_vpn_paor_rd; > + u32 mmu_vpn_paor_wr; > + u32 mmu_ppn_paor_rd; > + u32 mmu_ppn_paor_wr; > + u32 mmu_reg_au_manage; > + u32 mmu_page_rd_ch; > + u32 mmu_page_wr_ch; > + u32 mmu_read_page_cmd_cnt; > + u32 mmu_read_page_latency_cnt; > + u32 mmu_page_max_latency; > +}; > + > +static void dpu_dump(struct dpu_context *ctx) > +{ > + u32 *reg = (u32 *)ctx->base; > + int i; > + > + pr_info(" 0 4 8 C\n"); > + for (i = 0; i < 256; i += 4) { > + pr_info("%04x: 0x%08x 0x%08x 0x%08x 0x%08x\n", > + i * 4, reg[i], reg[i + 1], reg[i + 2], reg[i + 3]); > + } > +} > + > +static u32 check_mmu_isr(struct dpu_context *ctx, u32 reg_val) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + u32 mmu_mask = DISPC_INT_MMU_VAOR_RD_MASK | > + DISPC_INT_MMU_VAOR_WR_MASK | > + DISPC_INT_MMU_INV_RD_MASK | > + DISPC_INT_MMU_INV_WR_MASK; > + u32 val = reg_val & mmu_mask; > + > + if (val) { > + pr_err("--- iommu interrupt err: 0x%04x ---\n", val); > + > + pr_err("iommu invalid read error, addr: 0x%08x\n", > + reg->mmu_inv_addr_rd); > + pr_err("iommu invalid write error, addr: 0x%08x\n", > + reg->mmu_inv_addr_wr); > + pr_err("iommu va out of range read error, addr: 0x%08x\n", > + reg->mmu_vaor_addr_rd); > + pr_err("iommu va out of range write error, addr: 0x%08x\n", > + reg->mmu_vaor_addr_wr); > + pr_err("BUG: iommu failure at %s:%d/%s()!\n", > + __FILE__, __LINE__, __func__); > + > + dpu_dump(ctx); > + } > + > + return val; > +} > + > +static void dpu_clean_all(struct dpu_context *ctx) > +{ > + int i; > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + for (i = 0; i < 8; i++) > + reg->layers[i].ctrl = 0; > +} > + > +static u32 dpu_isr(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + u32 reg_val, int_mask = 0; > + > + reg_val = reg->dpu_int_sts; > + > + /* disable err interrupt */ > + if (reg_val & DISPC_INT_ERR_MASK) > + int_mask |= DISPC_INT_ERR_MASK; > + > + /* dpu update done isr */ > + if (reg_val & DISPC_INT_UPDATE_DONE_MASK) { > + ctx->evt_update = true; > + wake_up_interruptible_all(&ctx->wait_queue); > + } > + > + /* dpu stop done isr */ > + if (reg_val & DISPC_INT_DONE_MASK) { > + ctx->evt_stop = true; > + wake_up_interruptible_all(&ctx->wait_queue); > + } > + > + /* dpu ifbc payload error isr */ > + if (reg_val & DISPC_INT_FBC_PLD_ERR_MASK) { > + int_mask |= DISPC_INT_FBC_PLD_ERR_MASK; > + pr_err("dpu ifbc payload error\n"); > + } > + > + /* dpu ifbc header error isr */ > + if (reg_val & DISPC_INT_FBC_HDR_ERR_MASK) { > + int_mask |= DISPC_INT_FBC_HDR_ERR_MASK; > + pr_err("dpu ifbc header error\n"); > + } > + > + int_mask |= check_mmu_isr(ctx, reg_val); > + > + reg->dpu_int_clr = reg_val; > + reg->dpu_int_en &= ~int_mask; Also, not coding style. Use readl/writel{_relaxed}. This is fragile because the compiler has a lot of freedom in what it can do here including reordering the accesses, changing the access size or skipping the register access altogether. I don't think the latter would happen in this case, but if a read has side effects such as causing bits to clear you'll see problems. > + > + return reg_val; > +} > + > +static int dpu_wait_stop_done(struct dpu_context *ctx) > +{ > + int rc; > + > + if (ctx->stopped) > + return 0; > + > + rc = wait_event_interruptible_timeout(ctx->wait_queue, ctx->evt_stop, > + msecs_to_jiffies(500)); > + ctx->evt_stop = false; > + > + ctx->stopped = true; > + > + if (!rc) { > + pr_err("dpu wait for stop done time out!\n"); > + return -ETIMEDOUT; > + } > + > + return 0; > +} > + > +static int dpu_wait_update_done(struct dpu_context *ctx) > +{ > + int rc; > + > + ctx->evt_update = false; > + > + rc = wait_event_interruptible_timeout(ctx->wait_queue, ctx->evt_update, > + msecs_to_jiffies(500)); > + > + if (!rc) { > + pr_err("dpu wait for reg update done time out!\n"); > + return -ETIMEDOUT; > + } > + > + return 0; > +} > + > +static void dpu_stop(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + if (ctx->stopped) > + return; > + > + if (ctx->if_type == SPRD_DISPC_IF_DPI) > + reg->dpu_ctrl |= BIT(1); > + > + dpu_wait_stop_done(ctx); > + > + pr_info("dpu stop\n"); > +} > + > +static void dpu_run(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + if (!ctx->stopped) > + return; > + > + reg->dpu_ctrl |= BIT(0); > + > + ctx->stopped = false; > + > + pr_info("dpu run\n"); > +} > + > +static int dpu_init(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + u32 size; > + > + reg->bg_color = 0; > + > + size = (ctx->vm.vactive << 16) | ctx->vm.hactive; > + reg->panel_size = size; > + reg->blend_size = size; > + > + reg->dpu_cfg0 = BIT(4) | BIT(5); > + > + reg->dpu_cfg1 = 0x004466da; > + reg->dpu_cfg2 = 0; > + > + if (ctx->stopped) > + dpu_clean_all(ctx); > + > + reg->mmu_en = 0; > + reg->mmu_min_ppn1 = 0; > + reg->mmu_ppn_range1 = 0xffff; > + reg->mmu_min_ppn2 = 0; > + reg->mmu_ppn_range2 = 0xffff; > + reg->mmu_vpn_range = 0x1ffff; > + > + reg->dpu_int_clr = 0xffff; > + > + init_waitqueue_head(&ctx->wait_queue); > + > + return 0; > +} > + > +static void dpu_fini(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + reg->dpu_int_en = 0; > + reg->dpu_int_clr = 0xff; > +} > + > +enum { > + DPU_LAYER_FORMAT_YUV422_2PLANE, > + DPU_LAYER_FORMAT_YUV420_2PLANE, > + DPU_LAYER_FORMAT_YUV420_3PLANE, > + DPU_LAYER_FORMAT_ARGB8888, > + DPU_LAYER_FORMAT_RGB565, > + DPU_LAYER_FORMAT_XFBC_ARGB8888 = 8, > + DPU_LAYER_FORMAT_XFBC_RGB565, > + DPU_LAYER_FORMAT_MAX_TYPES, > +}; > + > +enum { > + DPU_LAYER_ROTATION_0, > + DPU_LAYER_ROTATION_90, > + DPU_LAYER_ROTATION_180, > + DPU_LAYER_ROTATION_270, > + DPU_LAYER_ROTATION_0_M, > + DPU_LAYER_ROTATION_90_M, > + DPU_LAYER_ROTATION_180_M, > + DPU_LAYER_ROTATION_270_M, > +}; > + > +static u32 to_dpu_rotation(u32 angle) > +{ > + u32 rot = DPU_LAYER_ROTATION_0; > + > + switch (angle) { > + case 0: > + case DRM_MODE_ROTATE_0: > + rot = DPU_LAYER_ROTATION_0; > + break; > + case DRM_MODE_ROTATE_90: > + rot = DPU_LAYER_ROTATION_90; > + break; > + case DRM_MODE_ROTATE_180: > + rot = DPU_LAYER_ROTATION_180; > + break; > + case DRM_MODE_ROTATE_270: > + rot = DPU_LAYER_ROTATION_270; > + break; > + case DRM_MODE_REFLECT_Y: > + rot = DPU_LAYER_ROTATION_180_M; > + break; > + case (DRM_MODE_REFLECT_Y | DRM_MODE_ROTATE_90): > + rot = DPU_LAYER_ROTATION_90_M; > + break; > + case DRM_MODE_REFLECT_X: > + rot = DPU_LAYER_ROTATION_0_M; > + break; > + case (DRM_MODE_REFLECT_X | DRM_MODE_ROTATE_90): > + rot = DPU_LAYER_ROTATION_270_M; > + break; > + default: > + pr_err("rotation convert unsupport angle (drm)= 0x%x\n", angle); > + break; > + } > + > + return rot; > +} > + > +static u32 dpu_img_ctrl(u32 format, u32 blending, u32 rotation) > +{ > + int reg_val = 0; > + > + /* layer enable */ > + reg_val |= BIT(0); > + > + switch (format) { > + case DRM_FORMAT_BGRA8888: > + /* BGRA8888 -> ARGB8888 */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B3B2B1B0 << 8; > + reg_val |= (DPU_LAYER_FORMAT_ARGB8888 << 4); > + break; > + case DRM_FORMAT_RGBX8888: > + case DRM_FORMAT_RGBA8888: > + /* RGBA8888 -> ABGR8888 */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B3B2B1B0 << 8; > + /* FALLTHRU */ > + case DRM_FORMAT_ABGR8888: > + /* rb switch */ > + reg_val |= BIT(10); > + /* FALLTHRU */ > + case DRM_FORMAT_ARGB8888: > + reg_val |= (DPU_LAYER_FORMAT_ARGB8888 << 4); > + break; > + case DRM_FORMAT_XBGR8888: > + /* rb switch */ > + reg_val |= BIT(10); > + /* FALLTHRU */ > + case DRM_FORMAT_XRGB8888: > + reg_val |= (DPU_LAYER_FORMAT_ARGB8888 << 4); > + break; > + case DRM_FORMAT_BGR565: > + /* rb switch */ > + reg_val |= BIT(10); > + /* FALLTHRU */ > + case DRM_FORMAT_RGB565: > + reg_val |= (DPU_LAYER_FORMAT_RGB565 << 4); > + break; > + case DRM_FORMAT_NV12: > + /* 2-Lane: Yuv420 */ > + reg_val |= DPU_LAYER_FORMAT_YUV420_2PLANE << 4; > + /* Y endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 8; > + /* UV endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 10; > + break; > + case DRM_FORMAT_NV21: > + /* 2-Lane: Yuv420 */ > + reg_val |= DPU_LAYER_FORMAT_YUV420_2PLANE << 4; > + /* Y endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 8; > + /* UV endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B3B2B1B0 << 10; > + break; > + case DRM_FORMAT_NV16: > + /* 2-Lane: Yuv422 */ > + reg_val |= DPU_LAYER_FORMAT_YUV422_2PLANE << 4; > + /* Y endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B3B2B1B0 << 8; > + /* UV endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B3B2B1B0 << 10; > + break; > + case DRM_FORMAT_NV61: > + /* 2-Lane: Yuv422 */ > + reg_val |= DPU_LAYER_FORMAT_YUV422_2PLANE << 4; > + /* Y endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 8; > + /* UV endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 10; > + break; > + case DRM_FORMAT_YUV420: > + reg_val |= DPU_LAYER_FORMAT_YUV420_3PLANE << 4; > + /* Y endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 8; > + /* UV endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 10; > + break; > + case DRM_FORMAT_YVU420: > + reg_val |= DPU_LAYER_FORMAT_YUV420_3PLANE << 4; > + /* Y endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B0B1B2B3 << 8; > + /* UV endian */ > + reg_val |= SPRD_IMG_DATA_ENDIAN_B3B2B1B0 << 10; > + break; > + default: > + pr_err("error: invalid format %c%c%c%c\n", format, > + format >> 8, > + format >> 16, > + format >> 24); > + break; > + } > + > + switch (blending) { > + case DRM_MODE_BLEND_PIXEL_NONE: > + /* don't do blending, maybe RGBX */ > + /* alpha mode select - layer alpha */ > + reg_val |= BIT(2); > + break; > + case DRM_MODE_BLEND_COVERAGE: > + /* alpha mode select - combo alpha */ > + reg_val |= BIT(3); > + /*Normal mode*/ > + reg_val &= (~BIT(16)); > + break; > + case DRM_MODE_BLEND_PREMULTI: > + /* alpha mode select - combo alpha */ > + reg_val |= BIT(3); > + /*Pre-mult mode*/ > + reg_val |= BIT(16); > + break; > + default: > + /* alpha mode select - layer alpha */ > + reg_val |= BIT(2); > + break; > + } > + > + rotation = to_dpu_rotation(rotation); > + reg_val |= (rotation & 0x7) << 20; > + > + return reg_val; > +} > + > +static void dpu_bgcolor(struct dpu_context *ctx, u32 color) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + if (ctx->if_type == SPRD_DISPC_IF_EDPI) > + dpu_wait_stop_done(ctx); > + > + reg->bg_color = color; > + > + dpu_clean_all(ctx); > + > + if ((ctx->if_type == SPRD_DISPC_IF_DPI) && !ctx->stopped) { > + reg->dpu_ctrl |= BIT(2); > + dpu_wait_update_done(ctx); > + } else if (ctx->if_type == SPRD_DISPC_IF_EDPI) { > + reg->dpu_ctrl |= BIT(0); > + ctx->stopped = false; > + } > +} > + > +static void dpu_layer(struct dpu_context *ctx, > + struct sprd_dpu_layer *hwlayer) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + const struct drm_format_info *info; > + struct layer_reg *layer; > + u32 addr, size, offset; > + int i; > + > + layer = ®->layers[hwlayer->index]; > + offset = (hwlayer->dst_x & 0xffff) | ((hwlayer->dst_y) << 16); > + > + if (hwlayer->src_w && hwlayer->src_h) > + size = (hwlayer->src_w & 0xffff) | ((hwlayer->src_h) << 16); > + else > + size = (hwlayer->dst_w & 0xffff) | ((hwlayer->dst_h) << 16); > + > + for (i = 0; i < hwlayer->planes; i++) { > + addr = hwlayer->addr[i]; > + > + if (addr % 16) > + pr_err("layer addr[%d] is not 16 bytes align, it's 0x%08x\n", > + i, addr); > + layer->addr[i] = addr; > + } > + > + layer->pos = offset; > + layer->size = size; > + layer->crop_start = (hwlayer->src_y << 16) | hwlayer->src_x; > + layer->alpha = hwlayer->alpha; > + > + info = drm_format_info(hwlayer->format); > + if (info->cpp[0] == 0) { > + pr_err("layer[%d] bytes per pixel is invalid\n", hwlayer->index); > + return; > + } > + > + if (hwlayer->planes == 3) > + /* UV pitch is 1/2 of Y pitch*/ > + layer->pitch = (hwlayer->pitch[0] / info->cpp[0]) | > + (hwlayer->pitch[0] / info->cpp[0] << 15); > + else > + layer->pitch = hwlayer->pitch[0] / info->cpp[0]; > + > + layer->ctrl = dpu_img_ctrl(hwlayer->format, hwlayer->blending, > + hwlayer->rotation); > + > + pr_debug("dst_x = %d, dst_y = %d, dst_w = %d, dst_h = %d\n", > + hwlayer->dst_x, hwlayer->dst_y, > + hwlayer->dst_w, hwlayer->dst_h); > + pr_debug("start_x = %d, start_y = %d, start_w = %d, start_h = %d\n", > + hwlayer->src_x, hwlayer->src_y, > + hwlayer->src_w, hwlayer->src_h); > +} > + > +static void dpu_flip(struct dpu_context *ctx, > + struct sprd_dpu_layer layers[], u8 count) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + int i; > + > + /* > + * Make sure the dpu is in stop status. DPU_R2P0 has no shadow > + * registers in EDPI mode. So the config registers can only be > + * updated in the rising edge of DPU_RUN bit. > + */ > + if (ctx->if_type == SPRD_DISPC_IF_EDPI) > + dpu_wait_stop_done(ctx); > + > + /* reset the bgcolor to black */ > + reg->bg_color = 0; > + > + /* disable all the layers */ > + dpu_clean_all(ctx); > + > + /* start configure dpu layers */ > + for (i = 0; i < count; i++) > + dpu_layer(ctx, &layers[i]); > + > + /* update trigger and wait */ > + if (ctx->if_type == SPRD_DISPC_IF_DPI) { > + if (!ctx->stopped) { > + reg->dpu_ctrl |= BIT(2); > + dpu_wait_update_done(ctx); > + } > + > + reg->dpu_int_en |= DISPC_INT_ERR_MASK; > + > + } else if (ctx->if_type == SPRD_DISPC_IF_EDPI) { > + reg->dpu_ctrl |= BIT(0); > + > + ctx->stopped = false; > + } > + > + /* > + * If the following interrupt was disabled in isr, > + * re-enable it. > + */ > + reg->dpu_int_en |= DISPC_INT_FBC_PLD_ERR_MASK | > + DISPC_INT_FBC_HDR_ERR_MASK | > + DISPC_INT_MMU_VAOR_RD_MASK | > + DISPC_INT_MMU_VAOR_WR_MASK | > + DISPC_INT_MMU_INV_RD_MASK | > + DISPC_INT_MMU_INV_WR_MASK; > +} > + > +static void dpu_dpi_init(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + u32 int_mask = 0; > + > + if (ctx->if_type == SPRD_DISPC_IF_DPI) { > + /* use dpi as interface */ > + reg->dpu_cfg0 &= ~BIT(0); > + > + /* disable Halt function for SPRD DSI */ > + reg->dpi_ctrl &= ~BIT(16); > + > + /* select te from external pad */ > + reg->dpi_ctrl |= BIT(10); > + > + /* set dpi timing */ > + reg->dpi_h_timing = (ctx->vm.hsync_len << 0) | > + (ctx->vm.hback_porch << 8) | > + (ctx->vm.hfront_porch << 20); > + reg->dpi_v_timing = (ctx->vm.vsync_len << 0) | > + (ctx->vm.vback_porch << 8) | > + (ctx->vm.vfront_porch << 20); > + if (ctx->vm.vsync_len + ctx->vm.vback_porch < 32) > + pr_warn("Warning: (vsync + vbp) < 32, " > + "underflow risk!\n"); > + > + /* enable dpu update done INT */ > + int_mask |= DISPC_INT_UPDATE_DONE_MASK; > + /* enable dpu DONE INT */ > + int_mask |= DISPC_INT_DONE_MASK; > + /* enable dpu dpi vsync */ > + int_mask |= DISPC_INT_DPI_VSYNC_MASK; > + /* enable dpu TE INT */ > + int_mask |= DISPC_INT_TE_MASK; > + /* enable underflow err INT */ > + int_mask |= DISPC_INT_ERR_MASK; > + /* enable write back done INT */ > + int_mask |= DISPC_INT_WB_DONE_MASK; > + /* enable write back fail INT */ > + int_mask |= DISPC_INT_WB_FAIL_MASK; > + > + } else if (ctx->if_type == SPRD_DISPC_IF_EDPI) { > + /* use edpi as interface */ > + reg->dpu_cfg0 |= BIT(0); > + > + /* use external te */ > + reg->dpi_ctrl |= BIT(10); > + > + /* enable te */ > + reg->dpi_ctrl |= BIT(8); > + > + /* enable stop DONE INT */ > + int_mask |= DISPC_INT_DONE_MASK; > + /* enable TE INT */ > + int_mask |= DISPC_INT_TE_MASK; > + } > + > + /* enable ifbc payload error INT */ > + int_mask |= DISPC_INT_FBC_PLD_ERR_MASK; > + /* enable ifbc header error INT */ > + int_mask |= DISPC_INT_FBC_HDR_ERR_MASK; > + /* enable iommu va out of range read error INT */ > + int_mask |= DISPC_INT_MMU_VAOR_RD_MASK; > + /* enable iommu va out of range write error INT */ > + int_mask |= DISPC_INT_MMU_VAOR_WR_MASK; > + /* enable iommu invalid read error INT */ > + int_mask |= DISPC_INT_MMU_INV_RD_MASK; > + /* enable iommu invalid write error INT */ > + int_mask |= DISPC_INT_MMU_INV_WR_MASK; > + > + reg->dpu_int_en = int_mask; > +} > + > +static void enable_vsync(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + reg->dpu_int_en |= DISPC_INT_DPI_VSYNC_MASK; > +} > + > +static void disable_vsync(struct dpu_context *ctx) > +{ > + struct dpu_reg *reg = (struct dpu_reg *)ctx->base; > + > + reg->dpu_int_en &= ~DISPC_INT_DPI_VSYNC_MASK; > +} > + > +static const u32 primary_fmts[] = { > + DRM_FORMAT_XRGB8888, DRM_FORMAT_XBGR8888, > + DRM_FORMAT_ARGB8888, DRM_FORMAT_ABGR8888, > + DRM_FORMAT_RGBA8888, DRM_FORMAT_BGRA8888, > + DRM_FORMAT_RGBX8888, DRM_FORMAT_BGRX8888, > + DRM_FORMAT_RGB565, DRM_FORMAT_BGR565, > + DRM_FORMAT_NV12, DRM_FORMAT_NV21, > + DRM_FORMAT_NV16, DRM_FORMAT_NV61, > + DRM_FORMAT_YUV420, DRM_FORMAT_YVU420, > +}; > + > +static int dpu_capability(struct dpu_context *ctx, > + struct dpu_capability *cap) > +{ > + if (!cap) > + return -EINVAL; > + > + cap->max_layers = 6; > + cap->fmts_ptr = primary_fmts; > + cap->fmts_cnt = ARRAY_SIZE(primary_fmts); > + > + return 0; > +} > + > +const struct dpu_core_ops sharkl3_dpu_core_ops = { > + .init = dpu_init, > + .fini = dpu_fini, > + .run = dpu_run, > + .stop = dpu_stop, > + .isr = dpu_isr, > + .ifconfig = dpu_dpi_init, > + .capability = dpu_capability, > + .flip = dpu_flip, > + .bg_color = dpu_bgcolor, > + .enable_vsync = enable_vsync, > + .disable_vsync = disable_vsync, > +}; > diff --git a/drivers/gpu/drm/sprd/sprd_dpu.c b/drivers/gpu/drm/sprd/sprd_dpu.c > new file mode 100644 > index 0000000..f122b0e > --- /dev/null > +++ b/drivers/gpu/drm/sprd/sprd_dpu.c > @@ -0,0 +1,586 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (C) 2020 Unisoc Inc. > + */ > + > +#include > +#include > +#include > +#include > +#include > +#include > +#include > + > +#include > +#include > +#include > +#include > +#include > +#include > + > +#include "sprd_drm.h" > +#include "sprd_dpu.h" > + > +struct sprd_plane { > + struct drm_plane plane; > + u32 index; > +}; > + > +static int sprd_dpu_init(struct sprd_dpu *dpu); > +static int sprd_dpu_fini(struct sprd_dpu *dpu); > + > +static inline struct sprd_plane *to_sprd_plane(struct drm_plane *plane) > +{ > + return container_of(plane, struct sprd_plane, plane); > +} > + > +static int sprd_plane_atomic_check(struct drm_plane *plane, > + struct drm_plane_state *state) > +{ > + DRM_DEBUG("%s()\n", __func__); > + > + return 0; > +} > + > +static void sprd_plane_atomic_update(struct drm_plane *plane, > + struct drm_plane_state *old_state) > +{ > + struct drm_plane_state *state = plane->state; > + struct drm_framebuffer *fb = plane->state->fb; > + struct drm_gem_cma_object *cma_obj; > + struct sprd_plane *p = to_sprd_plane(plane); > + struct sprd_dpu *dpu = crtc_to_dpu(plane->state->crtc); > + struct sprd_dpu_layer *layer = &dpu->layers[p->index]; > + int i; > + > + if (plane->state->crtc->state->active_changed) { > + DRM_DEBUG("resume or suspend, no need to update plane\n"); > + return; > + } > + > + layer->index = p->index; > + layer->src_x = state->src_x >> 16; > + layer->src_y = state->src_y >> 16; > + layer->src_w = state->src_w >> 16; > + layer->src_h = state->src_h >> 16; > + layer->dst_x = state->crtc_x; > + layer->dst_y = state->crtc_y; > + layer->dst_w = state->crtc_w; > + layer->dst_h = state->crtc_h; > + layer->alpha = state->alpha; > + layer->blending = state->pixel_blend_mode; > + layer->rotation = state->rotation; > + layer->planes = fb->format->num_planes; > + layer->format = fb->format->format; > + > + DRM_DEBUG("%s() alpha = %u, blending = %u, rotation = %u\n", > + __func__, layer->alpha, layer->blending, layer->rotation); > + > + for (i = 0; i < layer->planes; i++) { > + cma_obj = drm_fb_cma_get_gem_obj(fb, i); > + layer->addr[i] = cma_obj->paddr + fb->offsets[i]; > + layer->pitch[i] = fb->pitches[i]; > + } > + > + dpu->pending_planes++; > +} > + > +static void sprd_plane_atomic_disable(struct drm_plane *plane, > + struct drm_plane_state *old_state) > +{ > + struct sprd_plane *p = to_sprd_plane(plane); > + > + /* > + * NOTE: > + * The dpu->core->flip() will disable all the planes each time. > + * So there is no need to impliment the atomic_disable() function. > + * But this function can not be removed, because it will change > + * to call atomic_update() callback instead. Which will cause > + * kernel panic in sprd_plane_atomic_update(). > + * > + * We do nothing here but just print a debug log. > + */ > + DRM_DEBUG("%s() layer_id = %u\n", __func__, p->index); > +} > + > +static int sprd_plane_create_properties(struct sprd_plane *p, int index) > +{ > + unsigned int supported_modes = BIT(DRM_MODE_BLEND_PIXEL_NONE) | > + BIT(DRM_MODE_BLEND_PREMULTI) | > + BIT(DRM_MODE_BLEND_COVERAGE); > + > + /* create rotation property */ > + drm_plane_create_rotation_property(&p->plane, > + DRM_MODE_ROTATE_0, > + DRM_MODE_ROTATE_MASK | > + DRM_MODE_REFLECT_MASK); > + > + /* create alpha property */ > + drm_plane_create_alpha_property(&p->plane); > + > + /* create blend mode property */ > + drm_plane_create_blend_mode_property(&p->plane, supported_modes); > + > + /* create zpos property */ > + drm_plane_create_zpos_immutable_property(&p->plane, index); > + > + return 0; > +} > + > +static const struct drm_plane_helper_funcs sprd_plane_helper_funcs = { > + .atomic_check = sprd_plane_atomic_check, > + .atomic_update = sprd_plane_atomic_update, > + .atomic_disable = sprd_plane_atomic_disable, > +}; > + > +static const struct drm_plane_funcs sprd_plane_funcs = { > + .update_plane = drm_atomic_helper_update_plane, > + .disable_plane = drm_atomic_helper_disable_plane, > + .destroy = drm_plane_cleanup, > + .reset = drm_atomic_helper_plane_reset, > + .atomic_duplicate_state = drm_atomic_helper_plane_duplicate_state, > + .atomic_destroy_state = drm_atomic_helper_plane_destroy_state, > +}; > + > +static struct drm_plane *sprd_plane_init(struct drm_device *drm, > + struct sprd_dpu *dpu) > +{ > + struct drm_plane *primary = NULL; > + struct sprd_plane *p = NULL; > + struct dpu_capability cap = {}; > + int err, i; > + > + if (dpu->core && dpu->core->capability) > + dpu->core->capability(&dpu->ctx, &cap); > + > + dpu->layers = devm_kcalloc(drm->dev, cap.max_layers, > + sizeof(struct sprd_dpu_layer), GFP_KERNEL); > + if (!dpu->layers) > + return ERR_PTR(-ENOMEM); > + > + for (i = 0; i < cap.max_layers; i++) { > + > + p = devm_kzalloc(drm->dev, sizeof(*p), GFP_KERNEL); > + if (!p) > + return ERR_PTR(-ENOMEM); > + > + err = drm_universal_plane_init(drm, &p->plane, 1, > + &sprd_plane_funcs, cap.fmts_ptr, > + cap.fmts_cnt, NULL, > + DRM_PLANE_TYPE_PRIMARY, NULL); > + if (err) { > + DRM_ERROR("fail to init primary plane\n"); > + return ERR_PTR(err); > + } > + > + drm_plane_helper_add(&p->plane, &sprd_plane_helper_funcs); > + > + sprd_plane_create_properties(p, i); > + > + p->index = i; > + if (i == 0) > + primary = &p->plane; > + } > + > + if (p) > + DRM_INFO("dpu plane init ok\n"); > + > + return primary; > +} > + > +static void sprd_crtc_mode_set_nofb(struct drm_crtc *crtc) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + > + if ((dpu->mode->hdisplay == dpu->mode->htotal) || > + (dpu->mode->vdisplay == dpu->mode->vtotal)) > + dpu->ctx.if_type = SPRD_DISPC_IF_EDPI; > + else > + dpu->ctx.if_type = SPRD_DISPC_IF_DPI; > +} > + > +static enum drm_mode_status sprd_crtc_mode_valid(struct drm_crtc *crtc, > + const struct drm_display_mode *mode) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + > + DRM_INFO("%s() mode: "DRM_MODE_FMT"\n", __func__, DRM_MODE_ARG(mode)); > + > + if (mode->type & DRM_MODE_TYPE_DEFAULT) > + dpu->mode = (struct drm_display_mode *)mode; > + > + if (mode->type & DRM_MODE_TYPE_PREFERRED) { > + dpu->mode = (struct drm_display_mode *)mode; > + drm_display_mode_to_videomode(dpu->mode, &dpu->ctx.vm); > + } > + > + return MODE_OK; > +} > + > +static void sprd_crtc_atomic_enable(struct drm_crtc *crtc, > + struct drm_crtc_state *old_state) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + > + DRM_INFO("%s()\n", __func__); > + > + sprd_dpu_init(dpu); > + > + enable_irq(dpu->ctx.irq); > + > + drm_crtc_vblank_on(crtc); > +} > + > +static void sprd_crtc_atomic_disable(struct drm_crtc *crtc, > + struct drm_crtc_state *old_state) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + struct drm_device *drm = dpu->crtc.dev; > + > + DRM_INFO("%s()\n", __func__); > + > + drm_crtc_vblank_off(crtc); > + > + disable_irq(dpu->ctx.irq); > + > + sprd_dpu_fini(dpu); > + > + spin_lock_irq(&drm->event_lock); > + if (crtc->state->event) { > + drm_crtc_send_vblank_event(crtc, crtc->state->event); > + crtc->state->event = NULL; > + } > + spin_unlock_irq(&drm->event_lock); > +} > + > +static int sprd_crtc_atomic_check(struct drm_crtc *crtc, > + struct drm_crtc_state *state) > +{ > + DRM_DEBUG("%s()\n", __func__); > + > + return 0; > +} > + > +static void sprd_crtc_atomic_begin(struct drm_crtc *crtc, > + struct drm_crtc_state *old_state) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + > + DRM_DEBUG("%s()\n", __func__); > + > + memset(dpu->layers, 0, sizeof(*dpu->layers) * dpu->pending_planes); > + > + dpu->pending_planes = 0; > +} > + > +static void sprd_crtc_atomic_flush(struct drm_crtc *crtc, > + struct drm_crtc_state *old_state) > + > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + struct drm_device *drm = dpu->crtc.dev; > + > + DRM_DEBUG("%s()\n", __func__); > + > + if (dpu->core && dpu->core->flip && dpu->pending_planes) > + dpu->core->flip(&dpu->ctx, dpu->layers, dpu->pending_planes); > + > + spin_lock_irq(&drm->event_lock); > + if (crtc->state->event) { > + drm_crtc_send_vblank_event(crtc, crtc->state->event); > + crtc->state->event = NULL; > + } > + spin_unlock_irq(&drm->event_lock); > +} > + > +static int sprd_crtc_enable_vblank(struct drm_crtc *crtc) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + > + DRM_DEBUG("%s()\n", __func__); > + > + if (dpu->core && dpu->core->enable_vsync) > + dpu->core->enable_vsync(&dpu->ctx); > + > + return 0; > +} > + > +static void sprd_crtc_disable_vblank(struct drm_crtc *crtc) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + > + DRM_DEBUG("%s()\n", __func__); > + > + if (dpu->core && dpu->core->disable_vsync) > + dpu->core->disable_vsync(&dpu->ctx); > +} > + > +static int sprd_crtc_create_properties(struct drm_crtc *crtc) > +{ > + struct sprd_dpu *dpu = crtc_to_dpu(crtc); > + struct drm_device *drm = dpu->crtc.dev; > + struct drm_property *prop; > + struct drm_property_blob *blob; > + size_t blob_size; > + > + blob_size = strlen(dpu->ctx.version) + 1; > + > + blob = drm_property_create_blob(dpu->crtc.dev, blob_size, > + dpu->ctx.version); > + if (IS_ERR(blob)) { > + DRM_ERROR("drm_property_create_blob dpu version failed\n"); > + return PTR_ERR(blob); > + } > + > + /* create dpu version property */ > + prop = drm_property_create(drm, > + DRM_MODE_PROP_IMMUTABLE | DRM_MODE_PROP_BLOB, > + "dpu version", 0); > + if (!prop) { > + DRM_ERROR("drm_property_create dpu version failed\n"); > + return -ENOMEM; > + } > + drm_object_attach_property(&crtc->base, prop, blob->base.id); > + > + return 0; > +} > + > +static const struct drm_crtc_helper_funcs sprd_crtc_helper_funcs = { > + .mode_set_nofb = sprd_crtc_mode_set_nofb, > + .mode_valid = sprd_crtc_mode_valid, > + .atomic_check = sprd_crtc_atomic_check, > + .atomic_begin = sprd_crtc_atomic_begin, > + .atomic_flush = sprd_crtc_atomic_flush, > + .atomic_enable = sprd_crtc_atomic_enable, > + .atomic_disable = sprd_crtc_atomic_disable, > +}; > + > +static const struct drm_crtc_funcs sprd_crtc_funcs = { > + .destroy = drm_crtc_cleanup, > + .set_config = drm_atomic_helper_set_config, > + .page_flip = drm_atomic_helper_page_flip, > + .reset = drm_atomic_helper_crtc_reset, > + .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state, > + .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state, > + .enable_vblank = sprd_crtc_enable_vblank, > + .disable_vblank = sprd_crtc_disable_vblank, > +}; > + > +static int sprd_crtc_init(struct drm_device *drm, struct drm_crtc *crtc, > + struct drm_plane *primary) > +{ > + struct device_node *port; > + int err; > + > + /* > + * set crtc port so that drm_of_find_possible_crtcs call works > + */ > + port = of_parse_phandle(drm->dev->of_node, "ports", 0); > + if (!port) { > + DRM_ERROR("find 'ports' phandle of %s failed\n", > + drm->dev->of_node->full_name); > + return -EINVAL; > + } > + of_node_put(port); > + crtc->port = port; > + > + err = drm_crtc_init_with_planes(drm, crtc, primary, NULL, > + &sprd_crtc_funcs, NULL); > + if (err) { > + DRM_ERROR("failed to init crtc.\n"); > + return err; > + } > + > + drm_mode_crtc_set_gamma_size(crtc, 256); > + > + drm_crtc_helper_add(crtc, &sprd_crtc_helper_funcs); > + > + sprd_crtc_create_properties(crtc); > + > + DRM_INFO("%s() ok\n", __func__); > + return 0; > +} > + > +static int sprd_dpu_init(struct sprd_dpu *dpu) > +{ > + struct dpu_context *ctx = &dpu->ctx; > + > + if (dpu->core && dpu->core->init) > + dpu->core->init(ctx); > + if (dpu->core && dpu->core->ifconfig) > + dpu->core->ifconfig(ctx); > + > + return 0; > +} > + > +static int sprd_dpu_fini(struct sprd_dpu *dpu) > +{ > + struct dpu_context *ctx = &dpu->ctx; > + > + if (dpu->core && dpu->core->fini) > + dpu->core->fini(ctx); > + > + return 0; > +} > + > +static irqreturn_t sprd_dpu_isr(int irq, void *data) > +{ > + struct sprd_dpu *dpu = data; > + struct dpu_context *ctx = &dpu->ctx; > + u32 int_mask = 0; > + > + if (dpu->core && dpu->core->isr) > + int_mask = dpu->core->isr(ctx); > + > + if (int_mask & DISPC_INT_ERR_MASK) > + DRM_WARN("Warning: dpu underflow!\n"); > + > + if ((int_mask & DISPC_INT_DPI_VSYNC_MASK)) > + drm_crtc_handle_vblank(&dpu->crtc); > + > + return IRQ_HANDLED; > +} > + > +static int sprd_dpu_irq_request(struct sprd_dpu *dpu) > +{ > + int err; > + int irq_num; > + > + irq_num = irq_of_parse_and_map(dpu->dev.of_node, 0); Use platform_get_irq instead. > + if (!irq_num) { > + DRM_ERROR("error: dpu parse irq num failed\n"); > + return -EINVAL; > + } > + DRM_INFO("dpu irq_num = %d\n", irq_num); > + > + irq_set_status_flags(irq_num, IRQ_NOAUTOEN); I think you shouldn't need this. Make sure you've disabled interrupts in the h/w first or that the handler can handle it if you haven't. > + err = devm_request_irq(&dpu->dev, irq_num, sprd_dpu_isr, > + 0, "DISPC", dpu); > + if (err) { > + DRM_ERROR("error: dpu request irq failed\n"); > + return -EINVAL; > + } > + dpu->ctx.irq = irq_num; > + > + return 0; > +} > + > +static int sprd_dpu_bind(struct device *dev, struct device *master, void *data) > +{ > + struct drm_device *drm = data; > + struct sprd_dpu *dpu = dev_get_drvdata(dev); > + struct drm_plane *plane; > + int err; > + > + DRM_INFO("%s()\n", __func__); > + > + plane = sprd_plane_init(drm, dpu); > + if (IS_ERR_OR_NULL(plane)) { > + err = PTR_ERR(plane); > + return err; > + } > + > + err = sprd_crtc_init(drm, &dpu->crtc, plane); > + if (err) > + return err; > + > + sprd_dpu_irq_request(dpu); > + > + return 0; > +} > + > +static void sprd_dpu_unbind(struct device *dev, struct device *master, > + void *data) > +{ > + struct sprd_dpu *dpu = dev_get_drvdata(dev); > + > + DRM_INFO("%s()\n", __func__); > + > + drm_crtc_cleanup(&dpu->crtc); > +} > + > +static const struct component_ops dpu_component_ops = { > + .bind = sprd_dpu_bind, > + .unbind = sprd_dpu_unbind, > +}; > + > +static int sprd_dpu_context_init(struct sprd_dpu *dpu, > + struct device_node *np) > +{ > + struct resource r; > + struct dpu_context *ctx = &dpu->ctx; > + > + if (of_address_to_resource(np, 0, &r)) { > + DRM_ERROR("parse dt base address failed\n"); > + return -ENODEV; > + } > + ctx->base = (unsigned long)ioremap(r.start, > + resource_size(&r)); > + if (ctx->base == 0) { > + DRM_ERROR("ioremap base address failed\n"); > + return -EFAULT; > + } > + > + return 0; > +} > + > +static const struct sprd_dpu_ops sharkl3_dpu = { > + .core = &sharkl3_dpu_core_ops, > +}; > + > +static const struct of_device_id dpu_match_table[] = { > + { .compatible = "sprd,sharkl3-dpu", > + .data = &sharkl3_dpu }, > + { /* sentinel */ }, > +}; > + > +static int sprd_dpu_probe(struct platform_device *pdev) > +{ > + struct device_node *np = pdev->dev.of_node; > + const struct of_device_id *of_id = > + of_match_node(dpu_match_table, np); Don't need this. > + const struct sprd_dpu_ops *pdata; > + struct sprd_dpu *dpu; > + int ret; > + > + dpu = devm_kzalloc(&pdev->dev, sizeof(*dpu), GFP_KERNEL); > + if (!dpu) > + return -ENOMEM; > + > + pdata = of_device_get_match_data(&pdev->dev); > + if (pdata) { > + dpu->core = pdata->core; > + dpu->ctx.version = "dpu-r2p0"; > + } else { > + DRM_ERROR("Can't get %s ops data\n", of_id->name); > + return -EINVAL; > + } > + > + ret = sprd_dpu_context_init(dpu, np); > + if (ret) > + return ret; > + > + platform_set_drvdata(pdev, dpu); > + > + return component_add(&pdev->dev, &dpu_component_ops); > +} > + > +static int sprd_dpu_remove(struct platform_device *pdev) > +{ > + component_del(&pdev->dev, &dpu_component_ops); > + return 0; > +} > + > +struct platform_driver sprd_dpu_driver = { > + .probe = sprd_dpu_probe, > + .remove = sprd_dpu_remove, > + .driver = { > + .name = "sprd-dpu-drv", > + .of_match_table = dpu_match_table, > + }, > +}; > + > +MODULE_AUTHOR("Leon He "); > +MODULE_AUTHOR("Kevin Tang "); > +MODULE_DESCRIPTION("Unisoc Display Controller Driver"); > +MODULE_LICENSE("GPL v2"); > diff --git a/drivers/gpu/drm/sprd/sprd_dpu.h b/drivers/gpu/drm/sprd/sprd_dpu.h > new file mode 100644 > index 0000000..3e7f91f > --- /dev/null > +++ b/drivers/gpu/drm/sprd/sprd_dpu.h > @@ -0,0 +1,127 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (C) 2020 Unisoc Inc. > + */ > + > +#ifndef __SPRD_DPU_H__ > +#define __SPRD_DPU_H__ > + > +#include > +#include > +#include > +#include > +#include > +#include > +#include