From: Wen He <[email protected]>
This patch implement a standard macro call functions is
used to NXP dma drivers.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
---
change in v10:
- no
drivers/dma/fsldma.c | 16 ++++++++--------
drivers/dma/fsldma.h | 4 ++--
2 files changed, 10 insertions(+), 10 deletions(-)
diff --git a/drivers/dma/fsldma.c b/drivers/dma/fsldma.c
index 1117b51..39871e0 100644
--- a/drivers/dma/fsldma.c
+++ b/drivers/dma/fsldma.c
@@ -53,42 +53,42 @@
static void set_sr(struct fsldma_chan *chan, u32 val)
{
- DMA_OUT(chan, &chan->regs->sr, val, 32);
+ FSL_DMA_OUT(chan, &chan->regs->sr, val, 32);
}
static u32 get_sr(struct fsldma_chan *chan)
{
- return DMA_IN(chan, &chan->regs->sr, 32);
+ return FSL_DMA_IN(chan, &chan->regs->sr, 32);
}
static void set_mr(struct fsldma_chan *chan, u32 val)
{
- DMA_OUT(chan, &chan->regs->mr, val, 32);
+ FSL_DMA_OUT(chan, &chan->regs->mr, val, 32);
}
static u32 get_mr(struct fsldma_chan *chan)
{
- return DMA_IN(chan, &chan->regs->mr, 32);
+ return FSL_DMA_IN(chan, &chan->regs->mr, 32);
}
static void set_cdar(struct fsldma_chan *chan, dma_addr_t addr)
{
- DMA_OUT(chan, &chan->regs->cdar, addr | FSL_DMA_SNEN, 64);
+ FSL_DMA_OUT(chan, &chan->regs->cdar, addr | FSL_DMA_SNEN, 64);
}
static dma_addr_t get_cdar(struct fsldma_chan *chan)
{
- return DMA_IN(chan, &chan->regs->cdar, 64) & ~FSL_DMA_SNEN;
+ return FSL_DMA_IN(chan, &chan->regs->cdar, 64) & ~FSL_DMA_SNEN;
}
static void set_bcr(struct fsldma_chan *chan, u32 val)
{
- DMA_OUT(chan, &chan->regs->bcr, val, 32);
+ FSL_DMA_OUT(chan, &chan->regs->bcr, val, 32);
}
static u32 get_bcr(struct fsldma_chan *chan)
{
- return DMA_IN(chan, &chan->regs->bcr, 32);
+ return FSL_DMA_IN(chan, &chan->regs->bcr, 32);
}
/*
diff --git a/drivers/dma/fsldma.h b/drivers/dma/fsldma.h
index 4787d48..982845b 100644
--- a/drivers/dma/fsldma.h
+++ b/drivers/dma/fsldma.h
@@ -223,10 +223,10 @@ static void out_le64(u64 __iomem *addr, u64 val)
}
#endif
-#define DMA_IN(fsl_chan, addr, width) \
+#define FSL_DMA_IN(fsl_chan, addr, width) \
(((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ? \
in_be##width(addr) : in_le##width(addr))
-#define DMA_OUT(fsl_chan, addr, val, width) \
+#define FSL_DMA_OUT(fsl_chan, addr, val, width) \
(((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ? \
out_be##width(addr, val) : out_le##width(addr, val))
--
1.7.1
This patch add the macro FSL_DMA_IN/OUT implement for ARM platform.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
---
change in v10:
- fixed compile warning on powerpc
drivers/dma/fsldma.h | 61 ++++++++++++++++++++++++++++++++++---------------
1 files changed, 42 insertions(+), 19 deletions(-)
diff --git a/drivers/dma/fsldma.h b/drivers/dma/fsldma.h
index 982845b..88db939 100644
--- a/drivers/dma/fsldma.h
+++ b/drivers/dma/fsldma.h
@@ -196,39 +196,62 @@ struct fsldma_chan {
#define to_fsl_desc(lh) container_of(lh, struct fsl_desc_sw, node)
#define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx)
+#ifdef CONFIG_PPC
+#define fsl_ioread32(p) in_le32(p)
+#define fsl_ioread32be(p) in_be32(p)
+#define fsl_iowrite32(v, p) out_le32(p, v)
+#define fsl_iowrite32be(v, p) out_be32(p, v)
+
#ifndef __powerpc64__
-static u64 in_be64(const u64 __iomem *addr)
+static u64 fsl_ioread64(const u64 __iomem *addr)
{
- return ((u64)in_be32((u32 __iomem *)addr) << 32) |
- (in_be32((u32 __iomem *)addr + 1));
+ u32 fsl_addr = lower_32_bits(addr);
+ u64 fsl_addr_hi = (u64)in_le32((u32 *)(fsl_addr + 1)) << 32;
+
+ return fsl_addr_hi | in_le32((u32 *)fsl_addr);
}
-static void out_be64(u64 __iomem *addr, u64 val)
+static void fsl_iowrite64(u64 val, u64 __iomem *addr)
{
- out_be32((u32 __iomem *)addr, val >> 32);
- out_be32((u32 __iomem *)addr + 1, (u32)val);
+ out_le32((u32 __iomem *)addr + 1, val >> 32);
+ out_le32((u32 __iomem *)addr, (u32)val);
}
-/* There is no asm instructions for 64 bits reverse loads and stores */
-static u64 in_le64(const u64 __iomem *addr)
+static u64 fsl_ioread64be(const u64 __iomem *addr)
{
- return ((u64)in_le32((u32 __iomem *)addr + 1) << 32) |
- (in_le32((u32 __iomem *)addr));
+ u32 fsl_addr = lower_32_bits(addr);
+ u64 fsl_addr_hi = (u64)in_be32((u32 *)fsl_addr) << 32;
+
+ return fsl_addr_hi | in_be32((u32 *)(fsl_addr + 1));
}
-static void out_le64(u64 __iomem *addr, u64 val)
+static void fsl_iowrite64be(u64 val, u64 __iomem *addr)
{
- out_le32((u32 __iomem *)addr + 1, val >> 32);
- out_le32((u32 __iomem *)addr, (u32)val);
+ out_be32((u32 __iomem *)addr, val >> 32);
+ out_be32((u32 __iomem *)addr + 1, (u32)val);
}
#endif
+#endif
-#define FSL_DMA_IN(fsl_chan, addr, width) \
- (((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ? \
- in_be##width(addr) : in_le##width(addr))
-#define FSL_DMA_OUT(fsl_chan, addr, val, width) \
- (((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ? \
- out_be##width(addr, val) : out_le##width(addr, val))
+#if defined(CONFIG_ARM64) || defined(CONFIG_ARM)
+#define fsl_ioread32(p) ioread32(p)
+#define fsl_ioread32be(p) ioread32be(p)
+#define fsl_iowrite32(v, p) iowrite32(v, p)
+#define fsl_iowrite32be(v, p) iowrite32be(v, p)
+#define fsl_ioread64(p) ioread64(p)
+#define fsl_ioread64be(p) ioread64be(p)
+#define fsl_iowrite64(v, p) iowrite64(v, p)
+#define fsl_iowrite64be(v, p) iowrite64be(v, p)
+#endif
+
+#define FSL_DMA_IN(fsl_dma, addr, width) \
+ (((fsl_dma)->feature & FSL_DMA_BIG_ENDIAN) ? \
+ fsl_ioread##width##be(addr) : fsl_ioread##width(addr))
+
+#define FSL_DMA_OUT(fsl_dma, addr, val, width) \
+ (((fsl_dma)->feature & FSL_DMA_BIG_ENDIAN) ? \
+ fsl_iowrite##width##be(val, addr) : fsl_iowrite \
+ ##width(val, addr))
#define DMA_TO_CPU(fsl_chan, d, width) \
(((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ? \
--
1.7.1
Document the devicetree bindings for NXP Layerscape qDMA controller
which could be found on NXP QorIQ Layerscape SoCs.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
Reviewed-by: Rob Herring <[email protected]>
---
change in v10:
- no
Documentation/devicetree/bindings/dma/fsl-qdma.txt | 57 ++++++++++++++++++++
1 files changed, 57 insertions(+), 0 deletions(-)
create mode 100644 Documentation/devicetree/bindings/dma/fsl-qdma.txt
diff --git a/Documentation/devicetree/bindings/dma/fsl-qdma.txt b/Documentation/devicetree/bindings/dma/fsl-qdma.txt
new file mode 100644
index 0000000..6a0ff90
--- /dev/null
+++ b/Documentation/devicetree/bindings/dma/fsl-qdma.txt
@@ -0,0 +1,57 @@
+NXP Layerscape SoC qDMA Controller
+==================================
+
+This device follows the generic DMA bindings defined in dma/dma.txt.
+
+Required properties:
+
+- compatible: Must be one of
+ "fsl,ls1021a-qdma": for LS1021A Board
+ "fsl,ls1043a-qdma": for ls1043A Board
+ "fsl,ls1046a-qdma": for ls1046A Board
+- reg: Should contain the register's base address and length.
+- interrupts: Should contain a reference to the interrupt used by this
+ device.
+- interrupt-names: Should contain interrupt names:
+ "qdma-queue0": the block0 interrupt
+ "qdma-queue1": the block1 interrupt
+ "qdma-queue2": the block2 interrupt
+ "qdma-queue3": the block3 interrupt
+ "qdma-error": the error interrupt
+- fsl,dma-queues: Should contain number of queues supported.
+- dma-channels: Number of DMA channels supported
+- block-number: the virtual block number
+- block-offset: the offset of different virtual block
+- status-sizes: status queue size of per virtual block
+- queue-sizes: command queue size of per virtual block, the size number
+ based on queues
+
+Optional properties:
+
+- dma-channels: Number of DMA channels supported by the controller.
+- big-endian: If present registers and hardware scatter/gather descriptors
+ of the qDMA are implemented in big endian mode, otherwise in little
+ mode.
+
+Examples:
+
+ qdma: dma-controller@8390000 {
+ compatible = "fsl,ls1021a-qdma";
+ reg = <0x0 0x8388000 0x0 0x1000>, /* Controller regs */
+ <0x0 0x8389000 0x0 0x1000>, /* Status regs */
+ <0x0 0x838a000 0x0 0x2000>; /* Block regs */
+ interrupts = <GIC_SPI 185 IRQ_TYPE_LEVEL_HIGH>,
+ <GIC_SPI 76 IRQ_TYPE_LEVEL_HIGH>,
+ <GIC_SPI 77 IRQ_TYPE_LEVEL_HIGH>;
+ interrupt-names = "qdma-error",
+ "qdma-queue0", "qdma-queue1";
+ dma-channels = <8>;
+ block-number = <2>;
+ block-offset = <0x1000>;
+ fsl,dma-queues = <2>;
+ status-sizes = <64>;
+ queue-sizes = <64 64>;
+ big-endian;
+ };
+
+DMA clients must use the format described in dma/dma.txt file.
--
1.7.1
NXP Queue DMA controller(qDMA) on Layerscape SoCs supports channel
virtuallization by allowing DMA jobs to be enqueued into different
command queues.
Note that this module depends on NXP DPAA.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Jiaheng Fan <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
---
change in v10:
- no
drivers/dma/Kconfig | 13 +
drivers/dma/Makefile | 1 +
drivers/dma/fsl-qdma.c | 1257 ++++++++++++++++++++++++++++++++++++++++++++++++
3 files changed, 1271 insertions(+), 0 deletions(-)
create mode 100644 drivers/dma/fsl-qdma.c
diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
index dacf3f4..50e19d7 100644
--- a/drivers/dma/Kconfig
+++ b/drivers/dma/Kconfig
@@ -218,6 +218,19 @@ config FSL_EDMA
multiplexing capability for DMA request sources(slot).
This module can be found on Freescale Vybrid and LS-1 SoCs.
+config FSL_QDMA
+ tristate "NXP Layerscape qDMA engine support"
+ depends on ARM || ARM64
+ select DMA_ENGINE
+ select DMA_VIRTUAL_CHANNELS
+ select DMA_ENGINE_RAID
+ select ASYNC_TX_ENABLE_CHANNEL_SWITCH
+ help
+ Support the NXP Layerscape qDMA engine with command queue and legacy mode.
+ Channel virtualization is supported through enqueuing of DMA jobs to,
+ or dequeuing DMA jobs from, different work queues.
+ This module can be found on NXP Layerscape SoCs.
+
config FSL_RAID
tristate "Freescale RAID engine Support"
depends on FSL_SOC && !ASYNC_TX_ENABLE_CHANNEL_SWITCH
diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
index c91702d..2d1b586 100644
--- a/drivers/dma/Makefile
+++ b/drivers/dma/Makefile
@@ -32,6 +32,7 @@ obj-$(CONFIG_DW_DMAC_CORE) += dw/
obj-$(CONFIG_EP93XX_DMA) += ep93xx_dma.o
obj-$(CONFIG_FSL_DMA) += fsldma.o
obj-$(CONFIG_FSL_EDMA) += fsl-edma.o
+obj-$(CONFIG_FSL_QDMA) += fsl-qdma.o
obj-$(CONFIG_FSL_RAID) += fsl_raid.o
obj-$(CONFIG_HSU_DMA) += hsu/
obj-$(CONFIG_IMG_MDC_DMA) += img-mdc-dma.o
diff --git a/drivers/dma/fsl-qdma.c b/drivers/dma/fsl-qdma.c
new file mode 100644
index 0000000..404869e
--- /dev/null
+++ b/drivers/dma/fsl-qdma.c
@@ -0,0 +1,1257 @@
+// SPDX-License-Identifier: GPL-2.0
+// Copyright 2018 NXP
+
+/*
+ * Driver for NXP Layerscape Queue Direct Memory Access Controller
+ *
+ * Author:
+ * Wen He <[email protected]>
+ * Jiaheng Fan <[email protected]>
+ *
+ */
+
+#include <linux/module.h>
+#include <linux/delay.h>
+#include <linux/of_irq.h>
+#include <linux/of_platform.h>
+#include <linux/of_dma.h>
+#include <linux/dma-mapping.h>
+
+#include "virt-dma.h"
+#include "fsldma.h"
+
+/* Register related definition */
+#define FSL_QDMA_DMR 0x0
+#define FSL_QDMA_DSR 0x4
+#define FSL_QDMA_DEIER 0xe00
+#define FSL_QDMA_DEDR 0xe04
+#define FSL_QDMA_DECFDW0R 0xe10
+#define FSL_QDMA_DECFDW1R 0xe14
+#define FSL_QDMA_DECFDW2R 0xe18
+#define FSL_QDMA_DECFDW3R 0xe1c
+#define FSL_QDMA_DECFQIDR 0xe30
+#define FSL_QDMA_DECBR 0xe34
+
+#define FSL_QDMA_BCQMR(x) (0xc0 + 0x100 * (x))
+#define FSL_QDMA_BCQSR(x) (0xc4 + 0x100 * (x))
+#define FSL_QDMA_BCQEDPA_SADDR(x) (0xc8 + 0x100 * (x))
+#define FSL_QDMA_BCQDPA_SADDR(x) (0xcc + 0x100 * (x))
+#define FSL_QDMA_BCQEEPA_SADDR(x) (0xd0 + 0x100 * (x))
+#define FSL_QDMA_BCQEPA_SADDR(x) (0xd4 + 0x100 * (x))
+#define FSL_QDMA_BCQIER(x) (0xe0 + 0x100 * (x))
+#define FSL_QDMA_BCQIDR(x) (0xe4 + 0x100 * (x))
+
+#define FSL_QDMA_SQDPAR 0x80c
+#define FSL_QDMA_SQEPAR 0x814
+#define FSL_QDMA_BSQMR 0x800
+#define FSL_QDMA_BSQSR 0x804
+#define FSL_QDMA_BSQICR 0x828
+#define FSL_QDMA_CQMR 0xa00
+#define FSL_QDMA_CQDSCR1 0xa08
+#define FSL_QDMA_CQDSCR2 0xa0c
+#define FSL_QDMA_CQIER 0xa10
+#define FSL_QDMA_CQEDR 0xa14
+#define FSL_QDMA_SQCCMR 0xa20
+
+/* Registers for bit and genmask */
+#define FSL_QDMA_CQIDR_SQT BIT(15)
+#define QDMA_CCDF_FOTMAT BIT(29)
+#define QDMA_CCDF_SER BIT(30)
+#define QDMA_SG_FIN BIT(30)
+#define QDMA_SG_LEN_MASK GENMASK(29, 0)
+#define QDMA_CCDF_MASK GENMASK(28, 20)
+
+#define FSL_QDMA_DEDR_CLEAR GENMASK(31, 0)
+#define FSL_QDMA_BCQIDR_CLEAR GENMASK(31, 0)
+#define FSL_QDMA_DEIER_CLEAR GENMASK(31, 0)
+
+#define FSL_QDMA_BCQIER_CQTIE BIT(15)
+#define FSL_QDMA_BCQIER_CQPEIE BIT(23)
+#define FSL_QDMA_BSQICR_ICEN BIT(31)
+
+#define FSL_QDMA_BSQICR_ICST(x) ((x) << 16)
+#define FSL_QDMA_CQIER_MEIE BIT(31)
+#define FSL_QDMA_CQIER_TEIE BIT(0)
+#define FSL_QDMA_SQCCMR_ENTER_WM BIT(21)
+
+#define FSL_QDMA_BCQMR_EN BIT(31)
+#define FSL_QDMA_BCQMR_EI BIT(30)
+#define FSL_QDMA_BCQMR_CD_THLD(x) ((x) << 20)
+#define FSL_QDMA_BCQMR_CQ_SIZE(x) ((x) << 16)
+
+#define FSL_QDMA_BCQSR_QF BIT(16)
+#define FSL_QDMA_BCQSR_XOFF BIT(0)
+
+#define FSL_QDMA_BSQMR_EN BIT(31)
+#define FSL_QDMA_BSQMR_DI BIT(30)
+#define FSL_QDMA_BSQMR_CQ_SIZE(x) ((x) << 16)
+
+#define FSL_QDMA_BSQSR_QE BIT(17)
+
+#define FSL_QDMA_DMR_DQD BIT(30)
+#define FSL_QDMA_DSR_DB BIT(31)
+
+/* Size related definition */
+#define FSL_QDMA_QUEUE_MAX 8
+#define FSL_QDMA_COMMAND_BUFFER_SIZE 64
+#define FSL_QDMA_DESCRIPTOR_BUFFER_SIZE 32
+#define FSL_QDMA_CIRCULAR_DESC_SIZE_MIN 64
+#define FSL_QDMA_CIRCULAR_DESC_SIZE_MAX 16384
+#define FSL_QDMA_QUEUE_NUM_MAX 8
+
+/* Field definition for CMD */
+#define FSL_QDMA_CMD_RWTTYPE 0x4
+#define FSL_QDMA_CMD_LWC 0x2
+#define FSL_QDMA_CMD_RWTTYPE_OFFSET 28
+#define FSL_QDMA_CMD_NS_OFFSET 27
+#define FSL_QDMA_CMD_DQOS_OFFSET 24
+#define FSL_QDMA_CMD_WTHROTL_OFFSET 20
+#define FSL_QDMA_CMD_DSEN_OFFSET 19
+#define FSL_QDMA_CMD_LWC_OFFSET 16
+
+/* Field definition for Descriptor offset */
+#define QDMA_CCDF_STATUS 20
+#define QDMA_CCDF_OFFSET 20
+
+/* Field definition for safe loop count*/
+#define FSL_QDMA_HALT_COUNT 1500
+#define FSL_QDMA_MAX_SIZE 16385
+#define FSL_QDMA_COMP_TIMEOUT 1000
+#define FSL_COMMAND_QUEUE_OVERFLLOW 10
+
+#define FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma_engine, x) \
+ (((fsl_qdma_engine)->block_offset) * (x))
+
+/**
+ * struct fsl_qdma_format - This is the struct holding describing compound
+ * descriptor format with qDMA.
+ * @status: Command status and enqueue status notification.
+ * @cfg: Frame offset and frame format.
+ * @addr_lo: Holding the compound descriptor of the lower
+ * 32-bits address in memory 40-bit address.
+ * @addr_hi: Same as above member, but point high 8-bits in
+ * memory 40-bit address.
+ * @__reserved1: Reserved field.
+ * @cfg8b_w1: Compound descriptor command queue origin produced
+ * by qDMA and dynamic debug field.
+ * @data Pointer to the memory 40-bit address, describes DMA
+ * source information and DMA destination information.
+ */
+struct fsl_qdma_format {
+ __le32 status;
+ __le32 cfg;
+ union {
+ struct {
+ __le32 addr_lo;
+ u8 addr_hi;
+ u8 __reserved1[2];
+ u8 cfg8b_w1;
+ } __packed;
+ __le64 data;
+ };
+} __packed;
+
+/* qDMA status notification pre information */
+struct fsl_pre_status {
+ u64 addr;
+ u8 queue;
+};
+
+static DEFINE_PER_CPU(struct fsl_pre_status, pre);
+
+struct fsl_qdma_chan {
+ struct virt_dma_chan vchan;
+ struct virt_dma_desc vdesc;
+ enum dma_status status;
+ struct fsl_qdma_engine *qdma;
+ struct fsl_qdma_queue *queue;
+};
+
+struct fsl_qdma_queue {
+ struct fsl_qdma_format *virt_head;
+ struct fsl_qdma_format *virt_tail;
+ struct list_head comp_used;
+ struct list_head comp_free;
+ struct dma_pool *comp_pool;
+ struct dma_pool *desc_pool;
+ spinlock_t queue_lock;
+ dma_addr_t bus_addr;
+ u32 n_cq;
+ u32 id;
+ struct fsl_qdma_format *cq;
+ void __iomem *block_base;
+};
+
+struct fsl_qdma_comp {
+ dma_addr_t bus_addr;
+ dma_addr_t desc_bus_addr;
+ struct fsl_qdma_format *virt_addr;
+ struct fsl_qdma_format *desc_virt_addr;
+ struct fsl_qdma_chan *qchan;
+ struct virt_dma_desc vdesc;
+ struct list_head list;
+};
+
+struct fsl_qdma_engine {
+ struct dma_device dma_dev;
+ void __iomem *ctrl_base;
+ void __iomem *status_base;
+ void __iomem *block_base;
+ u32 n_chans;
+ u32 n_queues;
+ struct mutex fsl_qdma_mutex;
+ int error_irq;
+ int *queue_irq;
+ u32 feature;
+ struct fsl_qdma_queue *queue;
+ struct fsl_qdma_queue **status;
+ struct fsl_qdma_chan *chans;
+ int block_number;
+ int block_offset;
+ int irq_base;
+ int desc_allocated;
+
+};
+
+static inline u64
+qdma_ccdf_addr_get64(const struct fsl_qdma_format *ccdf)
+{
+ return le64_to_cpu(ccdf->data) & (U64_MAX >> 24);
+}
+
+static inline void
+qdma_desc_addr_set64(struct fsl_qdma_format *ccdf, u64 addr)
+{
+ ccdf->addr_hi = upper_32_bits(addr);
+ ccdf->addr_lo = cpu_to_le32(lower_32_bits(addr));
+}
+
+static inline u8
+qdma_ccdf_get_queue(const struct fsl_qdma_format *ccdf)
+{
+ return ccdf->cfg8b_w1 & U8_MAX;
+}
+
+static inline int
+qdma_ccdf_get_offset(const struct fsl_qdma_format *ccdf)
+{
+ return (le32_to_cpu(ccdf->cfg) & QDMA_CCDF_MASK) >> QDMA_CCDF_OFFSET;
+}
+
+static inline void
+qdma_ccdf_set_format(struct fsl_qdma_format *ccdf, int offset)
+{
+ ccdf->cfg = cpu_to_le32(QDMA_CCDF_FOTMAT | offset);
+}
+
+static inline int
+qdma_ccdf_get_status(const struct fsl_qdma_format *ccdf)
+{
+ return (le32_to_cpu(ccdf->status) & QDMA_CCDF_MASK) >> QDMA_CCDF_STATUS;
+}
+
+static inline void
+qdma_ccdf_set_ser(struct fsl_qdma_format *ccdf, int status)
+{
+ ccdf->status = cpu_to_le32(QDMA_CCDF_SER | status);
+}
+
+static inline void qdma_csgf_set_len(struct fsl_qdma_format *csgf, int len)
+{
+ csgf->cfg = cpu_to_le32(len & QDMA_SG_LEN_MASK);
+}
+
+static inline void qdma_csgf_set_f(struct fsl_qdma_format *csgf, int len)
+{
+ csgf->cfg = cpu_to_le32(QDMA_SG_FIN | (len & QDMA_SG_LEN_MASK));
+}
+
+static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem *addr)
+{
+ return FSL_DMA_IN(qdma, addr, 32);
+}
+
+static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val,
+ void __iomem *addr)
+{
+ FSL_DMA_OUT(qdma, addr, val, 32);
+}
+
+static struct fsl_qdma_chan *to_fsl_qdma_chan(struct dma_chan *chan)
+{
+ return container_of(chan, struct fsl_qdma_chan, vchan.chan);
+}
+
+static struct fsl_qdma_comp *to_fsl_qdma_comp(struct virt_dma_desc *vd)
+{
+ return container_of(vd, struct fsl_qdma_comp, vdesc);
+}
+
+static void fsl_qdma_free_chan_resources(struct dma_chan *chan)
+{
+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
+ struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
+ struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
+ struct fsl_qdma_comp *comp_temp, *_comp_temp;
+ unsigned long flags;
+ LIST_HEAD(head);
+
+ spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
+ vchan_get_all_descriptors(&fsl_chan->vchan, &head);
+ spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
+
+ vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
+
+ if (!fsl_queue->comp_pool && !fsl_queue->comp_pool)
+ return;
+
+ list_for_each_entry_safe(comp_temp, _comp_temp,
+ &fsl_queue->comp_used, list) {
+ dma_pool_free(fsl_queue->comp_pool,
+ comp_temp->virt_addr,
+ comp_temp->bus_addr);
+ dma_pool_free(fsl_queue->desc_pool,
+ comp_temp->desc_virt_addr,
+ comp_temp->desc_bus_addr);
+ list_del(&comp_temp->list);
+ kfree(comp_temp);
+ }
+
+ list_for_each_entry_safe(comp_temp, _comp_temp,
+ &fsl_queue->comp_free, list) {
+ dma_pool_free(fsl_queue->comp_pool,
+ comp_temp->virt_addr,
+ comp_temp->bus_addr);
+ dma_pool_free(fsl_queue->desc_pool,
+ comp_temp->desc_virt_addr,
+ comp_temp->desc_bus_addr);
+ list_del(&comp_temp->list);
+ kfree(comp_temp);
+ }
+
+ dma_pool_destroy(fsl_queue->comp_pool);
+ dma_pool_destroy(fsl_queue->desc_pool);
+
+ fsl_qdma->desc_allocated--;
+ fsl_queue->comp_pool = NULL;
+ fsl_queue->desc_pool = NULL;
+}
+
+static void fsl_qdma_comp_fill_memcpy(struct fsl_qdma_comp *fsl_comp,
+ dma_addr_t dst, dma_addr_t src, u32 len)
+{
+ struct fsl_qdma_format *sdf, *ddf;
+ struct fsl_qdma_format *ccdf, *csgf_desc, *csgf_src, *csgf_dest;
+
+ ccdf = fsl_comp->virt_addr;
+ csgf_desc = fsl_comp->virt_addr + 1;
+ csgf_src = fsl_comp->virt_addr + 2;
+ csgf_dest = fsl_comp->virt_addr + 3;
+ sdf = fsl_comp->desc_virt_addr;
+ ddf = fsl_comp->desc_virt_addr + 1;
+
+ memset(fsl_comp->virt_addr, 0, FSL_QDMA_COMMAND_BUFFER_SIZE);
+ memset(fsl_comp->desc_virt_addr, 0, FSL_QDMA_DESCRIPTOR_BUFFER_SIZE);
+ /* Head Command Descriptor(Frame Descriptor) */
+ qdma_desc_addr_set64(ccdf, fsl_comp->bus_addr + 16);
+ qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf));
+ qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf));
+ /* Status notification is enqueued to status queue. */
+ /* Compound Command Descriptor(Frame List Table) */
+ qdma_desc_addr_set64(csgf_desc, fsl_comp->desc_bus_addr);
+ /* It must be 32 as Compound S/G Descriptor */
+ qdma_csgf_set_len(csgf_desc, 32);
+ qdma_desc_addr_set64(csgf_src, src);
+ qdma_csgf_set_len(csgf_src, len);
+ qdma_desc_addr_set64(csgf_dest, dst);
+ qdma_csgf_set_len(csgf_dest, len);
+ /* This entry is the last entry. */
+ qdma_csgf_set_f(csgf_dest, len);
+ /* Descriptor Buffer */
+ sdf->data =
+ cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
+ FSL_QDMA_CMD_RWTTYPE_OFFSET);
+ ddf->data =
+ cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
+ FSL_QDMA_CMD_RWTTYPE_OFFSET);
+ ddf->data |=
+ cpu_to_le64(FSL_QDMA_CMD_LWC << FSL_QDMA_CMD_LWC_OFFSET);
+}
+
+/*
+ * Pre-request full command descriptor for enqueue.
+ */
+static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue *queue)
+{
+ int i;
+ struct fsl_qdma_comp *comp_temp, *_comp_temp;
+
+ for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW; i++) {
+ comp_temp = kzalloc(sizeof(*comp_temp), GFP_KERNEL);
+ if (!comp_temp)
+ goto err_alloc;
+ comp_temp->virt_addr =
+ dma_pool_alloc(queue->comp_pool, GFP_KERNEL,
+ &comp_temp->bus_addr);
+ if (!comp_temp->virt_addr)
+ goto err_dma_alloc;
+
+ comp_temp->desc_virt_addr =
+ dma_pool_alloc(queue->desc_pool, GFP_KERNEL,
+ &comp_temp->desc_bus_addr);
+ if (!comp_temp->desc_virt_addr)
+ goto err_desc_dma_alloc;
+
+ list_add_tail(&comp_temp->list, &queue->comp_free);
+ }
+
+ return 0;
+
+err_desc_dma_alloc:
+ dma_pool_free(queue->comp_pool, comp_temp->virt_addr,
+ comp_temp->bus_addr);
+
+err_dma_alloc:
+ kfree(comp_temp);
+
+err_alloc:
+ list_for_each_entry_safe(comp_temp, _comp_temp,
+ &queue->comp_free, list) {
+ if (comp_temp->virt_addr)
+ dma_pool_free(queue->comp_pool,
+ comp_temp->virt_addr,
+ comp_temp->bus_addr);
+ if (comp_temp->desc_virt_addr)
+ dma_pool_free(queue->desc_pool,
+ comp_temp->desc_virt_addr,
+ comp_temp->desc_bus_addr);
+
+ list_del(&comp_temp->list);
+ kfree(comp_temp);
+ }
+
+ return -ENOMEM;
+}
+
+/*
+ * Request a command descriptor for enqueue.
+ */
+static struct fsl_qdma_comp
+*fsl_qdma_request_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
+{
+ unsigned long flags;
+ struct fsl_qdma_comp *comp_temp;
+ int timeout = FSL_QDMA_COMP_TIMEOUT;
+ struct fsl_qdma_queue *queue = fsl_chan->queue;
+
+ while (timeout--) {
+ spin_lock_irqsave(&queue->queue_lock, flags);
+ if (!list_empty(&queue->comp_free)) {
+ comp_temp = list_first_entry(&queue->comp_free,
+ struct fsl_qdma_comp,
+ list);
+ list_del(&comp_temp->list);
+
+ spin_unlock_irqrestore(&queue->queue_lock, flags);
+ comp_temp->qchan = fsl_chan;
+ return comp_temp;
+ }
+ spin_unlock_irqrestore(&queue->queue_lock, flags);
+ udelay(1);
+ }
+
+ return NULL;
+}
+
+static struct fsl_qdma_queue
+*fsl_qdma_alloc_queue_resources(struct platform_device *pdev,
+ struct fsl_qdma_engine *fsl_qdma)
+{
+ int ret, len, i, j;
+ int queue_num, block_number;
+ unsigned int queue_size[FSL_QDMA_QUEUE_MAX];
+ struct fsl_qdma_queue *queue_head, *queue_temp;
+
+ queue_num = fsl_qdma->n_queues;
+ block_number = fsl_qdma->block_number;
+
+ if (queue_num > FSL_QDMA_QUEUE_MAX)
+ queue_num = FSL_QDMA_QUEUE_MAX;
+ len = sizeof(*queue_head) * queue_num * block_number;
+ queue_head = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
+ if (!queue_head)
+ return NULL;
+
+ ret = device_property_read_u32_array(&pdev->dev, "queue-sizes",
+ queue_size, queue_num);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't get queue-sizes.\n");
+ return NULL;
+ }
+ for (j = 0; j < block_number; j++) {
+ for (i = 0; i < queue_num; i++) {
+ if (queue_size[i] > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
+ queue_size[i] < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
+ dev_err(&pdev->dev,
+ "Get wrong queue-sizes.\n");
+ return NULL;
+ }
+ queue_temp = queue_head + i + (j * queue_num);
+
+ queue_temp->cq =
+ dma_alloc_coherent(&pdev->dev,
+ sizeof(struct fsl_qdma_format) *
+ queue_size[i],
+ &queue_temp->bus_addr,
+ GFP_KERNEL);
+ if (!queue_temp->cq)
+ return NULL;
+ queue_temp->block_base = fsl_qdma->block_base +
+ FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
+ queue_temp->n_cq = queue_size[i];
+ queue_temp->id = i;
+ queue_temp->virt_head = queue_temp->cq;
+ queue_temp->virt_tail = queue_temp->cq;
+ /*
+ * List for queue command buffer
+ */
+ INIT_LIST_HEAD(&queue_temp->comp_used);
+ spin_lock_init(&queue_temp->queue_lock);
+ }
+ }
+ return queue_head;
+}
+
+static struct fsl_qdma_queue
+*fsl_qdma_prep_status_queue(struct platform_device *pdev)
+{
+ int ret;
+ unsigned int status_size;
+ struct fsl_qdma_queue *status_head;
+ struct device_node *np = pdev->dev.of_node;
+
+ ret = of_property_read_u32(np, "status-sizes", &status_size);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't get status-sizes.\n");
+ return NULL;
+ }
+ if (status_size > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
+ status_size < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
+ dev_err(&pdev->dev, "Get wrong status_size.\n");
+ return NULL;
+ }
+ status_head = devm_kzalloc(&pdev->dev,
+ sizeof(*status_head), GFP_KERNEL);
+ if (!status_head)
+ return NULL;
+
+ /*
+ * Buffer for queue command
+ */
+ status_head->cq = dma_alloc_coherent(&pdev->dev,
+ sizeof(struct fsl_qdma_format) *
+ status_size,
+ &status_head->bus_addr,
+ GFP_KERNEL);
+ if (!status_head->cq) {
+ devm_kfree(&pdev->dev, status_head);
+ return NULL;
+ }
+ status_head->n_cq = status_size;
+ status_head->virt_head = status_head->cq;
+ status_head->virt_tail = status_head->cq;
+ status_head->comp_pool = NULL;
+
+ return status_head;
+}
+
+static int fsl_qdma_halt(struct fsl_qdma_engine *fsl_qdma)
+{
+ u32 reg;
+ int i, j, count = FSL_QDMA_HALT_COUNT;
+ void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
+
+ /* Disable the command queue and wait for idle state. */
+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
+ reg |= FSL_QDMA_DMR_DQD;
+ qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
+ for (j = 0; j < fsl_qdma->block_number; j++) {
+ block = fsl_qdma->block_base +
+ FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
+ for (i = 0; i < FSL_QDMA_QUEUE_NUM_MAX; i++)
+ qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQMR(i));
+ }
+ while (1) {
+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DSR);
+ if (!(reg & FSL_QDMA_DSR_DB))
+ break;
+ if (count-- < 0)
+ return -EBUSY;
+ udelay(100);
+ }
+
+ for (j = 0; j < fsl_qdma->block_number; j++) {
+ block = fsl_qdma->block_base +
+ FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
+
+ /* Disable status queue. */
+ qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BSQMR);
+
+ /*
+ * clear the command queue interrupt detect register for
+ * all queues.
+ */
+ qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
+ block + FSL_QDMA_BCQIDR(0));
+ }
+
+ return 0;
+}
+
+static int
+fsl_qdma_queue_transfer_complete(struct fsl_qdma_engine *fsl_qdma,
+ void *block,
+ int id)
+{
+ bool duplicate;
+ u32 reg, i, count;
+ struct fsl_qdma_queue *temp_queue;
+ struct fsl_qdma_format *status_addr;
+ struct fsl_qdma_comp *fsl_comp = NULL;
+ struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
+ struct fsl_qdma_queue *fsl_status = fsl_qdma->status[id];
+
+ count = FSL_QDMA_MAX_SIZE;
+
+ while (count--) {
+ duplicate = 0;
+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQSR);
+ if (reg & FSL_QDMA_BSQSR_QE)
+ return 0;
+
+ status_addr = fsl_status->virt_head;
+
+ if (qdma_ccdf_get_queue(status_addr) ==
+ __this_cpu_read(pre.queue) &&
+ qdma_ccdf_addr_get64(status_addr) ==
+ __this_cpu_read(pre.addr))
+ duplicate = 1;
+ i = qdma_ccdf_get_queue(status_addr) +
+ id * fsl_qdma->n_queues;
+ __this_cpu_write(pre.addr, qdma_ccdf_addr_get64(status_addr));
+ __this_cpu_write(pre.queue, qdma_ccdf_get_queue(status_addr));
+ temp_queue = fsl_queue + i;
+
+ spin_lock(&temp_queue->queue_lock);
+ if (list_empty(&temp_queue->comp_used)) {
+ if (!duplicate) {
+ spin_unlock(&temp_queue->queue_lock);
+ return -EAGAIN;
+ }
+ } else {
+ fsl_comp = list_first_entry(&temp_queue->comp_used,
+ struct fsl_qdma_comp, list);
+ if (fsl_comp->bus_addr + 16 !=
+ __this_cpu_read(pre.addr)) {
+ if (!duplicate) {
+ spin_unlock(&temp_queue->queue_lock);
+ return -EAGAIN;
+ }
+ }
+ }
+
+ if (duplicate) {
+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
+ reg |= FSL_QDMA_BSQMR_DI;
+ qdma_desc_addr_set64(status_addr, 0x0);
+ fsl_status->virt_head++;
+ if (fsl_status->virt_head == fsl_status->cq
+ + fsl_status->n_cq)
+ fsl_status->virt_head = fsl_status->cq;
+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
+ spin_unlock(&temp_queue->queue_lock);
+ continue;
+ }
+ list_del(&fsl_comp->list);
+
+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
+ reg |= FSL_QDMA_BSQMR_DI;
+ qdma_desc_addr_set64(status_addr, 0x0);
+ fsl_status->virt_head++;
+ if (fsl_status->virt_head == fsl_status->cq + fsl_status->n_cq)
+ fsl_status->virt_head = fsl_status->cq;
+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
+ spin_unlock(&temp_queue->queue_lock);
+
+ spin_lock(&fsl_comp->qchan->vchan.lock);
+ vchan_cookie_complete(&fsl_comp->vdesc);
+ fsl_comp->qchan->status = DMA_COMPLETE;
+ spin_unlock(&fsl_comp->qchan->vchan.lock);
+ }
+
+ return 0;
+}
+
+static irqreturn_t fsl_qdma_error_handler(int irq, void *dev_id)
+{
+ unsigned int intr;
+ struct fsl_qdma_engine *fsl_qdma = dev_id;
+ void __iomem *status = fsl_qdma->status_base;
+
+ intr = qdma_readl(fsl_qdma, status + FSL_QDMA_DEDR);
+
+ if (intr) {
+ dev_err(fsl_qdma->dma_dev.dev, "DMA transaction error!\n");
+ return IRQ_NONE;
+ }
+
+ qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status + FSL_QDMA_DEDR);
+ return IRQ_HANDLED;
+}
+
+static irqreturn_t fsl_qdma_queue_handler(int irq, void *dev_id)
+{
+ int id;
+ unsigned int intr, reg;
+ struct fsl_qdma_engine *fsl_qdma = dev_id;
+ void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
+
+ id = irq - fsl_qdma->irq_base;
+ if (id < 0 && id > fsl_qdma->block_number) {
+ dev_err(fsl_qdma->dma_dev.dev,
+ "irq %d is wrong irq_base is %d\n",
+ irq, fsl_qdma->irq_base);
+ }
+
+ block = fsl_qdma->block_base +
+ FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, id);
+
+ intr = qdma_readl(fsl_qdma, block + FSL_QDMA_BCQIDR(0));
+
+ if ((intr & FSL_QDMA_CQIDR_SQT) != 0)
+ intr = fsl_qdma_queue_transfer_complete(fsl_qdma, block, id);
+
+ if (intr != 0) {
+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
+ reg |= FSL_QDMA_DMR_DQD;
+ qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
+ qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQIER(0));
+ dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n");
+ }
+
+ /* Clear all detected events and interrupts. */
+ qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
+ block + FSL_QDMA_BCQIDR(0));
+
+ return IRQ_HANDLED;
+}
+
+static int
+fsl_qdma_irq_init(struct platform_device *pdev,
+ struct fsl_qdma_engine *fsl_qdma)
+{
+ int i;
+ int cpu;
+ int ret;
+ char irq_name[20];
+
+ fsl_qdma->error_irq =
+ platform_get_irq_byname(pdev, "qdma-error");
+ if (fsl_qdma->error_irq < 0) {
+ dev_err(&pdev->dev, "Can't get qdma controller irq.\n");
+ return fsl_qdma->error_irq;
+ }
+
+ ret = devm_request_irq(&pdev->dev, fsl_qdma->error_irq,
+ fsl_qdma_error_handler, 0,
+ "qDMA error", fsl_qdma);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't register qDMA controller IRQ.\n");
+ return ret;
+ }
+
+ for (i = 0; i < fsl_qdma->block_number; i++) {
+ sprintf(irq_name, "qdma-queue%d", i);
+ fsl_qdma->queue_irq[i] =
+ platform_get_irq_byname(pdev, irq_name);
+
+ if (fsl_qdma->queue_irq[i] < 0) {
+ dev_err(&pdev->dev,
+ "Can't get qdma queue %d irq.\n", i);
+ return fsl_qdma->queue_irq[i];
+ }
+
+ ret = devm_request_irq(&pdev->dev,
+ fsl_qdma->queue_irq[i],
+ fsl_qdma_queue_handler,
+ 0,
+ "qDMA queue",
+ fsl_qdma);
+ if (ret) {
+ dev_err(&pdev->dev,
+ "Can't register qDMA queue IRQ.\n");
+ return ret;
+ }
+
+ cpu = i % num_online_cpus();
+ ret = irq_set_affinity_hint(fsl_qdma->queue_irq[i],
+ get_cpu_mask(cpu));
+ if (ret) {
+ dev_err(&pdev->dev,
+ "Can't set cpu %d affinity to IRQ %d.\n",
+ cpu,
+ fsl_qdma->queue_irq[i]);
+ return ret;
+ }
+ }
+
+ return 0;
+}
+
+static void fsl_qdma_irq_exit(struct platform_device *pdev,
+ struct fsl_qdma_engine *fsl_qdma)
+{
+ int i;
+
+ devm_free_irq(&pdev->dev, fsl_qdma->error_irq, fsl_qdma);
+ for (i = 0; i < fsl_qdma->block_number; i++)
+ devm_free_irq(&pdev->dev, fsl_qdma->queue_irq[i], fsl_qdma);
+}
+
+static int fsl_qdma_reg_init(struct fsl_qdma_engine *fsl_qdma)
+{
+ u32 reg;
+ int i, j, ret;
+ struct fsl_qdma_queue *temp;
+ void __iomem *status = fsl_qdma->status_base;
+ void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
+ struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
+
+ /* Try to halt the qDMA engine first. */
+ ret = fsl_qdma_halt(fsl_qdma);
+ if (ret) {
+ dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
+ return ret;
+ }
+
+ for (i = 0; i < fsl_qdma->block_number; i++) {
+ /*
+ * Clear the command queue interrupt detect register for
+ * all queues.
+ */
+
+ block = fsl_qdma->block_base +
+ FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, i);
+ qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
+ block + FSL_QDMA_BCQIDR(0));
+ }
+
+ for (j = 0; j < fsl_qdma->block_number; j++) {
+ block = fsl_qdma->block_base +
+ FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
+ for (i = 0; i < fsl_qdma->n_queues; i++) {
+ temp = fsl_queue + i + (j * fsl_qdma->n_queues);
+ /*
+ * Initialize Command Queue registers to
+ * point to the first
+ * command descriptor in memory.
+ * Dequeue Pointer Address Registers
+ * Enqueue Pointer Address Registers
+ */
+
+ qdma_writel(fsl_qdma, temp->bus_addr,
+ block + FSL_QDMA_BCQDPA_SADDR(i));
+ qdma_writel(fsl_qdma, temp->bus_addr,
+ block + FSL_QDMA_BCQEPA_SADDR(i));
+
+ /* Initialize the queue mode. */
+ reg = FSL_QDMA_BCQMR_EN;
+ reg |= FSL_QDMA_BCQMR_CD_THLD(ilog2(temp->n_cq) - 4);
+ reg |= FSL_QDMA_BCQMR_CQ_SIZE(ilog2(temp->n_cq) - 6);
+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BCQMR(i));
+ }
+
+ /*
+ * Workaround for erratum: ERR010812.
+ * We must enable XOFF to avoid the enqueue rejection occurs.
+ * Setting SQCCMR ENTER_WM to 0x20.
+ */
+
+ qdma_writel(fsl_qdma, FSL_QDMA_SQCCMR_ENTER_WM,
+ block + FSL_QDMA_SQCCMR);
+
+ /*
+ * Initialize status queue registers to point to the first
+ * command descriptor in memory.
+ * Dequeue Pointer Address Registers
+ * Enqueue Pointer Address Registers
+ */
+
+ qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
+ block + FSL_QDMA_SQEPAR);
+ qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
+ block + FSL_QDMA_SQDPAR);
+ /* Initialize status queue interrupt. */
+ qdma_writel(fsl_qdma, FSL_QDMA_BCQIER_CQTIE,
+ block + FSL_QDMA_BCQIER(0));
+ qdma_writel(fsl_qdma, FSL_QDMA_BSQICR_ICEN |
+ FSL_QDMA_BSQICR_ICST(5) | 0x8000,
+ block + FSL_QDMA_BSQICR);
+ qdma_writel(fsl_qdma, FSL_QDMA_CQIER_MEIE |
+ FSL_QDMA_CQIER_TEIE,
+ block + FSL_QDMA_CQIER);
+
+ /* Initialize the status queue mode. */
+ reg = FSL_QDMA_BSQMR_EN;
+ reg |= FSL_QDMA_BSQMR_CQ_SIZE(ilog2
+ (fsl_qdma->status[j]->n_cq) - 6);
+
+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
+ }
+
+ /* Initialize controller interrupt register. */
+ qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status + FSL_QDMA_DEDR);
+ qdma_writel(fsl_qdma, FSL_QDMA_DEIER_CLEAR, status + FSL_QDMA_DEIER);
+
+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
+ reg &= ~FSL_QDMA_DMR_DQD;
+ qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
+
+ return 0;
+}
+
+static struct dma_async_tx_descriptor *
+fsl_qdma_prep_memcpy(struct dma_chan *chan, dma_addr_t dst,
+ dma_addr_t src, size_t len, unsigned long flags)
+{
+ struct fsl_qdma_comp *fsl_comp;
+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
+
+ fsl_comp = fsl_qdma_request_enqueue_desc(fsl_chan);
+
+ if (!fsl_comp)
+ return NULL;
+
+ fsl_qdma_comp_fill_memcpy(fsl_comp, dst, src, len);
+
+ return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
+}
+
+static void fsl_qdma_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
+{
+ u32 reg;
+ struct virt_dma_desc *vdesc;
+ struct fsl_qdma_comp *fsl_comp;
+ struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
+ void __iomem *block = fsl_queue->block_base;
+
+ reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQSR(fsl_queue->id));
+ if (reg & (FSL_QDMA_BCQSR_QF | FSL_QDMA_BCQSR_XOFF))
+ return;
+ vdesc = vchan_next_desc(&fsl_chan->vchan);
+ if (!vdesc)
+ return;
+ list_del(&vdesc->node);
+ fsl_comp = to_fsl_qdma_comp(vdesc);
+
+ memcpy(fsl_queue->virt_head++,
+ fsl_comp->virt_addr, sizeof(struct fsl_qdma_format));
+ if (fsl_queue->virt_head == fsl_queue->cq + fsl_queue->n_cq)
+ fsl_queue->virt_head = fsl_queue->cq;
+
+ list_add_tail(&fsl_comp->list, &fsl_queue->comp_used);
+ barrier();
+ reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQMR(fsl_queue->id));
+ reg |= FSL_QDMA_BCQMR_EI;
+ qdma_writel(fsl_chan->qdma, reg, block + FSL_QDMA_BCQMR(fsl_queue->id));
+ fsl_chan->status = DMA_IN_PROGRESS;
+}
+
+static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc)
+{
+ unsigned long flags;
+ struct fsl_qdma_comp *fsl_comp;
+ struct fsl_qdma_queue *fsl_queue;
+
+ fsl_comp = to_fsl_qdma_comp(vdesc);
+ fsl_queue = fsl_comp->qchan->queue;
+
+ spin_lock_irqsave(&fsl_queue->queue_lock, flags);
+ list_add_tail(&fsl_comp->list, &fsl_queue->comp_free);
+ spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
+}
+
+static void fsl_qdma_issue_pending(struct dma_chan *chan)
+{
+ unsigned long flags;
+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
+ struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
+
+ spin_lock_irqsave(&fsl_queue->queue_lock, flags);
+ spin_lock(&fsl_chan->vchan.lock);
+ if (vchan_issue_pending(&fsl_chan->vchan))
+ fsl_qdma_enqueue_desc(fsl_chan);
+ spin_unlock(&fsl_chan->vchan.lock);
+ spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
+}
+
+static void fsl_qdma_synchronize(struct dma_chan *chan)
+{
+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
+
+ vchan_synchronize(&fsl_chan->vchan);
+}
+
+static int fsl_qdma_terminate_all(struct dma_chan *chan)
+{
+ LIST_HEAD(head);
+ unsigned long flags;
+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
+
+ spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
+ vchan_get_all_descriptors(&fsl_chan->vchan, &head);
+ spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
+ vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
+ return 0;
+}
+
+static int fsl_qdma_alloc_chan_resources(struct dma_chan *chan)
+{
+ int ret;
+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
+ struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
+ struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
+
+ if (fsl_queue->comp_pool && fsl_queue->desc_pool)
+ return fsl_qdma->desc_allocated;
+
+ INIT_LIST_HEAD(&fsl_queue->comp_free);
+
+ /*
+ * The dma pool for queue command buffer
+ */
+ fsl_queue->comp_pool =
+ dma_pool_create("comp_pool",
+ chan->device->dev,
+ FSL_QDMA_COMMAND_BUFFER_SIZE,
+ 64, 0);
+ if (!fsl_queue->comp_pool)
+ return -ENOMEM;
+
+ /*
+ * The dma pool for Descriptor(SD/DD) buffer
+ */
+ fsl_queue->desc_pool =
+ dma_pool_create("desc_pool",
+ chan->device->dev,
+ FSL_QDMA_DESCRIPTOR_BUFFER_SIZE,
+ 32, 0);
+ if (!fsl_queue->desc_pool)
+ goto err_desc_pool;
+
+ ret = fsl_qdma_pre_request_enqueue_desc(fsl_queue);
+ if (ret) {
+ dev_err(chan->device->dev,
+ "failed to alloc dma buffer for S/G descriptor\n");
+ goto err_mem;
+ }
+
+ fsl_qdma->desc_allocated++;
+ return fsl_qdma->desc_allocated;
+
+err_mem:
+ dma_pool_destroy(fsl_queue->desc_pool);
+err_desc_pool:
+ dma_pool_destroy(fsl_queue->comp_pool);
+ return -ENOMEM;
+}
+
+static int fsl_qdma_probe(struct platform_device *pdev)
+{
+ int ret, i;
+ int blk_num, blk_off;
+ u32 len, chans, queues;
+ struct resource *res;
+ struct fsl_qdma_chan *fsl_chan;
+ struct fsl_qdma_engine *fsl_qdma;
+ struct device_node *np = pdev->dev.of_node;
+
+ ret = of_property_read_u32(np, "dma-channels", &chans);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't get dma-channels.\n");
+ return ret;
+ }
+
+ ret = of_property_read_u32(np, "block-offset", &blk_off);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't get block-offset.\n");
+ return ret;
+ }
+
+ ret = of_property_read_u32(np, "block-number", &blk_num);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't get block-number.\n");
+ return ret;
+ }
+
+ blk_num = min_t(int, blk_num, num_online_cpus());
+
+ len = sizeof(*fsl_qdma);
+ fsl_qdma = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
+ if (!fsl_qdma)
+ return -ENOMEM;
+
+ len = sizeof(*fsl_chan) * chans;
+ fsl_qdma->chans = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
+ if (!fsl_qdma->chans)
+ return -ENOMEM;
+
+ len = sizeof(struct fsl_qdma_queue *) * blk_num;
+ fsl_qdma->status = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
+ if (!fsl_qdma->status)
+ return -ENOMEM;
+
+ len = sizeof(int) * blk_num;
+ fsl_qdma->queue_irq = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
+ if (!fsl_qdma->queue_irq)
+ return -ENOMEM;
+
+ ret = of_property_read_u32(np, "fsl,dma-queues", &queues);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't get queues.\n");
+ return ret;
+ }
+
+ fsl_qdma->desc_allocated = 0;
+ fsl_qdma->n_chans = chans;
+ fsl_qdma->n_queues = queues;
+ fsl_qdma->block_number = blk_num;
+ fsl_qdma->block_offset = blk_off;
+
+ mutex_init(&fsl_qdma->fsl_qdma_mutex);
+
+ for (i = 0; i < fsl_qdma->block_number; i++) {
+ fsl_qdma->status[i] = fsl_qdma_prep_status_queue(pdev);
+ if (!fsl_qdma->status[i])
+ return -ENOMEM;
+ }
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
+ fsl_qdma->ctrl_base = devm_ioremap_resource(&pdev->dev, res);
+ if (IS_ERR(fsl_qdma->ctrl_base))
+ return PTR_ERR(fsl_qdma->ctrl_base);
+
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
+ fsl_qdma->status_base = devm_ioremap_resource(&pdev->dev, res);
+ if (IS_ERR(fsl_qdma->status_base))
+ return PTR_ERR(fsl_qdma->status_base);
+
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 2);
+ fsl_qdma->block_base = devm_ioremap_resource(&pdev->dev, res);
+ if (IS_ERR(fsl_qdma->block_base))
+ return PTR_ERR(fsl_qdma->block_base);
+ fsl_qdma->queue = fsl_qdma_alloc_queue_resources(pdev, fsl_qdma);
+ if (!fsl_qdma->queue)
+ return -ENOMEM;
+
+ ret = fsl_qdma_irq_init(pdev, fsl_qdma);
+ if (ret)
+ return ret;
+
+ fsl_qdma->irq_base = platform_get_irq_byname(pdev, "qdma-queue0");
+ fsl_qdma->feature = of_property_read_bool(np, "big-endian");
+ INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
+
+ for (i = 0; i < fsl_qdma->n_chans; i++) {
+ struct fsl_qdma_chan *fsl_chan = &fsl_qdma->chans[i];
+
+ fsl_chan->qdma = fsl_qdma;
+ fsl_chan->queue = fsl_qdma->queue + i % (fsl_qdma->n_queues *
+ fsl_qdma->block_number);
+ fsl_chan->vchan.desc_free = fsl_qdma_free_desc;
+ vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
+ }
+
+ dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
+
+ fsl_qdma->dma_dev.dev = &pdev->dev;
+ fsl_qdma->dma_dev.device_free_chan_resources =
+ fsl_qdma_free_chan_resources;
+ fsl_qdma->dma_dev.device_alloc_chan_resources =
+ fsl_qdma_alloc_chan_resources;
+ fsl_qdma->dma_dev.device_tx_status = dma_cookie_status;
+ fsl_qdma->dma_dev.device_prep_dma_memcpy = fsl_qdma_prep_memcpy;
+ fsl_qdma->dma_dev.device_issue_pending = fsl_qdma_issue_pending;
+ fsl_qdma->dma_dev.device_synchronize = fsl_qdma_synchronize;
+ fsl_qdma->dma_dev.device_terminate_all = fsl_qdma_terminate_all;
+
+ dma_set_mask(&pdev->dev, DMA_BIT_MASK(40));
+
+ platform_set_drvdata(pdev, fsl_qdma);
+
+ ret = dma_async_device_register(&fsl_qdma->dma_dev);
+ if (ret) {
+ dev_err(&pdev->dev,
+ "Can't register NXP Layerscape qDMA engine.\n");
+ return ret;
+ }
+
+ ret = fsl_qdma_reg_init(fsl_qdma);
+ if (ret) {
+ dev_err(&pdev->dev, "Can't Initialize the qDMA engine.\n");
+ return ret;
+ }
+
+ return 0;
+}
+
+static void fsl_qdma_cleanup_vchan(struct dma_device *dmadev)
+{
+ struct fsl_qdma_chan *chan, *_chan;
+
+ list_for_each_entry_safe(chan, _chan,
+ &dmadev->channels, vchan.chan.device_node) {
+ list_del(&chan->vchan.chan.device_node);
+ tasklet_kill(&chan->vchan.task);
+ }
+}
+
+static int fsl_qdma_remove(struct platform_device *pdev)
+{
+ int i;
+ struct fsl_qdma_queue *status;
+ struct device_node *np = pdev->dev.of_node;
+ struct fsl_qdma_engine *fsl_qdma = platform_get_drvdata(pdev);
+
+ fsl_qdma_irq_exit(pdev, fsl_qdma);
+ fsl_qdma_cleanup_vchan(&fsl_qdma->dma_dev);
+ of_dma_controller_free(np);
+ dma_async_device_unregister(&fsl_qdma->dma_dev);
+
+ for (i = 0; i < fsl_qdma->block_number; i++) {
+ status = fsl_qdma->status[i];
+ dma_free_coherent(&pdev->dev, sizeof(struct fsl_qdma_format) *
+ status->n_cq, status->cq, status->bus_addr);
+ }
+ return 0;
+}
+
+static const struct of_device_id fsl_qdma_dt_ids[] = {
+ { .compatible = "fsl,ls1021a-qdma", },
+ { /* sentinel */ }
+};
+MODULE_DEVICE_TABLE(of, fsl_qdma_dt_ids);
+
+static struct platform_driver fsl_qdma_driver = {
+ .driver = {
+ .name = "fsl-qdma",
+ .of_match_table = fsl_qdma_dt_ids,
+ },
+ .probe = fsl_qdma_probe,
+ .remove = fsl_qdma_remove,
+};
+
+module_platform_driver(fsl_qdma_driver);
+
+MODULE_ALIAS("platform:fsl-qdma");
+MODULE_DESCRIPTION("NXP Layerscape qDMA engine driver");
--
1.7.1
add the qDMA device tree nodes for LS1043A devices.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
---
change in v10:
- no
arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi | 22 ++++++++++++++++++++++
1 files changed, 22 insertions(+), 0 deletions(-)
diff --git a/arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi b/arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi
index 7881e3d..d560141 100644
--- a/arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi
+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi
@@ -734,6 +734,28 @@
<0000 0 0 3 &gic 0 156 0x4>,
<0000 0 0 4 &gic 0 157 0x4>;
};
+
+ qdma: dma-controller@8380000 {
+ compatible = "fsl,ls1021a-qdma", "fsl,ls1043a-qdma";
+ reg = <0x0 0x8380000 0x0 0x1000>, /* Controller regs */
+ <0x0 0x8390000 0x0 0x10000>, /* Status regs */
+ <0x0 0x83a0000 0x0 0x40000>; /* Block regs */
+ interrupts = <0 153 0x4>,
+ <0 39 0x4>,
+ <0 40 0x4>,
+ <0 41 0x4>,
+ <0 42 0x4>;
+ interrupt-names = "qdma-error", "qdma-queue0",
+ "qdma-queue1", "qdma-queue2", "qdma-queue3";
+ dma-channels = <8>;
+ block-number = <1>;
+ block-offset = <0x10000>;
+ fsl,dma-queues = <2>;
+ status-sizes = <64>;
+ queue-sizes = <64 64>;
+ big-endian;
+ };
+
};
firmware {
--
1.7.1
add the qDMA device tree nodes for LS1046A devices.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
---
change in v10:
- no
arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi | 21 +++++++++++++++++++++
1 files changed, 21 insertions(+), 0 deletions(-)
diff --git a/arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi b/arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi
index ef83786..dc65318 100644
--- a/arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi
+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi
@@ -704,6 +704,27 @@
<0000 0 0 4 &gic GIC_SPI 154 IRQ_TYPE_LEVEL_HIGH>;
};
+ qdma: dma-controller@8380000 {
+ compatible = "fsl,ls1046a-qdma", "fsl,ls1021a-qdma";
+ reg = <0x0 0x8380000 0x0 0x1000>, /* Controller regs */
+ <0x0 0x8390000 0x0 0x10000>, /* Status regs */
+ <0x0 0x83a0000 0x0 0x40000>; /* Block regs */
+ interrupts = <0 153 0x4>,
+ <0 39 0x4>,
+ <0 40 0x4>,
+ <0 41 0x4>,
+ <0 42 0x4>;
+ interrupt-names = "qdma-error", "qdma-queue0",
+ "qdma-queue1", "qdma-queue2", "qdma-queue3";
+ dma-channels = <8>;
+ block-number = <1>;
+ block-offset = <0x10000>;
+ fsl,dma-queues = <2>;
+ status-sizes = <64>;
+ queue-sizes = <64 64>;
+ big-endian;
+ };
+
};
reserved-memory {
--
1.7.1
add the qDMA device tree nodes for LS1021A devices.
Signed-off-by: Wen He <[email protected]>
Signed-off-by: Peng Ma <[email protected]>
---
change in v10:
- no
arch/arm/boot/dts/ls1021a.dtsi | 20 ++++++++++++++++++++
1 files changed, 20 insertions(+), 0 deletions(-)
diff --git a/arch/arm/boot/dts/ls1021a.dtsi b/arch/arm/boot/dts/ls1021a.dtsi
index f184905..c0ed6be 100644
--- a/arch/arm/boot/dts/ls1021a.dtsi
+++ b/arch/arm/boot/dts/ls1021a.dtsi
@@ -806,5 +806,25 @@
#size-cells = <1>;
ranges = <0x0 0x0 0x10010000 0x10000>;
};
+
+ qdma: dma-controller@8390000 {
+ compatible = "fsl,ls1021a-qdma";
+ reg = <0x0 0x8388000 0x0 0x1000>, /* Controller regs */
+ <0x0 0x8389000 0x0 0x1000>, /* Status regs */
+ <0x0 0x838a000 0x0 0x2000>; /* Block regs */
+ interrupts = <GIC_SPI 185 IRQ_TYPE_LEVEL_HIGH>,
+ <GIC_SPI 76 IRQ_TYPE_LEVEL_HIGH>,
+ <GIC_SPI 77 IRQ_TYPE_LEVEL_HIGH>;
+ interrupt-names = "qdma-error",
+ "qdma-queue0", "qdma-queue1";
+ dma-channels = <8>;
+ block-number = <1>;
+ block-offset = <0x1000>;
+ fsl,dma-queues = <2>;
+ status-sizes = <64>;
+ queue-sizes = <64 64>;
+ big-endian;
+ };
+
};
};
--
1.7.1
On Fri, Oct 26, 2018 at 4:57 AM Peng Ma <[email protected]> wrote:
>
> NXP Queue DMA controller(qDMA) on Layerscape SoCs supports channel
> virtuallization by allowing DMA jobs to be enqueued into different
> command queues.
>
> Note that this module depends on NXP DPAA.
It is not clear if you are saying that the driver can only work on
SoCs with a DPAA hardware block, or the driver is actually depending
on the DPAA drivers also. If it is the later case, you also should
express that in the Kconfig you added below.
>
> Signed-off-by: Wen He <[email protected]>
> Signed-off-by: Jiaheng Fan <[email protected]>
> Signed-off-by: Peng Ma <[email protected]>
> ---
> change in v10:
> - no
>
> drivers/dma/Kconfig | 13 +
> drivers/dma/Makefile | 1 +
> drivers/dma/fsl-qdma.c | 1257 ++++++++++++++++++++++++++++++++++++++++++++++++
> 3 files changed, 1271 insertions(+), 0 deletions(-)
> create mode 100644 drivers/dma/fsl-qdma.c
>
> diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
> index dacf3f4..50e19d7 100644
> --- a/drivers/dma/Kconfig
> +++ b/drivers/dma/Kconfig
> @@ -218,6 +218,19 @@ config FSL_EDMA
> multiplexing capability for DMA request sources(slot).
> This module can be found on Freescale Vybrid and LS-1 SoCs.
>
> +config FSL_QDMA
> + tristate "NXP Layerscape qDMA engine support"
> + depends on ARM || ARM64
> + select DMA_ENGINE
> + select DMA_VIRTUAL_CHANNELS
> + select DMA_ENGINE_RAID
> + select ASYNC_TX_ENABLE_CHANNEL_SWITCH
> + help
> + Support the NXP Layerscape qDMA engine with command queue and legacy mode.
> + Channel virtualization is supported through enqueuing of DMA jobs to,
> + or dequeuing DMA jobs from, different work queues.
> + This module can be found on NXP Layerscape SoCs.
> +
> config FSL_RAID
> tristate "Freescale RAID engine Support"
> depends on FSL_SOC && !ASYNC_TX_ENABLE_CHANNEL_SWITCH
> diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
> index c91702d..2d1b586 100644
> --- a/drivers/dma/Makefile
> +++ b/drivers/dma/Makefile
> @@ -32,6 +32,7 @@ obj-$(CONFIG_DW_DMAC_CORE) += dw/
> obj-$(CONFIG_EP93XX_DMA) += ep93xx_dma.o
> obj-$(CONFIG_FSL_DMA) += fsldma.o
> obj-$(CONFIG_FSL_EDMA) += fsl-edma.o
> +obj-$(CONFIG_FSL_QDMA) += fsl-qdma.o
> obj-$(CONFIG_FSL_RAID) += fsl_raid.o
> obj-$(CONFIG_HSU_DMA) += hsu/
> obj-$(CONFIG_IMG_MDC_DMA) += img-mdc-dma.o
> diff --git a/drivers/dma/fsl-qdma.c b/drivers/dma/fsl-qdma.c
> new file mode 100644
> index 0000000..404869e
> --- /dev/null
> +++ b/drivers/dma/fsl-qdma.c
> @@ -0,0 +1,1257 @@
> +// SPDX-License-Identifier: GPL-2.0
> +// Copyright 2018 NXP
I'm not sure if this is really the case. The driver at least has been
sent out in 2015. We should keep these copyright claims, even the
legacy Freescale copyright claims.
> +
> +/*
> + * Driver for NXP Layerscape Queue Direct Memory Access Controller
> + *
> + * Author:
> + * Wen He <[email protected]>
> + * Jiaheng Fan <[email protected]>
> + *
> + */
> +
> +#include <linux/module.h>
> +#include <linux/delay.h>
> +#include <linux/of_irq.h>
> +#include <linux/of_platform.h>
> +#include <linux/of_dma.h>
> +#include <linux/dma-mapping.h>
> +
> +#include "virt-dma.h"
> +#include "fsldma.h"
> +
> +/* Register related definition */
> +#define FSL_QDMA_DMR 0x0
> +#define FSL_QDMA_DSR 0x4
> +#define FSL_QDMA_DEIER 0xe00
> +#define FSL_QDMA_DEDR 0xe04
> +#define FSL_QDMA_DECFDW0R 0xe10
> +#define FSL_QDMA_DECFDW1R 0xe14
> +#define FSL_QDMA_DECFDW2R 0xe18
> +#define FSL_QDMA_DECFDW3R 0xe1c
> +#define FSL_QDMA_DECFQIDR 0xe30
> +#define FSL_QDMA_DECBR 0xe34
> +
> +#define FSL_QDMA_BCQMR(x) (0xc0 + 0x100 * (x))
> +#define FSL_QDMA_BCQSR(x) (0xc4 + 0x100 * (x))
> +#define FSL_QDMA_BCQEDPA_SADDR(x) (0xc8 + 0x100 * (x))
> +#define FSL_QDMA_BCQDPA_SADDR(x) (0xcc + 0x100 * (x))
> +#define FSL_QDMA_BCQEEPA_SADDR(x) (0xd0 + 0x100 * (x))
> +#define FSL_QDMA_BCQEPA_SADDR(x) (0xd4 + 0x100 * (x))
> +#define FSL_QDMA_BCQIER(x) (0xe0 + 0x100 * (x))
> +#define FSL_QDMA_BCQIDR(x) (0xe4 + 0x100 * (x))
> +
> +#define FSL_QDMA_SQDPAR 0x80c
> +#define FSL_QDMA_SQEPAR 0x814
> +#define FSL_QDMA_BSQMR 0x800
> +#define FSL_QDMA_BSQSR 0x804
> +#define FSL_QDMA_BSQICR 0x828
> +#define FSL_QDMA_CQMR 0xa00
> +#define FSL_QDMA_CQDSCR1 0xa08
> +#define FSL_QDMA_CQDSCR2 0xa0c
> +#define FSL_QDMA_CQIER 0xa10
> +#define FSL_QDMA_CQEDR 0xa14
> +#define FSL_QDMA_SQCCMR 0xa20
> +
> +/* Registers for bit and genmask */
> +#define FSL_QDMA_CQIDR_SQT BIT(15)
> +#define QDMA_CCDF_FOTMAT BIT(29)
> +#define QDMA_CCDF_SER BIT(30)
> +#define QDMA_SG_FIN BIT(30)
> +#define QDMA_SG_LEN_MASK GENMASK(29, 0)
> +#define QDMA_CCDF_MASK GENMASK(28, 20)
> +
> +#define FSL_QDMA_DEDR_CLEAR GENMASK(31, 0)
> +#define FSL_QDMA_BCQIDR_CLEAR GENMASK(31, 0)
> +#define FSL_QDMA_DEIER_CLEAR GENMASK(31, 0)
> +
> +#define FSL_QDMA_BCQIER_CQTIE BIT(15)
> +#define FSL_QDMA_BCQIER_CQPEIE BIT(23)
> +#define FSL_QDMA_BSQICR_ICEN BIT(31)
> +
> +#define FSL_QDMA_BSQICR_ICST(x) ((x) << 16)
> +#define FSL_QDMA_CQIER_MEIE BIT(31)
> +#define FSL_QDMA_CQIER_TEIE BIT(0)
> +#define FSL_QDMA_SQCCMR_ENTER_WM BIT(21)
> +
> +#define FSL_QDMA_BCQMR_EN BIT(31)
> +#define FSL_QDMA_BCQMR_EI BIT(30)
> +#define FSL_QDMA_BCQMR_CD_THLD(x) ((x) << 20)
> +#define FSL_QDMA_BCQMR_CQ_SIZE(x) ((x) << 16)
> +
> +#define FSL_QDMA_BCQSR_QF BIT(16)
> +#define FSL_QDMA_BCQSR_XOFF BIT(0)
> +
> +#define FSL_QDMA_BSQMR_EN BIT(31)
> +#define FSL_QDMA_BSQMR_DI BIT(30)
> +#define FSL_QDMA_BSQMR_CQ_SIZE(x) ((x) << 16)
> +
> +#define FSL_QDMA_BSQSR_QE BIT(17)
> +
> +#define FSL_QDMA_DMR_DQD BIT(30)
> +#define FSL_QDMA_DSR_DB BIT(31)
> +
> +/* Size related definition */
> +#define FSL_QDMA_QUEUE_MAX 8
> +#define FSL_QDMA_COMMAND_BUFFER_SIZE 64
> +#define FSL_QDMA_DESCRIPTOR_BUFFER_SIZE 32
> +#define FSL_QDMA_CIRCULAR_DESC_SIZE_MIN 64
> +#define FSL_QDMA_CIRCULAR_DESC_SIZE_MAX 16384
> +#define FSL_QDMA_QUEUE_NUM_MAX 8
> +
> +/* Field definition for CMD */
> +#define FSL_QDMA_CMD_RWTTYPE 0x4
> +#define FSL_QDMA_CMD_LWC 0x2
> +#define FSL_QDMA_CMD_RWTTYPE_OFFSET 28
> +#define FSL_QDMA_CMD_NS_OFFSET 27
> +#define FSL_QDMA_CMD_DQOS_OFFSET 24
> +#define FSL_QDMA_CMD_WTHROTL_OFFSET 20
> +#define FSL_QDMA_CMD_DSEN_OFFSET 19
> +#define FSL_QDMA_CMD_LWC_OFFSET 16
> +
> +/* Field definition for Descriptor offset */
> +#define QDMA_CCDF_STATUS 20
> +#define QDMA_CCDF_OFFSET 20
> +
> +/* Field definition for safe loop count*/
> +#define FSL_QDMA_HALT_COUNT 1500
> +#define FSL_QDMA_MAX_SIZE 16385
> +#define FSL_QDMA_COMP_TIMEOUT 1000
> +#define FSL_COMMAND_QUEUE_OVERFLLOW 10
> +
> +#define FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma_engine, x) \
> + (((fsl_qdma_engine)->block_offset) * (x))
> +
> +/**
> + * struct fsl_qdma_format - This is the struct holding describing compound
> + * descriptor format with qDMA.
> + * @status: Command status and enqueue status notification.
> + * @cfg: Frame offset and frame format.
> + * @addr_lo: Holding the compound descriptor of the lower
> + * 32-bits address in memory 40-bit address.
> + * @addr_hi: Same as above member, but point high 8-bits in
> + * memory 40-bit address.
> + * @__reserved1: Reserved field.
> + * @cfg8b_w1: Compound descriptor command queue origin produced
> + * by qDMA and dynamic debug field.
> + * @data Pointer to the memory 40-bit address, describes DMA
> + * source information and DMA destination information.
> + */
> +struct fsl_qdma_format {
> + __le32 status;
> + __le32 cfg;
> + union {
> + struct {
> + __le32 addr_lo;
> + u8 addr_hi;
> + u8 __reserved1[2];
> + u8 cfg8b_w1;
> + } __packed;
> + __le64 data;
> + };
> +} __packed;
> +
> +/* qDMA status notification pre information */
> +struct fsl_pre_status {
> + u64 addr;
> + u8 queue;
> +};
> +
> +static DEFINE_PER_CPU(struct fsl_pre_status, pre);
> +
> +struct fsl_qdma_chan {
> + struct virt_dma_chan vchan;
> + struct virt_dma_desc vdesc;
> + enum dma_status status;
> + struct fsl_qdma_engine *qdma;
> + struct fsl_qdma_queue *queue;
> +};
> +
> +struct fsl_qdma_queue {
> + struct fsl_qdma_format *virt_head;
> + struct fsl_qdma_format *virt_tail;
> + struct list_head comp_used;
> + struct list_head comp_free;
> + struct dma_pool *comp_pool;
> + struct dma_pool *desc_pool;
> + spinlock_t queue_lock;
> + dma_addr_t bus_addr;
> + u32 n_cq;
> + u32 id;
> + struct fsl_qdma_format *cq;
> + void __iomem *block_base;
> +};
> +
> +struct fsl_qdma_comp {
> + dma_addr_t bus_addr;
> + dma_addr_t desc_bus_addr;
> + struct fsl_qdma_format *virt_addr;
> + struct fsl_qdma_format *desc_virt_addr;
> + struct fsl_qdma_chan *qchan;
> + struct virt_dma_desc vdesc;
> + struct list_head list;
> +};
> +
> +struct fsl_qdma_engine {
> + struct dma_device dma_dev;
> + void __iomem *ctrl_base;
> + void __iomem *status_base;
> + void __iomem *block_base;
> + u32 n_chans;
> + u32 n_queues;
> + struct mutex fsl_qdma_mutex;
> + int error_irq;
> + int *queue_irq;
> + u32 feature;
> + struct fsl_qdma_queue *queue;
> + struct fsl_qdma_queue **status;
> + struct fsl_qdma_chan *chans;
> + int block_number;
> + int block_offset;
> + int irq_base;
> + int desc_allocated;
> +
> +};
> +
> +static inline u64
> +qdma_ccdf_addr_get64(const struct fsl_qdma_format *ccdf)
> +{
> + return le64_to_cpu(ccdf->data) & (U64_MAX >> 24);
> +}
> +
> +static inline void
> +qdma_desc_addr_set64(struct fsl_qdma_format *ccdf, u64 addr)
> +{
> + ccdf->addr_hi = upper_32_bits(addr);
> + ccdf->addr_lo = cpu_to_le32(lower_32_bits(addr));
> +}
> +
> +static inline u8
> +qdma_ccdf_get_queue(const struct fsl_qdma_format *ccdf)
> +{
> + return ccdf->cfg8b_w1 & U8_MAX;
> +}
> +
> +static inline int
> +qdma_ccdf_get_offset(const struct fsl_qdma_format *ccdf)
> +{
> + return (le32_to_cpu(ccdf->cfg) & QDMA_CCDF_MASK) >> QDMA_CCDF_OFFSET;
> +}
> +
> +static inline void
> +qdma_ccdf_set_format(struct fsl_qdma_format *ccdf, int offset)
> +{
> + ccdf->cfg = cpu_to_le32(QDMA_CCDF_FOTMAT | offset);
> +}
> +
> +static inline int
> +qdma_ccdf_get_status(const struct fsl_qdma_format *ccdf)
> +{
> + return (le32_to_cpu(ccdf->status) & QDMA_CCDF_MASK) >> QDMA_CCDF_STATUS;
> +}
> +
> +static inline void
> +qdma_ccdf_set_ser(struct fsl_qdma_format *ccdf, int status)
> +{
> + ccdf->status = cpu_to_le32(QDMA_CCDF_SER | status);
> +}
> +
> +static inline void qdma_csgf_set_len(struct fsl_qdma_format *csgf, int len)
> +{
> + csgf->cfg = cpu_to_le32(len & QDMA_SG_LEN_MASK);
> +}
> +
> +static inline void qdma_csgf_set_f(struct fsl_qdma_format *csgf, int len)
> +{
> + csgf->cfg = cpu_to_le32(QDMA_SG_FIN | (len & QDMA_SG_LEN_MASK));
> +}
> +
> +static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem *addr)
> +{
> + return FSL_DMA_IN(qdma, addr, 32);
> +}
> +
> +static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val,
> + void __iomem *addr)
> +{
> + FSL_DMA_OUT(qdma, addr, val, 32);
> +}
> +
> +static struct fsl_qdma_chan *to_fsl_qdma_chan(struct dma_chan *chan)
> +{
> + return container_of(chan, struct fsl_qdma_chan, vchan.chan);
> +}
> +
> +static struct fsl_qdma_comp *to_fsl_qdma_comp(struct virt_dma_desc *vd)
> +{
> + return container_of(vd, struct fsl_qdma_comp, vdesc);
> +}
> +
> +static void fsl_qdma_free_chan_resources(struct dma_chan *chan)
> +{
> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
> + struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
> + struct fsl_qdma_comp *comp_temp, *_comp_temp;
> + unsigned long flags;
> + LIST_HEAD(head);
> +
> + spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
> + vchan_get_all_descriptors(&fsl_chan->vchan, &head);
> + spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
> +
> + vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
> +
> + if (!fsl_queue->comp_pool && !fsl_queue->comp_pool)
> + return;
> +
> + list_for_each_entry_safe(comp_temp, _comp_temp,
> + &fsl_queue->comp_used, list) {
> + dma_pool_free(fsl_queue->comp_pool,
> + comp_temp->virt_addr,
> + comp_temp->bus_addr);
> + dma_pool_free(fsl_queue->desc_pool,
> + comp_temp->desc_virt_addr,
> + comp_temp->desc_bus_addr);
> + list_del(&comp_temp->list);
> + kfree(comp_temp);
> + }
> +
> + list_for_each_entry_safe(comp_temp, _comp_temp,
> + &fsl_queue->comp_free, list) {
> + dma_pool_free(fsl_queue->comp_pool,
> + comp_temp->virt_addr,
> + comp_temp->bus_addr);
> + dma_pool_free(fsl_queue->desc_pool,
> + comp_temp->desc_virt_addr,
> + comp_temp->desc_bus_addr);
> + list_del(&comp_temp->list);
> + kfree(comp_temp);
> + }
> +
> + dma_pool_destroy(fsl_queue->comp_pool);
> + dma_pool_destroy(fsl_queue->desc_pool);
> +
> + fsl_qdma->desc_allocated--;
> + fsl_queue->comp_pool = NULL;
> + fsl_queue->desc_pool = NULL;
> +}
> +
> +static void fsl_qdma_comp_fill_memcpy(struct fsl_qdma_comp *fsl_comp,
> + dma_addr_t dst, dma_addr_t src, u32 len)
> +{
> + struct fsl_qdma_format *sdf, *ddf;
> + struct fsl_qdma_format *ccdf, *csgf_desc, *csgf_src, *csgf_dest;
> +
> + ccdf = fsl_comp->virt_addr;
> + csgf_desc = fsl_comp->virt_addr + 1;
> + csgf_src = fsl_comp->virt_addr + 2;
> + csgf_dest = fsl_comp->virt_addr + 3;
> + sdf = fsl_comp->desc_virt_addr;
> + ddf = fsl_comp->desc_virt_addr + 1;
> +
> + memset(fsl_comp->virt_addr, 0, FSL_QDMA_COMMAND_BUFFER_SIZE);
> + memset(fsl_comp->desc_virt_addr, 0, FSL_QDMA_DESCRIPTOR_BUFFER_SIZE);
> + /* Head Command Descriptor(Frame Descriptor) */
> + qdma_desc_addr_set64(ccdf, fsl_comp->bus_addr + 16);
> + qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf));
> + qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf));
> + /* Status notification is enqueued to status queue. */
> + /* Compound Command Descriptor(Frame List Table) */
> + qdma_desc_addr_set64(csgf_desc, fsl_comp->desc_bus_addr);
> + /* It must be 32 as Compound S/G Descriptor */
> + qdma_csgf_set_len(csgf_desc, 32);
> + qdma_desc_addr_set64(csgf_src, src);
> + qdma_csgf_set_len(csgf_src, len);
> + qdma_desc_addr_set64(csgf_dest, dst);
> + qdma_csgf_set_len(csgf_dest, len);
> + /* This entry is the last entry. */
> + qdma_csgf_set_f(csgf_dest, len);
> + /* Descriptor Buffer */
> + sdf->data =
> + cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
> + FSL_QDMA_CMD_RWTTYPE_OFFSET);
> + ddf->data =
> + cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
> + FSL_QDMA_CMD_RWTTYPE_OFFSET);
> + ddf->data |=
> + cpu_to_le64(FSL_QDMA_CMD_LWC << FSL_QDMA_CMD_LWC_OFFSET);
> +}
> +
> +/*
> + * Pre-request full command descriptor for enqueue.
> + */
> +static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue *queue)
> +{
> + int i;
> + struct fsl_qdma_comp *comp_temp, *_comp_temp;
> +
> + for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW; i++) {
> + comp_temp = kzalloc(sizeof(*comp_temp), GFP_KERNEL);
> + if (!comp_temp)
> + goto err_alloc;
> + comp_temp->virt_addr =
> + dma_pool_alloc(queue->comp_pool, GFP_KERNEL,
> + &comp_temp->bus_addr);
> + if (!comp_temp->virt_addr)
> + goto err_dma_alloc;
> +
> + comp_temp->desc_virt_addr =
> + dma_pool_alloc(queue->desc_pool, GFP_KERNEL,
> + &comp_temp->desc_bus_addr);
> + if (!comp_temp->desc_virt_addr)
> + goto err_desc_dma_alloc;
> +
> + list_add_tail(&comp_temp->list, &queue->comp_free);
> + }
> +
> + return 0;
> +
> +err_desc_dma_alloc:
> + dma_pool_free(queue->comp_pool, comp_temp->virt_addr,
> + comp_temp->bus_addr);
> +
> +err_dma_alloc:
> + kfree(comp_temp);
> +
> +err_alloc:
> + list_for_each_entry_safe(comp_temp, _comp_temp,
> + &queue->comp_free, list) {
> + if (comp_temp->virt_addr)
> + dma_pool_free(queue->comp_pool,
> + comp_temp->virt_addr,
> + comp_temp->bus_addr);
> + if (comp_temp->desc_virt_addr)
> + dma_pool_free(queue->desc_pool,
> + comp_temp->desc_virt_addr,
> + comp_temp->desc_bus_addr);
> +
> + list_del(&comp_temp->list);
> + kfree(comp_temp);
> + }
> +
> + return -ENOMEM;
> +}
> +
> +/*
> + * Request a command descriptor for enqueue.
> + */
> +static struct fsl_qdma_comp
> +*fsl_qdma_request_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
> +{
> + unsigned long flags;
> + struct fsl_qdma_comp *comp_temp;
> + int timeout = FSL_QDMA_COMP_TIMEOUT;
> + struct fsl_qdma_queue *queue = fsl_chan->queue;
> +
> + while (timeout--) {
> + spin_lock_irqsave(&queue->queue_lock, flags);
> + if (!list_empty(&queue->comp_free)) {
> + comp_temp = list_first_entry(&queue->comp_free,
> + struct fsl_qdma_comp,
> + list);
> + list_del(&comp_temp->list);
> +
> + spin_unlock_irqrestore(&queue->queue_lock, flags);
> + comp_temp->qchan = fsl_chan;
> + return comp_temp;
> + }
> + spin_unlock_irqrestore(&queue->queue_lock, flags);
> + udelay(1);
> + }
> +
> + return NULL;
> +}
> +
> +static struct fsl_qdma_queue
> +*fsl_qdma_alloc_queue_resources(struct platform_device *pdev,
> + struct fsl_qdma_engine *fsl_qdma)
> +{
> + int ret, len, i, j;
> + int queue_num, block_number;
> + unsigned int queue_size[FSL_QDMA_QUEUE_MAX];
> + struct fsl_qdma_queue *queue_head, *queue_temp;
> +
> + queue_num = fsl_qdma->n_queues;
> + block_number = fsl_qdma->block_number;
> +
> + if (queue_num > FSL_QDMA_QUEUE_MAX)
> + queue_num = FSL_QDMA_QUEUE_MAX;
> + len = sizeof(*queue_head) * queue_num * block_number;
> + queue_head = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
> + if (!queue_head)
> + return NULL;
> +
> + ret = device_property_read_u32_array(&pdev->dev, "queue-sizes",
> + queue_size, queue_num);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't get queue-sizes.\n");
> + return NULL;
> + }
> + for (j = 0; j < block_number; j++) {
> + for (i = 0; i < queue_num; i++) {
> + if (queue_size[i] > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
> + queue_size[i] < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
> + dev_err(&pdev->dev,
> + "Get wrong queue-sizes.\n");
> + return NULL;
> + }
> + queue_temp = queue_head + i + (j * queue_num);
> +
> + queue_temp->cq =
> + dma_alloc_coherent(&pdev->dev,
> + sizeof(struct fsl_qdma_format) *
> + queue_size[i],
> + &queue_temp->bus_addr,
> + GFP_KERNEL);
> + if (!queue_temp->cq)
> + return NULL;
> + queue_temp->block_base = fsl_qdma->block_base +
> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
> + queue_temp->n_cq = queue_size[i];
> + queue_temp->id = i;
> + queue_temp->virt_head = queue_temp->cq;
> + queue_temp->virt_tail = queue_temp->cq;
> + /*
> + * List for queue command buffer
> + */
> + INIT_LIST_HEAD(&queue_temp->comp_used);
> + spin_lock_init(&queue_temp->queue_lock);
> + }
> + }
> + return queue_head;
> +}
> +
> +static struct fsl_qdma_queue
> +*fsl_qdma_prep_status_queue(struct platform_device *pdev)
> +{
> + int ret;
> + unsigned int status_size;
> + struct fsl_qdma_queue *status_head;
> + struct device_node *np = pdev->dev.of_node;
> +
> + ret = of_property_read_u32(np, "status-sizes", &status_size);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't get status-sizes.\n");
> + return NULL;
> + }
> + if (status_size > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
> + status_size < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
> + dev_err(&pdev->dev, "Get wrong status_size.\n");
> + return NULL;
> + }
> + status_head = devm_kzalloc(&pdev->dev,
> + sizeof(*status_head), GFP_KERNEL);
> + if (!status_head)
> + return NULL;
> +
> + /*
> + * Buffer for queue command
> + */
> + status_head->cq = dma_alloc_coherent(&pdev->dev,
> + sizeof(struct fsl_qdma_format) *
> + status_size,
> + &status_head->bus_addr,
> + GFP_KERNEL);
> + if (!status_head->cq) {
> + devm_kfree(&pdev->dev, status_head);
> + return NULL;
> + }
> + status_head->n_cq = status_size;
> + status_head->virt_head = status_head->cq;
> + status_head->virt_tail = status_head->cq;
> + status_head->comp_pool = NULL;
> +
> + return status_head;
> +}
> +
> +static int fsl_qdma_halt(struct fsl_qdma_engine *fsl_qdma)
> +{
> + u32 reg;
> + int i, j, count = FSL_QDMA_HALT_COUNT;
> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
> +
> + /* Disable the command queue and wait for idle state. */
> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
> + reg |= FSL_QDMA_DMR_DQD;
> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
> + for (j = 0; j < fsl_qdma->block_number; j++) {
> + block = fsl_qdma->block_base +
> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
> + for (i = 0; i < FSL_QDMA_QUEUE_NUM_MAX; i++)
> + qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQMR(i));
> + }
> + while (1) {
> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DSR);
> + if (!(reg & FSL_QDMA_DSR_DB))
> + break;
> + if (count-- < 0)
> + return -EBUSY;
> + udelay(100);
> + }
> +
> + for (j = 0; j < fsl_qdma->block_number; j++) {
> + block = fsl_qdma->block_base +
> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
> +
> + /* Disable status queue. */
> + qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BSQMR);
> +
> + /*
> + * clear the command queue interrupt detect register for
> + * all queues.
> + */
> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
> + block + FSL_QDMA_BCQIDR(0));
> + }
> +
> + return 0;
> +}
> +
> +static int
> +fsl_qdma_queue_transfer_complete(struct fsl_qdma_engine *fsl_qdma,
> + void *block,
> + int id)
> +{
> + bool duplicate;
> + u32 reg, i, count;
> + struct fsl_qdma_queue *temp_queue;
> + struct fsl_qdma_format *status_addr;
> + struct fsl_qdma_comp *fsl_comp = NULL;
> + struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
> + struct fsl_qdma_queue *fsl_status = fsl_qdma->status[id];
> +
> + count = FSL_QDMA_MAX_SIZE;
> +
> + while (count--) {
> + duplicate = 0;
> + reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQSR);
> + if (reg & FSL_QDMA_BSQSR_QE)
> + return 0;
> +
> + status_addr = fsl_status->virt_head;
> +
> + if (qdma_ccdf_get_queue(status_addr) ==
> + __this_cpu_read(pre.queue) &&
> + qdma_ccdf_addr_get64(status_addr) ==
> + __this_cpu_read(pre.addr))
> + duplicate = 1;
> + i = qdma_ccdf_get_queue(status_addr) +
> + id * fsl_qdma->n_queues;
> + __this_cpu_write(pre.addr, qdma_ccdf_addr_get64(status_addr));
> + __this_cpu_write(pre.queue, qdma_ccdf_get_queue(status_addr));
> + temp_queue = fsl_queue + i;
> +
> + spin_lock(&temp_queue->queue_lock);
> + if (list_empty(&temp_queue->comp_used)) {
> + if (!duplicate) {
> + spin_unlock(&temp_queue->queue_lock);
> + return -EAGAIN;
> + }
> + } else {
> + fsl_comp = list_first_entry(&temp_queue->comp_used,
> + struct fsl_qdma_comp, list);
> + if (fsl_comp->bus_addr + 16 !=
> + __this_cpu_read(pre.addr)) {
> + if (!duplicate) {
> + spin_unlock(&temp_queue->queue_lock);
> + return -EAGAIN;
> + }
> + }
> + }
> +
> + if (duplicate) {
> + reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
> + reg |= FSL_QDMA_BSQMR_DI;
> + qdma_desc_addr_set64(status_addr, 0x0);
> + fsl_status->virt_head++;
> + if (fsl_status->virt_head == fsl_status->cq
> + + fsl_status->n_cq)
> + fsl_status->virt_head = fsl_status->cq;
> + qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
> + spin_unlock(&temp_queue->queue_lock);
> + continue;
> + }
> + list_del(&fsl_comp->list);
> +
> + reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
> + reg |= FSL_QDMA_BSQMR_DI;
> + qdma_desc_addr_set64(status_addr, 0x0);
> + fsl_status->virt_head++;
> + if (fsl_status->virt_head == fsl_status->cq + fsl_status->n_cq)
> + fsl_status->virt_head = fsl_status->cq;
> + qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
> + spin_unlock(&temp_queue->queue_lock);
> +
> + spin_lock(&fsl_comp->qchan->vchan.lock);
> + vchan_cookie_complete(&fsl_comp->vdesc);
> + fsl_comp->qchan->status = DMA_COMPLETE;
> + spin_unlock(&fsl_comp->qchan->vchan.lock);
> + }
> +
> + return 0;
> +}
> +
> +static irqreturn_t fsl_qdma_error_handler(int irq, void *dev_id)
> +{
> + unsigned int intr;
> + struct fsl_qdma_engine *fsl_qdma = dev_id;
> + void __iomem *status = fsl_qdma->status_base;
> +
> + intr = qdma_readl(fsl_qdma, status + FSL_QDMA_DEDR);
> +
> + if (intr) {
> + dev_err(fsl_qdma->dma_dev.dev, "DMA transaction error!\n");
> + return IRQ_NONE;
> + }
> +
> + qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status + FSL_QDMA_DEDR);
> + return IRQ_HANDLED;
> +}
> +
> +static irqreturn_t fsl_qdma_queue_handler(int irq, void *dev_id)
> +{
> + int id;
> + unsigned int intr, reg;
> + struct fsl_qdma_engine *fsl_qdma = dev_id;
> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
> +
> + id = irq - fsl_qdma->irq_base;
> + if (id < 0 && id > fsl_qdma->block_number) {
> + dev_err(fsl_qdma->dma_dev.dev,
> + "irq %d is wrong irq_base is %d\n",
> + irq, fsl_qdma->irq_base);
> + }
> +
> + block = fsl_qdma->block_base +
> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, id);
> +
> + intr = qdma_readl(fsl_qdma, block + FSL_QDMA_BCQIDR(0));
> +
> + if ((intr & FSL_QDMA_CQIDR_SQT) != 0)
> + intr = fsl_qdma_queue_transfer_complete(fsl_qdma, block, id);
> +
> + if (intr != 0) {
> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
> + reg |= FSL_QDMA_DMR_DQD;
> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
> + qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQIER(0));
> + dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n");
> + }
> +
> + /* Clear all detected events and interrupts. */
> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
> + block + FSL_QDMA_BCQIDR(0));
> +
> + return IRQ_HANDLED;
> +}
> +
> +static int
> +fsl_qdma_irq_init(struct platform_device *pdev,
> + struct fsl_qdma_engine *fsl_qdma)
> +{
> + int i;
> + int cpu;
> + int ret;
> + char irq_name[20];
> +
> + fsl_qdma->error_irq =
> + platform_get_irq_byname(pdev, "qdma-error");
> + if (fsl_qdma->error_irq < 0) {
> + dev_err(&pdev->dev, "Can't get qdma controller irq.\n");
> + return fsl_qdma->error_irq;
> + }
> +
> + ret = devm_request_irq(&pdev->dev, fsl_qdma->error_irq,
> + fsl_qdma_error_handler, 0,
> + "qDMA error", fsl_qdma);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't register qDMA controller IRQ.\n");
> + return ret;
> + }
> +
> + for (i = 0; i < fsl_qdma->block_number; i++) {
> + sprintf(irq_name, "qdma-queue%d", i);
> + fsl_qdma->queue_irq[i] =
> + platform_get_irq_byname(pdev, irq_name);
> +
> + if (fsl_qdma->queue_irq[i] < 0) {
> + dev_err(&pdev->dev,
> + "Can't get qdma queue %d irq.\n", i);
> + return fsl_qdma->queue_irq[i];
> + }
> +
> + ret = devm_request_irq(&pdev->dev,
> + fsl_qdma->queue_irq[i],
> + fsl_qdma_queue_handler,
> + 0,
> + "qDMA queue",
> + fsl_qdma);
> + if (ret) {
> + dev_err(&pdev->dev,
> + "Can't register qDMA queue IRQ.\n");
> + return ret;
> + }
> +
> + cpu = i % num_online_cpus();
> + ret = irq_set_affinity_hint(fsl_qdma->queue_irq[i],
> + get_cpu_mask(cpu));
> + if (ret) {
> + dev_err(&pdev->dev,
> + "Can't set cpu %d affinity to IRQ %d.\n",
> + cpu,
> + fsl_qdma->queue_irq[i]);
> + return ret;
> + }
> + }
> +
> + return 0;
> +}
> +
> +static void fsl_qdma_irq_exit(struct platform_device *pdev,
> + struct fsl_qdma_engine *fsl_qdma)
> +{
> + int i;
> +
> + devm_free_irq(&pdev->dev, fsl_qdma->error_irq, fsl_qdma);
> + for (i = 0; i < fsl_qdma->block_number; i++)
> + devm_free_irq(&pdev->dev, fsl_qdma->queue_irq[i], fsl_qdma);
> +}
> +
> +static int fsl_qdma_reg_init(struct fsl_qdma_engine *fsl_qdma)
> +{
> + u32 reg;
> + int i, j, ret;
> + struct fsl_qdma_queue *temp;
> + void __iomem *status = fsl_qdma->status_base;
> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
> + struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
> +
> + /* Try to halt the qDMA engine first. */
> + ret = fsl_qdma_halt(fsl_qdma);
> + if (ret) {
> + dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
> + return ret;
> + }
> +
> + for (i = 0; i < fsl_qdma->block_number; i++) {
> + /*
> + * Clear the command queue interrupt detect register for
> + * all queues.
> + */
> +
> + block = fsl_qdma->block_base +
> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, i);
> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
> + block + FSL_QDMA_BCQIDR(0));
> + }
> +
> + for (j = 0; j < fsl_qdma->block_number; j++) {
> + block = fsl_qdma->block_base +
> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
> + for (i = 0; i < fsl_qdma->n_queues; i++) {
> + temp = fsl_queue + i + (j * fsl_qdma->n_queues);
> + /*
> + * Initialize Command Queue registers to
> + * point to the first
> + * command descriptor in memory.
> + * Dequeue Pointer Address Registers
> + * Enqueue Pointer Address Registers
> + */
> +
> + qdma_writel(fsl_qdma, temp->bus_addr,
> + block + FSL_QDMA_BCQDPA_SADDR(i));
> + qdma_writel(fsl_qdma, temp->bus_addr,
> + block + FSL_QDMA_BCQEPA_SADDR(i));
> +
> + /* Initialize the queue mode. */
> + reg = FSL_QDMA_BCQMR_EN;
> + reg |= FSL_QDMA_BCQMR_CD_THLD(ilog2(temp->n_cq) - 4);
> + reg |= FSL_QDMA_BCQMR_CQ_SIZE(ilog2(temp->n_cq) - 6);
> + qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BCQMR(i));
> + }
> +
> + /*
> + * Workaround for erratum: ERR010812.
> + * We must enable XOFF to avoid the enqueue rejection occurs.
> + * Setting SQCCMR ENTER_WM to 0x20.
> + */
> +
> + qdma_writel(fsl_qdma, FSL_QDMA_SQCCMR_ENTER_WM,
> + block + FSL_QDMA_SQCCMR);
> +
> + /*
> + * Initialize status queue registers to point to the first
> + * command descriptor in memory.
> + * Dequeue Pointer Address Registers
> + * Enqueue Pointer Address Registers
> + */
> +
> + qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
> + block + FSL_QDMA_SQEPAR);
> + qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
> + block + FSL_QDMA_SQDPAR);
> + /* Initialize status queue interrupt. */
> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIER_CQTIE,
> + block + FSL_QDMA_BCQIER(0));
> + qdma_writel(fsl_qdma, FSL_QDMA_BSQICR_ICEN |
> + FSL_QDMA_BSQICR_ICST(5) | 0x8000,
> + block + FSL_QDMA_BSQICR);
> + qdma_writel(fsl_qdma, FSL_QDMA_CQIER_MEIE |
> + FSL_QDMA_CQIER_TEIE,
> + block + FSL_QDMA_CQIER);
> +
> + /* Initialize the status queue mode. */
> + reg = FSL_QDMA_BSQMR_EN;
> + reg |= FSL_QDMA_BSQMR_CQ_SIZE(ilog2
> + (fsl_qdma->status[j]->n_cq) - 6);
> +
> + qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
> + reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
> + }
> +
> + /* Initialize controller interrupt register. */
> + qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status + FSL_QDMA_DEDR);
> + qdma_writel(fsl_qdma, FSL_QDMA_DEIER_CLEAR, status + FSL_QDMA_DEIER);
> +
> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
> + reg &= ~FSL_QDMA_DMR_DQD;
> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
> +
> + return 0;
> +}
> +
> +static struct dma_async_tx_descriptor *
> +fsl_qdma_prep_memcpy(struct dma_chan *chan, dma_addr_t dst,
> + dma_addr_t src, size_t len, unsigned long flags)
> +{
> + struct fsl_qdma_comp *fsl_comp;
> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
> +
> + fsl_comp = fsl_qdma_request_enqueue_desc(fsl_chan);
> +
> + if (!fsl_comp)
> + return NULL;
> +
> + fsl_qdma_comp_fill_memcpy(fsl_comp, dst, src, len);
> +
> + return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
> +}
> +
> +static void fsl_qdma_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
> +{
> + u32 reg;
> + struct virt_dma_desc *vdesc;
> + struct fsl_qdma_comp *fsl_comp;
> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
> + void __iomem *block = fsl_queue->block_base;
> +
> + reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQSR(fsl_queue->id));
> + if (reg & (FSL_QDMA_BCQSR_QF | FSL_QDMA_BCQSR_XOFF))
> + return;
> + vdesc = vchan_next_desc(&fsl_chan->vchan);
> + if (!vdesc)
> + return;
> + list_del(&vdesc->node);
> + fsl_comp = to_fsl_qdma_comp(vdesc);
> +
> + memcpy(fsl_queue->virt_head++,
> + fsl_comp->virt_addr, sizeof(struct fsl_qdma_format));
> + if (fsl_queue->virt_head == fsl_queue->cq + fsl_queue->n_cq)
> + fsl_queue->virt_head = fsl_queue->cq;
> +
> + list_add_tail(&fsl_comp->list, &fsl_queue->comp_used);
> + barrier();
> + reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQMR(fsl_queue->id));
> + reg |= FSL_QDMA_BCQMR_EI;
> + qdma_writel(fsl_chan->qdma, reg, block + FSL_QDMA_BCQMR(fsl_queue->id));
> + fsl_chan->status = DMA_IN_PROGRESS;
> +}
> +
> +static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc)
> +{
> + unsigned long flags;
> + struct fsl_qdma_comp *fsl_comp;
> + struct fsl_qdma_queue *fsl_queue;
> +
> + fsl_comp = to_fsl_qdma_comp(vdesc);
> + fsl_queue = fsl_comp->qchan->queue;
> +
> + spin_lock_irqsave(&fsl_queue->queue_lock, flags);
> + list_add_tail(&fsl_comp->list, &fsl_queue->comp_free);
> + spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
> +}
> +
> +static void fsl_qdma_issue_pending(struct dma_chan *chan)
> +{
> + unsigned long flags;
> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
> +
> + spin_lock_irqsave(&fsl_queue->queue_lock, flags);
> + spin_lock(&fsl_chan->vchan.lock);
> + if (vchan_issue_pending(&fsl_chan->vchan))
> + fsl_qdma_enqueue_desc(fsl_chan);
> + spin_unlock(&fsl_chan->vchan.lock);
> + spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
> +}
> +
> +static void fsl_qdma_synchronize(struct dma_chan *chan)
> +{
> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
> +
> + vchan_synchronize(&fsl_chan->vchan);
> +}
> +
> +static int fsl_qdma_terminate_all(struct dma_chan *chan)
> +{
> + LIST_HEAD(head);
> + unsigned long flags;
> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
> +
> + spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
> + vchan_get_all_descriptors(&fsl_chan->vchan, &head);
> + spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
> + vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
> + return 0;
> +}
> +
> +static int fsl_qdma_alloc_chan_resources(struct dma_chan *chan)
> +{
> + int ret;
> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
> + struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
> +
> + if (fsl_queue->comp_pool && fsl_queue->desc_pool)
> + return fsl_qdma->desc_allocated;
> +
> + INIT_LIST_HEAD(&fsl_queue->comp_free);
> +
> + /*
> + * The dma pool for queue command buffer
> + */
> + fsl_queue->comp_pool =
> + dma_pool_create("comp_pool",
> + chan->device->dev,
> + FSL_QDMA_COMMAND_BUFFER_SIZE,
> + 64, 0);
> + if (!fsl_queue->comp_pool)
> + return -ENOMEM;
> +
> + /*
> + * The dma pool for Descriptor(SD/DD) buffer
> + */
> + fsl_queue->desc_pool =
> + dma_pool_create("desc_pool",
> + chan->device->dev,
> + FSL_QDMA_DESCRIPTOR_BUFFER_SIZE,
> + 32, 0);
> + if (!fsl_queue->desc_pool)
> + goto err_desc_pool;
> +
> + ret = fsl_qdma_pre_request_enqueue_desc(fsl_queue);
> + if (ret) {
> + dev_err(chan->device->dev,
> + "failed to alloc dma buffer for S/G descriptor\n");
> + goto err_mem;
> + }
> +
> + fsl_qdma->desc_allocated++;
> + return fsl_qdma->desc_allocated;
> +
> +err_mem:
> + dma_pool_destroy(fsl_queue->desc_pool);
> +err_desc_pool:
> + dma_pool_destroy(fsl_queue->comp_pool);
> + return -ENOMEM;
> +}
> +
> +static int fsl_qdma_probe(struct platform_device *pdev)
> +{
> + int ret, i;
> + int blk_num, blk_off;
> + u32 len, chans, queues;
> + struct resource *res;
> + struct fsl_qdma_chan *fsl_chan;
> + struct fsl_qdma_engine *fsl_qdma;
> + struct device_node *np = pdev->dev.of_node;
> +
> + ret = of_property_read_u32(np, "dma-channels", &chans);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't get dma-channels.\n");
> + return ret;
> + }
> +
> + ret = of_property_read_u32(np, "block-offset", &blk_off);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't get block-offset.\n");
> + return ret;
> + }
> +
> + ret = of_property_read_u32(np, "block-number", &blk_num);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't get block-number.\n");
> + return ret;
> + }
> +
> + blk_num = min_t(int, blk_num, num_online_cpus());
> +
> + len = sizeof(*fsl_qdma);
> + fsl_qdma = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
> + if (!fsl_qdma)
> + return -ENOMEM;
> +
> + len = sizeof(*fsl_chan) * chans;
> + fsl_qdma->chans = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
> + if (!fsl_qdma->chans)
> + return -ENOMEM;
> +
> + len = sizeof(struct fsl_qdma_queue *) * blk_num;
> + fsl_qdma->status = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
> + if (!fsl_qdma->status)
> + return -ENOMEM;
> +
> + len = sizeof(int) * blk_num;
> + fsl_qdma->queue_irq = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
> + if (!fsl_qdma->queue_irq)
> + return -ENOMEM;
> +
> + ret = of_property_read_u32(np, "fsl,dma-queues", &queues);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't get queues.\n");
> + return ret;
> + }
> +
> + fsl_qdma->desc_allocated = 0;
> + fsl_qdma->n_chans = chans;
> + fsl_qdma->n_queues = queues;
> + fsl_qdma->block_number = blk_num;
> + fsl_qdma->block_offset = blk_off;
> +
> + mutex_init(&fsl_qdma->fsl_qdma_mutex);
> +
> + for (i = 0; i < fsl_qdma->block_number; i++) {
> + fsl_qdma->status[i] = fsl_qdma_prep_status_queue(pdev);
> + if (!fsl_qdma->status[i])
> + return -ENOMEM;
> + }
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> + fsl_qdma->ctrl_base = devm_ioremap_resource(&pdev->dev, res);
> + if (IS_ERR(fsl_qdma->ctrl_base))
> + return PTR_ERR(fsl_qdma->ctrl_base);
> +
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
> + fsl_qdma->status_base = devm_ioremap_resource(&pdev->dev, res);
> + if (IS_ERR(fsl_qdma->status_base))
> + return PTR_ERR(fsl_qdma->status_base);
> +
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 2);
> + fsl_qdma->block_base = devm_ioremap_resource(&pdev->dev, res);
> + if (IS_ERR(fsl_qdma->block_base))
> + return PTR_ERR(fsl_qdma->block_base);
> + fsl_qdma->queue = fsl_qdma_alloc_queue_resources(pdev, fsl_qdma);
> + if (!fsl_qdma->queue)
> + return -ENOMEM;
> +
> + ret = fsl_qdma_irq_init(pdev, fsl_qdma);
> + if (ret)
> + return ret;
> +
> + fsl_qdma->irq_base = platform_get_irq_byname(pdev, "qdma-queue0");
> + fsl_qdma->feature = of_property_read_bool(np, "big-endian");
> + INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
> +
> + for (i = 0; i < fsl_qdma->n_chans; i++) {
> + struct fsl_qdma_chan *fsl_chan = &fsl_qdma->chans[i];
> +
> + fsl_chan->qdma = fsl_qdma;
> + fsl_chan->queue = fsl_qdma->queue + i % (fsl_qdma->n_queues *
> + fsl_qdma->block_number);
> + fsl_chan->vchan.desc_free = fsl_qdma_free_desc;
> + vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
> + }
> +
> + dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
> +
> + fsl_qdma->dma_dev.dev = &pdev->dev;
> + fsl_qdma->dma_dev.device_free_chan_resources =
> + fsl_qdma_free_chan_resources;
> + fsl_qdma->dma_dev.device_alloc_chan_resources =
> + fsl_qdma_alloc_chan_resources;
> + fsl_qdma->dma_dev.device_tx_status = dma_cookie_status;
> + fsl_qdma->dma_dev.device_prep_dma_memcpy = fsl_qdma_prep_memcpy;
> + fsl_qdma->dma_dev.device_issue_pending = fsl_qdma_issue_pending;
> + fsl_qdma->dma_dev.device_synchronize = fsl_qdma_synchronize;
> + fsl_qdma->dma_dev.device_terminate_all = fsl_qdma_terminate_all;
> +
> + dma_set_mask(&pdev->dev, DMA_BIT_MASK(40));
> +
> + platform_set_drvdata(pdev, fsl_qdma);
> +
> + ret = dma_async_device_register(&fsl_qdma->dma_dev);
> + if (ret) {
> + dev_err(&pdev->dev,
> + "Can't register NXP Layerscape qDMA engine.\n");
> + return ret;
> + }
> +
> + ret = fsl_qdma_reg_init(fsl_qdma);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't Initialize the qDMA engine.\n");
> + return ret;
> + }
> +
> + return 0;
> +}
> +
> +static void fsl_qdma_cleanup_vchan(struct dma_device *dmadev)
> +{
> + struct fsl_qdma_chan *chan, *_chan;
> +
> + list_for_each_entry_safe(chan, _chan,
> + &dmadev->channels, vchan.chan.device_node) {
> + list_del(&chan->vchan.chan.device_node);
> + tasklet_kill(&chan->vchan.task);
> + }
> +}
> +
> +static int fsl_qdma_remove(struct platform_device *pdev)
> +{
> + int i;
> + struct fsl_qdma_queue *status;
> + struct device_node *np = pdev->dev.of_node;
> + struct fsl_qdma_engine *fsl_qdma = platform_get_drvdata(pdev);
> +
> + fsl_qdma_irq_exit(pdev, fsl_qdma);
> + fsl_qdma_cleanup_vchan(&fsl_qdma->dma_dev);
> + of_dma_controller_free(np);
> + dma_async_device_unregister(&fsl_qdma->dma_dev);
> +
> + for (i = 0; i < fsl_qdma->block_number; i++) {
> + status = fsl_qdma->status[i];
> + dma_free_coherent(&pdev->dev, sizeof(struct fsl_qdma_format) *
> + status->n_cq, status->cq, status->bus_addr);
> + }
> + return 0;
> +}
> +
> +static const struct of_device_id fsl_qdma_dt_ids[] = {
> + { .compatible = "fsl,ls1021a-qdma", },
> + { /* sentinel */ }
> +};
> +MODULE_DEVICE_TABLE(of, fsl_qdma_dt_ids);
> +
> +static struct platform_driver fsl_qdma_driver = {
> + .driver = {
> + .name = "fsl-qdma",
> + .of_match_table = fsl_qdma_dt_ids,
> + },
> + .probe = fsl_qdma_probe,
> + .remove = fsl_qdma_remove,
> +};
> +
> +module_platform_driver(fsl_qdma_driver);
> +
> +MODULE_ALIAS("platform:fsl-qdma");
> +MODULE_DESCRIPTION("NXP Layerscape qDMA engine driver");
> --
> 1.7.1
>
>-----Original Message-----
>From: Li Yang <[email protected]>
>Sent: 2018年10月27日 4:48
>To: Peng Ma <[email protected]>
>Cc: Vinod <[email protected]>; Rob Herring <[email protected]>; Mark
>Rutland <[email protected]>; Shawn Guo <[email protected]>; Dan
>Williams <[email protected]>; [email protected]; open
>list:OPEN FIRMWARE AND FLATTENED DEVICE TREE BINDINGS
><[email protected]>; lkml <[email protected]>;
>moderated list:ARM/FREESCALE IMX / MXC ARM ARCHITECTURE
><[email protected]>; linuxppc-dev
><[email protected]>; Wen He <[email protected]>; Jiaheng Fan
><[email protected]>
>Subject: Re: [PATCH 3/7] dmaengine: fsl-qdma: Add qDMA controller driver for
>Layerscape SoCs
>
>On Fri, Oct 26, 2018 at 4:57 AM Peng Ma <[email protected]> wrote:
>>
>> NXP Queue DMA controller(qDMA) on Layerscape SoCs supports channel
>> virtuallization by allowing DMA jobs to be enqueued into different
>> command queues.
>>
>> Note that this module depends on NXP DPAA.
>
>It is not clear if you are saying that the driver can only work on
>SoCs with a DPAA hardware block, or the driver is actually depending
>on the DPAA drivers also. If it is the later case, you also should
>express that in the Kconfig you added below.
>
[Peng Ma] Ok, I will express it in the Kconfig.
>>
>> Signed-off-by: Wen He <[email protected]>
>> Signed-off-by: Jiaheng Fan <[email protected]>
>> Signed-off-by: Peng Ma <[email protected]>
>> ---
>> change in v10:
>> - no
>>
>> drivers/dma/Kconfig | 13 +
>> drivers/dma/Makefile | 1 +
>> drivers/dma/fsl-qdma.c | 1257
>++++++++++++++++++++++++++++++++++++++++++++++++
>> 3 files changed, 1271 insertions(+), 0 deletions(-)
>> create mode 100644 drivers/dma/fsl-qdma.c
>>
>> diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
>> index dacf3f4..50e19d7 100644
>> --- a/drivers/dma/Kconfig
>> +++ b/drivers/dma/Kconfig
>> @@ -218,6 +218,19 @@ config FSL_EDMA
>> multiplexing capability for DMA request sources(slot).
>> This module can be found on Freescale Vybrid and LS-1 SoCs.
>>
>> +config FSL_QDMA
>> + tristate "NXP Layerscape qDMA engine support"
>> + depends on ARM || ARM64
>> + select DMA_ENGINE
>> + select DMA_VIRTUAL_CHANNELS
>> + select DMA_ENGINE_RAID
>> + select ASYNC_TX_ENABLE_CHANNEL_SWITCH
>> + help
>> + Support the NXP Layerscape qDMA engine with command queue
>and legacy mode.
>> + Channel virtualization is supported through enqueuing of DMA
>jobs to,
>> + or dequeuing DMA jobs from, different work queues.
>> + This module can be found on NXP Layerscape SoCs.
>> +
>> config FSL_RAID
>> tristate "Freescale RAID engine Support"
>> depends on FSL_SOC
>&& !ASYNC_TX_ENABLE_CHANNEL_SWITCH
>> diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
>> index c91702d..2d1b586 100644
>> --- a/drivers/dma/Makefile
>> +++ b/drivers/dma/Makefile
>> @@ -32,6 +32,7 @@ obj-$(CONFIG_DW_DMAC_CORE) += dw/
>> obj-$(CONFIG_EP93XX_DMA) += ep93xx_dma.o
>> obj-$(CONFIG_FSL_DMA) += fsldma.o
>> obj-$(CONFIG_FSL_EDMA) += fsl-edma.o
>> +obj-$(CONFIG_FSL_QDMA) += fsl-qdma.o
>> obj-$(CONFIG_FSL_RAID) += fsl_raid.o
>> obj-$(CONFIG_HSU_DMA) += hsu/
>> obj-$(CONFIG_IMG_MDC_DMA) += img-mdc-dma.o
>> diff --git a/drivers/dma/fsl-qdma.c b/drivers/dma/fsl-qdma.c
>> new file mode 100644
>> index 0000000..404869e
>> --- /dev/null
>> +++ b/drivers/dma/fsl-qdma.c
>> @@ -0,0 +1,1257 @@
>> +// SPDX-License-Identifier: GPL-2.0
>> +// Copyright 2018 NXP
>
>I'm not sure if this is really the case. The driver at least has been
>sent out in 2015. We should keep these copyright claims, even the
>legacy Freescale copyright claims.
>
[Peng Ma]
I am not sure this patch sent out in 2015, but the git log earliest shows the patch cteated at Dec 20 2017. so if i changed the "Copyright 2018 NXP" to " Copyright 2017-2018 NXP "
Best regards
Peng Ma
>> +
>> +/*
>> + * Driver for NXP Layerscape Queue Direct Memory Access Controller
>> + *
>> + * Author:
>> + * Wen He <[email protected]>
>> + * Jiaheng Fan <[email protected]>
>> + *
>> + */
>> +
>> +#include <linux/module.h>
>> +#include <linux/delay.h>
>> +#include <linux/of_irq.h>
>> +#include <linux/of_platform.h>
>> +#include <linux/of_dma.h>
>> +#include <linux/dma-mapping.h>
>> +
>> +#include "virt-dma.h"
>> +#include "fsldma.h"
>> +
>> +/* Register related definition */
>> +#define FSL_QDMA_DMR 0x0
>> +#define FSL_QDMA_DSR 0x4
>> +#define FSL_QDMA_DEIER 0xe00
>> +#define FSL_QDMA_DEDR 0xe04
>> +#define FSL_QDMA_DECFDW0R 0xe10
>> +#define FSL_QDMA_DECFDW1R 0xe14
>> +#define FSL_QDMA_DECFDW2R 0xe18
>> +#define FSL_QDMA_DECFDW3R 0xe1c
>> +#define FSL_QDMA_DECFQIDR 0xe30
>> +#define FSL_QDMA_DECBR 0xe34
>> +
>> +#define FSL_QDMA_BCQMR(x) (0xc0 + 0x100 * (x))
>> +#define FSL_QDMA_BCQSR(x) (0xc4 + 0x100 * (x))
>> +#define FSL_QDMA_BCQEDPA_SADDR(x) (0xc8 + 0x100 * (x))
>> +#define FSL_QDMA_BCQDPA_SADDR(x) (0xcc + 0x100 * (x))
>> +#define FSL_QDMA_BCQEEPA_SADDR(x) (0xd0 + 0x100 * (x))
>> +#define FSL_QDMA_BCQEPA_SADDR(x) (0xd4 + 0x100 * (x))
>> +#define FSL_QDMA_BCQIER(x) (0xe0 + 0x100 * (x))
>> +#define FSL_QDMA_BCQIDR(x) (0xe4 + 0x100 * (x))
>> +
>> +#define FSL_QDMA_SQDPAR 0x80c
>> +#define FSL_QDMA_SQEPAR 0x814
>> +#define FSL_QDMA_BSQMR 0x800
>> +#define FSL_QDMA_BSQSR 0x804
>> +#define FSL_QDMA_BSQICR 0x828
>> +#define FSL_QDMA_CQMR 0xa00
>> +#define FSL_QDMA_CQDSCR1 0xa08
>> +#define FSL_QDMA_CQDSCR2 0xa0c
>> +#define FSL_QDMA_CQIER 0xa10
>> +#define FSL_QDMA_CQEDR 0xa14
>> +#define FSL_QDMA_SQCCMR 0xa20
>> +
>> +/* Registers for bit and genmask */
>> +#define FSL_QDMA_CQIDR_SQT BIT(15)
>> +#define QDMA_CCDF_FOTMAT BIT(29)
>> +#define QDMA_CCDF_SER BIT(30)
>> +#define QDMA_SG_FIN BIT(30)
>> +#define QDMA_SG_LEN_MASK GENMASK(29, 0)
>> +#define QDMA_CCDF_MASK GENMASK(28, 20)
>> +
>> +#define FSL_QDMA_DEDR_CLEAR GENMASK(31, 0)
>> +#define FSL_QDMA_BCQIDR_CLEAR GENMASK(31, 0)
>> +#define FSL_QDMA_DEIER_CLEAR GENMASK(31, 0)
>> +
>> +#define FSL_QDMA_BCQIER_CQTIE BIT(15)
>> +#define FSL_QDMA_BCQIER_CQPEIE BIT(23)
>> +#define FSL_QDMA_BSQICR_ICEN BIT(31)
>> +
>> +#define FSL_QDMA_BSQICR_ICST(x) ((x) << 16)
>> +#define FSL_QDMA_CQIER_MEIE BIT(31)
>> +#define FSL_QDMA_CQIER_TEIE BIT(0)
>> +#define FSL_QDMA_SQCCMR_ENTER_WM BIT(21)
>> +
>> +#define FSL_QDMA_BCQMR_EN BIT(31)
>> +#define FSL_QDMA_BCQMR_EI BIT(30)
>> +#define FSL_QDMA_BCQMR_CD_THLD(x) ((x) << 20)
>> +#define FSL_QDMA_BCQMR_CQ_SIZE(x) ((x) << 16)
>> +
>> +#define FSL_QDMA_BCQSR_QF BIT(16)
>> +#define FSL_QDMA_BCQSR_XOFF BIT(0)
>> +
>> +#define FSL_QDMA_BSQMR_EN BIT(31)
>> +#define FSL_QDMA_BSQMR_DI BIT(30)
>> +#define FSL_QDMA_BSQMR_CQ_SIZE(x) ((x) << 16)
>> +
>> +#define FSL_QDMA_BSQSR_QE BIT(17)
>> +
>> +#define FSL_QDMA_DMR_DQD BIT(30)
>> +#define FSL_QDMA_DSR_DB BIT(31)
>> +
>> +/* Size related definition */
>> +#define FSL_QDMA_QUEUE_MAX 8
>> +#define FSL_QDMA_COMMAND_BUFFER_SIZE 64
>> +#define FSL_QDMA_DESCRIPTOR_BUFFER_SIZE 32
>> +#define FSL_QDMA_CIRCULAR_DESC_SIZE_MIN 64
>> +#define FSL_QDMA_CIRCULAR_DESC_SIZE_MAX 16384
>> +#define FSL_QDMA_QUEUE_NUM_MAX 8
>> +
>> +/* Field definition for CMD */
>> +#define FSL_QDMA_CMD_RWTTYPE 0x4
>> +#define FSL_QDMA_CMD_LWC 0x2
>> +#define FSL_QDMA_CMD_RWTTYPE_OFFSET 28
>> +#define FSL_QDMA_CMD_NS_OFFSET 27
>> +#define FSL_QDMA_CMD_DQOS_OFFSET 24
>> +#define FSL_QDMA_CMD_WTHROTL_OFFSET 20
>> +#define FSL_QDMA_CMD_DSEN_OFFSET 19
>> +#define FSL_QDMA_CMD_LWC_OFFSET 16
>> +
>> +/* Field definition for Descriptor offset */
>> +#define QDMA_CCDF_STATUS 20
>> +#define QDMA_CCDF_OFFSET 20
>> +
>> +/* Field definition for safe loop count*/
>> +#define FSL_QDMA_HALT_COUNT 1500
>> +#define FSL_QDMA_MAX_SIZE 16385
>> +#define FSL_QDMA_COMP_TIMEOUT 1000
>> +#define FSL_COMMAND_QUEUE_OVERFLLOW 10
>> +
>> +#define FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma_engine, x)
>\
>> + (((fsl_qdma_engine)->block_offset) * (x))
>> +
>> +/**
>> + * struct fsl_qdma_format - This is the struct holding describing compound
>> + * descriptor format with qDMA.
>> + * @status: Command status and enqueue status
>notification.
>> + * @cfg: Frame offset and frame format.
>> + * @addr_lo: Holding the compound descriptor of the
>lower
>> + * 32-bits address in memory 40-bit address.
>> + * @addr_hi: Same as above member, but point high
>8-bits in
>> + * memory 40-bit address.
>> + * @__reserved1: Reserved field.
>> + * @cfg8b_w1: Compound descriptor command queue
>origin produced
>> + * by qDMA and dynamic debug field.
>> + * @data Pointer to the memory 40-bit address,
>describes DMA
>> + * source information and DMA destination
>information.
>> + */
>> +struct fsl_qdma_format {
>> + __le32 status;
>> + __le32 cfg;
>> + union {
>> + struct {
>> + __le32 addr_lo;
>> + u8 addr_hi;
>> + u8 __reserved1[2];
>> + u8 cfg8b_w1;
>> + } __packed;
>> + __le64 data;
>> + };
>> +} __packed;
>> +
>> +/* qDMA status notification pre information */
>> +struct fsl_pre_status {
>> + u64 addr;
>> + u8 queue;
>> +};
>> +
>> +static DEFINE_PER_CPU(struct fsl_pre_status, pre);
>> +
>> +struct fsl_qdma_chan {
>> + struct virt_dma_chan vchan;
>> + struct virt_dma_desc vdesc;
>> + enum dma_status status;
>> + struct fsl_qdma_engine *qdma;
>> + struct fsl_qdma_queue *queue;
>> +};
>> +
>> +struct fsl_qdma_queue {
>> + struct fsl_qdma_format *virt_head;
>> + struct fsl_qdma_format *virt_tail;
>> + struct list_head comp_used;
>> + struct list_head comp_free;
>> + struct dma_pool *comp_pool;
>> + struct dma_pool *desc_pool;
>> + spinlock_t queue_lock;
>> + dma_addr_t bus_addr;
>> + u32 n_cq;
>> + u32 id;
>> + struct fsl_qdma_format *cq;
>> + void __iomem *block_base;
>> +};
>> +
>> +struct fsl_qdma_comp {
>> + dma_addr_t bus_addr;
>> + dma_addr_t desc_bus_addr;
>> + struct fsl_qdma_format *virt_addr;
>> + struct fsl_qdma_format *desc_virt_addr;
>> + struct fsl_qdma_chan *qchan;
>> + struct virt_dma_desc vdesc;
>> + struct list_head list;
>> +};
>> +
>> +struct fsl_qdma_engine {
>> + struct dma_device dma_dev;
>> + void __iomem *ctrl_base;
>> + void __iomem *status_base;
>> + void __iomem *block_base;
>> + u32 n_chans;
>> + u32 n_queues;
>> + struct mutex fsl_qdma_mutex;
>> + int error_irq;
>> + int *queue_irq;
>> + u32 feature;
>> + struct fsl_qdma_queue *queue;
>> + struct fsl_qdma_queue **status;
>> + struct fsl_qdma_chan *chans;
>> + int block_number;
>> + int block_offset;
>> + int irq_base;
>> + int desc_allocated;
>> +
>> +};
>> +
>> +static inline u64
>> +qdma_ccdf_addr_get64(const struct fsl_qdma_format *ccdf)
>> +{
>> + return le64_to_cpu(ccdf->data) & (U64_MAX >> 24);
>> +}
>> +
>> +static inline void
>> +qdma_desc_addr_set64(struct fsl_qdma_format *ccdf, u64 addr)
>> +{
>> + ccdf->addr_hi = upper_32_bits(addr);
>> + ccdf->addr_lo = cpu_to_le32(lower_32_bits(addr));
>> +}
>> +
>> +static inline u8
>> +qdma_ccdf_get_queue(const struct fsl_qdma_format *ccdf)
>> +{
>> + return ccdf->cfg8b_w1 & U8_MAX;
>> +}
>> +
>> +static inline int
>> +qdma_ccdf_get_offset(const struct fsl_qdma_format *ccdf)
>> +{
>> + return (le32_to_cpu(ccdf->cfg) & QDMA_CCDF_MASK) >>
>QDMA_CCDF_OFFSET;
>> +}
>> +
>> +static inline void
>> +qdma_ccdf_set_format(struct fsl_qdma_format *ccdf, int offset)
>> +{
>> + ccdf->cfg = cpu_to_le32(QDMA_CCDF_FOTMAT | offset);
>> +}
>> +
>> +static inline int
>> +qdma_ccdf_get_status(const struct fsl_qdma_format *ccdf)
>> +{
>> + return (le32_to_cpu(ccdf->status) & QDMA_CCDF_MASK) >>
>QDMA_CCDF_STATUS;
>> +}
>> +
>> +static inline void
>> +qdma_ccdf_set_ser(struct fsl_qdma_format *ccdf, int status)
>> +{
>> + ccdf->status = cpu_to_le32(QDMA_CCDF_SER | status);
>> +}
>> +
>> +static inline void qdma_csgf_set_len(struct fsl_qdma_format *csgf, int len)
>> +{
>> + csgf->cfg = cpu_to_le32(len & QDMA_SG_LEN_MASK);
>> +}
>> +
>> +static inline void qdma_csgf_set_f(struct fsl_qdma_format *csgf, int len)
>> +{
>> + csgf->cfg = cpu_to_le32(QDMA_SG_FIN | (len &
>QDMA_SG_LEN_MASK));
>> +}
>> +
>> +static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem
>*addr)
>> +{
>> + return FSL_DMA_IN(qdma, addr, 32);
>> +}
>> +
>> +static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val,
>> + void __iomem *addr)
>> +{
>> + FSL_DMA_OUT(qdma, addr, val, 32);
>> +}
>> +
>> +static struct fsl_qdma_chan *to_fsl_qdma_chan(struct dma_chan *chan)
>> +{
>> + return container_of(chan, struct fsl_qdma_chan, vchan.chan);
>> +}
>> +
>> +static struct fsl_qdma_comp *to_fsl_qdma_comp(struct virt_dma_desc *vd)
>> +{
>> + return container_of(vd, struct fsl_qdma_comp, vdesc);
>> +}
>> +
>> +static void fsl_qdma_free_chan_resources(struct dma_chan *chan)
>> +{
>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>> + struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
>> + struct fsl_qdma_comp *comp_temp, *_comp_temp;
>> + unsigned long flags;
>> + LIST_HEAD(head);
>> +
>> + spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
>> + vchan_get_all_descriptors(&fsl_chan->vchan, &head);
>> + spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
>> +
>> + vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
>> +
>> + if (!fsl_queue->comp_pool && !fsl_queue->comp_pool)
>> + return;
>> +
>> + list_for_each_entry_safe(comp_temp, _comp_temp,
>> + &fsl_queue->comp_used, list) {
>> + dma_pool_free(fsl_queue->comp_pool,
>> + comp_temp->virt_addr,
>> + comp_temp->bus_addr);
>> + dma_pool_free(fsl_queue->desc_pool,
>> + comp_temp->desc_virt_addr,
>> + comp_temp->desc_bus_addr);
>> + list_del(&comp_temp->list);
>> + kfree(comp_temp);
>> + }
>> +
>> + list_for_each_entry_safe(comp_temp, _comp_temp,
>> + &fsl_queue->comp_free, list) {
>> + dma_pool_free(fsl_queue->comp_pool,
>> + comp_temp->virt_addr,
>> + comp_temp->bus_addr);
>> + dma_pool_free(fsl_queue->desc_pool,
>> + comp_temp->desc_virt_addr,
>> + comp_temp->desc_bus_addr);
>> + list_del(&comp_temp->list);
>> + kfree(comp_temp);
>> + }
>> +
>> + dma_pool_destroy(fsl_queue->comp_pool);
>> + dma_pool_destroy(fsl_queue->desc_pool);
>> +
>> + fsl_qdma->desc_allocated--;
>> + fsl_queue->comp_pool = NULL;
>> + fsl_queue->desc_pool = NULL;
>> +}
>> +
>> +static void fsl_qdma_comp_fill_memcpy(struct fsl_qdma_comp *fsl_comp,
>> + dma_addr_t dst, dma_addr_t
>src, u32 len)
>> +{
>> + struct fsl_qdma_format *sdf, *ddf;
>> + struct fsl_qdma_format *ccdf, *csgf_desc, *csgf_src, *csgf_dest;
>> +
>> + ccdf = fsl_comp->virt_addr;
>> + csgf_desc = fsl_comp->virt_addr + 1;
>> + csgf_src = fsl_comp->virt_addr + 2;
>> + csgf_dest = fsl_comp->virt_addr + 3;
>> + sdf = fsl_comp->desc_virt_addr;
>> + ddf = fsl_comp->desc_virt_addr + 1;
>> +
>> + memset(fsl_comp->virt_addr, 0,
>FSL_QDMA_COMMAND_BUFFER_SIZE);
>> + memset(fsl_comp->desc_virt_addr, 0,
>FSL_QDMA_DESCRIPTOR_BUFFER_SIZE);
>> + /* Head Command Descriptor(Frame Descriptor) */
>> + qdma_desc_addr_set64(ccdf, fsl_comp->bus_addr + 16);
>> + qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf));
>> + qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf));
>> + /* Status notification is enqueued to status queue. */
>> + /* Compound Command Descriptor(Frame List Table) */
>> + qdma_desc_addr_set64(csgf_desc, fsl_comp->desc_bus_addr);
>> + /* It must be 32 as Compound S/G Descriptor */
>> + qdma_csgf_set_len(csgf_desc, 32);
>> + qdma_desc_addr_set64(csgf_src, src);
>> + qdma_csgf_set_len(csgf_src, len);
>> + qdma_desc_addr_set64(csgf_dest, dst);
>> + qdma_csgf_set_len(csgf_dest, len);
>> + /* This entry is the last entry. */
>> + qdma_csgf_set_f(csgf_dest, len);
>> + /* Descriptor Buffer */
>> + sdf->data =
>> + cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
>> + FSL_QDMA_CMD_RWTTYPE_OFFSET);
>> + ddf->data =
>> + cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
>> + FSL_QDMA_CMD_RWTTYPE_OFFSET);
>> + ddf->data |=
>> + cpu_to_le64(FSL_QDMA_CMD_LWC <<
>FSL_QDMA_CMD_LWC_OFFSET);
>> +}
>> +
>> +/*
>> + * Pre-request full command descriptor for enqueue.
>> + */
>> +static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue
>*queue)
>> +{
>> + int i;
>> + struct fsl_qdma_comp *comp_temp, *_comp_temp;
>> +
>> + for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW;
>i++) {
>> + comp_temp = kzalloc(sizeof(*comp_temp),
>GFP_KERNEL);
>> + if (!comp_temp)
>> + goto err_alloc;
>> + comp_temp->virt_addr =
>> + dma_pool_alloc(queue->comp_pool,
>GFP_KERNEL,
>> + &comp_temp->bus_addr);
>> + if (!comp_temp->virt_addr)
>> + goto err_dma_alloc;
>> +
>> + comp_temp->desc_virt_addr =
>> + dma_pool_alloc(queue->desc_pool,
>GFP_KERNEL,
>> +
>&comp_temp->desc_bus_addr);
>> + if (!comp_temp->desc_virt_addr)
>> + goto err_desc_dma_alloc;
>> +
>> + list_add_tail(&comp_temp->list, &queue->comp_free);
>> + }
>> +
>> + return 0;
>> +
>> +err_desc_dma_alloc:
>> + dma_pool_free(queue->comp_pool, comp_temp->virt_addr,
>> + comp_temp->bus_addr);
>> +
>> +err_dma_alloc:
>> + kfree(comp_temp);
>> +
>> +err_alloc:
>> + list_for_each_entry_safe(comp_temp, _comp_temp,
>> + &queue->comp_free, list) {
>> + if (comp_temp->virt_addr)
>> + dma_pool_free(queue->comp_pool,
>> + comp_temp->virt_addr,
>> + comp_temp->bus_addr);
>> + if (comp_temp->desc_virt_addr)
>> + dma_pool_free(queue->desc_pool,
>> + comp_temp->desc_virt_addr,
>> +
>comp_temp->desc_bus_addr);
>> +
>> + list_del(&comp_temp->list);
>> + kfree(comp_temp);
>> + }
>> +
>> + return -ENOMEM;
>> +}
>> +
>> +/*
>> + * Request a command descriptor for enqueue.
>> + */
>> +static struct fsl_qdma_comp
>> +*fsl_qdma_request_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
>> +{
>> + unsigned long flags;
>> + struct fsl_qdma_comp *comp_temp;
>> + int timeout = FSL_QDMA_COMP_TIMEOUT;
>> + struct fsl_qdma_queue *queue = fsl_chan->queue;
>> +
>> + while (timeout--) {
>> + spin_lock_irqsave(&queue->queue_lock, flags);
>> + if (!list_empty(&queue->comp_free)) {
>> + comp_temp =
>list_first_entry(&queue->comp_free,
>> + struct
>fsl_qdma_comp,
>> + list);
>> + list_del(&comp_temp->list);
>> +
>> + spin_unlock_irqrestore(&queue->queue_lock,
>flags);
>> + comp_temp->qchan = fsl_chan;
>> + return comp_temp;
>> + }
>> + spin_unlock_irqrestore(&queue->queue_lock, flags);
>> + udelay(1);
>> + }
>> +
>> + return NULL;
>> +}
>> +
>> +static struct fsl_qdma_queue
>> +*fsl_qdma_alloc_queue_resources(struct platform_device *pdev,
>> + struct fsl_qdma_engine *fsl_qdma)
>> +{
>> + int ret, len, i, j;
>> + int queue_num, block_number;
>> + unsigned int queue_size[FSL_QDMA_QUEUE_MAX];
>> + struct fsl_qdma_queue *queue_head, *queue_temp;
>> +
>> + queue_num = fsl_qdma->n_queues;
>> + block_number = fsl_qdma->block_number;
>> +
>> + if (queue_num > FSL_QDMA_QUEUE_MAX)
>> + queue_num = FSL_QDMA_QUEUE_MAX;
>> + len = sizeof(*queue_head) * queue_num * block_number;
>> + queue_head = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>> + if (!queue_head)
>> + return NULL;
>> +
>> + ret = device_property_read_u32_array(&pdev->dev, "queue-sizes",
>> + queue_size,
>queue_num);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't get queue-sizes.\n");
>> + return NULL;
>> + }
>> + for (j = 0; j < block_number; j++) {
>> + for (i = 0; i < queue_num; i++) {
>> + if (queue_size[i] >
>FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
>> + queue_size[i] <
>FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
>> + dev_err(&pdev->dev,
>> + "Get wrong
>queue-sizes.\n");
>> + return NULL;
>> + }
>> + queue_temp = queue_head + i + (j *
>queue_num);
>> +
>> + queue_temp->cq =
>> + dma_alloc_coherent(&pdev->dev,
>> + sizeof(struct
>fsl_qdma_format) *
>> + queue_size[i],
>> +
>&queue_temp->bus_addr,
>> + GFP_KERNEL);
>> + if (!queue_temp->cq)
>> + return NULL;
>> + queue_temp->block_base =
>fsl_qdma->block_base +
>> +
>FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>> + queue_temp->n_cq = queue_size[i];
>> + queue_temp->id = i;
>> + queue_temp->virt_head = queue_temp->cq;
>> + queue_temp->virt_tail = queue_temp->cq;
>> + /*
>> + * List for queue command buffer
>> + */
>> + INIT_LIST_HEAD(&queue_temp->comp_used);
>> + spin_lock_init(&queue_temp->queue_lock);
>> + }
>> + }
>> + return queue_head;
>> +}
>> +
>> +static struct fsl_qdma_queue
>> +*fsl_qdma_prep_status_queue(struct platform_device *pdev)
>> +{
>> + int ret;
>> + unsigned int status_size;
>> + struct fsl_qdma_queue *status_head;
>> + struct device_node *np = pdev->dev.of_node;
>> +
>> + ret = of_property_read_u32(np, "status-sizes", &status_size);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't get status-sizes.\n");
>> + return NULL;
>> + }
>> + if (status_size > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
>> + status_size < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
>> + dev_err(&pdev->dev, "Get wrong status_size.\n");
>> + return NULL;
>> + }
>> + status_head = devm_kzalloc(&pdev->dev,
>> + sizeof(*status_head),
>GFP_KERNEL);
>> + if (!status_head)
>> + return NULL;
>> +
>> + /*
>> + * Buffer for queue command
>> + */
>> + status_head->cq = dma_alloc_coherent(&pdev->dev,
>> + sizeof(struct
>fsl_qdma_format) *
>> + status_size,
>> +
>&status_head->bus_addr,
>> + GFP_KERNEL);
>> + if (!status_head->cq) {
>> + devm_kfree(&pdev->dev, status_head);
>> + return NULL;
>> + }
>> + status_head->n_cq = status_size;
>> + status_head->virt_head = status_head->cq;
>> + status_head->virt_tail = status_head->cq;
>> + status_head->comp_pool = NULL;
>> +
>> + return status_head;
>> +}
>> +
>> +static int fsl_qdma_halt(struct fsl_qdma_engine *fsl_qdma)
>> +{
>> + u32 reg;
>> + int i, j, count = FSL_QDMA_HALT_COUNT;
>> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
>> +
>> + /* Disable the command queue and wait for idle state. */
>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
>> + reg |= FSL_QDMA_DMR_DQD;
>> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
>> + for (j = 0; j < fsl_qdma->block_number; j++) {
>> + block = fsl_qdma->block_base +
>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>> + for (i = 0; i < FSL_QDMA_QUEUE_NUM_MAX; i++)
>> + qdma_writel(fsl_qdma, 0, block +
>FSL_QDMA_BCQMR(i));
>> + }
>> + while (1) {
>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DSR);
>> + if (!(reg & FSL_QDMA_DSR_DB))
>> + break;
>> + if (count-- < 0)
>> + return -EBUSY;
>> + udelay(100);
>> + }
>> +
>> + for (j = 0; j < fsl_qdma->block_number; j++) {
>> + block = fsl_qdma->block_base +
>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>> +
>> + /* Disable status queue. */
>> + qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BSQMR);
>> +
>> + /*
>> + * clear the command queue interrupt detect register for
>> + * all queues.
>> + */
>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
>> + block + FSL_QDMA_BCQIDR(0));
>> + }
>> +
>> + return 0;
>> +}
>> +
>> +static int
>> +fsl_qdma_queue_transfer_complete(struct fsl_qdma_engine *fsl_qdma,
>> + void *block,
>> + int id)
>> +{
>> + bool duplicate;
>> + u32 reg, i, count;
>> + struct fsl_qdma_queue *temp_queue;
>> + struct fsl_qdma_format *status_addr;
>> + struct fsl_qdma_comp *fsl_comp = NULL;
>> + struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
>> + struct fsl_qdma_queue *fsl_status = fsl_qdma->status[id];
>> +
>> + count = FSL_QDMA_MAX_SIZE;
>> +
>> + while (count--) {
>> + duplicate = 0;
>> + reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQSR);
>> + if (reg & FSL_QDMA_BSQSR_QE)
>> + return 0;
>> +
>> + status_addr = fsl_status->virt_head;
>> +
>> + if (qdma_ccdf_get_queue(status_addr) ==
>> + __this_cpu_read(pre.queue) &&
>> + qdma_ccdf_addr_get64(status_addr) ==
>> + __this_cpu_read(pre.addr))
>> + duplicate = 1;
>> + i = qdma_ccdf_get_queue(status_addr) +
>> + id * fsl_qdma->n_queues;
>> + __this_cpu_write(pre.addr,
>qdma_ccdf_addr_get64(status_addr));
>> + __this_cpu_write(pre.queue,
>qdma_ccdf_get_queue(status_addr));
>> + temp_queue = fsl_queue + i;
>> +
>> + spin_lock(&temp_queue->queue_lock);
>> + if (list_empty(&temp_queue->comp_used)) {
>> + if (!duplicate) {
>> +
>spin_unlock(&temp_queue->queue_lock);
>> + return -EAGAIN;
>> + }
>> + } else {
>> + fsl_comp =
>list_first_entry(&temp_queue->comp_used,
>> + struct
>fsl_qdma_comp, list);
>> + if (fsl_comp->bus_addr + 16 !=
>> + __this_cpu_read(pre.addr)) {
>> + if (!duplicate) {
>> +
>spin_unlock(&temp_queue->queue_lock);
>> + return -EAGAIN;
>> + }
>> + }
>> + }
>> +
>> + if (duplicate) {
>> + reg = qdma_readl(fsl_qdma, block +
>FSL_QDMA_BSQMR);
>> + reg |= FSL_QDMA_BSQMR_DI;
>> + qdma_desc_addr_set64(status_addr, 0x0);
>> + fsl_status->virt_head++;
>> + if (fsl_status->virt_head == fsl_status->cq
>> + +
>fsl_status->n_cq)
>> + fsl_status->virt_head =
>fsl_status->cq;
>> + qdma_writel(fsl_qdma, reg, block +
>FSL_QDMA_BSQMR);
>> + spin_unlock(&temp_queue->queue_lock);
>> + continue;
>> + }
>> + list_del(&fsl_comp->list);
>> +
>> + reg = qdma_readl(fsl_qdma, block +
>FSL_QDMA_BSQMR);
>> + reg |= FSL_QDMA_BSQMR_DI;
>> + qdma_desc_addr_set64(status_addr, 0x0);
>> + fsl_status->virt_head++;
>> + if (fsl_status->virt_head == fsl_status->cq +
>fsl_status->n_cq)
>> + fsl_status->virt_head = fsl_status->cq;
>> + qdma_writel(fsl_qdma, reg, block +
>FSL_QDMA_BSQMR);
>> + spin_unlock(&temp_queue->queue_lock);
>> +
>> + spin_lock(&fsl_comp->qchan->vchan.lock);
>> + vchan_cookie_complete(&fsl_comp->vdesc);
>> + fsl_comp->qchan->status = DMA_COMPLETE;
>> + spin_unlock(&fsl_comp->qchan->vchan.lock);
>> + }
>> +
>> + return 0;
>> +}
>> +
>> +static irqreturn_t fsl_qdma_error_handler(int irq, void *dev_id)
>> +{
>> + unsigned int intr;
>> + struct fsl_qdma_engine *fsl_qdma = dev_id;
>> + void __iomem *status = fsl_qdma->status_base;
>> +
>> + intr = qdma_readl(fsl_qdma, status + FSL_QDMA_DEDR);
>> +
>> + if (intr) {
>> + dev_err(fsl_qdma->dma_dev.dev, "DMA transaction
>error!\n");
>> + return IRQ_NONE;
>> + }
>> +
>> + qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status +
>FSL_QDMA_DEDR);
>> + return IRQ_HANDLED;
>> +}
>> +
>> +static irqreturn_t fsl_qdma_queue_handler(int irq, void *dev_id)
>> +{
>> + int id;
>> + unsigned int intr, reg;
>> + struct fsl_qdma_engine *fsl_qdma = dev_id;
>> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
>> +
>> + id = irq - fsl_qdma->irq_base;
>> + if (id < 0 && id > fsl_qdma->block_number) {
>> + dev_err(fsl_qdma->dma_dev.dev,
>> + "irq %d is wrong irq_base is %d\n",
>> + irq, fsl_qdma->irq_base);
>> + }
>> +
>> + block = fsl_qdma->block_base +
>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, id);
>> +
>> + intr = qdma_readl(fsl_qdma, block + FSL_QDMA_BCQIDR(0));
>> +
>> + if ((intr & FSL_QDMA_CQIDR_SQT) != 0)
>> + intr = fsl_qdma_queue_transfer_complete(fsl_qdma,
>block, id);
>> +
>> + if (intr != 0) {
>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
>> + reg |= FSL_QDMA_DMR_DQD;
>> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
>> + qdma_writel(fsl_qdma, 0, block +
>FSL_QDMA_BCQIER(0));
>> + dev_err(fsl_qdma->dma_dev.dev, "QDMA: status
>err!\n");
>> + }
>> +
>> + /* Clear all detected events and interrupts. */
>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
>> + block + FSL_QDMA_BCQIDR(0));
>> +
>> + return IRQ_HANDLED;
>> +}
>> +
>> +static int
>> +fsl_qdma_irq_init(struct platform_device *pdev,
>> + struct fsl_qdma_engine *fsl_qdma)
>> +{
>> + int i;
>> + int cpu;
>> + int ret;
>> + char irq_name[20];
>> +
>> + fsl_qdma->error_irq =
>> + platform_get_irq_byname(pdev, "qdma-error");
>> + if (fsl_qdma->error_irq < 0) {
>> + dev_err(&pdev->dev, "Can't get qdma controller irq.\n");
>> + return fsl_qdma->error_irq;
>> + }
>> +
>> + ret = devm_request_irq(&pdev->dev, fsl_qdma->error_irq,
>> + fsl_qdma_error_handler, 0,
>> + "qDMA error", fsl_qdma);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't register qDMA controller
>IRQ.\n");
>> + return ret;
>> + }
>> +
>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>> + sprintf(irq_name, "qdma-queue%d", i);
>> + fsl_qdma->queue_irq[i] =
>> + platform_get_irq_byname(pdev,
>irq_name);
>> +
>> + if (fsl_qdma->queue_irq[i] < 0) {
>> + dev_err(&pdev->dev,
>> + "Can't get qdma queue %d irq.\n", i);
>> + return fsl_qdma->queue_irq[i];
>> + }
>> +
>> + ret = devm_request_irq(&pdev->dev,
>> + fsl_qdma->queue_irq[i],
>> + fsl_qdma_queue_handler,
>> + 0,
>> + "qDMA queue",
>> + fsl_qdma);
>> + if (ret) {
>> + dev_err(&pdev->dev,
>> + "Can't register qDMA queue
>IRQ.\n");
>> + return ret;
>> + }
>> +
>> + cpu = i % num_online_cpus();
>> + ret = irq_set_affinity_hint(fsl_qdma->queue_irq[i],
>> + get_cpu_mask(cpu));
>> + if (ret) {
>> + dev_err(&pdev->dev,
>> + "Can't set cpu %d affinity to
>IRQ %d.\n",
>> + cpu,
>> + fsl_qdma->queue_irq[i]);
>> + return ret;
>> + }
>> + }
>> +
>> + return 0;
>> +}
>> +
>> +static void fsl_qdma_irq_exit(struct platform_device *pdev,
>> + struct fsl_qdma_engine *fsl_qdma)
>> +{
>> + int i;
>> +
>> + devm_free_irq(&pdev->dev, fsl_qdma->error_irq, fsl_qdma);
>> + for (i = 0; i < fsl_qdma->block_number; i++)
>> + devm_free_irq(&pdev->dev, fsl_qdma->queue_irq[i],
>fsl_qdma);
>> +}
>> +
>> +static int fsl_qdma_reg_init(struct fsl_qdma_engine *fsl_qdma)
>> +{
>> + u32 reg;
>> + int i, j, ret;
>> + struct fsl_qdma_queue *temp;
>> + void __iomem *status = fsl_qdma->status_base;
>> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
>> + struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
>> +
>> + /* Try to halt the qDMA engine first. */
>> + ret = fsl_qdma_halt(fsl_qdma);
>> + if (ret) {
>> + dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
>> + return ret;
>> + }
>> +
>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>> + /*
>> + * Clear the command queue interrupt detect register
>for
>> + * all queues.
>> + */
>> +
>> + block = fsl_qdma->block_base +
>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, i);
>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
>> + block + FSL_QDMA_BCQIDR(0));
>> + }
>> +
>> + for (j = 0; j < fsl_qdma->block_number; j++) {
>> + block = fsl_qdma->block_base +
>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>> + for (i = 0; i < fsl_qdma->n_queues; i++) {
>> + temp = fsl_queue + i + (j *
>fsl_qdma->n_queues);
>> + /*
>> + * Initialize Command Queue registers to
>> + * point to the first
>> + * command descriptor in memory.
>> + * Dequeue Pointer Address Registers
>> + * Enqueue Pointer Address Registers
>> + */
>> +
>> + qdma_writel(fsl_qdma, temp->bus_addr,
>> + block +
>FSL_QDMA_BCQDPA_SADDR(i));
>> + qdma_writel(fsl_qdma, temp->bus_addr,
>> + block +
>FSL_QDMA_BCQEPA_SADDR(i));
>> +
>> + /* Initialize the queue mode. */
>> + reg = FSL_QDMA_BCQMR_EN;
>> + reg |=
>FSL_QDMA_BCQMR_CD_THLD(ilog2(temp->n_cq) - 4);
>> + reg |=
>FSL_QDMA_BCQMR_CQ_SIZE(ilog2(temp->n_cq) - 6);
>> + qdma_writel(fsl_qdma, reg, block +
>FSL_QDMA_BCQMR(i));
>> + }
>> +
>> + /*
>> + * Workaround for erratum: ERR010812.
>> + * We must enable XOFF to avoid the enqueue rejection
>occurs.
>> + * Setting SQCCMR ENTER_WM to 0x20.
>> + */
>> +
>> + qdma_writel(fsl_qdma,
>FSL_QDMA_SQCCMR_ENTER_WM,
>> + block + FSL_QDMA_SQCCMR);
>> +
>> + /*
>> + * Initialize status queue registers to point to the first
>> + * command descriptor in memory.
>> + * Dequeue Pointer Address Registers
>> + * Enqueue Pointer Address Registers
>> + */
>> +
>> + qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
>> + block + FSL_QDMA_SQEPAR);
>> + qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
>> + block + FSL_QDMA_SQDPAR);
>> + /* Initialize status queue interrupt. */
>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIER_CQTIE,
>> + block + FSL_QDMA_BCQIER(0));
>> + qdma_writel(fsl_qdma, FSL_QDMA_BSQICR_ICEN |
>> + FSL_QDMA_BSQICR_ICST(5) |
>0x8000,
>> + block + FSL_QDMA_BSQICR);
>> + qdma_writel(fsl_qdma, FSL_QDMA_CQIER_MEIE |
>> + FSL_QDMA_CQIER_TEIE,
>> + block + FSL_QDMA_CQIER);
>> +
>> + /* Initialize the status queue mode. */
>> + reg = FSL_QDMA_BSQMR_EN;
>> + reg |= FSL_QDMA_BSQMR_CQ_SIZE(ilog2
>> + (fsl_qdma->status[j]->n_cq) - 6);
>> +
>> + qdma_writel(fsl_qdma, reg, block +
>FSL_QDMA_BSQMR);
>> + reg = qdma_readl(fsl_qdma, block +
>FSL_QDMA_BSQMR);
>> + }
>> +
>> + /* Initialize controller interrupt register. */
>> + qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status +
>FSL_QDMA_DEDR);
>> + qdma_writel(fsl_qdma, FSL_QDMA_DEIER_CLEAR, status +
>FSL_QDMA_DEIER);
>> +
>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
>> + reg &= ~FSL_QDMA_DMR_DQD;
>> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
>> +
>> + return 0;
>> +}
>> +
>> +static struct dma_async_tx_descriptor *
>> +fsl_qdma_prep_memcpy(struct dma_chan *chan, dma_addr_t dst,
>> + dma_addr_t src, size_t len, unsigned long flags)
>> +{
>> + struct fsl_qdma_comp *fsl_comp;
>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>> +
>> + fsl_comp = fsl_qdma_request_enqueue_desc(fsl_chan);
>> +
>> + if (!fsl_comp)
>> + return NULL;
>> +
>> + fsl_qdma_comp_fill_memcpy(fsl_comp, dst, src, len);
>> +
>> + return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
>> +}
>> +
>> +static void fsl_qdma_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
>> +{
>> + u32 reg;
>> + struct virt_dma_desc *vdesc;
>> + struct fsl_qdma_comp *fsl_comp;
>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>> + void __iomem *block = fsl_queue->block_base;
>> +
>> + reg = qdma_readl(fsl_chan->qdma, block +
>FSL_QDMA_BCQSR(fsl_queue->id));
>> + if (reg & (FSL_QDMA_BCQSR_QF | FSL_QDMA_BCQSR_XOFF))
>> + return;
>> + vdesc = vchan_next_desc(&fsl_chan->vchan);
>> + if (!vdesc)
>> + return;
>> + list_del(&vdesc->node);
>> + fsl_comp = to_fsl_qdma_comp(vdesc);
>> +
>> + memcpy(fsl_queue->virt_head++,
>> + fsl_comp->virt_addr, sizeof(struct fsl_qdma_format));
>> + if (fsl_queue->virt_head == fsl_queue->cq + fsl_queue->n_cq)
>> + fsl_queue->virt_head = fsl_queue->cq;
>> +
>> + list_add_tail(&fsl_comp->list, &fsl_queue->comp_used);
>> + barrier();
>> + reg = qdma_readl(fsl_chan->qdma, block +
>FSL_QDMA_BCQMR(fsl_queue->id));
>> + reg |= FSL_QDMA_BCQMR_EI;
>> + qdma_writel(fsl_chan->qdma, reg, block +
>FSL_QDMA_BCQMR(fsl_queue->id));
>> + fsl_chan->status = DMA_IN_PROGRESS;
>> +}
>> +
>> +static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc)
>> +{
>> + unsigned long flags;
>> + struct fsl_qdma_comp *fsl_comp;
>> + struct fsl_qdma_queue *fsl_queue;
>> +
>> + fsl_comp = to_fsl_qdma_comp(vdesc);
>> + fsl_queue = fsl_comp->qchan->queue;
>> +
>> + spin_lock_irqsave(&fsl_queue->queue_lock, flags);
>> + list_add_tail(&fsl_comp->list, &fsl_queue->comp_free);
>> + spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
>> +}
>> +
>> +static void fsl_qdma_issue_pending(struct dma_chan *chan)
>> +{
>> + unsigned long flags;
>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>> +
>> + spin_lock_irqsave(&fsl_queue->queue_lock, flags);
>> + spin_lock(&fsl_chan->vchan.lock);
>> + if (vchan_issue_pending(&fsl_chan->vchan))
>> + fsl_qdma_enqueue_desc(fsl_chan);
>> + spin_unlock(&fsl_chan->vchan.lock);
>> + spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
>> +}
>> +
>> +static void fsl_qdma_synchronize(struct dma_chan *chan)
>> +{
>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>> +
>> + vchan_synchronize(&fsl_chan->vchan);
>> +}
>> +
>> +static int fsl_qdma_terminate_all(struct dma_chan *chan)
>> +{
>> + LIST_HEAD(head);
>> + unsigned long flags;
>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>> +
>> + spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
>> + vchan_get_all_descriptors(&fsl_chan->vchan, &head);
>> + spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
>> + vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
>> + return 0;
>> +}
>> +
>> +static int fsl_qdma_alloc_chan_resources(struct dma_chan *chan)
>> +{
>> + int ret;
>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>> + struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>> +
>> + if (fsl_queue->comp_pool && fsl_queue->desc_pool)
>> + return fsl_qdma->desc_allocated;
>> +
>> + INIT_LIST_HEAD(&fsl_queue->comp_free);
>> +
>> + /*
>> + * The dma pool for queue command buffer
>> + */
>> + fsl_queue->comp_pool =
>> + dma_pool_create("comp_pool",
>> + chan->device->dev,
>> + FSL_QDMA_COMMAND_BUFFER_SIZE,
>> + 64, 0);
>> + if (!fsl_queue->comp_pool)
>> + return -ENOMEM;
>> +
>> + /*
>> + * The dma pool for Descriptor(SD/DD) buffer
>> + */
>> + fsl_queue->desc_pool =
>> + dma_pool_create("desc_pool",
>> + chan->device->dev,
>> + FSL_QDMA_DESCRIPTOR_BUFFER_SIZE,
>> + 32, 0);
>> + if (!fsl_queue->desc_pool)
>> + goto err_desc_pool;
>> +
>> + ret = fsl_qdma_pre_request_enqueue_desc(fsl_queue);
>> + if (ret) {
>> + dev_err(chan->device->dev,
>> + "failed to alloc dma buffer for S/G
>descriptor\n");
>> + goto err_mem;
>> + }
>> +
>> + fsl_qdma->desc_allocated++;
>> + return fsl_qdma->desc_allocated;
>> +
>> +err_mem:
>> + dma_pool_destroy(fsl_queue->desc_pool);
>> +err_desc_pool:
>> + dma_pool_destroy(fsl_queue->comp_pool);
>> + return -ENOMEM;
>> +}
>> +
>> +static int fsl_qdma_probe(struct platform_device *pdev)
>> +{
>> + int ret, i;
>> + int blk_num, blk_off;
>> + u32 len, chans, queues;
>> + struct resource *res;
>> + struct fsl_qdma_chan *fsl_chan;
>> + struct fsl_qdma_engine *fsl_qdma;
>> + struct device_node *np = pdev->dev.of_node;
>> +
>> + ret = of_property_read_u32(np, "dma-channels", &chans);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't get dma-channels.\n");
>> + return ret;
>> + }
>> +
>> + ret = of_property_read_u32(np, "block-offset", &blk_off);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't get block-offset.\n");
>> + return ret;
>> + }
>> +
>> + ret = of_property_read_u32(np, "block-number", &blk_num);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't get block-number.\n");
>> + return ret;
>> + }
>> +
>> + blk_num = min_t(int, blk_num, num_online_cpus());
>> +
>> + len = sizeof(*fsl_qdma);
>> + fsl_qdma = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>> + if (!fsl_qdma)
>> + return -ENOMEM;
>> +
>> + len = sizeof(*fsl_chan) * chans;
>> + fsl_qdma->chans = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>> + if (!fsl_qdma->chans)
>> + return -ENOMEM;
>> +
>> + len = sizeof(struct fsl_qdma_queue *) * blk_num;
>> + fsl_qdma->status = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>> + if (!fsl_qdma->status)
>> + return -ENOMEM;
>> +
>> + len = sizeof(int) * blk_num;
>> + fsl_qdma->queue_irq = devm_kzalloc(&pdev->dev, len,
>GFP_KERNEL);
>> + if (!fsl_qdma->queue_irq)
>> + return -ENOMEM;
>> +
>> + ret = of_property_read_u32(np, "fsl,dma-queues", &queues);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't get queues.\n");
>> + return ret;
>> + }
>> +
>> + fsl_qdma->desc_allocated = 0;
>> + fsl_qdma->n_chans = chans;
>> + fsl_qdma->n_queues = queues;
>> + fsl_qdma->block_number = blk_num;
>> + fsl_qdma->block_offset = blk_off;
>> +
>> + mutex_init(&fsl_qdma->fsl_qdma_mutex);
>> +
>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>> + fsl_qdma->status[i] =
>fsl_qdma_prep_status_queue(pdev);
>> + if (!fsl_qdma->status[i])
>> + return -ENOMEM;
>> + }
>> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
>> + fsl_qdma->ctrl_base = devm_ioremap_resource(&pdev->dev, res);
>> + if (IS_ERR(fsl_qdma->ctrl_base))
>> + return PTR_ERR(fsl_qdma->ctrl_base);
>> +
>> + res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
>> + fsl_qdma->status_base = devm_ioremap_resource(&pdev->dev,
>res);
>> + if (IS_ERR(fsl_qdma->status_base))
>> + return PTR_ERR(fsl_qdma->status_base);
>> +
>> + res = platform_get_resource(pdev, IORESOURCE_MEM, 2);
>> + fsl_qdma->block_base = devm_ioremap_resource(&pdev->dev,
>res);
>> + if (IS_ERR(fsl_qdma->block_base))
>> + return PTR_ERR(fsl_qdma->block_base);
>> + fsl_qdma->queue = fsl_qdma_alloc_queue_resources(pdev,
>fsl_qdma);
>> + if (!fsl_qdma->queue)
>> + return -ENOMEM;
>> +
>> + ret = fsl_qdma_irq_init(pdev, fsl_qdma);
>> + if (ret)
>> + return ret;
>> +
>> + fsl_qdma->irq_base = platform_get_irq_byname(pdev,
>"qdma-queue0");
>> + fsl_qdma->feature = of_property_read_bool(np, "big-endian");
>> + INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
>> +
>> + for (i = 0; i < fsl_qdma->n_chans; i++) {
>> + struct fsl_qdma_chan *fsl_chan = &fsl_qdma->chans[i];
>> +
>> + fsl_chan->qdma = fsl_qdma;
>> + fsl_chan->queue = fsl_qdma->queue + i %
>(fsl_qdma->n_queues *
>> +
>fsl_qdma->block_number);
>> + fsl_chan->vchan.desc_free = fsl_qdma_free_desc;
>> + vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
>> + }
>> +
>> + dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
>> +
>> + fsl_qdma->dma_dev.dev = &pdev->dev;
>> + fsl_qdma->dma_dev.device_free_chan_resources =
>> + fsl_qdma_free_chan_resources;
>> + fsl_qdma->dma_dev.device_alloc_chan_resources =
>> + fsl_qdma_alloc_chan_resources;
>> + fsl_qdma->dma_dev.device_tx_status = dma_cookie_status;
>> + fsl_qdma->dma_dev.device_prep_dma_memcpy =
>fsl_qdma_prep_memcpy;
>> + fsl_qdma->dma_dev.device_issue_pending =
>fsl_qdma_issue_pending;
>> + fsl_qdma->dma_dev.device_synchronize = fsl_qdma_synchronize;
>> + fsl_qdma->dma_dev.device_terminate_all =
>fsl_qdma_terminate_all;
>> +
>> + dma_set_mask(&pdev->dev, DMA_BIT_MASK(40));
>> +
>> + platform_set_drvdata(pdev, fsl_qdma);
>> +
>> + ret = dma_async_device_register(&fsl_qdma->dma_dev);
>> + if (ret) {
>> + dev_err(&pdev->dev,
>> + "Can't register NXP Layerscape qDMA
>engine.\n");
>> + return ret;
>> + }
>> +
>> + ret = fsl_qdma_reg_init(fsl_qdma);
>> + if (ret) {
>> + dev_err(&pdev->dev, "Can't Initialize the qDMA
>engine.\n");
>> + return ret;
>> + }
>> +
>> + return 0;
>> +}
>> +
>> +static void fsl_qdma_cleanup_vchan(struct dma_device *dmadev)
>> +{
>> + struct fsl_qdma_chan *chan, *_chan;
>> +
>> + list_for_each_entry_safe(chan, _chan,
>> + &dmadev->channels,
>vchan.chan.device_node) {
>> + list_del(&chan->vchan.chan.device_node);
>> + tasklet_kill(&chan->vchan.task);
>> + }
>> +}
>> +
>> +static int fsl_qdma_remove(struct platform_device *pdev)
>> +{
>> + int i;
>> + struct fsl_qdma_queue *status;
>> + struct device_node *np = pdev->dev.of_node;
>> + struct fsl_qdma_engine *fsl_qdma = platform_get_drvdata(pdev);
>> +
>> + fsl_qdma_irq_exit(pdev, fsl_qdma);
>> + fsl_qdma_cleanup_vchan(&fsl_qdma->dma_dev);
>> + of_dma_controller_free(np);
>> + dma_async_device_unregister(&fsl_qdma->dma_dev);
>> +
>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>> + status = fsl_qdma->status[i];
>> + dma_free_coherent(&pdev->dev, sizeof(struct
>fsl_qdma_format) *
>> + status->n_cq, status->cq,
>status->bus_addr);
>> + }
>> + return 0;
>> +}
>> +
>> +static const struct of_device_id fsl_qdma_dt_ids[] = {
>> + { .compatible = "fsl,ls1021a-qdma", },
>> + { /* sentinel */ }
>> +};
>> +MODULE_DEVICE_TABLE(of, fsl_qdma_dt_ids);
>> +
>> +static struct platform_driver fsl_qdma_driver = {
>> + .driver = {
>> + .name = "fsl-qdma",
>> + .of_match_table = fsl_qdma_dt_ids,
>> + },
>> + .probe = fsl_qdma_probe,
>> + .remove = fsl_qdma_remove,
>> +};
>> +
>> +module_platform_driver(fsl_qdma_driver);
>> +
>> +MODULE_ALIAS("platform:fsl-qdma");
>> +MODULE_DESCRIPTION("NXP Layerscape qDMA engine driver");
>> --
>> 1.7.1
>>
> On Oct 29, 2018, at 4:51 AM, Peng Ma <[email protected]> wrote:
>
>
>
>> -----Original Message-----
>> From: Li Yang <[email protected]>
>> Sent: 2018年10月27日 4:48
>> To: Peng Ma <[email protected]>
>> Cc: Vinod <[email protected]>; Rob Herring <[email protected]>; Mark
>> Rutland <[email protected]>; Shawn Guo <[email protected]>; Dan
>> Williams <[email protected]>; [email protected]; open
>> list:OPEN FIRMWARE AND FLATTENED DEVICE TREE BINDINGS
>> <[email protected]>; lkml <[email protected]>;
>> moderated list:ARM/FREESCALE IMX / MXC ARM ARCHITECTURE
>> <[email protected]>; linuxppc-dev
>> <[email protected]>; Wen He <[email protected]>; Jiaheng Fan
>> <[email protected]>
>> Subject: Re: [PATCH 3/7] dmaengine: fsl-qdma: Add qDMA controller driver for
>> Layerscape SoCs
>>
>>> On Fri, Oct 26, 2018 at 4:57 AM Peng Ma <[email protected]> wrote:
>>>
>>> NXP Queue DMA controller(qDMA) on Layerscape SoCs supports channel
>>> virtuallization by allowing DMA jobs to be enqueued into different
>>> command queues.
>>>
>>> Note that this module depends on NXP DPAA.
>>
>> It is not clear if you are saying that the driver can only work on
>> SoCs with a DPAA hardware block, or the driver is actually depending
>> on the DPAA drivers also. If it is the later case, you also should
>> express that in the Kconfig you added below.
>>
> [Peng Ma] Ok, I will express it in the Kconfig.
>>>
>>> Signed-off-by: Wen He <[email protected]>
>>> Signed-off-by: Jiaheng Fan <[email protected]>
>>> Signed-off-by: Peng Ma <[email protected]>
>>> ---
>>> change in v10:
>>> - no
>>>
>>> drivers/dma/Kconfig | 13 +
>>> drivers/dma/Makefile | 1 +
>>> drivers/dma/fsl-qdma.c | 1257
>> ++++++++++++++++++++++++++++++++++++++++++++++++
>>> 3 files changed, 1271 insertions(+), 0 deletions(-)
>>> create mode 100644 drivers/dma/fsl-qdma.c
>>>
>>> diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
>>> index dacf3f4..50e19d7 100644
>>> --- a/drivers/dma/Kconfig
>>> +++ b/drivers/dma/Kconfig
>>> @@ -218,6 +218,19 @@ config FSL_EDMA
>>> multiplexing capability for DMA request sources(slot).
>>> This module can be found on Freescale Vybrid and LS-1 SoCs.
>>>
>>> +config FSL_QDMA
>>> + tristate "NXP Layerscape qDMA engine support"
>>> + depends on ARM || ARM64
>>> + select DMA_ENGINE
>>> + select DMA_VIRTUAL_CHANNELS
>>> + select DMA_ENGINE_RAID
>>> + select ASYNC_TX_ENABLE_CHANNEL_SWITCH
>>> + help
>>> + Support the NXP Layerscape qDMA engine with command queue
>> and legacy mode.
>>> + Channel virtualization is supported through enqueuing of DMA
>> jobs to,
>>> + or dequeuing DMA jobs from, different work queues.
>>> + This module can be found on NXP Layerscape SoCs.
>>> +
>>> config FSL_RAID
>>> tristate "Freescale RAID engine Support"
>>> depends on FSL_SOC
>> && !ASYNC_TX_ENABLE_CHANNEL_SWITCH
>>> diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
>>> index c91702d..2d1b586 100644
>>> --- a/drivers/dma/Makefile
>>> +++ b/drivers/dma/Makefile
>>> @@ -32,6 +32,7 @@ obj-$(CONFIG_DW_DMAC_CORE) += dw/
>>> obj-$(CONFIG_EP93XX_DMA) += ep93xx_dma.o
>>> obj-$(CONFIG_FSL_DMA) += fsldma.o
>>> obj-$(CONFIG_FSL_EDMA) += fsl-edma.o
>>> +obj-$(CONFIG_FSL_QDMA) += fsl-qdma.o
>>> obj-$(CONFIG_FSL_RAID) += fsl_raid.o
>>> obj-$(CONFIG_HSU_DMA) += hsu/
>>> obj-$(CONFIG_IMG_MDC_DMA) += img-mdc-dma.o
>>> diff --git a/drivers/dma/fsl-qdma.c b/drivers/dma/fsl-qdma.c
>>> new file mode 100644
>>> index 0000000..404869e
>>> --- /dev/null
>>> +++ b/drivers/dma/fsl-qdma.c
>>> @@ -0,0 +1,1257 @@
>>> +// SPDX-License-Identifier: GPL-2.0
>>> +// Copyright 2018 NXP
>>
>> I'm not sure if this is really the case. The driver at least has been
>> sent out in 2015. We should keep these copyright claims, even the
>> legacy Freescale copyright claims.
>>
> [Peng Ma]
> I am not sure this patch sent out in 2015, but the git log earliest shows the patch cteated at Dec 20 2017. so if i changed the "Copyright 2018 NXP" to " Copyright 2017-2018 NXP "
You can find early versions of this patch in previous SDK releases or the upstream versions with a google search. These existing copyright claims should not be removed in the first place.
>
> Best regards
> Peng Ma
>>> +
>>> +/*
>>> + * Driver for NXP Layerscape Queue Direct Memory Access Controller
>>> + *
>>> + * Author:
>>> + * Wen He <[email protected]>
>>> + * Jiaheng Fan <[email protected]>
>>> + *
>>> + */
>>> +
>>> +#include <linux/module.h>
>>> +#include <linux/delay.h>
>>> +#include <linux/of_irq.h>
>>> +#include <linux/of_platform.h>
>>> +#include <linux/of_dma.h>
>>> +#include <linux/dma-mapping.h>
>>> +
>>> +#include "virt-dma.h"
>>> +#include "fsldma.h"
>>> +
>>> +/* Register related definition */
>>> +#define FSL_QDMA_DMR 0x0
>>> +#define FSL_QDMA_DSR 0x4
>>> +#define FSL_QDMA_DEIER 0xe00
>>> +#define FSL_QDMA_DEDR 0xe04
>>> +#define FSL_QDMA_DECFDW0R 0xe10
>>> +#define FSL_QDMA_DECFDW1R 0xe14
>>> +#define FSL_QDMA_DECFDW2R 0xe18
>>> +#define FSL_QDMA_DECFDW3R 0xe1c
>>> +#define FSL_QDMA_DECFQIDR 0xe30
>>> +#define FSL_QDMA_DECBR 0xe34
>>> +
>>> +#define FSL_QDMA_BCQMR(x) (0xc0 + 0x100 * (x))
>>> +#define FSL_QDMA_BCQSR(x) (0xc4 + 0x100 * (x))
>>> +#define FSL_QDMA_BCQEDPA_SADDR(x) (0xc8 + 0x100 * (x))
>>> +#define FSL_QDMA_BCQDPA_SADDR(x) (0xcc + 0x100 * (x))
>>> +#define FSL_QDMA_BCQEEPA_SADDR(x) (0xd0 + 0x100 * (x))
>>> +#define FSL_QDMA_BCQEPA_SADDR(x) (0xd4 + 0x100 * (x))
>>> +#define FSL_QDMA_BCQIER(x) (0xe0 + 0x100 * (x))
>>> +#define FSL_QDMA_BCQIDR(x) (0xe4 + 0x100 * (x))
>>> +
>>> +#define FSL_QDMA_SQDPAR 0x80c
>>> +#define FSL_QDMA_SQEPAR 0x814
>>> +#define FSL_QDMA_BSQMR 0x800
>>> +#define FSL_QDMA_BSQSR 0x804
>>> +#define FSL_QDMA_BSQICR 0x828
>>> +#define FSL_QDMA_CQMR 0xa00
>>> +#define FSL_QDMA_CQDSCR1 0xa08
>>> +#define FSL_QDMA_CQDSCR2 0xa0c
>>> +#define FSL_QDMA_CQIER 0xa10
>>> +#define FSL_QDMA_CQEDR 0xa14
>>> +#define FSL_QDMA_SQCCMR 0xa20
>>> +
>>> +/* Registers for bit and genmask */
>>> +#define FSL_QDMA_CQIDR_SQT BIT(15)
>>> +#define QDMA_CCDF_FOTMAT BIT(29)
>>> +#define QDMA_CCDF_SER BIT(30)
>>> +#define QDMA_SG_FIN BIT(30)
>>> +#define QDMA_SG_LEN_MASK GENMASK(29, 0)
>>> +#define QDMA_CCDF_MASK GENMASK(28, 20)
>>> +
>>> +#define FSL_QDMA_DEDR_CLEAR GENMASK(31, 0)
>>> +#define FSL_QDMA_BCQIDR_CLEAR GENMASK(31, 0)
>>> +#define FSL_QDMA_DEIER_CLEAR GENMASK(31, 0)
>>> +
>>> +#define FSL_QDMA_BCQIER_CQTIE BIT(15)
>>> +#define FSL_QDMA_BCQIER_CQPEIE BIT(23)
>>> +#define FSL_QDMA_BSQICR_ICEN BIT(31)
>>> +
>>> +#define FSL_QDMA_BSQICR_ICST(x) ((x) << 16)
>>> +#define FSL_QDMA_CQIER_MEIE BIT(31)
>>> +#define FSL_QDMA_CQIER_TEIE BIT(0)
>>> +#define FSL_QDMA_SQCCMR_ENTER_WM BIT(21)
>>> +
>>> +#define FSL_QDMA_BCQMR_EN BIT(31)
>>> +#define FSL_QDMA_BCQMR_EI BIT(30)
>>> +#define FSL_QDMA_BCQMR_CD_THLD(x) ((x) << 20)
>>> +#define FSL_QDMA_BCQMR_CQ_SIZE(x) ((x) << 16)
>>> +
>>> +#define FSL_QDMA_BCQSR_QF BIT(16)
>>> +#define FSL_QDMA_BCQSR_XOFF BIT(0)
>>> +
>>> +#define FSL_QDMA_BSQMR_EN BIT(31)
>>> +#define FSL_QDMA_BSQMR_DI BIT(30)
>>> +#define FSL_QDMA_BSQMR_CQ_SIZE(x) ((x) << 16)
>>> +
>>> +#define FSL_QDMA_BSQSR_QE BIT(17)
>>> +
>>> +#define FSL_QDMA_DMR_DQD BIT(30)
>>> +#define FSL_QDMA_DSR_DB BIT(31)
>>> +
>>> +/* Size related definition */
>>> +#define FSL_QDMA_QUEUE_MAX 8
>>> +#define FSL_QDMA_COMMAND_BUFFER_SIZE 64
>>> +#define FSL_QDMA_DESCRIPTOR_BUFFER_SIZE 32
>>> +#define FSL_QDMA_CIRCULAR_DESC_SIZE_MIN 64
>>> +#define FSL_QDMA_CIRCULAR_DESC_SIZE_MAX 16384
>>> +#define FSL_QDMA_QUEUE_NUM_MAX 8
>>> +
>>> +/* Field definition for CMD */
>>> +#define FSL_QDMA_CMD_RWTTYPE 0x4
>>> +#define FSL_QDMA_CMD_LWC 0x2
>>> +#define FSL_QDMA_CMD_RWTTYPE_OFFSET 28
>>> +#define FSL_QDMA_CMD_NS_OFFSET 27
>>> +#define FSL_QDMA_CMD_DQOS_OFFSET 24
>>> +#define FSL_QDMA_CMD_WTHROTL_OFFSET 20
>>> +#define FSL_QDMA_CMD_DSEN_OFFSET 19
>>> +#define FSL_QDMA_CMD_LWC_OFFSET 16
>>> +
>>> +/* Field definition for Descriptor offset */
>>> +#define QDMA_CCDF_STATUS 20
>>> +#define QDMA_CCDF_OFFSET 20
>>> +
>>> +/* Field definition for safe loop count*/
>>> +#define FSL_QDMA_HALT_COUNT 1500
>>> +#define FSL_QDMA_MAX_SIZE 16385
>>> +#define FSL_QDMA_COMP_TIMEOUT 1000
>>> +#define FSL_COMMAND_QUEUE_OVERFLLOW 10
>>> +
>>> +#define FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma_engine, x)
>> \
>>> + (((fsl_qdma_engine)->block_offset) * (x))
>>> +
>>> +/**
>>> + * struct fsl_qdma_format - This is the struct holding describing compound
>>> + * descriptor format with qDMA.
>>> + * @status: Command status and enqueue status
>> notification.
>>> + * @cfg: Frame offset and frame format.
>>> + * @addr_lo: Holding the compound descriptor of the
>> lower
>>> + * 32-bits address in memory 40-bit address.
>>> + * @addr_hi: Same as above member, but point high
>> 8-bits in
>>> + * memory 40-bit address.
>>> + * @__reserved1: Reserved field.
>>> + * @cfg8b_w1: Compound descriptor command queue
>> origin produced
>>> + * by qDMA and dynamic debug field.
>>> + * @data Pointer to the memory 40-bit address,
>> describes DMA
>>> + * source information and DMA destination
>> information.
>>> + */
>>> +struct fsl_qdma_format {
>>> + __le32 status;
>>> + __le32 cfg;
>>> + union {
>>> + struct {
>>> + __le32 addr_lo;
>>> + u8 addr_hi;
>>> + u8 __reserved1[2];
>>> + u8 cfg8b_w1;
>>> + } __packed;
>>> + __le64 data;
>>> + };
>>> +} __packed;
>>> +
>>> +/* qDMA status notification pre information */
>>> +struct fsl_pre_status {
>>> + u64 addr;
>>> + u8 queue;
>>> +};
>>> +
>>> +static DEFINE_PER_CPU(struct fsl_pre_status, pre);
>>> +
>>> +struct fsl_qdma_chan {
>>> + struct virt_dma_chan vchan;
>>> + struct virt_dma_desc vdesc;
>>> + enum dma_status status;
>>> + struct fsl_qdma_engine *qdma;
>>> + struct fsl_qdma_queue *queue;
>>> +};
>>> +
>>> +struct fsl_qdma_queue {
>>> + struct fsl_qdma_format *virt_head;
>>> + struct fsl_qdma_format *virt_tail;
>>> + struct list_head comp_used;
>>> + struct list_head comp_free;
>>> + struct dma_pool *comp_pool;
>>> + struct dma_pool *desc_pool;
>>> + spinlock_t queue_lock;
>>> + dma_addr_t bus_addr;
>>> + u32 n_cq;
>>> + u32 id;
>>> + struct fsl_qdma_format *cq;
>>> + void __iomem *block_base;
>>> +};
>>> +
>>> +struct fsl_qdma_comp {
>>> + dma_addr_t bus_addr;
>>> + dma_addr_t desc_bus_addr;
>>> + struct fsl_qdma_format *virt_addr;
>>> + struct fsl_qdma_format *desc_virt_addr;
>>> + struct fsl_qdma_chan *qchan;
>>> + struct virt_dma_desc vdesc;
>>> + struct list_head list;
>>> +};
>>> +
>>> +struct fsl_qdma_engine {
>>> + struct dma_device dma_dev;
>>> + void __iomem *ctrl_base;
>>> + void __iomem *status_base;
>>> + void __iomem *block_base;
>>> + u32 n_chans;
>>> + u32 n_queues;
>>> + struct mutex fsl_qdma_mutex;
>>> + int error_irq;
>>> + int *queue_irq;
>>> + u32 feature;
>>> + struct fsl_qdma_queue *queue;
>>> + struct fsl_qdma_queue **status;
>>> + struct fsl_qdma_chan *chans;
>>> + int block_number;
>>> + int block_offset;
>>> + int irq_base;
>>> + int desc_allocated;
>>> +
>>> +};
>>> +
>>> +static inline u64
>>> +qdma_ccdf_addr_get64(const struct fsl_qdma_format *ccdf)
>>> +{
>>> + return le64_to_cpu(ccdf->data) & (U64_MAX >> 24);
>>> +}
>>> +
>>> +static inline void
>>> +qdma_desc_addr_set64(struct fsl_qdma_format *ccdf, u64 addr)
>>> +{
>>> + ccdf->addr_hi = upper_32_bits(addr);
>>> + ccdf->addr_lo = cpu_to_le32(lower_32_bits(addr));
>>> +}
>>> +
>>> +static inline u8
>>> +qdma_ccdf_get_queue(const struct fsl_qdma_format *ccdf)
>>> +{
>>> + return ccdf->cfg8b_w1 & U8_MAX;
>>> +}
>>> +
>>> +static inline int
>>> +qdma_ccdf_get_offset(const struct fsl_qdma_format *ccdf)
>>> +{
>>> + return (le32_to_cpu(ccdf->cfg) & QDMA_CCDF_MASK) >>
>> QDMA_CCDF_OFFSET;
>>> +}
>>> +
>>> +static inline void
>>> +qdma_ccdf_set_format(struct fsl_qdma_format *ccdf, int offset)
>>> +{
>>> + ccdf->cfg = cpu_to_le32(QDMA_CCDF_FOTMAT | offset);
>>> +}
>>> +
>>> +static inline int
>>> +qdma_ccdf_get_status(const struct fsl_qdma_format *ccdf)
>>> +{
>>> + return (le32_to_cpu(ccdf->status) & QDMA_CCDF_MASK) >>
>> QDMA_CCDF_STATUS;
>>> +}
>>> +
>>> +static inline void
>>> +qdma_ccdf_set_ser(struct fsl_qdma_format *ccdf, int status)
>>> +{
>>> + ccdf->status = cpu_to_le32(QDMA_CCDF_SER | status);
>>> +}
>>> +
>>> +static inline void qdma_csgf_set_len(struct fsl_qdma_format *csgf, int len)
>>> +{
>>> + csgf->cfg = cpu_to_le32(len & QDMA_SG_LEN_MASK);
>>> +}
>>> +
>>> +static inline void qdma_csgf_set_f(struct fsl_qdma_format *csgf, int len)
>>> +{
>>> + csgf->cfg = cpu_to_le32(QDMA_SG_FIN | (len &
>> QDMA_SG_LEN_MASK));
>>> +}
>>> +
>>> +static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem
>> *addr)
>>> +{
>>> + return FSL_DMA_IN(qdma, addr, 32);
>>> +}
>>> +
>>> +static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val,
>>> + void __iomem *addr)
>>> +{
>>> + FSL_DMA_OUT(qdma, addr, val, 32);
>>> +}
>>> +
>>> +static struct fsl_qdma_chan *to_fsl_qdma_chan(struct dma_chan *chan)
>>> +{
>>> + return container_of(chan, struct fsl_qdma_chan, vchan.chan);
>>> +}
>>> +
>>> +static struct fsl_qdma_comp *to_fsl_qdma_comp(struct virt_dma_desc *vd)
>>> +{
>>> + return container_of(vd, struct fsl_qdma_comp, vdesc);
>>> +}
>>> +
>>> +static void fsl_qdma_free_chan_resources(struct dma_chan *chan)
>>> +{
>>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>>> + struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
>>> + struct fsl_qdma_comp *comp_temp, *_comp_temp;
>>> + unsigned long flags;
>>> + LIST_HEAD(head);
>>> +
>>> + spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
>>> + vchan_get_all_descriptors(&fsl_chan->vchan, &head);
>>> + spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
>>> +
>>> + vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
>>> +
>>> + if (!fsl_queue->comp_pool && !fsl_queue->comp_pool)
>>> + return;
>>> +
>>> + list_for_each_entry_safe(comp_temp, _comp_temp,
>>> + &fsl_queue->comp_used, list) {
>>> + dma_pool_free(fsl_queue->comp_pool,
>>> + comp_temp->virt_addr,
>>> + comp_temp->bus_addr);
>>> + dma_pool_free(fsl_queue->desc_pool,
>>> + comp_temp->desc_virt_addr,
>>> + comp_temp->desc_bus_addr);
>>> + list_del(&comp_temp->list);
>>> + kfree(comp_temp);
>>> + }
>>> +
>>> + list_for_each_entry_safe(comp_temp, _comp_temp,
>>> + &fsl_queue->comp_free, list) {
>>> + dma_pool_free(fsl_queue->comp_pool,
>>> + comp_temp->virt_addr,
>>> + comp_temp->bus_addr);
>>> + dma_pool_free(fsl_queue->desc_pool,
>>> + comp_temp->desc_virt_addr,
>>> + comp_temp->desc_bus_addr);
>>> + list_del(&comp_temp->list);
>>> + kfree(comp_temp);
>>> + }
>>> +
>>> + dma_pool_destroy(fsl_queue->comp_pool);
>>> + dma_pool_destroy(fsl_queue->desc_pool);
>>> +
>>> + fsl_qdma->desc_allocated--;
>>> + fsl_queue->comp_pool = NULL;
>>> + fsl_queue->desc_pool = NULL;
>>> +}
>>> +
>>> +static void fsl_qdma_comp_fill_memcpy(struct fsl_qdma_comp *fsl_comp,
>>> + dma_addr_t dst, dma_addr_t
>> src, u32 len)
>>> +{
>>> + struct fsl_qdma_format *sdf, *ddf;
>>> + struct fsl_qdma_format *ccdf, *csgf_desc, *csgf_src, *csgf_dest;
>>> +
>>> + ccdf = fsl_comp->virt_addr;
>>> + csgf_desc = fsl_comp->virt_addr + 1;
>>> + csgf_src = fsl_comp->virt_addr + 2;
>>> + csgf_dest = fsl_comp->virt_addr + 3;
>>> + sdf = fsl_comp->desc_virt_addr;
>>> + ddf = fsl_comp->desc_virt_addr + 1;
>>> +
>>> + memset(fsl_comp->virt_addr, 0,
>> FSL_QDMA_COMMAND_BUFFER_SIZE);
>>> + memset(fsl_comp->desc_virt_addr, 0,
>> FSL_QDMA_DESCRIPTOR_BUFFER_SIZE);
>>> + /* Head Command Descriptor(Frame Descriptor) */
>>> + qdma_desc_addr_set64(ccdf, fsl_comp->bus_addr + 16);
>>> + qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf));
>>> + qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf));
>>> + /* Status notification is enqueued to status queue. */
>>> + /* Compound Command Descriptor(Frame List Table) */
>>> + qdma_desc_addr_set64(csgf_desc, fsl_comp->desc_bus_addr);
>>> + /* It must be 32 as Compound S/G Descriptor */
>>> + qdma_csgf_set_len(csgf_desc, 32);
>>> + qdma_desc_addr_set64(csgf_src, src);
>>> + qdma_csgf_set_len(csgf_src, len);
>>> + qdma_desc_addr_set64(csgf_dest, dst);
>>> + qdma_csgf_set_len(csgf_dest, len);
>>> + /* This entry is the last entry. */
>>> + qdma_csgf_set_f(csgf_dest, len);
>>> + /* Descriptor Buffer */
>>> + sdf->data =
>>> + cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
>>> + FSL_QDMA_CMD_RWTTYPE_OFFSET);
>>> + ddf->data =
>>> + cpu_to_le64(FSL_QDMA_CMD_RWTTYPE <<
>>> + FSL_QDMA_CMD_RWTTYPE_OFFSET);
>>> + ddf->data |=
>>> + cpu_to_le64(FSL_QDMA_CMD_LWC <<
>> FSL_QDMA_CMD_LWC_OFFSET);
>>> +}
>>> +
>>> +/*
>>> + * Pre-request full command descriptor for enqueue.
>>> + */
>>> +static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue
>> *queue)
>>> +{
>>> + int i;
>>> + struct fsl_qdma_comp *comp_temp, *_comp_temp;
>>> +
>>> + for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW;
>> i++) {
>>> + comp_temp = kzalloc(sizeof(*comp_temp),
>> GFP_KERNEL);
>>> + if (!comp_temp)
>>> + goto err_alloc;
>>> + comp_temp->virt_addr =
>>> + dma_pool_alloc(queue->comp_pool,
>> GFP_KERNEL,
>>> + &comp_temp->bus_addr);
>>> + if (!comp_temp->virt_addr)
>>> + goto err_dma_alloc;
>>> +
>>> + comp_temp->desc_virt_addr =
>>> + dma_pool_alloc(queue->desc_pool,
>> GFP_KERNEL,
>>> +
>> &comp_temp->desc_bus_addr);
>>> + if (!comp_temp->desc_virt_addr)
>>> + goto err_desc_dma_alloc;
>>> +
>>> + list_add_tail(&comp_temp->list, &queue->comp_free);
>>> + }
>>> +
>>> + return 0;
>>> +
>>> +err_desc_dma_alloc:
>>> + dma_pool_free(queue->comp_pool, comp_temp->virt_addr,
>>> + comp_temp->bus_addr);
>>> +
>>> +err_dma_alloc:
>>> + kfree(comp_temp);
>>> +
>>> +err_alloc:
>>> + list_for_each_entry_safe(comp_temp, _comp_temp,
>>> + &queue->comp_free, list) {
>>> + if (comp_temp->virt_addr)
>>> + dma_pool_free(queue->comp_pool,
>>> + comp_temp->virt_addr,
>>> + comp_temp->bus_addr);
>>> + if (comp_temp->desc_virt_addr)
>>> + dma_pool_free(queue->desc_pool,
>>> + comp_temp->desc_virt_addr,
>>> +
>> comp_temp->desc_bus_addr);
>>> +
>>> + list_del(&comp_temp->list);
>>> + kfree(comp_temp);
>>> + }
>>> +
>>> + return -ENOMEM;
>>> +}
>>> +
>>> +/*
>>> + * Request a command descriptor for enqueue.
>>> + */
>>> +static struct fsl_qdma_comp
>>> +*fsl_qdma_request_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
>>> +{
>>> + unsigned long flags;
>>> + struct fsl_qdma_comp *comp_temp;
>>> + int timeout = FSL_QDMA_COMP_TIMEOUT;
>>> + struct fsl_qdma_queue *queue = fsl_chan->queue;
>>> +
>>> + while (timeout--) {
>>> + spin_lock_irqsave(&queue->queue_lock, flags);
>>> + if (!list_empty(&queue->comp_free)) {
>>> + comp_temp =
>> list_first_entry(&queue->comp_free,
>>> + struct
>> fsl_qdma_comp,
>>> + list);
>>> + list_del(&comp_temp->list);
>>> +
>>> + spin_unlock_irqrestore(&queue->queue_lock,
>> flags);
>>> + comp_temp->qchan = fsl_chan;
>>> + return comp_temp;
>>> + }
>>> + spin_unlock_irqrestore(&queue->queue_lock, flags);
>>> + udelay(1);
>>> + }
>>> +
>>> + return NULL;
>>> +}
>>> +
>>> +static struct fsl_qdma_queue
>>> +*fsl_qdma_alloc_queue_resources(struct platform_device *pdev,
>>> + struct fsl_qdma_engine *fsl_qdma)
>>> +{
>>> + int ret, len, i, j;
>>> + int queue_num, block_number;
>>> + unsigned int queue_size[FSL_QDMA_QUEUE_MAX];
>>> + struct fsl_qdma_queue *queue_head, *queue_temp;
>>> +
>>> + queue_num = fsl_qdma->n_queues;
>>> + block_number = fsl_qdma->block_number;
>>> +
>>> + if (queue_num > FSL_QDMA_QUEUE_MAX)
>>> + queue_num = FSL_QDMA_QUEUE_MAX;
>>> + len = sizeof(*queue_head) * queue_num * block_number;
>>> + queue_head = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>>> + if (!queue_head)
>>> + return NULL;
>>> +
>>> + ret = device_property_read_u32_array(&pdev->dev, "queue-sizes",
>>> + queue_size,
>> queue_num);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't get queue-sizes.\n");
>>> + return NULL;
>>> + }
>>> + for (j = 0; j < block_number; j++) {
>>> + for (i = 0; i < queue_num; i++) {
>>> + if (queue_size[i] >
>> FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
>>> + queue_size[i] <
>> FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
>>> + dev_err(&pdev->dev,
>>> + "Get wrong
>> queue-sizes.\n");
>>> + return NULL;
>>> + }
>>> + queue_temp = queue_head + i + (j *
>> queue_num);
>>> +
>>> + queue_temp->cq =
>>> + dma_alloc_coherent(&pdev->dev,
>>> + sizeof(struct
>> fsl_qdma_format) *
>>> + queue_size[i],
>>> +
>> &queue_temp->bus_addr,
>>> + GFP_KERNEL);
>>> + if (!queue_temp->cq)
>>> + return NULL;
>>> + queue_temp->block_base =
>> fsl_qdma->block_base +
>>> +
>> FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>>> + queue_temp->n_cq = queue_size[i];
>>> + queue_temp->id = i;
>>> + queue_temp->virt_head = queue_temp->cq;
>>> + queue_temp->virt_tail = queue_temp->cq;
>>> + /*
>>> + * List for queue command buffer
>>> + */
>>> + INIT_LIST_HEAD(&queue_temp->comp_used);
>>> + spin_lock_init(&queue_temp->queue_lock);
>>> + }
>>> + }
>>> + return queue_head;
>>> +}
>>> +
>>> +static struct fsl_qdma_queue
>>> +*fsl_qdma_prep_status_queue(struct platform_device *pdev)
>>> +{
>>> + int ret;
>>> + unsigned int status_size;
>>> + struct fsl_qdma_queue *status_head;
>>> + struct device_node *np = pdev->dev.of_node;
>>> +
>>> + ret = of_property_read_u32(np, "status-sizes", &status_size);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't get status-sizes.\n");
>>> + return NULL;
>>> + }
>>> + if (status_size > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||
>>> + status_size < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
>>> + dev_err(&pdev->dev, "Get wrong status_size.\n");
>>> + return NULL;
>>> + }
>>> + status_head = devm_kzalloc(&pdev->dev,
>>> + sizeof(*status_head),
>> GFP_KERNEL);
>>> + if (!status_head)
>>> + return NULL;
>>> +
>>> + /*
>>> + * Buffer for queue command
>>> + */
>>> + status_head->cq = dma_alloc_coherent(&pdev->dev,
>>> + sizeof(struct
>> fsl_qdma_format) *
>>> + status_size,
>>> +
>> &status_head->bus_addr,
>>> + GFP_KERNEL);
>>> + if (!status_head->cq) {
>>> + devm_kfree(&pdev->dev, status_head);
>>> + return NULL;
>>> + }
>>> + status_head->n_cq = status_size;
>>> + status_head->virt_head = status_head->cq;
>>> + status_head->virt_tail = status_head->cq;
>>> + status_head->comp_pool = NULL;
>>> +
>>> + return status_head;
>>> +}
>>> +
>>> +static int fsl_qdma_halt(struct fsl_qdma_engine *fsl_qdma)
>>> +{
>>> + u32 reg;
>>> + int i, j, count = FSL_QDMA_HALT_COUNT;
>>> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
>>> +
>>> + /* Disable the command queue and wait for idle state. */
>>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
>>> + reg |= FSL_QDMA_DMR_DQD;
>>> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
>>> + for (j = 0; j < fsl_qdma->block_number; j++) {
>>> + block = fsl_qdma->block_base +
>>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>>> + for (i = 0; i < FSL_QDMA_QUEUE_NUM_MAX; i++)
>>> + qdma_writel(fsl_qdma, 0, block +
>> FSL_QDMA_BCQMR(i));
>>> + }
>>> + while (1) {
>>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DSR);
>>> + if (!(reg & FSL_QDMA_DSR_DB))
>>> + break;
>>> + if (count-- < 0)
>>> + return -EBUSY;
>>> + udelay(100);
>>> + }
>>> +
>>> + for (j = 0; j < fsl_qdma->block_number; j++) {
>>> + block = fsl_qdma->block_base +
>>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>>> +
>>> + /* Disable status queue. */
>>> + qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BSQMR);
>>> +
>>> + /*
>>> + * clear the command queue interrupt detect register for
>>> + * all queues.
>>> + */
>>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
>>> + block + FSL_QDMA_BCQIDR(0));
>>> + }
>>> +
>>> + return 0;
>>> +}
>>> +
>>> +static int
>>> +fsl_qdma_queue_transfer_complete(struct fsl_qdma_engine *fsl_qdma,
>>> + void *block,
>>> + int id)
>>> +{
>>> + bool duplicate;
>>> + u32 reg, i, count;
>>> + struct fsl_qdma_queue *temp_queue;
>>> + struct fsl_qdma_format *status_addr;
>>> + struct fsl_qdma_comp *fsl_comp = NULL;
>>> + struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
>>> + struct fsl_qdma_queue *fsl_status = fsl_qdma->status[id];
>>> +
>>> + count = FSL_QDMA_MAX_SIZE;
>>> +
>>> + while (count--) {
>>> + duplicate = 0;
>>> + reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQSR);
>>> + if (reg & FSL_QDMA_BSQSR_QE)
>>> + return 0;
>>> +
>>> + status_addr = fsl_status->virt_head;
>>> +
>>> + if (qdma_ccdf_get_queue(status_addr) ==
>>> + __this_cpu_read(pre.queue) &&
>>> + qdma_ccdf_addr_get64(status_addr) ==
>>> + __this_cpu_read(pre.addr))
>>> + duplicate = 1;
>>> + i = qdma_ccdf_get_queue(status_addr) +
>>> + id * fsl_qdma->n_queues;
>>> + __this_cpu_write(pre.addr,
>> qdma_ccdf_addr_get64(status_addr));
>>> + __this_cpu_write(pre.queue,
>> qdma_ccdf_get_queue(status_addr));
>>> + temp_queue = fsl_queue + i;
>>> +
>>> + spin_lock(&temp_queue->queue_lock);
>>> + if (list_empty(&temp_queue->comp_used)) {
>>> + if (!duplicate) {
>>> +
>> spin_unlock(&temp_queue->queue_lock);
>>> + return -EAGAIN;
>>> + }
>>> + } else {
>>> + fsl_comp =
>> list_first_entry(&temp_queue->comp_used,
>>> + struct
>> fsl_qdma_comp, list);
>>> + if (fsl_comp->bus_addr + 16 !=
>>> + __this_cpu_read(pre.addr)) {
>>> + if (!duplicate) {
>>> +
>> spin_unlock(&temp_queue->queue_lock);
>>> + return -EAGAIN;
>>> + }
>>> + }
>>> + }
>>> +
>>> + if (duplicate) {
>>> + reg = qdma_readl(fsl_qdma, block +
>> FSL_QDMA_BSQMR);
>>> + reg |= FSL_QDMA_BSQMR_DI;
>>> + qdma_desc_addr_set64(status_addr, 0x0);
>>> + fsl_status->virt_head++;
>>> + if (fsl_status->virt_head == fsl_status->cq
>>> + +
>> fsl_status->n_cq)
>>> + fsl_status->virt_head =
>> fsl_status->cq;
>>> + qdma_writel(fsl_qdma, reg, block +
>> FSL_QDMA_BSQMR);
>>> + spin_unlock(&temp_queue->queue_lock);
>>> + continue;
>>> + }
>>> + list_del(&fsl_comp->list);
>>> +
>>> + reg = qdma_readl(fsl_qdma, block +
>> FSL_QDMA_BSQMR);
>>> + reg |= FSL_QDMA_BSQMR_DI;
>>> + qdma_desc_addr_set64(status_addr, 0x0);
>>> + fsl_status->virt_head++;
>>> + if (fsl_status->virt_head == fsl_status->cq +
>> fsl_status->n_cq)
>>> + fsl_status->virt_head = fsl_status->cq;
>>> + qdma_writel(fsl_qdma, reg, block +
>> FSL_QDMA_BSQMR);
>>> + spin_unlock(&temp_queue->queue_lock);
>>> +
>>> + spin_lock(&fsl_comp->qchan->vchan.lock);
>>> + vchan_cookie_complete(&fsl_comp->vdesc);
>>> + fsl_comp->qchan->status = DMA_COMPLETE;
>>> + spin_unlock(&fsl_comp->qchan->vchan.lock);
>>> + }
>>> +
>>> + return 0;
>>> +}
>>> +
>>> +static irqreturn_t fsl_qdma_error_handler(int irq, void *dev_id)
>>> +{
>>> + unsigned int intr;
>>> + struct fsl_qdma_engine *fsl_qdma = dev_id;
>>> + void __iomem *status = fsl_qdma->status_base;
>>> +
>>> + intr = qdma_readl(fsl_qdma, status + FSL_QDMA_DEDR);
>>> +
>>> + if (intr) {
>>> + dev_err(fsl_qdma->dma_dev.dev, "DMA transaction
>> error!\n");
>>> + return IRQ_NONE;
>>> + }
>>> +
>>> + qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status +
>> FSL_QDMA_DEDR);
>>> + return IRQ_HANDLED;
>>> +}
>>> +
>>> +static irqreturn_t fsl_qdma_queue_handler(int irq, void *dev_id)
>>> +{
>>> + int id;
>>> + unsigned int intr, reg;
>>> + struct fsl_qdma_engine *fsl_qdma = dev_id;
>>> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
>>> +
>>> + id = irq - fsl_qdma->irq_base;
>>> + if (id < 0 && id > fsl_qdma->block_number) {
>>> + dev_err(fsl_qdma->dma_dev.dev,
>>> + "irq %d is wrong irq_base is %d\n",
>>> + irq, fsl_qdma->irq_base);
>>> + }
>>> +
>>> + block = fsl_qdma->block_base +
>>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, id);
>>> +
>>> + intr = qdma_readl(fsl_qdma, block + FSL_QDMA_BCQIDR(0));
>>> +
>>> + if ((intr & FSL_QDMA_CQIDR_SQT) != 0)
>>> + intr = fsl_qdma_queue_transfer_complete(fsl_qdma,
>> block, id);
>>> +
>>> + if (intr != 0) {
>>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
>>> + reg |= FSL_QDMA_DMR_DQD;
>>> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
>>> + qdma_writel(fsl_qdma, 0, block +
>> FSL_QDMA_BCQIER(0));
>>> + dev_err(fsl_qdma->dma_dev.dev, "QDMA: status
>> err!\n");
>>> + }
>>> +
>>> + /* Clear all detected events and interrupts. */
>>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
>>> + block + FSL_QDMA_BCQIDR(0));
>>> +
>>> + return IRQ_HANDLED;
>>> +}
>>> +
>>> +static int
>>> +fsl_qdma_irq_init(struct platform_device *pdev,
>>> + struct fsl_qdma_engine *fsl_qdma)
>>> +{
>>> + int i;
>>> + int cpu;
>>> + int ret;
>>> + char irq_name[20];
>>> +
>>> + fsl_qdma->error_irq =
>>> + platform_get_irq_byname(pdev, "qdma-error");
>>> + if (fsl_qdma->error_irq < 0) {
>>> + dev_err(&pdev->dev, "Can't get qdma controller irq.\n");
>>> + return fsl_qdma->error_irq;
>>> + }
>>> +
>>> + ret = devm_request_irq(&pdev->dev, fsl_qdma->error_irq,
>>> + fsl_qdma_error_handler, 0,
>>> + "qDMA error", fsl_qdma);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't register qDMA controller
>> IRQ.\n");
>>> + return ret;
>>> + }
>>> +
>>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>>> + sprintf(irq_name, "qdma-queue%d", i);
>>> + fsl_qdma->queue_irq[i] =
>>> + platform_get_irq_byname(pdev,
>> irq_name);
>>> +
>>> + if (fsl_qdma->queue_irq[i] < 0) {
>>> + dev_err(&pdev->dev,
>>> + "Can't get qdma queue %d irq.\n", i);
>>> + return fsl_qdma->queue_irq[i];
>>> + }
>>> +
>>> + ret = devm_request_irq(&pdev->dev,
>>> + fsl_qdma->queue_irq[i],
>>> + fsl_qdma_queue_handler,
>>> + 0,
>>> + "qDMA queue",
>>> + fsl_qdma);
>>> + if (ret) {
>>> + dev_err(&pdev->dev,
>>> + "Can't register qDMA queue
>> IRQ.\n");
>>> + return ret;
>>> + }
>>> +
>>> + cpu = i % num_online_cpus();
>>> + ret = irq_set_affinity_hint(fsl_qdma->queue_irq[i],
>>> + get_cpu_mask(cpu));
>>> + if (ret) {
>>> + dev_err(&pdev->dev,
>>> + "Can't set cpu %d affinity to
>> IRQ %d.\n",
>>> + cpu,
>>> + fsl_qdma->queue_irq[i]);
>>> + return ret;
>>> + }
>>> + }
>>> +
>>> + return 0;
>>> +}
>>> +
>>> +static void fsl_qdma_irq_exit(struct platform_device *pdev,
>>> + struct fsl_qdma_engine *fsl_qdma)
>>> +{
>>> + int i;
>>> +
>>> + devm_free_irq(&pdev->dev, fsl_qdma->error_irq, fsl_qdma);
>>> + for (i = 0; i < fsl_qdma->block_number; i++)
>>> + devm_free_irq(&pdev->dev, fsl_qdma->queue_irq[i],
>> fsl_qdma);
>>> +}
>>> +
>>> +static int fsl_qdma_reg_init(struct fsl_qdma_engine *fsl_qdma)
>>> +{
>>> + u32 reg;
>>> + int i, j, ret;
>>> + struct fsl_qdma_queue *temp;
>>> + void __iomem *status = fsl_qdma->status_base;
>>> + void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
>>> + struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
>>> +
>>> + /* Try to halt the qDMA engine first. */
>>> + ret = fsl_qdma_halt(fsl_qdma);
>>> + if (ret) {
>>> + dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
>>> + return ret;
>>> + }
>>> +
>>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>>> + /*
>>> + * Clear the command queue interrupt detect register
>> for
>>> + * all queues.
>>> + */
>>> +
>>> + block = fsl_qdma->block_base +
>>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, i);
>>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIDR_CLEAR,
>>> + block + FSL_QDMA_BCQIDR(0));
>>> + }
>>> +
>>> + for (j = 0; j < fsl_qdma->block_number; j++) {
>>> + block = fsl_qdma->block_base +
>>> + FSL_QDMA_BLOCK_BASE_OFFSET(fsl_qdma, j);
>>> + for (i = 0; i < fsl_qdma->n_queues; i++) {
>>> + temp = fsl_queue + i + (j *
>> fsl_qdma->n_queues);
>>> + /*
>>> + * Initialize Command Queue registers to
>>> + * point to the first
>>> + * command descriptor in memory.
>>> + * Dequeue Pointer Address Registers
>>> + * Enqueue Pointer Address Registers
>>> + */
>>> +
>>> + qdma_writel(fsl_qdma, temp->bus_addr,
>>> + block +
>> FSL_QDMA_BCQDPA_SADDR(i));
>>> + qdma_writel(fsl_qdma, temp->bus_addr,
>>> + block +
>> FSL_QDMA_BCQEPA_SADDR(i));
>>> +
>>> + /* Initialize the queue mode. */
>>> + reg = FSL_QDMA_BCQMR_EN;
>>> + reg |=
>> FSL_QDMA_BCQMR_CD_THLD(ilog2(temp->n_cq) - 4);
>>> + reg |=
>> FSL_QDMA_BCQMR_CQ_SIZE(ilog2(temp->n_cq) - 6);
>>> + qdma_writel(fsl_qdma, reg, block +
>> FSL_QDMA_BCQMR(i));
>>> + }
>>> +
>>> + /*
>>> + * Workaround for erratum: ERR010812.
>>> + * We must enable XOFF to avoid the enqueue rejection
>> occurs.
>>> + * Setting SQCCMR ENTER_WM to 0x20.
>>> + */
>>> +
>>> + qdma_writel(fsl_qdma,
>> FSL_QDMA_SQCCMR_ENTER_WM,
>>> + block + FSL_QDMA_SQCCMR);
>>> +
>>> + /*
>>> + * Initialize status queue registers to point to the first
>>> + * command descriptor in memory.
>>> + * Dequeue Pointer Address Registers
>>> + * Enqueue Pointer Address Registers
>>> + */
>>> +
>>> + qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
>>> + block + FSL_QDMA_SQEPAR);
>>> + qdma_writel(fsl_qdma, fsl_qdma->status[j]->bus_addr,
>>> + block + FSL_QDMA_SQDPAR);
>>> + /* Initialize status queue interrupt. */
>>> + qdma_writel(fsl_qdma, FSL_QDMA_BCQIER_CQTIE,
>>> + block + FSL_QDMA_BCQIER(0));
>>> + qdma_writel(fsl_qdma, FSL_QDMA_BSQICR_ICEN |
>>> + FSL_QDMA_BSQICR_ICST(5) |
>> 0x8000,
>>> + block + FSL_QDMA_BSQICR);
>>> + qdma_writel(fsl_qdma, FSL_QDMA_CQIER_MEIE |
>>> + FSL_QDMA_CQIER_TEIE,
>>> + block + FSL_QDMA_CQIER);
>>> +
>>> + /* Initialize the status queue mode. */
>>> + reg = FSL_QDMA_BSQMR_EN;
>>> + reg |= FSL_QDMA_BSQMR_CQ_SIZE(ilog2
>>> + (fsl_qdma->status[j]->n_cq) - 6);
>>> +
>>> + qdma_writel(fsl_qdma, reg, block +
>> FSL_QDMA_BSQMR);
>>> + reg = qdma_readl(fsl_qdma, block +
>> FSL_QDMA_BSQMR);
>>> + }
>>> +
>>> + /* Initialize controller interrupt register. */
>>> + qdma_writel(fsl_qdma, FSL_QDMA_DEDR_CLEAR, status +
>> FSL_QDMA_DEDR);
>>> + qdma_writel(fsl_qdma, FSL_QDMA_DEIER_CLEAR, status +
>> FSL_QDMA_DEIER);
>>> +
>>> + reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
>>> + reg &= ~FSL_QDMA_DMR_DQD;
>>> + qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
>>> +
>>> + return 0;
>>> +}
>>> +
>>> +static struct dma_async_tx_descriptor *
>>> +fsl_qdma_prep_memcpy(struct dma_chan *chan, dma_addr_t dst,
>>> + dma_addr_t src, size_t len, unsigned long flags)
>>> +{
>>> + struct fsl_qdma_comp *fsl_comp;
>>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>>> +
>>> + fsl_comp = fsl_qdma_request_enqueue_desc(fsl_chan);
>>> +
>>> + if (!fsl_comp)
>>> + return NULL;
>>> +
>>> + fsl_qdma_comp_fill_memcpy(fsl_comp, dst, src, len);
>>> +
>>> + return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
>>> +}
>>> +
>>> +static void fsl_qdma_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
>>> +{
>>> + u32 reg;
>>> + struct virt_dma_desc *vdesc;
>>> + struct fsl_qdma_comp *fsl_comp;
>>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>>> + void __iomem *block = fsl_queue->block_base;
>>> +
>>> + reg = qdma_readl(fsl_chan->qdma, block +
>> FSL_QDMA_BCQSR(fsl_queue->id));
>>> + if (reg & (FSL_QDMA_BCQSR_QF | FSL_QDMA_BCQSR_XOFF))
>>> + return;
>>> + vdesc = vchan_next_desc(&fsl_chan->vchan);
>>> + if (!vdesc)
>>> + return;
>>> + list_del(&vdesc->node);
>>> + fsl_comp = to_fsl_qdma_comp(vdesc);
>>> +
>>> + memcpy(fsl_queue->virt_head++,
>>> + fsl_comp->virt_addr, sizeof(struct fsl_qdma_format));
>>> + if (fsl_queue->virt_head == fsl_queue->cq + fsl_queue->n_cq)
>>> + fsl_queue->virt_head = fsl_queue->cq;
>>> +
>>> + list_add_tail(&fsl_comp->list, &fsl_queue->comp_used);
>>> + barrier();
>>> + reg = qdma_readl(fsl_chan->qdma, block +
>> FSL_QDMA_BCQMR(fsl_queue->id));
>>> + reg |= FSL_QDMA_BCQMR_EI;
>>> + qdma_writel(fsl_chan->qdma, reg, block +
>> FSL_QDMA_BCQMR(fsl_queue->id));
>>> + fsl_chan->status = DMA_IN_PROGRESS;
>>> +}
>>> +
>>> +static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc)
>>> +{
>>> + unsigned long flags;
>>> + struct fsl_qdma_comp *fsl_comp;
>>> + struct fsl_qdma_queue *fsl_queue;
>>> +
>>> + fsl_comp = to_fsl_qdma_comp(vdesc);
>>> + fsl_queue = fsl_comp->qchan->queue;
>>> +
>>> + spin_lock_irqsave(&fsl_queue->queue_lock, flags);
>>> + list_add_tail(&fsl_comp->list, &fsl_queue->comp_free);
>>> + spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
>>> +}
>>> +
>>> +static void fsl_qdma_issue_pending(struct dma_chan *chan)
>>> +{
>>> + unsigned long flags;
>>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>>> +
>>> + spin_lock_irqsave(&fsl_queue->queue_lock, flags);
>>> + spin_lock(&fsl_chan->vchan.lock);
>>> + if (vchan_issue_pending(&fsl_chan->vchan))
>>> + fsl_qdma_enqueue_desc(fsl_chan);
>>> + spin_unlock(&fsl_chan->vchan.lock);
>>> + spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
>>> +}
>>> +
>>> +static void fsl_qdma_synchronize(struct dma_chan *chan)
>>> +{
>>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>>> +
>>> + vchan_synchronize(&fsl_chan->vchan);
>>> +}
>>> +
>>> +static int fsl_qdma_terminate_all(struct dma_chan *chan)
>>> +{
>>> + LIST_HEAD(head);
>>> + unsigned long flags;
>>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>>> +
>>> + spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
>>> + vchan_get_all_descriptors(&fsl_chan->vchan, &head);
>>> + spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
>>> + vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
>>> + return 0;
>>> +}
>>> +
>>> +static int fsl_qdma_alloc_chan_resources(struct dma_chan *chan)
>>> +{
>>> + int ret;
>>> + struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
>>> + struct fsl_qdma_engine *fsl_qdma = fsl_chan->qdma;
>>> + struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
>>> +
>>> + if (fsl_queue->comp_pool && fsl_queue->desc_pool)
>>> + return fsl_qdma->desc_allocated;
>>> +
>>> + INIT_LIST_HEAD(&fsl_queue->comp_free);
>>> +
>>> + /*
>>> + * The dma pool for queue command buffer
>>> + */
>>> + fsl_queue->comp_pool =
>>> + dma_pool_create("comp_pool",
>>> + chan->device->dev,
>>> + FSL_QDMA_COMMAND_BUFFER_SIZE,
>>> + 64, 0);
>>> + if (!fsl_queue->comp_pool)
>>> + return -ENOMEM;
>>> +
>>> + /*
>>> + * The dma pool for Descriptor(SD/DD) buffer
>>> + */
>>> + fsl_queue->desc_pool =
>>> + dma_pool_create("desc_pool",
>>> + chan->device->dev,
>>> + FSL_QDMA_DESCRIPTOR_BUFFER_SIZE,
>>> + 32, 0);
>>> + if (!fsl_queue->desc_pool)
>>> + goto err_desc_pool;
>>> +
>>> + ret = fsl_qdma_pre_request_enqueue_desc(fsl_queue);
>>> + if (ret) {
>>> + dev_err(chan->device->dev,
>>> + "failed to alloc dma buffer for S/G
>> descriptor\n");
>>> + goto err_mem;
>>> + }
>>> +
>>> + fsl_qdma->desc_allocated++;
>>> + return fsl_qdma->desc_allocated;
>>> +
>>> +err_mem:
>>> + dma_pool_destroy(fsl_queue->desc_pool);
>>> +err_desc_pool:
>>> + dma_pool_destroy(fsl_queue->comp_pool);
>>> + return -ENOMEM;
>>> +}
>>> +
>>> +static int fsl_qdma_probe(struct platform_device *pdev)
>>> +{
>>> + int ret, i;
>>> + int blk_num, blk_off;
>>> + u32 len, chans, queues;
>>> + struct resource *res;
>>> + struct fsl_qdma_chan *fsl_chan;
>>> + struct fsl_qdma_engine *fsl_qdma;
>>> + struct device_node *np = pdev->dev.of_node;
>>> +
>>> + ret = of_property_read_u32(np, "dma-channels", &chans);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't get dma-channels.\n");
>>> + return ret;
>>> + }
>>> +
>>> + ret = of_property_read_u32(np, "block-offset", &blk_off);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't get block-offset.\n");
>>> + return ret;
>>> + }
>>> +
>>> + ret = of_property_read_u32(np, "block-number", &blk_num);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't get block-number.\n");
>>> + return ret;
>>> + }
>>> +
>>> + blk_num = min_t(int, blk_num, num_online_cpus());
>>> +
>>> + len = sizeof(*fsl_qdma);
>>> + fsl_qdma = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>>> + if (!fsl_qdma)
>>> + return -ENOMEM;
>>> +
>>> + len = sizeof(*fsl_chan) * chans;
>>> + fsl_qdma->chans = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>>> + if (!fsl_qdma->chans)
>>> + return -ENOMEM;
>>> +
>>> + len = sizeof(struct fsl_qdma_queue *) * blk_num;
>>> + fsl_qdma->status = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
>>> + if (!fsl_qdma->status)
>>> + return -ENOMEM;
>>> +
>>> + len = sizeof(int) * blk_num;
>>> + fsl_qdma->queue_irq = devm_kzalloc(&pdev->dev, len,
>> GFP_KERNEL);
>>> + if (!fsl_qdma->queue_irq)
>>> + return -ENOMEM;
>>> +
>>> + ret = of_property_read_u32(np, "fsl,dma-queues", &queues);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't get queues.\n");
>>> + return ret;
>>> + }
>>> +
>>> + fsl_qdma->desc_allocated = 0;
>>> + fsl_qdma->n_chans = chans;
>>> + fsl_qdma->n_queues = queues;
>>> + fsl_qdma->block_number = blk_num;
>>> + fsl_qdma->block_offset = blk_off;
>>> +
>>> + mutex_init(&fsl_qdma->fsl_qdma_mutex);
>>> +
>>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>>> + fsl_qdma->status[i] =
>> fsl_qdma_prep_status_queue(pdev);
>>> + if (!fsl_qdma->status[i])
>>> + return -ENOMEM;
>>> + }
>>> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
>>> + fsl_qdma->ctrl_base = devm_ioremap_resource(&pdev->dev, res);
>>> + if (IS_ERR(fsl_qdma->ctrl_base))
>>> + return PTR_ERR(fsl_qdma->ctrl_base);
>>> +
>>> + res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
>>> + fsl_qdma->status_base = devm_ioremap_resource(&pdev->dev,
>> res);
>>> + if (IS_ERR(fsl_qdma->status_base))
>>> + return PTR_ERR(fsl_qdma->status_base);
>>> +
>>> + res = platform_get_resource(pdev, IORESOURCE_MEM, 2);
>>> + fsl_qdma->block_base = devm_ioremap_resource(&pdev->dev,
>> res);
>>> + if (IS_ERR(fsl_qdma->block_base))
>>> + return PTR_ERR(fsl_qdma->block_base);
>>> + fsl_qdma->queue = fsl_qdma_alloc_queue_resources(pdev,
>> fsl_qdma);
>>> + if (!fsl_qdma->queue)
>>> + return -ENOMEM;
>>> +
>>> + ret = fsl_qdma_irq_init(pdev, fsl_qdma);
>>> + if (ret)
>>> + return ret;
>>> +
>>> + fsl_qdma->irq_base = platform_get_irq_byname(pdev,
>> "qdma-queue0");
>>> + fsl_qdma->feature = of_property_read_bool(np, "big-endian");
>>> + INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
>>> +
>>> + for (i = 0; i < fsl_qdma->n_chans; i++) {
>>> + struct fsl_qdma_chan *fsl_chan = &fsl_qdma->chans[i];
>>> +
>>> + fsl_chan->qdma = fsl_qdma;
>>> + fsl_chan->queue = fsl_qdma->queue + i %
>> (fsl_qdma->n_queues *
>>> +
>> fsl_qdma->block_number);
>>> + fsl_chan->vchan.desc_free = fsl_qdma_free_desc;
>>> + vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
>>> + }
>>> +
>>> + dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
>>> +
>>> + fsl_qdma->dma_dev.dev = &pdev->dev;
>>> + fsl_qdma->dma_dev.device_free_chan_resources =
>>> + fsl_qdma_free_chan_resources;
>>> + fsl_qdma->dma_dev.device_alloc_chan_resources =
>>> + fsl_qdma_alloc_chan_resources;
>>> + fsl_qdma->dma_dev.device_tx_status = dma_cookie_status;
>>> + fsl_qdma->dma_dev.device_prep_dma_memcpy =
>> fsl_qdma_prep_memcpy;
>>> + fsl_qdma->dma_dev.device_issue_pending =
>> fsl_qdma_issue_pending;
>>> + fsl_qdma->dma_dev.device_synchronize = fsl_qdma_synchronize;
>>> + fsl_qdma->dma_dev.device_terminate_all =
>> fsl_qdma_terminate_all;
>>> +
>>> + dma_set_mask(&pdev->dev, DMA_BIT_MASK(40));
>>> +
>>> + platform_set_drvdata(pdev, fsl_qdma);
>>> +
>>> + ret = dma_async_device_register(&fsl_qdma->dma_dev);
>>> + if (ret) {
>>> + dev_err(&pdev->dev,
>>> + "Can't register NXP Layerscape qDMA
>> engine.\n");
>>> + return ret;
>>> + }
>>> +
>>> + ret = fsl_qdma_reg_init(fsl_qdma);
>>> + if (ret) {
>>> + dev_err(&pdev->dev, "Can't Initialize the qDMA
>> engine.\n");
>>> + return ret;
>>> + }
>>> +
>>> + return 0;
>>> +}
>>> +
>>> +static void fsl_qdma_cleanup_vchan(struct dma_device *dmadev)
>>> +{
>>> + struct fsl_qdma_chan *chan, *_chan;
>>> +
>>> + list_for_each_entry_safe(chan, _chan,
>>> + &dmadev->channels,
>> vchan.chan.device_node) {
>>> + list_del(&chan->vchan.chan.device_node);
>>> + tasklet_kill(&chan->vchan.task);
>>> + }
>>> +}
>>> +
>>> +static int fsl_qdma_remove(struct platform_device *pdev)
>>> +{
>>> + int i;
>>> + struct fsl_qdma_queue *status;
>>> + struct device_node *np = pdev->dev.of_node;
>>> + struct fsl_qdma_engine *fsl_qdma = platform_get_drvdata(pdev);
>>> +
>>> + fsl_qdma_irq_exit(pdev, fsl_qdma);
>>> + fsl_qdma_cleanup_vchan(&fsl_qdma->dma_dev);
>>> + of_dma_controller_free(np);
>>> + dma_async_device_unregister(&fsl_qdma->dma_dev);
>>> +
>>> + for (i = 0; i < fsl_qdma->block_number; i++) {
>>> + status = fsl_qdma->status[i];
>>> + dma_free_coherent(&pdev->dev, sizeof(struct
>> fsl_qdma_format) *
>>> + status->n_cq, status->cq,
>> status->bus_addr);
>>> + }
>>> + return 0;
>>> +}
>>> +
>>> +static const struct of_device_id fsl_qdma_dt_ids[] = {
>>> + { .compatible = "fsl,ls1021a-qdma", },
>>> + { /* sentinel */ }
>>> +};
>>> +MODULE_DEVICE_TABLE(of, fsl_qdma_dt_ids);
>>> +
>>> +static struct platform_driver fsl_qdma_driver = {
>>> + .driver = {
>>> + .name = "fsl-qdma",
>>> + .of_match_table = fsl_qdma_dt_ids,
>>> + },
>>> + .probe = fsl_qdma_probe,
>>> + .remove = fsl_qdma_remove,
>>> +};
>>> +
>>> +module_platform_driver(fsl_qdma_driver);
>>> +
>>> +MODULE_ALIAS("platform:fsl-qdma");
>>> +MODULE_DESCRIPTION("NXP Layerscape qDMA engine driver");
>>> --
>>> 1.7.1
>>>