|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | /* | 
|  | * Driver for EIP97 cryptographic accelerator. | 
|  | * | 
|  | * Copyright (c) 2016 Ryder Lee <ryder.lee@mediatek.com> | 
|  | */ | 
|  |  | 
|  | #include <linux/clk.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/mod_devicetable.h> | 
|  | #include <linux/platform_device.h> | 
|  | #include <linux/pm_runtime.h> | 
|  | #include "mtk-platform.h" | 
|  |  | 
|  | #define MTK_BURST_SIZE_MSK		GENMASK(7, 4) | 
|  | #define MTK_BURST_SIZE(x)		((x) << 4) | 
|  | #define MTK_DESC_SIZE(x)		((x) << 0) | 
|  | #define MTK_DESC_OFFSET(x)		((x) << 16) | 
|  | #define MTK_DESC_FETCH_SIZE(x)		((x) << 0) | 
|  | #define MTK_DESC_FETCH_THRESH(x)	((x) << 16) | 
|  | #define MTK_DESC_OVL_IRQ_EN		BIT(25) | 
|  | #define MTK_DESC_ATP_PRESENT		BIT(30) | 
|  |  | 
|  | #define MTK_DFSE_IDLE			GENMASK(3, 0) | 
|  | #define MTK_DFSE_THR_CTRL_EN		BIT(30) | 
|  | #define MTK_DFSE_THR_CTRL_RESET		BIT(31) | 
|  | #define MTK_DFSE_RING_ID(x)		(((x) >> 12) & GENMASK(3, 0)) | 
|  | #define MTK_DFSE_MIN_DATA(x)		((x) << 0) | 
|  | #define MTK_DFSE_MAX_DATA(x)		((x) << 8) | 
|  | #define MTK_DFE_MIN_CTRL(x)		((x) << 16) | 
|  | #define MTK_DFE_MAX_CTRL(x)		((x) << 24) | 
|  |  | 
|  | #define MTK_IN_BUF_MIN_THRESH(x)	((x) << 8) | 
|  | #define MTK_IN_BUF_MAX_THRESH(x)	((x) << 12) | 
|  | #define MTK_OUT_BUF_MIN_THRESH(x)	((x) << 0) | 
|  | #define MTK_OUT_BUF_MAX_THRESH(x)	((x) << 4) | 
|  | #define MTK_IN_TBUF_SIZE(x)		(((x) >> 4) & GENMASK(3, 0)) | 
|  | #define MTK_IN_DBUF_SIZE(x)		(((x) >> 8) & GENMASK(3, 0)) | 
|  | #define MTK_OUT_DBUF_SIZE(x)		(((x) >> 16) & GENMASK(3, 0)) | 
|  | #define MTK_CMD_FIFO_SIZE(x)		(((x) >> 8) & GENMASK(3, 0)) | 
|  | #define MTK_RES_FIFO_SIZE(x)		(((x) >> 12) & GENMASK(3, 0)) | 
|  |  | 
|  | #define MTK_PE_TK_LOC_AVL		BIT(2) | 
|  | #define MTK_PE_PROC_HELD		BIT(14) | 
|  | #define MTK_PE_TK_TIMEOUT_EN		BIT(22) | 
|  | #define MTK_PE_INPUT_DMA_ERR		BIT(0) | 
|  | #define MTK_PE_OUTPUT_DMA_ERR		BIT(1) | 
|  | #define MTK_PE_PKT_PORC_ERR		BIT(2) | 
|  | #define MTK_PE_PKT_TIMEOUT		BIT(3) | 
|  | #define MTK_PE_FATAL_ERR		BIT(14) | 
|  | #define MTK_PE_INPUT_DMA_ERR_EN		BIT(16) | 
|  | #define MTK_PE_OUTPUT_DMA_ERR_EN	BIT(17) | 
|  | #define MTK_PE_PKT_PORC_ERR_EN		BIT(18) | 
|  | #define MTK_PE_PKT_TIMEOUT_EN		BIT(19) | 
|  | #define MTK_PE_FATAL_ERR_EN		BIT(30) | 
|  | #define MTK_PE_INT_OUT_EN		BIT(31) | 
|  |  | 
|  | #define MTK_HIA_SIGNATURE		((u16)0x35ca) | 
|  | #define MTK_HIA_DATA_WIDTH(x)		(((x) >> 25) & GENMASK(1, 0)) | 
|  | #define MTK_HIA_DMA_LENGTH(x)		(((x) >> 20) & GENMASK(4, 0)) | 
|  | #define MTK_CDR_STAT_CLR		GENMASK(4, 0) | 
|  | #define MTK_RDR_STAT_CLR		GENMASK(7, 0) | 
|  |  | 
|  | #define MTK_AIC_INT_MSK			GENMASK(5, 0) | 
|  | #define MTK_AIC_VER_MSK			(GENMASK(15, 0) | GENMASK(27, 20)) | 
|  | #define MTK_AIC_VER11			0x011036c9 | 
|  | #define MTK_AIC_VER12			0x012036c9 | 
|  | #define MTK_AIC_G_CLR			GENMASK(30, 20) | 
|  |  | 
|  | /** | 
|  | * EIP97 is an integrated security subsystem to accelerate cryptographic | 
|  | * functions and protocols to offload the host processor. | 
|  | * Some important hardware modules are briefly introduced below: | 
|  | * | 
|  | * Host Interface Adapter(HIA) - the main interface between the host | 
|  | * system and the hardware subsystem. It is responsible for attaching | 
|  | * processing engine to the specific host bus interface and provides a | 
|  | * standardized software view for off loading tasks to the engine. | 
|  | * | 
|  | * Command Descriptor Ring Manager(CDR Manager) - keeps track of how many | 
|  | * CD the host has prepared in the CDR. It monitors the fill level of its | 
|  | * CD-FIFO and if there's sufficient space for the next block of descriptors, | 
|  | * then it fires off a DMA request to fetch a block of CDs. | 
|  | * | 
|  | * Data fetch engine(DFE) - It is responsible for parsing the CD and | 
|  | * setting up the required control and packet data DMA transfers from | 
|  | * system memory to the processing engine. | 
|  | * | 
|  | * Result Descriptor Ring Manager(RDR Manager) - same as CDR Manager, | 
|  | * but target is result descriptors, Moreover, it also handles the RD | 
|  | * updates under control of the DSE. For each packet data segment | 
|  | * processed, the DSE triggers the RDR Manager to write the updated RD. | 
|  | * If triggered to update, the RDR Manager sets up a DMA operation to | 
|  | * copy the RD from the DSE to the correct location in the RDR. | 
|  | * | 
|  | * Data Store Engine(DSE) - It is responsible for parsing the prepared RD | 
|  | * and setting up the required control and packet data DMA transfers from | 
|  | * the processing engine to system memory. | 
|  | * | 
|  | * Advanced Interrupt Controllers(AICs) - receive interrupt request signals | 
|  | * from various sources and combine them into one interrupt output. | 
|  | * The AICs are used by: | 
|  | * - One for the HIA global and processing engine interrupts. | 
|  | * - The others for the descriptor ring interrupts. | 
|  | */ | 
|  |  | 
|  | /* Cryptographic engine capabilities */ | 
|  | struct mtk_sys_cap { | 
|  | /* host interface adapter */ | 
|  | u32 hia_ver; | 
|  | u32 hia_opt; | 
|  | /* packet engine */ | 
|  | u32 pkt_eng_opt; | 
|  | /* global hardware */ | 
|  | u32 hw_opt; | 
|  | }; | 
|  |  | 
|  | static void mtk_desc_ring_link(struct mtk_cryp *cryp, u32 mask) | 
|  | { | 
|  | /* Assign rings to DFE/DSE thread and enable it */ | 
|  | writel(MTK_DFSE_THR_CTRL_EN | mask, cryp->base + DFE_THR_CTRL); | 
|  | writel(MTK_DFSE_THR_CTRL_EN | mask, cryp->base + DSE_THR_CTRL); | 
|  | } | 
|  |  | 
|  | static void mtk_dfe_dse_buf_setup(struct mtk_cryp *cryp, | 
|  | struct mtk_sys_cap *cap) | 
|  | { | 
|  | u32 width = MTK_HIA_DATA_WIDTH(cap->hia_opt) + 2; | 
|  | u32 len = MTK_HIA_DMA_LENGTH(cap->hia_opt) - 1; | 
|  | u32 ipbuf = min((u32)MTK_IN_DBUF_SIZE(cap->hw_opt) + width, len); | 
|  | u32 opbuf = min((u32)MTK_OUT_DBUF_SIZE(cap->hw_opt) + width, len); | 
|  | u32 itbuf = min((u32)MTK_IN_TBUF_SIZE(cap->hw_opt) + width, len); | 
|  |  | 
|  | writel(MTK_DFSE_MIN_DATA(ipbuf - 1) | | 
|  | MTK_DFSE_MAX_DATA(ipbuf) | | 
|  | MTK_DFE_MIN_CTRL(itbuf - 1) | | 
|  | MTK_DFE_MAX_CTRL(itbuf), | 
|  | cryp->base + DFE_CFG); | 
|  |  | 
|  | writel(MTK_DFSE_MIN_DATA(opbuf - 1) | | 
|  | MTK_DFSE_MAX_DATA(opbuf), | 
|  | cryp->base + DSE_CFG); | 
|  |  | 
|  | writel(MTK_IN_BUF_MIN_THRESH(ipbuf - 1) | | 
|  | MTK_IN_BUF_MAX_THRESH(ipbuf), | 
|  | cryp->base + PE_IN_DBUF_THRESH); | 
|  |  | 
|  | writel(MTK_IN_BUF_MIN_THRESH(itbuf - 1) | | 
|  | MTK_IN_BUF_MAX_THRESH(itbuf), | 
|  | cryp->base + PE_IN_TBUF_THRESH); | 
|  |  | 
|  | writel(MTK_OUT_BUF_MIN_THRESH(opbuf - 1) | | 
|  | MTK_OUT_BUF_MAX_THRESH(opbuf), | 
|  | cryp->base + PE_OUT_DBUF_THRESH); | 
|  |  | 
|  | writel(0, cryp->base + PE_OUT_TBUF_THRESH); | 
|  | writel(0, cryp->base + PE_OUT_BUF_CTRL); | 
|  | } | 
|  |  | 
|  | static int mtk_dfe_dse_state_check(struct mtk_cryp *cryp) | 
|  | { | 
|  | int ret = -EINVAL; | 
|  | u32 val; | 
|  |  | 
|  | /* Check for completion of all DMA transfers */ | 
|  | val = readl(cryp->base + DFE_THR_STAT); | 
|  | if (MTK_DFSE_RING_ID(val) == MTK_DFSE_IDLE) { | 
|  | val = readl(cryp->base + DSE_THR_STAT); | 
|  | if (MTK_DFSE_RING_ID(val) == MTK_DFSE_IDLE) | 
|  | ret = 0; | 
|  | } | 
|  |  | 
|  | if (!ret) { | 
|  | /* Take DFE/DSE thread out of reset */ | 
|  | writel(0, cryp->base + DFE_THR_CTRL); | 
|  | writel(0, cryp->base + DSE_THR_CTRL); | 
|  | } else { | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int mtk_dfe_dse_reset(struct mtk_cryp *cryp) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | /* Reset DSE/DFE and correct system priorities for all rings. */ | 
|  | writel(MTK_DFSE_THR_CTRL_RESET, cryp->base + DFE_THR_CTRL); | 
|  | writel(0, cryp->base + DFE_PRIO_0); | 
|  | writel(0, cryp->base + DFE_PRIO_1); | 
|  | writel(0, cryp->base + DFE_PRIO_2); | 
|  | writel(0, cryp->base + DFE_PRIO_3); | 
|  |  | 
|  | writel(MTK_DFSE_THR_CTRL_RESET, cryp->base + DSE_THR_CTRL); | 
|  | writel(0, cryp->base + DSE_PRIO_0); | 
|  | writel(0, cryp->base + DSE_PRIO_1); | 
|  | writel(0, cryp->base + DSE_PRIO_2); | 
|  | writel(0, cryp->base + DSE_PRIO_3); | 
|  |  | 
|  | err = mtk_dfe_dse_state_check(cryp); | 
|  | if (err) | 
|  | return err; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void mtk_cmd_desc_ring_setup(struct mtk_cryp *cryp, | 
|  | int i, struct mtk_sys_cap *cap) | 
|  | { | 
|  | /* Full descriptor that fits FIFO minus one */ | 
|  | u32 count = | 
|  | ((1 << MTK_CMD_FIFO_SIZE(cap->hia_opt)) / MTK_DESC_SZ) - 1; | 
|  |  | 
|  | /* Temporarily disable external triggering */ | 
|  | writel(0, cryp->base + CDR_CFG(i)); | 
|  |  | 
|  | /* Clear CDR count */ | 
|  | writel(MTK_CNT_RST, cryp->base + CDR_PREP_COUNT(i)); | 
|  | writel(MTK_CNT_RST, cryp->base + CDR_PROC_COUNT(i)); | 
|  |  | 
|  | writel(0, cryp->base + CDR_PREP_PNTR(i)); | 
|  | writel(0, cryp->base + CDR_PROC_PNTR(i)); | 
|  | writel(0, cryp->base + CDR_DMA_CFG(i)); | 
|  |  | 
|  | /* Configure CDR host address space */ | 
|  | writel(0, cryp->base + CDR_BASE_ADDR_HI(i)); | 
|  | writel(cryp->ring[i]->cmd_dma, cryp->base + CDR_BASE_ADDR_LO(i)); | 
|  |  | 
|  | writel(MTK_DESC_RING_SZ, cryp->base + CDR_RING_SIZE(i)); | 
|  |  | 
|  | /* Clear and disable all CDR interrupts */ | 
|  | writel(MTK_CDR_STAT_CLR, cryp->base + CDR_STAT(i)); | 
|  |  | 
|  | /* | 
|  | * Set command descriptor offset and enable additional | 
|  | * token present in descriptor. | 
|  | */ | 
|  | writel(MTK_DESC_SIZE(MTK_DESC_SZ) | | 
|  | MTK_DESC_OFFSET(MTK_DESC_OFF) | | 
|  | MTK_DESC_ATP_PRESENT, | 
|  | cryp->base + CDR_DESC_SIZE(i)); | 
|  |  | 
|  | writel(MTK_DESC_FETCH_SIZE(count * MTK_DESC_OFF) | | 
|  | MTK_DESC_FETCH_THRESH(count * MTK_DESC_SZ), | 
|  | cryp->base + CDR_CFG(i)); | 
|  | } | 
|  |  | 
|  | static void mtk_res_desc_ring_setup(struct mtk_cryp *cryp, | 
|  | int i, struct mtk_sys_cap *cap) | 
|  | { | 
|  | u32 rndup = 2; | 
|  | u32 count = ((1 << MTK_RES_FIFO_SIZE(cap->hia_opt)) / rndup) - 1; | 
|  |  | 
|  | /* Temporarily disable external triggering */ | 
|  | writel(0, cryp->base + RDR_CFG(i)); | 
|  |  | 
|  | /* Clear RDR count */ | 
|  | writel(MTK_CNT_RST, cryp->base + RDR_PREP_COUNT(i)); | 
|  | writel(MTK_CNT_RST, cryp->base + RDR_PROC_COUNT(i)); | 
|  |  | 
|  | writel(0, cryp->base + RDR_PREP_PNTR(i)); | 
|  | writel(0, cryp->base + RDR_PROC_PNTR(i)); | 
|  | writel(0, cryp->base + RDR_DMA_CFG(i)); | 
|  |  | 
|  | /* Configure RDR host address space */ | 
|  | writel(0, cryp->base + RDR_BASE_ADDR_HI(i)); | 
|  | writel(cryp->ring[i]->res_dma, cryp->base + RDR_BASE_ADDR_LO(i)); | 
|  |  | 
|  | writel(MTK_DESC_RING_SZ, cryp->base + RDR_RING_SIZE(i)); | 
|  | writel(MTK_RDR_STAT_CLR, cryp->base + RDR_STAT(i)); | 
|  |  | 
|  | /* | 
|  | * RDR manager generates update interrupts on a per-completed-packet, | 
|  | * and the rd_proc_thresh_irq interrupt is fired when proc_pkt_count | 
|  | * for the RDR exceeds the number of packets. | 
|  | */ | 
|  | writel(MTK_RDR_PROC_THRESH | MTK_RDR_PROC_MODE, | 
|  | cryp->base + RDR_THRESH(i)); | 
|  |  | 
|  | /* | 
|  | * Configure a threshold and time-out value for the processed | 
|  | * result descriptors (or complete packets) that are written to | 
|  | * the RDR. | 
|  | */ | 
|  | writel(MTK_DESC_SIZE(MTK_DESC_SZ) | MTK_DESC_OFFSET(MTK_DESC_OFF), | 
|  | cryp->base + RDR_DESC_SIZE(i)); | 
|  |  | 
|  | /* | 
|  | * Configure HIA fetch size and fetch threshold that are used to | 
|  | * fetch blocks of multiple descriptors. | 
|  | */ | 
|  | writel(MTK_DESC_FETCH_SIZE(count * MTK_DESC_OFF) | | 
|  | MTK_DESC_FETCH_THRESH(count * rndup) | | 
|  | MTK_DESC_OVL_IRQ_EN, | 
|  | cryp->base + RDR_CFG(i)); | 
|  | } | 
|  |  | 
|  | static int mtk_packet_engine_setup(struct mtk_cryp *cryp) | 
|  | { | 
|  | struct mtk_sys_cap cap; | 
|  | int i, err; | 
|  | u32 val; | 
|  |  | 
|  | cap.hia_ver = readl(cryp->base + HIA_VERSION); | 
|  | cap.hia_opt = readl(cryp->base + HIA_OPTIONS); | 
|  | cap.hw_opt = readl(cryp->base + EIP97_OPTIONS); | 
|  |  | 
|  | if (!(((u16)cap.hia_ver) == MTK_HIA_SIGNATURE)) | 
|  | return -EINVAL; | 
|  |  | 
|  | /* Configure endianness conversion method for master (DMA) interface */ | 
|  | writel(0, cryp->base + EIP97_MST_CTRL); | 
|  |  | 
|  | /* Set HIA burst size */ | 
|  | val = readl(cryp->base + HIA_MST_CTRL); | 
|  | val &= ~MTK_BURST_SIZE_MSK; | 
|  | val |= MTK_BURST_SIZE(5); | 
|  | writel(val, cryp->base + HIA_MST_CTRL); | 
|  |  | 
|  | err = mtk_dfe_dse_reset(cryp); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Failed to reset DFE and DSE.\n"); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | mtk_dfe_dse_buf_setup(cryp, &cap); | 
|  |  | 
|  | /* Enable the 4 rings for the packet engines. */ | 
|  | mtk_desc_ring_link(cryp, 0xf); | 
|  |  | 
|  | for (i = 0; i < MTK_RING_MAX; i++) { | 
|  | mtk_cmd_desc_ring_setup(cryp, i, &cap); | 
|  | mtk_res_desc_ring_setup(cryp, i, &cap); | 
|  | } | 
|  |  | 
|  | writel(MTK_PE_TK_LOC_AVL | MTK_PE_PROC_HELD | MTK_PE_TK_TIMEOUT_EN, | 
|  | cryp->base + PE_TOKEN_CTRL_STAT); | 
|  |  | 
|  | /* Clear all pending interrupts */ | 
|  | writel(MTK_AIC_G_CLR, cryp->base + AIC_G_ACK); | 
|  | writel(MTK_PE_INPUT_DMA_ERR | MTK_PE_OUTPUT_DMA_ERR | | 
|  | MTK_PE_PKT_PORC_ERR | MTK_PE_PKT_TIMEOUT | | 
|  | MTK_PE_FATAL_ERR | MTK_PE_INPUT_DMA_ERR_EN | | 
|  | MTK_PE_OUTPUT_DMA_ERR_EN | MTK_PE_PKT_PORC_ERR_EN | | 
|  | MTK_PE_PKT_TIMEOUT_EN | MTK_PE_FATAL_ERR_EN | | 
|  | MTK_PE_INT_OUT_EN, | 
|  | cryp->base + PE_INTERRUPT_CTRL_STAT); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int mtk_aic_cap_check(struct mtk_cryp *cryp, int hw) | 
|  | { | 
|  | u32 val; | 
|  |  | 
|  | if (hw == MTK_RING_MAX) | 
|  | val = readl(cryp->base + AIC_G_VERSION); | 
|  | else | 
|  | val = readl(cryp->base + AIC_VERSION(hw)); | 
|  |  | 
|  | val &= MTK_AIC_VER_MSK; | 
|  | if (val != MTK_AIC_VER11 && val != MTK_AIC_VER12) | 
|  | return -ENXIO; | 
|  |  | 
|  | if (hw == MTK_RING_MAX) | 
|  | val = readl(cryp->base + AIC_G_OPTIONS); | 
|  | else | 
|  | val = readl(cryp->base + AIC_OPTIONS(hw)); | 
|  |  | 
|  | val &= MTK_AIC_INT_MSK; | 
|  | if (!val || val > 32) | 
|  | return -ENXIO; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int mtk_aic_init(struct mtk_cryp *cryp, int hw) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | err = mtk_aic_cap_check(cryp, hw); | 
|  | if (err) | 
|  | return err; | 
|  |  | 
|  | /* Disable all interrupts and set initial configuration */ | 
|  | if (hw == MTK_RING_MAX) { | 
|  | writel(0, cryp->base + AIC_G_ENABLE_CTRL); | 
|  | writel(0, cryp->base + AIC_G_POL_CTRL); | 
|  | writel(0, cryp->base + AIC_G_TYPE_CTRL); | 
|  | writel(0, cryp->base + AIC_G_ENABLE_SET); | 
|  | } else { | 
|  | writel(0, cryp->base + AIC_ENABLE_CTRL(hw)); | 
|  | writel(0, cryp->base + AIC_POL_CTRL(hw)); | 
|  | writel(0, cryp->base + AIC_TYPE_CTRL(hw)); | 
|  | writel(0, cryp->base + AIC_ENABLE_SET(hw)); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int mtk_accelerator_init(struct mtk_cryp *cryp) | 
|  | { | 
|  | int i, err; | 
|  |  | 
|  | /* Initialize advanced interrupt controller(AIC) */ | 
|  | for (i = 0; i < MTK_IRQ_NUM; i++) { | 
|  | err = mtk_aic_init(cryp, i); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Failed to initialize AIC.\n"); | 
|  | return err; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Initialize packet engine */ | 
|  | err = mtk_packet_engine_setup(cryp); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Failed to configure packet engine.\n"); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void mtk_desc_dma_free(struct mtk_cryp *cryp) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < MTK_RING_MAX; i++) { | 
|  | dma_free_coherent(cryp->dev, MTK_DESC_RING_SZ, | 
|  | cryp->ring[i]->res_base, | 
|  | cryp->ring[i]->res_dma); | 
|  | dma_free_coherent(cryp->dev, MTK_DESC_RING_SZ, | 
|  | cryp->ring[i]->cmd_base, | 
|  | cryp->ring[i]->cmd_dma); | 
|  | kfree(cryp->ring[i]); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int mtk_desc_ring_alloc(struct mtk_cryp *cryp) | 
|  | { | 
|  | struct mtk_ring **ring = cryp->ring; | 
|  | int i, err = ENOMEM; | 
|  |  | 
|  | for (i = 0; i < MTK_RING_MAX; i++) { | 
|  | ring[i] = kzalloc(sizeof(**ring), GFP_KERNEL); | 
|  | if (!ring[i]) | 
|  | goto err_cleanup; | 
|  |  | 
|  | ring[i]->cmd_base = dma_alloc_coherent(cryp->dev, | 
|  | MTK_DESC_RING_SZ, | 
|  | &ring[i]->cmd_dma, | 
|  | GFP_KERNEL); | 
|  | if (!ring[i]->cmd_base) | 
|  | goto err_cleanup; | 
|  |  | 
|  | ring[i]->res_base = dma_alloc_coherent(cryp->dev, | 
|  | MTK_DESC_RING_SZ, | 
|  | &ring[i]->res_dma, | 
|  | GFP_KERNEL); | 
|  | if (!ring[i]->res_base) | 
|  | goto err_cleanup; | 
|  |  | 
|  | ring[i]->cmd_next = ring[i]->cmd_base; | 
|  | ring[i]->res_next = ring[i]->res_base; | 
|  | } | 
|  | return 0; | 
|  |  | 
|  | err_cleanup: | 
|  | for (; i--; ) { | 
|  | dma_free_coherent(cryp->dev, MTK_DESC_RING_SZ, | 
|  | ring[i]->res_base, ring[i]->res_dma); | 
|  | dma_free_coherent(cryp->dev, MTK_DESC_RING_SZ, | 
|  | ring[i]->cmd_base, ring[i]->cmd_dma); | 
|  | kfree(ring[i]); | 
|  | } | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static int mtk_crypto_probe(struct platform_device *pdev) | 
|  | { | 
|  | struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | 
|  | struct mtk_cryp *cryp; | 
|  | int i, err; | 
|  |  | 
|  | cryp = devm_kzalloc(&pdev->dev, sizeof(*cryp), GFP_KERNEL); | 
|  | if (!cryp) | 
|  | return -ENOMEM; | 
|  |  | 
|  | cryp->base = devm_ioremap_resource(&pdev->dev, res); | 
|  | if (IS_ERR(cryp->base)) | 
|  | return PTR_ERR(cryp->base); | 
|  |  | 
|  | for (i = 0; i < MTK_IRQ_NUM; i++) { | 
|  | cryp->irq[i] = platform_get_irq(pdev, i); | 
|  | if (cryp->irq[i] < 0) { | 
|  | dev_err(cryp->dev, "no IRQ:%d resource info\n", i); | 
|  | return cryp->irq[i]; | 
|  | } | 
|  | } | 
|  |  | 
|  | cryp->clk_cryp = devm_clk_get(&pdev->dev, "cryp"); | 
|  | if (IS_ERR(cryp->clk_cryp)) | 
|  | return -EPROBE_DEFER; | 
|  |  | 
|  | cryp->dev = &pdev->dev; | 
|  | pm_runtime_enable(cryp->dev); | 
|  | pm_runtime_get_sync(cryp->dev); | 
|  |  | 
|  | err = clk_prepare_enable(cryp->clk_cryp); | 
|  | if (err) | 
|  | goto err_clk_cryp; | 
|  |  | 
|  | /* Allocate four command/result descriptor rings */ | 
|  | err = mtk_desc_ring_alloc(cryp); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Unable to allocate descriptor rings.\n"); | 
|  | goto err_resource; | 
|  | } | 
|  |  | 
|  | /* Initialize hardware modules */ | 
|  | err = mtk_accelerator_init(cryp); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Failed to initialize cryptographic engine.\n"); | 
|  | goto err_engine; | 
|  | } | 
|  |  | 
|  | err = mtk_cipher_alg_register(cryp); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Unable to register cipher algorithm.\n"); | 
|  | goto err_cipher; | 
|  | } | 
|  |  | 
|  | err = mtk_hash_alg_register(cryp); | 
|  | if (err) { | 
|  | dev_err(cryp->dev, "Unable to register hash algorithm.\n"); | 
|  | goto err_hash; | 
|  | } | 
|  |  | 
|  | platform_set_drvdata(pdev, cryp); | 
|  | return 0; | 
|  |  | 
|  | err_hash: | 
|  | mtk_cipher_alg_release(cryp); | 
|  | err_cipher: | 
|  | mtk_dfe_dse_reset(cryp); | 
|  | err_engine: | 
|  | mtk_desc_dma_free(cryp); | 
|  | err_resource: | 
|  | clk_disable_unprepare(cryp->clk_cryp); | 
|  | err_clk_cryp: | 
|  | pm_runtime_put_sync(cryp->dev); | 
|  | pm_runtime_disable(cryp->dev); | 
|  |  | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static int mtk_crypto_remove(struct platform_device *pdev) | 
|  | { | 
|  | struct mtk_cryp *cryp = platform_get_drvdata(pdev); | 
|  |  | 
|  | mtk_hash_alg_release(cryp); | 
|  | mtk_cipher_alg_release(cryp); | 
|  | mtk_desc_dma_free(cryp); | 
|  |  | 
|  | clk_disable_unprepare(cryp->clk_cryp); | 
|  |  | 
|  | pm_runtime_put_sync(cryp->dev); | 
|  | pm_runtime_disable(cryp->dev); | 
|  | platform_set_drvdata(pdev, NULL); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static const struct of_device_id of_crypto_id[] = { | 
|  | { .compatible = "mediatek,eip97-crypto" }, | 
|  | {}, | 
|  | }; | 
|  | MODULE_DEVICE_TABLE(of, of_crypto_id); | 
|  |  | 
|  | static struct platform_driver mtk_crypto_driver = { | 
|  | .probe = mtk_crypto_probe, | 
|  | .remove = mtk_crypto_remove, | 
|  | .driver = { | 
|  | .name = "mtk-crypto", | 
|  | .of_match_table = of_crypto_id, | 
|  | }, | 
|  | }; | 
|  | module_platform_driver(mtk_crypto_driver); | 
|  |  | 
|  | MODULE_LICENSE("GPL"); | 
|  | MODULE_AUTHOR("Ryder Lee <ryder.lee@mediatek.com>"); | 
|  | MODULE_DESCRIPTION("Cryptographic accelerator driver for EIP97"); |