blob: ec59ef58044707f1f7a0f9120fe6079f25b114cb [file] [log] [blame]
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +05301// SPDX-License-Identifier: GPL-2.0+
2/*
3 * (C) Copyright 2018 Xilinx
4 *
5 * Xilinx ZynqMP Generic Quad-SPI(QSPI) controller driver(master mode only)
6 */
7
8#include <common.h>
Simon Glass1eb69ae2019-11-14 12:57:39 -07009#include <cpu_func.h>
Simon Glassf7ae49f2020-05-10 11:40:05 -060010#include <log.h>
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053011#include <asm/arch/sys_proto.h>
Simon Glass90526e92020-05-10 11:39:56 -060012#include <asm/cache.h>
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053013#include <asm/io.h>
14#include <clk.h>
15#include <dm.h>
16#include <malloc.h>
17#include <memalign.h>
18#include <spi.h>
Brandon Maierf1fd79a2021-01-20 10:39:46 -060019#include <spi-mem.h>
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053020#include <ubi_uboot.h>
21#include <wait_bit.h>
Simon Glass336d4612020-02-03 07:36:16 -070022#include <dm/device_compat.h>
Simon Glasscd93d622020-05-10 11:40:13 -060023#include <linux/bitops.h>
Simon Glass61b29b82020-02-03 07:36:15 -070024#include <linux/err.h>
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -060025#include <linux/sizes.h>
Ashok Reddy Somaf4f1b652022-08-25 06:59:01 -060026#include <zynqmp_firmware.h>
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053027
28#define GQSPI_GFIFO_STRT_MODE_MASK BIT(29)
29#define GQSPI_CONFIG_MODE_EN_MASK (3 << 30)
30#define GQSPI_CONFIG_DMA_MODE (2 << 30)
31#define GQSPI_CONFIG_CPHA_MASK BIT(2)
32#define GQSPI_CONFIG_CPOL_MASK BIT(1)
33
34/*
35 * QSPI Interrupt Registers bit Masks
36 *
37 * All the four interrupt registers (Status/Mask/Enable/Disable) have the same
38 * bit definitions.
39 */
40#define GQSPI_IXR_TXNFULL_MASK 0x00000004 /* QSPI TX FIFO Overflow */
41#define GQSPI_IXR_TXFULL_MASK 0x00000008 /* QSPI TX FIFO is full */
Ashok Reddy Soma4f9d2552021-10-19 19:43:00 +053042#define GQSPI_IXR_TXFIFOEMPTY_MASK 0x00000100 /* QSPI TX FIFO is Empty */
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053043#define GQSPI_IXR_RXNEMTY_MASK 0x00000010 /* QSPI RX FIFO Not Empty */
44#define GQSPI_IXR_GFEMTY_MASK 0x00000080 /* QSPI Generic FIFO Empty */
Ashok Reddy Soma2ffa6532021-05-25 06:36:27 -060045#define GQSPI_IXR_GFNFULL_MASK 0x00000200 /* QSPI GENFIFO not full */
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053046#define GQSPI_IXR_ALL_MASK (GQSPI_IXR_TXNFULL_MASK | \
47 GQSPI_IXR_RXNEMTY_MASK)
48
49/*
50 * QSPI Enable Register bit Masks
51 *
52 * This register is used to enable or disable the QSPI controller
53 */
54#define GQSPI_ENABLE_ENABLE_MASK 0x00000001 /* QSPI Enable Bit Mask */
55
56#define GQSPI_GFIFO_LOW_BUS BIT(14)
57#define GQSPI_GFIFO_CS_LOWER BIT(12)
58#define GQSPI_GFIFO_UP_BUS BIT(15)
59#define GQSPI_GFIFO_CS_UPPER BIT(13)
60#define GQSPI_SPI_MODE_QSPI (3 << 10)
61#define GQSPI_SPI_MODE_SPI BIT(10)
62#define GQSPI_SPI_MODE_DUAL_SPI (2 << 10)
63#define GQSPI_IMD_DATA_CS_ASSERT 5
64#define GQSPI_IMD_DATA_CS_DEASSERT 5
65#define GQSPI_GFIFO_TX BIT(16)
66#define GQSPI_GFIFO_RX BIT(17)
67#define GQSPI_GFIFO_STRIPE_MASK BIT(18)
68#define GQSPI_GFIFO_IMD_MASK 0xFF
69#define GQSPI_GFIFO_EXP_MASK BIT(9)
70#define GQSPI_GFIFO_DATA_XFR_MASK BIT(8)
71#define GQSPI_STRT_GEN_FIFO BIT(28)
72#define GQSPI_GEN_FIFO_STRT_MOD BIT(29)
73#define GQSPI_GFIFO_WP_HOLD BIT(19)
74#define GQSPI_BAUD_DIV_MASK (7 << 3)
75#define GQSPI_DFLT_BAUD_RATE_DIV BIT(3)
76#define GQSPI_GFIFO_ALL_INT_MASK 0xFBE
77#define GQSPI_DMA_DST_I_STS_DONE BIT(1)
78#define GQSPI_DMA_DST_I_STS_MASK 0xFE
79#define MODEBITS 0x6
80
81#define GQSPI_GFIFO_SELECT BIT(0)
82#define GQSPI_FIFO_THRESHOLD 1
Ashok Reddy Soma020b3532021-08-20 07:43:17 -060083#define GQSPI_GENFIFO_THRESHOLD 31
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053084
85#define SPI_XFER_ON_BOTH 0
86#define SPI_XFER_ON_LOWER 1
87#define SPI_XFER_ON_UPPER 2
88
89#define GQSPI_DMA_ALIGN 0x4
90#define GQSPI_MAX_BAUD_RATE_VAL 7
91#define GQSPI_DFLT_BAUD_RATE_VAL 2
92
93#define GQSPI_TIMEOUT 100000000
94
95#define GQSPI_BAUD_DIV_SHIFT 2
96#define GQSPI_LPBK_DLY_ADJ_LPBK_SHIFT 5
T Karthik Reddya5e770b2022-11-23 02:04:51 -070097#define GQSPI_LPBK_DLY_ADJ_DLY_1 0x1
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +053098#define GQSPI_LPBK_DLY_ADJ_DLY_1_SHIFT 3
99#define GQSPI_LPBK_DLY_ADJ_DLY_0 0x3
100#define GQSPI_USE_DATA_DLY 0x1
101#define GQSPI_USE_DATA_DLY_SHIFT 31
102#define GQSPI_DATA_DLY_ADJ_VALUE 0x2
103#define GQSPI_DATA_DLY_ADJ_SHIFT 28
104#define TAP_DLY_BYPASS_LQSPI_RX_VALUE 0x1
105#define TAP_DLY_BYPASS_LQSPI_RX_SHIFT 2
106#define GQSPI_DATA_DLY_ADJ_OFST 0x000001F8
Ashok Reddy Soma450d8eb2022-11-16 16:40:30 +0100107#define IOU_TAPDLY_BYPASS_OFST !(IS_ENABLED(CONFIG_ARCH_VERSAL) || \
108 IS_ENABLED(CONFIG_ARCH_VERSAL_NET)) ? \
Ashok Reddy Somaf4f1b652022-08-25 06:59:01 -0600109 0xFF180390 : 0xF103003C
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530110#define GQSPI_LPBK_DLY_ADJ_LPBK_MASK 0x00000020
Ashok Reddy Somaf4f1b652022-08-25 06:59:01 -0600111#define GQSPI_FREQ_37_5MHZ 37500000
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530112#define GQSPI_FREQ_40MHZ 40000000
113#define GQSPI_FREQ_100MHZ 100000000
114#define GQSPI_FREQ_150MHZ 150000000
115#define IOU_TAPDLY_BYPASS_MASK 0x7
116
117#define GQSPI_REG_OFFSET 0x100
118#define GQSPI_DMA_REG_OFFSET 0x800
119
120/* QSPI register offsets */
121struct zynqmp_qspi_regs {
122 u32 confr; /* 0x00 */
123 u32 isr; /* 0x04 */
124 u32 ier; /* 0x08 */
125 u32 idisr; /* 0x0C */
126 u32 imaskr; /* 0x10 */
127 u32 enbr; /* 0x14 */
128 u32 dr; /* 0x18 */
129 u32 txd0r; /* 0x1C */
130 u32 drxr; /* 0x20 */
131 u32 sicr; /* 0x24 */
132 u32 txftr; /* 0x28 */
133 u32 rxftr; /* 0x2C */
134 u32 gpior; /* 0x30 */
135 u32 reserved0; /* 0x34 */
136 u32 lpbkdly; /* 0x38 */
137 u32 reserved1; /* 0x3C */
138 u32 genfifo; /* 0x40 */
139 u32 gqspisel; /* 0x44 */
140 u32 reserved2; /* 0x48 */
141 u32 gqfifoctrl; /* 0x4C */
142 u32 gqfthr; /* 0x50 */
143 u32 gqpollcfg; /* 0x54 */
144 u32 gqpollto; /* 0x58 */
145 u32 gqxfersts; /* 0x5C */
146 u32 gqfifosnap; /* 0x60 */
147 u32 gqrxcpy; /* 0x64 */
148 u32 reserved3[36]; /* 0x68 */
149 u32 gqspidlyadj; /* 0xF8 */
150};
151
152struct zynqmp_qspi_dma_regs {
153 u32 dmadst; /* 0x00 */
154 u32 dmasize; /* 0x04 */
155 u32 dmasts; /* 0x08 */
156 u32 dmactrl; /* 0x0C */
157 u32 reserved0; /* 0x10 */
158 u32 dmaisr; /* 0x14 */
159 u32 dmaier; /* 0x18 */
160 u32 dmaidr; /* 0x1C */
161 u32 dmaimr; /* 0x20 */
162 u32 dmactrl2; /* 0x24 */
163 u32 dmadstmsb; /* 0x28 */
164};
165
Simon Glass8a8d24b2020-12-03 16:55:23 -0700166struct zynqmp_qspi_plat {
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530167 struct zynqmp_qspi_regs *regs;
168 struct zynqmp_qspi_dma_regs *dma_regs;
169 u32 frequency;
170 u32 speed_hz;
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600171 unsigned int io_mode;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530172};
173
174struct zynqmp_qspi_priv {
175 struct zynqmp_qspi_regs *regs;
176 struct zynqmp_qspi_dma_regs *dma_regs;
177 const void *tx_buf;
178 void *rx_buf;
179 unsigned int len;
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600180 unsigned int io_mode;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530181 int bytes_to_transfer;
182 int bytes_to_receive;
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600183 const struct spi_mem_op *op;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530184};
185
Algapally Santosh Sagarcc24fd72023-03-01 03:33:33 -0700186__weak int zynqmp_mmio_write(const u32 address, const u32 mask, const u32 value)
187{
188 return 0;
189}
190
Simon Glassd1998a92020-12-03 16:55:21 -0700191static int zynqmp_qspi_of_to_plat(struct udevice *bus)
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530192{
Simon Glass0fd3d912020-12-22 19:30:28 -0700193 struct zynqmp_qspi_plat *plat = dev_get_plat(bus);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530194
195 debug("%s\n", __func__);
196
Masahiro Yamada25484932020-07-17 14:36:48 +0900197 plat->regs = (struct zynqmp_qspi_regs *)(dev_read_addr(bus) +
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530198 GQSPI_REG_OFFSET);
199 plat->dma_regs = (struct zynqmp_qspi_dma_regs *)
Masahiro Yamada25484932020-07-17 14:36:48 +0900200 (dev_read_addr(bus) + GQSPI_DMA_REG_OFFSET);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530201
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600202 plat->io_mode = dev_read_bool(bus, "has-io-mode");
203
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530204 return 0;
205}
206
207static void zynqmp_qspi_init_hw(struct zynqmp_qspi_priv *priv)
208{
209 u32 config_reg;
210 struct zynqmp_qspi_regs *regs = priv->regs;
211
212 writel(GQSPI_GFIFO_SELECT, &regs->gqspisel);
213 writel(GQSPI_GFIFO_ALL_INT_MASK, &regs->idisr);
214 writel(GQSPI_FIFO_THRESHOLD, &regs->txftr);
215 writel(GQSPI_FIFO_THRESHOLD, &regs->rxftr);
Ashok Reddy Soma020b3532021-08-20 07:43:17 -0600216 writel(GQSPI_GENFIFO_THRESHOLD, &regs->gqfthr);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530217 writel(GQSPI_GFIFO_ALL_INT_MASK, &regs->isr);
Ashok Reddy Soma020b3532021-08-20 07:43:17 -0600218 writel(~GQSPI_ENABLE_ENABLE_MASK, &regs->enbr);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530219
220 config_reg = readl(&regs->confr);
221 config_reg &= ~(GQSPI_GFIFO_STRT_MODE_MASK |
222 GQSPI_CONFIG_MODE_EN_MASK);
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600223 config_reg |= GQSPI_GFIFO_WP_HOLD | GQSPI_DFLT_BAUD_RATE_DIV;
224 config_reg |= GQSPI_GFIFO_STRT_MODE_MASK;
225 if (!priv->io_mode)
226 config_reg |= GQSPI_CONFIG_DMA_MODE;
227
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530228 writel(config_reg, &regs->confr);
229
230 writel(GQSPI_ENABLE_ENABLE_MASK, &regs->enbr);
231}
232
233static u32 zynqmp_qspi_bus_select(struct zynqmp_qspi_priv *priv)
234{
235 u32 gqspi_fifo_reg = 0;
236
237 gqspi_fifo_reg = GQSPI_GFIFO_LOW_BUS |
238 GQSPI_GFIFO_CS_LOWER;
239
240 return gqspi_fifo_reg;
241}
242
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600243static u32 zynqmp_qspi_genfifo_mode(u8 buswidth)
244{
245 switch (buswidth) {
246 case 1:
247 return GQSPI_SPI_MODE_SPI;
248 case 2:
249 return GQSPI_SPI_MODE_DUAL_SPI;
250 case 4:
251 return GQSPI_SPI_MODE_QSPI;
252 default:
253 debug("Unsupported bus width %u\n", buswidth);
254 return GQSPI_SPI_MODE_SPI;
255 }
256}
257
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530258static void zynqmp_qspi_fill_gen_fifo(struct zynqmp_qspi_priv *priv,
259 u32 gqspi_fifo_reg)
260{
261 struct zynqmp_qspi_regs *regs = priv->regs;
Ashok Reddy Soma2ffa6532021-05-25 06:36:27 -0600262 u32 config_reg, ier;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530263 int ret = 0;
264
Ashok Reddy Soma72022a52021-08-20 07:43:16 -0600265 writel(gqspi_fifo_reg, &regs->genfifo);
266
Ashok Reddy Soma2ffa6532021-05-25 06:36:27 -0600267 config_reg = readl(&regs->confr);
268 /* Manual start if needed */
269 config_reg |= GQSPI_STRT_GEN_FIFO;
270 writel(config_reg, &regs->confr);
271
272 /* Enable interrupts */
273 ier = readl(&regs->ier);
Ashok Reddy Soma72022a52021-08-20 07:43:16 -0600274 ier |= GQSPI_IXR_GFEMTY_MASK;
Ashok Reddy Soma2ffa6532021-05-25 06:36:27 -0600275 writel(ier, &regs->ier);
276
Ashok Reddy Soma72022a52021-08-20 07:43:16 -0600277 /* Wait until the gen fifo is empty to write the new command */
278 ret = wait_for_bit_le32(&regs->isr, GQSPI_IXR_GFEMTY_MASK, 1,
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530279 GQSPI_TIMEOUT, 1);
280 if (ret)
281 printf("%s Timeout\n", __func__);
282
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530283}
284
285static void zynqmp_qspi_chipselect(struct zynqmp_qspi_priv *priv, int is_on)
286{
287 u32 gqspi_fifo_reg = 0;
288
289 if (is_on) {
290 gqspi_fifo_reg = zynqmp_qspi_bus_select(priv);
291 gqspi_fifo_reg |= GQSPI_SPI_MODE_SPI |
292 GQSPI_IMD_DATA_CS_ASSERT;
293 } else {
294 gqspi_fifo_reg = GQSPI_GFIFO_LOW_BUS;
295 gqspi_fifo_reg |= GQSPI_IMD_DATA_CS_DEASSERT;
296 }
297
298 debug("GFIFO_CMD_CS: 0x%x\n", gqspi_fifo_reg);
299
300 zynqmp_qspi_fill_gen_fifo(priv, gqspi_fifo_reg);
301}
302
Venkatesh Yadav Abbarapub4513302022-10-04 11:07:30 +0530303static void zynqmp_qspi_set_tapdelay(struct udevice *bus, u32 baudrateval)
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530304{
Simon Glass0fd3d912020-12-22 19:30:28 -0700305 struct zynqmp_qspi_plat *plat = dev_get_plat(bus);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530306 struct zynqmp_qspi_priv *priv = dev_get_priv(bus);
307 struct zynqmp_qspi_regs *regs = priv->regs;
308 u32 tapdlybypass = 0, lpbkdlyadj = 0, datadlyadj = 0, clk_rate;
309 u32 reqhz = 0;
310
311 clk_rate = plat->frequency;
312 reqhz = (clk_rate / (GQSPI_BAUD_DIV_SHIFT << baudrateval));
313
314 debug("%s, req_hz:%d, clk_rate:%d, baudrateval:%d\n",
315 __func__, reqhz, clk_rate, baudrateval);
316
Michal Simek23896472022-09-19 14:21:04 +0200317 if (!(IS_ENABLED(CONFIG_ARCH_VERSAL) ||
318 IS_ENABLED(CONFIG_ARCH_VERSAL_NET))) {
Ashok Reddy Somaf4f1b652022-08-25 06:59:01 -0600319 if (reqhz <= GQSPI_FREQ_40MHZ) {
320 tapdlybypass = TAP_DLY_BYPASS_LQSPI_RX_VALUE <<
321 TAP_DLY_BYPASS_LQSPI_RX_SHIFT;
322 } else if (reqhz <= GQSPI_FREQ_100MHZ) {
323 tapdlybypass = TAP_DLY_BYPASS_LQSPI_RX_VALUE <<
324 TAP_DLY_BYPASS_LQSPI_RX_SHIFT;
325 lpbkdlyadj = GQSPI_LPBK_DLY_ADJ_LPBK_MASK;
326 datadlyadj = (GQSPI_USE_DATA_DLY <<
327 GQSPI_USE_DATA_DLY_SHIFT) |
328 (GQSPI_DATA_DLY_ADJ_VALUE <<
329 GQSPI_DATA_DLY_ADJ_SHIFT);
330 } else if (reqhz <= GQSPI_FREQ_150MHZ) {
331 lpbkdlyadj = GQSPI_LPBK_DLY_ADJ_LPBK_MASK |
332 GQSPI_LPBK_DLY_ADJ_DLY_0;
333 }
334 zynqmp_mmio_write(IOU_TAPDLY_BYPASS_OFST,
335 IOU_TAPDLY_BYPASS_MASK, tapdlybypass);
336 } else {
337 if (reqhz <= GQSPI_FREQ_37_5MHZ) {
338 tapdlybypass = TAP_DLY_BYPASS_LQSPI_RX_VALUE <<
339 TAP_DLY_BYPASS_LQSPI_RX_SHIFT;
340 } else if (reqhz <= GQSPI_FREQ_100MHZ) {
341 tapdlybypass = TAP_DLY_BYPASS_LQSPI_RX_VALUE <<
342 TAP_DLY_BYPASS_LQSPI_RX_SHIFT;
343 lpbkdlyadj = GQSPI_LPBK_DLY_ADJ_LPBK_MASK;
344 datadlyadj = GQSPI_USE_DATA_DLY <<
345 GQSPI_USE_DATA_DLY_SHIFT;
346 } else if (reqhz <= GQSPI_FREQ_150MHZ) {
347 lpbkdlyadj = GQSPI_LPBK_DLY_ADJ_LPBK_MASK |
348 (GQSPI_LPBK_DLY_ADJ_DLY_1 <<
349 GQSPI_LPBK_DLY_ADJ_DLY_1_SHIFT);
350 }
351 writel(tapdlybypass, IOU_TAPDLY_BYPASS_OFST);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530352 }
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530353 writel(lpbkdlyadj, &regs->lpbkdly);
354 writel(datadlyadj, &regs->gqspidlyadj);
355}
356
357static int zynqmp_qspi_set_speed(struct udevice *bus, uint speed)
358{
Simon Glass0fd3d912020-12-22 19:30:28 -0700359 struct zynqmp_qspi_plat *plat = dev_get_plat(bus);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530360 struct zynqmp_qspi_priv *priv = dev_get_priv(bus);
361 struct zynqmp_qspi_regs *regs = priv->regs;
362 u32 confr;
363 u8 baud_rate_val = 0;
364
365 debug("%s\n", __func__);
366 if (speed > plat->frequency)
367 speed = plat->frequency;
368
Brandon Maierd9aa19e2021-01-20 14:28:30 -0600369 if (plat->speed_hz != speed) {
370 /* Set the clock frequency */
371 /* If speed == 0, default to lowest speed */
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530372 while ((baud_rate_val < 8) &&
373 ((plat->frequency /
374 (2 << baud_rate_val)) > speed))
375 baud_rate_val++;
376
377 if (baud_rate_val > GQSPI_MAX_BAUD_RATE_VAL)
378 baud_rate_val = GQSPI_DFLT_BAUD_RATE_VAL;
379
380 plat->speed_hz = plat->frequency / (2 << baud_rate_val);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530381
Brandon Maierd9aa19e2021-01-20 14:28:30 -0600382 confr = readl(&regs->confr);
383 confr &= ~GQSPI_BAUD_DIV_MASK;
384 confr |= (baud_rate_val << 3);
385 writel(confr, &regs->confr);
386 zynqmp_qspi_set_tapdelay(bus, baud_rate_val);
387
388 debug("regs=%p, speed=%d\n", priv->regs, plat->speed_hz);
389 }
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530390
391 return 0;
392}
393
394static int zynqmp_qspi_probe(struct udevice *bus)
395{
Simon Glass8a8d24b2020-12-03 16:55:23 -0700396 struct zynqmp_qspi_plat *plat = dev_get_plat(bus);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530397 struct zynqmp_qspi_priv *priv = dev_get_priv(bus);
398 struct clk clk;
399 unsigned long clock;
400 int ret;
401
402 debug("%s: bus:%p, priv:%p\n", __func__, bus, priv);
403
404 priv->regs = plat->regs;
405 priv->dma_regs = plat->dma_regs;
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600406 priv->io_mode = plat->io_mode;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530407
408 ret = clk_get_by_index(bus, 0, &clk);
409 if (ret < 0) {
Sean Anderson49dfbe92020-09-15 10:45:12 -0400410 dev_err(bus, "failed to get clock\n");
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530411 return ret;
412 }
413
414 clock = clk_get_rate(&clk);
415 if (IS_ERR_VALUE(clock)) {
Sean Anderson49dfbe92020-09-15 10:45:12 -0400416 dev_err(bus, "failed to get rate\n");
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530417 return clock;
418 }
419 debug("%s: CLK %ld\n", __func__, clock);
420
421 ret = clk_enable(&clk);
Michal Simek9b7aac72021-02-09 15:28:15 +0100422 if (ret) {
Sean Anderson49dfbe92020-09-15 10:45:12 -0400423 dev_err(bus, "failed to enable clock\n");
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530424 return ret;
425 }
426 plat->frequency = clock;
427 plat->speed_hz = plat->frequency / 2;
428
429 /* init the zynq spi hw */
430 zynqmp_qspi_init_hw(priv);
431
432 return 0;
433}
434
435static int zynqmp_qspi_set_mode(struct udevice *bus, uint mode)
436{
437 struct zynqmp_qspi_priv *priv = dev_get_priv(bus);
438 struct zynqmp_qspi_regs *regs = priv->regs;
439 u32 confr;
440
441 debug("%s\n", __func__);
442 /* Set the SPI Clock phase and polarities */
443 confr = readl(&regs->confr);
Ashok Reddy Somaafe03862022-08-25 06:59:05 -0600444 confr &= ~(GQSPI_CONFIG_CPHA_MASK | GQSPI_CONFIG_CPOL_MASK);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530445
446 if (mode & SPI_CPHA)
447 confr |= GQSPI_CONFIG_CPHA_MASK;
448 if (mode & SPI_CPOL)
449 confr |= GQSPI_CONFIG_CPOL_MASK;
450
451 writel(confr, &regs->confr);
452
453 return 0;
454}
455
456static int zynqmp_qspi_fill_tx_fifo(struct zynqmp_qspi_priv *priv, u32 size)
457{
458 u32 data;
459 int ret = 0;
460 struct zynqmp_qspi_regs *regs = priv->regs;
461 u32 *buf = (u32 *)priv->tx_buf;
462 u32 len = size;
463
464 debug("TxFIFO: 0x%x, size: 0x%x\n", readl(&regs->isr),
465 size);
466
467 while (size) {
468 ret = wait_for_bit_le32(&regs->isr, GQSPI_IXR_TXNFULL_MASK, 1,
469 GQSPI_TIMEOUT, 1);
470 if (ret) {
471 printf("%s: Timeout\n", __func__);
472 return ret;
473 }
474
475 if (size >= 4) {
476 writel(*buf, &regs->txd0r);
477 buf++;
478 size -= 4;
479 } else {
480 switch (size) {
481 case 1:
482 data = *((u8 *)buf);
483 buf += 1;
484 data |= GENMASK(31, 8);
485 break;
486 case 2:
487 data = *((u16 *)buf);
488 buf += 2;
489 data |= GENMASK(31, 16);
490 break;
491 case 3:
T Karthik Reddy90217482020-11-19 05:00:36 -0700492 data = *buf;
493 buf += 3;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530494 data |= GENMASK(31, 24);
495 break;
496 }
497 writel(data, &regs->txd0r);
498 size = 0;
499 }
500 }
501
Ashok Reddy Soma4f9d2552021-10-19 19:43:00 +0530502 ret = wait_for_bit_le32(&regs->isr, GQSPI_IXR_TXFIFOEMPTY_MASK, 1,
503 GQSPI_TIMEOUT, 1);
504 if (ret) {
505 printf("%s: Timeout\n", __func__);
506 return ret;
507 }
508
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530509 priv->tx_buf += len;
510 return 0;
511}
512
513static void zynqmp_qspi_genfifo_cmd(struct zynqmp_qspi_priv *priv)
514{
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600515 const struct spi_mem_op *op = priv->op;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530516 u32 gen_fifo_cmd;
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600517 u8 i, dummy_cycles, addr;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530518
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600519 /* Send opcode */
520 gen_fifo_cmd = zynqmp_qspi_bus_select(priv);
521 gen_fifo_cmd |= zynqmp_qspi_genfifo_mode(op->cmd.buswidth);
522 gen_fifo_cmd |= GQSPI_GFIFO_TX;
523 gen_fifo_cmd |= op->cmd.opcode;
524 zynqmp_qspi_fill_gen_fifo(priv, gen_fifo_cmd);
525
526 /* Send address */
527 for (i = 0; i < op->addr.nbytes; i++) {
528 addr = op->addr.val >> (8 * (op->addr.nbytes - i - 1));
529
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530530 gen_fifo_cmd = zynqmp_qspi_bus_select(priv);
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600531 gen_fifo_cmd |= zynqmp_qspi_genfifo_mode(op->addr.buswidth);
532 gen_fifo_cmd |= GQSPI_GFIFO_TX;
533 gen_fifo_cmd |= addr;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530534
535 debug("GFIFO_CMD_Cmd = 0x%x\n", gen_fifo_cmd);
536
537 zynqmp_qspi_fill_gen_fifo(priv, gen_fifo_cmd);
538 }
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600539
540 /* Send dummy */
541 if (op->dummy.nbytes) {
542 dummy_cycles = op->dummy.nbytes * 8 / op->dummy.buswidth;
543
544 gen_fifo_cmd = zynqmp_qspi_bus_select(priv);
545 gen_fifo_cmd |= zynqmp_qspi_genfifo_mode(op->dummy.buswidth);
546 gen_fifo_cmd &= ~(GQSPI_GFIFO_TX | GQSPI_GFIFO_RX);
547 gen_fifo_cmd |= GQSPI_GFIFO_DATA_XFR_MASK;
548 gen_fifo_cmd |= dummy_cycles;
549 zynqmp_qspi_fill_gen_fifo(priv, gen_fifo_cmd);
550 }
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530551}
552
553static u32 zynqmp_qspi_calc_exp(struct zynqmp_qspi_priv *priv,
554 u32 *gen_fifo_cmd)
555{
556 u32 expval = 8;
557 u32 len;
558
559 while (1) {
560 if (priv->len > 255) {
561 if (priv->len & (1 << expval)) {
562 *gen_fifo_cmd &= ~GQSPI_GFIFO_IMD_MASK;
563 *gen_fifo_cmd |= GQSPI_GFIFO_EXP_MASK;
564 *gen_fifo_cmd |= expval;
565 priv->len -= (1 << expval);
566 return expval;
567 }
568 expval++;
569 } else {
570 *gen_fifo_cmd &= ~(GQSPI_GFIFO_IMD_MASK |
571 GQSPI_GFIFO_EXP_MASK);
572 *gen_fifo_cmd |= (u8)priv->len;
573 len = (u8)priv->len;
574 priv->len = 0;
575 return len;
576 }
577 }
578}
579
580static int zynqmp_qspi_genfifo_fill_tx(struct zynqmp_qspi_priv *priv)
581{
582 u32 gen_fifo_cmd;
583 u32 len;
584 int ret = 0;
585
586 gen_fifo_cmd = zynqmp_qspi_bus_select(priv);
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600587 gen_fifo_cmd |= zynqmp_qspi_genfifo_mode(priv->op->data.buswidth);
Ashok Reddy Somaafe03862022-08-25 06:59:05 -0600588 gen_fifo_cmd |= GQSPI_GFIFO_TX | GQSPI_GFIFO_DATA_XFR_MASK;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530589
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530590 while (priv->len) {
591 len = zynqmp_qspi_calc_exp(priv, &gen_fifo_cmd);
592 zynqmp_qspi_fill_gen_fifo(priv, gen_fifo_cmd);
593
594 debug("GFIFO_CMD_TX:0x%x\n", gen_fifo_cmd);
595
596 if (gen_fifo_cmd & GQSPI_GFIFO_EXP_MASK)
Ashok Reddy Somaafe03862022-08-25 06:59:05 -0600597 ret = zynqmp_qspi_fill_tx_fifo(priv, 1 << len);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530598 else
Ashok Reddy Somaafe03862022-08-25 06:59:05 -0600599 ret = zynqmp_qspi_fill_tx_fifo(priv, len);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530600
601 if (ret)
602 return ret;
603 }
604 return ret;
605}
606
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600607static int zynqmp_qspi_start_io(struct zynqmp_qspi_priv *priv,
608 u32 gen_fifo_cmd, u32 *buf)
609{
610 u32 len;
611 u32 actuallen = priv->len;
612 u32 config_reg, ier, isr;
613 u32 timeout = GQSPI_TIMEOUT;
614 struct zynqmp_qspi_regs *regs = priv->regs;
615 u32 last_bits;
616 u32 *traverse = buf;
617
618 while (priv->len) {
619 len = zynqmp_qspi_calc_exp(priv, &gen_fifo_cmd);
620 /* If exponent bit is set, reset immediate to be 2^len */
621 if (gen_fifo_cmd & GQSPI_GFIFO_EXP_MASK)
622 priv->bytes_to_receive = (1 << len);
623 else
624 priv->bytes_to_receive = len;
625 zynqmp_qspi_fill_gen_fifo(priv, gen_fifo_cmd);
626 debug("GFIFO_CMD_RX:0x%x\n", gen_fifo_cmd);
627 /* Manual start */
628 config_reg = readl(&regs->confr);
629 config_reg |= GQSPI_STRT_GEN_FIFO;
630 writel(config_reg, &regs->confr);
631 /* Enable RX interrupts for IO mode */
632 ier = readl(&regs->ier);
633 ier |= GQSPI_IXR_ALL_MASK;
634 writel(ier, &regs->ier);
635 while (priv->bytes_to_receive && timeout) {
636 isr = readl(&regs->isr);
637 if (isr & GQSPI_IXR_RXNEMTY_MASK) {
638 if (priv->bytes_to_receive >= 4) {
639 *traverse = readl(&regs->drxr);
640 traverse++;
641 priv->bytes_to_receive -= 4;
642 } else {
643 last_bits = readl(&regs->drxr);
644 memcpy(traverse, &last_bits,
645 priv->bytes_to_receive);
646 priv->bytes_to_receive = 0;
647 }
648 timeout = GQSPI_TIMEOUT;
649 } else {
650 udelay(1);
651 timeout--;
652 }
653 }
654
655 debug("buf:0x%lx, rxbuf:0x%lx, *buf:0x%x len: 0x%x\n",
656 (unsigned long)buf, (unsigned long)priv->rx_buf,
657 *buf, actuallen);
658 if (!timeout) {
659 printf("IO timeout: %d\n", readl(&regs->isr));
660 return -1;
661 }
662 }
663
664 return 0;
665}
666
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530667static int zynqmp_qspi_start_dma(struct zynqmp_qspi_priv *priv,
668 u32 gen_fifo_cmd, u32 *buf)
669{
Venkatesh Yadav Abbarapu906e20a2022-11-25 16:14:13 +0530670 unsigned long addr;
Ashok Reddy Soma020b3532021-08-20 07:43:17 -0600671 u32 size;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530672 u32 actuallen = priv->len;
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -0600673 u32 totallen = priv->len;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530674 int ret = 0;
675 struct zynqmp_qspi_dma_regs *dma_regs = priv->dma_regs;
676
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -0600677 while (totallen) {
678 if (totallen >= SZ_512M)
679 priv->len = SZ_256M;
680 else
681 priv->len = totallen;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530682
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -0600683 totallen -= priv->len; /* Save remaining bytes length to read */
684 actuallen = priv->len; /* Actual number of bytes reading */
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530685
Venkatesh Yadav Abbarapu906e20a2022-11-25 16:14:13 +0530686 writel(lower_32_bits((unsigned long)buf), &dma_regs->dmadst);
687 writel(upper_32_bits((unsigned long)buf) & GENMASK(11, 0),
688 &dma_regs->dmadstmsb);
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -0600689 writel(roundup(priv->len, GQSPI_DMA_ALIGN), &dma_regs->dmasize);
690 writel(GQSPI_DMA_DST_I_STS_MASK, &dma_regs->dmaier);
691 addr = (unsigned long)buf;
692 size = roundup(priv->len, GQSPI_DMA_ALIGN);
Ashok Reddy Soma638189d2023-09-15 08:47:58 +0530693 invalidate_dcache_range(addr, addr + size);
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -0600694
695 while (priv->len) {
696 zynqmp_qspi_calc_exp(priv, &gen_fifo_cmd);
697 zynqmp_qspi_fill_gen_fifo(priv, gen_fifo_cmd);
698
699 debug("GFIFO_CMD_RX:0x%x\n", gen_fifo_cmd);
700 }
701
702 ret = wait_for_bit_le32(&dma_regs->dmaisr,
703 GQSPI_DMA_DST_I_STS_DONE, 1,
704 GQSPI_TIMEOUT, 1);
705 if (ret) {
706 printf("DMA Timeout:0x%x\n", readl(&dma_regs->dmaisr));
707 return -ETIMEDOUT;
708 }
709
Venkatesh Yadav Abbarapua3ade3d2023-09-15 08:47:59 +0530710 invalidate_dcache_range(addr, addr + size);
711
Ashok Reddy Somaa3d4bfb2022-08-25 06:59:04 -0600712 writel(GQSPI_DMA_DST_I_STS_DONE, &dma_regs->dmaisr);
713
714 debug("buf:0x%lx, rxbuf:0x%lx, *buf:0x%x len: 0x%x\n",
715 (unsigned long)buf, (unsigned long)priv->rx_buf, *buf,
716 actuallen);
717
718 if (buf != priv->rx_buf)
719 memcpy(priv->rx_buf, buf, actuallen);
720
721 buf = (u32 *)((u8 *)buf + actuallen);
722 priv->rx_buf = (u8 *)priv->rx_buf + actuallen;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530723 }
724
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530725 return 0;
726}
727
728static int zynqmp_qspi_genfifo_fill_rx(struct zynqmp_qspi_priv *priv)
729{
730 u32 gen_fifo_cmd;
731 u32 *buf;
732 u32 actuallen = priv->len;
733
734 gen_fifo_cmd = zynqmp_qspi_bus_select(priv);
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600735 gen_fifo_cmd |= zynqmp_qspi_genfifo_mode(priv->op->data.buswidth);
Ashok Reddy Somaafe03862022-08-25 06:59:05 -0600736 gen_fifo_cmd |= GQSPI_GFIFO_RX | GQSPI_GFIFO_DATA_XFR_MASK;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530737
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530738 /*
739 * Check if receive buffer is aligned to 4 byte and length
740 * is multiples of four byte as we are using dma to receive.
741 */
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600742 if ((!((unsigned long)priv->rx_buf & (GQSPI_DMA_ALIGN - 1)) &&
743 !(actuallen % GQSPI_DMA_ALIGN)) || priv->io_mode) {
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530744 buf = (u32 *)priv->rx_buf;
Ashok Reddy Somad91b0f42022-08-25 06:59:03 -0600745 if (priv->io_mode)
746 return zynqmp_qspi_start_io(priv, gen_fifo_cmd, buf);
747 else
748 return zynqmp_qspi_start_dma(priv, gen_fifo_cmd, buf);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530749 }
750
Ashok Reddy Somaafe03862022-08-25 06:59:05 -0600751 ALLOC_CACHE_ALIGN_BUFFER(u8, tmp, roundup(priv->len, GQSPI_DMA_ALIGN));
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530752 buf = (u32 *)tmp;
753 return zynqmp_qspi_start_dma(priv, gen_fifo_cmd, buf);
754}
755
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530756static int zynqmp_qspi_claim_bus(struct udevice *dev)
757{
758 struct udevice *bus = dev->parent;
759 struct zynqmp_qspi_priv *priv = dev_get_priv(bus);
760 struct zynqmp_qspi_regs *regs = priv->regs;
761
762 writel(GQSPI_ENABLE_ENABLE_MASK, &regs->enbr);
763
764 return 0;
765}
766
767static int zynqmp_qspi_release_bus(struct udevice *dev)
768{
769 struct udevice *bus = dev->parent;
770 struct zynqmp_qspi_priv *priv = dev_get_priv(bus);
771 struct zynqmp_qspi_regs *regs = priv->regs;
772
773 writel(~GQSPI_ENABLE_ENABLE_MASK, &regs->enbr);
774
775 return 0;
776}
777
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600778static int zynqmp_qspi_exec_op(struct spi_slave *slave,
779 const struct spi_mem_op *op)
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530780{
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600781 struct zynqmp_qspi_priv *priv = dev_get_priv(slave->dev->parent);
782 int ret = 0;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530783
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600784 priv->op = op;
785 priv->tx_buf = op->data.buf.out;
786 priv->rx_buf = op->data.buf.in;
787 priv->len = op->data.nbytes;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530788
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600789 zynqmp_qspi_chipselect(priv, 1);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530790
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600791 /* Send opcode, addr, dummy */
792 zynqmp_qspi_genfifo_cmd(priv);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530793
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600794 /* Request the transfer */
795 if (op->data.dir == SPI_MEM_DATA_IN)
796 ret = zynqmp_qspi_genfifo_fill_rx(priv);
797 else if (op->data.dir == SPI_MEM_DATA_OUT)
798 ret = zynqmp_qspi_genfifo_fill_tx(priv);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530799
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600800 zynqmp_qspi_chipselect(priv, 0);
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530801
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600802 return ret;
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530803}
804
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600805static const struct spi_controller_mem_ops zynqmp_qspi_mem_ops = {
806 .exec_op = zynqmp_qspi_exec_op,
807};
808
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530809static const struct dm_spi_ops zynqmp_qspi_ops = {
810 .claim_bus = zynqmp_qspi_claim_bus,
811 .release_bus = zynqmp_qspi_release_bus,
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530812 .set_speed = zynqmp_qspi_set_speed,
813 .set_mode = zynqmp_qspi_set_mode,
Brandon Maierf1fd79a2021-01-20 10:39:46 -0600814 .mem_ops = &zynqmp_qspi_mem_ops,
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530815};
816
817static const struct udevice_id zynqmp_qspi_ids[] = {
818 { .compatible = "xlnx,zynqmp-qspi-1.0" },
Michal Simekf3976cc2018-11-29 08:48:28 +0100819 { .compatible = "xlnx,versal-qspi-1.0" },
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530820 { }
821};
822
823U_BOOT_DRIVER(zynqmp_qspi) = {
824 .name = "zynqmp_qspi",
825 .id = UCLASS_SPI,
826 .of_match = zynqmp_qspi_ids,
827 .ops = &zynqmp_qspi_ops,
Simon Glassd1998a92020-12-03 16:55:21 -0700828 .of_to_plat = zynqmp_qspi_of_to_plat,
Simon Glass8a8d24b2020-12-03 16:55:23 -0700829 .plat_auto = sizeof(struct zynqmp_qspi_plat),
Simon Glass41575d82020-12-03 16:55:17 -0700830 .priv_auto = sizeof(struct zynqmp_qspi_priv),
Siva Durga Prasad Paladugu22cca172018-07-04 17:31:23 +0530831 .probe = zynqmp_qspi_probe,
832};