1 // SPDX-License-Identifier: GPL-2.0+
3 * Copyright (C) 2015-2017 Socionext Inc.
4 * Author: Masahiro Yamada <yamada.masahiro@socionext.com>
6 * based on commit 21b6e480f92ccc38fe0502e3116411d6509d3bf2 of Diag by:
7 * Copyright (C) 2015 Socionext Inc.
11 #include <linux/delay.h>
12 #include <linux/errno.h>
14 #include <linux/printk.h>
15 #include <linux/sizes.h>
16 #include <asm/processor.h>
20 #include "../soc-info.h"
21 #include "ddrmphy-regs.h"
39 static u32 ddrphy_pgcr2[DRAM_FREQ_NR] = {0x00FC7E5D, 0x00FC90AB};
40 static u32 ddrphy_ptr0[DRAM_FREQ_NR] = {0x0EA09205, 0x10C0A6C6};
41 static u32 ddrphy_ptr1[DRAM_FREQ_NR] = {0x0DAC041B, 0x0FA104B1};
42 static u32 ddrphy_ptr3[DRAM_FREQ_NR] = {0x15171e45, 0x18182357};
43 static u32 ddrphy_ptr4[DRAM_FREQ_NR] = {0x0e9ad8e9, 0x10b34157};
44 static u32 ddrphy_dtpr0[DRAM_FREQ_NR] = {0x35a00d88, 0x39e40e88};
45 static u32 ddrphy_dtpr1[DRAM_FREQ_NR] = {0x2288cc2c, 0x228a04d0};
46 static u32 ddrphy_dtpr2[DRAM_FREQ_NR] = {0x50005e00, 0x50006a00};
47 static u32 ddrphy_dtpr3[DRAM_FREQ_NR] = {0x0010cb49, 0x0010ec89};
48 static u32 ddrphy_mr0[DRAM_FREQ_NR] = {0x00000115, 0x00000125};
49 static u32 ddrphy_mr2[DRAM_FREQ_NR] = {0x000002a0, 0x000002a8};
51 /* dependent on package and board design */
52 static u32 ddrphy_acbdlr0[DRAM_CH_NR] = {0x0000000c, 0x0000000c, 0x00000009};
55 static inline int ddrphy_get_rank(int dx)
60 static void ddrphy_fifo_reset(void __iomem *phy_base)
64 tmp = readl(phy_base + MPHY_PGCR0);
65 tmp &= ~MPHY_PGCR0_PHYFRST;
66 writel(tmp, phy_base + MPHY_PGCR0);
70 tmp |= MPHY_PGCR0_PHYFRST;
71 writel(tmp, phy_base + MPHY_PGCR0);
76 static void ddrphy_vt_ctrl(void __iomem *phy_base, int enable)
80 tmp = readl(phy_base + MPHY_PGCR1);
83 tmp &= ~MPHY_PGCR1_INHVT;
85 tmp |= MPHY_PGCR1_INHVT;
87 writel(tmp, phy_base + MPHY_PGCR1);
90 while (!(readl(phy_base + MPHY_PGSR1) & MPHY_PGSR1_VTSTOP))
95 static void ddrphy_dqs_delay_fixup(void __iomem *phy_base, int nr_dx, int step)
99 void __iomem *dx_base = phy_base + MPHY_DX_BASE;
101 ddrphy_vt_ctrl(phy_base, 0);
103 for (dx = 0; dx < nr_dx; dx++) {
104 lcdlr1 = readl(dx_base + MPHY_DX_LCDLR1);
105 rdqsd = (lcdlr1 >> 8) & 0xff;
106 rdqsd = clamp(rdqsd + step, 0U, 0xffU);
107 lcdlr1 = (lcdlr1 & ~(0xff << 8)) | (rdqsd << 8);
108 writel(lcdlr1, dx_base + MPHY_DX_LCDLR1);
109 readl(dx_base + MPHY_DX_LCDLR1); /* relax */
110 dx_base += MPHY_DX_STRIDE;
113 ddrphy_vt_ctrl(phy_base, 1);
116 static int ddrphy_get_system_latency(void __iomem *phy_base, int width)
118 void __iomem *dx_base = phy_base + MPHY_DX_BASE;
119 const int nr_dx = width / 8;
122 int dgsl, dgsl_min = INT_MAX, dgsl_max = 0;
124 for (dx = 0; dx < nr_dx; dx++) {
125 gtr = readl(dx_base + MPHY_DX_GTR);
126 for (rank = 0; rank < 4; rank++) {
128 /* if dgsl is zero, this rank was not trained. skip. */
130 dgsl_min = min(dgsl_min, dgsl);
131 dgsl_max = max(dgsl_max, dgsl);
135 dx_base += MPHY_DX_STRIDE;
138 if (dgsl_min != dgsl_max)
139 pr_warn("DQS Gateing System Latencies are not all leveled.\n");
144 static void ddrphy_init(void __iomem *phy_base, enum dram_freq freq, int width,
148 void __iomem *zq_base, *dx_base;
154 writel(MPHY_PIR_ZCALBYP, phy_base + MPHY_PIR);
156 * Disable RGLVT bit (Read DQS Gating LCDL Delay VT Compensation)
157 * to avoid read error issue.
159 writel(0x07d81e37, phy_base + MPHY_PGCR0);
160 writel(0x0200c4e0, phy_base + MPHY_PGCR1);
162 tmp = ddrphy_pgcr2[freq];
164 tmp |= MPHY_PGCR2_DUALCHN | MPHY_PGCR2_ACPDDC;
165 writel(tmp, phy_base + MPHY_PGCR2);
167 writel(ddrphy_ptr0[freq], phy_base + MPHY_PTR0);
168 writel(ddrphy_ptr1[freq], phy_base + MPHY_PTR1);
169 writel(0x00083def, phy_base + MPHY_PTR2);
170 writel(ddrphy_ptr3[freq], phy_base + MPHY_PTR3);
171 writel(ddrphy_ptr4[freq], phy_base + MPHY_PTR4);
173 writel(ddrphy_acbdlr0[ch], phy_base + MPHY_ACBDLR0);
175 writel(0x55555555, phy_base + MPHY_ACIOCR1);
176 writel(0x00000000, phy_base + MPHY_ACIOCR2);
177 writel(0x55555555, phy_base + MPHY_ACIOCR3);
178 writel(0x00000000, phy_base + MPHY_ACIOCR4);
179 writel(0x00000055, phy_base + MPHY_ACIOCR5);
180 writel(0x00181aa4, phy_base + MPHY_DXCCR);
182 writel(0x0024641e, phy_base + MPHY_DSGCR);
183 writel(0x0000040b, phy_base + MPHY_DCR);
184 writel(ddrphy_dtpr0[freq], phy_base + MPHY_DTPR0);
185 writel(ddrphy_dtpr1[freq], phy_base + MPHY_DTPR1);
186 writel(ddrphy_dtpr2[freq], phy_base + MPHY_DTPR2);
187 writel(ddrphy_dtpr3[freq], phy_base + MPHY_DTPR3);
188 writel(ddrphy_mr0[freq], phy_base + MPHY_MR0);
189 writel(0x00000006, phy_base + MPHY_MR1);
190 writel(ddrphy_mr2[freq], phy_base + MPHY_MR2);
191 writel(0x00000000, phy_base + MPHY_MR3);
194 for (dx = 0; dx < nr_dx; dx++)
195 tmp |= BIT(MPHY_DTCR_RANKEN_SHIFT + ddrphy_get_rank(dx));
196 writel(0x90003087 | tmp, phy_base + MPHY_DTCR);
198 writel(0x00000000, phy_base + MPHY_DTAR0);
199 writel(0x00000008, phy_base + MPHY_DTAR1);
200 writel(0x00000010, phy_base + MPHY_DTAR2);
201 writel(0x00000018, phy_base + MPHY_DTAR3);
202 writel(0xdd22ee11, phy_base + MPHY_DTDR0);
203 writel(0x7788bb44, phy_base + MPHY_DTDR1);
205 /* impedance control settings */
206 writel(0x04048900, phy_base + MPHY_ZQCR);
208 zq_base = phy_base + MPHY_ZQ_BASE;
209 for (zq = 0; zq < 4; zq++) {
212 * PXS2: CH0ZQ0=0x5B, CH1ZQ0=0x5B, CH2ZQ0=0x59, others=0x5D
214 writel(0x0007BB5D, zq_base + MPHY_ZQ_PR);
215 zq_base += MPHY_ZQ_STRIDE;
219 dx_base = phy_base + MPHY_DX_BASE;
220 for (dx = 0; dx < 4; dx++) {
221 tmp = readl(dx_base + MPHY_DX_GCR0);
222 tmp &= ~MPHY_DX_GCR0_WLRKEN_MASK;
223 tmp |= BIT(MPHY_DX_GCR0_WLRKEN_SHIFT + ddrphy_get_rank(dx)) &
224 MPHY_DX_GCR0_WLRKEN_MASK;
225 writel(tmp, dx_base + MPHY_DX_GCR0);
227 writel(0x00000000, dx_base + MPHY_DX_GCR1);
228 writel(0x00000000, dx_base + MPHY_DX_GCR2);
229 writel(0x00000000, dx_base + MPHY_DX_GCR3);
230 dx_base += MPHY_DX_STRIDE;
233 while (!(readl(phy_base + MPHY_PGSR0) & MPHY_PGSR0_IDONE))
236 ddrphy_dqs_delay_fixup(phy_base, nr_dx, -4);
239 struct ddrphy_init_sequence {
246 static const struct ddrphy_init_sequence impedance_calibration_sequence[] = {
248 "Impedance Calibration",
256 static const struct ddrphy_init_sequence dram_init_sequence[] = {
258 "DRAM Initialization",
259 MPHY_PIR_DRAMRST | MPHY_PIR_DRAMINIT,
266 static const struct ddrphy_init_sequence training_sequence[] = {
274 "Read DQS Gate Training",
280 "Write Leveling Adjustment",
304 "Write Eye Training",
312 static int __ddrphy_training(void __iomem *phy_base,
313 const struct ddrphy_init_sequence *seq)
315 const struct ddrphy_init_sequence *s;
317 u32 init_flag = MPHY_PIR_INIT;
318 u32 done_flag = MPHY_PGSR0_IDONE;
319 int timeout = 50000; /* 50 msec is long enough */
320 unsigned long start = 0;
323 start = get_timer(0);
326 for (s = seq; s->description; s++) {
327 init_flag |= s->init_flag;
328 done_flag |= s->done_flag;
331 writel(init_flag, phy_base + MPHY_PIR);
335 pr_err("%s: error: timeout during DDR training\n",
340 pgsr0 = readl(phy_base + MPHY_PGSR0);
341 } while ((pgsr0 & done_flag) != done_flag);
343 for (s = seq; s->description; s++) {
344 if (pgsr0 & s->err_flag) {
345 pr_err("%s: error: %s failed\n", __func__,
351 pr_debug("DDRPHY training: elapsed time %ld msec\n", get_timer(start));
356 static int ddrphy_impedance_calibration(void __iomem *phy_base)
361 ret = __ddrphy_training(phy_base, impedance_calibration_sequence);
366 * Because of a hardware bug, IDONE flag is set when the first ZQ block
367 * is calibrated. The flag does not guarantee the completion for all
368 * the ZQ blocks. Wait a little more just in case.
372 /* reflect ZQ settings and enable average algorithm*/
373 tmp = readl(phy_base + MPHY_ZQCR);
374 tmp |= MPHY_ZQCR_FORCE_ZCAL_VT_UPDATE;
375 writel(tmp, phy_base + MPHY_ZQCR);
376 tmp &= ~MPHY_ZQCR_FORCE_ZCAL_VT_UPDATE;
377 tmp |= MPHY_ZQCR_AVGEN;
378 writel(tmp, phy_base + MPHY_ZQCR);
383 static int ddrphy_dram_init(void __iomem *phy_base)
385 return __ddrphy_training(phy_base, dram_init_sequence);
388 static int ddrphy_training(void __iomem *phy_base)
390 return __ddrphy_training(phy_base, training_sequence);
394 static u32 umc_cmdctla[DRAM_FREQ_NR] = {0x66DD131D, 0x77EE1722};
396 * The ch2 is a different generation UMC core.
397 * The register spec is different, unfortunately.
399 static u32 umc_cmdctlb_ch01[DRAM_FREQ_NR] = {0x13E87C44, 0x18F88C44};
400 static u32 umc_cmdctlb_ch2[DRAM_FREQ_NR] = {0x19E8DC44, 0x1EF8EC44};
401 static u32 umc_spcctla[DRAM_FREQ_NR][DRAM_SZ_NR] = {
402 {0x004A071D, 0x0078071D},
403 {0x0055081E, 0x0089081E},
406 static u32 umc_spcctlb[] = {0x00FF000A, 0x00FF000B};
407 /* The ch2 is different for some reason only hardware guys know... */
408 static u32 umc_flowctla_ch01[] = {0x0800001E, 0x08000022};
409 static u32 umc_flowctla_ch2[] = {0x0800001E, 0x0800001E};
411 static void umc_set_system_latency(void __iomem *dc_base, int phy_latency)
416 val = readl(dc_base + UMC_RDATACTL_D0);
417 latency = (val & UMC_RDATACTL_RADLTY_MASK) >> UMC_RDATACTL_RADLTY_SHIFT;
418 latency += (val & UMC_RDATACTL_RAD2LTY_MASK) >>
419 UMC_RDATACTL_RAD2LTY_SHIFT;
421 * UMC works at the half clock rate of the PHY.
422 * The LSB of latency is ignored
424 latency += phy_latency & ~1;
426 val &= ~(UMC_RDATACTL_RADLTY_MASK | UMC_RDATACTL_RAD2LTY_MASK);
428 val |= 0xf << UMC_RDATACTL_RADLTY_SHIFT;
429 val |= (latency - 0xf) << UMC_RDATACTL_RAD2LTY_SHIFT;
431 val |= latency << UMC_RDATACTL_RADLTY_SHIFT;
434 writel(val, dc_base + UMC_RDATACTL_D0);
435 writel(val, dc_base + UMC_RDATACTL_D1);
437 readl(dc_base + UMC_RDATACTL_D1); /* relax */
440 /* enable/disable auto refresh */
441 static void umc_refresh_ctrl(void __iomem *dc_base, int enable)
445 tmp = readl(dc_base + UMC_SPCSETB);
446 tmp &= ~UMC_SPCSETB_AREFMD_MASK;
449 tmp |= UMC_SPCSETB_AREFMD_ARB;
451 tmp |= UMC_SPCSETB_AREFMD_REG;
453 writel(tmp, dc_base + UMC_SPCSETB);
457 static void umc_ud_init(void __iomem *umc_base, int ch)
459 writel(0x00000003, umc_base + UMC_BITPERPIXELMODE_D0);
462 writel(0x00000033, umc_base + UMC_PAIR1DOFF_D0);
465 static int umc_dc_init(void __iomem *dc_base, enum dram_freq freq,
466 unsigned long size, int width, int ch)
468 enum dram_size size_e;
476 size_e = DRAM_SZ_256M;
479 size_e = DRAM_SZ_512M;
482 pr_err("unsupported DRAM size 0x%08lx (per 16bit) for ch%d\n",
487 writel(umc_cmdctla[freq], dc_base + UMC_CMDCTLA);
489 writel(ch == 2 ? umc_cmdctlb_ch2[freq] : umc_cmdctlb_ch01[freq],
490 dc_base + UMC_CMDCTLB);
492 writel(umc_spcctla[freq][size_e], dc_base + UMC_SPCCTLA);
493 writel(umc_spcctlb[freq], dc_base + UMC_SPCCTLB);
497 /* ES2 inserted one more FF to the logic. */
498 if (uniphier_get_soc_model() >= 2)
502 val |= 0xf << UMC_RDATACTL_RADLTY_SHIFT;
503 val |= (latency - 0xf) << UMC_RDATACTL_RAD2LTY_SHIFT;
505 val |= latency << UMC_RDATACTL_RADLTY_SHIFT;
508 writel(val, dc_base + UMC_RDATACTL_D0);
510 writel(val, dc_base + UMC_RDATACTL_D1);
512 writel(0x04060A02, dc_base + UMC_WDATACTL_D0);
514 writel(0x04060A02, dc_base + UMC_WDATACTL_D1);
515 writel(0x04000000, dc_base + UMC_DATASET);
516 writel(0x00400020, dc_base + UMC_DCCGCTL);
517 writel(0x00000084, dc_base + UMC_FLOWCTLG);
518 writel(0x00000000, dc_base + UMC_ACSSETA);
520 writel(ch == 2 ? umc_flowctla_ch2[freq] : umc_flowctla_ch01[freq],
521 dc_base + UMC_FLOWCTLA);
523 writel(0x00004400, dc_base + UMC_FLOWCTLC);
524 writel(0x200A0A00, dc_base + UMC_SPCSETB);
525 writel(0x00000520, dc_base + UMC_DFICUPDCTLA);
526 writel(0x0000000D, dc_base + UMC_RESPCTL);
529 writel(0x00202000, dc_base + UMC_FLOWCTLB);
530 writel(0xFDBFFFFF, dc_base + UMC_FLOWCTLOB0);
531 writel(0xFFFFFFFF, dc_base + UMC_FLOWCTLOB1);
532 writel(0x00080700, dc_base + UMC_BSICMAPSET);
534 writel(0x00200000, dc_base + UMC_FLOWCTLB);
535 writel(0x00000000, dc_base + UMC_BSICMAPSET);
538 writel(0x00000000, dc_base + UMC_ERRMASKA);
539 writel(0x00000000, dc_base + UMC_ERRMASKB);
544 static int umc_ch_init(void __iomem *umc_ch_base, enum dram_freq freq,
545 unsigned long size, unsigned int width, int ch)
547 void __iomem *dc_base = umc_ch_base + 0x00011000;
548 void __iomem *phy_base = umc_ch_base + 0x00030000;
551 writel(0x00000002, dc_base + UMC_INITSET);
552 while (readl(dc_base + UMC_INITSTAT) & BIT(2))
555 /* deassert PHY reset signals */
556 writel(UMC_DIOCTLA_CTL_NRST | UMC_DIOCTLA_CFG_NRST,
557 dc_base + UMC_DIOCTLA);
559 ddrphy_init(phy_base, freq, width, ch);
561 ret = ddrphy_impedance_calibration(phy_base);
565 ddrphy_dram_init(phy_base);
569 ret = umc_dc_init(dc_base, freq, size, width, ch);
573 umc_ud_init(umc_ch_base, ch);
575 ret = ddrphy_training(phy_base);
581 /* match the system latency between UMC and PHY */
582 umc_set_system_latency(dc_base,
583 ddrphy_get_system_latency(phy_base, width));
587 /* stop auto refresh before clearing FIFO in PHY */
588 umc_refresh_ctrl(dc_base, 0);
589 ddrphy_fifo_reset(phy_base);
590 umc_refresh_ctrl(dc_base, 1);
597 static void um_init(void __iomem *um_base)
599 writel(0x000000ff, um_base + UMC_MBUS0);
600 writel(0x000000ff, um_base + UMC_MBUS1);
601 writel(0x000000ff, um_base + UMC_MBUS2);
602 writel(0x000000ff, um_base + UMC_MBUS3);
605 int uniphier_pxs2_umc_init(const struct uniphier_board_data *bd)
607 void __iomem *um_base = (void __iomem *)0x5b600000;
608 void __iomem *umc_ch_base = (void __iomem *)0x5b800000;
612 switch (bd->dram_freq) {
614 freq = DRAM_FREQ_1866M;
617 freq = DRAM_FREQ_2133M;
620 pr_err("unsupported DRAM frequency %d MHz\n", bd->dram_freq);
624 for (ch = 0; ch < DRAM_CH_NR; ch++) {
625 unsigned long size = bd->dram_ch[ch].size;
626 unsigned int width = bd->dram_ch[ch].width;
629 ret = umc_ch_init(umc_ch_base, freq,
630 size / (width / 16), width, ch);
632 pr_err("failed to initialize UMC ch%d\n", ch);
637 umc_ch_base += 0x00200000;