blob: 35d98faf58f3d2a62511ecf83add479d3c10f243 [file] [log] [blame]
Tom Rini10e47792018-05-06 17:58:06 -04001// SPDX-License-Identifier: GPL-2.0
Stefan Roeseae6223d2015-01-19 11:33:40 +01002/*
3 * Copyright (C) Marvell International Ltd. and its affiliates
Stefan Roeseae6223d2015-01-19 11:33:40 +01004 */
5
6#include <common.h>
7#include <i2c.h>
Simon Glass0f2af882020-05-10 11:40:05 -06008#include <log.h>
Stefan Roeseae6223d2015-01-19 11:33:40 +01009#include <spl.h>
10#include <asm/io.h>
11#include <asm/arch/cpu.h>
12#include <asm/arch/soc.h>
Simon Glassdbd79542020-05-10 11:40:11 -060013#include <linux/delay.h>
Stefan Roeseae6223d2015-01-19 11:33:40 +010014
15#include "ddr3_init.h"
16#include "ddr3_hw_training.h"
17#include "xor.h"
18
19#ifdef MV88F78X60
20#include "ddr3_patterns_64bit.h"
21#else
22#include "ddr3_patterns_16bit.h"
23#if defined(MV88F672X)
24#include "ddr3_patterns_16bit.h"
25#endif
26#endif
27
28/*
29 * Debug
30 */
31
32#define DEBUG_MAIN_C(s, d, l) \
33 DEBUG_MAIN_S(s); DEBUG_MAIN_D(d, l); DEBUG_MAIN_S("\n")
34#define DEBUG_MAIN_FULL_C(s, d, l) \
35 DEBUG_MAIN_FULL_S(s); DEBUG_MAIN_FULL_D(d, l); DEBUG_MAIN_FULL_S("\n")
36
37#ifdef MV_DEBUG_MAIN
38#define DEBUG_MAIN_S(s) puts(s)
39#define DEBUG_MAIN_D(d, l) printf("%x", d)
40#else
41#define DEBUG_MAIN_S(s)
42#define DEBUG_MAIN_D(d, l)
43#endif
44
45#ifdef MV_DEBUG_MAIN_FULL
46#define DEBUG_MAIN_FULL_S(s) puts(s)
47#define DEBUG_MAIN_FULL_D(d, l) printf("%x", d)
48#else
49#define DEBUG_MAIN_FULL_S(s)
50#define DEBUG_MAIN_FULL_D(d, l)
51#endif
52
53#ifdef MV_DEBUG_SUSPEND_RESUME
54#define DEBUG_SUSPEND_RESUME_S(s) puts(s)
55#define DEBUG_SUSPEND_RESUME_D(d, l) printf("%x", d)
56#else
57#define DEBUG_SUSPEND_RESUME_S(s)
58#define DEBUG_SUSPEND_RESUME_D(d, l)
59#endif
60
61static u32 ddr3_sw_wl_rl_debug;
62static u32 ddr3_run_pbs = 1;
63
64void ddr3_print_version(void)
65{
66 puts("DDR3 Training Sequence - Ver 5.7.");
67}
68
69void ddr3_set_sw_wl_rl_debug(u32 val)
70{
71 ddr3_sw_wl_rl_debug = val;
72}
73
74void ddr3_set_pbs(u32 val)
75{
76 ddr3_run_pbs = val;
77}
78
79int ddr3_hw_training(u32 target_freq, u32 ddr_width, int xor_bypass,
80 u32 scrub_offs, u32 scrub_size, int dqs_clk_aligned,
81 int debug_mode, int reg_dimm_skip_wl)
82{
83 /* A370 has no PBS mechanism */
84 __maybe_unused u32 first_loop_flag = 0;
85 u32 freq, reg;
86 MV_DRAM_INFO dram_info;
87 int ratio_2to1 = 0;
88 int tmp_ratio = 1;
89 int status;
90
91 if (debug_mode)
92 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 1\n");
93
94 memset(&dram_info, 0, sizeof(dram_info));
95 dram_info.num_cs = ddr3_get_cs_num_from_reg();
96 dram_info.cs_ena = ddr3_get_cs_ena_from_reg();
97 dram_info.target_frequency = target_freq;
98 dram_info.ddr_width = ddr_width;
99 dram_info.num_of_std_pups = ddr_width / PUP_SIZE;
100 dram_info.rl400_bug = 0;
101 dram_info.multi_cs_mr_support = 0;
102#ifdef MV88F67XX
103 dram_info.rl400_bug = 1;
104#endif
105
106 /* Ignore ECC errors - if ECC is enabled */
107 reg = reg_read(REG_SDRAM_CONFIG_ADDR);
108 if (reg & (1 << REG_SDRAM_CONFIG_ECC_OFFS)) {
109 dram_info.ecc_ena = 1;
110 reg |= (1 << REG_SDRAM_CONFIG_IERR_OFFS);
111 reg_write(REG_SDRAM_CONFIG_ADDR, reg);
112 } else {
113 dram_info.ecc_ena = 0;
114 }
115
116 reg = reg_read(REG_SDRAM_CONFIG_ADDR);
117 if (reg & (1 << REG_SDRAM_CONFIG_REGDIMM_OFFS))
118 dram_info.reg_dimm = 1;
119 else
120 dram_info.reg_dimm = 0;
121
122 dram_info.num_of_total_pups = ddr_width / PUP_SIZE + dram_info.ecc_ena;
123
124 /* Get target 2T value */
125 reg = reg_read(REG_DUNIT_CTRL_LOW_ADDR);
126 dram_info.mode_2t = (reg >> REG_DUNIT_CTRL_LOW_2T_OFFS) &
127 REG_DUNIT_CTRL_LOW_2T_MASK;
128
129 /* Get target CL value */
130#ifdef MV88F67XX
131 reg = reg_read(REG_DDR3_MR0_ADDR) >> 2;
132#else
133 reg = reg_read(REG_DDR3_MR0_CS_ADDR) >> 2;
134#endif
135
136 reg = (((reg >> 1) & 0xE) | (reg & 0x1)) & 0xF;
137 dram_info.cl = ddr3_valid_cl_to_cl(reg);
138
139 /* Get target CWL value */
140#ifdef MV88F67XX
141 reg = reg_read(REG_DDR3_MR2_ADDR) >> REG_DDR3_MR2_CWL_OFFS;
142#else
143 reg = reg_read(REG_DDR3_MR2_CS_ADDR) >> REG_DDR3_MR2_CWL_OFFS;
144#endif
145
146 reg &= REG_DDR3_MR2_CWL_MASK;
147 dram_info.cwl = reg;
148#if !defined(MV88F67XX)
149 /* A370 has no PBS mechanism */
150#if defined(MV88F78X60)
151 if ((dram_info.target_frequency > DDR_400) && (ddr3_run_pbs))
152 first_loop_flag = 1;
153#else
154 /* first_loop_flag = 1; skip mid freq at ALP/A375 */
155 if ((dram_info.target_frequency > DDR_400) && (ddr3_run_pbs) &&
156 (mv_ctrl_revision_get() >= UMC_A0))
157 first_loop_flag = 1;
158 else
159 first_loop_flag = 0;
160#endif
161#endif
162
163 freq = dram_info.target_frequency;
164
165 /* Set ODT to always on */
166 ddr3_odt_activate(1);
167
168 /* Init XOR */
169 mv_sys_xor_init(&dram_info);
170
171 /* Get DRAM/HCLK ratio */
172 if (reg_read(REG_DDR_IO_ADDR) & (1 << REG_DDR_IO_CLK_RATIO_OFFS))
173 ratio_2to1 = 1;
174
175 /*
176 * Xor Bypass - ECC support in AXP is currently available for 1:1
177 * modes frequency modes.
178 * Not all frequency modes support the ddr3 training sequence
179 * (Only 1200/300).
180 * Xor Bypass allows using the Xor initializations and scrubbing
181 * inside the ddr3 training sequence without running the training
182 * itself.
183 */
184 if (xor_bypass == 0) {
185 if (ddr3_run_pbs) {
186 DEBUG_MAIN_S("DDR3 Training Sequence - Run with PBS.\n");
187 } else {
188 DEBUG_MAIN_S("DDR3 Training Sequence - Run without PBS.\n");
189 }
190
191 if (dram_info.target_frequency > DFS_MARGIN) {
192 tmp_ratio = 0;
193 freq = DDR_100;
194
195 if (dram_info.reg_dimm == 1)
196 freq = DDR_300;
197
198 if (MV_OK != ddr3_dfs_high_2_low(freq, &dram_info)) {
199 /* Set low - 100Mhz DDR Frequency by HW */
200 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs High2Low)\n");
201 return MV_DDR3_TRAINING_ERR_DFS_H2L;
202 }
203
204 if ((dram_info.reg_dimm == 1) &&
205 (reg_dimm_skip_wl == 0)) {
206 if (MV_OK !=
207 ddr3_write_leveling_hw_reg_dimm(freq,
208 &dram_info))
209 DEBUG_MAIN_S("DDR3 Training Sequence - Registered DIMM Low WL - SKIP\n");
210 }
211
212 if (ddr3_get_log_level() >= MV_LOG_LEVEL_1)
213 ddr3_print_freq(freq);
214
215 if (debug_mode)
216 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 2\n");
217 } else {
218 if (!dqs_clk_aligned) {
219#ifdef MV88F67XX
220 /*
221 * If running training sequence without DFS,
222 * we must run Write leveling before writing
223 * the patterns
224 */
225
226 /*
227 * ODT - Multi CS system use SW WL,
228 * Single CS System use HW WL
229 */
230 if (dram_info.cs_ena > 1) {
231 if (MV_OK !=
232 ddr3_write_leveling_sw(
233 freq, tmp_ratio,
234 &dram_info)) {
235 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
236 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
237 }
238 } else {
239 if (MV_OK !=
240 ddr3_write_leveling_hw(freq,
241 &dram_info)) {
242 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
243 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
244 }
245 }
246#else
247 if (MV_OK != ddr3_write_leveling_hw(
248 freq, &dram_info)) {
249 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
250 if (ddr3_sw_wl_rl_debug) {
251 if (MV_OK !=
252 ddr3_write_leveling_sw(
253 freq, tmp_ratio,
254 &dram_info)) {
255 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
256 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
257 }
258 } else {
259 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
260 }
261 }
262#endif
263 }
264
265 if (debug_mode)
266 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 3\n");
267 }
268
269 if (MV_OK != ddr3_load_patterns(&dram_info, 0)) {
270 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Loading Patterns)\n");
271 return MV_DDR3_TRAINING_ERR_LOAD_PATTERNS;
272 }
273
274 /*
275 * TODO:
276 * The mainline U-Boot port of the bin_hdr DDR training code
277 * needs a delay of minimum 20ms here (10ms is a bit too short
278 * and the CPU hangs). The bin_hdr code doesn't have this delay.
279 * To be save here, lets add a delay of 50ms here.
280 *
281 * Tested on the Marvell DB-MV784MP-GP board
282 */
283 mdelay(50);
284
285 do {
286 freq = dram_info.target_frequency;
287 tmp_ratio = ratio_2to1;
288 DEBUG_MAIN_FULL_S("DDR3 Training Sequence - DEBUG - 4\n");
289
290#if defined(MV88F78X60)
291 /*
292 * There is a difference on the DFS frequency at the
293 * first iteration of this loop
294 */
295 if (first_loop_flag) {
296 freq = DDR_400;
297 tmp_ratio = 0;
298 }
299#endif
300
301 if (MV_OK != ddr3_dfs_low_2_high(freq, tmp_ratio,
302 &dram_info)) {
303 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs Low2High)\n");
304 return MV_DDR3_TRAINING_ERR_DFS_H2L;
305 }
306
307 if (ddr3_get_log_level() >= MV_LOG_LEVEL_1) {
308 ddr3_print_freq(freq);
309 }
310
311 if (debug_mode)
312 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 5\n");
313
314 /* Write leveling */
315 if (!dqs_clk_aligned) {
316#ifdef MV88F67XX
317 /*
318 * ODT - Multi CS system that not support Multi
319 * CS MRS commands must use SW WL
320 */
321 if (dram_info.cs_ena > 1) {
322 if (MV_OK != ddr3_write_leveling_sw(
323 freq, tmp_ratio, &dram_info)) {
324 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
325 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
326 }
327 } else {
328 if (MV_OK != ddr3_write_leveling_hw(
329 freq, &dram_info)) {
330 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
331 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
332 }
333 }
334#else
335 if ((dram_info.reg_dimm == 1) &&
336 (freq == DDR_400)) {
337 if (reg_dimm_skip_wl == 0) {
338 if (MV_OK != ddr3_write_leveling_hw_reg_dimm(
339 freq, &dram_info))
340 DEBUG_MAIN_S("DDR3 Training Sequence - Registered DIMM WL - SKIP\n");
341 }
342 } else {
343 if (MV_OK != ddr3_write_leveling_hw(
344 freq, &dram_info)) {
345 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
346 if (ddr3_sw_wl_rl_debug) {
347 if (MV_OK != ddr3_write_leveling_sw(
348 freq, tmp_ratio, &dram_info)) {
349 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
350 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
351 }
352 } else {
353 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
354 }
355 }
356 }
357#endif
358 if (debug_mode)
359 DEBUG_MAIN_S
360 ("DDR3 Training Sequence - DEBUG - 6\n");
361 }
362
363 /* Read Leveling */
364 /*
365 * Armada 370 - Support for HCLK @ 400MHZ - must use
366 * SW read leveling
367 */
368 if (freq == DDR_400 && dram_info.rl400_bug) {
369 status = ddr3_read_leveling_sw(freq, tmp_ratio,
370 &dram_info);
371 if (MV_OK != status) {
372 DEBUG_MAIN_S
373 ("DDR3 Training Sequence - FAILED (Read Leveling Sw)\n");
374 return status;
375 }
376 } else {
377 if (MV_OK != ddr3_read_leveling_hw(
378 freq, &dram_info)) {
379 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Hw)\n");
380 if (ddr3_sw_wl_rl_debug) {
381 if (MV_OK != ddr3_read_leveling_sw(
382 freq, tmp_ratio,
383 &dram_info)) {
384 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Sw)\n");
385 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
386 }
387 } else {
388 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
389 }
390 }
391 }
392
393 if (debug_mode)
394 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 7\n");
395
396 if (MV_OK != ddr3_wl_supplement(&dram_info)) {
397 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hi-Freq Sup)\n");
398 return MV_DDR3_TRAINING_ERR_WR_LVL_HI_FREQ;
399 }
400
401 if (debug_mode)
402 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 8\n");
403#if !defined(MV88F67XX)
404 /* A370 has no PBS mechanism */
405#if defined(MV88F78X60) || defined(MV88F672X)
406 if (first_loop_flag == 1) {
407 first_loop_flag = 0;
408
409 status = MV_OK;
410 status = ddr3_pbs_rx(&dram_info);
411 if (MV_OK != status) {
412 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (PBS RX)\n");
413 return status;
414 }
415
416 if (debug_mode)
417 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 9\n");
418
419 status = ddr3_pbs_tx(&dram_info);
420 if (MV_OK != status) {
421 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (PBS TX)\n");
422 return status;
423 }
424
425 if (debug_mode)
426 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 10\n");
427 }
428#endif
429#endif
430 } while (freq != dram_info.target_frequency);
431
432 status = ddr3_dqs_centralization_rx(&dram_info);
433 if (MV_OK != status) {
434 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (DQS Centralization RX)\n");
435 return status;
436 }
437
438 if (debug_mode)
439 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 11\n");
440
441 status = ddr3_dqs_centralization_tx(&dram_info);
442 if (MV_OK != status) {
443 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (DQS Centralization TX)\n");
444 return status;
445 }
446
447 if (debug_mode)
448 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 12\n");
449 }
450
451 ddr3_set_performance_params(&dram_info);
452
453 if (dram_info.ecc_ena) {
Bin Meng75574052016-02-05 19:30:11 -0800454 /* Need to SCRUB the DRAM memory area to load U-Boot */
Stefan Roeseae6223d2015-01-19 11:33:40 +0100455 mv_sys_xor_finish();
456 dram_info.num_cs = 1;
457 dram_info.cs_ena = 1;
458 mv_sys_xor_init(&dram_info);
459 mv_xor_mem_init(0, scrub_offs, scrub_size, 0xdeadbeef,
460 0xdeadbeef);
461
462 /* Wait for previous transfer completion */
463 while (mv_xor_state_get(0) != MV_IDLE)
464 ;
465
466 if (debug_mode)
467 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 13\n");
468 }
469
470 /* Return XOR State */
471 mv_sys_xor_finish();
472
473#if defined(MV88F78X60)
474 /* Save training results in memeory for resume state */
475 ddr3_save_training(&dram_info);
476#endif
477 /* Clear ODT always on */
478 ddr3_odt_activate(0);
479
480 /* Configure Dynamic read ODT */
481 ddr3_odt_read_dynamic_config(&dram_info);
482
483 return MV_OK;
484}
485
486void ddr3_set_performance_params(MV_DRAM_INFO *dram_info)
487{
488 u32 twr2wr, trd2rd, trd2wr_wr2rd;
489 u32 tmp1, tmp2, reg;
490
491 DEBUG_MAIN_FULL_C("Max WL Phase: ", dram_info->wl_max_phase, 2);
492 DEBUG_MAIN_FULL_C("Min WL Phase: ", dram_info->wl_min_phase, 2);
493 DEBUG_MAIN_FULL_C("Max RL Phase: ", dram_info->rl_max_phase, 2);
494 DEBUG_MAIN_FULL_C("Min RL Phase: ", dram_info->rl_min_phase, 2);
495
496 if (dram_info->wl_max_phase < 2)
497 twr2wr = 0x2;
498 else
499 twr2wr = 0x3;
500
501 trd2rd = 0x1 + (dram_info->rl_max_phase + 1) / 2 +
502 (dram_info->rl_max_phase + 1) % 2;
503
504 tmp1 = (dram_info->rl_max_phase - dram_info->wl_min_phase) / 2 +
505 (((dram_info->rl_max_phase - dram_info->wl_min_phase) % 2) >
506 0 ? 1 : 0);
507 tmp2 = (dram_info->wl_max_phase - dram_info->rl_min_phase) / 2 +
508 ((dram_info->wl_max_phase - dram_info->rl_min_phase) % 2 >
509 0 ? 1 : 0);
510 trd2wr_wr2rd = (tmp1 >= tmp2) ? tmp1 : tmp2;
511
512 trd2wr_wr2rd += 2;
513 trd2rd += 2;
514 twr2wr += 2;
515
516 DEBUG_MAIN_FULL_C("WR 2 WR: ", twr2wr, 2);
517 DEBUG_MAIN_FULL_C("RD 2 RD: ", trd2rd, 2);
518 DEBUG_MAIN_FULL_C("RD 2 WR / WR 2 RD: ", trd2wr_wr2rd, 2);
519
520 reg = reg_read(REG_SDRAM_TIMING_HIGH_ADDR);
521
522 reg &= ~(REG_SDRAM_TIMING_H_W2W_MASK << REG_SDRAM_TIMING_H_W2W_OFFS);
523 reg |= ((twr2wr & REG_SDRAM_TIMING_H_W2W_MASK) <<
524 REG_SDRAM_TIMING_H_W2W_OFFS);
525
526 reg &= ~(REG_SDRAM_TIMING_H_R2R_MASK << REG_SDRAM_TIMING_H_R2R_OFFS);
527 reg &= ~(REG_SDRAM_TIMING_H_R2R_H_MASK <<
528 REG_SDRAM_TIMING_H_R2R_H_OFFS);
529 reg |= ((trd2rd & REG_SDRAM_TIMING_H_R2R_MASK) <<
530 REG_SDRAM_TIMING_H_R2R_OFFS);
531 reg |= (((trd2rd >> 2) & REG_SDRAM_TIMING_H_R2R_H_MASK) <<
532 REG_SDRAM_TIMING_H_R2R_H_OFFS);
533
534 reg &= ~(REG_SDRAM_TIMING_H_R2W_W2R_MASK <<
535 REG_SDRAM_TIMING_H_R2W_W2R_OFFS);
536 reg &= ~(REG_SDRAM_TIMING_H_R2W_W2R_H_MASK <<
537 REG_SDRAM_TIMING_H_R2W_W2R_H_OFFS);
538 reg |= ((trd2wr_wr2rd & REG_SDRAM_TIMING_H_R2W_W2R_MASK) <<
539 REG_SDRAM_TIMING_H_R2W_W2R_OFFS);
540 reg |= (((trd2wr_wr2rd >> 2) & REG_SDRAM_TIMING_H_R2W_W2R_H_MASK) <<
541 REG_SDRAM_TIMING_H_R2W_W2R_H_OFFS);
542
543 reg_write(REG_SDRAM_TIMING_HIGH_ADDR, reg);
544}
545
546/*
547 * Perform DDR3 PUP Indirect Write
548 */
549void ddr3_write_pup_reg(u32 mode, u32 cs, u32 pup, u32 phase, u32 delay)
550{
551 u32 reg = 0;
552
553 if (pup == PUP_BC)
554 reg |= (1 << REG_PHY_BC_OFFS);
555 else
556 reg |= (pup << REG_PHY_PUP_OFFS);
557
558 reg |= ((0x4 * cs + mode) << REG_PHY_CS_OFFS);
559 reg |= (phase << REG_PHY_PHASE_OFFS) | delay;
560
561 if (mode == PUP_WL_MODE)
562 reg |= ((INIT_WL_DELAY + delay) << REG_PHY_DQS_REF_DLY_OFFS);
563
564 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
565 reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
566 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
567
568 do {
569 reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
570 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
571 } while (reg); /* Wait for '0' to mark the end of the transaction */
572
573 /* If read Leveling mode - need to write to register 3 separetly */
574 if (mode == PUP_RL_MODE) {
575 reg = 0;
576
577 if (pup == PUP_BC)
578 reg |= (1 << REG_PHY_BC_OFFS);
579 else
580 reg |= (pup << REG_PHY_PUP_OFFS);
581
582 reg |= ((0x4 * cs + mode + 1) << REG_PHY_CS_OFFS);
583 reg |= (INIT_RL_DELAY);
584
585 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
586 reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
587 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
588
589 do {
590 reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
591 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
592 } while (reg);
593 }
594}
595
596/*
597 * Perform DDR3 PUP Indirect Read
598 */
599u32 ddr3_read_pup_reg(u32 mode, u32 cs, u32 pup)
600{
601 u32 reg;
602
603 reg = (pup << REG_PHY_PUP_OFFS) |
604 ((0x4 * cs + mode) << REG_PHY_CS_OFFS);
605 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
606
607 reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_RD;
608 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
609
610 do {
611 reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
612 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
613 } while (reg); /* Wait for '0' to mark the end of the transaction */
614
615 return reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR); /* 0x16A0 */
616}
617
618/*
619 * Set training patterns
620 */
621int ddr3_load_patterns(MV_DRAM_INFO *dram_info, int resume)
622{
623 u32 reg;
624
625 /* Enable SW override - Required for the ECC Pup */
626 reg = reg_read(REG_DRAM_TRAINING_2_ADDR) |
627 (1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS);
628
629 /* [0] = 1 - Enable SW override */
630 /* 0x15B8 - Training SW 2 Register */
631 reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
632
633 reg = (1 << REG_DRAM_TRAINING_AUTO_OFFS);
634 reg_write(REG_DRAM_TRAINING_ADDR, reg); /* 0x15B0 - Training Register */
635
636 if (resume == 0) {
637#if defined(MV88F78X60) || defined(MV88F672X)
638 ddr3_load_pbs_patterns(dram_info);
639#endif
640 ddr3_load_dqs_patterns(dram_info);
641 }
642
643 /* Disable SW override - Must be in a different stage */
644 /* [0]=0 - Enable SW override */
645 reg = reg_read(REG_DRAM_TRAINING_2_ADDR);
646 reg &= ~(1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS);
647 /* 0x15B8 - Training SW 2 Register */
648 reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
649
650 reg = reg_read(REG_DRAM_TRAINING_1_ADDR) |
651 (1 << REG_DRAM_TRAINING_1_TRNBPOINT_OFFS);
652 reg_write(REG_DRAM_TRAINING_1_ADDR, reg);
653
654 /* Set Base Addr */
655#if defined(MV88F67XX)
656 reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR, 0);
657#else
658 if (resume == 0)
659 reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR, 0);
660 else
661 reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR,
662 RESUME_RL_PATTERNS_ADDR);
663#endif
664
665 /* Set Patterns */
666 if (resume == 0) {
667 reg = (dram_info->cs_ena << REG_DRAM_TRAINING_CS_OFFS) |
668 (1 << REG_DRAM_TRAINING_PATTERNS_OFFS);
669 } else {
670 reg = (0x1 << REG_DRAM_TRAINING_CS_OFFS) |
671 (1 << REG_DRAM_TRAINING_PATTERNS_OFFS);
672 }
673
674 reg |= (1 << REG_DRAM_TRAINING_AUTO_OFFS);
675
676 reg_write(REG_DRAM_TRAINING_ADDR, reg);
677
678 udelay(100);
679
680 /* Check if Successful */
681 if (reg_read(REG_DRAM_TRAINING_ADDR) &
682 (1 << REG_DRAM_TRAINING_ERROR_OFFS))
683 return MV_OK;
684 else
685 return MV_FAIL;
686}
687
688#if !defined(MV88F67XX)
689/*
690 * Name: ddr3_save_training(MV_DRAM_INFO *dram_info)
691 * Desc: saves the training results to memeory (RL,WL,PBS,Rx/Tx
692 * Centeralization)
693 * Args: MV_DRAM_INFO *dram_info
694 * Notes:
695 * Returns: None.
696 */
697void ddr3_save_training(MV_DRAM_INFO *dram_info)
698{
699 u32 val, pup, tmp_cs, cs, i, dq;
700 u32 crc = 0;
701 u32 regs = 0;
702 u32 *sdram_offset = (u32 *)RESUME_TRAINING_VALUES_ADDR;
703 u32 mode_config[MAX_TRAINING_MODE];
704
705 mode_config[DQS_WR_MODE] = PUP_DQS_WR;
706 mode_config[WL_MODE_] = PUP_WL_MODE;
707 mode_config[RL_MODE_] = PUP_RL_MODE;
708 mode_config[DQS_RD_MODE] = PUP_DQS_RD;
709 mode_config[PBS_TX_DM_MODE] = PUP_PBS_TX_DM;
710 mode_config[PBS_TX_MODE] = PUP_PBS_TX;
711 mode_config[PBS_RX_MODE] = PUP_PBS_RX;
712
713 /* num of training modes */
714 for (i = 0; i < MAX_TRAINING_MODE; i++) {
715 tmp_cs = dram_info->cs_ena;
716 /* num of CS */
717 for (cs = 0; cs < MAX_CS; cs++) {
718 if (tmp_cs & (1 << cs)) {
719 /* num of PUPs */
720 for (pup = 0; pup < dram_info->num_of_total_pups;
721 pup++) {
722 if (pup == dram_info->num_of_std_pups &&
723 dram_info->ecc_ena)
724 pup = ECC_PUP;
725 if (i == PBS_TX_DM_MODE) {
726 /*
727 * Change CS bitmask because
728 * PBS works only with CS0
729 */
730 tmp_cs = 0x1;
731 val = ddr3_read_pup_reg(
732 mode_config[i], CS0, pup);
733 } else if (i == PBS_TX_MODE ||
734 i == PBS_RX_MODE) {
735 /*
736 * Change CS bitmask because
737 * PBS works only with CS0
738 */
739 tmp_cs = 0x1;
740 for (dq = 0; dq <= DQ_NUM;
741 dq++) {
742 val = ddr3_read_pup_reg(
743 mode_config[i] + dq,
744 CS0,
745 pup);
746 (*sdram_offset) = val;
747 crc += *sdram_offset;
748 sdram_offset++;
749 regs++;
750 }
751 continue;
752 } else {
753 val = ddr3_read_pup_reg(
754 mode_config[i], cs, pup);
755 }
756
757 *sdram_offset = val;
758 crc += *sdram_offset;
759 sdram_offset++;
760 regs++;
761 }
762 }
763 }
764 }
765
766 *sdram_offset = reg_read(REG_READ_DATA_SAMPLE_DELAYS_ADDR);
767 crc += *sdram_offset;
768 sdram_offset++;
769 regs++;
770 *sdram_offset = reg_read(REG_READ_DATA_READY_DELAYS_ADDR);
771 crc += *sdram_offset;
772 sdram_offset++;
773 regs++;
774 sdram_offset = (u32 *)NUM_OF_REGISTER_ADDR;
775 *sdram_offset = regs;
776 DEBUG_SUSPEND_RESUME_S("Training Results CheckSum write= ");
777 DEBUG_SUSPEND_RESUME_D(crc, 8);
778 DEBUG_SUSPEND_RESUME_S("\n");
779 sdram_offset = (u32 *)CHECKSUM_RESULT_ADDR;
780 *sdram_offset = crc;
781}
782
783/*
784 * Name: ddr3_read_training_results()
785 * Desc: Reads the training results from memeory (RL,WL,PBS,Rx/Tx
786 * Centeralization)
787 * and writes them to the relevant registers
788 * Args: MV_DRAM_INFO *dram_info
789 * Notes:
790 * Returns: None.
791 */
792int ddr3_read_training_results(void)
793{
794 u32 val, reg, idx, dqs_wr_idx = 0, crc = 0;
795 u32 *sdram_offset = (u32 *)RESUME_TRAINING_VALUES_ADDR;
796 u32 training_val[RESUME_TRAINING_VALUES_MAX] = { 0 };
797 u32 regs = *((u32 *)NUM_OF_REGISTER_ADDR);
798
799 /*
800 * Read Training results & Dunit registers from memory and write
801 * it to an array
802 */
803 for (idx = 0; idx < regs; idx++) {
804 training_val[idx] = *sdram_offset;
805 crc += *sdram_offset;
806 sdram_offset++;
807 }
808
809 sdram_offset = (u32 *)CHECKSUM_RESULT_ADDR;
810
811 if ((*sdram_offset) == crc) {
812 DEBUG_SUSPEND_RESUME_S("Training Results CheckSum read PASS= ");
813 DEBUG_SUSPEND_RESUME_D(crc, 8);
814 DEBUG_SUSPEND_RESUME_S("\n");
815 } else {
816 DEBUG_MAIN_S("Wrong Training Results CheckSum\n");
817 return MV_FAIL;
818 }
819
820 /*
821 * We iterate through all the registers except for the last 2 since
822 * they are Dunit registers (and not PHY registers)
823 */
824 for (idx = 0; idx < (regs - 2); idx++) {
825 val = training_val[idx];
826 reg = (val >> REG_PHY_CS_OFFS) & 0x3F; /*read the phy address */
827
828 /* Check if the values belongs to the DQS WR */
829 if (reg == PUP_WL_MODE) {
830 /* bit[5:0] in DQS_WR are delay */
831 val = (training_val[dqs_wr_idx++] & 0x3F);
832 /*
833 * bit[15:10] are DQS_WR delay & bit[9:0] are
834 * WL phase & delay
835 */
836 val = (val << REG_PHY_DQS_REF_DLY_OFFS) |
837 (training_val[idx] & 0x3C003FF);
838 /* Add Request pending and write operation bits */
839 val |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
840 } else if (reg == PUP_DQS_WR) {
841 /*
842 * Do nothing since DQS_WR will be done in PUP_WL_MODE
843 */
844 continue;
845 }
846
847 val |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
848 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, val);
849 do {
850 val = (reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR)) &
851 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
852 } while (val); /* Wait for '0' to mark the end of the transaction */
853 }
854
855 /* write last 2 Dunit configurations */
856 val = training_val[idx];
857 reg_write(REG_READ_DATA_SAMPLE_DELAYS_ADDR, val); /* reg 0x1538 */
858 val = training_val[idx + 1];
859 reg_write(REG_READ_DATA_READY_DELAYS_ADDR, val); /* reg 0x153c */
860
861 return MV_OK;
862}
863
864/*
865 * Name: ddr3_check_if_resume_mode()
866 * Desc: Reads the address (0x3000) of the Resume Magic word (0xDEADB002)
867 * Args: MV_DRAM_INFO *dram_info
868 * Notes:
869 * Returns: return (magic_word == SUSPEND_MAGIC_WORD)
870 */
871int ddr3_check_if_resume_mode(MV_DRAM_INFO *dram_info, u32 freq)
872{
873 u32 magic_word;
874 u32 *sdram_offset = (u32 *)BOOT_INFO_ADDR;
875
876 if (dram_info->reg_dimm != 1) {
877 /*
878 * Perform write levleling in order initiate the phy with
879 * low frequency
880 */
881 if (MV_OK != ddr3_write_leveling_hw(freq, dram_info)) {
882 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
883 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
884 }
885 }
886
887 if (MV_OK != ddr3_load_patterns(dram_info, 1)) {
888 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Loading Patterns)\n");
889 return MV_DDR3_TRAINING_ERR_LOAD_PATTERNS;
890 }
891
892 /* Enable CS0 only for RL */
893 dram_info->cs_ena = 0x1;
894
895 /* Perform Read levleling in order to get stable memory */
896 if (MV_OK != ddr3_read_leveling_hw(freq, dram_info)) {
897 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Hw)\n");
898 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
899 }
900
901 /* Back to relevant CS */
902 dram_info->cs_ena = ddr3_get_cs_ena_from_reg();
903
904 magic_word = *sdram_offset;
905 return magic_word == SUSPEND_MAGIC_WORD;
906}
907
908/*
909 * Name: ddr3_training_suspend_resume()
910 * Desc: Execute the Resume state
911 * Args: MV_DRAM_INFO *dram_info
912 * Notes:
913 * Returns: return (magic_word == SUSPEND_MAGIC_WORD)
914 */
915int ddr3_training_suspend_resume(MV_DRAM_INFO *dram_info)
916{
917 u32 freq, reg;
918 int tmp_ratio;
919
920 /* Configure DDR */
921 if (MV_OK != ddr3_read_training_results())
922 return MV_FAIL;
923
924 /* Reset read FIFO */
925 reg = reg_read(REG_DRAM_TRAINING_ADDR);
926
927 /* Start Auto Read Leveling procedure */
928 reg |= (1 << REG_DRAM_TRAINING_RL_OFFS);
929 reg_write(REG_DRAM_TRAINING_ADDR, reg); /* 0x15B0 - Training Register */
930
931 reg = reg_read(REG_DRAM_TRAINING_2_ADDR);
932 reg |= ((1 << REG_DRAM_TRAINING_2_FIFO_RST_OFFS) +
933 (1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS));
934
935 /* [0] = 1 - Enable SW override, [4] = 1 - FIFO reset */
936 /* 0x15B8 - Training SW 2 Register */
937 reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
938
939 udelay(2);
940
941 reg = reg_read(REG_DRAM_TRAINING_ADDR);
942 /* Clear Auto Read Leveling procedure */
943 reg &= ~(1 << REG_DRAM_TRAINING_RL_OFFS);
944 reg_write(REG_DRAM_TRAINING_ADDR, reg); /* 0x15B0 - Training Register */
945
946 /* Return to target frequency */
947 freq = dram_info->target_frequency;
948 tmp_ratio = 1;
949 if (MV_OK != ddr3_dfs_low_2_high(freq, tmp_ratio, dram_info)) {
950 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs Low2High)\n");
951 return MV_DDR3_TRAINING_ERR_DFS_H2L;
952 }
953
954 if (dram_info->ecc_ena) {
955 /* Scabbling the RL area pattern and the training area */
956 mv_sys_xor_finish();
957 dram_info->num_cs = 1;
958 dram_info->cs_ena = 1;
959 mv_sys_xor_init(dram_info);
960 mv_xor_mem_init(0, RESUME_RL_PATTERNS_ADDR,
961 RESUME_RL_PATTERNS_SIZE, 0xFFFFFFFF, 0xFFFFFFFF);
962
963 /* Wait for previous transfer completion */
964
965 while (mv_xor_state_get(0) != MV_IDLE)
966 ;
967
968 /* Return XOR State */
969 mv_sys_xor_finish();
970 }
971
972 return MV_OK;
973}
974#endif
975
976void ddr3_print_freq(u32 freq)
977{
978 u32 tmp_freq;
979
980 switch (freq) {
981 case 0:
982 tmp_freq = 100;
983 break;
984 case 1:
985 tmp_freq = 300;
986 break;
987 case 2:
988 tmp_freq = 360;
989 break;
990 case 3:
991 tmp_freq = 400;
992 break;
993 case 4:
994 tmp_freq = 444;
995 break;
996 case 5:
997 tmp_freq = 500;
998 break;
999 case 6:
1000 tmp_freq = 533;
1001 break;
1002 case 7:
1003 tmp_freq = 600;
1004 break;
1005 case 8:
1006 tmp_freq = 666;
1007 break;
1008 case 9:
1009 tmp_freq = 720;
1010 break;
1011 case 10:
1012 tmp_freq = 800;
1013 break;
1014 default:
1015 tmp_freq = 100;
1016 }
1017
1018 printf("Current frequency is: %dMHz\n", tmp_freq);
1019}
1020
1021int ddr3_get_min_max_read_sample_delay(u32 cs_enable, u32 reg, u32 *min,
1022 u32 *max, u32 *cs_max)
1023{
1024 u32 cs, delay;
1025
1026 *min = 0xFFFFFFFF;
1027 *max = 0x0;
1028
1029 for (cs = 0; cs < MAX_CS; cs++) {
1030 if ((cs_enable & (1 << cs)) == 0)
1031 continue;
1032
1033 delay = ((reg >> (cs * 8)) & 0x1F);
1034
1035 if (delay < *min)
1036 *min = delay;
1037
1038 if (delay > *max) {
1039 *max = delay;
1040 *cs_max = cs;
1041 }
1042 }
1043
1044 return MV_OK;
1045}
1046
1047int ddr3_get_min_max_rl_phase(MV_DRAM_INFO *dram_info, u32 *min, u32 *max,
1048 u32 cs)
1049{
1050 u32 pup, reg, phase;
1051
1052 *min = 0xFFFFFFFF;
1053 *max = 0x0;
1054
1055 for (pup = 0; pup < dram_info->num_of_total_pups; pup++) {
1056 reg = ddr3_read_pup_reg(PUP_RL_MODE, cs, pup);
1057 phase = ((reg >> 8) & 0x7);
1058
1059 if (phase < *min)
1060 *min = phase;
1061
1062 if (phase > *max)
1063 *max = phase;
1064 }
1065
1066 return MV_OK;
1067}
1068
1069int ddr3_odt_activate(int activate)
1070{
1071 u32 reg, mask;
1072
1073 mask = (1 << REG_DUNIT_ODT_CTRL_OVRD_OFFS) |
1074 (1 << REG_DUNIT_ODT_CTRL_OVRD_VAL_OFFS);
1075 /* {0x0000149C} - DDR Dunit ODT Control Register */
1076 reg = reg_read(REG_DUNIT_ODT_CTRL_ADDR);
1077 if (activate)
1078 reg |= mask;
1079 else
1080 reg &= ~mask;
1081
1082 reg_write(REG_DUNIT_ODT_CTRL_ADDR, reg);
1083
1084 return MV_OK;
1085}
1086
1087int ddr3_odt_read_dynamic_config(MV_DRAM_INFO *dram_info)
1088{
1089 u32 min_read_sample_delay, max_read_sample_delay, max_rl_phase;
1090 u32 min, max, cs_max;
1091 u32 cs_ena, reg;
1092
1093 reg = reg_read(REG_READ_DATA_SAMPLE_DELAYS_ADDR);
1094 cs_ena = ddr3_get_cs_ena_from_reg();
1095
1096 /* Get minimum and maximum of read sample delay of all CS */
1097 ddr3_get_min_max_read_sample_delay(cs_ena, reg, &min_read_sample_delay,
1098 &max_read_sample_delay, &cs_max);
1099
1100 /*
1101 * Get minimum and maximum read leveling phase which belongs to the
1102 * maximal read sample delay
1103 */
1104 ddr3_get_min_max_rl_phase(dram_info, &min, &max, cs_max);
1105 max_rl_phase = max;
1106
1107 /* DDR ODT Timing (Low) Register calculation */
1108 reg = reg_read(REG_ODT_TIME_LOW_ADDR);
1109 reg &= ~(0x1FF << REG_ODT_ON_CTL_RD_OFFS);
1110 reg |= (((min_read_sample_delay - 1) & 0xF) << REG_ODT_ON_CTL_RD_OFFS);
1111 reg |= (((max_read_sample_delay + 4 + (((max_rl_phase + 1) / 2) + 1)) &
1112 0x1F) << REG_ODT_OFF_CTL_RD_OFFS);
1113 reg_write(REG_ODT_TIME_LOW_ADDR, reg);
1114
1115 return MV_OK;
1116}