blob: 17c2bbb2a1731b008f154d6aff5f4f13742dc1fc [file] [log] [blame]
Pankaj Guptac518de42020-12-09 14:02:39 +05301/*
2 * Copyright 2021 NXP
3 *
4 * SPDX-License-Identifier: BSD-3-Clause
5 */
6
7#include <errno.h>
Scott Brandene5dcf982020-08-25 13:49:32 -07008#include <inttypes.h>
Pankaj Guptac518de42020-12-09 14:02:39 +05309#include <stdint.h>
10#include <stdio.h>
11#include <stdlib.h>
12#include <string.h>
13
14#include <common/debug.h>
15#include <ddr.h>
16#ifndef CONFIG_DDR_NODIMM
17#include <i2c.h>
18#endif
19#include <nxp_timer.h>
20
21struct dynamic_odt {
22 unsigned int odt_rd_cfg;
23 unsigned int odt_wr_cfg;
24 unsigned int odt_rtt_norm;
25 unsigned int odt_rtt_wr;
26};
27
28#ifndef CONFIG_STATIC_DDR
29#if defined(PHY_GEN2_FW_IMAGE_BUFFER) && !defined(NXP_DDR_PHY_GEN2)
30#error Missing NXP_DDR_PHY_GEN2
31#endif
32#ifdef NXP_DDR_PHY_GEN2
33static const struct dynamic_odt single_D[4] = {
34 { /* cs0 */
35 DDR_ODT_NEVER,
36 DDR_ODT_ALL,
37 DDR4_RTT_80_OHM,
38 DDR4_RTT_WR_OFF
39 },
40 { /* cs1 */
41 DDR_ODT_NEVER,
42 DDR_ODT_NEVER,
43 DDR4_RTT_OFF,
44 DDR4_RTT_WR_OFF
45 },
46 {},
47 {}
48};
49
50static const struct dynamic_odt single_S[4] = {
51 { /* cs0 */
52 DDR_ODT_NEVER,
53 DDR_ODT_ALL,
54 DDR4_RTT_80_OHM,
55 DDR4_RTT_WR_OFF
56 },
57 {},
58 {},
59 {},
60};
61
62static const struct dynamic_odt dual_DD[4] = {
63 { /* cs0 */
64 DDR_ODT_OTHER_DIMM,
65 DDR_ODT_ALL,
66 DDR4_RTT_60_OHM,
67 DDR4_RTT_WR_240_OHM
68 },
69 { /* cs1 */
70 DDR_ODT_OTHER_DIMM,
71 DDR_ODT_ALL,
72 DDR4_RTT_60_OHM,
73 DDR4_RTT_WR_240_OHM
74 },
75 { /* cs2 */
76 DDR_ODT_OTHER_DIMM,
77 DDR_ODT_ALL,
78 DDR4_RTT_60_OHM,
79 DDR4_RTT_WR_240_OHM
80 },
81 { /* cs3 */
82 DDR_ODT_OTHER_DIMM,
83 DDR_ODT_ALL,
84 DDR4_RTT_60_OHM,
85 DDR4_RTT_WR_240_OHM
86 }
87};
88
89static const struct dynamic_odt dual_SS[4] = {
90 { /* cs0 */
91 DDR_ODT_NEVER,
92 DDR_ODT_ALL,
93 DDR4_RTT_80_OHM,
94 DDR4_RTT_WR_OFF
95 },
96 {},
97 { /* cs2 */
98 DDR_ODT_NEVER,
99 DDR_ODT_ALL,
100 DDR4_RTT_80_OHM,
101 DDR4_RTT_WR_OFF
102 },
103 {}
104};
105
106static const struct dynamic_odt dual_D0[4] = {
107 { /* cs0 */
108 DDR_ODT_NEVER,
109 DDR_ODT_SAME_DIMM,
110 DDR4_RTT_80_OHM,
111 DDR4_RTT_WR_OFF
112 },
113 { /* cs1 */
114 DDR_ODT_NEVER,
115 DDR_ODT_NEVER,
116 DDR4_RTT_80_OHM,
117 DDR4_RTT_WR_OFF
118 },
119 {},
120 {}
121};
122
123static const struct dynamic_odt dual_S0[4] = {
124 { /* cs0 */
125 DDR_ODT_NEVER,
126 DDR_ODT_CS,
127 DDR4_RTT_80_OHM,
128 DDR4_RTT_WR_OFF
129 },
130 {},
131 {},
132 {}
133};
134#else
135static const struct dynamic_odt single_D[4] = {
136 { /* cs0 */
137 DDR_ODT_NEVER,
138 DDR_ODT_ALL,
139 DDR4_RTT_40_OHM,
140 DDR4_RTT_WR_OFF
141 },
142 { /* cs1 */
143 DDR_ODT_NEVER,
144 DDR_ODT_NEVER,
145 DDR4_RTT_OFF,
146 DDR4_RTT_WR_OFF
147 },
148 {},
149 {}
150};
151
152static const struct dynamic_odt single_S[4] = {
153 { /* cs0 */
154 DDR_ODT_NEVER,
155 DDR_ODT_ALL,
156 DDR4_RTT_40_OHM,
157 DDR4_RTT_WR_OFF
158 },
159 {},
160 {},
161 {},
162};
163
164static const struct dynamic_odt dual_DD[4] = {
165 { /* cs0 */
166 DDR_ODT_NEVER,
167 DDR_ODT_SAME_DIMM,
168 DDR4_RTT_120_OHM,
169 DDR4_RTT_WR_OFF
170 },
171 { /* cs1 */
172 DDR_ODT_OTHER_DIMM,
173 DDR_ODT_OTHER_DIMM,
174 DDR4_RTT_34_OHM,
175 DDR4_RTT_WR_OFF
176 },
177 { /* cs2 */
178 DDR_ODT_NEVER,
179 DDR_ODT_SAME_DIMM,
180 DDR4_RTT_120_OHM,
181 DDR4_RTT_WR_OFF
182 },
183 { /* cs3 */
184 DDR_ODT_OTHER_DIMM,
185 DDR_ODT_OTHER_DIMM,
186 DDR4_RTT_34_OHM,
187 DDR4_RTT_WR_OFF
188 }
189};
190
191static const struct dynamic_odt dual_SS[4] = {
192 { /* cs0 */
193 DDR_ODT_OTHER_DIMM,
194 DDR_ODT_ALL,
195 DDR4_RTT_34_OHM,
196 DDR4_RTT_WR_120_OHM
197 },
198 {},
199 { /* cs2 */
200 DDR_ODT_OTHER_DIMM,
201 DDR_ODT_ALL,
202 DDR4_RTT_34_OHM,
203 DDR4_RTT_WR_120_OHM
204 },
205 {}
206};
207
208static const struct dynamic_odt dual_D0[4] = {
209 { /* cs0 */
210 DDR_ODT_NEVER,
211 DDR_ODT_SAME_DIMM,
212 DDR4_RTT_40_OHM,
213 DDR4_RTT_WR_OFF
214 },
215 { /* cs1 */
216 DDR_ODT_NEVER,
217 DDR_ODT_NEVER,
218 DDR4_RTT_OFF,
219 DDR4_RTT_WR_OFF
220 },
221 {},
222 {}
223};
224
225static const struct dynamic_odt dual_S0[4] = {
226 { /* cs0 */
227 DDR_ODT_NEVER,
228 DDR_ODT_CS,
229 DDR4_RTT_40_OHM,
230 DDR4_RTT_WR_OFF
231 },
232 {},
233 {},
234 {}
235};
236#endif /* NXP_DDR_PHY_GEN2 */
237
238/*
239 * Automatically select bank interleaving mode based on DIMMs
240 * in this order: cs0_cs1_cs2_cs3, cs0_cs1, null.
241 * This function only deal with one or two slots per controller.
242 */
243static inline unsigned int auto_bank_intlv(const int cs_in_use,
244 const struct dimm_params *pdimm)
245{
246 switch (cs_in_use) {
247 case 0xf:
248 return DDR_BA_INTLV_CS0123;
249 case 0x3:
250 return DDR_BA_INTLV_CS01;
251 case 0x1:
252 return DDR_BA_NONE;
253 case 0x5:
254 return DDR_BA_NONE;
255 default:
256 break;
257 }
258
259 return 0U;
260}
261
262static int cal_odt(const unsigned int clk,
263 struct memctl_opt *popts,
264 struct ddr_conf *conf,
265 struct dimm_params *pdimm,
266 const int dimm_slot_per_ctrl)
267
268{
269 unsigned int i;
270 const struct dynamic_odt *pdodt = NULL;
271
Boyan Karatotev05e9d4d2022-11-22 14:31:41 +0000272 static const struct dynamic_odt *table[2][5] = {
Pankaj Guptac518de42020-12-09 14:02:39 +0530273 {single_S, single_D, NULL, NULL},
274 {dual_SS, dual_DD, NULL, NULL},
275 };
276
277 if (dimm_slot_per_ctrl != 1 && dimm_slot_per_ctrl != 2) {
278 ERROR("Unsupported number of DIMMs\n");
279 return -EINVAL;
280 }
281
282 pdodt = table[dimm_slot_per_ctrl - 1][pdimm->n_ranks - 1];
283 if (pdodt == dual_SS) {
284 pdodt = (conf->cs_in_use == 0x5) ? dual_SS :
285 ((conf->cs_in_use == 0x1) ? dual_S0 : NULL);
286 } else if (pdodt == dual_DD) {
287 pdodt = (conf->cs_in_use == 0xf) ? dual_DD :
288 ((conf->cs_in_use == 0x3) ? dual_D0 : NULL);
289 }
290 if (pdodt == dual_DD && pdimm->package_3ds) {
291 ERROR("Too many 3DS DIMMs.\n");
292 return -EINVAL;
293 }
294
295 if (pdodt == NULL) {
Elyes Haouas2be03c02023-02-13 09:14:48 +0100296 ERROR("Error determining ODT.\n");
Pankaj Guptac518de42020-12-09 14:02:39 +0530297 return -EINVAL;
298 }
299
300 /* Pick chip-select local options. */
301 for (i = 0U; i < DDRC_NUM_CS; i++) {
302 debug("cs %d\n", i);
303 popts->cs_odt[i].odt_rd_cfg = pdodt[i].odt_rd_cfg;
304 debug(" odt_rd_cfg 0x%x\n",
305 popts->cs_odt[i].odt_rd_cfg);
306 popts->cs_odt[i].odt_wr_cfg = pdodt[i].odt_wr_cfg;
307 debug(" odt_wr_cfg 0x%x\n",
308 popts->cs_odt[i].odt_wr_cfg);
309 popts->cs_odt[i].odt_rtt_norm = pdodt[i].odt_rtt_norm;
310 debug(" odt_rtt_norm 0x%x\n",
311 popts->cs_odt[i].odt_rtt_norm);
312 popts->cs_odt[i].odt_rtt_wr = pdodt[i].odt_rtt_wr;
313 debug(" odt_rtt_wr 0x%x\n",
314 popts->cs_odt[i].odt_rtt_wr);
315 popts->cs_odt[i].auto_precharge = 0;
316 debug(" auto_precharge %d\n",
317 popts->cs_odt[i].auto_precharge);
318 }
319
320 return 0;
321}
322
323static int cal_opts(const unsigned int clk,
324 struct memctl_opt *popts,
325 struct ddr_conf *conf,
326 struct dimm_params *pdimm,
327 const int dimm_slot_per_ctrl,
328 const unsigned int ip_rev)
329{
330 popts->rdimm = pdimm->rdimm;
331 popts->mirrored_dimm = pdimm->mirrored_dimm;
332#ifdef CONFIG_DDR_ECC_EN
333 popts->ecc_mode = pdimm->edc_config == 0x02 ? 1 : 0;
334#endif
335 popts->ctlr_init_ecc = popts->ecc_mode;
336 debug("ctlr_init_ecc %d\n", popts->ctlr_init_ecc);
337 popts->self_refresh_in_sleep = 1;
338 popts->dynamic_power = 0;
339
340 /*
341 * check sdram width, allow platform override
342 * 0 = 64-bit, 1 = 32-bit, 2 = 16-bit
343 */
344 if (pdimm->primary_sdram_width == 64) {
345 popts->data_bus_dimm = DDR_DBUS_64;
346 popts->otf_burst_chop_en = 1;
347 } else if (pdimm->primary_sdram_width == 32) {
348 popts->data_bus_dimm = DDR_DBUS_32;
349 popts->otf_burst_chop_en = 0;
350 } else if (pdimm->primary_sdram_width == 16) {
351 popts->data_bus_dimm = DDR_DBUS_16;
352 popts->otf_burst_chop_en = 0;
353 } else {
354 ERROR("primary sdram width invalid!\n");
355 return -EINVAL;
356 }
357 popts->data_bus_used = popts->data_bus_dimm;
358 popts->x4_en = (pdimm->device_width == 4) ? 1 : 0;
359 debug("x4_en %d\n", popts->x4_en);
360
361 /* for RDIMM and DDR4 UDIMM/discrete memory, address parity enable */
362 if (popts->rdimm != 0) {
363 popts->ap_en = 1; /* 0 = disable, 1 = enable */
364 } else {
365 popts->ap_en = 0; /* disabled for DDR4 UDIMM/discrete default */
366 }
367
368 if (ip_rev == 0x50500) {
369 popts->ap_en = 0;
370 }
371
372 debug("ap_en %d\n", popts->ap_en);
373
374 /* BSTTOPRE precharge interval uses 1/4 of refint value. */
375 popts->bstopre = picos_to_mclk(clk, pdimm->refresh_rate_ps) >> 2;
376 popts->tfaw_ps = pdimm->tfaw_ps;
377
378 return 0;
379}
380
381static void cal_intlv(const int num_ctlrs,
382 struct memctl_opt *popts,
383 struct ddr_conf *conf,
384 struct dimm_params *pdimm)
385{
386#ifdef NXP_DDR_INTLV_256B
387 if (num_ctlrs == 2) {
388 popts->ctlr_intlv = 1;
389 popts->ctlr_intlv_mode = DDR_256B_INTLV;
390 }
391#endif
392 debug("ctlr_intlv %d\n", popts->ctlr_intlv);
393 debug("ctlr_intlv_mode %d\n", popts->ctlr_intlv_mode);
394
395 popts->ba_intlv = auto_bank_intlv(conf->cs_in_use, pdimm);
396 debug("ba_intlv 0x%x\n", popts->ba_intlv);
397}
398
399static int update_burst_length(struct memctl_opt *popts)
400{
401 /* Choose burst length. */
402 if ((popts->data_bus_used == DDR_DBUS_32) ||
403 (popts->data_bus_used == DDR_DBUS_16)) {
404 /* 32-bit or 16-bit bus */
405 popts->otf_burst_chop_en = 0;
406 popts->burst_length = DDR_BL8;
407 } else if (popts->otf_burst_chop_en != 0) { /* on-the-fly burst chop */
408 popts->burst_length = DDR_OTF; /* on-the-fly BC4 and BL8 */
409 } else {
410 popts->burst_length = DDR_BL8;
411 }
412 debug("data_bus_used %d\n", popts->data_bus_used);
413 debug("otf_burst_chop_en %d\n", popts->otf_burst_chop_en);
414 debug("burst_length 0x%x\n", popts->burst_length);
415 /*
416 * If a reduced data width is requested, but the SPD
417 * specifies a physically wider device, adjust the
418 * computed dimm capacities accordingly before
419 * assigning addresses.
420 * 0 = 64-bit, 1 = 32-bit, 2 = 16-bit
421 */
422 if (popts->data_bus_dimm > popts->data_bus_used) {
423 ERROR("Data bus configuration error\n");
424 return -EINVAL;
425 }
426 popts->dbw_cap_shift = popts->data_bus_used - popts->data_bus_dimm;
427 debug("dbw_cap_shift %d\n", popts->dbw_cap_shift);
428
429 return 0;
430}
431
432int cal_board_params(struct ddr_info *priv,
433 const struct board_timing *dimm,
434 int len)
435{
436 const unsigned long speed = priv->clk / 1000000;
437 const struct dimm_params *pdimm = &priv->dimm;
438 struct memctl_opt *popts = &priv->opt;
439 struct rc_timing const *prt = NULL;
440 struct rc_timing const *chosen = NULL;
441 int i;
442
443 for (i = 0; i < len; i++) {
444 if (pdimm->rc == dimm[i].rc) {
445 prt = dimm[i].p;
446 break;
447 }
448 }
449 if (prt == NULL) {
450 ERROR("Board parameters no match.\n");
451 return -EINVAL;
452 }
453 while (prt->speed_bin != 0) {
454 if (speed <= prt->speed_bin) {
455 chosen = prt;
456 break;
457 }
458 prt++;
459 }
460 if (chosen == NULL) {
461 ERROR("timing no match for speed %lu\n", speed);
462 return -EINVAL;
463 }
464 popts->clk_adj = prt->clk_adj;
465 popts->wrlvl_start = prt->wrlvl;
466 popts->wrlvl_ctl_2 = (prt->wrlvl * 0x01010101 + dimm[i].add1) &
467 0xFFFFFFFF;
468 popts->wrlvl_ctl_3 = (prt->wrlvl * 0x01010101 + dimm[i].add2) &
469 0xFFFFFFFF;
470
471 return 0;
472}
473
474static int synthesize_ctlr(struct ddr_info *priv)
475{
476 int ret;
477
478 ret = cal_odt(priv->clk,
479 &priv->opt,
480 &priv->conf,
481 &priv->dimm,
482 priv->dimm_on_ctlr);
483 if (ret != 0) {
484 return ret;
485 }
486
487 ret = cal_opts(priv->clk,
488 &priv->opt,
489 &priv->conf,
490 &priv->dimm,
491 priv->dimm_on_ctlr,
492 priv->ip_rev);
493
494 if (ret != 0) {
495 return ret;
496 }
497
498 cal_intlv(priv->num_ctlrs, &priv->opt, &priv->conf, &priv->dimm);
499 ret = ddr_board_options(priv);
500 if (ret != 0) {
501 ERROR("Failed matching board timing.\n");
502 }
503
504 ret = update_burst_length(&priv->opt);
505
506 return ret;
507}
508
509/* Return the bit mask of valid DIMMs found */
510static int parse_spd(struct ddr_info *priv)
511{
512 struct ddr_conf *conf = &priv->conf;
513 struct dimm_params *dimm = &priv->dimm;
514 int j, valid_mask = 0;
515
516#ifdef CONFIG_DDR_NODIMM
517 valid_mask = ddr_get_ddr_params(dimm, conf);
518 if (valid_mask < 0) {
519 ERROR("DDR params error\n");
520 return valid_mask;
521 }
522#else
523 const int *spd_addr = priv->spd_addr;
524 const int num_ctlrs = priv->num_ctlrs;
525 const int num_dimm = priv->dimm_on_ctlr;
526 struct ddr4_spd spd[2];
527 unsigned int spd_checksum[2];
528 int addr_idx = 0;
529 int spd_idx = 0;
530 int ret, addr, i;
531
532 /* Scan all DIMMs */
533 for (i = 0; i < num_ctlrs; i++) {
534 debug("Controller %d\n", i);
535 for (j = 0; j < num_dimm; j++, addr_idx++) {
536 debug("DIMM %d\n", j);
537 addr = spd_addr[addr_idx];
538 if (addr == 0) {
539 if (j == 0) {
540 ERROR("First SPD addr wrong.\n");
541 return -EINVAL;
542 }
543 continue;
544 }
545 debug("addr 0x%x\n", addr);
546 ret = read_spd(addr, &spd[spd_idx],
547 sizeof(struct ddr4_spd));
548 if (ret != 0) { /* invalid */
549 debug("Invalid SPD at address 0x%x\n", addr);
550 continue;
551 }
552
553 spd_checksum[spd_idx] =
554 (spd[spd_idx].crc[1] << 24) |
555 (spd[spd_idx].crc[0] << 16) |
556 (spd[spd_idx].mod_section.uc[127] << 8) |
557 (spd[spd_idx].mod_section.uc[126] << 0);
558 debug("checksum 0x%x\n", spd_checksum[spd_idx]);
559 if (spd_checksum[spd_idx] == 0) {
560 debug("Bad checksum, ignored.\n");
561 continue;
562 }
563 if (spd_idx == 0) {
564 /* first valid SPD */
565 ret = cal_dimm_params(&spd[0], dimm);
566 if (ret != 0) {
567 ERROR("SPD calculation error\n");
568 return -EINVAL;
569 }
570 }
571
572 if (spd_idx != 0 && spd_checksum[0] !=
573 spd_checksum[spd_idx]) {
574 ERROR("Not identical DIMMs.\n");
575 return -EINVAL;
576 }
577 conf->dimm_in_use[j] = 1;
578 valid_mask |= 1 << addr_idx;
579 spd_idx = 1;
580 }
581 debug("done with controller %d\n", i);
582 }
583 switch (num_ctlrs) {
584 case 1:
585 if ((valid_mask & 0x1) == 0) {
586 ERROR("First slot cannot be empty.\n");
587 return -EINVAL;
588 }
589 break;
590 case 2:
591 switch (num_dimm) {
592 case 1:
593 if (valid_mask == 0) {
594 ERROR("Both slot empty\n");
595 return -EINVAL;
596 }
597 break;
598 case 2:
599 if (valid_mask != 0x5 &&
600 valid_mask != 0xf &&
601 (valid_mask & 0x7) != 0x4 &&
602 (valid_mask & 0xd) != 0x1) {
603 ERROR("Invalid DIMM combination.\n");
604 return -EINVAL;
605 }
606 break;
607 default:
608 ERROR("Invalid number of DIMMs.\n");
609 return -EINVAL;
610 }
611 break;
612 default:
613 ERROR("Invalid number of controllers.\n");
614 return -EINVAL;
615 }
616 /* now we have valid and identical DIMMs on controllers */
617#endif /* CONFIG_DDR_NODIMM */
618
619 debug("cal cs\n");
620 conf->cs_in_use = 0;
621 for (j = 0; j < DDRC_NUM_DIMM; j++) {
622 if (conf->dimm_in_use[j] == 0) {
623 continue;
624 }
625 switch (dimm->n_ranks) {
626 case 4:
627 ERROR("Quad-rank DIMM not supported\n");
628 return -EINVAL;
629 case 2:
630 conf->cs_on_dimm[j] = 0x3 << (j * CONFIG_CS_PER_SLOT);
631 conf->cs_in_use |= conf->cs_on_dimm[j];
632 break;
633 case 1:
634 conf->cs_on_dimm[j] = 0x1 << (j * CONFIG_CS_PER_SLOT);
635 conf->cs_in_use |= conf->cs_on_dimm[j];
636 break;
637 default:
638 ERROR("SPD error with n_ranks\n");
639 return -EINVAL;
640 }
641 debug("cs_in_use = %x\n", conf->cs_in_use);
642 debug("cs_on_dimm[%d] = %x\n", j, conf->cs_on_dimm[j]);
643 }
644#ifndef CONFIG_DDR_NODIMM
645 if (priv->dimm.rdimm != 0) {
646 NOTICE("RDIMM %s\n", priv->dimm.mpart);
647 } else {
648 NOTICE("UDIMM %s\n", priv->dimm.mpart);
649 }
650#else
651 NOTICE("%s\n", priv->dimm.mpart);
652#endif
653
654 return valid_mask;
655}
656
657static unsigned long long assign_intlv_addr(
658 const struct dimm_params *pdimm,
659 const struct memctl_opt *opt,
660 struct ddr_conf *conf,
661 const unsigned long long current_mem_base)
662{
663 int i;
664 int ctlr_density_mul = 0;
665 const unsigned long long rank_density = pdimm->rank_density >>
666 opt->dbw_cap_shift;
667 unsigned long long total_ctlr_mem;
668
669 debug("rank density 0x%llx\n", rank_density);
670 switch (opt->ba_intlv & DDR_BA_INTLV_CS0123) {
671 case DDR_BA_INTLV_CS0123:
672 ctlr_density_mul = 4;
673 break;
674 case DDR_BA_INTLV_CS01:
675 ctlr_density_mul = 2;
676 break;
677 default:
678 ctlr_density_mul = 1;
679 break;
680 }
681 debug("ctlr density mul %d\n", ctlr_density_mul);
682 switch (opt->ctlr_intlv_mode) {
683 case DDR_256B_INTLV:
684 total_ctlr_mem = 2 * ctlr_density_mul * rank_density;
685 break;
686 default:
687 ERROR("Unknown interleaving mode");
688 return 0;
689 }
690 conf->base_addr = current_mem_base;
691 conf->total_mem = total_ctlr_mem;
692
693 /* overwrite cs_in_use bitmask with controller interleaving */
694 conf->cs_in_use = (1 << ctlr_density_mul) - 1;
695 debug("Overwrite cs_in_use as %x\n", conf->cs_in_use);
696
697 /* Fill addr with each cs in use */
698 for (i = 0; i < ctlr_density_mul; i++) {
699 conf->cs_base_addr[i] = current_mem_base;
700 conf->cs_size[i] = total_ctlr_mem;
701 debug("CS %d\n", i);
702 debug(" base_addr 0x%llx\n", conf->cs_base_addr[i]);
703 debug(" size 0x%llx\n", conf->cs_size[i]);
704 }
705
706 return total_ctlr_mem;
707}
708
709static unsigned long long assign_non_intlv_addr(
710 const struct dimm_params *pdimm,
711 const struct memctl_opt *opt,
712 struct ddr_conf *conf,
713 unsigned long long current_mem_base)
714{
715 int i;
716 const unsigned long long rank_density = pdimm->rank_density >>
717 opt->dbw_cap_shift;
718 unsigned long long total_ctlr_mem = 0ULL;
719
720 debug("rank density 0x%llx\n", rank_density);
721 conf->base_addr = current_mem_base;
722
723 /* assign each cs */
724 switch (opt->ba_intlv & DDR_BA_INTLV_CS0123) {
725 case DDR_BA_INTLV_CS0123:
726 for (i = 0; i < DDRC_NUM_CS; i++) {
727 conf->cs_base_addr[i] = current_mem_base;
728 conf->cs_size[i] = rank_density << 2;
729 total_ctlr_mem += rank_density;
730 }
731 break;
732 case DDR_BA_INTLV_CS01:
733 for (i = 0; ((conf->cs_in_use & (1 << i)) != 0) && i < 2; i++) {
734 conf->cs_base_addr[i] = current_mem_base;
735 conf->cs_size[i] = rank_density << 1;
736 total_ctlr_mem += rank_density;
737 }
738 current_mem_base += total_ctlr_mem;
739 for (; ((conf->cs_in_use & (1 << i)) != 0) && i < DDRC_NUM_CS;
740 i++) {
741 conf->cs_base_addr[i] = current_mem_base;
742 conf->cs_size[i] = rank_density;
743 total_ctlr_mem += rank_density;
744 current_mem_base += rank_density;
745 }
746 break;
747 case DDR_BA_NONE:
748 for (i = 0; ((conf->cs_in_use & (1 << i)) != 0) &&
749 (i < DDRC_NUM_CS); i++) {
750 conf->cs_base_addr[i] = current_mem_base;
751 conf->cs_size[i] = rank_density;
752 current_mem_base += rank_density;
753 total_ctlr_mem += rank_density;
754 }
755 break;
756 default:
757 ERROR("Unsupported bank interleaving\n");
758 return 0;
759 }
760 for (i = 0; ((conf->cs_in_use & (1 << i)) != 0) &&
761 (i < DDRC_NUM_CS); i++) {
762 debug("CS %d\n", i);
763 debug(" base_addr 0x%llx\n", conf->cs_base_addr[i]);
764 debug(" size 0x%llx\n", conf->cs_size[i]);
765 }
766
767 return total_ctlr_mem;
768}
769
770unsigned long long assign_addresses(struct ddr_info *priv)
771 __attribute__ ((weak));
772
773unsigned long long assign_addresses(struct ddr_info *priv)
774{
775 struct memctl_opt *opt = &priv->opt;
776 const struct dimm_params *dimm = &priv->dimm;
777 struct ddr_conf *conf = &priv->conf;
778 unsigned long long current_mem_base = priv->mem_base;
779 unsigned long long total_mem;
780
781 total_mem = 0ULL;
782 debug("ctlr_intlv %d\n", opt->ctlr_intlv);
783 if (opt->ctlr_intlv != 0) {
784 total_mem = assign_intlv_addr(dimm, opt, conf,
785 current_mem_base);
786 } else {
787 /*
788 * Simple linear assignment if memory controllers are not
789 * interleaved. This is only valid for SoCs with single DDRC.
790 */
791 total_mem = assign_non_intlv_addr(dimm, opt, conf,
792 current_mem_base);
793 }
794 conf->total_mem = total_mem;
795 debug("base 0x%llx\n", current_mem_base);
796 debug("Total mem by assignment is 0x%llx\n", total_mem);
797
798 return total_mem;
799}
800
801static int cal_ddrc_regs(struct ddr_info *priv)
802{
803 int ret;
804
805 ret = compute_ddrc(priv->clk,
806 &priv->opt,
807 &priv->conf,
808 &priv->ddr_reg,
809 &priv->dimm,
810 priv->ip_rev);
811 if (ret != 0) {
812 ERROR("Calculating DDR registers failed\n");
813 }
814
815 return ret;
816}
817
818#endif /* CONFIG_STATIC_DDR */
819
820static int write_ddrc_regs(struct ddr_info *priv)
821{
822 int i;
823 int ret;
824
825 for (i = 0; i < priv->num_ctlrs; i++) {
826 ret = ddrc_set_regs(priv->clk, &priv->ddr_reg, priv->ddr[i], 0);
827 if (ret != 0) {
828 ERROR("Writing DDR register(s) failed\n");
829 return ret;
830 }
831 }
832
833 return 0;
834}
835
836long long dram_init(struct ddr_info *priv
837#if defined(NXP_HAS_CCN504) || defined(NXP_HAS_CCN508)
838 , uintptr_t nxp_ccn_hn_f0_addr
839#endif
840 )
841{
842 uint64_t time __unused;
843 long long dram_size;
844 int ret;
845 const uint64_t time_base = get_timer_val(0);
846 unsigned int ip_rev = get_ddrc_version(priv->ddr[0]);
847
848 int valid_spd_mask __unused;
849 int scratch = 0x0;
850
851 priv->ip_rev = ip_rev;
852
853#ifndef CONFIG_STATIC_DDR
Scott Brandene5dcf982020-08-25 13:49:32 -0700854 INFO("time base %" PRIu64 " ms\n", time_base);
Pankaj Guptac518de42020-12-09 14:02:39 +0530855 debug("Parse DIMM SPD(s)\n");
856 valid_spd_mask = parse_spd(priv);
857
858 if (valid_spd_mask < 0) {
859 ERROR("Parsing DIMM Error\n");
860 return valid_spd_mask;
861 }
862
863#if defined(NXP_HAS_CCN504) || defined(NXP_HAS_CCN508)
864 if (priv->num_ctlrs == 2 || priv->num_ctlrs == 1) {
865 ret = disable_unused_ddrc(priv, valid_spd_mask,
866 nxp_ccn_hn_f0_addr);
867 if (ret != 0) {
868 return ret;
869 }
870 }
871#endif
872
873 time = get_timer_val(time_base);
Scott Brandene5dcf982020-08-25 13:49:32 -0700874 INFO("Time after parsing SPD %" PRIu64 " ms\n", time);
Pankaj Guptac518de42020-12-09 14:02:39 +0530875 debug("Synthesize configurations\n");
876 ret = synthesize_ctlr(priv);
877 if (ret != 0) {
878 ERROR("Synthesize config error\n");
879 return ret;
880 }
881
882 debug("Assign binding addresses\n");
883 dram_size = assign_addresses(priv);
884 if (dram_size == 0) {
885 ERROR("Assigning address error\n");
886 return -EINVAL;
887 }
888
889 debug("Calculate controller registers\n");
890 ret = cal_ddrc_regs(priv);
891 if (ret != 0) {
892 ERROR("Calculate register error\n");
893 return ret;
894 }
895
896 ret = compute_ddr_phy(priv);
897 if (ret != 0)
898 ERROR("Calculating DDR PHY registers failed.\n");
899
900#else
901 dram_size = board_static_ddr(priv);
902 if (dram_size == 0) {
903 ERROR("Error getting static DDR settings.\n");
904 return -EINVAL;
905 }
906#endif
907
908 if (priv->warm_boot_flag == DDR_WARM_BOOT) {
909 scratch = (priv->ddr_reg).sdram_cfg[1];
910 scratch = scratch & ~(SDRAM_CFG2_D_INIT);
911 priv->ddr_reg.sdram_cfg[1] = scratch;
912 }
913
914 time = get_timer_val(time_base);
Scott Brandene5dcf982020-08-25 13:49:32 -0700915 INFO("Time before programming controller %" PRIu64 " ms\n", time);
Pankaj Guptac518de42020-12-09 14:02:39 +0530916 debug("Program controller registers\n");
917 ret = write_ddrc_regs(priv);
918 if (ret != 0) {
Elyes Haouas2be03c02023-02-13 09:14:48 +0100919 ERROR("Programming DDRC error\n");
Pankaj Guptac518de42020-12-09 14:02:39 +0530920 return ret;
921 }
922
923 puts("");
924 NOTICE("%lld GB ", dram_size >> 30);
925 print_ddr_info(priv->ddr[0]);
926
927 time = get_timer_val(time_base);
Scott Brandene5dcf982020-08-25 13:49:32 -0700928 INFO("Time used by DDR driver %" PRIu64 " ms\n", time);
Pankaj Guptac518de42020-12-09 14:02:39 +0530929
930 return dram_size;
931}