sdram_s10.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Copyright (C) 2016-2018 Intel Corporation <www.intel.com>
  4. *
  5. */
  6. #include <common.h>
  7. #include <cpu_func.h>
  8. #include <dm.h>
  9. #include <errno.h>
  10. #include <div64.h>
  11. #include <fdtdec.h>
  12. #include <hang.h>
  13. #include <init.h>
  14. #include <log.h>
  15. #include <ram.h>
  16. #include <reset.h>
  17. #include "sdram_s10.h"
  18. #include <wait_bit.h>
  19. #include <asm/arch/firewall.h>
  20. #include <asm/arch/reset_manager.h>
  21. #include <asm/io.h>
  22. #include <linux/sizes.h>
  23. DECLARE_GLOBAL_DATA_PTR;
  24. #define DDR_CONFIG(A, B, C, R) (((A) << 24) | ((B) << 16) | ((C) << 8) | (R))
  25. /* The followring are the supported configurations */
  26. u32 ddr_config[] = {
  27. /* DDR_CONFIG(Address order,Bank,Column,Row) */
  28. /* List for DDR3 or LPDDR3 (pinout order > chip, row, bank, column) */
  29. DDR_CONFIG(0, 3, 10, 12),
  30. DDR_CONFIG(0, 3, 9, 13),
  31. DDR_CONFIG(0, 3, 10, 13),
  32. DDR_CONFIG(0, 3, 9, 14),
  33. DDR_CONFIG(0, 3, 10, 14),
  34. DDR_CONFIG(0, 3, 10, 15),
  35. DDR_CONFIG(0, 3, 11, 14),
  36. DDR_CONFIG(0, 3, 11, 15),
  37. DDR_CONFIG(0, 3, 10, 16),
  38. DDR_CONFIG(0, 3, 11, 16),
  39. DDR_CONFIG(0, 3, 12, 15), /* 0xa */
  40. /* List for DDR4 only (pinout order > chip, bank, row, column) */
  41. DDR_CONFIG(1, 3, 10, 14),
  42. DDR_CONFIG(1, 4, 10, 14),
  43. DDR_CONFIG(1, 3, 10, 15),
  44. DDR_CONFIG(1, 4, 10, 15),
  45. DDR_CONFIG(1, 3, 10, 16),
  46. DDR_CONFIG(1, 4, 10, 16),
  47. DDR_CONFIG(1, 3, 10, 17),
  48. DDR_CONFIG(1, 4, 10, 17),
  49. };
  50. int match_ddr_conf(u32 ddr_conf)
  51. {
  52. int i;
  53. for (i = 0; i < ARRAY_SIZE(ddr_config); i++) {
  54. if (ddr_conf == ddr_config[i])
  55. return i;
  56. }
  57. return 0;
  58. }
  59. /**
  60. * sdram_mmr_init_full() - Function to initialize SDRAM MMR
  61. *
  62. * Initialize the SDRAM MMR.
  63. */
  64. int sdram_mmr_init_full(struct udevice *dev)
  65. {
  66. struct altera_sdram_plat *plat = dev->plat;
  67. struct altera_sdram_priv *priv = dev_get_priv(dev);
  68. u32 update_value, io48_value, ddrioctl;
  69. u32 i;
  70. int ret;
  71. phys_size_t hw_size;
  72. struct bd_info bd = {0};
  73. /* Enable access to DDR from CPU master */
  74. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_DDRREG),
  75. CCU_ADBASE_DI_MASK);
  76. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_MEMSPACE0),
  77. CCU_ADBASE_DI_MASK);
  78. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_MEMSPACE1A),
  79. CCU_ADBASE_DI_MASK);
  80. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_MEMSPACE1B),
  81. CCU_ADBASE_DI_MASK);
  82. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_MEMSPACE1C),
  83. CCU_ADBASE_DI_MASK);
  84. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_MEMSPACE1D),
  85. CCU_ADBASE_DI_MASK);
  86. clrbits_le32(CCU_REG_ADDR(CCU_CPU0_MPRT_ADBASE_MEMSPACE1E),
  87. CCU_ADBASE_DI_MASK);
  88. /* Enable access to DDR from IO master */
  89. clrbits_le32(CCU_REG_ADDR(CCU_IOM_MPRT_ADBASE_MEMSPACE0),
  90. CCU_ADBASE_DI_MASK);
  91. clrbits_le32(CCU_REG_ADDR(CCU_IOM_MPRT_ADBASE_MEMSPACE1A),
  92. CCU_ADBASE_DI_MASK);
  93. clrbits_le32(CCU_REG_ADDR(CCU_IOM_MPRT_ADBASE_MEMSPACE1B),
  94. CCU_ADBASE_DI_MASK);
  95. clrbits_le32(CCU_REG_ADDR(CCU_IOM_MPRT_ADBASE_MEMSPACE1C),
  96. CCU_ADBASE_DI_MASK);
  97. clrbits_le32(CCU_REG_ADDR(CCU_IOM_MPRT_ADBASE_MEMSPACE1D),
  98. CCU_ADBASE_DI_MASK);
  99. clrbits_le32(CCU_REG_ADDR(CCU_IOM_MPRT_ADBASE_MEMSPACE1E),
  100. CCU_ADBASE_DI_MASK);
  101. /* Enable access to DDR from TCU */
  102. clrbits_le32(CCU_REG_ADDR(CCU_TCU_MPRT_ADBASE_MEMSPACE0),
  103. CCU_ADBASE_DI_MASK);
  104. clrbits_le32(CCU_REG_ADDR(CCU_TCU_MPRT_ADBASE_MEMSPACE1A),
  105. CCU_ADBASE_DI_MASK);
  106. clrbits_le32(CCU_REG_ADDR(CCU_TCU_MPRT_ADBASE_MEMSPACE1B),
  107. CCU_ADBASE_DI_MASK);
  108. clrbits_le32(CCU_REG_ADDR(CCU_TCU_MPRT_ADBASE_MEMSPACE1C),
  109. CCU_ADBASE_DI_MASK);
  110. clrbits_le32(CCU_REG_ADDR(CCU_TCU_MPRT_ADBASE_MEMSPACE1D),
  111. CCU_ADBASE_DI_MASK);
  112. clrbits_le32(CCU_REG_ADDR(CCU_TCU_MPRT_ADBASE_MEMSPACE1E),
  113. CCU_ADBASE_DI_MASK);
  114. /* this enables nonsecure access to DDR */
  115. /* mpuregion0addr_limit */
  116. FW_MPU_DDR_SCR_WRITEL(0xFFFF0000, FW_MPU_DDR_SCR_MPUREGION0ADDR_LIMIT);
  117. FW_MPU_DDR_SCR_WRITEL(0x1F, FW_MPU_DDR_SCR_MPUREGION0ADDR_LIMITEXT);
  118. /* nonmpuregion0addr_limit */
  119. FW_MPU_DDR_SCR_WRITEL(0xFFFF0000,
  120. FW_MPU_DDR_SCR_NONMPUREGION0ADDR_LIMIT);
  121. FW_MPU_DDR_SCR_WRITEL(0x1F, FW_MPU_DDR_SCR_NONMPUREGION0ADDR_LIMITEXT);
  122. /* Enable mpuregion0enable and nonmpuregion0enable */
  123. FW_MPU_DDR_SCR_WRITEL(MPUREGION0_ENABLE | NONMPUREGION0_ENABLE,
  124. FW_MPU_DDR_SCR_EN_SET);
  125. /* Ensure HMC clock is running */
  126. if (poll_hmc_clock_status()) {
  127. puts("DDR: Error as HMC clock not running\n");
  128. return -1;
  129. }
  130. /* Try 3 times to do a calibration */
  131. for (i = 0; i < 3; i++) {
  132. ret = wait_for_bit_le32((const void *)(plat->hmc +
  133. DDRCALSTAT),
  134. DDR_HMC_DDRCALSTAT_CAL_MSK, true, 1000,
  135. false);
  136. if (!ret)
  137. break;
  138. emif_reset(plat);
  139. }
  140. if (ret) {
  141. puts("DDR: Error as SDRAM calibration failed\n");
  142. return -1;
  143. }
  144. debug("DDR: Calibration success\n");
  145. u32 ctrlcfg0 = hmc_readl(plat, CTRLCFG0);
  146. u32 ctrlcfg1 = hmc_readl(plat, CTRLCFG1);
  147. u32 dramaddrw = hmc_readl(plat, DRAMADDRW);
  148. u32 dramtim0 = hmc_readl(plat, DRAMTIMING0);
  149. u32 caltim0 = hmc_readl(plat, CALTIMING0);
  150. u32 caltim1 = hmc_readl(plat, CALTIMING1);
  151. u32 caltim2 = hmc_readl(plat, CALTIMING2);
  152. u32 caltim3 = hmc_readl(plat, CALTIMING3);
  153. u32 caltim4 = hmc_readl(plat, CALTIMING4);
  154. u32 caltim9 = hmc_readl(plat, CALTIMING9);
  155. /*
  156. * Configure the DDR IO size [0xFFCFB008]
  157. * niosreserve0: Used to indicate DDR width &
  158. * bit[7:0] = Number of data bits (bit[6:5] 0x01=32bit, 0x10=64bit)
  159. * bit[8] = 1 if user-mode OCT is present
  160. * bit[9] = 1 if warm reset compiled into EMIF Cal Code
  161. * bit[10] = 1 if warm reset is on during generation in EMIF Cal
  162. * niosreserve1: IP ADCDS version encoded as 16 bit value
  163. * bit[2:0] = Variant (0=not special,1=FAE beta, 2=Customer beta,
  164. * 3=EAP, 4-6 are reserved)
  165. * bit[5:3] = Service Pack # (e.g. 1)
  166. * bit[9:6] = Minor Release #
  167. * bit[14:10] = Major Release #
  168. */
  169. update_value = hmc_readl(plat, NIOSRESERVED0);
  170. hmc_ecc_writel(plat, ((update_value & 0xFF) >> 5), DDRIOCTRL);
  171. ddrioctl = hmc_ecc_readl(plat, DDRIOCTRL);
  172. /* enable HPS interface to HMC */
  173. hmc_ecc_writel(plat, DDR_HMC_HPSINTFCSEL_ENABLE_MASK, HPSINTFCSEL);
  174. /* Set the DDR Configuration */
  175. io48_value = DDR_CONFIG(CTRLCFG1_CFG_ADDR_ORDER(ctrlcfg1),
  176. (DRAMADDRW_CFG_BANK_ADDR_WIDTH(dramaddrw) +
  177. DRAMADDRW_CFG_BANK_GRP_ADDR_WIDTH(dramaddrw)),
  178. DRAMADDRW_CFG_COL_ADDR_WIDTH(dramaddrw),
  179. DRAMADDRW_CFG_ROW_ADDR_WIDTH(dramaddrw));
  180. update_value = match_ddr_conf(io48_value);
  181. if (update_value)
  182. ddr_sch_writel(plat, update_value, DDR_SCH_DDRCONF);
  183. /* Configure HMC dramaddrw */
  184. hmc_ecc_writel(plat, hmc_readl(plat, DRAMADDRW), DRAMADDRWIDTH);
  185. /*
  186. * Configure DDR timing
  187. * RDTOMISS = tRTP + tRP + tRCD - BL/2
  188. * WRTOMISS = WL + tWR + tRP + tRCD and
  189. * WL = RL + BL/2 + 2 - rd-to-wr ; tWR = 15ns so...
  190. * First part of equation is in memory clock units so divide by 2
  191. * for HMC clock units. 1066MHz is close to 1ns so use 15 directly.
  192. * WRTOMISS = ((RL + BL/2 + 2 + tWR) >> 1)- rd-to-wr + tRP + tRCD
  193. */
  194. u32 burst_len = CTRLCFG0_CFG_CTRL_BURST_LEN(ctrlcfg0);
  195. update_value = CALTIMING2_CFG_RD_TO_WR_PCH(caltim2) +
  196. CALTIMING4_CFG_PCH_TO_VALID(caltim4) +
  197. CALTIMING0_CFG_ACT_TO_RDWR(caltim0) -
  198. (burst_len >> 2);
  199. io48_value = (((DRAMTIMING0_CFG_TCL(dramtim0) + 2 + DDR_TWR +
  200. (burst_len >> 1)) >> 1) -
  201. /* Up to here was in memory cycles so divide by 2 */
  202. CALTIMING1_CFG_RD_TO_WR(caltim1) +
  203. CALTIMING0_CFG_ACT_TO_RDWR(caltim0) +
  204. CALTIMING4_CFG_PCH_TO_VALID(caltim4));
  205. ddr_sch_writel(plat, ((CALTIMING0_CFG_ACT_TO_ACT(caltim0) <<
  206. DDR_SCH_DDRTIMING_ACTTOACT_OFF) |
  207. (update_value << DDR_SCH_DDRTIMING_RDTOMISS_OFF) |
  208. (io48_value << DDR_SCH_DDRTIMING_WRTOMISS_OFF) |
  209. ((burst_len >> 2) << DDR_SCH_DDRTIMING_BURSTLEN_OFF) |
  210. (CALTIMING1_CFG_RD_TO_WR(caltim1) <<
  211. DDR_SCH_DDRTIMING_RDTOWR_OFF) |
  212. (CALTIMING3_CFG_WR_TO_RD(caltim3) <<
  213. DDR_SCH_DDRTIMING_WRTORD_OFF) |
  214. (((ddrioctl == 1) ? 1 : 0) <<
  215. DDR_SCH_DDRTIMING_BWRATIO_OFF)),
  216. DDR_SCH_DDRTIMING);
  217. /* Configure DDR mode [precharge = 0] */
  218. ddr_sch_writel(plat, ((ddrioctl ? 0 : 1) <<
  219. DDR_SCH_DDRMOD_BWRATIOEXTENDED_OFF),
  220. DDR_SCH_DDRMODE);
  221. /* Configure the read latency */
  222. ddr_sch_writel(plat, (DRAMTIMING0_CFG_TCL(dramtim0) >> 1) +
  223. DDR_READ_LATENCY_DELAY,
  224. DDR_SCH_READ_LATENCY);
  225. /*
  226. * Configuring timing values concerning activate commands
  227. * [FAWBANK alway 1 because always 4 bank DDR]
  228. */
  229. ddr_sch_writel(plat, ((CALTIMING0_CFG_ACT_TO_ACT_DB(caltim0) <<
  230. DDR_SCH_ACTIVATE_RRD_OFF) |
  231. (CALTIMING9_CFG_4_ACT_TO_ACT(caltim9) <<
  232. DDR_SCH_ACTIVATE_FAW_OFF) |
  233. (DDR_ACTIVATE_FAWBANK <<
  234. DDR_SCH_ACTIVATE_FAWBANK_OFF)),
  235. DDR_SCH_ACTIVATE);
  236. /*
  237. * Configuring timing values concerning device to device data bus
  238. * ownership change
  239. */
  240. ddr_sch_writel(plat, ((CALTIMING1_CFG_RD_TO_RD_DC(caltim1) <<
  241. DDR_SCH_DEVTODEV_BUSRDTORD_OFF) |
  242. (CALTIMING1_CFG_RD_TO_WR_DC(caltim1) <<
  243. DDR_SCH_DEVTODEV_BUSRDTOWR_OFF) |
  244. (CALTIMING3_CFG_WR_TO_RD_DC(caltim3) <<
  245. DDR_SCH_DEVTODEV_BUSWRTORD_OFF)),
  246. DDR_SCH_DEVTODEV);
  247. /* assigning the SDRAM size */
  248. unsigned long long size = sdram_calculate_size(plat);
  249. /* If the size is invalid, use default Config size */
  250. if (size <= 0)
  251. hw_size = PHYS_SDRAM_1_SIZE;
  252. else
  253. hw_size = size;
  254. /* Get bank configuration from devicetree */
  255. ret = fdtdec_decode_ram_size(gd->fdt_blob, NULL, 0, NULL,
  256. (phys_size_t *)&gd->ram_size, &bd);
  257. if (ret) {
  258. puts("DDR: Failed to decode memory node\n");
  259. return -1;
  260. }
  261. if (gd->ram_size != hw_size)
  262. printf("DDR: Warning: DRAM size from device tree mismatch with hardware.\n");
  263. printf("DDR: %lld MiB\n", gd->ram_size >> 20);
  264. /* Enable or disable the SDRAM ECC */
  265. if (CTRLCFG1_CFG_CTRL_EN_ECC(ctrlcfg1)) {
  266. setbits_le32(plat->hmc + ECCCTRL1,
  267. (DDR_HMC_ECCCTL_AWB_CNT_RST_SET_MSK |
  268. DDR_HMC_ECCCTL_CNT_RST_SET_MSK |
  269. DDR_HMC_ECCCTL_ECC_EN_SET_MSK));
  270. clrbits_le32(plat->hmc + ECCCTRL1,
  271. (DDR_HMC_ECCCTL_AWB_CNT_RST_SET_MSK |
  272. DDR_HMC_ECCCTL_CNT_RST_SET_MSK));
  273. setbits_le32(plat->hmc + ECCCTRL2,
  274. (DDR_HMC_ECCCTL2_RMW_EN_SET_MSK |
  275. DDR_HMC_ECCCTL2_AWB_EN_SET_MSK));
  276. hmc_ecc_writel(plat, DDR_HMC_ERRINTEN_INTMASK, ERRINTENS);
  277. /* Initialize memory content if not from warm reset */
  278. if (!cpu_has_been_warmreset())
  279. sdram_init_ecc_bits(&bd);
  280. } else {
  281. clrbits_le32(plat->hmc + ECCCTRL1,
  282. (DDR_HMC_ECCCTL_AWB_CNT_RST_SET_MSK |
  283. DDR_HMC_ECCCTL_CNT_RST_SET_MSK |
  284. DDR_HMC_ECCCTL_ECC_EN_SET_MSK));
  285. clrbits_le32(plat->hmc + ECCCTRL2,
  286. (DDR_HMC_ECCCTL2_RMW_EN_SET_MSK |
  287. DDR_HMC_ECCCTL2_AWB_EN_SET_MSK));
  288. }
  289. /* Enable non-secure reads/writes to HMC Adapter for SDRAM ECC */
  290. writel(FW_HMC_ADAPTOR_MPU_MASK, FW_HMC_ADAPTOR_REG_ADDR);
  291. sdram_size_check(&bd);
  292. priv->info.base = bd.bi_dram[0].start;
  293. priv->info.size = gd->ram_size;
  294. debug("DDR: HMC init success\n");
  295. return 0;
  296. }