1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) Marvell International Ltd. and its affiliates
4  */
5 
6 #include "ddr3_init.h"
7 #include "mv_ddr_common.h"
8 #include "mv_ddr_training_db.h"
9 #include "mv_ddr_regs.h"
10 #include "mv_ddr_sys_env_lib.h"
11 
12 #define DDR_INTERFACES_NUM		1
13 #define DDR_INTERFACE_OCTETS_NUM	5
14 
15 /*
16  * 1. L2 filter should be set at binary header to 0xD000000,
17  *    to avoid conflict with internal register IO.
18  * 2. U-Boot modifies internal registers base to 0xf100000,
19  *    and than should update L2 filter accordingly to 0xf000000 (3.75 GB)
20  */
21 #define L2_FILTER_FOR_MAX_MEMORY_SIZE	0xC0000000 /* temporary limit l2 filter to 3gb (LSP issue) */
22 #define ADDRESS_FILTERING_END_REGISTER	0x8c04
23 
24 #define DYNAMIC_CS_SIZE_CONFIG
25 #define DISABLE_L2_FILTERING_DURING_DDR_TRAINING
26 
27 /* Termal Sensor Registers */
28 #define TSEN_CONTROL_LSB_REG		0xE4070
29 #define TSEN_CONTROL_LSB_TC_TRIM_OFFSET	0
30 #define TSEN_CONTROL_LSB_TC_TRIM_MASK	(0x7 << TSEN_CONTROL_LSB_TC_TRIM_OFFSET)
31 #define TSEN_CONTROL_MSB_REG		0xE4074
32 #define TSEN_CONTROL_MSB_RST_OFFSET	8
33 #define TSEN_CONTROL_MSB_RST_MASK	(0x1 << TSEN_CONTROL_MSB_RST_OFFSET)
34 #define TSEN_STATUS_REG			0xe4078
35 #define TSEN_STATUS_READOUT_VALID_OFFSET	10
36 #define TSEN_STATUS_READOUT_VALID_MASK	(0x1 <<				\
37 					 TSEN_STATUS_READOUT_VALID_OFFSET)
38 #define TSEN_STATUS_TEMP_OUT_OFFSET	0
39 #define TSEN_STATUS_TEMP_OUT_MASK	(0x3ff << TSEN_STATUS_TEMP_OUT_OFFSET)
40 
41 #if defined(CONFIG_DDR4)
42 static struct dlb_config ddr3_dlb_config_table[] = {
43 	{DLB_CTRL_REG, 0x2000005f},
44 	{DLB_BUS_OPT_WT_REG, 0x00880000},
45 	{DLB_AGING_REG, 0x3f7f007f},
46 	{DLB_EVICTION_CTRL_REG, 0x0000129f},
47 	{DLB_EVICTION_TIMERS_REG, 0x00ff0000},
48 	{DLB_WTS_DIFF_CS_REG, 0x04030803},
49 	{DLB_WTS_DIFF_BG_REG, 0x00000A02},
50 	{DLB_WTS_SAME_BG_REG, 0x08000901},
51 	{DLB_WTS_CMDS_REG,  0x00020005},
52 	{DLB_WTS_ATTR_PRIO_REG, 0x00060f10},
53 	{DLB_QUEUE_MAP_REG, 0x00000543},
54 	{DLB_SPLIT_REG, 0x0000000f},
55 	{DLB_USER_CMD_REG, 0x00000000},
56 	{0x0, 0x0}
57 };
58 #else /* !CONFIG_DDR4 */
59 static struct dlb_config ddr3_dlb_config_table[] = {
60 	{DLB_CTRL_REG, 0x2000005c},
61 	{DLB_BUS_OPT_WT_REG, 0x00880000},
62 	{DLB_AGING_REG, 0x0f7f007f},
63 	{DLB_EVICTION_CTRL_REG, 0x0000129f},
64 	{DLB_EVICTION_TIMERS_REG, 0x00ff0000},
65 	{DLB_WTS_DIFF_CS_REG, 0x04030802},
66 	{DLB_WTS_DIFF_BG_REG, 0x00000a02},
67 	{DLB_WTS_SAME_BG_REG, 0x09000a01},
68 	{DLB_WTS_CMDS_REG, 0x00020005},
69 	{DLB_WTS_ATTR_PRIO_REG, 0x00060f10},
70 	{DLB_QUEUE_MAP_REG, 0x00000543},
71 	{DLB_SPLIT_REG, 0x00000000},
72 	{DLB_USER_CMD_REG, 0x00000000},
73 	{0x0, 0x0}
74 };
75 #endif /* CONFIG_DDR4 */
76 
sys_env_dlb_config_ptr_get(void)77 static struct dlb_config *sys_env_dlb_config_ptr_get(void)
78 {
79 	return &ddr3_dlb_config_table[0];
80 }
81 
82 static u8 a38x_bw_per_freq[MV_DDR_FREQ_LAST] = {
83 	0x3,			/* MV_DDR_FREQ_100 */
84 #if !defined(CONFIG_DDR4)
85 	0x4,			/* MV_DDR_FREQ_400 */
86 	0x4,			/* MV_DDR_FREQ_533 */
87 #endif /* CONFIG_DDR4 */
88 	0x5,			/* MV_DDR_FREQ_667 */
89 	0x5,			/* MV_DDR_FREQ_800 */
90 	0x5,			/* MV_DDR_FREQ_933 */
91 	0x5,			/* MV_DDR_FREQ_1066 */
92 #if defined(CONFIG_DDR4)
93 	0x5,			/*MV_DDR_FREQ_900*/
94 	0x5,			/*MV_DDR_FREQ_1000*/
95 #else /* CONFIG_DDR4 */
96 	0x3,			/* MV_DDR_FREQ_311 */
97 	0x3,			/* MV_DDR_FREQ_333 */
98 	0x4,			/* MV_DDR_FREQ_467 */
99 	0x5,			/* MV_DDR_FREQ_850 */
100 	0x5,			/* MV_DDR_FREQ_600 */
101 	0x3,			/* MV_DDR_FREQ_300 */
102 	0x5,			/* MV_DDR_FREQ_900 */
103 	0x3,			/* MV_DDR_FREQ_360 */
104 	0x5			/* MV_DDR_FREQ_1000 */
105 #endif /* CONFIG_DDR4 */
106 };
107 
108 static u8 a38x_rate_per_freq[MV_DDR_FREQ_LAST] = {
109 	0x1,			/* MV_DDR_FREQ_100 */
110 #if !defined(CONFIG_DDR4)
111 	0x2,			/* MV_DDR_FREQ_400 */
112 	0x2,			/* MV_DDR_FREQ_533 */
113 #endif /* CONFIG_DDR4 */
114 	0x2,			/* MV_DDR_FREQ_667 */
115 	0x2,			/* MV_DDR_FREQ_800 */
116 	0x3,			/* MV_DDR_FREQ_933 */
117 	0x3,			/* MV_DDR_FREQ_1066 */
118 #ifdef CONFIG_DDR4
119 	0x2,			/*MV_DDR_FREQ_900*/
120 	0x2,			/*MV_DDR_FREQ_1000*/
121 #else /* CONFIG_DDR4 */
122 	0x1,			/* MV_DDR_FREQ_311 */
123 	0x1,			/* MV_DDR_FREQ_333 */
124 	0x2,			/* MV_DDR_FREQ_467 */
125 	0x2,			/* MV_DDR_FREQ_850 */
126 	0x2,			/* MV_DDR_FREQ_600 */
127 	0x1,			/* MV_DDR_FREQ_300 */
128 	0x2,			/* MV_DDR_FREQ_900 */
129 	0x1,			/* MV_DDR_FREQ_360 */
130 	0x2			/* MV_DDR_FREQ_1000 */
131 #endif /* CONFIG_DDR4 */
132 };
133 
134 static u16 a38x_vco_freq_per_sar_ref_clk_25_mhz[] = {
135 	666,			/* 0 */
136 	1332,
137 	800,
138 	1600,
139 	1066,
140 	2132,
141 	1200,
142 	2400,
143 	1332,
144 	1332,
145 	1500,
146 	1500,
147 	1600,			/* 12 */
148 	1600,
149 	1700,
150 	1700,
151 	1866,
152 	1866,
153 	1800,			/* 18 */
154 	2000,
155 	2000,
156 	4000,
157 	2132,
158 	2132,
159 	2300,
160 	2300,
161 	2400,
162 	2400,
163 	2500,
164 	2500,
165 	800
166 };
167 
168 static u16 a38x_vco_freq_per_sar_ref_clk_40_mhz[] = {
169 	666,			/* 0 */
170 	1332,
171 	800,
172 	800,			/* 0x3 */
173 	1066,
174 	1066,			/* 0x5 */
175 	1200,
176 	2400,
177 	1332,
178 	1332,
179 	1500,			/* 10 */
180 	1600,			/* 0xB */
181 	1600,
182 	1600,
183 	1700,
184 	1560,			/* 0xF */
185 	1866,
186 	1866,
187 	1800,
188 	2000,
189 	2000,			/* 20 */
190 	4000,
191 	2132,
192 	2132,
193 	2300,
194 	2300,
195 	2400,
196 	2400,
197 	2500,
198 	2500,
199 	1800			/* 30 - 0x1E */
200 };
201 
202 #if defined(CONFIG_DDR4)
203 u16 odt_slope[] = {
204 	21443,
205 	1452,
206 	482,
207 	240,
208 	141,
209 	90,
210 	67,
211 	52
212 };
213 
214 u16 odt_intercept[] = {
215 	1517,
216 	328,
217 	186,
218 	131,
219 	100,
220 	80,
221 	69,
222 	61
223 };
224 
225 /* Map of scratch PHY registers used to store stability value */
226 u32 dmin_phy_reg_table[MAX_BUS_NUM * MAX_CS_NUM][2] = {
227 	/* subphy, addr */
228 	{0, 0xc0},	/* cs 0, subphy 0 */
229 	{0, 0xc1},	/* cs 0, subphy 1 */
230 	{0, 0xc2},	/* cs 0, subphy 2 */
231 	{0, 0xc3},	/* cs 0, subphy 3 */
232 	{0, 0xc4},	/* cs 0, subphy 4 */
233 	{1, 0xc0},	/* cs 1, subphy 0 */
234 	{1, 0xc1},	/* cs 1, subphy 1 */
235 	{1, 0xc2},	/* cs 1, subphy 2 */
236 	{1, 0xc3},	/* cs 1, subphy 3 */
237 	{1, 0xc4},	/* cs 1, subphy 4 */
238 	{2, 0xc0},	/* cs 2, subphy 0 */
239 	{2, 0xc1},	/* cs 2, subphy 1 */
240 	{2, 0xc2},	/* cs 2, subphy 2 */
241 	{2, 0xc3},	/* cs 2, subphy 3 */
242 	{2, 0xc4},	/* cs 2, subphy 4 */
243 	{0, 0xc5},	/* cs 3, subphy 0 */
244 	{1, 0xc5},	/* cs 3, subphy 1 */
245 	{2, 0xc5},	/* cs 3, subphy 2 */
246 	{0, 0xc6},	/* cs 3, subphy 3 */
247 	{1, 0xc6}	/* cs 3, subphy 4 */
248 };
249 #endif /* CONFIG_DDR4 */
250 
251 static u32 dq_bit_map_2_phy_pin[] = {
252 	1, 0, 2, 6, 9, 8, 3, 7,	/* 0 */
253 	8, 9, 1, 7, 2, 6, 3, 0,	/* 1 */
254 	3, 9, 7, 8, 1, 0, 2, 6,	/* 2 */
255 	1, 0, 6, 2, 8, 3, 7, 9,	/* 3 */
256 	0, 1, 2, 9, 7, 8, 3, 6,	/* 4 */
257 };
258 
mv_ddr_mem_scrubbing(void)259 void mv_ddr_mem_scrubbing(void)
260 {
261 	ddr3_new_tip_ecc_scrub();
262 }
263 
264 static int ddr3_tip_a38x_set_divider(u8 dev_num, u32 if_id,
265 				     enum mv_ddr_freq freq);
266 
267 /*
268  * Read temperature TJ value
269  */
ddr3_ctrl_get_junc_temp(u8 dev_num)270 static u32 ddr3_ctrl_get_junc_temp(u8 dev_num)
271 {
272 	int reg = 0;
273 
274 	/* Initiates TSEN hardware reset once */
275 	if ((reg_read(TSEN_CONTROL_MSB_REG) & TSEN_CONTROL_MSB_RST_MASK) == 0) {
276 		reg_bit_set(TSEN_CONTROL_MSB_REG, TSEN_CONTROL_MSB_RST_MASK);
277 		/* set Tsen Tc Trim to correct default value (errata #132698) */
278 		reg = reg_read(TSEN_CONTROL_LSB_REG);
279 		reg &= ~TSEN_CONTROL_LSB_TC_TRIM_MASK;
280 		reg |= 0x3 << TSEN_CONTROL_LSB_TC_TRIM_OFFSET;
281 		reg_write(TSEN_CONTROL_LSB_REG, reg);
282 	}
283 	mdelay(10);
284 
285 	/* Check if the readout field is valid */
286 	if ((reg_read(TSEN_STATUS_REG) & TSEN_STATUS_READOUT_VALID_MASK) == 0) {
287 		printf("%s: TSEN not ready\n", __func__);
288 		return 0;
289 	}
290 
291 	reg = reg_read(TSEN_STATUS_REG);
292 	reg = (reg & TSEN_STATUS_TEMP_OUT_MASK) >> TSEN_STATUS_TEMP_OUT_OFFSET;
293 
294 	return ((((10000 * reg) / 21445) * 1000) - 272674) / 1000;
295 }
296 
297 /*
298  * Name:     ddr3_tip_a38x_get_freq_config.
299  * Desc:
300  * Args:
301  * Notes:
302  * Returns:  MV_OK if success, other error code if fail.
303  */
ddr3_tip_a38x_get_freq_config(u8 dev_num,enum mv_ddr_freq freq,struct hws_tip_freq_config_info * freq_config_info)304 static int ddr3_tip_a38x_get_freq_config(u8 dev_num, enum mv_ddr_freq freq,
305 				  struct hws_tip_freq_config_info
306 				  *freq_config_info)
307 {
308 	if (a38x_bw_per_freq[freq] == 0xff)
309 		return MV_NOT_SUPPORTED;
310 
311 	if (freq_config_info == NULL)
312 		return MV_BAD_PARAM;
313 
314 	freq_config_info->bw_per_freq = a38x_bw_per_freq[freq];
315 	freq_config_info->rate_per_freq = a38x_rate_per_freq[freq];
316 	freq_config_info->is_supported = 1;
317 
318 	return MV_OK;
319 }
320 
dunit_read(u32 addr,u32 mask,u32 * data)321 static void dunit_read(u32 addr, u32 mask, u32 *data)
322 {
323 	*data = reg_read(addr) & mask;
324 }
325 
dunit_write(u32 addr,u32 mask,u32 data)326 static void dunit_write(u32 addr, u32 mask, u32 data)
327 {
328 	u32 reg_val = data;
329 
330 	if (mask != MASK_ALL_BITS) {
331 		dunit_read(addr, MASK_ALL_BITS, &reg_val);
332 		reg_val &= (~mask);
333 		reg_val |= (data & mask);
334 	}
335 
336 	reg_write(addr, reg_val);
337 }
338 
339 #define ODPG_ENABLE_REG				0x186d4
340 #define ODPG_EN_OFFS				0
341 #define ODPG_EN_MASK				0x1
342 #define ODPG_EN_ENA				1
343 #define ODPG_EN_DONE				0
344 #define ODPG_DIS_OFFS				8
345 #define ODPG_DIS_MASK				0x1
346 #define ODPG_DIS_DIS				1
mv_ddr_odpg_enable(void)347 void mv_ddr_odpg_enable(void)
348 {
349 	dunit_write(ODPG_ENABLE_REG,
350 		    ODPG_EN_MASK << ODPG_EN_OFFS,
351 		    ODPG_EN_ENA << ODPG_EN_OFFS);
352 }
353 
mv_ddr_odpg_disable(void)354 void mv_ddr_odpg_disable(void)
355 {
356 	dunit_write(ODPG_ENABLE_REG,
357 		    ODPG_DIS_MASK << ODPG_DIS_OFFS,
358 		    ODPG_DIS_DIS << ODPG_DIS_OFFS);
359 }
360 
mv_ddr_odpg_done_clr(void)361 void mv_ddr_odpg_done_clr(void)
362 {
363 	return;
364 }
365 
mv_ddr_is_odpg_done(u32 count)366 int mv_ddr_is_odpg_done(u32 count)
367 {
368 	u32 i, data;
369 
370 	for (i = 0; i < count; i++) {
371 		dunit_read(ODPG_ENABLE_REG, MASK_ALL_BITS, &data);
372 		if (((data >> ODPG_EN_OFFS) & ODPG_EN_MASK) ==
373 		     ODPG_EN_DONE)
374 			break;
375 	}
376 
377 	if (i >= count) {
378 		printf("%s: timeout\n", __func__);
379 		return MV_FAIL;
380 	}
381 
382 	return MV_OK;
383 }
384 
mv_ddr_training_enable(void)385 void mv_ddr_training_enable(void)
386 {
387 	dunit_write(GLOB_CTRL_STATUS_REG,
388 		    TRAINING_TRIGGER_MASK << TRAINING_TRIGGER_OFFS,
389 		    TRAINING_TRIGGER_ENA << TRAINING_TRIGGER_OFFS);
390 }
391 
392 #define DRAM_INIT_CTRL_STATUS_REG	0x18488
393 #define TRAINING_TRIGGER_OFFS		0
394 #define TRAINING_TRIGGER_MASK		0x1
395 #define TRAINING_TRIGGER_ENA		1
396 #define TRAINING_DONE_OFFS		1
397 #define TRAINING_DONE_MASK		0x1
398 #define TRAINING_DONE_DONE		1
399 #define TRAINING_DONE_NOT_DONE		0
400 #define TRAINING_RESULT_OFFS		2
401 #define TRAINING_RESULT_MASK		0x1
402 #define TRAINING_RESULT_PASS		0
403 #define TRAINING_RESULT_FAIL		1
mv_ddr_is_training_done(u32 count,u32 * result)404 int mv_ddr_is_training_done(u32 count, u32 *result)
405 {
406 	u32 i, data;
407 
408 	if (result == NULL) {
409 		printf("%s: NULL result pointer found\n", __func__);
410 		return MV_FAIL;
411 	}
412 
413 	for (i = 0; i < count; i++) {
414 		dunit_read(DRAM_INIT_CTRL_STATUS_REG, MASK_ALL_BITS, &data);
415 		if (((data >> TRAINING_DONE_OFFS) & TRAINING_DONE_MASK) ==
416 		     TRAINING_DONE_DONE)
417 			break;
418 	}
419 
420 	if (i >= count) {
421 		printf("%s: timeout\n", __func__);
422 		return MV_FAIL;
423 	}
424 
425 	*result = (data >> TRAINING_RESULT_OFFS) & TRAINING_RESULT_MASK;
426 
427 	return MV_OK;
428 }
429 
430 #define DM_PAD	10
mv_ddr_dm_pad_get(void)431 u32 mv_ddr_dm_pad_get(void)
432 {
433 	return DM_PAD;
434 }
435 
436 /*
437  * Name:     ddr3_tip_a38x_select_ddr_controller.
438  * Desc:     Enable/Disable access to Marvell's server.
439  * Args:     dev_num     - device number
440  *           enable        - whether to enable or disable the server
441  * Notes:
442  * Returns:  MV_OK if success, other error code if fail.
443  */
ddr3_tip_a38x_select_ddr_controller(u8 dev_num,int enable)444 static int ddr3_tip_a38x_select_ddr_controller(u8 dev_num, int enable)
445 {
446 	u32 reg;
447 
448 	reg = reg_read(DUAL_DUNIT_CFG_REG);
449 
450 	if (enable)
451 		reg |= (1 << 6);
452 	else
453 		reg &= ~(1 << 6);
454 
455 	reg_write(DUAL_DUNIT_CFG_REG, reg);
456 
457 	return MV_OK;
458 }
459 
ddr3_tip_clock_mode(u32 frequency)460 static u8 ddr3_tip_clock_mode(u32 frequency)
461 {
462 	if ((frequency == MV_DDR_FREQ_LOW_FREQ) || (mv_ddr_freq_get(frequency) <= 400))
463 		return 1;
464 
465 	return 2;
466 }
467 
mv_ddr_sar_freq_get(int dev_num,enum mv_ddr_freq * freq)468 static int mv_ddr_sar_freq_get(int dev_num, enum mv_ddr_freq *freq)
469 {
470 	u32 reg, ref_clk_satr;
471 
472 	/* Read sample at reset setting */
473 	reg = (reg_read(REG_DEVICE_SAR1_ADDR) >>
474 	       RST2_CPU_DDR_CLOCK_SELECT_IN_OFFSET) &
475 		RST2_CPU_DDR_CLOCK_SELECT_IN_MASK;
476 
477 	ref_clk_satr = reg_read(DEVICE_SAMPLE_AT_RESET2_REG);
478 	if (((ref_clk_satr >> DEVICE_SAMPLE_AT_RESET2_REG_REFCLK_OFFSET) & 0x1) ==
479 	    DEVICE_SAMPLE_AT_RESET2_REG_REFCLK_25MHZ) {
480 		switch (reg) {
481 #if !defined(CONFIG_DDR4)
482 		case 0x1:
483 			DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
484 					      ("Warning: Unsupported freq mode for 333Mhz configured(%d)\n",
485 					      reg));
486 			/* fallthrough */
487 		case 0x0:
488 			*freq = MV_DDR_FREQ_333;
489 			break;
490 		case 0x3:
491 			DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
492 					      ("Warning: Unsupported freq mode for 400Mhz configured(%d)\n",
493 					      reg));
494 			/* fallthrough */
495 		case 0x2:
496 			*freq = MV_DDR_FREQ_400;
497 			break;
498 		case 0xd:
499 			DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
500 					      ("Warning: Unsupported freq mode for 533Mhz configured(%d)\n",
501 					      reg));
502 			/* fallthrough */
503 		case 0x4:
504 			*freq = MV_DDR_FREQ_533;
505 			break;
506 		case 0x6:
507 			*freq = MV_DDR_FREQ_600;
508 			break;
509 #endif /* CONFIG_DDR4 */
510 		case 0x11:
511 		case 0x14:
512 			DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
513 					      ("Warning: Unsupported freq mode for 667Mhz configured(%d)\n",
514 					      reg));
515 			/* fallthrough */
516 		case 0x8:
517 			*freq = MV_DDR_FREQ_667;
518 			break;
519 		case 0x15:
520 		case 0x1b:
521 			DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
522 					      ("Warning: Unsupported freq mode for 800Mhz configured(%d)\n",
523 					      reg));
524 			/* fallthrough */
525 		case 0xc:
526 			*freq = MV_DDR_FREQ_800;
527 			break;
528 		case 0x10:
529 			*freq = MV_DDR_FREQ_933;
530 			break;
531 		case 0x12:
532 			*freq = MV_DDR_FREQ_900;
533 			break;
534 #if defined(CONFIG_DDR4)
535 		case 0x13:
536 			*freq = MV_DDR_FREQ_1000;
537 			DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
538 					      ("Warning: Unsupported freq mode for 1000Mhz configured(%d)\n",
539 					      reg));
540 			break;
541 #else /* CONFIG_DDR4 */
542 		case 0x13:
543 			*freq = MV_DDR_FREQ_933;
544 			break;
545 #endif /* CONFIG_DDR4 */
546 		default:
547 			*freq = 0;
548 			return MV_NOT_SUPPORTED;
549 		}
550 	} else { /* REFCLK 40MHz case */
551 		switch (reg) {
552 #if !defined(CONFIG_DDR4)
553 		case 0x3:
554 			*freq = MV_DDR_FREQ_400;
555 			break;
556 		case 0x5:
557 			*freq = MV_DDR_FREQ_533;
558 			break;
559 #endif /* CONFIG_DDR4 */
560 		case 0xb:
561 			*freq = MV_DDR_FREQ_800;
562 			break;
563 		case 0x1e:
564 			*freq = MV_DDR_FREQ_900;
565 			break;
566 		default:
567 			*freq = 0;
568 			return MV_NOT_SUPPORTED;
569 		}
570 	}
571 
572 	return MV_OK;
573 }
574 
575 #if !defined(CONFIG_DDR4)
ddr3_tip_a38x_get_medium_freq(int dev_num,enum mv_ddr_freq * freq)576 static int ddr3_tip_a38x_get_medium_freq(int dev_num, enum mv_ddr_freq *freq)
577 {
578 	u32 reg, ref_clk_satr;
579 
580 	/* Read sample at reset setting */
581 	reg = (reg_read(REG_DEVICE_SAR1_ADDR) >>
582 	RST2_CPU_DDR_CLOCK_SELECT_IN_OFFSET) &
583 	RST2_CPU_DDR_CLOCK_SELECT_IN_MASK;
584 
585 	ref_clk_satr = reg_read(DEVICE_SAMPLE_AT_RESET2_REG);
586 	if (((ref_clk_satr >> DEVICE_SAMPLE_AT_RESET2_REG_REFCLK_OFFSET) & 0x1) ==
587 	    DEVICE_SAMPLE_AT_RESET2_REG_REFCLK_25MHZ) {
588 		switch (reg) {
589 		case 0x0:
590 		case 0x1:
591 			/* Medium is same as TF to run PBS in this freq */
592 			*freq = MV_DDR_FREQ_333;
593 			break;
594 		case 0x2:
595 		case 0x3:
596 			/* Medium is same as TF to run PBS in this freq */
597 			*freq = MV_DDR_FREQ_400;
598 			break;
599 		case 0x4:
600 		case 0xd:
601 			/* Medium is same as TF to run PBS in this freq */
602 			*freq = MV_DDR_FREQ_533;
603 			break;
604 		case 0x8:
605 		case 0x10:
606 		case 0x11:
607 		case 0x14:
608 			*freq = MV_DDR_FREQ_333;
609 			break;
610 		case 0xc:
611 		case 0x15:
612 		case 0x1b:
613 			*freq = MV_DDR_FREQ_400;
614 			break;
615 		case 0x6:
616 			*freq = MV_DDR_FREQ_300;
617 			break;
618 		case 0x12:
619 			*freq = MV_DDR_FREQ_360;
620 			break;
621 		case 0x13:
622 			*freq = MV_DDR_FREQ_400;
623 			break;
624 		default:
625 			*freq = 0;
626 			return MV_NOT_SUPPORTED;
627 		}
628 	} else { /* REFCLK 40MHz case */
629 		switch (reg) {
630 		case 0x3:
631 			/* Medium is same as TF to run PBS in this freq */
632 			*freq = MV_DDR_FREQ_400;
633 			break;
634 		case 0x5:
635 			/* Medium is same as TF to run PBS in this freq */
636 			*freq = MV_DDR_FREQ_533;
637 			break;
638 		case 0xb:
639 			*freq = MV_DDR_FREQ_400;
640 			break;
641 		case 0x1e:
642 			*freq = MV_DDR_FREQ_360;
643 			break;
644 		default:
645 			*freq = 0;
646 			return MV_NOT_SUPPORTED;
647 		}
648 	}
649 
650 	return MV_OK;
651 }
652 #endif /* CONFIG_DDR4 */
653 
ddr3_tip_a38x_get_device_info(u8 dev_num,struct ddr3_device_info * info_ptr)654 static int ddr3_tip_a38x_get_device_info(u8 dev_num, struct ddr3_device_info *info_ptr)
655 {
656 	info_ptr->device_id = 0x6800;
657 	info_ptr->ck_delay = ck_delay;
658 
659 	return MV_OK;
660 }
661 
662 /* check indirect access to phy register file completed */
is_prfa_done(void)663 static int is_prfa_done(void)
664 {
665 	u32 reg_val;
666 	u32 iter = 0;
667 
668 	do {
669 		if (iter++ > MAX_POLLING_ITERATIONS) {
670 			printf("error: %s: polling timeout\n", __func__);
671 			return MV_FAIL;
672 		}
673 		dunit_read(PHY_REG_FILE_ACCESS_REG, MASK_ALL_BITS, &reg_val);
674 		reg_val >>= PRFA_REQ_OFFS;
675 		reg_val &= PRFA_REQ_MASK;
676 	} while (reg_val == PRFA_REQ_ENA); /* request pending */
677 
678 	return MV_OK;
679 }
680 
681 /* write to phy register thru indirect access */
prfa_write(enum hws_access_type phy_access,u32 phy,enum hws_ddr_phy phy_type,u32 addr,u32 data,enum hws_operation op_type)682 static int prfa_write(enum hws_access_type phy_access, u32 phy,
683 		      enum hws_ddr_phy phy_type, u32 addr,
684 		      u32 data, enum hws_operation op_type)
685 {
686 	u32 reg_val = ((data & PRFA_DATA_MASK) << PRFA_DATA_OFFS) |
687 		      ((addr & PRFA_REG_NUM_MASK) << PRFA_REG_NUM_OFFS) |
688 		      ((phy & PRFA_PUP_NUM_MASK) << PRFA_PUP_NUM_OFFS) |
689 		      ((phy_type & PRFA_PUP_CTRL_DATA_MASK) << PRFA_PUP_CTRL_DATA_OFFS) |
690 		      ((phy_access & PRFA_PUP_BCAST_WR_ENA_MASK) << PRFA_PUP_BCAST_WR_ENA_OFFS) |
691 		      (((addr >> 6) & PRFA_REG_NUM_HI_MASK) << PRFA_REG_NUM_HI_OFFS) |
692 		      ((op_type & PRFA_TYPE_MASK) << PRFA_TYPE_OFFS);
693 	dunit_write(PHY_REG_FILE_ACCESS_REG, MASK_ALL_BITS, reg_val);
694 	reg_val |= (PRFA_REQ_ENA << PRFA_REQ_OFFS);
695 	dunit_write(PHY_REG_FILE_ACCESS_REG, MASK_ALL_BITS, reg_val);
696 
697 	/* polling for prfa request completion */
698 	if (is_prfa_done() != MV_OK)
699 		return MV_FAIL;
700 
701 	return MV_OK;
702 }
703 
704 /* read from phy register thru indirect access */
prfa_read(enum hws_access_type phy_access,u32 phy,enum hws_ddr_phy phy_type,u32 addr,u32 * data)705 static int prfa_read(enum hws_access_type phy_access, u32 phy,
706 		     enum hws_ddr_phy phy_type, u32 addr, u32 *data)
707 {
708 	struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get();
709 	u32 max_phy = ddr3_tip_dev_attr_get(0, MV_ATTR_OCTET_PER_INTERFACE);
710 	u32 i, reg_val;
711 
712 	if (phy_access == ACCESS_TYPE_MULTICAST) {
713 		for (i = 0; i < max_phy; i++) {
714 			VALIDATE_BUS_ACTIVE(tm->bus_act_mask, i);
715 			if (prfa_write(ACCESS_TYPE_UNICAST, i, phy_type, addr, 0, OPERATION_READ) != MV_OK)
716 				return MV_FAIL;
717 			dunit_read(PHY_REG_FILE_ACCESS_REG, MASK_ALL_BITS, &reg_val);
718 			data[i] = (reg_val >> PRFA_DATA_OFFS) & PRFA_DATA_MASK;
719 		}
720 	} else {
721 		if (prfa_write(phy_access, phy, phy_type, addr, 0, OPERATION_READ) != MV_OK)
722 			return MV_FAIL;
723 		dunit_read(PHY_REG_FILE_ACCESS_REG, MASK_ALL_BITS, &reg_val);
724 		*data = (reg_val >> PRFA_DATA_OFFS) & PRFA_DATA_MASK;
725 	}
726 
727 	return MV_OK;
728 }
729 
mv_ddr_sw_db_init(u32 dev_num,u32 board_id)730 static int mv_ddr_sw_db_init(u32 dev_num, u32 board_id)
731 {
732 	struct hws_tip_config_func_db config_func;
733 
734 	/* new read leveling version */
735 	config_func.mv_ddr_dunit_read = dunit_read;
736 	config_func.mv_ddr_dunit_write = dunit_write;
737 	config_func.tip_dunit_mux_select_func =
738 		ddr3_tip_a38x_select_ddr_controller;
739 	config_func.tip_get_freq_config_info_func =
740 		ddr3_tip_a38x_get_freq_config;
741 	config_func.tip_set_freq_divider_func = ddr3_tip_a38x_set_divider;
742 	config_func.tip_get_device_info_func = ddr3_tip_a38x_get_device_info;
743 	config_func.tip_get_temperature = ddr3_ctrl_get_junc_temp;
744 	config_func.tip_get_clock_ratio = ddr3_tip_clock_mode;
745 	config_func.tip_external_read = ddr3_tip_ext_read;
746 	config_func.tip_external_write = ddr3_tip_ext_write;
747 	config_func.mv_ddr_phy_read = prfa_read;
748 	config_func.mv_ddr_phy_write = prfa_write;
749 
750 	ddr3_tip_init_config_func(dev_num, &config_func);
751 
752 	ddr3_tip_register_dq_table(dev_num, dq_bit_map_2_phy_pin);
753 
754 	/* set device attributes*/
755 	ddr3_tip_dev_attr_init(dev_num);
756 	ddr3_tip_dev_attr_set(dev_num, MV_ATTR_TIP_REV, MV_TIP_REV_4);
757 	ddr3_tip_dev_attr_set(dev_num, MV_ATTR_PHY_EDGE, MV_DDR_PHY_EDGE_POSITIVE);
758 	ddr3_tip_dev_attr_set(dev_num, MV_ATTR_OCTET_PER_INTERFACE, DDR_INTERFACE_OCTETS_NUM);
759 	ddr3_tip_dev_attr_set(dev_num, MV_ATTR_INTERLEAVE_WA, 0);
760 
761 	ca_delay = 0;
762 	delay_enable = 1;
763 	dfs_low_freq = DFS_LOW_FREQ_VALUE;
764 	calibration_update_control = 1;
765 
766 #if !defined(CONFIG_DDR4)
767 	ddr3_tip_a38x_get_medium_freq(dev_num, &medium_freq);
768 #endif /* CONFIG_DDR4 */
769 
770 	return MV_OK;
771 }
772 
mv_ddr_training_mask_set(void)773 static int mv_ddr_training_mask_set(void)
774 {
775 	struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get();
776 #if defined(CONFIG_DDR4)
777 	mask_tune_func = (SET_LOW_FREQ_MASK_BIT |
778 			  LOAD_PATTERN_MASK_BIT |
779 			  SET_TARGET_FREQ_MASK_BIT |
780 			  WRITE_LEVELING_TF_MASK_BIT |
781 			  READ_LEVELING_TF_MASK_BIT |
782 			  RECEIVER_CALIBRATION_MASK_BIT |
783 			  WL_PHASE_CORRECTION_MASK_BIT |
784 			  DQ_VREF_CALIBRATION_MASK_BIT);
785 	/* Temporarily disable the DQ_MAPPING stage */
786 	/*		  DQ_MAPPING_MASK_BIT */
787 	rl_mid_freq_wa = 0;
788 
789 	/* In case A382, Vref calibration workaround isn't required */
790 	if (((reg_read(DEV_ID_REG) & 0xFFFF0000) >> 16) == 0x6811) {
791 		printf("vref_calibration_wa is disabled\n");
792 		vref_calibration_wa = 0;
793 	}
794 
795 	if (DDR3_IS_16BIT_DRAM_MODE(tm->bus_act_mask) == 1)
796 		mask_tune_func &= ~WL_PHASE_CORRECTION_MASK_BIT;
797 
798 #else /* CONFIG_DDR4 */
799 	enum mv_ddr_freq ddr_freq = tm->interface_params[0].memory_freq;
800 
801 	mask_tune_func = (SET_LOW_FREQ_MASK_BIT |
802 			  LOAD_PATTERN_MASK_BIT |
803 			  SET_MEDIUM_FREQ_MASK_BIT | WRITE_LEVELING_MASK_BIT |
804 			  WRITE_LEVELING_SUPP_MASK_BIT |
805 			  READ_LEVELING_MASK_BIT |
806 			  PBS_RX_MASK_BIT |
807 			  PBS_TX_MASK_BIT |
808 			  SET_TARGET_FREQ_MASK_BIT |
809 			  WRITE_LEVELING_TF_MASK_BIT |
810 			  WRITE_LEVELING_SUPP_TF_MASK_BIT |
811 			  READ_LEVELING_TF_MASK_BIT |
812 			  CENTRALIZATION_RX_MASK_BIT |
813 			  CENTRALIZATION_TX_MASK_BIT);
814 	rl_mid_freq_wa = 1;
815 
816 	if ((ddr_freq == MV_DDR_FREQ_333) || (ddr_freq == MV_DDR_FREQ_400)) {
817 		mask_tune_func = (WRITE_LEVELING_MASK_BIT |
818 				  LOAD_PATTERN_2_MASK_BIT |
819 				  WRITE_LEVELING_SUPP_MASK_BIT |
820 				  READ_LEVELING_MASK_BIT |
821 				  PBS_RX_MASK_BIT |
822 				  PBS_TX_MASK_BIT |
823 				  CENTRALIZATION_RX_MASK_BIT |
824 				  CENTRALIZATION_TX_MASK_BIT);
825 		rl_mid_freq_wa = 0; /* WA not needed if 333/400 is TF */
826 	}
827 
828 	/* Supplementary not supported for ECC modes */
829 	if (mv_ddr_is_ecc_ena()) {
830 		mask_tune_func &= ~WRITE_LEVELING_SUPP_TF_MASK_BIT;
831 		mask_tune_func &= ~WRITE_LEVELING_SUPP_MASK_BIT;
832 		mask_tune_func &= ~PBS_TX_MASK_BIT;
833 		mask_tune_func &= ~PBS_RX_MASK_BIT;
834 	}
835 #endif /* CONFIG_DDR4 */
836 
837 	return MV_OK;
838 }
839 
840 /* function: mv_ddr_set_calib_controller
841  * this function sets the controller which will control
842  * the calibration cycle in the end of the training.
843  * 1 - internal controller
844  * 2 - external controller
845  */
mv_ddr_set_calib_controller(void)846 void mv_ddr_set_calib_controller(void)
847 {
848 	calibration_update_control = CAL_UPDATE_CTRL_INT;
849 }
850 
ddr3_tip_a38x_set_divider(u8 dev_num,u32 if_id,enum mv_ddr_freq frequency)851 static int ddr3_tip_a38x_set_divider(u8 dev_num, u32 if_id,
852 				     enum mv_ddr_freq frequency)
853 {
854 	u32 divider = 0;
855 	u32 sar_val, ref_clk_satr;
856 	u32 async_val;
857 	u32 cpu_freq;
858 	u32 ddr_freq = mv_ddr_freq_get(frequency);
859 
860 	if (if_id != 0) {
861 		DEBUG_TRAINING_ACCESS(DEBUG_LEVEL_ERROR,
862 				      ("A38x does not support interface 0x%x\n",
863 				       if_id));
864 		return MV_BAD_PARAM;
865 	}
866 
867 	/* get VCO freq index */
868 	sar_val = (reg_read(REG_DEVICE_SAR1_ADDR) >>
869 		   RST2_CPU_DDR_CLOCK_SELECT_IN_OFFSET) &
870 		RST2_CPU_DDR_CLOCK_SELECT_IN_MASK;
871 
872 	ref_clk_satr = reg_read(DEVICE_SAMPLE_AT_RESET2_REG);
873 	if (((ref_clk_satr >> DEVICE_SAMPLE_AT_RESET2_REG_REFCLK_OFFSET) & 0x1) ==
874 	    DEVICE_SAMPLE_AT_RESET2_REG_REFCLK_25MHZ)
875 		cpu_freq = a38x_vco_freq_per_sar_ref_clk_25_mhz[sar_val];
876 	else
877 		cpu_freq = a38x_vco_freq_per_sar_ref_clk_40_mhz[sar_val];
878 
879 	divider = cpu_freq / ddr_freq;
880 
881 	if (((cpu_freq % ddr_freq != 0) || (divider != 2 && divider != 3)) &&
882 	    (ddr_freq > 400)) {
883 		/* Set async mode */
884 		dunit_write(0x20220, 0x1000, 0x1000);
885 		dunit_write(0xe42f4, 0x200, 0x200);
886 
887 		/* Wait for async mode setup */
888 		mdelay(5);
889 
890 		/* Set KNL values */
891 		switch (frequency) {
892 #ifndef CONFIG_DDR4 /* CONFIG_DDR3 */
893 		case MV_DDR_FREQ_467:
894 			async_val = 0x806f012;
895 			break;
896 		case MV_DDR_FREQ_533:
897 			async_val = 0x807f012;
898 			break;
899 		case MV_DDR_FREQ_600:
900 			async_val = 0x805f00a;
901 			break;
902 #endif
903 		case MV_DDR_FREQ_667:
904 			async_val = 0x809f012;
905 			break;
906 		case MV_DDR_FREQ_800:
907 			async_val = 0x807f00a;
908 			break;
909 #ifndef CONFIG_DDR4 /* CONFIG_DDR3 */
910 		case MV_DDR_FREQ_850:
911 			async_val = 0x80cb012;
912 			break;
913 #endif
914 		case MV_DDR_FREQ_900:
915 			async_val = 0x80d7012;
916 			break;
917 		case MV_DDR_FREQ_933:
918 			async_val = 0x80df012;
919 			break;
920 		case MV_DDR_FREQ_1000:
921 			async_val = 0x80ef012;
922 			break;
923 		case MV_DDR_FREQ_1066:
924 			async_val = 0x80ff012;
925 			break;
926 		default:
927 			/* set MV_DDR_FREQ_667 as default */
928 			async_val = 0x809f012;
929 		}
930 		dunit_write(0xe42f0, 0xffffffff, async_val);
931 	} else {
932 		/* Set sync mode */
933 		dunit_write(0x20220, 0x1000, 0x0);
934 		dunit_write(0xe42f4, 0x200, 0x0);
935 
936 		/* cpupll_clkdiv_reset_mask */
937 		dunit_write(0xe4264, 0xff, 0x1f);
938 
939 		/* cpupll_clkdiv_reload_smooth */
940 		dunit_write(0xe4260, (0xff << 8), (0x2 << 8));
941 
942 		/* cpupll_clkdiv_relax_en */
943 		dunit_write(0xe4260, (0xff << 24), (0x2 << 24));
944 
945 		/* write the divider */
946 		dunit_write(0xe4268, (0x3f << 8), (divider << 8));
947 
948 		/* set cpupll_clkdiv_reload_ratio */
949 		dunit_write(0xe4264, (1 << 8), (1 << 8));
950 
951 		/* undet cpupll_clkdiv_reload_ratio */
952 		dunit_write(0xe4264, (1 << 8), 0x0);
953 
954 		/* clear cpupll_clkdiv_reload_force */
955 		dunit_write(0xe4260, (0xff << 8), 0x0);
956 
957 		/* clear cpupll_clkdiv_relax_en */
958 		dunit_write(0xe4260, (0xff << 24), 0x0);
959 
960 		/* clear cpupll_clkdiv_reset_mask */
961 		dunit_write(0xe4264, 0xff, 0x0);
962 	}
963 
964 	/* Dunit training clock + 1:1/2:1 mode */
965 	dunit_write(0x18488, (1 << 16), ((ddr3_tip_clock_mode(frequency) & 0x1) << 16));
966 	dunit_write(0x1524, (1 << 15), ((ddr3_tip_clock_mode(frequency) - 1) << 15));
967 
968 	return MV_OK;
969 }
970 
971 /*
972  * external read from memory
973  */
ddr3_tip_ext_read(u32 dev_num,u32 if_id,u32 reg_addr,u32 num_of_bursts,u32 * data)974 int ddr3_tip_ext_read(u32 dev_num, u32 if_id, u32 reg_addr,
975 		      u32 num_of_bursts, u32 *data)
976 {
977 	u32 burst_num;
978 
979 	for (burst_num = 0; burst_num < num_of_bursts * 8; burst_num++)
980 		data[burst_num] = readl(reg_addr + 4 * burst_num);
981 
982 	return MV_OK;
983 }
984 
985 /*
986  * external write to memory
987  */
ddr3_tip_ext_write(u32 dev_num,u32 if_id,u32 reg_addr,u32 num_of_bursts,u32 * data)988 int ddr3_tip_ext_write(u32 dev_num, u32 if_id, u32 reg_addr,
989 		       u32 num_of_bursts, u32 *data) {
990 	u32 burst_num;
991 
992 	for (burst_num = 0; burst_num < num_of_bursts * 8; burst_num++)
993 		writel(data[burst_num], reg_addr + 4 * burst_num);
994 
995 	return MV_OK;
996 }
997 
mv_ddr_early_init(void)998 int mv_ddr_early_init(void)
999 {
1000 	/* FIXME: change this configuration per ddr type
1001 	 * configure a380 and a390 to work with receiver odt timing
1002 	 * the odt_config is defined:
1003 	 * '1' in ddr4
1004 	 * '0' in ddr3
1005 	 * here the parameter is run over in ddr4 and ddr3 to '1' (in ddr4 the default is '1')
1006 	 * to configure the odt to work with timing restrictions
1007 	 */
1008 
1009 	mv_ddr_sw_db_init(0, 0);
1010 
1011 	return MV_OK;
1012 }
1013 
mv_ddr_early_init2(void)1014 int mv_ddr_early_init2(void)
1015 {
1016 	mv_ddr_training_mask_set();
1017 
1018 	return MV_OK;
1019 }
1020 
mv_ddr_pre_training_fixup(void)1021 int mv_ddr_pre_training_fixup(void)
1022 {
1023 	return 0;
1024 }
1025 
mv_ddr_post_training_fixup(void)1026 int mv_ddr_post_training_fixup(void)
1027 {
1028 	return 0;
1029 }
1030 
ddr3_post_run_alg(void)1031 int ddr3_post_run_alg(void)
1032 {
1033 	return MV_OK;
1034 }
1035 
ddr3_silicon_post_init(void)1036 int ddr3_silicon_post_init(void)
1037 {
1038 	struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get();
1039 
1040 	/* Set half bus width */
1041 	if (DDR3_IS_16BIT_DRAM_MODE(tm->bus_act_mask)) {
1042 		CHECK_STATUS(ddr3_tip_if_write
1043 			     (0, ACCESS_TYPE_UNICAST, PARAM_NOT_CARE,
1044 			      SDRAM_CFG_REG, 0x0, 0x8000));
1045 	}
1046 
1047 	return MV_OK;
1048 }
1049 
mv_ddr_init_freq_get(void)1050 u32 mv_ddr_init_freq_get(void)
1051 {
1052 	enum mv_ddr_freq freq;
1053 
1054 	mv_ddr_sar_freq_get(0, &freq);
1055 
1056 	return freq;
1057 }
1058 
ddr3_get_bus_width(void)1059 static u32 ddr3_get_bus_width(void)
1060 {
1061 	u32 bus_width;
1062 
1063 	bus_width = (reg_read(SDRAM_CFG_REG) & 0x8000) >>
1064 		BUS_IN_USE_OFFS;
1065 
1066 	return (bus_width == 0) ? 16 : 32;
1067 }
1068 
ddr3_get_device_width(u32 cs)1069 static u32 ddr3_get_device_width(u32 cs)
1070 {
1071 	u32 device_width;
1072 
1073 	device_width = (reg_read(SDRAM_ADDR_CTRL_REG) &
1074 			(CS_STRUCT_MASK << CS_STRUCT_OFFS(cs))) >>
1075 			CS_STRUCT_OFFS(cs);
1076 
1077 	return (device_width == 0) ? 8 : 16;
1078 }
1079 
ddr3_get_device_size(u32 cs)1080 static u32 ddr3_get_device_size(u32 cs)
1081 {
1082 	u32 device_size_low, device_size_high, device_size;
1083 	u32 data, cs_low_offset, cs_high_offset;
1084 
1085 	cs_low_offset = CS_SIZE_OFFS(cs);
1086 	cs_high_offset = CS_SIZE_HIGH_OFFS(cs);
1087 
1088 	data = reg_read(SDRAM_ADDR_CTRL_REG);
1089 	device_size_low = (data >> cs_low_offset) & 0x3;
1090 	device_size_high = (data >> cs_high_offset) & 0x1;
1091 
1092 	device_size = device_size_low | (device_size_high << 2);
1093 
1094 	switch (device_size) {
1095 	case 0:
1096 		return 2048;
1097 	case 2:
1098 		return 512;
1099 	case 3:
1100 		return 1024;
1101 	case 4:
1102 		return 4096;
1103 	case 5:
1104 		return 8192;
1105 	case 1:
1106 	default:
1107 		DEBUG_INIT_C("Error: Wrong device size of Cs: ", cs, 1);
1108 		/* zeroes mem size in ddr3_calc_mem_cs_size */
1109 		return 0;
1110 	}
1111 }
1112 
ddr3_calc_mem_cs_size(u32 cs,uint64_t * cs_size)1113 int ddr3_calc_mem_cs_size(u32 cs, uint64_t *cs_size)
1114 {
1115 	u32 cs_mem_size;
1116 
1117 	/* Calculate in MiB */
1118 	cs_mem_size = ((ddr3_get_bus_width() / ddr3_get_device_width(cs)) *
1119 		       ddr3_get_device_size(cs)) / 8;
1120 
1121 	/*
1122 	 * Multiple controller bus width, 2x for 64 bit
1123 	 * (SoC controller may be 32 or 64 bit,
1124 	 * so bit 15 in 0x1400, that means if whole bus used or only half,
1125 	 * have a differnt meaning
1126 	 */
1127 	cs_mem_size *= DDR_CONTROLLER_BUS_WIDTH_MULTIPLIER;
1128 
1129 	if ((cs_mem_size < 128) || (cs_mem_size > 4096)) {
1130 		DEBUG_INIT_C("Error: Wrong Memory size of Cs: ", cs, 1);
1131 		return MV_BAD_VALUE;
1132 	}
1133 
1134 	*cs_size = cs_mem_size;
1135 
1136 	return MV_OK;
1137 }
1138 
ddr3_fast_path_dynamic_cs_size_config(u32 cs_ena)1139 static int ddr3_fast_path_dynamic_cs_size_config(u32 cs_ena)
1140 {
1141 	u32 reg, cs;
1142 	uint64_t mem_total_size = 0;
1143 	uint64_t cs_mem_size_mb = 0;
1144 	uint64_t cs_mem_size = 0;
1145 	uint64_t mem_total_size_c, cs_mem_size_c;
1146 
1147 
1148 #ifdef DEVICE_MAX_DRAM_ADDRESS_SIZE
1149 	u32 physical_mem_size;
1150 	u32 max_mem_size = DEVICE_MAX_DRAM_ADDRESS_SIZE;
1151 	struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get();
1152 #endif
1153 
1154 	/* Open fast path windows */
1155 	for (cs = 0; cs < MAX_CS_NUM; cs++) {
1156 		if (cs_ena & (1 << cs)) {
1157 			/* get CS size */
1158 			if (ddr3_calc_mem_cs_size(cs, &cs_mem_size_mb) != MV_OK)
1159 				return MV_FAIL;
1160 			cs_mem_size = cs_mem_size_mb * _1M;
1161 
1162 #ifdef DEVICE_MAX_DRAM_ADDRESS_SIZE
1163 			/*
1164 			 * if number of address pins doesn't allow to use max
1165 			 * mem size that is defined in topology
1166 			 * mem size is defined by DEVICE_MAX_DRAM_ADDRESS_SIZE
1167 			 */
1168 			physical_mem_size = mem_size
1169 				[tm->interface_params[0].memory_size];
1170 
1171 			if (ddr3_get_device_width(cs) == 16) {
1172 				/*
1173 				 * 16bit mem device can be twice more - no need
1174 				 * in less significant pin
1175 				 */
1176 				max_mem_size = DEVICE_MAX_DRAM_ADDRESS_SIZE * 2;
1177 			}
1178 
1179 			if (physical_mem_size > max_mem_size) {
1180 				cs_mem_size = max_mem_size *
1181 					(ddr3_get_bus_width() /
1182 					 ddr3_get_device_width(cs));
1183 				printf("Updated Physical Mem size is from 0x%x to %x\n",
1184 				       physical_mem_size,
1185 				       DEVICE_MAX_DRAM_ADDRESS_SIZE);
1186 			}
1187 #endif
1188 
1189 			/* set fast path window control for the cs */
1190 			reg = 0xffffe1;
1191 			reg |= (cs << 2);
1192 			reg |= (cs_mem_size - 1) & 0xffff0000;
1193 			/*Open fast path Window */
1194 			reg_write(REG_FASTPATH_WIN_CTRL_ADDR(cs), reg);
1195 
1196 			/* Set fast path window base address for the cs */
1197 			reg = ((cs_mem_size) * cs) & 0xffff0000;
1198 			/* Set base address */
1199 			reg_write(REG_FASTPATH_WIN_BASE_ADDR(cs), reg);
1200 
1201 			/*
1202 			 * Since memory size may be bigger than 4G the summ may
1203 			 * be more than 32 bit word,
1204 			 * so to estimate the result divide mem_total_size and
1205 			 * cs_mem_size by 0x10000 (it is equal to >> 16)
1206 			 */
1207 			mem_total_size_c = (mem_total_size >> 16) & 0xffffffffffff;
1208 			cs_mem_size_c = (cs_mem_size >> 16) & 0xffffffffffff;
1209 
1210 			/* if the sum less than 2 G - calculate the value */
1211 			if (mem_total_size_c + cs_mem_size_c < 0x10000)
1212 				mem_total_size += cs_mem_size;
1213 			else	/* put max possible size */
1214 				mem_total_size = L2_FILTER_FOR_MAX_MEMORY_SIZE;
1215 		}
1216 	}
1217 
1218 	/* Set L2 filtering to Max Memory size */
1219 	reg_write(ADDRESS_FILTERING_END_REGISTER, mem_total_size);
1220 
1221 	return MV_OK;
1222 }
1223 
ddr3_restore_and_set_final_windows(u32 * win,const char * ddr_type)1224 static int ddr3_restore_and_set_final_windows(u32 *win, const char *ddr_type)
1225 {
1226 	u32 win_ctrl_reg, num_of_win_regs;
1227 	u32 cs_ena = mv_ddr_sys_env_get_cs_ena_from_reg();
1228 	u32 ui;
1229 
1230 	win_ctrl_reg = REG_XBAR_WIN_4_CTRL_ADDR;
1231 	num_of_win_regs = 16;
1232 
1233 	/* Return XBAR windows 4-7 or 16-19 init configuration */
1234 	for (ui = 0; ui < num_of_win_regs; ui++)
1235 		reg_write((win_ctrl_reg + 0x4 * ui), win[ui]);
1236 
1237 	printf("%s Training Sequence - Switching XBAR Window to FastPath Window\n",
1238 	       ddr_type);
1239 
1240 #if defined DYNAMIC_CS_SIZE_CONFIG
1241 	if (ddr3_fast_path_dynamic_cs_size_config(cs_ena) != MV_OK)
1242 		printf("ddr3_fast_path_dynamic_cs_size_config FAILED\n");
1243 #else
1244 	u32 reg, cs;
1245 	reg = 0x1fffffe1;
1246 	for (cs = 0; cs < MAX_CS_NUM; cs++) {
1247 		if (cs_ena & (1 << cs)) {
1248 			reg |= (cs << 2);
1249 			break;
1250 		}
1251 	}
1252 	/* Open fast path Window to - 0.5G */
1253 	reg_write(REG_FASTPATH_WIN_CTRL_ADDR(0), reg);
1254 #endif
1255 
1256 	return MV_OK;
1257 }
1258 
ddr3_save_and_set_training_windows(u32 * win)1259 static int ddr3_save_and_set_training_windows(u32 *win)
1260 {
1261 	u32 cs_ena;
1262 	u32 reg, tmp_count, cs, ui;
1263 	u32 win_ctrl_reg, win_base_reg, win_remap_reg;
1264 	u32 num_of_win_regs, win_jump_index;
1265 	win_ctrl_reg = REG_XBAR_WIN_4_CTRL_ADDR;
1266 	win_base_reg = REG_XBAR_WIN_4_BASE_ADDR;
1267 	win_remap_reg = REG_XBAR_WIN_4_REMAP_ADDR;
1268 	win_jump_index = 0x10;
1269 	num_of_win_regs = 16;
1270 	struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get();
1271 
1272 #ifdef DISABLE_L2_FILTERING_DURING_DDR_TRAINING
1273 	/*
1274 	 * Disable L2 filtering during DDR training
1275 	 * (when Cross Bar window is open)
1276 	 */
1277 	reg_write(ADDRESS_FILTERING_END_REGISTER, 0);
1278 #endif
1279 
1280 	cs_ena = tm->interface_params[0].as_bus_params[0].cs_bitmask;
1281 
1282 	/* Close XBAR Window 19 - Not needed */
1283 	/* {0x000200e8}  -   Open Mbus Window - 2G */
1284 	reg_write(REG_XBAR_WIN_19_CTRL_ADDR, 0);
1285 
1286 	/* Save XBAR Windows 4-19 init configurations */
1287 	for (ui = 0; ui < num_of_win_regs; ui++)
1288 		win[ui] = reg_read(win_ctrl_reg + 0x4 * ui);
1289 
1290 	/* Open XBAR Windows 4-7 or 16-19 for other CS */
1291 	reg = 0;
1292 	tmp_count = 0;
1293 	for (cs = 0; cs < MAX_CS_NUM; cs++) {
1294 		if (cs_ena & (1 << cs)) {
1295 			switch (cs) {
1296 			case 0:
1297 				reg = 0x0e00;
1298 				break;
1299 			case 1:
1300 				reg = 0x0d00;
1301 				break;
1302 			case 2:
1303 				reg = 0x0b00;
1304 				break;
1305 			case 3:
1306 				reg = 0x0700;
1307 				break;
1308 			}
1309 			reg |= (1 << 0);
1310 			reg |= (SDRAM_CS_SIZE & 0xffff0000);
1311 
1312 			reg_write(win_ctrl_reg + win_jump_index * tmp_count,
1313 				  reg);
1314 			reg = (((SDRAM_CS_SIZE + 1) * (tmp_count)) &
1315 			       0xffff0000);
1316 			reg_write(win_base_reg + win_jump_index * tmp_count,
1317 				  reg);
1318 
1319 			if (win_remap_reg <= REG_XBAR_WIN_7_REMAP_ADDR)
1320 				reg_write(win_remap_reg +
1321 					  win_jump_index * tmp_count, 0);
1322 
1323 			tmp_count++;
1324 		}
1325 	}
1326 
1327 	return MV_OK;
1328 }
1329 
1330 static u32 win[16];
1331 
mv_ddr_pre_training_soc_config(const char * ddr_type)1332 int mv_ddr_pre_training_soc_config(const char *ddr_type)
1333 {
1334 	u32 soc_num;
1335 	u32 reg_val;
1336 
1337 	/* Switching CPU to MRVL ID */
1338 	soc_num = (reg_read(REG_SAMPLE_RESET_HIGH_ADDR) & SAR1_CPU_CORE_MASK) >>
1339 		SAR1_CPU_CORE_OFFSET;
1340 	switch (soc_num) {
1341 	case 0x3:
1342 		reg_bit_set(CPU_CONFIGURATION_REG(3), CPU_MRVL_ID_OFFSET);
1343 		reg_bit_set(CPU_CONFIGURATION_REG(2), CPU_MRVL_ID_OFFSET);
1344 		/* fallthrough */
1345 	case 0x1:
1346 		reg_bit_set(CPU_CONFIGURATION_REG(1), CPU_MRVL_ID_OFFSET);
1347 		/* fallthrough */
1348 	case 0x0:
1349 		reg_bit_set(CPU_CONFIGURATION_REG(0), CPU_MRVL_ID_OFFSET);
1350 		/* fallthrough */
1351 	default:
1352 		break;
1353 	}
1354 
1355 	/*
1356 	 * Set DRAM Reset Mask in case detected GPIO indication of wakeup from
1357 	 * suspend i.e the DRAM values will not be overwritten / reset when
1358 	 * waking from suspend
1359 	 */
1360 	if (mv_ddr_sys_env_suspend_wakeup_check() ==
1361 	    SUSPEND_WAKEUP_ENABLED_GPIO_DETECTED) {
1362 		reg_bit_set(SDRAM_INIT_CTRL_REG,
1363 			    DRAM_RESET_MASK_MASKED << DRAM_RESET_MASK_OFFS);
1364 	}
1365 
1366 	/* Fix read ready phases for all SOC in reg 0x15c8 */
1367 	reg_val = reg_read(TRAINING_DBG_3_REG);
1368 
1369 	reg_val &= ~(TRN_DBG_RDY_INC_PH_2TO1_MASK << TRN_DBG_RDY_INC_PH_2TO1_OFFS(0));
1370 	reg_val |= (0x4 << TRN_DBG_RDY_INC_PH_2TO1_OFFS(0));	/* phase 0 */
1371 
1372 	reg_val &= ~(TRN_DBG_RDY_INC_PH_2TO1_MASK << TRN_DBG_RDY_INC_PH_2TO1_OFFS(1));
1373 	reg_val |= (0x4 << TRN_DBG_RDY_INC_PH_2TO1_OFFS(1));	/* phase 1 */
1374 
1375 	reg_val &= ~(TRN_DBG_RDY_INC_PH_2TO1_MASK << TRN_DBG_RDY_INC_PH_2TO1_OFFS(3));
1376 	reg_val |= (0x6 << TRN_DBG_RDY_INC_PH_2TO1_OFFS(3));	/* phase 3 */
1377 
1378 	reg_val &= ~(TRN_DBG_RDY_INC_PH_2TO1_MASK << TRN_DBG_RDY_INC_PH_2TO1_OFFS(4));
1379 	reg_val |= (0x6 << TRN_DBG_RDY_INC_PH_2TO1_OFFS(4));	/* phase 4 */
1380 
1381 	reg_val &= ~(TRN_DBG_RDY_INC_PH_2TO1_MASK << TRN_DBG_RDY_INC_PH_2TO1_OFFS(5));
1382 	reg_val |= (0x6 << TRN_DBG_RDY_INC_PH_2TO1_OFFS(5));	/* phase 5 */
1383 
1384 	reg_write(TRAINING_DBG_3_REG, reg_val);
1385 
1386 	/*
1387 	 * Axi_bresp_mode[8] = Compliant,
1388 	 * Axi_addr_decode_cntrl[11] = Internal,
1389 	 * Axi_data_bus_width[0] = 128bit
1390 	 * */
1391 	/* 0x14a8 - AXI Control Register */
1392 	reg_write(AXI_CTRL_REG, 0);
1393 
1394 	/*
1395 	 * Stage 2 - Training Values Setup
1396 	 */
1397 	/* Set X-BAR windows for the training sequence */
1398 	ddr3_save_and_set_training_windows(win);
1399 
1400 	return MV_OK;
1401 }
1402 
ddr3_new_tip_dlb_config(void)1403 static int ddr3_new_tip_dlb_config(void)
1404 {
1405 	u32 reg, i = 0;
1406 	struct dlb_config *config_table_ptr = sys_env_dlb_config_ptr_get();
1407 
1408 	/* Write the configuration */
1409 	while (config_table_ptr[i].reg_addr != 0) {
1410 		reg_write(config_table_ptr[i].reg_addr,
1411 			  config_table_ptr[i].reg_data);
1412 		i++;
1413 	}
1414 
1415 #if defined(CONFIG_DDR4)
1416 	reg = reg_read(DUNIT_CTRL_HIGH_REG);
1417 	reg &= ~(CPU_INTERJECTION_ENA_MASK << CPU_INTERJECTION_ENA_OFFS);
1418 	reg |= CPU_INTERJECTION_ENA_SPLIT_DIS << CPU_INTERJECTION_ENA_OFFS;
1419 	reg_write(DUNIT_CTRL_HIGH_REG, reg);
1420 #endif /* CONFIG_DDR4 */
1421 
1422 	/* Enable DLB */
1423 	reg = reg_read(DLB_CTRL_REG);
1424 	reg &= ~(DLB_EN_MASK << DLB_EN_OFFS) &
1425 	       ~(WR_COALESCE_EN_MASK << WR_COALESCE_EN_OFFS) &
1426 	       ~(AXI_PREFETCH_EN_MASK << AXI_PREFETCH_EN_OFFS) &
1427 	       ~(MBUS_PREFETCH_EN_MASK << MBUS_PREFETCH_EN_OFFS) &
1428 	       ~(PREFETCH_NXT_LN_SZ_TRIG_MASK << PREFETCH_NXT_LN_SZ_TRIG_OFFS);
1429 
1430 	reg |= (DLB_EN_ENA << DLB_EN_OFFS) |
1431 	       (WR_COALESCE_EN_ENA << WR_COALESCE_EN_OFFS) |
1432 	       (AXI_PREFETCH_EN_ENA << AXI_PREFETCH_EN_OFFS) |
1433 	       (MBUS_PREFETCH_EN_ENA << MBUS_PREFETCH_EN_OFFS) |
1434 	       (PREFETCH_NXT_LN_SZ_TRIG_ENA << PREFETCH_NXT_LN_SZ_TRIG_OFFS);
1435 
1436 	reg_write(DLB_CTRL_REG, reg);
1437 
1438 	return MV_OK;
1439 }
1440 
mv_ddr_post_training_soc_config(const char * ddr_type)1441 int mv_ddr_post_training_soc_config(const char *ddr_type)
1442 {
1443 	u32 reg_val;
1444 
1445 	/* Restore and set windows */
1446 	ddr3_restore_and_set_final_windows(win, ddr_type);
1447 
1448 	/* Update DRAM init indication in bootROM register */
1449 	reg_val = reg_read(REG_BOOTROM_ROUTINE_ADDR);
1450 	reg_write(REG_BOOTROM_ROUTINE_ADDR,
1451 		  reg_val | (1 << REG_BOOTROM_ROUTINE_DRAM_INIT_OFFS));
1452 
1453 	/* DLB config */
1454 	ddr3_new_tip_dlb_config();
1455 
1456 	return MV_OK;
1457 }
1458 
mv_ddr_mc_config(void)1459 void mv_ddr_mc_config(void)
1460 {
1461 	/* Memory controller initializations */
1462 	struct init_cntr_param init_param;
1463 	int status;
1464 
1465 	init_param.do_mrs_phy = 1;
1466 	init_param.is_ctrl64_bit = 0;
1467 	init_param.init_phy = 1;
1468 	init_param.msys_init = 1;
1469 	status = hws_ddr3_tip_init_controller(0, &init_param);
1470 	if (status != MV_OK)
1471 		printf("DDR3 init controller - FAILED 0x%x\n", status);
1472 
1473 	status = mv_ddr_mc_init();
1474 	if (status != MV_OK)
1475 		printf("DDR3 init_sequence - FAILED 0x%x\n", status);
1476 }
1477 /* function: mv_ddr_mc_init
1478  * this function enables the dunit after init controller configuration
1479  */
mv_ddr_mc_init(void)1480 int mv_ddr_mc_init(void)
1481 {
1482 	CHECK_STATUS(ddr3_tip_enable_init_sequence(0));
1483 
1484 	return MV_OK;
1485 }
1486 
1487 /* function: ddr3_tip_configure_phy
1488  * configures phy and electrical parameters
1489  */
ddr3_tip_configure_phy(u32 dev_num)1490 int ddr3_tip_configure_phy(u32 dev_num)
1491 {
1492 	u32 if_id, phy_id;
1493 	u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE);
1494 	struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get();
1495 
1496 	CHECK_STATUS(ddr3_tip_bus_write
1497 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1498 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_DATA,
1499 		PAD_ZRI_CAL_PHY_REG,
1500 		((0x7f & g_zpri_data) << 7 | (0x7f & g_znri_data))));
1501 	CHECK_STATUS(ddr3_tip_bus_write
1502 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1503 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_CONTROL,
1504 		PAD_ZRI_CAL_PHY_REG,
1505 		((0x7f & g_zpri_ctrl) << 7 | (0x7f & g_znri_ctrl))));
1506 	CHECK_STATUS(ddr3_tip_bus_write
1507 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1508 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_DATA,
1509 		PAD_ODT_CAL_PHY_REG,
1510 		((0x3f & g_zpodt_data) << 6 | (0x3f & g_znodt_data))));
1511 	CHECK_STATUS(ddr3_tip_bus_write
1512 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1513 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_CONTROL,
1514 		PAD_ODT_CAL_PHY_REG,
1515 		((0x3f & g_zpodt_ctrl) << 6 | (0x3f & g_znodt_ctrl))));
1516 
1517 	CHECK_STATUS(ddr3_tip_bus_write
1518 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1519 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_DATA,
1520 		PAD_PRE_DISABLE_PHY_REG, 0));
1521 	CHECK_STATUS(ddr3_tip_bus_write
1522 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1523 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_DATA,
1524 		CMOS_CONFIG_PHY_REG, 0));
1525 	CHECK_STATUS(ddr3_tip_bus_write
1526 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1527 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, DDR_PHY_CONTROL,
1528 		CMOS_CONFIG_PHY_REG, 0));
1529 
1530 	for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) {
1531 		/* check if the interface is enabled */
1532 		VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id);
1533 
1534 		for (phy_id = 0;
1535 			phy_id < octets_per_if_num;
1536 			phy_id++) {
1537 				VALIDATE_BUS_ACTIVE(tm->bus_act_mask, phy_id);
1538 				/* Vref & clamp */
1539 				CHECK_STATUS(ddr3_tip_bus_read_modify_write
1540 					(dev_num, ACCESS_TYPE_UNICAST,
1541 					if_id, phy_id, DDR_PHY_DATA,
1542 					PAD_CFG_PHY_REG,
1543 					((clamp_tbl[if_id] << 4) | vref_init_val),
1544 					((0x7 << 4) | 0x7)));
1545 				/* clamp not relevant for control */
1546 				CHECK_STATUS(ddr3_tip_bus_read_modify_write
1547 					(dev_num, ACCESS_TYPE_UNICAST,
1548 					if_id, phy_id, DDR_PHY_CONTROL,
1549 					PAD_CFG_PHY_REG, 0x4, 0x7));
1550 		}
1551 	}
1552 
1553 	if (ddr3_tip_dev_attr_get(dev_num, MV_ATTR_PHY_EDGE) ==
1554 		MV_DDR_PHY_EDGE_POSITIVE)
1555 		CHECK_STATUS(ddr3_tip_bus_write
1556 		(dev_num, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1557 		ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE,
1558 		DDR_PHY_DATA, 0x90, 0x6002));
1559 
1560 #if defined(CONFIG_DDR4)
1561 	mv_ddr4_phy_config(dev_num);
1562 #endif /* CONFIG_DDR4 */
1563 
1564 	return MV_OK;
1565 }
1566 
1567 #if defined(CONFIG_DDR4)
1568 /* function: ddr4TipCalibrationValidate
1569  * this function validates the calibration values
1570  * the function is per soc due to the different processes the calibration values are different
1571  */
mv_ddr4_calibration_validate(u32 dev_num)1572 int mv_ddr4_calibration_validate(u32 dev_num)
1573 {
1574 	int status = MV_OK;
1575 	u8 if_id = 0;
1576 	u32 read_data[MAX_INTERFACE_NUM];
1577 	u32 cal_n = 0, cal_p = 0;
1578 
1579 	/*
1580 	 * Pad calibration control enable: during training set the calibration to be internal
1581 	 * at the end of the training it should be fixed to external to be configured by the mc6
1582 	 * FIXME: set the calibration to external in the end of the training
1583 	 */
1584 
1585 	/* pad calibration control enable */
1586 	CHECK_STATUS(ddr3_tip_if_write
1587 			(0, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, MAIN_PADS_CAL_MACH_CTRL_REG,
1588 			DYN_PADS_CAL_ENABLE_ENA << DYN_PADS_CAL_ENABLE_OFFS |
1589 			CAL_UPDATE_CTRL_INT << CAL_UPDATE_CTRL_OFFS,
1590 			DYN_PADS_CAL_ENABLE_MASK << DYN_PADS_CAL_ENABLE_OFFS |
1591 			CAL_UPDATE_CTRL_MASK << CAL_UPDATE_CTRL_OFFS));
1592 
1593 	/* Polling initial calibration is done*/
1594 	if (ddr3_tip_if_polling(dev_num, ACCESS_TYPE_UNICAST, if_id,
1595 				CAL_MACH_RDY << CAL_MACH_STATUS_OFFS,
1596 				CAL_MACH_STATUS_MASK << CAL_MACH_STATUS_OFFS,
1597 				MAIN_PADS_CAL_MACH_CTRL_REG, MAX_POLLING_ITERATIONS) != MV_OK)
1598 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, ("ddr4TipCalibrationAdjust: DDR4 calibration poll failed(0)\n"));
1599 
1600 	/* Polling that calibration propagate to io */
1601 	if (ddr3_tip_if_polling(dev_num, ACCESS_TYPE_UNICAST, if_id, 0x3FFFFFF, 0x3FFFFFF, PHY_LOCK_STATUS_REG,
1602 				MAX_POLLING_ITERATIONS) != MV_OK)
1603 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, ("ddr4TipCalibrationAdjust: DDR4 calibration poll failed(1)\n"));
1604 
1605 	/* TODO - debug why polling not enough*/
1606 	mdelay(10);
1607 
1608 	/* pad calibration control disable */
1609 	CHECK_STATUS(ddr3_tip_if_write
1610 			(0, ACCESS_TYPE_MULTICAST, PARAM_NOT_CARE, MAIN_PADS_CAL_MACH_CTRL_REG,
1611 			DYN_PADS_CAL_ENABLE_DIS << DYN_PADS_CAL_ENABLE_OFFS |
1612 			CAL_UPDATE_CTRL_INT << CAL_UPDATE_CTRL_OFFS,
1613 			DYN_PADS_CAL_ENABLE_MASK << DYN_PADS_CAL_ENABLE_OFFS |
1614 			CAL_UPDATE_CTRL_MASK << CAL_UPDATE_CTRL_OFFS));
1615 
1616 	/* Polling initial calibration is done */
1617 	if (ddr3_tip_if_polling(dev_num, ACCESS_TYPE_UNICAST, if_id,
1618 				CAL_MACH_RDY << CAL_MACH_STATUS_OFFS,
1619 				CAL_MACH_STATUS_MASK << CAL_MACH_STATUS_OFFS,
1620 				MAIN_PADS_CAL_MACH_CTRL_REG, MAX_POLLING_ITERATIONS) != MV_OK)
1621 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, ("ddr4TipCalibrationAdjust: DDR4 calibration poll failed(0)\n"));
1622 
1623 	/* Polling that calibration propagate to io */
1624 	if (ddr3_tip_if_polling(dev_num, ACCESS_TYPE_UNICAST, if_id, 0x3FFFFFF, 0x3FFFFFF, PHY_LOCK_STATUS_REG,
1625 				MAX_POLLING_ITERATIONS) != MV_OK)
1626 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, ("ddr4TipCalibrationAdjust: DDR4 calibration poll failed(1)\n"));
1627 
1628 	/* TODO - debug why polling not enough */
1629 	mdelay(10);
1630 
1631 	/* Read Cal value and set to manual val */
1632 	CHECK_STATUS(ddr3_tip_if_read(dev_num, ACCESS_TYPE_UNICAST, if_id, 0x1DC8, read_data, MASK_ALL_BITS));
1633 	cal_n = (read_data[if_id] & ((0x3F) << 10)) >> 10;
1634 	cal_p = (read_data[if_id] & ((0x3F) << 4)) >> 4;
1635 	DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO,
1636 			  ("ddr4TipCalibrationValidate::DDR4 SSTL calib val - Pcal = 0x%x , Ncal = 0x%x\n",
1637 			   cal_p, cal_n));
1638 	if ((cal_n >= 56) || (cal_n <= 6) || (cal_p >= 59) || (cal_p <= 7)) {
1639 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR,
1640 				  ("%s: Error:DDR4 SSTL calib val - Pcal = 0x%x, Ncal = 0x%x are out of range\n",
1641 				  __func__, cal_p, cal_n));
1642 		status = MV_FAIL;
1643 	}
1644 
1645 	/* 14C8 - Vertical */
1646 	CHECK_STATUS(ddr3_tip_if_read(dev_num, ACCESS_TYPE_UNICAST, if_id, 0x14C8, read_data, MASK_ALL_BITS));
1647 	cal_n = (read_data[if_id] & ((0x3F) << 10)) >> 10;
1648 	cal_p = (read_data[if_id] & ((0x3F) << 4)) >> 4;
1649 	DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO,
1650 			  ("ddr4TipCalibrationValidate::DDR4 POD-V calib val - Pcal = 0x%x , Ncal = 0x%x\n",
1651 			  cal_p, cal_n));
1652 	if ((cal_n >= 56) || (cal_n <= 6) || (cal_p >= 59) || (cal_p <= 7)) {
1653 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR,
1654 				  ("%s: Error:DDR4 POD-V calib val - Pcal = 0x%x , Ncal= 0x%x are out of range\n",
1655 				  __func__, cal_p, cal_n));
1656 		status = MV_FAIL;
1657 	}
1658 
1659 	/* 17C8 - Horizontal */
1660 	CHECK_STATUS(ddr3_tip_if_read(dev_num, ACCESS_TYPE_UNICAST, if_id, 0x17C8, read_data, MASK_ALL_BITS));
1661 	cal_n = (read_data[if_id] & ((0x3F) << 10)) >> 10;
1662 	cal_p = (read_data[if_id] & ((0x3F) << 4)) >> 4;
1663 	DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO,
1664 			  ("ddr4TipCalibrationValidate::DDR4 POD-H calib val - Pcal = 0x%x , Ncal = 0x%x\n",
1665 			  cal_p, cal_n));
1666 	if ((cal_n >= 56) || (cal_n <= 6) || (cal_p >= 59) || (cal_p <= 7)) {
1667 		DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR,
1668 				  ("%s: Error:DDR4 POD-H calib val - Pcal = 0x%x, Ncal = 0x%x are out of range\n",
1669 				  __func__, cal_p, cal_n));
1670 		status = MV_FAIL;
1671 	}
1672 
1673 	return status;
1674 }
1675 #endif /* CONFIG_DDR4 */
1676 
mv_ddr_manual_cal_do(void)1677 int mv_ddr_manual_cal_do(void)
1678 {
1679 	return 0;
1680 }
1681