• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (c) 2017-2020, The Linux Foundation. All rights reserved.
4  */
5 
6 #define pr_fmt(fmt)	"[drm-dp] %s: " fmt, __func__
7 
8 #include <linux/rational.h>
9 #include <linux/delay.h>
10 #include <linux/iopoll.h>
11 #include <linux/phy/phy.h>
12 #include <linux/phy/phy-dp.h>
13 #include <linux/rational.h>
14 #include <drm/drm_dp_helper.h>
15 #include <drm/drm_print.h>
16 
17 #include "dp_catalog.h"
18 #include "dp_reg.h"
19 
20 #define POLLING_SLEEP_US			1000
21 #define POLLING_TIMEOUT_US			10000
22 
23 #define SCRAMBLER_RESET_COUNT_VALUE		0xFC
24 
25 #define DP_INTERRUPT_STATUS_ACK_SHIFT	1
26 #define DP_INTERRUPT_STATUS_MASK_SHIFT	2
27 
28 #define MSM_DP_CONTROLLER_AHB_OFFSET	0x0000
29 #define MSM_DP_CONTROLLER_AHB_SIZE	0x0200
30 #define MSM_DP_CONTROLLER_AUX_OFFSET	0x0200
31 #define MSM_DP_CONTROLLER_AUX_SIZE	0x0200
32 #define MSM_DP_CONTROLLER_LINK_OFFSET	0x0400
33 #define MSM_DP_CONTROLLER_LINK_SIZE	0x0C00
34 #define MSM_DP_CONTROLLER_P0_OFFSET	0x1000
35 #define MSM_DP_CONTROLLER_P0_SIZE	0x0400
36 
37 #define DP_INTERRUPT_STATUS1 \
38 	(DP_INTR_AUX_I2C_DONE| \
39 	DP_INTR_WRONG_ADDR | DP_INTR_TIMEOUT | \
40 	DP_INTR_NACK_DEFER | DP_INTR_WRONG_DATA_CNT | \
41 	DP_INTR_I2C_NACK | DP_INTR_I2C_DEFER | \
42 	DP_INTR_PLL_UNLOCKED | DP_INTR_AUX_ERROR)
43 
44 #define DP_INTERRUPT_STATUS1_ACK \
45 	(DP_INTERRUPT_STATUS1 << DP_INTERRUPT_STATUS_ACK_SHIFT)
46 #define DP_INTERRUPT_STATUS1_MASK \
47 	(DP_INTERRUPT_STATUS1 << DP_INTERRUPT_STATUS_MASK_SHIFT)
48 
49 #define DP_INTERRUPT_STATUS2 \
50 	(DP_INTR_READY_FOR_VIDEO | DP_INTR_IDLE_PATTERN_SENT | \
51 	DP_INTR_FRAME_END | DP_INTR_CRC_UPDATED)
52 
53 #define DP_INTERRUPT_STATUS2_ACK \
54 	(DP_INTERRUPT_STATUS2 << DP_INTERRUPT_STATUS_ACK_SHIFT)
55 #define DP_INTERRUPT_STATUS2_MASK \
56 	(DP_INTERRUPT_STATUS2 << DP_INTERRUPT_STATUS_MASK_SHIFT)
57 
58 struct dp_catalog_private {
59 	struct device *dev;
60 	struct dp_io *io;
61 	u32 (*audio_map)[DP_AUDIO_SDP_HEADER_MAX];
62 	struct dp_catalog dp_catalog;
63 	u8 aux_lut_cfg_index[PHY_AUX_CFG_MAX];
64 };
65 
dp_read_aux(struct dp_catalog_private * catalog,u32 offset)66 static inline u32 dp_read_aux(struct dp_catalog_private *catalog, u32 offset)
67 {
68 	offset += MSM_DP_CONTROLLER_AUX_OFFSET;
69 	return readl_relaxed(catalog->io->dp_controller.base + offset);
70 }
71 
dp_write_aux(struct dp_catalog_private * catalog,u32 offset,u32 data)72 static inline void dp_write_aux(struct dp_catalog_private *catalog,
73 			       u32 offset, u32 data)
74 {
75 	offset += MSM_DP_CONTROLLER_AUX_OFFSET;
76 	/*
77 	 * To make sure aux reg writes happens before any other operation,
78 	 * this function uses writel() instread of writel_relaxed()
79 	 */
80 	writel(data, catalog->io->dp_controller.base + offset);
81 }
82 
dp_read_ahb(struct dp_catalog_private * catalog,u32 offset)83 static inline u32 dp_read_ahb(struct dp_catalog_private *catalog, u32 offset)
84 {
85 	offset += MSM_DP_CONTROLLER_AHB_OFFSET;
86 	return readl_relaxed(catalog->io->dp_controller.base + offset);
87 }
88 
dp_write_ahb(struct dp_catalog_private * catalog,u32 offset,u32 data)89 static inline void dp_write_ahb(struct dp_catalog_private *catalog,
90 			       u32 offset, u32 data)
91 {
92 	offset += MSM_DP_CONTROLLER_AHB_OFFSET;
93 	/*
94 	 * To make sure phy reg writes happens before any other operation,
95 	 * this function uses writel() instread of writel_relaxed()
96 	 */
97 	writel(data, catalog->io->dp_controller.base + offset);
98 }
99 
dp_write_p0(struct dp_catalog_private * catalog,u32 offset,u32 data)100 static inline void dp_write_p0(struct dp_catalog_private *catalog,
101 			       u32 offset, u32 data)
102 {
103 	offset += MSM_DP_CONTROLLER_P0_OFFSET;
104 	/*
105 	 * To make sure interface reg writes happens before any other operation,
106 	 * this function uses writel() instread of writel_relaxed()
107 	 */
108 	writel(data, catalog->io->dp_controller.base + offset);
109 }
110 
dp_read_p0(struct dp_catalog_private * catalog,u32 offset)111 static inline u32 dp_read_p0(struct dp_catalog_private *catalog,
112 			       u32 offset)
113 {
114 	offset += MSM_DP_CONTROLLER_P0_OFFSET;
115 	/*
116 	 * To make sure interface reg writes happens before any other operation,
117 	 * this function uses writel() instread of writel_relaxed()
118 	 */
119 	return readl_relaxed(catalog->io->dp_controller.base + offset);
120 }
121 
dp_read_link(struct dp_catalog_private * catalog,u32 offset)122 static inline u32 dp_read_link(struct dp_catalog_private *catalog, u32 offset)
123 {
124 	offset += MSM_DP_CONTROLLER_LINK_OFFSET;
125 	return readl_relaxed(catalog->io->dp_controller.base + offset);
126 }
127 
dp_write_link(struct dp_catalog_private * catalog,u32 offset,u32 data)128 static inline void dp_write_link(struct dp_catalog_private *catalog,
129 			       u32 offset, u32 data)
130 {
131 	offset += MSM_DP_CONTROLLER_LINK_OFFSET;
132 	/*
133 	 * To make sure link reg writes happens before any other operation,
134 	 * this function uses writel() instread of writel_relaxed()
135 	 */
136 	writel(data, catalog->io->dp_controller.base + offset);
137 }
138 
139 /* aux related catalog functions */
dp_catalog_aux_read_data(struct dp_catalog * dp_catalog)140 u32 dp_catalog_aux_read_data(struct dp_catalog *dp_catalog)
141 {
142 	struct dp_catalog_private *catalog = container_of(dp_catalog,
143 				struct dp_catalog_private, dp_catalog);
144 
145 	return dp_read_aux(catalog, REG_DP_AUX_DATA);
146 }
147 
dp_catalog_aux_write_data(struct dp_catalog * dp_catalog)148 int dp_catalog_aux_write_data(struct dp_catalog *dp_catalog)
149 {
150 	struct dp_catalog_private *catalog = container_of(dp_catalog,
151 				struct dp_catalog_private, dp_catalog);
152 
153 	dp_write_aux(catalog, REG_DP_AUX_DATA, dp_catalog->aux_data);
154 	return 0;
155 }
156 
dp_catalog_aux_write_trans(struct dp_catalog * dp_catalog)157 int dp_catalog_aux_write_trans(struct dp_catalog *dp_catalog)
158 {
159 	struct dp_catalog_private *catalog = container_of(dp_catalog,
160 				struct dp_catalog_private, dp_catalog);
161 
162 	dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, dp_catalog->aux_data);
163 	return 0;
164 }
165 
dp_catalog_aux_clear_trans(struct dp_catalog * dp_catalog,bool read)166 int dp_catalog_aux_clear_trans(struct dp_catalog *dp_catalog, bool read)
167 {
168 	u32 data;
169 	struct dp_catalog_private *catalog = container_of(dp_catalog,
170 				struct dp_catalog_private, dp_catalog);
171 
172 	if (read) {
173 		data = dp_read_aux(catalog, REG_DP_AUX_TRANS_CTRL);
174 		data &= ~DP_AUX_TRANS_CTRL_GO;
175 		dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, data);
176 	} else {
177 		dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, 0);
178 	}
179 	return 0;
180 }
181 
dp_catalog_aux_clear_hw_interrupts(struct dp_catalog * dp_catalog)182 int dp_catalog_aux_clear_hw_interrupts(struct dp_catalog *dp_catalog)
183 {
184 	struct dp_catalog_private *catalog = container_of(dp_catalog,
185 				struct dp_catalog_private, dp_catalog);
186 
187 	dp_read_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_STATUS);
188 	dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0x1f);
189 	dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0x9f);
190 	dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0);
191 	return 0;
192 }
193 
dp_catalog_aux_reset(struct dp_catalog * dp_catalog)194 void dp_catalog_aux_reset(struct dp_catalog *dp_catalog)
195 {
196 	u32 aux_ctrl;
197 	struct dp_catalog_private *catalog = container_of(dp_catalog,
198 				struct dp_catalog_private, dp_catalog);
199 
200 	aux_ctrl = dp_read_aux(catalog, REG_DP_AUX_CTRL);
201 
202 	aux_ctrl |= DP_AUX_CTRL_RESET;
203 	dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl);
204 	usleep_range(1000, 1100); /* h/w recommended delay */
205 
206 	aux_ctrl &= ~DP_AUX_CTRL_RESET;
207 	dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl);
208 }
209 
dp_catalog_aux_enable(struct dp_catalog * dp_catalog,bool enable)210 void dp_catalog_aux_enable(struct dp_catalog *dp_catalog, bool enable)
211 {
212 	u32 aux_ctrl;
213 	struct dp_catalog_private *catalog = container_of(dp_catalog,
214 				struct dp_catalog_private, dp_catalog);
215 
216 	aux_ctrl = dp_read_aux(catalog, REG_DP_AUX_CTRL);
217 
218 	if (enable) {
219 		dp_write_aux(catalog, REG_DP_TIMEOUT_COUNT, 0xffff);
220 		dp_write_aux(catalog, REG_DP_AUX_LIMITS, 0xffff);
221 		aux_ctrl |= DP_AUX_CTRL_ENABLE;
222 	} else {
223 		aux_ctrl &= ~DP_AUX_CTRL_ENABLE;
224 	}
225 
226 	dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl);
227 }
228 
dp_catalog_aux_update_cfg(struct dp_catalog * dp_catalog)229 void dp_catalog_aux_update_cfg(struct dp_catalog *dp_catalog)
230 {
231 	struct dp_catalog_private *catalog = container_of(dp_catalog,
232 				struct dp_catalog_private, dp_catalog);
233 	struct dp_io *dp_io = catalog->io;
234 	struct phy *phy = dp_io->phy;
235 
236 	phy_calibrate(phy);
237 }
238 
dump_regs(void __iomem * base,int len)239 static void dump_regs(void __iomem *base, int len)
240 {
241 	int i;
242 	u32 x0, x4, x8, xc;
243 	u32 addr_off = 0;
244 
245 	len = DIV_ROUND_UP(len, 16);
246 	for (i = 0; i < len; i++) {
247 		x0 = readl_relaxed(base + addr_off);
248 		x4 = readl_relaxed(base + addr_off + 0x04);
249 		x8 = readl_relaxed(base + addr_off + 0x08);
250 		xc = readl_relaxed(base + addr_off + 0x0c);
251 
252 		pr_info("%08x: %08x %08x %08x %08x", addr_off, x0, x4, x8, xc);
253 		addr_off += 16;
254 	}
255 }
256 
dp_catalog_dump_regs(struct dp_catalog * dp_catalog)257 void dp_catalog_dump_regs(struct dp_catalog *dp_catalog)
258 {
259 	u32 offset, len;
260 	struct dp_catalog_private *catalog = container_of(dp_catalog,
261 		struct dp_catalog_private, dp_catalog);
262 
263 	pr_info("AHB regs\n");
264 	offset = MSM_DP_CONTROLLER_AHB_OFFSET;
265 	len = MSM_DP_CONTROLLER_AHB_SIZE;
266 	dump_regs(catalog->io->dp_controller.base + offset, len);
267 
268 	pr_info("AUXCLK regs\n");
269 	offset = MSM_DP_CONTROLLER_AUX_OFFSET;
270 	len = MSM_DP_CONTROLLER_AUX_SIZE;
271 	dump_regs(catalog->io->dp_controller.base + offset, len);
272 
273 	pr_info("LCLK regs\n");
274 	offset = MSM_DP_CONTROLLER_LINK_OFFSET;
275 	len = MSM_DP_CONTROLLER_LINK_SIZE;
276 	dump_regs(catalog->io->dp_controller.base + offset, len);
277 
278 	pr_info("P0CLK regs\n");
279 	offset = MSM_DP_CONTROLLER_P0_OFFSET;
280 	len = MSM_DP_CONTROLLER_P0_SIZE;
281 	dump_regs(catalog->io->dp_controller.base + offset, len);
282 }
283 
dp_catalog_aux_get_irq(struct dp_catalog * dp_catalog)284 int dp_catalog_aux_get_irq(struct dp_catalog *dp_catalog)
285 {
286 	struct dp_catalog_private *catalog = container_of(dp_catalog,
287 				struct dp_catalog_private, dp_catalog);
288 	u32 intr, intr_ack;
289 
290 	intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS);
291 	intr &= ~DP_INTERRUPT_STATUS1_MASK;
292 	intr_ack = (intr & DP_INTERRUPT_STATUS1)
293 			<< DP_INTERRUPT_STATUS_ACK_SHIFT;
294 	dp_write_ahb(catalog, REG_DP_INTR_STATUS, intr_ack |
295 			DP_INTERRUPT_STATUS1_MASK);
296 
297 	return intr;
298 
299 }
300 
301 /* controller related catalog functions */
dp_catalog_ctrl_update_transfer_unit(struct dp_catalog * dp_catalog,u32 dp_tu,u32 valid_boundary,u32 valid_boundary2)302 void dp_catalog_ctrl_update_transfer_unit(struct dp_catalog *dp_catalog,
303 				u32 dp_tu, u32 valid_boundary,
304 				u32 valid_boundary2)
305 {
306 	struct dp_catalog_private *catalog = container_of(dp_catalog,
307 				struct dp_catalog_private, dp_catalog);
308 
309 	dp_write_link(catalog, REG_DP_VALID_BOUNDARY, valid_boundary);
310 	dp_write_link(catalog, REG_DP_TU, dp_tu);
311 	dp_write_link(catalog, REG_DP_VALID_BOUNDARY_2, valid_boundary2);
312 }
313 
dp_catalog_ctrl_state_ctrl(struct dp_catalog * dp_catalog,u32 state)314 void dp_catalog_ctrl_state_ctrl(struct dp_catalog *dp_catalog, u32 state)
315 {
316 	struct dp_catalog_private *catalog = container_of(dp_catalog,
317 				struct dp_catalog_private, dp_catalog);
318 
319 	dp_write_link(catalog, REG_DP_STATE_CTRL, state);
320 }
321 
dp_catalog_ctrl_config_ctrl(struct dp_catalog * dp_catalog,u32 cfg)322 void dp_catalog_ctrl_config_ctrl(struct dp_catalog *dp_catalog, u32 cfg)
323 {
324 	struct dp_catalog_private *catalog = container_of(dp_catalog,
325 				struct dp_catalog_private, dp_catalog);
326 
327 	DRM_DEBUG_DP("DP_CONFIGURATION_CTRL=0x%x\n", cfg);
328 
329 	dp_write_link(catalog, REG_DP_CONFIGURATION_CTRL, cfg);
330 }
331 
dp_catalog_ctrl_lane_mapping(struct dp_catalog * dp_catalog)332 void dp_catalog_ctrl_lane_mapping(struct dp_catalog *dp_catalog)
333 {
334 	struct dp_catalog_private *catalog = container_of(dp_catalog,
335 				struct dp_catalog_private, dp_catalog);
336 	u32 ln_0 = 0, ln_1 = 1, ln_2 = 2, ln_3 = 3; /* One-to-One mapping */
337 	u32 ln_mapping;
338 
339 	ln_mapping = ln_0 << LANE0_MAPPING_SHIFT;
340 	ln_mapping |= ln_1 << LANE1_MAPPING_SHIFT;
341 	ln_mapping |= ln_2 << LANE2_MAPPING_SHIFT;
342 	ln_mapping |= ln_3 << LANE3_MAPPING_SHIFT;
343 
344 	dp_write_link(catalog, REG_DP_LOGICAL2PHYSICAL_LANE_MAPPING,
345 			ln_mapping);
346 }
347 
dp_catalog_ctrl_mainlink_ctrl(struct dp_catalog * dp_catalog,bool enable)348 void dp_catalog_ctrl_mainlink_ctrl(struct dp_catalog *dp_catalog,
349 						bool enable)
350 {
351 	u32 mainlink_ctrl;
352 	struct dp_catalog_private *catalog = container_of(dp_catalog,
353 				struct dp_catalog_private, dp_catalog);
354 
355 	if (enable) {
356 		/*
357 		 * To make sure link reg writes happens before other operation,
358 		 * dp_write_link() function uses writel()
359 		 */
360 		mainlink_ctrl = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
361 
362 		mainlink_ctrl &= ~(DP_MAINLINK_CTRL_RESET |
363 						DP_MAINLINK_CTRL_ENABLE);
364 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
365 
366 		mainlink_ctrl |= DP_MAINLINK_CTRL_RESET;
367 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
368 
369 		mainlink_ctrl &= ~DP_MAINLINK_CTRL_RESET;
370 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
371 
372 		mainlink_ctrl |= (DP_MAINLINK_CTRL_ENABLE |
373 					DP_MAINLINK_FB_BOUNDARY_SEL);
374 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
375 	} else {
376 		mainlink_ctrl = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
377 		mainlink_ctrl &= ~DP_MAINLINK_CTRL_ENABLE;
378 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
379 	}
380 }
381 
dp_catalog_ctrl_config_misc(struct dp_catalog * dp_catalog,u32 colorimetry_cfg,u32 test_bits_depth)382 void dp_catalog_ctrl_config_misc(struct dp_catalog *dp_catalog,
383 					u32 colorimetry_cfg,
384 					u32 test_bits_depth)
385 {
386 	u32 misc_val;
387 	struct dp_catalog_private *catalog = container_of(dp_catalog,
388 				struct dp_catalog_private, dp_catalog);
389 
390 	misc_val = dp_read_link(catalog, REG_DP_MISC1_MISC0);
391 
392 	/* clear bpp bits */
393 	misc_val &= ~(0x07 << DP_MISC0_TEST_BITS_DEPTH_SHIFT);
394 	misc_val |= colorimetry_cfg << DP_MISC0_COLORIMETRY_CFG_SHIFT;
395 	misc_val |= test_bits_depth << DP_MISC0_TEST_BITS_DEPTH_SHIFT;
396 	/* Configure clock to synchronous mode */
397 	misc_val |= DP_MISC0_SYNCHRONOUS_CLK;
398 
399 	DRM_DEBUG_DP("misc settings = 0x%x\n", misc_val);
400 	dp_write_link(catalog, REG_DP_MISC1_MISC0, misc_val);
401 }
402 
dp_catalog_ctrl_config_msa(struct dp_catalog * dp_catalog,u32 rate,u32 stream_rate_khz,bool fixed_nvid)403 void dp_catalog_ctrl_config_msa(struct dp_catalog *dp_catalog,
404 					u32 rate, u32 stream_rate_khz,
405 					bool fixed_nvid)
406 {
407 	u32 pixel_m, pixel_n;
408 	u32 mvid, nvid, pixel_div = 0, dispcc_input_rate;
409 	u32 const nvid_fixed = DP_LINK_CONSTANT_N_VALUE;
410 	u32 const link_rate_hbr2 = 540000;
411 	u32 const link_rate_hbr3 = 810000;
412 	unsigned long den, num;
413 
414 	struct dp_catalog_private *catalog = container_of(dp_catalog,
415 				struct dp_catalog_private, dp_catalog);
416 
417 	if (rate == link_rate_hbr3)
418 		pixel_div = 6;
419 	else if (rate == 1620000 || rate == 270000)
420 		pixel_div = 2;
421 	else if (rate == link_rate_hbr2)
422 		pixel_div = 4;
423 	else
424 		DRM_ERROR("Invalid pixel mux divider\n");
425 
426 	dispcc_input_rate = (rate * 10) / pixel_div;
427 
428 	rational_best_approximation(dispcc_input_rate, stream_rate_khz,
429 			(unsigned long)(1 << 16) - 1,
430 			(unsigned long)(1 << 16) - 1, &den, &num);
431 
432 	den = ~(den - num);
433 	den = den & 0xFFFF;
434 	pixel_m = num;
435 	pixel_n = den;
436 
437 	mvid = (pixel_m & 0xFFFF) * 5;
438 	nvid = (0xFFFF & (~pixel_n)) + (pixel_m & 0xFFFF);
439 
440 	if (nvid < nvid_fixed) {
441 		u32 temp;
442 
443 		temp = (nvid_fixed / nvid) * nvid;
444 		mvid = (nvid_fixed / nvid) * mvid;
445 		nvid = temp;
446 	}
447 
448 	if (link_rate_hbr2 == rate)
449 		nvid *= 2;
450 
451 	if (link_rate_hbr3 == rate)
452 		nvid *= 3;
453 
454 	DRM_DEBUG_DP("mvid=0x%x, nvid=0x%x\n", mvid, nvid);
455 	dp_write_link(catalog, REG_DP_SOFTWARE_MVID, mvid);
456 	dp_write_link(catalog, REG_DP_SOFTWARE_NVID, nvid);
457 	dp_write_p0(catalog, MMSS_DP_DSC_DTO, 0x0);
458 }
459 
dp_catalog_ctrl_set_pattern(struct dp_catalog * dp_catalog,u32 pattern)460 int dp_catalog_ctrl_set_pattern(struct dp_catalog *dp_catalog,
461 					u32 pattern)
462 {
463 	int bit, ret;
464 	u32 data;
465 	struct dp_catalog_private *catalog = container_of(dp_catalog,
466 				struct dp_catalog_private, dp_catalog);
467 
468 	bit = BIT(pattern - 1);
469 	DRM_DEBUG_DP("hw: bit=%d train=%d\n", bit, pattern);
470 	dp_catalog_ctrl_state_ctrl(dp_catalog, bit);
471 
472 	bit = BIT(pattern - 1) << DP_MAINLINK_READY_LINK_TRAINING_SHIFT;
473 
474 	/* Poll for mainlink ready status */
475 	ret = readx_poll_timeout(readl, catalog->io->dp_controller.base +
476 					MSM_DP_CONTROLLER_LINK_OFFSET +
477 					REG_DP_MAINLINK_READY,
478 					data, data & bit,
479 					POLLING_SLEEP_US, POLLING_TIMEOUT_US);
480 	if (ret < 0) {
481 		DRM_ERROR("set pattern for link_train=%d failed\n", pattern);
482 		return ret;
483 	}
484 	return 0;
485 }
486 
dp_catalog_ctrl_reset(struct dp_catalog * dp_catalog)487 void dp_catalog_ctrl_reset(struct dp_catalog *dp_catalog)
488 {
489 	u32 sw_reset;
490 	struct dp_catalog_private *catalog = container_of(dp_catalog,
491 				struct dp_catalog_private, dp_catalog);
492 
493 	sw_reset = dp_read_ahb(catalog, REG_DP_SW_RESET);
494 
495 	sw_reset |= DP_SW_RESET;
496 	dp_write_ahb(catalog, REG_DP_SW_RESET, sw_reset);
497 	usleep_range(1000, 1100); /* h/w recommended delay */
498 
499 	sw_reset &= ~DP_SW_RESET;
500 	dp_write_ahb(catalog, REG_DP_SW_RESET, sw_reset);
501 }
502 
dp_catalog_ctrl_mainlink_ready(struct dp_catalog * dp_catalog)503 bool dp_catalog_ctrl_mainlink_ready(struct dp_catalog *dp_catalog)
504 {
505 	u32 data;
506 	int ret;
507 	struct dp_catalog_private *catalog = container_of(dp_catalog,
508 				struct dp_catalog_private, dp_catalog);
509 
510 	/* Poll for mainlink ready status */
511 	ret = readl_poll_timeout(catalog->io->dp_controller.base +
512 				MSM_DP_CONTROLLER_LINK_OFFSET +
513 				REG_DP_MAINLINK_READY,
514 				data, data & DP_MAINLINK_READY_FOR_VIDEO,
515 				POLLING_SLEEP_US, POLLING_TIMEOUT_US);
516 	if (ret < 0) {
517 		DRM_ERROR("mainlink not ready\n");
518 		return false;
519 	}
520 
521 	return true;
522 }
523 
dp_catalog_ctrl_enable_irq(struct dp_catalog * dp_catalog,bool enable)524 void dp_catalog_ctrl_enable_irq(struct dp_catalog *dp_catalog,
525 						bool enable)
526 {
527 	struct dp_catalog_private *catalog = container_of(dp_catalog,
528 				struct dp_catalog_private, dp_catalog);
529 
530 	if (enable) {
531 		dp_write_ahb(catalog, REG_DP_INTR_STATUS,
532 				DP_INTERRUPT_STATUS1_MASK);
533 		dp_write_ahb(catalog, REG_DP_INTR_STATUS2,
534 				DP_INTERRUPT_STATUS2_MASK);
535 	} else {
536 		dp_write_ahb(catalog, REG_DP_INTR_STATUS, 0x00);
537 		dp_write_ahb(catalog, REG_DP_INTR_STATUS2, 0x00);
538 	}
539 }
540 
dp_catalog_hpd_config_intr(struct dp_catalog * dp_catalog,u32 intr_mask,bool en)541 void dp_catalog_hpd_config_intr(struct dp_catalog *dp_catalog,
542 			u32 intr_mask, bool en)
543 {
544 	struct dp_catalog_private *catalog = container_of(dp_catalog,
545 				struct dp_catalog_private, dp_catalog);
546 
547 	u32 config = dp_read_aux(catalog, REG_DP_DP_HPD_INT_MASK);
548 
549 	config = (en ? config | intr_mask : config & ~intr_mask);
550 
551 	dp_write_aux(catalog, REG_DP_DP_HPD_INT_MASK,
552 				config & DP_DP_HPD_INT_MASK);
553 }
554 
dp_catalog_ctrl_hpd_config(struct dp_catalog * dp_catalog)555 void dp_catalog_ctrl_hpd_config(struct dp_catalog *dp_catalog)
556 {
557 	struct dp_catalog_private *catalog = container_of(dp_catalog,
558 				struct dp_catalog_private, dp_catalog);
559 
560 	u32 reftimer = dp_read_aux(catalog, REG_DP_DP_HPD_REFTIMER);
561 
562 	/* enable HPD interrupts */
563 	dp_catalog_hpd_config_intr(dp_catalog,
564 		DP_DP_HPD_PLUG_INT_MASK | DP_DP_IRQ_HPD_INT_MASK
565 		| DP_DP_HPD_UNPLUG_INT_MASK | DP_DP_HPD_REPLUG_INT_MASK, true);
566 
567 	/* Configure REFTIMER and enable it */
568 	reftimer |= DP_DP_HPD_REFTIMER_ENABLE;
569 	dp_write_aux(catalog, REG_DP_DP_HPD_REFTIMER, reftimer);
570 
571 	/* Enable HPD */
572 	dp_write_aux(catalog, REG_DP_DP_HPD_CTRL, DP_DP_HPD_CTRL_HPD_EN);
573 }
574 
dp_catalog_hpd_get_state_status(struct dp_catalog * dp_catalog)575 u32 dp_catalog_hpd_get_state_status(struct dp_catalog *dp_catalog)
576 {
577 	struct dp_catalog_private *catalog = container_of(dp_catalog,
578 				struct dp_catalog_private, dp_catalog);
579 	u32 status;
580 
581 	status = dp_read_aux(catalog, REG_DP_DP_HPD_INT_STATUS);
582 	status >>= DP_DP_HPD_STATE_STATUS_BITS_SHIFT;
583 	status &= DP_DP_HPD_STATE_STATUS_BITS_MASK;
584 
585 	return status;
586 }
587 
dp_catalog_hpd_get_intr_status(struct dp_catalog * dp_catalog)588 u32 dp_catalog_hpd_get_intr_status(struct dp_catalog *dp_catalog)
589 {
590 	struct dp_catalog_private *catalog = container_of(dp_catalog,
591 				struct dp_catalog_private, dp_catalog);
592 	int isr = 0;
593 
594 	isr = dp_read_aux(catalog, REG_DP_DP_HPD_INT_STATUS);
595 	dp_write_aux(catalog, REG_DP_DP_HPD_INT_ACK,
596 				 (isr & DP_DP_HPD_INT_MASK));
597 
598 	return isr;
599 }
600 
dp_catalog_ctrl_get_interrupt(struct dp_catalog * dp_catalog)601 int dp_catalog_ctrl_get_interrupt(struct dp_catalog *dp_catalog)
602 {
603 	struct dp_catalog_private *catalog = container_of(dp_catalog,
604 				struct dp_catalog_private, dp_catalog);
605 	u32 intr, intr_ack;
606 
607 	intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS2);
608 	intr &= ~DP_INTERRUPT_STATUS2_MASK;
609 	intr_ack = (intr & DP_INTERRUPT_STATUS2)
610 			<< DP_INTERRUPT_STATUS_ACK_SHIFT;
611 	dp_write_ahb(catalog, REG_DP_INTR_STATUS2,
612 			intr_ack | DP_INTERRUPT_STATUS2_MASK);
613 
614 	return intr;
615 }
616 
dp_catalog_ctrl_phy_reset(struct dp_catalog * dp_catalog)617 void dp_catalog_ctrl_phy_reset(struct dp_catalog *dp_catalog)
618 {
619 	struct dp_catalog_private *catalog = container_of(dp_catalog,
620 				struct dp_catalog_private, dp_catalog);
621 
622 	dp_write_ahb(catalog, REG_DP_PHY_CTRL,
623 			DP_PHY_CTRL_SW_RESET | DP_PHY_CTRL_SW_RESET_PLL);
624 	usleep_range(1000, 1100); /* h/w recommended delay */
625 	dp_write_ahb(catalog, REG_DP_PHY_CTRL, 0x0);
626 }
627 
dp_catalog_ctrl_update_vx_px(struct dp_catalog * dp_catalog,u8 v_level,u8 p_level)628 int dp_catalog_ctrl_update_vx_px(struct dp_catalog *dp_catalog,
629 		u8 v_level, u8 p_level)
630 {
631 	struct dp_catalog_private *catalog = container_of(dp_catalog,
632 				struct dp_catalog_private, dp_catalog);
633 	struct dp_io *dp_io = catalog->io;
634 	struct phy *phy = dp_io->phy;
635 	struct phy_configure_opts_dp *opts_dp = &dp_io->phy_opts.dp;
636 
637 	/* TODO: Update for all lanes instead of just first one */
638 	opts_dp->voltage[0] = v_level;
639 	opts_dp->pre[0] = p_level;
640 	opts_dp->set_voltages = 1;
641 	phy_configure(phy, &dp_io->phy_opts);
642 	opts_dp->set_voltages = 0;
643 
644 	return 0;
645 }
646 
dp_catalog_ctrl_send_phy_pattern(struct dp_catalog * dp_catalog,u32 pattern)647 void dp_catalog_ctrl_send_phy_pattern(struct dp_catalog *dp_catalog,
648 			u32 pattern)
649 {
650 	struct dp_catalog_private *catalog = container_of(dp_catalog,
651 				struct dp_catalog_private, dp_catalog);
652 	u32 value = 0x0;
653 
654 	/* Make sure to clear the current pattern before starting a new one */
655 	dp_write_link(catalog, REG_DP_STATE_CTRL, 0x0);
656 
657 	switch (pattern) {
658 	case DP_PHY_TEST_PATTERN_D10_2:
659 		dp_write_link(catalog, REG_DP_STATE_CTRL,
660 				DP_STATE_CTRL_LINK_TRAINING_PATTERN1);
661 		break;
662 	case DP_PHY_TEST_PATTERN_ERROR_COUNT:
663 		value &= ~(1 << 16);
664 		dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
665 					value);
666 		value |= SCRAMBLER_RESET_COUNT_VALUE;
667 		dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
668 					value);
669 		dp_write_link(catalog, REG_DP_MAINLINK_LEVELS,
670 					DP_MAINLINK_SAFE_TO_EXIT_LEVEL_2);
671 		dp_write_link(catalog, REG_DP_STATE_CTRL,
672 					DP_STATE_CTRL_LINK_SYMBOL_ERR_MEASURE);
673 		break;
674 	case DP_PHY_TEST_PATTERN_PRBS7:
675 		dp_write_link(catalog, REG_DP_STATE_CTRL,
676 				DP_STATE_CTRL_LINK_PRBS7);
677 		break;
678 	case DP_PHY_TEST_PATTERN_80BIT_CUSTOM:
679 		dp_write_link(catalog, REG_DP_STATE_CTRL,
680 				DP_STATE_CTRL_LINK_TEST_CUSTOM_PATTERN);
681 		/* 00111110000011111000001111100000 */
682 		dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG0,
683 				0x3E0F83E0);
684 		/* 00001111100000111110000011111000 */
685 		dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG1,
686 				0x0F83E0F8);
687 		/* 1111100000111110 */
688 		dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG2,
689 				0x0000F83E);
690 		break;
691 	case DP_PHY_TEST_PATTERN_CP2520:
692 		value = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
693 		value &= ~DP_MAINLINK_CTRL_SW_BYPASS_SCRAMBLER;
694 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, value);
695 
696 		value = DP_HBR2_ERM_PATTERN;
697 		dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
698 				value);
699 		value |= SCRAMBLER_RESET_COUNT_VALUE;
700 		dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
701 					value);
702 		dp_write_link(catalog, REG_DP_MAINLINK_LEVELS,
703 					DP_MAINLINK_SAFE_TO_EXIT_LEVEL_2);
704 		dp_write_link(catalog, REG_DP_STATE_CTRL,
705 					DP_STATE_CTRL_LINK_SYMBOL_ERR_MEASURE);
706 		value = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
707 		value |= DP_MAINLINK_CTRL_ENABLE;
708 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL, value);
709 		break;
710 	case DP_PHY_TEST_PATTERN_SEL_MASK:
711 		dp_write_link(catalog, REG_DP_MAINLINK_CTRL,
712 				DP_MAINLINK_CTRL_ENABLE);
713 		dp_write_link(catalog, REG_DP_STATE_CTRL,
714 				DP_STATE_CTRL_LINK_TRAINING_PATTERN4);
715 		break;
716 	default:
717 		DRM_DEBUG_DP("No valid test pattern requested:0x%x\n", pattern);
718 		break;
719 	}
720 }
721 
dp_catalog_ctrl_read_phy_pattern(struct dp_catalog * dp_catalog)722 u32 dp_catalog_ctrl_read_phy_pattern(struct dp_catalog *dp_catalog)
723 {
724 	struct dp_catalog_private *catalog = container_of(dp_catalog,
725 				struct dp_catalog_private, dp_catalog);
726 
727 	return dp_read_link(catalog, REG_DP_MAINLINK_READY);
728 }
729 
730 /* panel related catalog functions */
dp_catalog_panel_timing_cfg(struct dp_catalog * dp_catalog)731 int dp_catalog_panel_timing_cfg(struct dp_catalog *dp_catalog)
732 {
733 	struct dp_catalog_private *catalog = container_of(dp_catalog,
734 				struct dp_catalog_private, dp_catalog);
735 
736 	dp_write_link(catalog, REG_DP_TOTAL_HOR_VER,
737 				dp_catalog->total);
738 	dp_write_link(catalog, REG_DP_START_HOR_VER_FROM_SYNC,
739 				dp_catalog->sync_start);
740 	dp_write_link(catalog, REG_DP_HSYNC_VSYNC_WIDTH_POLARITY,
741 				dp_catalog->width_blanking);
742 	dp_write_link(catalog, REG_DP_ACTIVE_HOR_VER, dp_catalog->dp_active);
743 	dp_write_p0(catalog, MMSS_DP_INTF_CONFIG, 0);
744 	return 0;
745 }
746 
dp_catalog_panel_tpg_enable(struct dp_catalog * dp_catalog,struct drm_display_mode * drm_mode)747 void dp_catalog_panel_tpg_enable(struct dp_catalog *dp_catalog,
748 				struct drm_display_mode *drm_mode)
749 {
750 	struct dp_catalog_private *catalog = container_of(dp_catalog,
751 				struct dp_catalog_private, dp_catalog);
752 	u32 hsync_period, vsync_period;
753 	u32 display_v_start, display_v_end;
754 	u32 hsync_start_x, hsync_end_x;
755 	u32 v_sync_width;
756 	u32 hsync_ctl;
757 	u32 display_hctl;
758 
759 	/* TPG config parameters*/
760 	hsync_period = drm_mode->htotal;
761 	vsync_period = drm_mode->vtotal;
762 
763 	display_v_start = ((drm_mode->vtotal - drm_mode->vsync_start) *
764 					hsync_period);
765 	display_v_end = ((vsync_period - (drm_mode->vsync_start -
766 					drm_mode->vdisplay))
767 					* hsync_period) - 1;
768 
769 	display_v_start += drm_mode->htotal - drm_mode->hsync_start;
770 	display_v_end -= (drm_mode->hsync_start - drm_mode->hdisplay);
771 
772 	hsync_start_x = drm_mode->htotal - drm_mode->hsync_start;
773 	hsync_end_x = hsync_period - (drm_mode->hsync_start -
774 					drm_mode->hdisplay) - 1;
775 
776 	v_sync_width = drm_mode->vsync_end - drm_mode->vsync_start;
777 
778 	hsync_ctl = (hsync_period << 16) |
779 			(drm_mode->hsync_end - drm_mode->hsync_start);
780 	display_hctl = (hsync_end_x << 16) | hsync_start_x;
781 
782 
783 	dp_write_p0(catalog, MMSS_DP_INTF_CONFIG, 0x0);
784 	dp_write_p0(catalog, MMSS_DP_INTF_HSYNC_CTL, hsync_ctl);
785 	dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PERIOD_F0, vsync_period *
786 			hsync_period);
787 	dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PULSE_WIDTH_F0, v_sync_width *
788 			hsync_period);
789 	dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PERIOD_F1, 0);
790 	dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PULSE_WIDTH_F1, 0);
791 	dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_HCTL, display_hctl);
792 	dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_HCTL, 0);
793 	dp_write_p0(catalog, MMSS_INTF_DISPLAY_V_START_F0, display_v_start);
794 	dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_V_END_F0, display_v_end);
795 	dp_write_p0(catalog, MMSS_INTF_DISPLAY_V_START_F1, 0);
796 	dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_V_END_F1, 0);
797 	dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_START_F0, 0);
798 	dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_END_F0, 0);
799 	dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_START_F1, 0);
800 	dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_END_F1, 0);
801 	dp_write_p0(catalog, MMSS_DP_INTF_POLARITY_CTL, 0);
802 
803 	dp_write_p0(catalog, MMSS_DP_TPG_MAIN_CONTROL,
804 				DP_TPG_CHECKERED_RECT_PATTERN);
805 	dp_write_p0(catalog, MMSS_DP_TPG_VIDEO_CONFIG,
806 				DP_TPG_VIDEO_CONFIG_BPP_8BIT |
807 				DP_TPG_VIDEO_CONFIG_RGB);
808 	dp_write_p0(catalog, MMSS_DP_BIST_ENABLE,
809 				DP_BIST_ENABLE_DPBIST_EN);
810 	dp_write_p0(catalog, MMSS_DP_TIMING_ENGINE_EN,
811 				DP_TIMING_ENGINE_EN_EN);
812 	DRM_DEBUG_DP("%s: enabled tpg\n", __func__);
813 }
814 
dp_catalog_panel_tpg_disable(struct dp_catalog * dp_catalog)815 void dp_catalog_panel_tpg_disable(struct dp_catalog *dp_catalog)
816 {
817 	struct dp_catalog_private *catalog = container_of(dp_catalog,
818 				struct dp_catalog_private, dp_catalog);
819 
820 	dp_write_p0(catalog, MMSS_DP_TPG_MAIN_CONTROL, 0x0);
821 	dp_write_p0(catalog, MMSS_DP_BIST_ENABLE, 0x0);
822 	dp_write_p0(catalog, MMSS_DP_TIMING_ENGINE_EN, 0x0);
823 }
824 
dp_catalog_get(struct device * dev,struct dp_io * io)825 struct dp_catalog *dp_catalog_get(struct device *dev, struct dp_io *io)
826 {
827 	struct dp_catalog_private *catalog;
828 
829 	if (!io) {
830 		DRM_ERROR("invalid input\n");
831 		return ERR_PTR(-EINVAL);
832 	}
833 
834 	catalog  = devm_kzalloc(dev, sizeof(*catalog), GFP_KERNEL);
835 	if (!catalog)
836 		return ERR_PTR(-ENOMEM);
837 
838 	catalog->dev = dev;
839 	catalog->io = io;
840 
841 	return &catalog->dp_catalog;
842 }
843 
dp_catalog_audio_get_header(struct dp_catalog * dp_catalog)844 void dp_catalog_audio_get_header(struct dp_catalog *dp_catalog)
845 {
846 	struct dp_catalog_private *catalog;
847 	u32 (*sdp_map)[DP_AUDIO_SDP_HEADER_MAX];
848 	enum dp_catalog_audio_sdp_type sdp;
849 	enum dp_catalog_audio_header_type header;
850 
851 	if (!dp_catalog)
852 		return;
853 
854 	catalog = container_of(dp_catalog,
855 		struct dp_catalog_private, dp_catalog);
856 
857 	sdp_map = catalog->audio_map;
858 	sdp     = dp_catalog->sdp_type;
859 	header  = dp_catalog->sdp_header;
860 
861 	dp_catalog->audio_data = dp_read_link(catalog,
862 			sdp_map[sdp][header]);
863 }
864 
dp_catalog_audio_set_header(struct dp_catalog * dp_catalog)865 void dp_catalog_audio_set_header(struct dp_catalog *dp_catalog)
866 {
867 	struct dp_catalog_private *catalog;
868 	u32 (*sdp_map)[DP_AUDIO_SDP_HEADER_MAX];
869 	enum dp_catalog_audio_sdp_type sdp;
870 	enum dp_catalog_audio_header_type header;
871 	u32 data;
872 
873 	if (!dp_catalog)
874 		return;
875 
876 	catalog = container_of(dp_catalog,
877 		struct dp_catalog_private, dp_catalog);
878 
879 	sdp_map = catalog->audio_map;
880 	sdp     = dp_catalog->sdp_type;
881 	header  = dp_catalog->sdp_header;
882 	data    = dp_catalog->audio_data;
883 
884 	dp_write_link(catalog, sdp_map[sdp][header], data);
885 }
886 
dp_catalog_audio_config_acr(struct dp_catalog * dp_catalog)887 void dp_catalog_audio_config_acr(struct dp_catalog *dp_catalog)
888 {
889 	struct dp_catalog_private *catalog;
890 	u32 acr_ctrl, select;
891 
892 	if (!dp_catalog)
893 		return;
894 
895 	catalog = container_of(dp_catalog,
896 		struct dp_catalog_private, dp_catalog);
897 
898 	select = dp_catalog->audio_data;
899 	acr_ctrl = select << 4 | BIT(31) | BIT(8) | BIT(14);
900 
901 	DRM_DEBUG_DP("select = 0x%x, acr_ctrl = 0x%x\n", select, acr_ctrl);
902 
903 	dp_write_link(catalog, MMSS_DP_AUDIO_ACR_CTRL, acr_ctrl);
904 }
905 
dp_catalog_audio_enable(struct dp_catalog * dp_catalog)906 void dp_catalog_audio_enable(struct dp_catalog *dp_catalog)
907 {
908 	struct dp_catalog_private *catalog;
909 	bool enable;
910 	u32 audio_ctrl;
911 
912 	if (!dp_catalog)
913 		return;
914 
915 	catalog = container_of(dp_catalog,
916 		struct dp_catalog_private, dp_catalog);
917 
918 	enable = !!dp_catalog->audio_data;
919 	audio_ctrl = dp_read_link(catalog, MMSS_DP_AUDIO_CFG);
920 
921 	if (enable)
922 		audio_ctrl |= BIT(0);
923 	else
924 		audio_ctrl &= ~BIT(0);
925 
926 	DRM_DEBUG_DP("dp_audio_cfg = 0x%x\n", audio_ctrl);
927 
928 	dp_write_link(catalog, MMSS_DP_AUDIO_CFG, audio_ctrl);
929 	/* make sure audio engine is disabled */
930 	wmb();
931 }
932 
dp_catalog_audio_config_sdp(struct dp_catalog * dp_catalog)933 void dp_catalog_audio_config_sdp(struct dp_catalog *dp_catalog)
934 {
935 	struct dp_catalog_private *catalog;
936 	u32 sdp_cfg = 0;
937 	u32 sdp_cfg2 = 0;
938 
939 	if (!dp_catalog)
940 		return;
941 
942 	catalog = container_of(dp_catalog,
943 		struct dp_catalog_private, dp_catalog);
944 
945 	sdp_cfg = dp_read_link(catalog, MMSS_DP_SDP_CFG);
946 	/* AUDIO_TIMESTAMP_SDP_EN */
947 	sdp_cfg |= BIT(1);
948 	/* AUDIO_STREAM_SDP_EN */
949 	sdp_cfg |= BIT(2);
950 	/* AUDIO_COPY_MANAGEMENT_SDP_EN */
951 	sdp_cfg |= BIT(5);
952 	/* AUDIO_ISRC_SDP_EN  */
953 	sdp_cfg |= BIT(6);
954 	/* AUDIO_INFOFRAME_SDP_EN  */
955 	sdp_cfg |= BIT(20);
956 
957 	DRM_DEBUG_DP("sdp_cfg = 0x%x\n", sdp_cfg);
958 
959 	dp_write_link(catalog, MMSS_DP_SDP_CFG, sdp_cfg);
960 
961 	sdp_cfg2 = dp_read_link(catalog, MMSS_DP_SDP_CFG2);
962 	/* IFRM_REGSRC -> Do not use reg values */
963 	sdp_cfg2 &= ~BIT(0);
964 	/* AUDIO_STREAM_HB3_REGSRC-> Do not use reg values */
965 	sdp_cfg2 &= ~BIT(1);
966 
967 	DRM_DEBUG_DP("sdp_cfg2 = 0x%x\n", sdp_cfg2);
968 
969 	dp_write_link(catalog, MMSS_DP_SDP_CFG2, sdp_cfg2);
970 }
971 
dp_catalog_audio_init(struct dp_catalog * dp_catalog)972 void dp_catalog_audio_init(struct dp_catalog *dp_catalog)
973 {
974 	struct dp_catalog_private *catalog;
975 
976 	static u32 sdp_map[][DP_AUDIO_SDP_HEADER_MAX] = {
977 		{
978 			MMSS_DP_AUDIO_STREAM_0,
979 			MMSS_DP_AUDIO_STREAM_1,
980 			MMSS_DP_AUDIO_STREAM_1,
981 		},
982 		{
983 			MMSS_DP_AUDIO_TIMESTAMP_0,
984 			MMSS_DP_AUDIO_TIMESTAMP_1,
985 			MMSS_DP_AUDIO_TIMESTAMP_1,
986 		},
987 		{
988 			MMSS_DP_AUDIO_INFOFRAME_0,
989 			MMSS_DP_AUDIO_INFOFRAME_1,
990 			MMSS_DP_AUDIO_INFOFRAME_1,
991 		},
992 		{
993 			MMSS_DP_AUDIO_COPYMANAGEMENT_0,
994 			MMSS_DP_AUDIO_COPYMANAGEMENT_1,
995 			MMSS_DP_AUDIO_COPYMANAGEMENT_1,
996 		},
997 		{
998 			MMSS_DP_AUDIO_ISRC_0,
999 			MMSS_DP_AUDIO_ISRC_1,
1000 			MMSS_DP_AUDIO_ISRC_1,
1001 		},
1002 	};
1003 
1004 	if (!dp_catalog)
1005 		return;
1006 
1007 	catalog = container_of(dp_catalog,
1008 		struct dp_catalog_private, dp_catalog);
1009 
1010 	catalog->audio_map = sdp_map;
1011 }
1012 
dp_catalog_audio_sfe_level(struct dp_catalog * dp_catalog)1013 void dp_catalog_audio_sfe_level(struct dp_catalog *dp_catalog)
1014 {
1015 	struct dp_catalog_private *catalog;
1016 	u32 mainlink_levels, safe_to_exit_level;
1017 
1018 	if (!dp_catalog)
1019 		return;
1020 
1021 	catalog = container_of(dp_catalog,
1022 		struct dp_catalog_private, dp_catalog);
1023 
1024 	safe_to_exit_level = dp_catalog->audio_data;
1025 	mainlink_levels = dp_read_link(catalog, REG_DP_MAINLINK_LEVELS);
1026 	mainlink_levels &= 0xFE0;
1027 	mainlink_levels |= safe_to_exit_level;
1028 
1029 	DRM_DEBUG_DP("mainlink_level = 0x%x, safe_to_exit_level = 0x%x\n",
1030 			 mainlink_levels, safe_to_exit_level);
1031 
1032 	dp_write_link(catalog, REG_DP_MAINLINK_LEVELS, mainlink_levels);
1033 }
1034