1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (c) 2017-2020, The Linux Foundation. All rights reserved.
6 #define pr_fmt(fmt) "[drm-dp] %s: " fmt, __func__
8 #include <linux/delay.h>
9 #include <linux/iopoll.h>
10 #include <linux/phy/phy.h>
11 #include <linux/phy/phy-dp.h>
12 #include <linux/rational.h>
13 #include <drm/display/drm_dp_helper.h>
14 #include <drm/drm_print.h>
16 #include "dp_catalog.h"
19 #define POLLING_SLEEP_US 1000
20 #define POLLING_TIMEOUT_US 10000
22 #define SCRAMBLER_RESET_COUNT_VALUE 0xFC
24 #define DP_INTERRUPT_STATUS_ACK_SHIFT 1
25 #define DP_INTERRUPT_STATUS_MASK_SHIFT 2
27 #define DP_INTF_CONFIG_DATABUS_WIDEN BIT(4)
29 #define DP_INTERRUPT_STATUS1 \
30 (DP_INTR_AUX_XFER_DONE| \
31 DP_INTR_WRONG_ADDR | DP_INTR_TIMEOUT | \
32 DP_INTR_NACK_DEFER | DP_INTR_WRONG_DATA_CNT | \
33 DP_INTR_I2C_NACK | DP_INTR_I2C_DEFER | \
34 DP_INTR_PLL_UNLOCKED | DP_INTR_AUX_ERROR)
36 #define DP_INTERRUPT_STATUS1_ACK \
37 (DP_INTERRUPT_STATUS1 << DP_INTERRUPT_STATUS_ACK_SHIFT)
38 #define DP_INTERRUPT_STATUS1_MASK \
39 (DP_INTERRUPT_STATUS1 << DP_INTERRUPT_STATUS_MASK_SHIFT)
41 #define DP_INTERRUPT_STATUS2 \
42 (DP_INTR_READY_FOR_VIDEO | DP_INTR_IDLE_PATTERN_SENT | \
43 DP_INTR_FRAME_END | DP_INTR_CRC_UPDATED)
45 #define DP_INTERRUPT_STATUS2_ACK \
46 (DP_INTERRUPT_STATUS2 << DP_INTERRUPT_STATUS_ACK_SHIFT)
47 #define DP_INTERRUPT_STATUS2_MASK \
48 (DP_INTERRUPT_STATUS2 << DP_INTERRUPT_STATUS_MASK_SHIFT)
50 #define DP_INTERRUPT_STATUS4 \
51 (PSR_UPDATE_INT | PSR_CAPTURE_INT | PSR_EXIT_INT | \
52 PSR_UPDATE_ERROR_INT | PSR_WAKE_ERROR_INT)
54 #define DP_INTERRUPT_MASK4 \
55 (PSR_UPDATE_MASK | PSR_CAPTURE_MASK | PSR_EXIT_MASK | \
56 PSR_UPDATE_ERROR_MASK | PSR_WAKE_ERROR_MASK)
58 struct dp_catalog_private {
60 struct drm_device *drm_dev;
62 u32 (*audio_map)[DP_AUDIO_SDP_HEADER_MAX];
63 struct dp_catalog dp_catalog;
64 u8 aux_lut_cfg_index[PHY_AUX_CFG_MAX];
67 void dp_catalog_snapshot(struct dp_catalog *dp_catalog, struct msm_disp_state *disp_state)
69 struct dp_catalog_private *catalog = container_of(dp_catalog,
70 struct dp_catalog_private, dp_catalog);
71 struct dss_io_data *dss = &catalog->io->dp_controller;
73 msm_disp_snapshot_add_block(disp_state, dss->ahb.len, dss->ahb.base, "dp_ahb");
74 msm_disp_snapshot_add_block(disp_state, dss->aux.len, dss->aux.base, "dp_aux");
75 msm_disp_snapshot_add_block(disp_state, dss->link.len, dss->link.base, "dp_link");
76 msm_disp_snapshot_add_block(disp_state, dss->p0.len, dss->p0.base, "dp_p0");
79 static inline u32 dp_read_aux(struct dp_catalog_private *catalog, u32 offset)
81 return readl_relaxed(catalog->io->dp_controller.aux.base + offset);
84 static inline void dp_write_aux(struct dp_catalog_private *catalog,
88 * To make sure aux reg writes happens before any other operation,
89 * this function uses writel() instread of writel_relaxed()
91 writel(data, catalog->io->dp_controller.aux.base + offset);
94 static inline u32 dp_read_ahb(const struct dp_catalog_private *catalog, u32 offset)
96 return readl_relaxed(catalog->io->dp_controller.ahb.base + offset);
99 static inline void dp_write_ahb(struct dp_catalog_private *catalog,
100 u32 offset, u32 data)
103 * To make sure phy reg writes happens before any other operation,
104 * this function uses writel() instread of writel_relaxed()
106 writel(data, catalog->io->dp_controller.ahb.base + offset);
109 static inline void dp_write_p0(struct dp_catalog_private *catalog,
110 u32 offset, u32 data)
113 * To make sure interface reg writes happens before any other operation,
114 * this function uses writel() instread of writel_relaxed()
116 writel(data, catalog->io->dp_controller.p0.base + offset);
119 static inline u32 dp_read_p0(struct dp_catalog_private *catalog,
123 * To make sure interface reg writes happens before any other operation,
124 * this function uses writel() instread of writel_relaxed()
126 return readl_relaxed(catalog->io->dp_controller.p0.base + offset);
129 static inline u32 dp_read_link(struct dp_catalog_private *catalog, u32 offset)
131 return readl_relaxed(catalog->io->dp_controller.link.base + offset);
134 static inline void dp_write_link(struct dp_catalog_private *catalog,
135 u32 offset, u32 data)
138 * To make sure link reg writes happens before any other operation,
139 * this function uses writel() instread of writel_relaxed()
141 writel(data, catalog->io->dp_controller.link.base + offset);
144 /* aux related catalog functions */
145 u32 dp_catalog_aux_read_data(struct dp_catalog *dp_catalog)
147 struct dp_catalog_private *catalog = container_of(dp_catalog,
148 struct dp_catalog_private, dp_catalog);
150 return dp_read_aux(catalog, REG_DP_AUX_DATA);
153 int dp_catalog_aux_write_data(struct dp_catalog *dp_catalog)
155 struct dp_catalog_private *catalog = container_of(dp_catalog,
156 struct dp_catalog_private, dp_catalog);
158 dp_write_aux(catalog, REG_DP_AUX_DATA, dp_catalog->aux_data);
162 int dp_catalog_aux_write_trans(struct dp_catalog *dp_catalog)
164 struct dp_catalog_private *catalog = container_of(dp_catalog,
165 struct dp_catalog_private, dp_catalog);
167 dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, dp_catalog->aux_data);
171 int dp_catalog_aux_clear_trans(struct dp_catalog *dp_catalog, bool read)
174 struct dp_catalog_private *catalog = container_of(dp_catalog,
175 struct dp_catalog_private, dp_catalog);
178 data = dp_read_aux(catalog, REG_DP_AUX_TRANS_CTRL);
179 data &= ~DP_AUX_TRANS_CTRL_GO;
180 dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, data);
182 dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, 0);
187 int dp_catalog_aux_clear_hw_interrupts(struct dp_catalog *dp_catalog)
189 struct dp_catalog_private *catalog = container_of(dp_catalog,
190 struct dp_catalog_private, dp_catalog);
192 dp_read_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_STATUS);
193 dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0x1f);
194 dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0x9f);
195 dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0);
200 * dp_catalog_aux_reset() - reset AUX controller
202 * @dp_catalog: DP catalog structure
206 * This function reset AUX controller
208 * NOTE: reset AUX controller will also clear any pending HPD related interrupts
211 void dp_catalog_aux_reset(struct dp_catalog *dp_catalog)
214 struct dp_catalog_private *catalog = container_of(dp_catalog,
215 struct dp_catalog_private, dp_catalog);
217 aux_ctrl = dp_read_aux(catalog, REG_DP_AUX_CTRL);
219 aux_ctrl |= DP_AUX_CTRL_RESET;
220 dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl);
221 usleep_range(1000, 1100); /* h/w recommended delay */
223 aux_ctrl &= ~DP_AUX_CTRL_RESET;
224 dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl);
227 void dp_catalog_aux_enable(struct dp_catalog *dp_catalog, bool enable)
230 struct dp_catalog_private *catalog = container_of(dp_catalog,
231 struct dp_catalog_private, dp_catalog);
233 aux_ctrl = dp_read_aux(catalog, REG_DP_AUX_CTRL);
236 dp_write_aux(catalog, REG_DP_TIMEOUT_COUNT, 0xffff);
237 dp_write_aux(catalog, REG_DP_AUX_LIMITS, 0xffff);
238 aux_ctrl |= DP_AUX_CTRL_ENABLE;
240 aux_ctrl &= ~DP_AUX_CTRL_ENABLE;
243 dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl);
246 void dp_catalog_aux_update_cfg(struct dp_catalog *dp_catalog)
248 struct dp_catalog_private *catalog = container_of(dp_catalog,
249 struct dp_catalog_private, dp_catalog);
250 struct dp_io *dp_io = catalog->io;
251 struct phy *phy = dp_io->phy;
256 int dp_catalog_aux_wait_for_hpd_connect_state(struct dp_catalog *dp_catalog)
259 struct dp_catalog_private *catalog = container_of(dp_catalog,
260 struct dp_catalog_private, dp_catalog);
262 /* poll for hpd connected status every 2ms and timeout after 500ms */
263 return readl_poll_timeout(catalog->io->dp_controller.aux.base +
264 REG_DP_DP_HPD_INT_STATUS,
265 state, state & DP_DP_HPD_STATE_STATUS_CONNECTED,
269 static void dump_regs(void __iomem *base, int len)
275 len = DIV_ROUND_UP(len, 16);
276 for (i = 0; i < len; i++) {
277 x0 = readl_relaxed(base + addr_off);
278 x4 = readl_relaxed(base + addr_off + 0x04);
279 x8 = readl_relaxed(base + addr_off + 0x08);
280 xc = readl_relaxed(base + addr_off + 0x0c);
282 pr_info("%08x: %08x %08x %08x %08x", addr_off, x0, x4, x8, xc);
287 void dp_catalog_dump_regs(struct dp_catalog *dp_catalog)
289 struct dp_catalog_private *catalog = container_of(dp_catalog,
290 struct dp_catalog_private, dp_catalog);
291 struct dss_io_data *io = &catalog->io->dp_controller;
293 pr_info("AHB regs\n");
294 dump_regs(io->ahb.base, io->ahb.len);
296 pr_info("AUXCLK regs\n");
297 dump_regs(io->aux.base, io->aux.len);
299 pr_info("LCLK regs\n");
300 dump_regs(io->link.base, io->link.len);
302 pr_info("P0CLK regs\n");
303 dump_regs(io->p0.base, io->p0.len);
306 u32 dp_catalog_aux_get_irq(struct dp_catalog *dp_catalog)
308 struct dp_catalog_private *catalog = container_of(dp_catalog,
309 struct dp_catalog_private, dp_catalog);
312 intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS);
313 intr &= ~DP_INTERRUPT_STATUS1_MASK;
314 intr_ack = (intr & DP_INTERRUPT_STATUS1)
315 << DP_INTERRUPT_STATUS_ACK_SHIFT;
316 dp_write_ahb(catalog, REG_DP_INTR_STATUS, intr_ack |
317 DP_INTERRUPT_STATUS1_MASK);
323 /* controller related catalog functions */
324 void dp_catalog_ctrl_update_transfer_unit(struct dp_catalog *dp_catalog,
325 u32 dp_tu, u32 valid_boundary,
328 struct dp_catalog_private *catalog = container_of(dp_catalog,
329 struct dp_catalog_private, dp_catalog);
331 dp_write_link(catalog, REG_DP_VALID_BOUNDARY, valid_boundary);
332 dp_write_link(catalog, REG_DP_TU, dp_tu);
333 dp_write_link(catalog, REG_DP_VALID_BOUNDARY_2, valid_boundary2);
336 void dp_catalog_ctrl_state_ctrl(struct dp_catalog *dp_catalog, u32 state)
338 struct dp_catalog_private *catalog = container_of(dp_catalog,
339 struct dp_catalog_private, dp_catalog);
341 dp_write_link(catalog, REG_DP_STATE_CTRL, state);
344 void dp_catalog_ctrl_config_ctrl(struct dp_catalog *dp_catalog, u32 cfg)
346 struct dp_catalog_private *catalog = container_of(dp_catalog,
347 struct dp_catalog_private, dp_catalog);
349 drm_dbg_dp(catalog->drm_dev, "DP_CONFIGURATION_CTRL=0x%x\n", cfg);
351 dp_write_link(catalog, REG_DP_CONFIGURATION_CTRL, cfg);
354 void dp_catalog_ctrl_lane_mapping(struct dp_catalog *dp_catalog)
356 struct dp_catalog_private *catalog = container_of(dp_catalog,
357 struct dp_catalog_private, dp_catalog);
358 u32 ln_0 = 0, ln_1 = 1, ln_2 = 2, ln_3 = 3; /* One-to-One mapping */
361 ln_mapping = ln_0 << LANE0_MAPPING_SHIFT;
362 ln_mapping |= ln_1 << LANE1_MAPPING_SHIFT;
363 ln_mapping |= ln_2 << LANE2_MAPPING_SHIFT;
364 ln_mapping |= ln_3 << LANE3_MAPPING_SHIFT;
366 dp_write_link(catalog, REG_DP_LOGICAL2PHYSICAL_LANE_MAPPING,
370 void dp_catalog_ctrl_psr_mainlink_enable(struct dp_catalog *dp_catalog,
374 struct dp_catalog_private *catalog = container_of(dp_catalog,
375 struct dp_catalog_private, dp_catalog);
377 val = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
380 val |= DP_MAINLINK_CTRL_ENABLE;
382 val &= ~DP_MAINLINK_CTRL_ENABLE;
384 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, val);
387 void dp_catalog_ctrl_mainlink_ctrl(struct dp_catalog *dp_catalog,
391 struct dp_catalog_private *catalog = container_of(dp_catalog,
392 struct dp_catalog_private, dp_catalog);
394 drm_dbg_dp(catalog->drm_dev, "enable=%d\n", enable);
397 * To make sure link reg writes happens before other operation,
398 * dp_write_link() function uses writel()
400 mainlink_ctrl = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
402 mainlink_ctrl &= ~(DP_MAINLINK_CTRL_RESET |
403 DP_MAINLINK_CTRL_ENABLE);
404 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
406 mainlink_ctrl |= DP_MAINLINK_CTRL_RESET;
407 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
409 mainlink_ctrl &= ~DP_MAINLINK_CTRL_RESET;
410 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
412 mainlink_ctrl |= (DP_MAINLINK_CTRL_ENABLE |
413 DP_MAINLINK_FB_BOUNDARY_SEL);
414 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
416 mainlink_ctrl = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
417 mainlink_ctrl &= ~DP_MAINLINK_CTRL_ENABLE;
418 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl);
422 void dp_catalog_ctrl_config_misc(struct dp_catalog *dp_catalog,
427 struct dp_catalog_private *catalog = container_of(dp_catalog,
428 struct dp_catalog_private, dp_catalog);
430 misc_val = dp_read_link(catalog, REG_DP_MISC1_MISC0);
433 misc_val &= ~(0x07 << DP_MISC0_TEST_BITS_DEPTH_SHIFT);
434 misc_val |= colorimetry_cfg << DP_MISC0_COLORIMETRY_CFG_SHIFT;
435 misc_val |= test_bits_depth << DP_MISC0_TEST_BITS_DEPTH_SHIFT;
436 /* Configure clock to synchronous mode */
437 misc_val |= DP_MISC0_SYNCHRONOUS_CLK;
439 drm_dbg_dp(catalog->drm_dev, "misc settings = 0x%x\n", misc_val);
440 dp_write_link(catalog, REG_DP_MISC1_MISC0, misc_val);
443 void dp_catalog_ctrl_config_msa(struct dp_catalog *dp_catalog,
444 u32 rate, u32 stream_rate_khz,
447 u32 pixel_m, pixel_n;
448 u32 mvid, nvid, pixel_div = 0, dispcc_input_rate;
449 u32 const nvid_fixed = DP_LINK_CONSTANT_N_VALUE;
450 u32 const link_rate_hbr2 = 540000;
451 u32 const link_rate_hbr3 = 810000;
452 unsigned long den, num;
454 struct dp_catalog_private *catalog = container_of(dp_catalog,
455 struct dp_catalog_private, dp_catalog);
457 if (rate == link_rate_hbr3)
459 else if (rate == 162000 || rate == 270000)
461 else if (rate == link_rate_hbr2)
464 DRM_ERROR("Invalid pixel mux divider\n");
466 dispcc_input_rate = (rate * 10) / pixel_div;
468 rational_best_approximation(dispcc_input_rate, stream_rate_khz,
469 (unsigned long)(1 << 16) - 1,
470 (unsigned long)(1 << 16) - 1, &den, &num);
477 mvid = (pixel_m & 0xFFFF) * 5;
478 nvid = (0xFFFF & (~pixel_n)) + (pixel_m & 0xFFFF);
480 if (nvid < nvid_fixed) {
483 temp = (nvid_fixed / nvid) * nvid;
484 mvid = (nvid_fixed / nvid) * mvid;
488 if (link_rate_hbr2 == rate)
491 if (link_rate_hbr3 == rate)
494 drm_dbg_dp(catalog->drm_dev, "mvid=0x%x, nvid=0x%x\n", mvid, nvid);
495 dp_write_link(catalog, REG_DP_SOFTWARE_MVID, mvid);
496 dp_write_link(catalog, REG_DP_SOFTWARE_NVID, nvid);
497 dp_write_p0(catalog, MMSS_DP_DSC_DTO, 0x0);
500 int dp_catalog_ctrl_set_pattern_state_bit(struct dp_catalog *dp_catalog,
505 struct dp_catalog_private *catalog = container_of(dp_catalog,
506 struct dp_catalog_private, dp_catalog);
508 bit = BIT(state_bit - 1);
509 drm_dbg_dp(catalog->drm_dev, "hw: bit=%d train=%d\n", bit, state_bit);
510 dp_catalog_ctrl_state_ctrl(dp_catalog, bit);
512 bit = BIT(state_bit - 1) << DP_MAINLINK_READY_LINK_TRAINING_SHIFT;
514 /* Poll for mainlink ready status */
515 ret = readx_poll_timeout(readl, catalog->io->dp_controller.link.base +
516 REG_DP_MAINLINK_READY,
518 POLLING_SLEEP_US, POLLING_TIMEOUT_US);
520 DRM_ERROR("set state_bit for link_train=%d failed\n", state_bit);
527 * dp_catalog_hw_revision() - retrieve DP hw revision
529 * @dp_catalog: DP catalog structure
531 * Return: DP controller hw revision
534 u32 dp_catalog_hw_revision(const struct dp_catalog *dp_catalog)
536 const struct dp_catalog_private *catalog = container_of(dp_catalog,
537 struct dp_catalog_private, dp_catalog);
539 return dp_read_ahb(catalog, REG_DP_HW_VERSION);
543 * dp_catalog_ctrl_reset() - reset DP controller
545 * @dp_catalog: DP catalog structure
549 * This function reset the DP controller
551 * NOTE: reset DP controller will also clear any pending HPD related interrupts
554 void dp_catalog_ctrl_reset(struct dp_catalog *dp_catalog)
557 struct dp_catalog_private *catalog = container_of(dp_catalog,
558 struct dp_catalog_private, dp_catalog);
560 sw_reset = dp_read_ahb(catalog, REG_DP_SW_RESET);
562 sw_reset |= DP_SW_RESET;
563 dp_write_ahb(catalog, REG_DP_SW_RESET, sw_reset);
564 usleep_range(1000, 1100); /* h/w recommended delay */
566 sw_reset &= ~DP_SW_RESET;
567 dp_write_ahb(catalog, REG_DP_SW_RESET, sw_reset);
570 bool dp_catalog_ctrl_mainlink_ready(struct dp_catalog *dp_catalog)
574 struct dp_catalog_private *catalog = container_of(dp_catalog,
575 struct dp_catalog_private, dp_catalog);
577 /* Poll for mainlink ready status */
578 ret = readl_poll_timeout(catalog->io->dp_controller.link.base +
579 REG_DP_MAINLINK_READY,
580 data, data & DP_MAINLINK_READY_FOR_VIDEO,
581 POLLING_SLEEP_US, POLLING_TIMEOUT_US);
583 DRM_ERROR("mainlink not ready\n");
590 void dp_catalog_ctrl_enable_irq(struct dp_catalog *dp_catalog,
593 struct dp_catalog_private *catalog = container_of(dp_catalog,
594 struct dp_catalog_private, dp_catalog);
597 dp_write_ahb(catalog, REG_DP_INTR_STATUS,
598 DP_INTERRUPT_STATUS1_MASK);
599 dp_write_ahb(catalog, REG_DP_INTR_STATUS2,
600 DP_INTERRUPT_STATUS2_MASK);
602 dp_write_ahb(catalog, REG_DP_INTR_STATUS, 0x00);
603 dp_write_ahb(catalog, REG_DP_INTR_STATUS2, 0x00);
607 void dp_catalog_hpd_config_intr(struct dp_catalog *dp_catalog,
608 u32 intr_mask, bool en)
610 struct dp_catalog_private *catalog = container_of(dp_catalog,
611 struct dp_catalog_private, dp_catalog);
613 u32 config = dp_read_aux(catalog, REG_DP_DP_HPD_INT_MASK);
615 config = (en ? config | intr_mask : config & ~intr_mask);
617 drm_dbg_dp(catalog->drm_dev, "intr_mask=%#x config=%#x\n",
619 dp_write_aux(catalog, REG_DP_DP_HPD_INT_MASK,
620 config & DP_DP_HPD_INT_MASK);
623 void dp_catalog_ctrl_hpd_config(struct dp_catalog *dp_catalog)
625 struct dp_catalog_private *catalog = container_of(dp_catalog,
626 struct dp_catalog_private, dp_catalog);
628 u32 reftimer = dp_read_aux(catalog, REG_DP_DP_HPD_REFTIMER);
630 /* Configure REFTIMER and enable it */
631 reftimer |= DP_DP_HPD_REFTIMER_ENABLE;
632 dp_write_aux(catalog, REG_DP_DP_HPD_REFTIMER, reftimer);
635 dp_write_aux(catalog, REG_DP_DP_HPD_CTRL, DP_DP_HPD_CTRL_HPD_EN);
638 static void dp_catalog_enable_sdp(struct dp_catalog_private *catalog)
641 dp_write_link(catalog, MMSS_DP_SDP_CFG3, UPDATE_SDP);
642 dp_write_link(catalog, MMSS_DP_SDP_CFG3, 0x0);
645 void dp_catalog_ctrl_config_psr(struct dp_catalog *dp_catalog)
647 struct dp_catalog_private *catalog = container_of(dp_catalog,
648 struct dp_catalog_private, dp_catalog);
651 /* enable PSR1 function */
652 config = dp_read_link(catalog, REG_PSR_CONFIG);
653 config |= PSR1_SUPPORTED;
654 dp_write_link(catalog, REG_PSR_CONFIG, config);
656 dp_write_ahb(catalog, REG_DP_INTR_MASK4, DP_INTERRUPT_MASK4);
657 dp_catalog_enable_sdp(catalog);
660 void dp_catalog_ctrl_set_psr(struct dp_catalog *dp_catalog, bool enter)
662 struct dp_catalog_private *catalog = container_of(dp_catalog,
663 struct dp_catalog_private, dp_catalog);
666 cmd = dp_read_link(catalog, REG_PSR_CMD);
668 cmd &= ~(PSR_ENTER | PSR_EXIT);
675 dp_catalog_enable_sdp(catalog);
676 dp_write_link(catalog, REG_PSR_CMD, cmd);
679 u32 dp_catalog_link_is_connected(struct dp_catalog *dp_catalog)
681 struct dp_catalog_private *catalog = container_of(dp_catalog,
682 struct dp_catalog_private, dp_catalog);
685 status = dp_read_aux(catalog, REG_DP_DP_HPD_INT_STATUS);
686 drm_dbg_dp(catalog->drm_dev, "aux status: %#x\n", status);
687 status >>= DP_DP_HPD_STATE_STATUS_BITS_SHIFT;
688 status &= DP_DP_HPD_STATE_STATUS_BITS_MASK;
693 u32 dp_catalog_hpd_get_intr_status(struct dp_catalog *dp_catalog)
695 struct dp_catalog_private *catalog = container_of(dp_catalog,
696 struct dp_catalog_private, dp_catalog);
699 isr = dp_read_aux(catalog, REG_DP_DP_HPD_INT_STATUS);
700 dp_write_aux(catalog, REG_DP_DP_HPD_INT_ACK,
701 (isr & DP_DP_HPD_INT_MASK));
702 mask = dp_read_aux(catalog, REG_DP_DP_HPD_INT_MASK);
705 * We only want to return interrupts that are unmasked to the caller.
706 * However, the interrupt status field also contains other
707 * informational bits about the HPD state status, so we only mask
708 * out the part of the register that tells us about which interrupts
711 return isr & (mask | ~DP_DP_HPD_INT_MASK);
714 u32 dp_catalog_ctrl_read_psr_interrupt_status(struct dp_catalog *dp_catalog)
716 struct dp_catalog_private *catalog = container_of(dp_catalog,
717 struct dp_catalog_private, dp_catalog);
720 intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS4);
721 intr_ack = (intr & DP_INTERRUPT_STATUS4)
722 << DP_INTERRUPT_STATUS_ACK_SHIFT;
723 dp_write_ahb(catalog, REG_DP_INTR_STATUS4, intr_ack);
728 int dp_catalog_ctrl_get_interrupt(struct dp_catalog *dp_catalog)
730 struct dp_catalog_private *catalog = container_of(dp_catalog,
731 struct dp_catalog_private, dp_catalog);
734 intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS2);
735 intr &= ~DP_INTERRUPT_STATUS2_MASK;
736 intr_ack = (intr & DP_INTERRUPT_STATUS2)
737 << DP_INTERRUPT_STATUS_ACK_SHIFT;
738 dp_write_ahb(catalog, REG_DP_INTR_STATUS2,
739 intr_ack | DP_INTERRUPT_STATUS2_MASK);
744 void dp_catalog_ctrl_phy_reset(struct dp_catalog *dp_catalog)
746 struct dp_catalog_private *catalog = container_of(dp_catalog,
747 struct dp_catalog_private, dp_catalog);
749 dp_write_ahb(catalog, REG_DP_PHY_CTRL,
750 DP_PHY_CTRL_SW_RESET | DP_PHY_CTRL_SW_RESET_PLL);
751 usleep_range(1000, 1100); /* h/w recommended delay */
752 dp_write_ahb(catalog, REG_DP_PHY_CTRL, 0x0);
755 int dp_catalog_ctrl_update_vx_px(struct dp_catalog *dp_catalog,
756 u8 v_level, u8 p_level)
758 struct dp_catalog_private *catalog = container_of(dp_catalog,
759 struct dp_catalog_private, dp_catalog);
760 struct dp_io *dp_io = catalog->io;
761 struct phy *phy = dp_io->phy;
762 struct phy_configure_opts_dp *opts_dp = &dp_io->phy_opts.dp;
764 /* TODO: Update for all lanes instead of just first one */
765 opts_dp->voltage[0] = v_level;
766 opts_dp->pre[0] = p_level;
767 opts_dp->set_voltages = 1;
768 phy_configure(phy, &dp_io->phy_opts);
769 opts_dp->set_voltages = 0;
774 void dp_catalog_ctrl_send_phy_pattern(struct dp_catalog *dp_catalog,
777 struct dp_catalog_private *catalog = container_of(dp_catalog,
778 struct dp_catalog_private, dp_catalog);
781 /* Make sure to clear the current pattern before starting a new one */
782 dp_write_link(catalog, REG_DP_STATE_CTRL, 0x0);
784 drm_dbg_dp(catalog->drm_dev, "pattern: %#x\n", pattern);
786 case DP_PHY_TEST_PATTERN_D10_2:
787 dp_write_link(catalog, REG_DP_STATE_CTRL,
788 DP_STATE_CTRL_LINK_TRAINING_PATTERN1);
790 case DP_PHY_TEST_PATTERN_ERROR_COUNT:
792 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
794 value |= SCRAMBLER_RESET_COUNT_VALUE;
795 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
797 dp_write_link(catalog, REG_DP_MAINLINK_LEVELS,
798 DP_MAINLINK_SAFE_TO_EXIT_LEVEL_2);
799 dp_write_link(catalog, REG_DP_STATE_CTRL,
800 DP_STATE_CTRL_LINK_SYMBOL_ERR_MEASURE);
802 case DP_PHY_TEST_PATTERN_PRBS7:
803 dp_write_link(catalog, REG_DP_STATE_CTRL,
804 DP_STATE_CTRL_LINK_PRBS7);
806 case DP_PHY_TEST_PATTERN_80BIT_CUSTOM:
807 dp_write_link(catalog, REG_DP_STATE_CTRL,
808 DP_STATE_CTRL_LINK_TEST_CUSTOM_PATTERN);
809 /* 00111110000011111000001111100000 */
810 dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG0,
812 /* 00001111100000111110000011111000 */
813 dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG1,
815 /* 1111100000111110 */
816 dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG2,
819 case DP_PHY_TEST_PATTERN_CP2520:
820 value = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
821 value &= ~DP_MAINLINK_CTRL_SW_BYPASS_SCRAMBLER;
822 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, value);
824 value = DP_HBR2_ERM_PATTERN;
825 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
827 value |= SCRAMBLER_RESET_COUNT_VALUE;
828 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET,
830 dp_write_link(catalog, REG_DP_MAINLINK_LEVELS,
831 DP_MAINLINK_SAFE_TO_EXIT_LEVEL_2);
832 dp_write_link(catalog, REG_DP_STATE_CTRL,
833 DP_STATE_CTRL_LINK_SYMBOL_ERR_MEASURE);
834 value = dp_read_link(catalog, REG_DP_MAINLINK_CTRL);
835 value |= DP_MAINLINK_CTRL_ENABLE;
836 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, value);
838 case DP_PHY_TEST_PATTERN_SEL_MASK:
839 dp_write_link(catalog, REG_DP_MAINLINK_CTRL,
840 DP_MAINLINK_CTRL_ENABLE);
841 dp_write_link(catalog, REG_DP_STATE_CTRL,
842 DP_STATE_CTRL_LINK_TRAINING_PATTERN4);
845 drm_dbg_dp(catalog->drm_dev,
846 "No valid test pattern requested: %#x\n", pattern);
851 u32 dp_catalog_ctrl_read_phy_pattern(struct dp_catalog *dp_catalog)
853 struct dp_catalog_private *catalog = container_of(dp_catalog,
854 struct dp_catalog_private, dp_catalog);
856 return dp_read_link(catalog, REG_DP_MAINLINK_READY);
859 /* panel related catalog functions */
860 int dp_catalog_panel_timing_cfg(struct dp_catalog *dp_catalog)
862 struct dp_catalog_private *catalog = container_of(dp_catalog,
863 struct dp_catalog_private, dp_catalog);
866 dp_write_link(catalog, REG_DP_TOTAL_HOR_VER,
868 dp_write_link(catalog, REG_DP_START_HOR_VER_FROM_SYNC,
869 dp_catalog->sync_start);
870 dp_write_link(catalog, REG_DP_HSYNC_VSYNC_WIDTH_POLARITY,
871 dp_catalog->width_blanking);
872 dp_write_link(catalog, REG_DP_ACTIVE_HOR_VER, dp_catalog->dp_active);
874 reg = dp_read_p0(catalog, MMSS_DP_INTF_CONFIG);
876 if (dp_catalog->wide_bus_en)
877 reg |= DP_INTF_CONFIG_DATABUS_WIDEN;
879 reg &= ~DP_INTF_CONFIG_DATABUS_WIDEN;
882 DRM_DEBUG_DP("wide_bus_en=%d reg=%#x\n", dp_catalog->wide_bus_en, reg);
884 dp_write_p0(catalog, MMSS_DP_INTF_CONFIG, reg);
888 void dp_catalog_panel_tpg_enable(struct dp_catalog *dp_catalog,
889 struct drm_display_mode *drm_mode)
891 struct dp_catalog_private *catalog = container_of(dp_catalog,
892 struct dp_catalog_private, dp_catalog);
893 u32 hsync_period, vsync_period;
894 u32 display_v_start, display_v_end;
895 u32 hsync_start_x, hsync_end_x;
900 /* TPG config parameters*/
901 hsync_period = drm_mode->htotal;
902 vsync_period = drm_mode->vtotal;
904 display_v_start = ((drm_mode->vtotal - drm_mode->vsync_start) *
906 display_v_end = ((vsync_period - (drm_mode->vsync_start -
910 display_v_start += drm_mode->htotal - drm_mode->hsync_start;
911 display_v_end -= (drm_mode->hsync_start - drm_mode->hdisplay);
913 hsync_start_x = drm_mode->htotal - drm_mode->hsync_start;
914 hsync_end_x = hsync_period - (drm_mode->hsync_start -
915 drm_mode->hdisplay) - 1;
917 v_sync_width = drm_mode->vsync_end - drm_mode->vsync_start;
919 hsync_ctl = (hsync_period << 16) |
920 (drm_mode->hsync_end - drm_mode->hsync_start);
921 display_hctl = (hsync_end_x << 16) | hsync_start_x;
924 dp_write_p0(catalog, MMSS_DP_INTF_CONFIG, 0x0);
925 dp_write_p0(catalog, MMSS_DP_INTF_HSYNC_CTL, hsync_ctl);
926 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PERIOD_F0, vsync_period *
928 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PULSE_WIDTH_F0, v_sync_width *
930 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PERIOD_F1, 0);
931 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PULSE_WIDTH_F1, 0);
932 dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_HCTL, display_hctl);
933 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_HCTL, 0);
934 dp_write_p0(catalog, MMSS_INTF_DISPLAY_V_START_F0, display_v_start);
935 dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_V_END_F0, display_v_end);
936 dp_write_p0(catalog, MMSS_INTF_DISPLAY_V_START_F1, 0);
937 dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_V_END_F1, 0);
938 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_START_F0, 0);
939 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_END_F0, 0);
940 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_START_F1, 0);
941 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_END_F1, 0);
942 dp_write_p0(catalog, MMSS_DP_INTF_POLARITY_CTL, 0);
944 dp_write_p0(catalog, MMSS_DP_TPG_MAIN_CONTROL,
945 DP_TPG_CHECKERED_RECT_PATTERN);
946 dp_write_p0(catalog, MMSS_DP_TPG_VIDEO_CONFIG,
947 DP_TPG_VIDEO_CONFIG_BPP_8BIT |
948 DP_TPG_VIDEO_CONFIG_RGB);
949 dp_write_p0(catalog, MMSS_DP_BIST_ENABLE,
950 DP_BIST_ENABLE_DPBIST_EN);
951 dp_write_p0(catalog, MMSS_DP_TIMING_ENGINE_EN,
952 DP_TIMING_ENGINE_EN_EN);
953 drm_dbg_dp(catalog->drm_dev, "%s: enabled tpg\n", __func__);
956 void dp_catalog_panel_tpg_disable(struct dp_catalog *dp_catalog)
958 struct dp_catalog_private *catalog = container_of(dp_catalog,
959 struct dp_catalog_private, dp_catalog);
961 dp_write_p0(catalog, MMSS_DP_TPG_MAIN_CONTROL, 0x0);
962 dp_write_p0(catalog, MMSS_DP_BIST_ENABLE, 0x0);
963 dp_write_p0(catalog, MMSS_DP_TIMING_ENGINE_EN, 0x0);
966 struct dp_catalog *dp_catalog_get(struct device *dev, struct dp_io *io)
968 struct dp_catalog_private *catalog;
971 DRM_ERROR("invalid input\n");
972 return ERR_PTR(-EINVAL);
975 catalog = devm_kzalloc(dev, sizeof(*catalog), GFP_KERNEL);
977 return ERR_PTR(-ENOMEM);
982 return &catalog->dp_catalog;
985 void dp_catalog_audio_get_header(struct dp_catalog *dp_catalog)
987 struct dp_catalog_private *catalog;
988 u32 (*sdp_map)[DP_AUDIO_SDP_HEADER_MAX];
989 enum dp_catalog_audio_sdp_type sdp;
990 enum dp_catalog_audio_header_type header;
995 catalog = container_of(dp_catalog,
996 struct dp_catalog_private, dp_catalog);
998 sdp_map = catalog->audio_map;
999 sdp = dp_catalog->sdp_type;
1000 header = dp_catalog->sdp_header;
1002 dp_catalog->audio_data = dp_read_link(catalog,
1003 sdp_map[sdp][header]);
1006 void dp_catalog_audio_set_header(struct dp_catalog *dp_catalog)
1008 struct dp_catalog_private *catalog;
1009 u32 (*sdp_map)[DP_AUDIO_SDP_HEADER_MAX];
1010 enum dp_catalog_audio_sdp_type sdp;
1011 enum dp_catalog_audio_header_type header;
1017 catalog = container_of(dp_catalog,
1018 struct dp_catalog_private, dp_catalog);
1020 sdp_map = catalog->audio_map;
1021 sdp = dp_catalog->sdp_type;
1022 header = dp_catalog->sdp_header;
1023 data = dp_catalog->audio_data;
1025 dp_write_link(catalog, sdp_map[sdp][header], data);
1028 void dp_catalog_audio_config_acr(struct dp_catalog *dp_catalog)
1030 struct dp_catalog_private *catalog;
1031 u32 acr_ctrl, select;
1036 catalog = container_of(dp_catalog,
1037 struct dp_catalog_private, dp_catalog);
1039 select = dp_catalog->audio_data;
1040 acr_ctrl = select << 4 | BIT(31) | BIT(8) | BIT(14);
1042 drm_dbg_dp(catalog->drm_dev, "select: %#x, acr_ctrl: %#x\n",
1045 dp_write_link(catalog, MMSS_DP_AUDIO_ACR_CTRL, acr_ctrl);
1048 void dp_catalog_audio_enable(struct dp_catalog *dp_catalog)
1050 struct dp_catalog_private *catalog;
1057 catalog = container_of(dp_catalog,
1058 struct dp_catalog_private, dp_catalog);
1060 enable = !!dp_catalog->audio_data;
1061 audio_ctrl = dp_read_link(catalog, MMSS_DP_AUDIO_CFG);
1064 audio_ctrl |= BIT(0);
1066 audio_ctrl &= ~BIT(0);
1068 drm_dbg_dp(catalog->drm_dev, "dp_audio_cfg = 0x%x\n", audio_ctrl);
1070 dp_write_link(catalog, MMSS_DP_AUDIO_CFG, audio_ctrl);
1071 /* make sure audio engine is disabled */
1075 void dp_catalog_audio_config_sdp(struct dp_catalog *dp_catalog)
1077 struct dp_catalog_private *catalog;
1084 catalog = container_of(dp_catalog,
1085 struct dp_catalog_private, dp_catalog);
1087 sdp_cfg = dp_read_link(catalog, MMSS_DP_SDP_CFG);
1088 /* AUDIO_TIMESTAMP_SDP_EN */
1090 /* AUDIO_STREAM_SDP_EN */
1092 /* AUDIO_COPY_MANAGEMENT_SDP_EN */
1094 /* AUDIO_ISRC_SDP_EN */
1096 /* AUDIO_INFOFRAME_SDP_EN */
1099 drm_dbg_dp(catalog->drm_dev, "sdp_cfg = 0x%x\n", sdp_cfg);
1101 dp_write_link(catalog, MMSS_DP_SDP_CFG, sdp_cfg);
1103 sdp_cfg2 = dp_read_link(catalog, MMSS_DP_SDP_CFG2);
1104 /* IFRM_REGSRC -> Do not use reg values */
1105 sdp_cfg2 &= ~BIT(0);
1106 /* AUDIO_STREAM_HB3_REGSRC-> Do not use reg values */
1107 sdp_cfg2 &= ~BIT(1);
1109 drm_dbg_dp(catalog->drm_dev, "sdp_cfg2 = 0x%x\n", sdp_cfg2);
1111 dp_write_link(catalog, MMSS_DP_SDP_CFG2, sdp_cfg2);
1114 void dp_catalog_audio_init(struct dp_catalog *dp_catalog)
1116 struct dp_catalog_private *catalog;
1118 static u32 sdp_map[][DP_AUDIO_SDP_HEADER_MAX] = {
1120 MMSS_DP_AUDIO_STREAM_0,
1121 MMSS_DP_AUDIO_STREAM_1,
1122 MMSS_DP_AUDIO_STREAM_1,
1125 MMSS_DP_AUDIO_TIMESTAMP_0,
1126 MMSS_DP_AUDIO_TIMESTAMP_1,
1127 MMSS_DP_AUDIO_TIMESTAMP_1,
1130 MMSS_DP_AUDIO_INFOFRAME_0,
1131 MMSS_DP_AUDIO_INFOFRAME_1,
1132 MMSS_DP_AUDIO_INFOFRAME_1,
1135 MMSS_DP_AUDIO_COPYMANAGEMENT_0,
1136 MMSS_DP_AUDIO_COPYMANAGEMENT_1,
1137 MMSS_DP_AUDIO_COPYMANAGEMENT_1,
1140 MMSS_DP_AUDIO_ISRC_0,
1141 MMSS_DP_AUDIO_ISRC_1,
1142 MMSS_DP_AUDIO_ISRC_1,
1149 catalog = container_of(dp_catalog,
1150 struct dp_catalog_private, dp_catalog);
1152 catalog->audio_map = sdp_map;
1155 void dp_catalog_audio_sfe_level(struct dp_catalog *dp_catalog)
1157 struct dp_catalog_private *catalog;
1158 u32 mainlink_levels, safe_to_exit_level;
1163 catalog = container_of(dp_catalog,
1164 struct dp_catalog_private, dp_catalog);
1166 safe_to_exit_level = dp_catalog->audio_data;
1167 mainlink_levels = dp_read_link(catalog, REG_DP_MAINLINK_LEVELS);
1168 mainlink_levels &= 0xFE0;
1169 mainlink_levels |= safe_to_exit_level;
1171 drm_dbg_dp(catalog->drm_dev,
1172 "mainlink_level = 0x%x, safe_to_exit_level = 0x%x\n",
1173 mainlink_levels, safe_to_exit_level);
1175 dp_write_link(catalog, REG_DP_MAINLINK_LEVELS, mainlink_levels);