1 // SPDX-License-Identifier: GPL-2.0
3 * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
5 * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6 * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
12 * - PM for serializer and remote peripherals. We need to manage:
14 * - Power domain? Regulator? Somehow any remote device should be able to
15 * cause the VPOC to be turned on.
16 * - Link between the deserializer and the serializer
17 * - Related to VPOC management. We probably always want to turn on the VPOC
18 * and then enable the link.
19 * - Serializer's services: i2c, gpios, power
20 * - The serializer needs to resume before the remote peripherals can
22 * - How to handle gpios? Reserving a gpio essentially keeps the provider
23 * (serializer) always powered on.
24 * - Do we need a new bus for the FPD-Link? At the moment the serializers
25 * are children of the same i2c-adapter where the deserializer resides.
26 * - i2c-atr could be made embeddable instead of allocatable.
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/workqueue.h>
48 #include <media/i2c/ds90ub9xx.h>
49 #include <media/mipi-csi2.h>
50 #include <media/v4l2-ctrls.h>
51 #include <media/v4l2-event.h>
52 #include <media/v4l2-fwnode.h>
53 #include <media/v4l2-subdev.h>
55 #define MHZ(v) ((u32)((v) * 1000000U))
57 #define UB960_POLL_TIME_MS 500
59 #define UB960_MAX_RX_NPORTS 4
60 #define UB960_MAX_TX_NPORTS 2
61 #define UB960_MAX_NPORTS (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
63 #define UB960_MAX_PORT_ALIASES 8
65 #define UB960_NUM_BC_GPIOS 4
70 * 0x00-0x32 Shared (UB960_SR)
71 * 0x33-0x3a CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
72 * 0x4c Shared (UB960_SR)
73 * 0x4d-0x7f FPD-Link RX, per-port paged (UB960_RR)
74 * 0xb0-0xbf Shared (UB960_SR)
75 * 0xd0-0xdf FPD-Link RX, per-port paged (UB960_RR)
76 * 0xf0-0xf5 Shared (UB960_SR)
77 * 0xf8-0xfb Shared (UB960_SR)
81 * UB960_SR_* = Shared register
82 * UB960_RR_* = FPD-Link RX, per-port paged register
83 * UB960_TR_* = CSI-2 TX, per-port paged register
84 * UB960_XR_* = Reserved register
85 * UB960_IR_* = Indirect register
88 #define UB960_SR_I2C_DEV_ID 0x00
89 #define UB960_SR_RESET 0x01
90 #define UB960_SR_RESET_DIGITAL_RESET1 BIT(1)
91 #define UB960_SR_RESET_DIGITAL_RESET0 BIT(0)
92 #define UB960_SR_RESET_GPIO_LOCK_RELEASE BIT(5)
94 #define UB960_SR_GEN_CONFIG 0x02
95 #define UB960_SR_REV_MASK 0x03
96 #define UB960_SR_DEVICE_STS 0x04
97 #define UB960_SR_PAR_ERR_THOLD_HI 0x05
98 #define UB960_SR_PAR_ERR_THOLD_LO 0x06
99 #define UB960_SR_BCC_WDOG_CTL 0x07
100 #define UB960_SR_I2C_CTL1 0x08
101 #define UB960_SR_I2C_CTL2 0x09
102 #define UB960_SR_SCL_HIGH_TIME 0x0a
103 #define UB960_SR_SCL_LOW_TIME 0x0b
104 #define UB960_SR_RX_PORT_CTL 0x0c
105 #define UB960_SR_IO_CTL 0x0d
106 #define UB960_SR_GPIO_PIN_STS 0x0e
107 #define UB960_SR_GPIO_INPUT_CTL 0x0f
108 #define UB960_SR_GPIO_PIN_CTL(n) (0x10 + (n)) /* n < UB960_NUM_GPIOS */
109 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL 5
110 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT 2
111 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN BIT(0)
113 #define UB960_SR_FS_CTL 0x18
114 #define UB960_SR_FS_HIGH_TIME_1 0x19
115 #define UB960_SR_FS_HIGH_TIME_0 0x1a
116 #define UB960_SR_FS_LOW_TIME_1 0x1b
117 #define UB960_SR_FS_LOW_TIME_0 0x1c
118 #define UB960_SR_MAX_FRM_HI 0x1d
119 #define UB960_SR_MAX_FRM_LO 0x1e
120 #define UB960_SR_CSI_PLL_CTL 0x1f
122 #define UB960_SR_FWD_CTL1 0x20
123 #define UB960_SR_FWD_CTL1_PORT_DIS(n) BIT((n) + 4)
125 #define UB960_SR_FWD_CTL2 0x21
126 #define UB960_SR_FWD_STS 0x22
128 #define UB960_SR_INTERRUPT_CTL 0x23
129 #define UB960_SR_INTERRUPT_CTL_INT_EN BIT(7)
130 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0 BIT(4)
131 #define UB960_SR_INTERRUPT_CTL_IE_RX(n) BIT((n)) /* rxport[n] IRQ */
133 #define UB960_SR_INTERRUPT_STS 0x24
134 #define UB960_SR_INTERRUPT_STS_INT BIT(7)
135 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n) BIT(4 + (n)) /* txport[n] IRQ */
136 #define UB960_SR_INTERRUPT_STS_IS_RX(n) BIT((n)) /* rxport[n] IRQ */
138 #define UB960_SR_TS_CONFIG 0x25
139 #define UB960_SR_TS_CONTROL 0x26
140 #define UB960_SR_TS_LINE_HI 0x27
141 #define UB960_SR_TS_LINE_LO 0x28
142 #define UB960_SR_TS_STATUS 0x29
143 #define UB960_SR_TIMESTAMP_P0_HI 0x2a
144 #define UB960_SR_TIMESTAMP_P0_LO 0x2b
145 #define UB960_SR_TIMESTAMP_P1_HI 0x2c
146 #define UB960_SR_TIMESTAMP_P1_LO 0x2d
148 #define UB960_SR_CSI_PORT_SEL 0x32
150 #define UB960_TR_CSI_CTL 0x33
151 #define UB960_TR_CSI_CTL_CSI_CAL_EN BIT(6)
152 #define UB960_TR_CSI_CTL_CSI_ENABLE BIT(0)
154 #define UB960_TR_CSI_CTL2 0x34
155 #define UB960_TR_CSI_STS 0x35
156 #define UB960_TR_CSI_TX_ICR 0x36
158 #define UB960_TR_CSI_TX_ISR 0x37
159 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR BIT(3)
160 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR BIT(1)
162 #define UB960_TR_CSI_TEST_CTL 0x38
163 #define UB960_TR_CSI_TEST_PATT_HI 0x39
164 #define UB960_TR_CSI_TEST_PATT_LO 0x3a
166 #define UB960_XR_SFILTER_CFG 0x41
167 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT 4
168 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT 0
170 #define UB960_XR_AEQ_CTL1 0x42
171 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK BIT(6)
172 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING BIT(5)
173 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY BIT(4)
174 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK \
175 (UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK | \
176 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
177 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
178 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN BIT(0)
180 #define UB960_XR_AEQ_ERR_THOLD 0x43
182 #define UB960_RR_BCC_ERR_CTL 0x46
183 #define UB960_RR_BCC_STATUS 0x47
184 #define UB960_RR_BCC_STATUS_SEQ_ERROR BIT(5)
185 #define UB960_RR_BCC_STATUS_MASTER_ERR BIT(4)
186 #define UB960_RR_BCC_STATUS_MASTER_TO BIT(3)
187 #define UB960_RR_BCC_STATUS_SLAVE_ERR BIT(2)
188 #define UB960_RR_BCC_STATUS_SLAVE_TO BIT(1)
189 #define UB960_RR_BCC_STATUS_RESP_ERR BIT(0)
190 #define UB960_RR_BCC_STATUS_ERROR_MASK \
191 (UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
192 UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR | \
193 UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
195 #define UB960_RR_FPD3_CAP 0x4a
196 #define UB960_RR_RAW_EMBED_DTYPE 0x4b
197 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT 6
199 #define UB960_SR_FPD3_PORT_SEL 0x4c
201 #define UB960_RR_RX_PORT_STS1 0x4d
202 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR BIT(5)
203 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG BIT(4)
204 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR BIT(3)
205 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR BIT(2)
206 #define UB960_RR_RX_PORT_STS1_PORT_PASS BIT(1)
207 #define UB960_RR_RX_PORT_STS1_LOCK_STS BIT(0)
208 #define UB960_RR_RX_PORT_STS1_ERROR_MASK \
209 (UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
210 UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
211 UB960_RR_RX_PORT_STS1_PARITY_ERROR)
213 #define UB960_RR_RX_PORT_STS2 0x4e
214 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
215 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG BIT(6)
216 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
217 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR BIT(4)
218 #define UB960_RR_RX_PORT_STS2_CSI_ERROR BIT(3)
219 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE BIT(2)
220 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT BIT(1)
221 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG BIT(0)
222 #define UB960_RR_RX_PORT_STS2_ERROR_MASK \
223 UB960_RR_RX_PORT_STS2_BUFFER_ERROR
225 #define UB960_RR_RX_FREQ_HIGH 0x4f
226 #define UB960_RR_RX_FREQ_LOW 0x50
227 #define UB960_RR_SENSOR_STS_0 0x51
228 #define UB960_RR_SENSOR_STS_1 0x52
229 #define UB960_RR_SENSOR_STS_2 0x53
230 #define UB960_RR_SENSOR_STS_3 0x54
231 #define UB960_RR_RX_PAR_ERR_HI 0x55
232 #define UB960_RR_RX_PAR_ERR_LO 0x56
233 #define UB960_RR_BIST_ERR_COUNT 0x57
235 #define UB960_RR_BCC_CONFIG 0x58
236 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH BIT(6)
237 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK GENMASK(2, 0)
239 #define UB960_RR_DATAPATH_CTL1 0x59
240 #define UB960_RR_DATAPATH_CTL2 0x5a
241 #define UB960_RR_SER_ID 0x5b
242 #define UB960_RR_SER_ALIAS_ID 0x5c
244 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
245 #define UB960_RR_SLAVE_ID(n) (0x5d + (n))
246 #define UB960_RR_SLAVE_ALIAS(n) (0x65 + (n))
248 #define UB960_RR_PORT_CONFIG 0x6d
249 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK GENMASK(1, 0)
251 #define UB960_RR_BC_GPIO_CTL(n) (0x6e + (n)) /* n < 2 */
252 #define UB960_RR_RAW10_ID 0x70
253 #define UB960_RR_RAW10_ID_VC_SHIFT 6
254 #define UB960_RR_RAW10_ID_DT_SHIFT 0
256 #define UB960_RR_RAW12_ID 0x71
257 #define UB960_RR_CSI_VC_MAP 0x72
258 #define UB960_RR_CSI_VC_MAP_SHIFT(x) ((x) * 2)
260 #define UB960_RR_LINE_COUNT_HI 0x73
261 #define UB960_RR_LINE_COUNT_LO 0x74
262 #define UB960_RR_LINE_LEN_1 0x75
263 #define UB960_RR_LINE_LEN_0 0x76
264 #define UB960_RR_FREQ_DET_CTL 0x77
265 #define UB960_RR_MAILBOX_1 0x78
266 #define UB960_RR_MAILBOX_2 0x79
268 #define UB960_RR_CSI_RX_STS 0x7a
269 #define UB960_RR_CSI_RX_STS_LENGTH_ERR BIT(3)
270 #define UB960_RR_CSI_RX_STS_CKSUM_ERR BIT(2)
271 #define UB960_RR_CSI_RX_STS_ECC2_ERR BIT(1)
272 #define UB960_RR_CSI_RX_STS_ECC1_ERR BIT(0)
273 #define UB960_RR_CSI_RX_STS_ERROR_MASK \
274 (UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
275 UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
277 #define UB960_RR_CSI_ERR_COUNTER 0x7b
278 #define UB960_RR_PORT_CONFIG2 0x7c
279 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
280 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
282 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW BIT(1)
283 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW BIT(0)
285 #define UB960_RR_PORT_PASS_CTL 0x7d
286 #define UB960_RR_SEN_INT_RISE_CTL 0x7e
287 #define UB960_RR_SEN_INT_FALL_CTL 0x7f
289 #define UB960_SR_CSI_FRAME_COUNT_HI(n) (0x90 + 8 * (n))
290 #define UB960_SR_CSI_FRAME_COUNT_LO(n) (0x91 + 8 * (n))
291 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n) (0x92 + 8 * (n))
292 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n) (0x93 + 8 * (n))
293 #define UB960_SR_CSI_LINE_COUNT_HI(n) (0x94 + 8 * (n))
294 #define UB960_SR_CSI_LINE_COUNT_LO(n) (0x95 + 8 * (n))
295 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n) (0x96 + 8 * (n))
296 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n) (0x97 + 8 * (n))
298 #define UB960_XR_REFCLK_FREQ 0xa5 /* UB960 */
300 #define UB960_RR_VC_ID_MAP(x) (0xa0 + (x)) /* UB9702 */
302 #define UB960_SR_IND_ACC_CTL 0xb0
303 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC BIT(1)
305 #define UB960_SR_IND_ACC_ADDR 0xb1
306 #define UB960_SR_IND_ACC_DATA 0xb2
307 #define UB960_SR_BIST_CONTROL 0xb3
308 #define UB960_SR_MODE_IDX_STS 0xb8
309 #define UB960_SR_LINK_ERROR_COUNT 0xb9
310 #define UB960_SR_FPD3_ENC_CTL 0xba
311 #define UB960_SR_FV_MIN_TIME 0xbc
312 #define UB960_SR_GPIO_PD_CTL 0xbe
314 #define UB960_SR_FPD_RATE_CFG 0xc2 /* UB9702 */
315 #define UB960_SR_CSI_PLL_DIV 0xc9 /* UB9702 */
317 #define UB960_RR_PORT_DEBUG 0xd0
318 #define UB960_RR_AEQ_CTL2 0xd2
319 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR BIT(2)
321 #define UB960_RR_AEQ_STATUS 0xd3
322 #define UB960_RR_AEQ_STATUS_STATUS_2 GENMASK(5, 3)
323 #define UB960_RR_AEQ_STATUS_STATUS_1 GENMASK(2, 0)
325 #define UB960_RR_AEQ_BYPASS 0xd4
326 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT 5
327 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK GENMASK(7, 5)
328 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT 1
329 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK GENMASK(3, 1)
330 #define UB960_RR_AEQ_BYPASS_ENABLE BIT(0)
332 #define UB960_RR_AEQ_MIN_MAX 0xd5
333 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT 4
334 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT 0
336 #define UB960_RR_SFILTER_STS_0 0xd6
337 #define UB960_RR_SFILTER_STS_1 0xd7
338 #define UB960_RR_PORT_ICR_HI 0xd8
339 #define UB960_RR_PORT_ICR_LO 0xd9
340 #define UB960_RR_PORT_ISR_HI 0xda
341 #define UB960_RR_PORT_ISR_LO 0xdb
342 #define UB960_RR_FC_GPIO_STS 0xdc
343 #define UB960_RR_FC_GPIO_ICR 0xdd
344 #define UB960_RR_SEN_INT_RISE_STS 0xde
345 #define UB960_RR_SEN_INT_FALL_STS 0xdf
347 #define UB960_RR_CHANNEL_MODE 0xe4 /* UB9702 */
349 #define UB960_SR_FPD3_RX_ID(n) (0xf0 + (n))
350 #define UB960_SR_FPD3_RX_ID_LEN 6
352 #define UB960_SR_I2C_RX_ID(n) (0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
354 /* Indirect register blocks */
355 #define UB960_IND_TARGET_PAT_GEN 0x00
356 #define UB960_IND_TARGET_RX_ANA(n) (0x01 + (n))
357 #define UB960_IND_TARGET_CSI_CSIPLL_REG_1 0x92 /* UB9702 */
358 #define UB960_IND_TARGET_CSI_ANA 0x07
360 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
362 #define UB960_IR_PGEN_CTL 0x01
363 #define UB960_IR_PGEN_CTL_PGEN_ENABLE BIT(0)
365 #define UB960_IR_PGEN_CFG 0x02
366 #define UB960_IR_PGEN_CSI_DI 0x03
367 #define UB960_IR_PGEN_LINE_SIZE1 0x04
368 #define UB960_IR_PGEN_LINE_SIZE0 0x05
369 #define UB960_IR_PGEN_BAR_SIZE1 0x06
370 #define UB960_IR_PGEN_BAR_SIZE0 0x07
371 #define UB960_IR_PGEN_ACT_LPF1 0x08
372 #define UB960_IR_PGEN_ACT_LPF0 0x09
373 #define UB960_IR_PGEN_TOT_LPF1 0x0a
374 #define UB960_IR_PGEN_TOT_LPF0 0x0b
375 #define UB960_IR_PGEN_LINE_PD1 0x0c
376 #define UB960_IR_PGEN_LINE_PD0 0x0d
377 #define UB960_IR_PGEN_VBP 0x0e
378 #define UB960_IR_PGEN_VFP 0x0f
379 #define UB960_IR_PGEN_COLOR(n) (0x10 + (n)) /* n < 15 */
381 #define UB960_IR_RX_ANA_STROBE_SET_CLK 0x08
382 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY BIT(3)
383 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK GENMASK(2, 0)
385 #define UB960_IR_RX_ANA_STROBE_SET_DATA 0x09
386 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY BIT(3)
387 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK GENMASK(2, 0)
391 #define UB960_MIN_AEQ_STROBE_POS -7
392 #define UB960_MAX_AEQ_STROBE_POS 7
394 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
396 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
397 #define UB960_MAX_MANUAL_STROBE_POS (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
398 #define UB960_NUM_MANUAL_STROBE_POS (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
400 #define UB960_MIN_EQ_LEVEL 0
401 #define UB960_MAX_EQ_LEVEL 14
402 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
404 struct ub960_hw_data {
412 enum ub960_rxport_mode {
413 RXPORT_MODE_RAW10 = 0,
414 RXPORT_MODE_RAW12_HF = 1,
415 RXPORT_MODE_RAW12_LF = 2,
416 RXPORT_MODE_CSI2_SYNC = 3,
417 RXPORT_MODE_CSI2_ASYNC = 4,
418 RXPORT_MODE_LAST = RXPORT_MODE_CSI2_ASYNC,
421 enum ub960_rxport_cdr {
424 RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
427 struct ub960_rxport {
428 struct ub960_data *priv;
429 u8 nport; /* RX port number, and index in priv->rxport[] */
432 struct v4l2_subdev *sd;
434 struct fwnode_handle *ep_fwnode;
439 struct fwnode_handle *fwnode;
440 struct i2c_client *client;
441 unsigned short alias; /* I2C alias (lower 7 bits) */
442 struct ds90ub9xx_platform_data pdata;
445 enum ub960_rxport_mode rx_mode;
446 enum ub960_rxport_cdr cdr_mode;
448 u8 lv_fv_pol; /* LV and FV polarities */
450 struct regulator *vpoc;
470 const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
474 struct v4l2_async_connection base;
475 struct ub960_rxport *rxport;
478 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
480 return container_of(asd, struct ub960_asd, base);
483 struct ub960_txport {
484 struct ub960_data *priv;
485 u8 nport; /* TX port number, and index in priv->txport[] */
491 const struct ub960_hw_data *hw_data;
492 struct i2c_client *client; /* for shared local registers */
493 struct regmap *regmap;
495 /* lock for register access */
496 struct mutex reg_lock;
500 struct regulator *vddio;
502 struct gpio_desc *pd_gpio;
503 struct delayed_work poll_work;
504 struct ub960_rxport *rxports[UB960_MAX_RX_NPORTS];
505 struct ub960_txport *txports[UB960_MAX_TX_NPORTS];
507 struct v4l2_subdev sd;
508 struct media_pad pads[UB960_MAX_NPORTS];
510 struct v4l2_ctrl_handler ctrl_handler;
511 struct v4l2_async_notifier notifier;
513 u32 tx_data_rate; /* Nominal data rate (Gb/s) */
528 u64 stream_enable_mask[UB960_MAX_NPORTS];
530 /* These are common to all ports */
539 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
541 return container_of(sd, struct ub960_data, sd);
544 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
546 return pad < priv->hw_data->num_rxports;
549 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
551 return pad >= priv->hw_data->num_rxports;
554 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
556 if (ub960_pad_is_sink(priv, pad))
559 return pad - priv->hw_data->num_rxports;
562 struct ub960_format_info {
569 static const struct ub960_format_info ub960_formats[] = {
570 { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
571 { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
572 { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
573 { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
575 { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
576 { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
577 { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
578 { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
581 static const struct ub960_format_info *ub960_find_format(u32 code)
585 for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
586 if (ub960_formats[i].code == code)
587 return &ub960_formats[i];
593 /* -----------------------------------------------------------------------------
594 * Basic device access
597 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
599 struct device *dev = &priv->client->dev;
603 mutex_lock(&priv->reg_lock);
605 ret = regmap_read(priv->regmap, reg, &v);
607 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
615 mutex_unlock(&priv->reg_lock);
620 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
622 struct device *dev = &priv->client->dev;
625 mutex_lock(&priv->reg_lock);
627 ret = regmap_write(priv->regmap, reg, val);
629 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
632 mutex_unlock(&priv->reg_lock);
637 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
639 struct device *dev = &priv->client->dev;
642 mutex_lock(&priv->reg_lock);
644 ret = regmap_update_bits(priv->regmap, reg, mask, val);
646 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
649 mutex_unlock(&priv->reg_lock);
654 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
656 struct device *dev = &priv->client->dev;
660 mutex_lock(&priv->reg_lock);
662 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
664 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
669 *val = be16_to_cpu(__v);
672 mutex_unlock(&priv->reg_lock);
677 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
679 struct device *dev = &priv->client->dev;
682 lockdep_assert_held(&priv->reg_lock);
684 if (priv->reg_current.rxport == nport)
687 ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
688 (nport << 4) | BIT(nport));
690 dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
695 priv->reg_current.rxport = nport;
700 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
702 struct device *dev = &priv->client->dev;
706 mutex_lock(&priv->reg_lock);
708 ret = ub960_rxport_select(priv, nport);
712 ret = regmap_read(priv->regmap, reg, &v);
714 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
722 mutex_unlock(&priv->reg_lock);
727 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
729 struct device *dev = &priv->client->dev;
732 mutex_lock(&priv->reg_lock);
734 ret = ub960_rxport_select(priv, nport);
738 ret = regmap_write(priv->regmap, reg, val);
740 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
744 mutex_unlock(&priv->reg_lock);
749 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
752 struct device *dev = &priv->client->dev;
755 mutex_lock(&priv->reg_lock);
757 ret = ub960_rxport_select(priv, nport);
761 ret = regmap_update_bits(priv->regmap, reg, mask, val);
763 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
767 mutex_unlock(&priv->reg_lock);
772 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
775 struct device *dev = &priv->client->dev;
779 mutex_lock(&priv->reg_lock);
781 ret = ub960_rxport_select(priv, nport);
785 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
787 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
792 *val = be16_to_cpu(__v);
795 mutex_unlock(&priv->reg_lock);
800 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
802 struct device *dev = &priv->client->dev;
805 lockdep_assert_held(&priv->reg_lock);
807 if (priv->reg_current.txport == nport)
810 ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
811 (nport << 4) | BIT(nport));
813 dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
818 priv->reg_current.txport = nport;
823 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
825 struct device *dev = &priv->client->dev;
829 mutex_lock(&priv->reg_lock);
831 ret = ub960_txport_select(priv, nport);
835 ret = regmap_read(priv->regmap, reg, &v);
837 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
845 mutex_unlock(&priv->reg_lock);
850 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
852 struct device *dev = &priv->client->dev;
855 mutex_lock(&priv->reg_lock);
857 ret = ub960_txport_select(priv, nport);
861 ret = regmap_write(priv->regmap, reg, val);
863 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
867 mutex_unlock(&priv->reg_lock);
872 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
875 struct device *dev = &priv->client->dev;
878 mutex_lock(&priv->reg_lock);
880 ret = ub960_txport_select(priv, nport);
884 ret = regmap_update_bits(priv->regmap, reg, mask, val);
886 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
890 mutex_unlock(&priv->reg_lock);
895 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
897 struct device *dev = &priv->client->dev;
900 lockdep_assert_held(&priv->reg_lock);
902 if (priv->reg_current.indirect_target == block)
905 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
907 dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
908 __func__, block, ret);
912 priv->reg_current.indirect_target = block;
917 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
919 struct device *dev = &priv->client->dev;
923 mutex_lock(&priv->reg_lock);
925 ret = ub960_select_ind_reg_block(priv, block);
929 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
932 "Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
937 ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
940 "Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
948 mutex_unlock(&priv->reg_lock);
953 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
955 struct device *dev = &priv->client->dev;
958 mutex_lock(&priv->reg_lock);
960 ret = ub960_select_ind_reg_block(priv, block);
964 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
967 "Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
972 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
975 "Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
981 mutex_unlock(&priv->reg_lock);
986 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
989 struct device *dev = &priv->client->dev;
992 mutex_lock(&priv->reg_lock);
994 ret = ub960_select_ind_reg_block(priv, block);
998 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
1001 "Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1006 ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1010 "Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1016 mutex_unlock(&priv->reg_lock);
1021 /* -----------------------------------------------------------------------------
1022 * I2C-ATR (address translator)
1025 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1026 const struct i2c_client *client, u16 alias)
1028 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1029 struct ub960_rxport *rxport = priv->rxports[chan_id];
1030 struct device *dev = &priv->client->dev;
1031 unsigned int reg_idx;
1033 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1034 if (!rxport->aliased_clients[reg_idx])
1038 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1039 dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1040 return -EADDRNOTAVAIL;
1043 rxport->aliased_clients[reg_idx] = client;
1045 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1047 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1050 dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1051 rxport->nport, client->addr, alias, reg_idx);
1056 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1057 const struct i2c_client *client)
1059 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1060 struct ub960_rxport *rxport = priv->rxports[chan_id];
1061 struct device *dev = &priv->client->dev;
1062 unsigned int reg_idx;
1064 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1065 if (rxport->aliased_clients[reg_idx] == client)
1069 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1070 dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1071 rxport->nport, client->addr);
1075 rxport->aliased_clients[reg_idx] = NULL;
1077 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1079 dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1080 client->addr, reg_idx);
1083 static const struct i2c_atr_ops ub960_atr_ops = {
1084 .attach_client = ub960_atr_attach_client,
1085 .detach_client = ub960_atr_detach_client,
1088 static int ub960_init_atr(struct ub960_data *priv)
1090 struct device *dev = &priv->client->dev;
1091 struct i2c_adapter *parent_adap = priv->client->adapter;
1093 priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1094 priv->hw_data->num_rxports);
1095 if (IS_ERR(priv->atr))
1096 return PTR_ERR(priv->atr);
1098 i2c_atr_set_driver_data(priv->atr, priv);
1103 static void ub960_uninit_atr(struct ub960_data *priv)
1105 i2c_atr_delete(priv->atr);
1109 /* -----------------------------------------------------------------------------
1113 static int ub960_parse_dt_txport(struct ub960_data *priv,
1114 struct fwnode_handle *ep_fwnode,
1117 struct device *dev = &priv->client->dev;
1118 struct v4l2_fwnode_endpoint vep = {};
1119 struct ub960_txport *txport;
1122 txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1126 txport->priv = priv;
1127 txport->nport = nport;
1129 vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1130 ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1132 dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1133 goto err_free_txport;
1136 txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1138 if (vep.nr_of_link_frequencies != 1) {
1143 priv->tx_link_freq[0] = vep.link_frequencies[0];
1144 priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1146 if (priv->tx_data_rate != MHZ(1600) &&
1147 priv->tx_data_rate != MHZ(1200) &&
1148 priv->tx_data_rate != MHZ(800) &&
1149 priv->tx_data_rate != MHZ(400)) {
1150 dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1155 v4l2_fwnode_endpoint_free(&vep);
1157 priv->txports[nport] = txport;
1162 v4l2_fwnode_endpoint_free(&vep);
1169 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1171 struct device *dev = &priv->client->dev;
1175 ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1179 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1180 dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1182 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1183 dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1186 /* -----------------------------------------------------------------------------
1190 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1195 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1196 struct ub960_rxport *rxport = priv->rxports[nport];
1198 if (!rxport || !rxport->vpoc)
1201 ret = regulator_enable(rxport->vpoc);
1203 goto err_disable_vpocs;
1210 struct ub960_rxport *rxport = priv->rxports[nport];
1212 if (!rxport || !rxport->vpoc)
1215 regulator_disable(rxport->vpoc);
1221 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1225 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1226 struct ub960_rxport *rxport = priv->rxports[nport];
1228 if (!rxport || !rxport->vpoc)
1231 regulator_disable(rxport->vpoc);
1235 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1240 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1241 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1242 ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1243 ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1245 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1246 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1248 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1251 static void ub960_clear_rx_errors(struct ub960_data *priv)
1255 for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1256 ub960_rxport_clear_errors(priv, nport);
1259 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1260 unsigned int nport, s8 *strobe_pos)
1263 u8 clk_delay, data_delay;
1266 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1267 UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1269 clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1270 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1272 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1273 UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1275 data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1276 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1278 ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1282 clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1284 ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1288 data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1290 *strobe_pos = data_delay - clk_delay;
1295 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1296 unsigned int nport, s8 strobe_pos)
1298 u8 clk_delay, data_delay;
1300 clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1301 data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1303 if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1304 clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1305 else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1306 data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1307 else if (strobe_pos < 0)
1308 clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1309 else if (strobe_pos > 0)
1310 data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1312 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1313 UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1315 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1316 UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1319 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1320 s8 strobe_min, s8 strobe_max)
1322 /* Convert the signed strobe pos to positive zero based value */
1323 strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1324 strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1326 ub960_write(priv, UB960_XR_SFILTER_CFG,
1327 ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1328 ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1331 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1332 unsigned int nport, u8 *eq_level)
1337 ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1341 *eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1342 (v & UB960_RR_AEQ_STATUS_STATUS_2);
1347 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1348 unsigned int nport, u8 eq_level)
1350 u8 eq_stage_1_select_value, eq_stage_2_select_value;
1351 const unsigned int eq_stage_max = 7;
1354 if (eq_level <= eq_stage_max) {
1355 eq_stage_1_select_value = eq_level;
1356 eq_stage_2_select_value = 0;
1358 eq_stage_1_select_value = eq_stage_max;
1359 eq_stage_2_select_value = eq_level - eq_stage_max;
1362 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1364 v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1365 UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1366 v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1367 v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1368 v |= UB960_RR_AEQ_BYPASS_ENABLE;
1370 ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1373 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1374 unsigned int nport, u8 eq_min, u8 eq_max)
1376 ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1377 (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1378 (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1380 /* Enable AEQ min setting */
1381 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1382 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1383 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1386 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1388 struct ub960_rxport *rxport = priv->rxports[nport];
1390 /* We also set common settings here. Should be moved elsewhere. */
1392 if (priv->strobe.manual) {
1393 /* Disable AEQ_SFILTER_EN */
1394 ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1395 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1397 /* Enable SFILTER and error control */
1398 ub960_write(priv, UB960_XR_AEQ_CTL1,
1399 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1400 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1402 /* Set AEQ strobe range */
1403 ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1407 /* The rest are port specific */
1409 if (priv->strobe.manual)
1410 ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1412 ub960_rxport_set_strobe_pos(priv, nport, 0);
1414 if (rxport->eq.manual_eq) {
1415 ub960_rxport_set_eq_level(priv, nport,
1416 rxport->eq.manual.eq_level);
1418 /* Enable AEQ Bypass */
1419 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1420 UB960_RR_AEQ_BYPASS_ENABLE,
1421 UB960_RR_AEQ_BYPASS_ENABLE);
1423 ub960_rxport_set_eq_range(priv, nport,
1424 rxport->eq.aeq.eq_level_min,
1425 rxport->eq.aeq.eq_level_max);
1427 /* Disable AEQ Bypass */
1428 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1429 UB960_RR_AEQ_BYPASS_ENABLE, 0);
1433 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1436 u8 rx_port_sts1, rx_port_sts2;
1444 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1449 if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1454 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1459 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1463 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1468 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1472 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1477 errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1478 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1479 (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1480 (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1489 * Wait for the RX ports to lock, have no errors and have stable strobe position
1492 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1493 unsigned long port_mask,
1494 unsigned int *lock_mask)
1496 struct device *dev = &priv->client->dev;
1497 unsigned long timeout;
1498 unsigned int link_ok_mask;
1499 unsigned int missing;
1504 if (port_mask == 0) {
1510 if (port_mask >= BIT(priv->hw_data->num_rxports))
1513 timeout = jiffies + msecs_to_jiffies(1000);
1517 while (time_before(jiffies, timeout)) {
1520 for_each_set_bit(nport, &port_mask,
1521 priv->hw_data->num_rxports) {
1522 struct ub960_rxport *rxport = priv->rxports[nport];
1528 ret = ub960_rxport_link_ok(priv, nport, &ok);
1533 * We want the link to be ok for two consecutive loops,
1534 * as a link could get established just before our test
1535 * and drop soon after.
1537 if (!ok || !(link_ok_mask & BIT(nport)))
1541 link_ok_mask |= BIT(nport);
1543 link_ok_mask &= ~BIT(nport);
1555 *lock_mask = link_ok_mask;
1557 dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1558 for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1559 struct ub960_rxport *rxport = priv->rxports[nport];
1560 s8 strobe_pos, eq_level;
1566 if (!(link_ok_mask & BIT(nport))) {
1567 dev_dbg(dev, "\trx%u: not locked\n", nport);
1571 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1573 ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
1577 ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1581 dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1582 nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
1588 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1589 struct ub960_rxport *rxport)
1594 switch (rxport->rx_mode) {
1595 case RXPORT_MODE_RAW10:
1596 case RXPORT_MODE_RAW12_HF:
1597 case RXPORT_MODE_RAW12_LF:
1602 case RXPORT_MODE_CSI2_SYNC:
1607 case RXPORT_MODE_CSI2_ASYNC:
1616 return clk_get_rate(priv->refclk) * mult / div;
1619 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1620 struct ub960_rxport *rxport)
1622 switch (rxport->rx_mode) {
1623 case RXPORT_MODE_RAW10:
1624 case RXPORT_MODE_RAW12_HF:
1625 case RXPORT_MODE_RAW12_LF:
1628 case RXPORT_MODE_CSI2_SYNC:
1631 case RXPORT_MODE_CSI2_ASYNC:
1639 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1641 struct ub960_rxport *rxport = priv->rxports[nport];
1642 struct device *dev = &priv->client->dev;
1643 struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1644 struct i2c_board_info ser_info = {
1645 .of_node = to_of_node(rxport->ser.fwnode),
1646 .fwnode = rxport->ser.fwnode,
1647 .platform_data = ser_pdata,
1650 ser_pdata->port = nport;
1651 ser_pdata->atr = priv->atr;
1652 if (priv->hw_data->is_ub9702)
1653 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1655 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1658 * The serializer is added under the same i2c adapter as the
1659 * deserializer. This is not quite right, as the serializer is behind
1662 ser_info.addr = rxport->ser.alias;
1663 rxport->ser.client =
1664 i2c_new_client_device(priv->client->adapter, &ser_info);
1665 if (IS_ERR(rxport->ser.client)) {
1666 dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1668 return PTR_ERR(rxport->ser.client);
1671 dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1672 nport, rxport->ser.client->addr,
1673 rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1678 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1680 struct ub960_rxport *rxport = priv->rxports[nport];
1682 i2c_unregister_device(rxport->ser.client);
1683 rxport->ser.client = NULL;
1686 /* Add serializer i2c devices for all initialized ports */
1687 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1692 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1693 struct ub960_rxport *rxport = priv->rxports[nport];
1698 ret = ub960_rxport_add_serializer(priv, nport);
1700 goto err_remove_sers;
1707 struct ub960_rxport *rxport = priv->rxports[nport];
1712 ub960_rxport_remove_serializer(priv, nport);
1718 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1722 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1723 struct ub960_rxport *rxport = priv->rxports[nport];
1728 ub960_rxport_remove_serializer(priv, nport);
1732 static void ub960_init_tx_port(struct ub960_data *priv,
1733 struct ub960_txport *txport)
1735 unsigned int nport = txport->nport;
1739 * From the datasheet: "initial CSI Skew-Calibration
1740 * sequence [...] should be set when operating at 1.6 Gbps"
1742 if (priv->tx_data_rate == MHZ(1600))
1743 csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1745 csi_ctl |= (4 - txport->num_data_lanes) << 4;
1747 ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1750 static int ub960_init_tx_ports(struct ub960_data *priv)
1758 switch (priv->tx_data_rate) {
1777 ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1779 if (priv->hw_data->is_ub9702) {
1780 ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1782 switch (priv->tx_data_rate) {
1785 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1786 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1789 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1790 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1791 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1794 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1799 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1800 struct ub960_txport *txport = priv->txports[nport];
1805 ub960_init_tx_port(priv, txport);
1811 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1812 struct ub960_rxport *rxport)
1814 unsigned int nport = rxport->nport;
1818 * Back channel frequency select.
1819 * Override FREQ_SELECT from the strap.
1820 * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1822 * 6 - 50 Mbps (DS90UB953-Q1)
1824 * Note that changing this setting will result in some errors on the back
1825 * channel for a short period of time.
1828 switch (rxport->rx_mode) {
1829 case RXPORT_MODE_RAW10:
1830 case RXPORT_MODE_RAW12_HF:
1831 case RXPORT_MODE_RAW12_LF:
1835 case RXPORT_MODE_CSI2_ASYNC:
1839 case RXPORT_MODE_CSI2_SYNC:
1847 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1848 UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1851 switch (rxport->rx_mode) {
1852 case RXPORT_MODE_RAW10:
1853 /* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1854 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1855 UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1859 * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1861 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1862 UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1863 0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1867 case RXPORT_MODE_RAW12_HF:
1868 case RXPORT_MODE_RAW12_LF:
1869 /* Not implemented */
1872 case RXPORT_MODE_CSI2_SYNC:
1873 case RXPORT_MODE_CSI2_ASYNC:
1874 /* CSI-2 Mode (DS90UB953-Q1 compatible) */
1875 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1881 /* LV_POLARITY & FV_POLARITY */
1882 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1885 /* Enable all interrupt sources from this port */
1886 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1887 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1889 /* Enable I2C_PASS_THROUGH */
1890 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1891 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1892 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1894 /* Enable I2C communication to the serializer via the alias addr */
1895 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1896 rxport->ser.alias << 1);
1898 /* Configure EQ related settings */
1899 ub960_rxport_config_eq(priv, nport);
1901 /* Enable RX port */
1902 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1905 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1906 struct ub960_rxport *rxport)
1908 unsigned int nport = rxport->nport;
1912 switch (rxport->rx_mode) {
1913 case RXPORT_MODE_RAW10:
1918 case RXPORT_MODE_RAW12_HF:
1923 case RXPORT_MODE_RAW12_LF:
1928 case RXPORT_MODE_CSI2_SYNC:
1933 case RXPORT_MODE_CSI2_ASYNC:
1942 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1944 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1946 /* set serdes_eq_mode = 1 */
1947 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1949 /* enable serdes driver */
1950 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1952 /* set serdes_eq_offset=4 */
1953 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1955 /* init default serdes_eq_max in 0xa9 */
1956 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
1958 /* init serdes_eq_min in 0xaa */
1959 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
1961 /* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
1962 ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
1965 /* RX port to half-rate */
1966 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
1970 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
1971 struct ub960_rxport *rxport)
1973 unsigned int nport = rxport->nport;
1974 bool first_time_power_up = true;
1976 if (first_time_power_up) {
1980 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
1982 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
1983 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
1985 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
1988 /* enable serdes_eq_ctl2 */
1989 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
1991 /* enable serdes_eq_ctl1 */
1992 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
1994 /* enable serdes_eq_en */
1995 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
1997 /* disable serdes_eq_override */
1998 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
2000 /* disable serdes_gain_override */
2001 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2004 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2005 struct ub960_rxport *rxport)
2007 unsigned int nport = rxport->nport;
2010 switch (rxport->rx_mode) {
2011 case RXPORT_MODE_RAW10:
2015 case RXPORT_MODE_RAW12_HF:
2019 case RXPORT_MODE_RAW12_LF:
2023 case RXPORT_MODE_CSI2_SYNC:
2027 case RXPORT_MODE_CSI2_ASYNC:
2035 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2038 /* FPD4 Sync Mode */
2039 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2041 /* add serdes_eq_offset of 4 */
2042 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2044 /* FPD4 serdes_start_eq in 0x27: assign default */
2045 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2046 /* FPD4 serdes_end_eq in 0x28: assign default */
2047 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2049 /* set serdes_driver_mode into FPD IV mode */
2050 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2051 /* set FPD PBC drv into FPD IV mode */
2052 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2054 /* set serdes_system_init to 0x2f */
2055 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2056 /* set serdes_system_rst in reset mode */
2057 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2059 /* RX port to 7.55G mode */
2060 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2063 ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2066 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2067 struct ub960_rxport *rxport)
2069 unsigned int nport = rxport->nport;
2071 if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2072 ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2073 else /* RXPORT_CDR_FPD4 */
2074 ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2076 switch (rxport->rx_mode) {
2077 case RXPORT_MODE_RAW10:
2079 * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2080 * 0b10 : 8-bit processing using upper 8 bits
2082 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2083 0x3 << 6, 0x2 << 6);
2087 case RXPORT_MODE_RAW12_HF:
2088 case RXPORT_MODE_RAW12_LF:
2089 /* Not implemented */
2092 case RXPORT_MODE_CSI2_SYNC:
2093 case RXPORT_MODE_CSI2_ASYNC:
2098 /* LV_POLARITY & FV_POLARITY */
2099 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2102 /* Enable all interrupt sources from this port */
2103 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2104 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2106 /* Enable I2C_PASS_THROUGH */
2107 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2108 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2109 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2111 /* Enable I2C communication to the serializer via the alias addr */
2112 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2113 rxport->ser.alias << 1);
2115 /* Enable RX port */
2116 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2118 if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2119 /* unreset 960 AEQ */
2120 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2124 static int ub960_init_rx_ports(struct ub960_data *priv)
2128 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2129 struct ub960_rxport *rxport = priv->rxports[nport];
2134 if (priv->hw_data->is_ub9702)
2135 ub960_init_rx_port_ub9702(priv, rxport);
2137 ub960_init_rx_port_ub960(priv, rxport);
2143 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2145 struct device *dev = &priv->client->dev;
2152 /* Read interrupts (also clears most of them) */
2154 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2157 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2160 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2163 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2169 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2172 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2175 dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2178 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2179 dev_err(dev, "rx%u BCC CRC error\n", nport);
2181 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2182 dev_err(dev, "rx%u BCC SEQ error\n", nport);
2184 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2185 dev_err(dev, "rx%u line length unstable\n", nport);
2187 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2188 dev_err(dev, "rx%u FPD3 encode error\n", nport);
2190 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2191 dev_err(dev, "rx%u buffer error\n", nport);
2194 dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2196 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2197 dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2199 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2200 dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2202 if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2203 dev_err(dev, "rx%u CSI checksum error\n", nport);
2205 if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2206 dev_err(dev, "rx%u CSI length error\n", nport);
2209 dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2211 if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2212 dev_err(dev, "rx%u BCC response error", nport);
2214 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2215 dev_err(dev, "rx%u BCC slave timeout", nport);
2217 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2218 dev_err(dev, "rx%u BCC slave error", nport);
2220 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2221 dev_err(dev, "rx%u BCC master timeout", nport);
2223 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2224 dev_err(dev, "rx%u BCC master error", nport);
2226 if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2227 dev_err(dev, "rx%u BCC sequence error", nport);
2229 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2232 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2234 dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2237 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2240 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2243 dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2246 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2247 dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2248 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2251 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2254 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2257 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2263 /* -----------------------------------------------------------------------------
2268 * The current implementation only supports a simple VC mapping, where all VCs
2269 * from a one RX port will be mapped to the same VC. Also, the hardware
2270 * dictates that all streams from an RX port must go to a single TX port.
2272 * This function decides the target VC numbers for each RX port with a simple
2273 * algorithm, so that for each TX port, we get VC numbers starting from 0,
2276 * E.g. if all four RX ports are in use, of which the first two go to the
2277 * first TX port and the secont two go to the second TX port, we would get
2278 * the following VCs for the four RX ports: 0, 1, 0, 1.
2280 * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2281 * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2282 * probably needs to be somehow configurable. Device tree?
2284 static void ub960_get_vc_maps(struct ub960_data *priv,
2285 struct v4l2_subdev_state *state, u8 *vc)
2287 u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2288 struct v4l2_subdev_route *route;
2289 u8 handled_mask = 0;
2291 for_each_active_route(&state->routing, route) {
2292 unsigned int rx, tx;
2294 rx = ub960_pad_to_port(priv, route->sink_pad);
2295 if (BIT(rx) & handled_mask)
2298 tx = ub960_pad_to_port(priv, route->source_pad);
2300 vc[rx] = cur_vc[tx]++;
2301 handled_mask |= BIT(rx);
2305 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2307 struct device *dev = &priv->client->dev;
2309 dev_dbg(dev, "enable TX port %u\n", nport);
2311 return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2312 UB960_TR_CSI_CTL_CSI_ENABLE,
2313 UB960_TR_CSI_CTL_CSI_ENABLE);
2316 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2318 struct device *dev = &priv->client->dev;
2320 dev_dbg(dev, "disable TX port %u\n", nport);
2322 ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2323 UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2326 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2328 struct device *dev = &priv->client->dev;
2330 dev_dbg(dev, "enable RX port %u\n", nport);
2332 /* Enable forwarding */
2333 return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2334 UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2337 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2339 struct device *dev = &priv->client->dev;
2341 dev_dbg(dev, "disable RX port %u\n", nport);
2343 /* Disable forwarding */
2344 ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2345 UB960_SR_FWD_CTL1_PORT_DIS(nport),
2346 UB960_SR_FWD_CTL1_PORT_DIS(nport));
2350 * The driver only supports using a single VC for each source. This function
2351 * checks that each source only provides streams using a single VC.
2353 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2358 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2359 struct ub960_rxport *rxport = priv->rxports[nport];
2360 struct v4l2_mbus_frame_desc desc;
2367 ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2368 rxport->source.pad, &desc);
2372 if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2375 if (desc.num_entries == 0)
2378 vc = desc.entry[0].bus.csi2.vc;
2380 for (i = 1; i < desc.num_entries; i++) {
2381 if (vc == desc.entry[i].bus.csi2.vc)
2384 dev_err(&priv->client->dev,
2385 "rx%u: source with multiple virtual-channels is not supported\n",
2394 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2395 struct v4l2_subdev_state *state)
2404 } rx_data[UB960_MAX_RX_NPORTS] = {};
2405 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2406 struct v4l2_subdev_route *route;
2410 ret = ub960_validate_stream_vcs(priv);
2414 ub960_get_vc_maps(priv, state, vc_map);
2416 for_each_active_route(&state->routing, route) {
2417 struct ub960_rxport *rxport;
2418 struct ub960_txport *txport;
2419 struct v4l2_mbus_framefmt *fmt;
2420 const struct ub960_format_info *ub960_fmt;
2423 nport = ub960_pad_to_port(priv, route->sink_pad);
2425 rxport = priv->rxports[nport];
2429 txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2433 rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2435 rx_data[nport].num_streams++;
2437 /* For the rest, we are only interested in parallel busses */
2438 if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2439 rxport->rx_mode == RXPORT_MODE_CSI2_ASYNC)
2442 if (rx_data[nport].num_streams > 2)
2445 fmt = v4l2_subdev_state_get_stream_format(state,
2447 route->sink_stream);
2451 ub960_fmt = ub960_find_format(fmt->code);
2455 if (ub960_fmt->meta) {
2456 if (fmt->height > 3) {
2457 dev_err(&priv->client->dev,
2458 "rx%u: unsupported metadata height %u\n",
2459 nport, fmt->height);
2463 rx_data[nport].meta_dt = ub960_fmt->datatype;
2464 rx_data[nport].meta_lines = fmt->height;
2466 rx_data[nport].pixel_dt = ub960_fmt->datatype;
2470 /* Configure RX ports */
2473 * Keep all port forwardings disabled by default. Forwarding will be
2474 * enabled in ub960_enable_rx_port.
2476 fwd_ctl = GENMASK(7, 4);
2478 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2479 struct ub960_rxport *rxport = priv->rxports[nport];
2480 u8 vc = vc_map[nport];
2482 if (rx_data[nport].num_streams == 0)
2485 switch (rxport->rx_mode) {
2486 case RXPORT_MODE_RAW10:
2487 ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2488 rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2490 ub960_rxport_write(priv, rxport->nport,
2491 UB960_RR_RAW_EMBED_DTYPE,
2492 (rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2493 rx_data[nport].meta_dt);
2497 case RXPORT_MODE_RAW12_HF:
2498 case RXPORT_MODE_RAW12_LF:
2499 /* Not implemented */
2502 case RXPORT_MODE_CSI2_SYNC:
2503 case RXPORT_MODE_CSI2_ASYNC:
2504 if (!priv->hw_data->is_ub9702) {
2505 /* Map all VCs from this port to the same VC */
2506 ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2507 (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2508 (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2509 (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2510 (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2514 /* Map all VCs from this port to VC(nport) */
2515 for (i = 0; i < 8; i++)
2516 ub960_rxport_write(priv, nport,
2517 UB960_RR_VC_ID_MAP(i),
2524 if (rx_data[nport].tx_port == 1)
2525 fwd_ctl |= BIT(nport); /* forward to TX1 */
2527 fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2530 ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2535 static void ub960_update_streaming_status(struct ub960_data *priv)
2539 for (i = 0; i < UB960_MAX_NPORTS; i++) {
2540 if (priv->stream_enable_mask[i])
2544 priv->streaming = i < UB960_MAX_NPORTS;
2547 static int ub960_enable_streams(struct v4l2_subdev *sd,
2548 struct v4l2_subdev_state *state, u32 source_pad,
2549 u64 source_streams_mask)
2551 struct ub960_data *priv = sd_to_ub960(sd);
2552 struct device *dev = &priv->client->dev;
2553 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2554 struct v4l2_subdev_route *route;
2555 unsigned int failed_port;
2559 if (!priv->streaming) {
2560 dev_dbg(dev, "Prepare for streaming\n");
2561 ret = ub960_configure_ports_for_streaming(priv, state);
2566 /* Enable TX port if not yet enabled */
2567 if (!priv->stream_enable_mask[source_pad]) {
2568 ret = ub960_enable_tx_port(priv,
2569 ub960_pad_to_port(priv, source_pad));
2574 priv->stream_enable_mask[source_pad] |= source_streams_mask;
2576 /* Collect sink streams per pad which we need to enable */
2577 for_each_active_route(&state->routing, route) {
2578 if (route->source_pad != source_pad)
2581 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2584 nport = ub960_pad_to_port(priv, route->sink_pad);
2586 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2589 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2590 if (!sink_streams[nport])
2593 /* Enable the RX port if not yet enabled */
2594 if (!priv->stream_enable_mask[nport]) {
2595 ret = ub960_enable_rx_port(priv, nport);
2597 failed_port = nport;
2602 priv->stream_enable_mask[nport] |= sink_streams[nport];
2604 dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2605 sink_streams[nport]);
2607 ret = v4l2_subdev_enable_streams(
2608 priv->rxports[nport]->source.sd,
2609 priv->rxports[nport]->source.pad,
2610 sink_streams[nport]);
2612 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2614 if (!priv->stream_enable_mask[nport])
2615 ub960_disable_rx_port(priv, nport);
2617 failed_port = nport;
2622 priv->streaming = true;
2627 for (nport = 0; nport < failed_port; nport++) {
2628 if (!sink_streams[nport])
2631 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2632 sink_streams[nport]);
2634 ret = v4l2_subdev_disable_streams(
2635 priv->rxports[nport]->source.sd,
2636 priv->rxports[nport]->source.pad,
2637 sink_streams[nport]);
2639 dev_err(dev, "Failed to disable streams: %d\n", ret);
2641 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2643 /* Disable RX port if no active streams */
2644 if (!priv->stream_enable_mask[nport])
2645 ub960_disable_rx_port(priv, nport);
2648 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2650 if (!priv->stream_enable_mask[source_pad])
2651 ub960_disable_tx_port(priv,
2652 ub960_pad_to_port(priv, source_pad));
2654 ub960_update_streaming_status(priv);
2659 static int ub960_disable_streams(struct v4l2_subdev *sd,
2660 struct v4l2_subdev_state *state,
2661 u32 source_pad, u64 source_streams_mask)
2663 struct ub960_data *priv = sd_to_ub960(sd);
2664 struct device *dev = &priv->client->dev;
2665 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2666 struct v4l2_subdev_route *route;
2670 /* Collect sink streams per pad which we need to disable */
2671 for_each_active_route(&state->routing, route) {
2672 if (route->source_pad != source_pad)
2675 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2678 nport = ub960_pad_to_port(priv, route->sink_pad);
2680 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2683 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2684 if (!sink_streams[nport])
2687 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2688 sink_streams[nport]);
2690 ret = v4l2_subdev_disable_streams(
2691 priv->rxports[nport]->source.sd,
2692 priv->rxports[nport]->source.pad,
2693 sink_streams[nport]);
2695 dev_err(dev, "Failed to disable streams: %d\n", ret);
2697 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2699 /* Disable RX port if no active streams */
2700 if (!priv->stream_enable_mask[nport])
2701 ub960_disable_rx_port(priv, nport);
2704 /* Disable TX port if no active streams */
2706 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2708 if (!priv->stream_enable_mask[source_pad])
2709 ub960_disable_tx_port(priv,
2710 ub960_pad_to_port(priv, source_pad));
2712 ub960_update_streaming_status(priv);
2717 static int _ub960_set_routing(struct v4l2_subdev *sd,
2718 struct v4l2_subdev_state *state,
2719 struct v4l2_subdev_krouting *routing)
2721 static const struct v4l2_mbus_framefmt format = {
2724 .code = MEDIA_BUS_FMT_UYVY8_1X16,
2725 .field = V4L2_FIELD_NONE,
2726 .colorspace = V4L2_COLORSPACE_SRGB,
2727 .ycbcr_enc = V4L2_YCBCR_ENC_601,
2728 .quantization = V4L2_QUANTIZATION_LIM_RANGE,
2729 .xfer_func = V4L2_XFER_FUNC_SRGB,
2734 * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2735 * frame desc is made dynamically allocated.
2738 if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2741 ret = v4l2_subdev_routing_validate(sd, routing,
2742 V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2743 V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2747 ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2754 static int ub960_set_routing(struct v4l2_subdev *sd,
2755 struct v4l2_subdev_state *state,
2756 enum v4l2_subdev_format_whence which,
2757 struct v4l2_subdev_krouting *routing)
2759 struct ub960_data *priv = sd_to_ub960(sd);
2761 if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2764 return _ub960_set_routing(sd, state, routing);
2767 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2768 struct v4l2_mbus_frame_desc *fd)
2770 struct ub960_data *priv = sd_to_ub960(sd);
2771 struct v4l2_subdev_route *route;
2772 struct v4l2_subdev_state *state;
2774 struct device *dev = &priv->client->dev;
2775 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2777 if (!ub960_pad_is_source(priv, pad))
2780 memset(fd, 0, sizeof(*fd));
2782 fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2784 state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2786 ub960_get_vc_maps(priv, state, vc_map);
2788 for_each_active_route(&state->routing, route) {
2789 struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2790 struct v4l2_mbus_frame_desc source_fd;
2794 if (route->source_pad != pad)
2797 nport = ub960_pad_to_port(priv, route->sink_pad);
2799 ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2801 priv->rxports[nport]->source.pad,
2805 "Failed to get source frame desc for pad %u\n",
2810 for (i = 0; i < source_fd.num_entries; i++) {
2811 if (source_fd.entry[i].stream == route->sink_stream) {
2812 source_entry = &source_fd.entry[i];
2817 if (!source_entry) {
2819 "Failed to find stream from source frame desc\n");
2824 fd->entry[fd->num_entries].stream = route->source_stream;
2825 fd->entry[fd->num_entries].flags = source_entry->flags;
2826 fd->entry[fd->num_entries].length = source_entry->length;
2827 fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2829 fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2831 if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2832 fd->entry[fd->num_entries].bus.csi2.dt =
2833 source_entry->bus.csi2.dt;
2835 const struct ub960_format_info *ub960_fmt;
2836 struct v4l2_mbus_framefmt *fmt;
2838 fmt = v4l2_subdev_state_get_stream_format(state, pad,
2839 route->source_stream);
2846 ub960_fmt = ub960_find_format(fmt->code);
2848 dev_err(dev, "Unable to find format\n");
2853 fd->entry[fd->num_entries].bus.csi2.dt =
2854 ub960_fmt->datatype;
2861 v4l2_subdev_unlock_state(state);
2866 static int ub960_set_fmt(struct v4l2_subdev *sd,
2867 struct v4l2_subdev_state *state,
2868 struct v4l2_subdev_format *format)
2870 struct ub960_data *priv = sd_to_ub960(sd);
2871 struct v4l2_mbus_framefmt *fmt;
2873 if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2876 /* No transcoding, source and sink formats must match. */
2877 if (ub960_pad_is_source(priv, format->pad))
2878 return v4l2_subdev_get_fmt(sd, state, format);
2881 * Default to the first format if the requested media bus code isn't
2884 if (!ub960_find_format(format->format.code))
2885 format->format.code = ub960_formats[0].code;
2887 fmt = v4l2_subdev_state_get_stream_format(state, format->pad,
2892 *fmt = format->format;
2894 fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2899 *fmt = format->format;
2904 static int ub960_init_cfg(struct v4l2_subdev *sd,
2905 struct v4l2_subdev_state *state)
2907 struct ub960_data *priv = sd_to_ub960(sd);
2909 struct v4l2_subdev_route routes[] = {
2913 .source_pad = priv->hw_data->num_rxports,
2915 .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2919 struct v4l2_subdev_krouting routing = {
2920 .num_routes = ARRAY_SIZE(routes),
2924 return _ub960_set_routing(sd, state, &routing);
2927 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2928 .enable_streams = ub960_enable_streams,
2929 .disable_streams = ub960_disable_streams,
2931 .set_routing = ub960_set_routing,
2932 .get_frame_desc = ub960_get_frame_desc,
2934 .get_fmt = v4l2_subdev_get_fmt,
2935 .set_fmt = ub960_set_fmt,
2937 .init_cfg = ub960_init_cfg,
2940 static int ub960_log_status(struct v4l2_subdev *sd)
2942 struct ub960_data *priv = sd_to_ub960(sd);
2943 struct device *dev = &priv->client->dev;
2944 struct v4l2_subdev_state *state;
2949 u8 id[UB960_SR_FPD3_RX_ID_LEN];
2951 state = v4l2_subdev_lock_and_get_active_state(sd);
2953 for (i = 0; i < sizeof(id); i++)
2954 ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
2956 dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
2958 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
2959 struct ub960_txport *txport = priv->txports[nport];
2961 dev_info(dev, "TX %u\n", nport);
2964 dev_info(dev, "\tNot initialized\n");
2968 ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
2969 dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
2972 ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
2973 dev_info(dev, "\tframe counter %u\n", v16);
2975 ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
2976 dev_info(dev, "\tframe error counter %u\n", v16);
2978 ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
2979 dev_info(dev, "\tline counter %u\n", v16);
2981 ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
2982 dev_info(dev, "\tline error counter %u\n", v16);
2985 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2986 struct ub960_rxport *rxport = priv->rxports[nport];
2991 dev_info(dev, "RX %u\n", nport);
2994 dev_info(dev, "\tNot initialized\n");
2998 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
3000 if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3001 dev_info(dev, "\tLocked\n");
3003 dev_info(dev, "\tNot locked\n");
3005 dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3006 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3007 dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3009 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3010 dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
3012 ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3013 dev_info(dev, "\tparity errors %u\n", v16);
3015 ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3016 dev_info(dev, "\tlines per frame %u\n", v16);
3018 ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3019 dev_info(dev, "\tbytes per line %u\n", v16);
3021 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3022 dev_info(dev, "\tcsi_err_counter %u\n", v);
3026 ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
3028 dev_info(dev, "\t%s strobe\n",
3029 (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
3032 if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
3033 ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
3035 dev_info(dev, "\tStrobe range [%d, %d]\n",
3036 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
3037 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
3040 ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
3042 dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
3046 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
3048 dev_info(dev, "\t%s EQ\n",
3049 (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
3052 if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
3053 ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
3055 dev_info(dev, "\tEQ range [%u, %u]\n",
3056 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
3057 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
3060 if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
3061 dev_info(dev, "\tEQ level %u\n", eq_level);
3064 for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3068 ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3069 ctl_shift = (i % 2) * 4;
3071 ub960_rxport_read(priv, nport, ctl_reg, &v);
3073 dev_info(dev, "\tGPIO%u: mode %u\n", i,
3074 (v >> ctl_shift) & 0xf);
3078 v4l2_subdev_unlock_state(state);
3083 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3084 .log_status = ub960_log_status,
3085 .subscribe_event = v4l2_ctrl_subdev_subscribe_event,
3086 .unsubscribe_event = v4l2_event_subdev_unsubscribe,
3089 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3090 .core = &ub960_subdev_core_ops,
3091 .pad = &ub960_pad_ops,
3094 static const struct media_entity_operations ub960_entity_ops = {
3095 .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3096 .link_validate = v4l2_subdev_link_validate,
3097 .has_pad_interdep = v4l2_subdev_has_pad_interdep,
3100 /* -----------------------------------------------------------------------------
3104 static irqreturn_t ub960_handle_events(int irq, void *arg)
3106 struct ub960_data *priv = arg;
3112 ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3113 if (ret || !int_sts)
3116 dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3118 ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3122 dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3124 for (i = 0; i < priv->hw_data->num_txports; i++) {
3125 if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3126 ub960_csi_handle_events(priv, i);
3129 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3130 if (!priv->rxports[i])
3133 if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3134 ub960_rxport_handle_events(priv, i);
3140 static void ub960_handler_work(struct work_struct *work)
3142 struct delayed_work *dwork = to_delayed_work(work);
3143 struct ub960_data *priv =
3144 container_of(dwork, struct ub960_data, poll_work);
3146 ub960_handle_events(0, priv);
3148 schedule_delayed_work(&priv->poll_work,
3149 msecs_to_jiffies(UB960_POLL_TIME_MS));
3152 static void ub960_txport_free_ports(struct ub960_data *priv)
3156 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3157 struct ub960_txport *txport = priv->txports[nport];
3163 priv->txports[nport] = NULL;
3167 static void ub960_rxport_free_ports(struct ub960_data *priv)
3171 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3172 struct ub960_rxport *rxport = priv->rxports[nport];
3177 fwnode_handle_put(rxport->source.ep_fwnode);
3178 fwnode_handle_put(rxport->ser.fwnode);
3181 priv->rxports[nport] = NULL;
3186 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3187 struct fwnode_handle *link_fwnode,
3188 struct ub960_rxport *rxport)
3190 struct device *dev = &priv->client->dev;
3191 unsigned int nport = rxport->nport;
3199 cdr_mode = RXPORT_CDR_FPD3;
3201 ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3202 if (ret < 0 && ret != -EINVAL) {
3203 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3204 "ti,cdr-mode", ret);
3208 if (cdr_mode > RXPORT_CDR_LAST) {
3209 dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3213 if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3214 dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3218 rxport->cdr_mode = cdr_mode;
3220 ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3222 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3227 if (rx_mode > RXPORT_MODE_LAST) {
3228 dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3233 case RXPORT_MODE_RAW12_HF:
3234 case RXPORT_MODE_RAW12_LF:
3235 case RXPORT_MODE_CSI2_ASYNC:
3236 dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3243 rxport->rx_mode = rx_mode;
3245 /* EQ & Strobe related */
3248 rxport->eq.manual_eq = false;
3249 rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3250 rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3252 ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3255 if (ret != -EINVAL) {
3256 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3257 "ti,strobe-pos", ret);
3261 if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3262 strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3263 dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3268 /* NOTE: ignored unless global manual strobe pos is also set */
3269 rxport->eq.strobe_pos = strobe_pos;
3270 if (!priv->strobe.manual)
3272 "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3276 ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3278 if (ret != -EINVAL) {
3279 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3280 "ti,eq-level", ret);
3284 if (eq_level > UB960_MAX_EQ_LEVEL) {
3285 dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3290 rxport->eq.manual_eq = true;
3291 rxport->eq.manual.eq_level = eq_level;
3294 ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3297 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3301 rxport->ser.alias = ser_i2c_alias;
3303 rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3304 if (!rxport->ser.fwnode) {
3305 dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3312 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3313 struct fwnode_handle *ep_fwnode,
3314 struct ub960_rxport *rxport)
3316 struct device *dev = &priv->client->dev;
3317 struct v4l2_fwnode_endpoint vep = {};
3318 unsigned int nport = rxport->nport;
3323 rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3324 if (!rxport->source.ep_fwnode) {
3325 dev_err(dev, "rx%u: no remote endpoint\n", nport);
3329 /* We currently have properties only for RAW modes */
3331 switch (rxport->rx_mode) {
3332 case RXPORT_MODE_RAW10:
3333 case RXPORT_MODE_RAW12_HF:
3334 case RXPORT_MODE_RAW12_LF:
3340 vep.bus_type = V4L2_MBUS_PARALLEL;
3341 ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3343 dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3344 goto err_put_source_ep_fwnode;
3347 hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3348 vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3350 /* LineValid and FrameValid are inverse to the h/vsync active */
3351 rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3352 (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3356 err_put_source_ep_fwnode:
3357 fwnode_handle_put(rxport->source.ep_fwnode);
3361 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3362 struct fwnode_handle *link_fwnode,
3363 struct fwnode_handle *ep_fwnode)
3365 static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3366 "vpoc0", "vpoc1", "vpoc2", "vpoc3"
3368 struct device *dev = &priv->client->dev;
3369 struct ub960_rxport *rxport;
3372 rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3376 priv->rxports[nport] = rxport;
3378 rxport->nport = nport;
3379 rxport->priv = priv;
3381 ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3383 goto err_free_rxport;
3385 rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3386 if (IS_ERR(rxport->vpoc)) {
3387 ret = PTR_ERR(rxport->vpoc);
3388 if (ret == -ENODEV) {
3389 rxport->vpoc = NULL;
3391 dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3393 goto err_put_remote_fwnode;
3397 ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3399 goto err_put_remote_fwnode;
3403 err_put_remote_fwnode:
3404 fwnode_handle_put(rxport->ser.fwnode);
3406 priv->rxports[nport] = NULL;
3411 static struct fwnode_handle *
3412 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3415 struct fwnode_handle *link_fwnode;
3418 fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3421 if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3424 ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3426 fwnode_handle_put(link_fwnode);
3430 if (nport == link_num)
3437 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3439 struct device *dev = &priv->client->dev;
3440 struct fwnode_handle *links_fwnode;
3444 links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3445 if (!links_fwnode) {
3446 dev_err(dev, "'links' node missing\n");
3450 /* Defaults, recommended by TI */
3451 priv->strobe.min = 2;
3452 priv->strobe.max = 3;
3454 priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3456 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3457 struct fwnode_handle *link_fwnode;
3458 struct fwnode_handle *ep_fwnode;
3460 link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3464 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3467 fwnode_handle_put(link_fwnode);
3471 ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3474 fwnode_handle_put(link_fwnode);
3475 fwnode_handle_put(ep_fwnode);
3478 dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3483 fwnode_handle_put(links_fwnode);
3488 fwnode_handle_put(links_fwnode);
3493 static int ub960_parse_dt_txports(struct ub960_data *priv)
3495 struct device *dev = &priv->client->dev;
3499 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3500 unsigned int port = nport + priv->hw_data->num_rxports;
3501 struct fwnode_handle *ep_fwnode;
3503 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3508 ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3510 fwnode_handle_put(ep_fwnode);
3519 static int ub960_parse_dt(struct ub960_data *priv)
3523 ret = ub960_parse_dt_rxports(priv);
3527 ret = ub960_parse_dt_txports(priv);
3529 goto err_free_rxports;
3534 ub960_rxport_free_ports(priv);
3539 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3540 struct v4l2_subdev *subdev,
3541 struct v4l2_async_connection *asd)
3543 struct ub960_data *priv = sd_to_ub960(notifier->sd);
3544 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3545 struct device *dev = &priv->client->dev;
3546 u8 nport = rxport->nport;
3550 ret = media_entity_get_fwnode_pad(&subdev->entity,
3551 rxport->source.ep_fwnode,
3552 MEDIA_PAD_FL_SOURCE);
3554 dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3558 rxport->source.sd = subdev;
3559 rxport->source.pad = ret;
3561 ret = media_create_pad_link(&rxport->source.sd->entity,
3562 rxport->source.pad, &priv->sd.entity, nport,
3563 MEDIA_LNK_FL_ENABLED |
3564 MEDIA_LNK_FL_IMMUTABLE);
3566 dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3567 rxport->source.sd->name, rxport->source.pad,
3568 priv->sd.name, nport);
3572 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3573 if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3574 dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3582 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3583 struct v4l2_subdev *subdev,
3584 struct v4l2_async_connection *asd)
3586 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3588 rxport->source.sd = NULL;
3591 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3592 .bound = ub960_notify_bound,
3593 .unbind = ub960_notify_unbind,
3596 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3598 struct device *dev = &priv->client->dev;
3602 v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
3604 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3605 struct ub960_rxport *rxport = priv->rxports[i];
3606 struct ub960_asd *asd;
3611 asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3612 rxport->source.ep_fwnode,
3615 dev_err(dev, "Failed to add subdev for source %u: %pe",
3617 v4l2_async_nf_cleanup(&priv->notifier);
3618 return PTR_ERR(asd);
3621 asd->rxport = rxport;
3624 priv->notifier.ops = &ub960_notify_ops;
3626 ret = v4l2_async_nf_register(&priv->notifier);
3628 dev_err(dev, "Failed to register subdev_notifier");
3629 v4l2_async_nf_cleanup(&priv->notifier);
3636 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3638 v4l2_async_nf_unregister(&priv->notifier);
3639 v4l2_async_nf_cleanup(&priv->notifier);
3642 static int ub960_create_subdev(struct ub960_data *priv)
3644 struct device *dev = &priv->client->dev;
3648 v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3650 v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3651 priv->sd.ctrl_handler = &priv->ctrl_handler;
3653 v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3654 ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3655 priv->tx_link_freq);
3657 if (priv->ctrl_handler.error) {
3658 ret = priv->ctrl_handler.error;
3662 priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3663 V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
3664 priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3665 priv->sd.entity.ops = &ub960_entity_ops;
3667 for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3668 priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3670 MEDIA_PAD_FL_SOURCE;
3673 ret = media_entity_pads_init(&priv->sd.entity,
3674 priv->hw_data->num_rxports +
3675 priv->hw_data->num_txports,
3680 priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3682 ret = v4l2_subdev_init_finalize(&priv->sd);
3684 goto err_entity_cleanup;
3686 ret = ub960_v4l2_notifier_register(priv);
3688 dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3689 goto err_subdev_cleanup;
3692 ret = v4l2_async_register_subdev(&priv->sd);
3694 dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3695 goto err_unreg_notif;
3701 ub960_v4l2_notifier_unregister(priv);
3703 v4l2_subdev_cleanup(&priv->sd);
3705 media_entity_cleanup(&priv->sd.entity);
3707 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3712 static void ub960_destroy_subdev(struct ub960_data *priv)
3714 ub960_v4l2_notifier_unregister(priv);
3715 v4l2_async_unregister_subdev(&priv->sd);
3717 v4l2_subdev_cleanup(&priv->sd);
3719 media_entity_cleanup(&priv->sd.entity);
3720 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3723 static const struct regmap_config ub960_regmap_config = {
3724 .name = "ds90ub960",
3729 .max_register = 0xff,
3732 * We do locking in the driver to cover the TX/RX port selection and the
3733 * indirect register access.
3735 .disable_locking = true,
3738 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3740 struct device *dev = &priv->client->dev;
3745 bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3746 UB960_SR_RESET_DIGITAL_RESET0;
3748 ub960_write(priv, UB960_SR_RESET, bit);
3750 mutex_lock(&priv->reg_lock);
3752 ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3753 (v & bit) == 0, 2000, 100000);
3755 mutex_unlock(&priv->reg_lock);
3758 dev_err(dev, "reset failed: %d\n", ret);
3761 static int ub960_get_hw_resources(struct ub960_data *priv)
3763 struct device *dev = &priv->client->dev;
3765 priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3766 if (IS_ERR(priv->regmap))
3767 return PTR_ERR(priv->regmap);
3769 priv->vddio = devm_regulator_get(dev, "vddio");
3770 if (IS_ERR(priv->vddio))
3771 return dev_err_probe(dev, PTR_ERR(priv->vddio),
3772 "cannot get VDDIO regulator\n");
3774 /* get power-down pin from DT */
3776 devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3777 if (IS_ERR(priv->pd_gpio))
3778 return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3779 "Cannot get powerdown GPIO\n");
3781 priv->refclk = devm_clk_get(dev, "refclk");
3782 if (IS_ERR(priv->refclk))
3783 return dev_err_probe(dev, PTR_ERR(priv->refclk),
3784 "Cannot get REFCLK\n");
3789 static int ub960_enable_core_hw(struct ub960_data *priv)
3791 struct device *dev = &priv->client->dev;
3797 ret = regulator_enable(priv->vddio);
3799 return dev_err_probe(dev, ret,
3800 "failed to enable VDDIO regulator\n");
3802 ret = clk_prepare_enable(priv->refclk);
3804 dev_err_probe(dev, ret, "Failed to enable refclk\n");
3805 goto err_disable_vddio;
3808 if (priv->pd_gpio) {
3809 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3810 /* wait min 2 ms for reset to complete */
3812 gpiod_set_value_cansleep(priv->pd_gpio, 0);
3813 /* wait min 2 ms for power up to finish */
3817 ub960_reset(priv, true);
3819 /* Runtime check register accessibility */
3820 ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3822 dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3826 dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3829 ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3833 ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3837 dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3838 !!(dev_sts & BIT(4)), refclk_freq,
3839 clk_get_rate(priv->refclk) / 1000000);
3841 /* Disable all RX ports by default */
3842 ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3846 /* release GPIO lock */
3847 if (priv->hw_data->is_ub9702) {
3848 ret = ub960_update_bits(priv, UB960_SR_RESET,
3849 UB960_SR_RESET_GPIO_LOCK_RELEASE,
3850 UB960_SR_RESET_GPIO_LOCK_RELEASE);
3858 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3859 clk_disable_unprepare(priv->refclk);
3861 regulator_disable(priv->vddio);
3866 static void ub960_disable_core_hw(struct ub960_data *priv)
3868 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3869 clk_disable_unprepare(priv->refclk);
3870 regulator_disable(priv->vddio);
3873 static int ub960_probe(struct i2c_client *client)
3875 struct device *dev = &client->dev;
3876 struct ub960_data *priv;
3877 unsigned int port_lock_mask;
3878 unsigned int port_mask;
3882 priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3886 priv->client = client;
3888 priv->hw_data = device_get_match_data(dev);
3890 mutex_init(&priv->reg_lock);
3892 INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3895 * Initialize these to invalid values so that the first reg writes will
3896 * configure the target.
3898 priv->reg_current.indirect_target = 0xff;
3899 priv->reg_current.rxport = 0xff;
3900 priv->reg_current.txport = 0xff;
3902 ret = ub960_get_hw_resources(priv);
3904 goto err_mutex_destroy;
3906 ret = ub960_enable_core_hw(priv);
3908 goto err_mutex_destroy;
3910 ret = ub960_parse_dt(priv);
3912 goto err_disable_core_hw;
3914 ret = ub960_init_tx_ports(priv);
3916 goto err_free_ports;
3918 ret = ub960_rxport_enable_vpocs(priv);
3920 goto err_free_ports;
3922 ret = ub960_init_rx_ports(priv);
3924 goto err_disable_vpocs;
3926 ub960_reset(priv, false);
3930 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3931 struct ub960_rxport *rxport = priv->rxports[nport];
3936 port_mask |= BIT(nport);
3939 ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
3941 goto err_disable_vpocs;
3943 if (port_mask != port_lock_mask) {
3945 dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
3946 goto err_disable_vpocs;
3950 * Clear any errors caused by switching the RX port settings while
3953 ub960_clear_rx_errors(priv);
3955 ret = ub960_init_atr(priv);
3957 goto err_disable_vpocs;
3959 ret = ub960_rxport_add_serializers(priv);
3961 goto err_uninit_atr;
3963 ret = ub960_create_subdev(priv);
3968 dev_warn(dev, "irq support not implemented, using polling\n");
3970 schedule_delayed_work(&priv->poll_work,
3971 msecs_to_jiffies(UB960_POLL_TIME_MS));
3976 ub960_rxport_remove_serializers(priv);
3978 ub960_uninit_atr(priv);
3980 ub960_rxport_disable_vpocs(priv);
3982 ub960_rxport_free_ports(priv);
3983 ub960_txport_free_ports(priv);
3984 err_disable_core_hw:
3985 ub960_disable_core_hw(priv);
3987 mutex_destroy(&priv->reg_lock);
3991 static void ub960_remove(struct i2c_client *client)
3993 struct v4l2_subdev *sd = i2c_get_clientdata(client);
3994 struct ub960_data *priv = sd_to_ub960(sd);
3996 cancel_delayed_work_sync(&priv->poll_work);
3998 ub960_destroy_subdev(priv);
3999 ub960_rxport_remove_serializers(priv);
4000 ub960_uninit_atr(priv);
4001 ub960_rxport_disable_vpocs(priv);
4002 ub960_rxport_free_ports(priv);
4003 ub960_txport_free_ports(priv);
4004 ub960_disable_core_hw(priv);
4005 mutex_destroy(&priv->reg_lock);
4008 static const struct ub960_hw_data ds90ub960_hw = {
4014 static const struct ub960_hw_data ds90ub9702_hw = {
4019 .is_fpdlink4 = true,
4022 static const struct i2c_device_id ub960_id[] = {
4023 { "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4024 { "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4027 MODULE_DEVICE_TABLE(i2c, ub960_id);
4029 static const struct of_device_id ub960_dt_ids[] = {
4030 { .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4031 { .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4034 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4036 static struct i2c_driver ds90ub960_driver = {
4037 .probe = ub960_probe,
4038 .remove = ub960_remove,
4039 .id_table = ub960_id,
4041 .name = "ds90ub960",
4042 .of_match_table = ub960_dt_ids,
4045 module_i2c_driver(ds90ub960_driver);
4047 MODULE_LICENSE("GPL");
4048 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4049 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4050 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4051 MODULE_IMPORT_NS(I2C_ATR);