1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
4 *
5 * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6 * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
7 */
8
9 /*
10 * (Possible) TODOs:
11 *
12 * - PM for serializer and remote peripherals. We need to manage:
13 * - VPOC
14 * - Power domain? Regulator? Somehow any remote device should be able to
15 * cause the VPOC to be turned on.
16 * - Link between the deserializer and the serializer
17 * - Related to VPOC management. We probably always want to turn on the VPOC
18 * and then enable the link.
19 * - Serializer's services: i2c, gpios, power
20 * - The serializer needs to resume before the remote peripherals can
21 * e.g. use the i2c.
22 * - How to handle gpios? Reserving a gpio essentially keeps the provider
23 * (serializer) always powered on.
24 * - Do we need a new bus for the FPD-Link? At the moment the serializers
25 * are children of the same i2c-adapter where the deserializer resides.
26 * - i2c-atr could be made embeddable instead of allocatable.
27 */
28
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/workqueue.h>
47
48 #include <media/i2c/ds90ub9xx.h>
49 #include <media/mipi-csi2.h>
50 #include <media/v4l2-ctrls.h>
51 #include <media/v4l2-event.h>
52 #include <media/v4l2-fwnode.h>
53 #include <media/v4l2-subdev.h>
54
55 #define MHZ(v) ((u32)((v) * 1000000U))
56
57 #define UB960_POLL_TIME_MS 500
58
59 #define UB960_MAX_RX_NPORTS 4
60 #define UB960_MAX_TX_NPORTS 2
61 #define UB960_MAX_NPORTS (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
62
63 #define UB960_MAX_PORT_ALIASES 8
64
65 #define UB960_NUM_BC_GPIOS 4
66
67 /*
68 * Register map
69 *
70 * 0x00-0x32 Shared (UB960_SR)
71 * 0x33-0x3a CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
72 * 0x4c Shared (UB960_SR)
73 * 0x4d-0x7f FPD-Link RX, per-port paged (UB960_RR)
74 * 0xb0-0xbf Shared (UB960_SR)
75 * 0xd0-0xdf FPD-Link RX, per-port paged (UB960_RR)
76 * 0xf0-0xf5 Shared (UB960_SR)
77 * 0xf8-0xfb Shared (UB960_SR)
78 * All others Reserved
79 *
80 * Register prefixes:
81 * UB960_SR_* = Shared register
82 * UB960_RR_* = FPD-Link RX, per-port paged register
83 * UB960_TR_* = CSI-2 TX, per-port paged register
84 * UB960_XR_* = Reserved register
85 * UB960_IR_* = Indirect register
86 */
87
88 #define UB960_SR_I2C_DEV_ID 0x00
89 #define UB960_SR_RESET 0x01
90 #define UB960_SR_RESET_DIGITAL_RESET1 BIT(1)
91 #define UB960_SR_RESET_DIGITAL_RESET0 BIT(0)
92 #define UB960_SR_RESET_GPIO_LOCK_RELEASE BIT(5)
93
94 #define UB960_SR_GEN_CONFIG 0x02
95 #define UB960_SR_REV_MASK 0x03
96 #define UB960_SR_DEVICE_STS 0x04
97 #define UB960_SR_PAR_ERR_THOLD_HI 0x05
98 #define UB960_SR_PAR_ERR_THOLD_LO 0x06
99 #define UB960_SR_BCC_WDOG_CTL 0x07
100 #define UB960_SR_I2C_CTL1 0x08
101 #define UB960_SR_I2C_CTL2 0x09
102 #define UB960_SR_SCL_HIGH_TIME 0x0a
103 #define UB960_SR_SCL_LOW_TIME 0x0b
104 #define UB960_SR_RX_PORT_CTL 0x0c
105 #define UB960_SR_IO_CTL 0x0d
106 #define UB960_SR_GPIO_PIN_STS 0x0e
107 #define UB960_SR_GPIO_INPUT_CTL 0x0f
108 #define UB960_SR_GPIO_PIN_CTL(n) (0x10 + (n)) /* n < UB960_NUM_GPIOS */
109 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL 5
110 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT 2
111 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN BIT(0)
112
113 #define UB960_SR_FS_CTL 0x18
114 #define UB960_SR_FS_HIGH_TIME_1 0x19
115 #define UB960_SR_FS_HIGH_TIME_0 0x1a
116 #define UB960_SR_FS_LOW_TIME_1 0x1b
117 #define UB960_SR_FS_LOW_TIME_0 0x1c
118 #define UB960_SR_MAX_FRM_HI 0x1d
119 #define UB960_SR_MAX_FRM_LO 0x1e
120 #define UB960_SR_CSI_PLL_CTL 0x1f
121
122 #define UB960_SR_FWD_CTL1 0x20
123 #define UB960_SR_FWD_CTL1_PORT_DIS(n) BIT((n) + 4)
124
125 #define UB960_SR_FWD_CTL2 0x21
126 #define UB960_SR_FWD_STS 0x22
127
128 #define UB960_SR_INTERRUPT_CTL 0x23
129 #define UB960_SR_INTERRUPT_CTL_INT_EN BIT(7)
130 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0 BIT(4)
131 #define UB960_SR_INTERRUPT_CTL_IE_RX(n) BIT((n)) /* rxport[n] IRQ */
132
133 #define UB960_SR_INTERRUPT_STS 0x24
134 #define UB960_SR_INTERRUPT_STS_INT BIT(7)
135 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n) BIT(4 + (n)) /* txport[n] IRQ */
136 #define UB960_SR_INTERRUPT_STS_IS_RX(n) BIT((n)) /* rxport[n] IRQ */
137
138 #define UB960_SR_TS_CONFIG 0x25
139 #define UB960_SR_TS_CONTROL 0x26
140 #define UB960_SR_TS_LINE_HI 0x27
141 #define UB960_SR_TS_LINE_LO 0x28
142 #define UB960_SR_TS_STATUS 0x29
143 #define UB960_SR_TIMESTAMP_P0_HI 0x2a
144 #define UB960_SR_TIMESTAMP_P0_LO 0x2b
145 #define UB960_SR_TIMESTAMP_P1_HI 0x2c
146 #define UB960_SR_TIMESTAMP_P1_LO 0x2d
147
148 #define UB960_SR_CSI_PORT_SEL 0x32
149
150 #define UB960_TR_CSI_CTL 0x33
151 #define UB960_TR_CSI_CTL_CSI_CAL_EN BIT(6)
152 #define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK BIT(1)
153 #define UB960_TR_CSI_CTL_CSI_ENABLE BIT(0)
154
155 #define UB960_TR_CSI_CTL2 0x34
156 #define UB960_TR_CSI_STS 0x35
157 #define UB960_TR_CSI_TX_ICR 0x36
158
159 #define UB960_TR_CSI_TX_ISR 0x37
160 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR BIT(3)
161 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR BIT(1)
162
163 #define UB960_TR_CSI_TEST_CTL 0x38
164 #define UB960_TR_CSI_TEST_PATT_HI 0x39
165 #define UB960_TR_CSI_TEST_PATT_LO 0x3a
166
167 #define UB960_XR_SFILTER_CFG 0x41
168 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT 4
169 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT 0
170
171 #define UB960_XR_AEQ_CTL1 0x42
172 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK BIT(6)
173 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING BIT(5)
174 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY BIT(4)
175 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK \
176 (UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK | \
177 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
178 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
179 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN BIT(0)
180
181 #define UB960_XR_AEQ_ERR_THOLD 0x43
182
183 #define UB960_RR_BCC_ERR_CTL 0x46
184 #define UB960_RR_BCC_STATUS 0x47
185 #define UB960_RR_BCC_STATUS_SEQ_ERROR BIT(5)
186 #define UB960_RR_BCC_STATUS_MASTER_ERR BIT(4)
187 #define UB960_RR_BCC_STATUS_MASTER_TO BIT(3)
188 #define UB960_RR_BCC_STATUS_SLAVE_ERR BIT(2)
189 #define UB960_RR_BCC_STATUS_SLAVE_TO BIT(1)
190 #define UB960_RR_BCC_STATUS_RESP_ERR BIT(0)
191 #define UB960_RR_BCC_STATUS_ERROR_MASK \
192 (UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
193 UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR | \
194 UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
195
196 #define UB960_RR_FPD3_CAP 0x4a
197 #define UB960_RR_RAW_EMBED_DTYPE 0x4b
198 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT 6
199
200 #define UB960_SR_FPD3_PORT_SEL 0x4c
201
202 #define UB960_RR_RX_PORT_STS1 0x4d
203 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR BIT(5)
204 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG BIT(4)
205 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR BIT(3)
206 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR BIT(2)
207 #define UB960_RR_RX_PORT_STS1_PORT_PASS BIT(1)
208 #define UB960_RR_RX_PORT_STS1_LOCK_STS BIT(0)
209 #define UB960_RR_RX_PORT_STS1_ERROR_MASK \
210 (UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
211 UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
212 UB960_RR_RX_PORT_STS1_PARITY_ERROR)
213
214 #define UB960_RR_RX_PORT_STS2 0x4e
215 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
216 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG BIT(6)
217 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
218 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR BIT(4)
219 #define UB960_RR_RX_PORT_STS2_CSI_ERROR BIT(3)
220 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE BIT(2)
221 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT BIT(1)
222 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG BIT(0)
223 #define UB960_RR_RX_PORT_STS2_ERROR_MASK \
224 UB960_RR_RX_PORT_STS2_BUFFER_ERROR
225
226 #define UB960_RR_RX_FREQ_HIGH 0x4f
227 #define UB960_RR_RX_FREQ_LOW 0x50
228 #define UB960_RR_SENSOR_STS_0 0x51
229 #define UB960_RR_SENSOR_STS_1 0x52
230 #define UB960_RR_SENSOR_STS_2 0x53
231 #define UB960_RR_SENSOR_STS_3 0x54
232 #define UB960_RR_RX_PAR_ERR_HI 0x55
233 #define UB960_RR_RX_PAR_ERR_LO 0x56
234 #define UB960_RR_BIST_ERR_COUNT 0x57
235
236 #define UB960_RR_BCC_CONFIG 0x58
237 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH BIT(6)
238 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK GENMASK(2, 0)
239
240 #define UB960_RR_DATAPATH_CTL1 0x59
241 #define UB960_RR_DATAPATH_CTL2 0x5a
242 #define UB960_RR_SER_ID 0x5b
243 #define UB960_RR_SER_ALIAS_ID 0x5c
244
245 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
246 #define UB960_RR_SLAVE_ID(n) (0x5d + (n))
247 #define UB960_RR_SLAVE_ALIAS(n) (0x65 + (n))
248
249 #define UB960_RR_PORT_CONFIG 0x6d
250 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK GENMASK(1, 0)
251
252 #define UB960_RR_BC_GPIO_CTL(n) (0x6e + (n)) /* n < 2 */
253 #define UB960_RR_RAW10_ID 0x70
254 #define UB960_RR_RAW10_ID_VC_SHIFT 6
255 #define UB960_RR_RAW10_ID_DT_SHIFT 0
256
257 #define UB960_RR_RAW12_ID 0x71
258 #define UB960_RR_CSI_VC_MAP 0x72
259 #define UB960_RR_CSI_VC_MAP_SHIFT(x) ((x) * 2)
260
261 #define UB960_RR_LINE_COUNT_HI 0x73
262 #define UB960_RR_LINE_COUNT_LO 0x74
263 #define UB960_RR_LINE_LEN_1 0x75
264 #define UB960_RR_LINE_LEN_0 0x76
265 #define UB960_RR_FREQ_DET_CTL 0x77
266 #define UB960_RR_MAILBOX_1 0x78
267 #define UB960_RR_MAILBOX_2 0x79
268
269 #define UB960_RR_CSI_RX_STS 0x7a
270 #define UB960_RR_CSI_RX_STS_LENGTH_ERR BIT(3)
271 #define UB960_RR_CSI_RX_STS_CKSUM_ERR BIT(2)
272 #define UB960_RR_CSI_RX_STS_ECC2_ERR BIT(1)
273 #define UB960_RR_CSI_RX_STS_ECC1_ERR BIT(0)
274 #define UB960_RR_CSI_RX_STS_ERROR_MASK \
275 (UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
276 UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
277
278 #define UB960_RR_CSI_ERR_COUNTER 0x7b
279 #define UB960_RR_PORT_CONFIG2 0x7c
280 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
281 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
282
283 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW BIT(1)
284 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW BIT(0)
285
286 #define UB960_RR_PORT_PASS_CTL 0x7d
287 #define UB960_RR_SEN_INT_RISE_CTL 0x7e
288 #define UB960_RR_SEN_INT_FALL_CTL 0x7f
289
290 #define UB960_SR_CSI_FRAME_COUNT_HI(n) (0x90 + 8 * (n))
291 #define UB960_SR_CSI_FRAME_COUNT_LO(n) (0x91 + 8 * (n))
292 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n) (0x92 + 8 * (n))
293 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n) (0x93 + 8 * (n))
294 #define UB960_SR_CSI_LINE_COUNT_HI(n) (0x94 + 8 * (n))
295 #define UB960_SR_CSI_LINE_COUNT_LO(n) (0x95 + 8 * (n))
296 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n) (0x96 + 8 * (n))
297 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n) (0x97 + 8 * (n))
298
299 #define UB960_XR_REFCLK_FREQ 0xa5 /* UB960 */
300
301 #define UB960_RR_VC_ID_MAP(x) (0xa0 + (x)) /* UB9702 */
302
303 #define UB960_SR_IND_ACC_CTL 0xb0
304 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC BIT(1)
305
306 #define UB960_SR_IND_ACC_ADDR 0xb1
307 #define UB960_SR_IND_ACC_DATA 0xb2
308 #define UB960_SR_BIST_CONTROL 0xb3
309 #define UB960_SR_MODE_IDX_STS 0xb8
310 #define UB960_SR_LINK_ERROR_COUNT 0xb9
311 #define UB960_SR_FPD3_ENC_CTL 0xba
312 #define UB960_SR_FV_MIN_TIME 0xbc
313 #define UB960_SR_GPIO_PD_CTL 0xbe
314
315 #define UB960_SR_FPD_RATE_CFG 0xc2 /* UB9702 */
316 #define UB960_SR_CSI_PLL_DIV 0xc9 /* UB9702 */
317
318 #define UB960_RR_PORT_DEBUG 0xd0
319 #define UB960_RR_AEQ_CTL2 0xd2
320 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR BIT(2)
321
322 #define UB960_RR_AEQ_STATUS 0xd3
323 #define UB960_RR_AEQ_STATUS_STATUS_2 GENMASK(5, 3)
324 #define UB960_RR_AEQ_STATUS_STATUS_1 GENMASK(2, 0)
325
326 #define UB960_RR_AEQ_BYPASS 0xd4
327 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT 5
328 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK GENMASK(7, 5)
329 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT 1
330 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK GENMASK(3, 1)
331 #define UB960_RR_AEQ_BYPASS_ENABLE BIT(0)
332
333 #define UB960_RR_AEQ_MIN_MAX 0xd5
334 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT 4
335 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT 0
336
337 #define UB960_RR_SFILTER_STS_0 0xd6
338 #define UB960_RR_SFILTER_STS_1 0xd7
339 #define UB960_RR_PORT_ICR_HI 0xd8
340 #define UB960_RR_PORT_ICR_LO 0xd9
341 #define UB960_RR_PORT_ISR_HI 0xda
342 #define UB960_RR_PORT_ISR_LO 0xdb
343 #define UB960_RR_FC_GPIO_STS 0xdc
344 #define UB960_RR_FC_GPIO_ICR 0xdd
345 #define UB960_RR_SEN_INT_RISE_STS 0xde
346 #define UB960_RR_SEN_INT_FALL_STS 0xdf
347
348 #define UB960_RR_CHANNEL_MODE 0xe4 /* UB9702 */
349
350 #define UB960_SR_FPD3_RX_ID(n) (0xf0 + (n))
351 #define UB960_SR_FPD3_RX_ID_LEN 6
352
353 #define UB960_SR_I2C_RX_ID(n) (0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
354
355 #define UB9702_SR_REFCLK_FREQ 0x3d
356
357 /* Indirect register blocks */
358 #define UB960_IND_TARGET_PAT_GEN 0x00
359 #define UB960_IND_TARGET_RX_ANA(n) (0x01 + (n))
360 #define UB960_IND_TARGET_CSI_CSIPLL_REG_1 0x92 /* UB9702 */
361 #define UB960_IND_TARGET_CSI_ANA 0x07
362
363 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
364
365 #define UB960_IR_PGEN_CTL 0x01
366 #define UB960_IR_PGEN_CTL_PGEN_ENABLE BIT(0)
367
368 #define UB960_IR_PGEN_CFG 0x02
369 #define UB960_IR_PGEN_CSI_DI 0x03
370 #define UB960_IR_PGEN_LINE_SIZE1 0x04
371 #define UB960_IR_PGEN_LINE_SIZE0 0x05
372 #define UB960_IR_PGEN_BAR_SIZE1 0x06
373 #define UB960_IR_PGEN_BAR_SIZE0 0x07
374 #define UB960_IR_PGEN_ACT_LPF1 0x08
375 #define UB960_IR_PGEN_ACT_LPF0 0x09
376 #define UB960_IR_PGEN_TOT_LPF1 0x0a
377 #define UB960_IR_PGEN_TOT_LPF0 0x0b
378 #define UB960_IR_PGEN_LINE_PD1 0x0c
379 #define UB960_IR_PGEN_LINE_PD0 0x0d
380 #define UB960_IR_PGEN_VBP 0x0e
381 #define UB960_IR_PGEN_VFP 0x0f
382 #define UB960_IR_PGEN_COLOR(n) (0x10 + (n)) /* n < 15 */
383
384 #define UB960_IR_RX_ANA_STROBE_SET_CLK 0x08
385 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY BIT(3)
386 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK GENMASK(2, 0)
387
388 #define UB960_IR_RX_ANA_STROBE_SET_DATA 0x09
389 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY BIT(3)
390 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK GENMASK(2, 0)
391
392 /* EQ related */
393
394 #define UB960_MIN_AEQ_STROBE_POS -7
395 #define UB960_MAX_AEQ_STROBE_POS 7
396
397 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
398
399 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
400 #define UB960_MAX_MANUAL_STROBE_POS (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
401 #define UB960_NUM_MANUAL_STROBE_POS (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
402
403 #define UB960_MIN_EQ_LEVEL 0
404 #define UB960_MAX_EQ_LEVEL 14
405 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
406
407 struct ub960_hw_data {
408 const char *model;
409 u8 num_rxports;
410 u8 num_txports;
411 bool is_ub9702;
412 bool is_fpdlink4;
413 };
414
415 enum ub960_rxport_mode {
416 RXPORT_MODE_RAW10 = 0,
417 RXPORT_MODE_RAW12_HF = 1,
418 RXPORT_MODE_RAW12_LF = 2,
419 RXPORT_MODE_CSI2_SYNC = 3,
420 RXPORT_MODE_CSI2_NONSYNC = 4,
421 RXPORT_MODE_LAST = RXPORT_MODE_CSI2_NONSYNC,
422 };
423
424 enum ub960_rxport_cdr {
425 RXPORT_CDR_FPD3 = 0,
426 RXPORT_CDR_FPD4 = 1,
427 RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
428 };
429
430 struct ub960_rxport {
431 struct ub960_data *priv;
432 u8 nport; /* RX port number, and index in priv->rxport[] */
433
434 struct {
435 struct v4l2_subdev *sd;
436 u16 pad;
437 struct fwnode_handle *ep_fwnode;
438 } source;
439
440 /* Serializer */
441 struct {
442 struct fwnode_handle *fwnode;
443 struct i2c_client *client;
444 unsigned short alias; /* I2C alias (lower 7 bits) */
445 struct ds90ub9xx_platform_data pdata;
446 } ser;
447
448 enum ub960_rxport_mode rx_mode;
449 enum ub960_rxport_cdr cdr_mode;
450
451 u8 lv_fv_pol; /* LV and FV polarities */
452
453 struct regulator *vpoc;
454
455 /* EQ settings */
456 struct {
457 bool manual_eq;
458
459 s8 strobe_pos;
460
461 union {
462 struct {
463 u8 eq_level_min;
464 u8 eq_level_max;
465 } aeq;
466
467 struct {
468 u8 eq_level;
469 } manual;
470 };
471 } eq;
472
473 const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
474 };
475
476 struct ub960_asd {
477 struct v4l2_async_connection base;
478 struct ub960_rxport *rxport;
479 };
480
to_ub960_asd(struct v4l2_async_connection * asd)481 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
482 {
483 return container_of(asd, struct ub960_asd, base);
484 }
485
486 struct ub960_txport {
487 struct ub960_data *priv;
488 u8 nport; /* TX port number, and index in priv->txport[] */
489
490 u32 num_data_lanes;
491 bool non_continous_clk;
492 };
493
494 struct ub960_data {
495 const struct ub960_hw_data *hw_data;
496 struct i2c_client *client; /* for shared local registers */
497 struct regmap *regmap;
498
499 /* lock for register access */
500 struct mutex reg_lock;
501
502 struct clk *refclk;
503
504 struct regulator *vddio;
505
506 struct gpio_desc *pd_gpio;
507 struct delayed_work poll_work;
508 struct ub960_rxport *rxports[UB960_MAX_RX_NPORTS];
509 struct ub960_txport *txports[UB960_MAX_TX_NPORTS];
510
511 struct v4l2_subdev sd;
512 struct media_pad pads[UB960_MAX_NPORTS];
513
514 struct v4l2_ctrl_handler ctrl_handler;
515 struct v4l2_async_notifier notifier;
516
517 u32 tx_data_rate; /* Nominal data rate (Gb/s) */
518 s64 tx_link_freq[1];
519
520 struct i2c_atr *atr;
521
522 struct {
523 u8 rxport;
524 u8 txport;
525 u8 indirect_target;
526 } reg_current;
527
528 bool streaming;
529
530 u8 stored_fwd_ctl;
531
532 u64 stream_enable_mask[UB960_MAX_NPORTS];
533
534 /* These are common to all ports */
535 struct {
536 bool manual;
537
538 s8 min;
539 s8 max;
540 } strobe;
541 };
542
sd_to_ub960(struct v4l2_subdev * sd)543 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
544 {
545 return container_of(sd, struct ub960_data, sd);
546 }
547
ub960_pad_is_sink(struct ub960_data * priv,u32 pad)548 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
549 {
550 return pad < priv->hw_data->num_rxports;
551 }
552
ub960_pad_is_source(struct ub960_data * priv,u32 pad)553 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
554 {
555 return pad >= priv->hw_data->num_rxports;
556 }
557
ub960_pad_to_port(struct ub960_data * priv,u32 pad)558 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
559 {
560 if (ub960_pad_is_sink(priv, pad))
561 return pad;
562 else
563 return pad - priv->hw_data->num_rxports;
564 }
565
566 struct ub960_format_info {
567 u32 code;
568 u32 bpp;
569 u8 datatype;
570 bool meta;
571 };
572
573 static const struct ub960_format_info ub960_formats[] = {
574 { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
575 { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
576 { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
577 { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
578
579 { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
580 { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
581 { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
582 { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
583 };
584
ub960_find_format(u32 code)585 static const struct ub960_format_info *ub960_find_format(u32 code)
586 {
587 unsigned int i;
588
589 for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
590 if (ub960_formats[i].code == code)
591 return &ub960_formats[i];
592 }
593
594 return NULL;
595 }
596
597 /* -----------------------------------------------------------------------------
598 * Basic device access
599 */
600
ub960_read(struct ub960_data * priv,u8 reg,u8 * val)601 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
602 {
603 struct device *dev = &priv->client->dev;
604 unsigned int v;
605 int ret;
606
607 mutex_lock(&priv->reg_lock);
608
609 ret = regmap_read(priv->regmap, reg, &v);
610 if (ret) {
611 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
612 __func__, reg, ret);
613 goto out_unlock;
614 }
615
616 *val = v;
617
618 out_unlock:
619 mutex_unlock(&priv->reg_lock);
620
621 return ret;
622 }
623
ub960_write(struct ub960_data * priv,u8 reg,u8 val)624 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
625 {
626 struct device *dev = &priv->client->dev;
627 int ret;
628
629 mutex_lock(&priv->reg_lock);
630
631 ret = regmap_write(priv->regmap, reg, val);
632 if (ret)
633 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
634 __func__, reg, ret);
635
636 mutex_unlock(&priv->reg_lock);
637
638 return ret;
639 }
640
ub960_update_bits(struct ub960_data * priv,u8 reg,u8 mask,u8 val)641 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
642 {
643 struct device *dev = &priv->client->dev;
644 int ret;
645
646 mutex_lock(&priv->reg_lock);
647
648 ret = regmap_update_bits(priv->regmap, reg, mask, val);
649 if (ret)
650 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
651 __func__, reg, ret);
652
653 mutex_unlock(&priv->reg_lock);
654
655 return ret;
656 }
657
ub960_read16(struct ub960_data * priv,u8 reg,u16 * val)658 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
659 {
660 struct device *dev = &priv->client->dev;
661 __be16 __v;
662 int ret;
663
664 mutex_lock(&priv->reg_lock);
665
666 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
667 if (ret) {
668 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
669 __func__, reg, ret);
670 goto out_unlock;
671 }
672
673 *val = be16_to_cpu(__v);
674
675 out_unlock:
676 mutex_unlock(&priv->reg_lock);
677
678 return ret;
679 }
680
ub960_rxport_select(struct ub960_data * priv,u8 nport)681 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
682 {
683 struct device *dev = &priv->client->dev;
684 int ret;
685
686 lockdep_assert_held(&priv->reg_lock);
687
688 if (priv->reg_current.rxport == nport)
689 return 0;
690
691 ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
692 (nport << 4) | BIT(nport));
693 if (ret) {
694 dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
695 nport, ret);
696 return ret;
697 }
698
699 priv->reg_current.rxport = nport;
700
701 return 0;
702 }
703
ub960_rxport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)704 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
705 {
706 struct device *dev = &priv->client->dev;
707 unsigned int v;
708 int ret;
709
710 mutex_lock(&priv->reg_lock);
711
712 ret = ub960_rxport_select(priv, nport);
713 if (ret)
714 goto out_unlock;
715
716 ret = regmap_read(priv->regmap, reg, &v);
717 if (ret) {
718 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
719 __func__, reg, ret);
720 goto out_unlock;
721 }
722
723 *val = v;
724
725 out_unlock:
726 mutex_unlock(&priv->reg_lock);
727
728 return ret;
729 }
730
ub960_rxport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)731 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
732 {
733 struct device *dev = &priv->client->dev;
734 int ret;
735
736 mutex_lock(&priv->reg_lock);
737
738 ret = ub960_rxport_select(priv, nport);
739 if (ret)
740 goto out_unlock;
741
742 ret = regmap_write(priv->regmap, reg, val);
743 if (ret)
744 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
745 __func__, reg, ret);
746
747 out_unlock:
748 mutex_unlock(&priv->reg_lock);
749
750 return ret;
751 }
752
ub960_rxport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)753 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
754 u8 mask, u8 val)
755 {
756 struct device *dev = &priv->client->dev;
757 int ret;
758
759 mutex_lock(&priv->reg_lock);
760
761 ret = ub960_rxport_select(priv, nport);
762 if (ret)
763 goto out_unlock;
764
765 ret = regmap_update_bits(priv->regmap, reg, mask, val);
766 if (ret)
767 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
768 __func__, reg, ret);
769
770 out_unlock:
771 mutex_unlock(&priv->reg_lock);
772
773 return ret;
774 }
775
ub960_rxport_read16(struct ub960_data * priv,u8 nport,u8 reg,u16 * val)776 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
777 u16 *val)
778 {
779 struct device *dev = &priv->client->dev;
780 __be16 __v;
781 int ret;
782
783 mutex_lock(&priv->reg_lock);
784
785 ret = ub960_rxport_select(priv, nport);
786 if (ret)
787 goto out_unlock;
788
789 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
790 if (ret) {
791 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
792 __func__, reg, ret);
793 goto out_unlock;
794 }
795
796 *val = be16_to_cpu(__v);
797
798 out_unlock:
799 mutex_unlock(&priv->reg_lock);
800
801 return ret;
802 }
803
ub960_txport_select(struct ub960_data * priv,u8 nport)804 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
805 {
806 struct device *dev = &priv->client->dev;
807 int ret;
808
809 lockdep_assert_held(&priv->reg_lock);
810
811 if (priv->reg_current.txport == nport)
812 return 0;
813
814 ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
815 (nport << 4) | BIT(nport));
816 if (ret) {
817 dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
818 nport, ret);
819 return ret;
820 }
821
822 priv->reg_current.txport = nport;
823
824 return 0;
825 }
826
ub960_txport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)827 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
828 {
829 struct device *dev = &priv->client->dev;
830 unsigned int v;
831 int ret;
832
833 mutex_lock(&priv->reg_lock);
834
835 ret = ub960_txport_select(priv, nport);
836 if (ret)
837 goto out_unlock;
838
839 ret = regmap_read(priv->regmap, reg, &v);
840 if (ret) {
841 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
842 __func__, reg, ret);
843 goto out_unlock;
844 }
845
846 *val = v;
847
848 out_unlock:
849 mutex_unlock(&priv->reg_lock);
850
851 return ret;
852 }
853
ub960_txport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)854 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
855 {
856 struct device *dev = &priv->client->dev;
857 int ret;
858
859 mutex_lock(&priv->reg_lock);
860
861 ret = ub960_txport_select(priv, nport);
862 if (ret)
863 goto out_unlock;
864
865 ret = regmap_write(priv->regmap, reg, val);
866 if (ret)
867 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
868 __func__, reg, ret);
869
870 out_unlock:
871 mutex_unlock(&priv->reg_lock);
872
873 return ret;
874 }
875
ub960_txport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)876 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
877 u8 mask, u8 val)
878 {
879 struct device *dev = &priv->client->dev;
880 int ret;
881
882 mutex_lock(&priv->reg_lock);
883
884 ret = ub960_txport_select(priv, nport);
885 if (ret)
886 goto out_unlock;
887
888 ret = regmap_update_bits(priv->regmap, reg, mask, val);
889 if (ret)
890 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
891 __func__, reg, ret);
892
893 out_unlock:
894 mutex_unlock(&priv->reg_lock);
895
896 return ret;
897 }
898
ub960_select_ind_reg_block(struct ub960_data * priv,u8 block)899 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
900 {
901 struct device *dev = &priv->client->dev;
902 int ret;
903
904 lockdep_assert_held(&priv->reg_lock);
905
906 if (priv->reg_current.indirect_target == block)
907 return 0;
908
909 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
910 if (ret) {
911 dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
912 __func__, block, ret);
913 return ret;
914 }
915
916 priv->reg_current.indirect_target = block;
917
918 return 0;
919 }
920
ub960_read_ind(struct ub960_data * priv,u8 block,u8 reg,u8 * val)921 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
922 {
923 struct device *dev = &priv->client->dev;
924 unsigned int v;
925 int ret;
926
927 mutex_lock(&priv->reg_lock);
928
929 ret = ub960_select_ind_reg_block(priv, block);
930 if (ret)
931 goto out_unlock;
932
933 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
934 if (ret) {
935 dev_err(dev,
936 "Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
937 block, reg, ret);
938 goto out_unlock;
939 }
940
941 ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
942 if (ret) {
943 dev_err(dev,
944 "Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
945 block, reg, ret);
946 goto out_unlock;
947 }
948
949 *val = v;
950
951 out_unlock:
952 mutex_unlock(&priv->reg_lock);
953
954 return ret;
955 }
956
ub960_write_ind(struct ub960_data * priv,u8 block,u8 reg,u8 val)957 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
958 {
959 struct device *dev = &priv->client->dev;
960 int ret;
961
962 mutex_lock(&priv->reg_lock);
963
964 ret = ub960_select_ind_reg_block(priv, block);
965 if (ret)
966 goto out_unlock;
967
968 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
969 if (ret) {
970 dev_err(dev,
971 "Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
972 block, reg, ret);
973 goto out_unlock;
974 }
975
976 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
977 if (ret) {
978 dev_err(dev,
979 "Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
980 block, reg, ret);
981 goto out_unlock;
982 }
983
984 out_unlock:
985 mutex_unlock(&priv->reg_lock);
986
987 return ret;
988 }
989
ub960_ind_update_bits(struct ub960_data * priv,u8 block,u8 reg,u8 mask,u8 val)990 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
991 u8 mask, u8 val)
992 {
993 struct device *dev = &priv->client->dev;
994 int ret;
995
996 mutex_lock(&priv->reg_lock);
997
998 ret = ub960_select_ind_reg_block(priv, block);
999 if (ret)
1000 goto out_unlock;
1001
1002 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
1003 if (ret) {
1004 dev_err(dev,
1005 "Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1006 block, reg, ret);
1007 goto out_unlock;
1008 }
1009
1010 ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1011 val);
1012 if (ret) {
1013 dev_err(dev,
1014 "Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1015 block, reg, ret);
1016 goto out_unlock;
1017 }
1018
1019 out_unlock:
1020 mutex_unlock(&priv->reg_lock);
1021
1022 return ret;
1023 }
1024
1025 /* -----------------------------------------------------------------------------
1026 * I2C-ATR (address translator)
1027 */
1028
ub960_atr_attach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client,u16 alias)1029 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1030 const struct i2c_client *client, u16 alias)
1031 {
1032 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1033 struct ub960_rxport *rxport = priv->rxports[chan_id];
1034 struct device *dev = &priv->client->dev;
1035 unsigned int reg_idx;
1036
1037 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1038 if (!rxport->aliased_clients[reg_idx])
1039 break;
1040 }
1041
1042 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1043 dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1044 return -EADDRNOTAVAIL;
1045 }
1046
1047 rxport->aliased_clients[reg_idx] = client;
1048
1049 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1050 client->addr << 1);
1051 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1052 alias << 1);
1053
1054 dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1055 rxport->nport, client->addr, alias, reg_idx);
1056
1057 return 0;
1058 }
1059
ub960_atr_detach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client)1060 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1061 const struct i2c_client *client)
1062 {
1063 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1064 struct ub960_rxport *rxport = priv->rxports[chan_id];
1065 struct device *dev = &priv->client->dev;
1066 unsigned int reg_idx;
1067
1068 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1069 if (rxport->aliased_clients[reg_idx] == client)
1070 break;
1071 }
1072
1073 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1074 dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1075 rxport->nport, client->addr);
1076 return;
1077 }
1078
1079 rxport->aliased_clients[reg_idx] = NULL;
1080
1081 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1082
1083 dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1084 client->addr, reg_idx);
1085 }
1086
1087 static const struct i2c_atr_ops ub960_atr_ops = {
1088 .attach_client = ub960_atr_attach_client,
1089 .detach_client = ub960_atr_detach_client,
1090 };
1091
ub960_init_atr(struct ub960_data * priv)1092 static int ub960_init_atr(struct ub960_data *priv)
1093 {
1094 struct device *dev = &priv->client->dev;
1095 struct i2c_adapter *parent_adap = priv->client->adapter;
1096
1097 priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1098 priv->hw_data->num_rxports);
1099 if (IS_ERR(priv->atr))
1100 return PTR_ERR(priv->atr);
1101
1102 i2c_atr_set_driver_data(priv->atr, priv);
1103
1104 return 0;
1105 }
1106
ub960_uninit_atr(struct ub960_data * priv)1107 static void ub960_uninit_atr(struct ub960_data *priv)
1108 {
1109 i2c_atr_delete(priv->atr);
1110 priv->atr = NULL;
1111 }
1112
1113 /* -----------------------------------------------------------------------------
1114 * TX ports
1115 */
1116
ub960_parse_dt_txport(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,u8 nport)1117 static int ub960_parse_dt_txport(struct ub960_data *priv,
1118 struct fwnode_handle *ep_fwnode,
1119 u8 nport)
1120 {
1121 struct device *dev = &priv->client->dev;
1122 struct v4l2_fwnode_endpoint vep = {};
1123 struct ub960_txport *txport;
1124 int ret;
1125
1126 txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1127 if (!txport)
1128 return -ENOMEM;
1129
1130 txport->priv = priv;
1131 txport->nport = nport;
1132
1133 vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1134 ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1135 if (ret) {
1136 dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1137 goto err_free_txport;
1138 }
1139
1140 txport->non_continous_clk = vep.bus.mipi_csi2.flags &
1141 V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK;
1142
1143 txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1144
1145 if (vep.nr_of_link_frequencies != 1) {
1146 ret = -EINVAL;
1147 goto err_free_vep;
1148 }
1149
1150 priv->tx_link_freq[0] = vep.link_frequencies[0];
1151 priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1152
1153 if (priv->tx_data_rate != MHZ(1600) &&
1154 priv->tx_data_rate != MHZ(1200) &&
1155 priv->tx_data_rate != MHZ(800) &&
1156 priv->tx_data_rate != MHZ(400)) {
1157 dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1158 ret = -EINVAL;
1159 goto err_free_vep;
1160 }
1161
1162 v4l2_fwnode_endpoint_free(&vep);
1163
1164 priv->txports[nport] = txport;
1165
1166 return 0;
1167
1168 err_free_vep:
1169 v4l2_fwnode_endpoint_free(&vep);
1170 err_free_txport:
1171 kfree(txport);
1172
1173 return ret;
1174 }
1175
ub960_csi_handle_events(struct ub960_data * priv,u8 nport)1176 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1177 {
1178 struct device *dev = &priv->client->dev;
1179 u8 csi_tx_isr;
1180 int ret;
1181
1182 ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1183 if (ret)
1184 return;
1185
1186 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1187 dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1188
1189 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1190 dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1191 }
1192
1193 /* -----------------------------------------------------------------------------
1194 * RX ports
1195 */
1196
ub960_rxport_enable_vpocs(struct ub960_data * priv)1197 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1198 {
1199 unsigned int nport;
1200 int ret;
1201
1202 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1203 struct ub960_rxport *rxport = priv->rxports[nport];
1204
1205 if (!rxport || !rxport->vpoc)
1206 continue;
1207
1208 ret = regulator_enable(rxport->vpoc);
1209 if (ret)
1210 goto err_disable_vpocs;
1211 }
1212
1213 return 0;
1214
1215 err_disable_vpocs:
1216 while (nport--) {
1217 struct ub960_rxport *rxport = priv->rxports[nport];
1218
1219 if (!rxport || !rxport->vpoc)
1220 continue;
1221
1222 regulator_disable(rxport->vpoc);
1223 }
1224
1225 return ret;
1226 }
1227
ub960_rxport_disable_vpocs(struct ub960_data * priv)1228 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1229 {
1230 unsigned int nport;
1231
1232 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1233 struct ub960_rxport *rxport = priv->rxports[nport];
1234
1235 if (!rxport || !rxport->vpoc)
1236 continue;
1237
1238 regulator_disable(rxport->vpoc);
1239 }
1240 }
1241
ub960_rxport_clear_errors(struct ub960_data * priv,unsigned int nport)1242 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1243 unsigned int nport)
1244 {
1245 u8 v;
1246
1247 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1248 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1249 ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1250 ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1251
1252 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1253 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1254
1255 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1256 }
1257
ub960_clear_rx_errors(struct ub960_data * priv)1258 static void ub960_clear_rx_errors(struct ub960_data *priv)
1259 {
1260 unsigned int nport;
1261
1262 for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1263 ub960_rxport_clear_errors(priv, nport);
1264 }
1265
ub960_rxport_get_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 * strobe_pos)1266 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1267 unsigned int nport, s8 *strobe_pos)
1268 {
1269 u8 v;
1270 u8 clk_delay, data_delay;
1271 int ret;
1272
1273 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1274 UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1275
1276 clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1277 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1278
1279 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1280 UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1281
1282 data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1283 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1284
1285 ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1286 if (ret)
1287 return ret;
1288
1289 clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1290
1291 ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1292 if (ret)
1293 return ret;
1294
1295 data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1296
1297 *strobe_pos = data_delay - clk_delay;
1298
1299 return 0;
1300 }
1301
ub960_rxport_set_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 strobe_pos)1302 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1303 unsigned int nport, s8 strobe_pos)
1304 {
1305 u8 clk_delay, data_delay;
1306
1307 clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1308 data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1309
1310 if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1311 clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1312 else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1313 data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1314 else if (strobe_pos < 0)
1315 clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1316 else if (strobe_pos > 0)
1317 data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1318
1319 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1320 UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1321
1322 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1323 UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1324 }
1325
ub960_rxport_set_strobe_range(struct ub960_data * priv,s8 strobe_min,s8 strobe_max)1326 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1327 s8 strobe_min, s8 strobe_max)
1328 {
1329 /* Convert the signed strobe pos to positive zero based value */
1330 strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1331 strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1332
1333 ub960_write(priv, UB960_XR_SFILTER_CFG,
1334 ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1335 ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1336 }
1337
ub960_rxport_get_eq_level(struct ub960_data * priv,unsigned int nport,u8 * eq_level)1338 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1339 unsigned int nport, u8 *eq_level)
1340 {
1341 int ret;
1342 u8 v;
1343
1344 ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1345 if (ret)
1346 return ret;
1347
1348 *eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1349 (v & UB960_RR_AEQ_STATUS_STATUS_2);
1350
1351 return 0;
1352 }
1353
ub960_rxport_set_eq_level(struct ub960_data * priv,unsigned int nport,u8 eq_level)1354 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1355 unsigned int nport, u8 eq_level)
1356 {
1357 u8 eq_stage_1_select_value, eq_stage_2_select_value;
1358 const unsigned int eq_stage_max = 7;
1359 u8 v;
1360
1361 if (eq_level <= eq_stage_max) {
1362 eq_stage_1_select_value = eq_level;
1363 eq_stage_2_select_value = 0;
1364 } else {
1365 eq_stage_1_select_value = eq_stage_max;
1366 eq_stage_2_select_value = eq_level - eq_stage_max;
1367 }
1368
1369 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1370
1371 v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1372 UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1373 v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1374 v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1375 v |= UB960_RR_AEQ_BYPASS_ENABLE;
1376
1377 ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1378 }
1379
ub960_rxport_set_eq_range(struct ub960_data * priv,unsigned int nport,u8 eq_min,u8 eq_max)1380 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1381 unsigned int nport, u8 eq_min, u8 eq_max)
1382 {
1383 ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1384 (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1385 (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1386
1387 /* Enable AEQ min setting */
1388 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1389 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1390 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1391 }
1392
ub960_rxport_config_eq(struct ub960_data * priv,unsigned int nport)1393 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1394 {
1395 struct ub960_rxport *rxport = priv->rxports[nport];
1396
1397 /* We also set common settings here. Should be moved elsewhere. */
1398
1399 if (priv->strobe.manual) {
1400 /* Disable AEQ_SFILTER_EN */
1401 ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1402 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1403 } else {
1404 /* Enable SFILTER and error control */
1405 ub960_write(priv, UB960_XR_AEQ_CTL1,
1406 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1407 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1408
1409 /* Set AEQ strobe range */
1410 ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1411 priv->strobe.max);
1412 }
1413
1414 /* The rest are port specific */
1415
1416 if (priv->strobe.manual)
1417 ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1418 else
1419 ub960_rxport_set_strobe_pos(priv, nport, 0);
1420
1421 if (rxport->eq.manual_eq) {
1422 ub960_rxport_set_eq_level(priv, nport,
1423 rxport->eq.manual.eq_level);
1424
1425 /* Enable AEQ Bypass */
1426 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1427 UB960_RR_AEQ_BYPASS_ENABLE,
1428 UB960_RR_AEQ_BYPASS_ENABLE);
1429 } else {
1430 ub960_rxport_set_eq_range(priv, nport,
1431 rxport->eq.aeq.eq_level_min,
1432 rxport->eq.aeq.eq_level_max);
1433
1434 /* Disable AEQ Bypass */
1435 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1436 UB960_RR_AEQ_BYPASS_ENABLE, 0);
1437 }
1438 }
1439
ub960_rxport_link_ok(struct ub960_data * priv,unsigned int nport,bool * ok)1440 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1441 bool *ok)
1442 {
1443 u8 rx_port_sts1, rx_port_sts2;
1444 u16 parity_errors;
1445 u8 csi_rx_sts;
1446 u8 csi_err_cnt;
1447 u8 bcc_sts;
1448 int ret;
1449 bool errors;
1450
1451 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1452 &rx_port_sts1);
1453 if (ret)
1454 return ret;
1455
1456 if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1457 *ok = false;
1458 return 0;
1459 }
1460
1461 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1462 &rx_port_sts2);
1463 if (ret)
1464 return ret;
1465
1466 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1467 if (ret)
1468 return ret;
1469
1470 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1471 &csi_err_cnt);
1472 if (ret)
1473 return ret;
1474
1475 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1476 if (ret)
1477 return ret;
1478
1479 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1480 &parity_errors);
1481 if (ret)
1482 return ret;
1483
1484 errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1485 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1486 (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1487 (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1488 parity_errors;
1489
1490 *ok = !errors;
1491
1492 return 0;
1493 }
1494
1495 /*
1496 * Wait for the RX ports to lock, have no errors and have stable strobe position
1497 * and EQ level.
1498 */
ub960_rxport_wait_locks(struct ub960_data * priv,unsigned long port_mask,unsigned int * lock_mask)1499 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1500 unsigned long port_mask,
1501 unsigned int *lock_mask)
1502 {
1503 struct device *dev = &priv->client->dev;
1504 unsigned long timeout;
1505 unsigned int link_ok_mask;
1506 unsigned int missing;
1507 unsigned int loops;
1508 u8 nport;
1509 int ret;
1510
1511 if (port_mask == 0) {
1512 if (lock_mask)
1513 *lock_mask = 0;
1514 return 0;
1515 }
1516
1517 if (port_mask >= BIT(priv->hw_data->num_rxports))
1518 return -EINVAL;
1519
1520 timeout = jiffies + msecs_to_jiffies(1000);
1521 loops = 0;
1522 link_ok_mask = 0;
1523
1524 while (time_before(jiffies, timeout)) {
1525 missing = 0;
1526
1527 for_each_set_bit(nport, &port_mask,
1528 priv->hw_data->num_rxports) {
1529 struct ub960_rxport *rxport = priv->rxports[nport];
1530 bool ok;
1531
1532 if (!rxport)
1533 continue;
1534
1535 ret = ub960_rxport_link_ok(priv, nport, &ok);
1536 if (ret)
1537 return ret;
1538
1539 /*
1540 * We want the link to be ok for two consecutive loops,
1541 * as a link could get established just before our test
1542 * and drop soon after.
1543 */
1544 if (!ok || !(link_ok_mask & BIT(nport)))
1545 missing++;
1546
1547 if (ok)
1548 link_ok_mask |= BIT(nport);
1549 else
1550 link_ok_mask &= ~BIT(nport);
1551 }
1552
1553 loops++;
1554
1555 if (missing == 0)
1556 break;
1557
1558 msleep(50);
1559 }
1560
1561 if (lock_mask)
1562 *lock_mask = link_ok_mask;
1563
1564 dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1565 for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1566 struct ub960_rxport *rxport = priv->rxports[nport];
1567 s8 strobe_pos, eq_level;
1568 u16 v;
1569
1570 if (!rxport)
1571 continue;
1572
1573 if (!(link_ok_mask & BIT(nport))) {
1574 dev_dbg(dev, "\trx%u: not locked\n", nport);
1575 continue;
1576 }
1577
1578 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1579
1580 if (priv->hw_data->is_ub9702) {
1581 dev_dbg(dev, "\trx%u: locked, freq %llu Hz\n",
1582 nport, (v * 1000000ULL) >> 8);
1583 } else {
1584 ret = ub960_rxport_get_strobe_pos(priv, nport,
1585 &strobe_pos);
1586 if (ret)
1587 return ret;
1588
1589 ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1590 if (ret)
1591 return ret;
1592
1593 dev_dbg(dev,
1594 "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1595 nport, strobe_pos, eq_level,
1596 (v * 1000000ULL) >> 8);
1597 }
1598 }
1599
1600 return 0;
1601 }
1602
ub960_calc_bc_clk_rate_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1603 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1604 struct ub960_rxport *rxport)
1605 {
1606 unsigned int mult;
1607 unsigned int div;
1608
1609 switch (rxport->rx_mode) {
1610 case RXPORT_MODE_RAW10:
1611 case RXPORT_MODE_RAW12_HF:
1612 case RXPORT_MODE_RAW12_LF:
1613 mult = 1;
1614 div = 10;
1615 break;
1616
1617 case RXPORT_MODE_CSI2_SYNC:
1618 mult = 2;
1619 div = 1;
1620 break;
1621
1622 case RXPORT_MODE_CSI2_NONSYNC:
1623 mult = 2;
1624 div = 5;
1625 break;
1626
1627 default:
1628 return 0;
1629 }
1630
1631 return clk_get_rate(priv->refclk) * mult / div;
1632 }
1633
ub960_calc_bc_clk_rate_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)1634 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1635 struct ub960_rxport *rxport)
1636 {
1637 switch (rxport->rx_mode) {
1638 case RXPORT_MODE_RAW10:
1639 case RXPORT_MODE_RAW12_HF:
1640 case RXPORT_MODE_RAW12_LF:
1641 return 2359400;
1642
1643 case RXPORT_MODE_CSI2_SYNC:
1644 return 47187500;
1645
1646 case RXPORT_MODE_CSI2_NONSYNC:
1647 return 9437500;
1648
1649 default:
1650 return 0;
1651 }
1652 }
1653
ub960_rxport_add_serializer(struct ub960_data * priv,u8 nport)1654 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1655 {
1656 struct ub960_rxport *rxport = priv->rxports[nport];
1657 struct device *dev = &priv->client->dev;
1658 struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1659 struct i2c_board_info ser_info = {
1660 .of_node = to_of_node(rxport->ser.fwnode),
1661 .fwnode = rxport->ser.fwnode,
1662 .platform_data = ser_pdata,
1663 };
1664
1665 ser_pdata->port = nport;
1666 ser_pdata->atr = priv->atr;
1667 if (priv->hw_data->is_ub9702)
1668 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1669 else
1670 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1671
1672 /*
1673 * The serializer is added under the same i2c adapter as the
1674 * deserializer. This is not quite right, as the serializer is behind
1675 * the FPD-Link.
1676 */
1677 ser_info.addr = rxport->ser.alias;
1678 rxport->ser.client =
1679 i2c_new_client_device(priv->client->adapter, &ser_info);
1680 if (IS_ERR(rxport->ser.client)) {
1681 dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1682 ser_info.type);
1683 return PTR_ERR(rxport->ser.client);
1684 }
1685
1686 dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1687 nport, rxport->ser.client->addr,
1688 rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1689
1690 return 0;
1691 }
1692
ub960_rxport_remove_serializer(struct ub960_data * priv,u8 nport)1693 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1694 {
1695 struct ub960_rxport *rxport = priv->rxports[nport];
1696
1697 i2c_unregister_device(rxport->ser.client);
1698 rxport->ser.client = NULL;
1699 }
1700
1701 /* Add serializer i2c devices for all initialized ports */
ub960_rxport_add_serializers(struct ub960_data * priv)1702 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1703 {
1704 unsigned int nport;
1705 int ret;
1706
1707 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1708 struct ub960_rxport *rxport = priv->rxports[nport];
1709
1710 if (!rxport)
1711 continue;
1712
1713 ret = ub960_rxport_add_serializer(priv, nport);
1714 if (ret)
1715 goto err_remove_sers;
1716 }
1717
1718 return 0;
1719
1720 err_remove_sers:
1721 while (nport--) {
1722 struct ub960_rxport *rxport = priv->rxports[nport];
1723
1724 if (!rxport)
1725 continue;
1726
1727 ub960_rxport_remove_serializer(priv, nport);
1728 }
1729
1730 return ret;
1731 }
1732
ub960_rxport_remove_serializers(struct ub960_data * priv)1733 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1734 {
1735 unsigned int nport;
1736
1737 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1738 struct ub960_rxport *rxport = priv->rxports[nport];
1739
1740 if (!rxport)
1741 continue;
1742
1743 ub960_rxport_remove_serializer(priv, nport);
1744 }
1745 }
1746
ub960_init_tx_port(struct ub960_data * priv,struct ub960_txport * txport)1747 static void ub960_init_tx_port(struct ub960_data *priv,
1748 struct ub960_txport *txport)
1749 {
1750 unsigned int nport = txport->nport;
1751 u8 csi_ctl = 0;
1752
1753 /*
1754 * From the datasheet: "initial CSI Skew-Calibration
1755 * sequence [...] should be set when operating at 1.6 Gbps"
1756 */
1757 if (priv->tx_data_rate == MHZ(1600))
1758 csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1759
1760 csi_ctl |= (4 - txport->num_data_lanes) << 4;
1761
1762 if (!txport->non_continous_clk)
1763 csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK;
1764
1765 ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1766 }
1767
ub960_init_tx_ports(struct ub960_data * priv)1768 static int ub960_init_tx_ports(struct ub960_data *priv)
1769 {
1770 unsigned int nport;
1771 u8 speed_select;
1772 u8 pll_div;
1773
1774 /* TX ports */
1775
1776 switch (priv->tx_data_rate) {
1777 case MHZ(1600):
1778 default:
1779 speed_select = 0;
1780 pll_div = 0x10;
1781 break;
1782 case MHZ(1200):
1783 speed_select = 1;
1784 pll_div = 0x18;
1785 break;
1786 case MHZ(800):
1787 speed_select = 2;
1788 pll_div = 0x10;
1789 break;
1790 case MHZ(400):
1791 speed_select = 3;
1792 pll_div = 0x10;
1793 break;
1794 }
1795
1796 ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1797
1798 if (priv->hw_data->is_ub9702) {
1799 ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1800
1801 switch (priv->tx_data_rate) {
1802 case MHZ(1600):
1803 default:
1804 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1805 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1806 break;
1807 case MHZ(800):
1808 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1809 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1810 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1811 break;
1812 case MHZ(400):
1813 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1814 break;
1815 }
1816 }
1817
1818 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1819 struct ub960_txport *txport = priv->txports[nport];
1820
1821 if (!txport)
1822 continue;
1823
1824 ub960_init_tx_port(priv, txport);
1825 }
1826
1827 return 0;
1828 }
1829
ub960_init_rx_port_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1830 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1831 struct ub960_rxport *rxport)
1832 {
1833 unsigned int nport = rxport->nport;
1834 u32 bc_freq_val;
1835
1836 /*
1837 * Back channel frequency select.
1838 * Override FREQ_SELECT from the strap.
1839 * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1840 * 2 - 10 Mbps
1841 * 6 - 50 Mbps (DS90UB953-Q1)
1842 *
1843 * Note that changing this setting will result in some errors on the back
1844 * channel for a short period of time.
1845 */
1846
1847 switch (rxport->rx_mode) {
1848 case RXPORT_MODE_RAW10:
1849 case RXPORT_MODE_RAW12_HF:
1850 case RXPORT_MODE_RAW12_LF:
1851 bc_freq_val = 0;
1852 break;
1853
1854 case RXPORT_MODE_CSI2_NONSYNC:
1855 bc_freq_val = 2;
1856 break;
1857
1858 case RXPORT_MODE_CSI2_SYNC:
1859 bc_freq_val = 6;
1860 break;
1861
1862 default:
1863 return;
1864 }
1865
1866 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1867 UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1868 bc_freq_val);
1869
1870 switch (rxport->rx_mode) {
1871 case RXPORT_MODE_RAW10:
1872 /* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1873 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1874 UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1875 0x3);
1876
1877 /*
1878 * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1879 */
1880 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1881 UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1882 0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1883
1884 break;
1885
1886 case RXPORT_MODE_RAW12_HF:
1887 case RXPORT_MODE_RAW12_LF:
1888 /* Not implemented */
1889 return;
1890
1891 case RXPORT_MODE_CSI2_SYNC:
1892 case RXPORT_MODE_CSI2_NONSYNC:
1893 /* CSI-2 Mode (DS90UB953-Q1 compatible) */
1894 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1895 0x0);
1896
1897 break;
1898 }
1899
1900 /* LV_POLARITY & FV_POLARITY */
1901 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1902 rxport->lv_fv_pol);
1903
1904 /* Enable all interrupt sources from this port */
1905 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1906 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1907
1908 /* Enable I2C_PASS_THROUGH */
1909 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1910 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1911 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1912
1913 /* Enable I2C communication to the serializer via the alias addr */
1914 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1915 rxport->ser.alias << 1);
1916
1917 /* Configure EQ related settings */
1918 ub960_rxport_config_eq(priv, nport);
1919
1920 /* Enable RX port */
1921 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1922 }
1923
ub960_init_rx_port_ub9702_fpd3(struct ub960_data * priv,struct ub960_rxport * rxport)1924 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1925 struct ub960_rxport *rxport)
1926 {
1927 unsigned int nport = rxport->nport;
1928 u8 bc_freq_val;
1929 u8 fpd_func_mode;
1930
1931 switch (rxport->rx_mode) {
1932 case RXPORT_MODE_RAW10:
1933 bc_freq_val = 0;
1934 fpd_func_mode = 5;
1935 break;
1936
1937 case RXPORT_MODE_RAW12_HF:
1938 bc_freq_val = 0;
1939 fpd_func_mode = 4;
1940 break;
1941
1942 case RXPORT_MODE_RAW12_LF:
1943 bc_freq_val = 0;
1944 fpd_func_mode = 6;
1945 break;
1946
1947 case RXPORT_MODE_CSI2_SYNC:
1948 bc_freq_val = 6;
1949 fpd_func_mode = 2;
1950 break;
1951
1952 case RXPORT_MODE_CSI2_NONSYNC:
1953 bc_freq_val = 2;
1954 fpd_func_mode = 2;
1955 break;
1956
1957 default:
1958 return;
1959 }
1960
1961 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1962 bc_freq_val);
1963 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1964
1965 /* set serdes_eq_mode = 1 */
1966 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1967
1968 /* enable serdes driver */
1969 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1970
1971 /* set serdes_eq_offset=4 */
1972 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1973
1974 /* init default serdes_eq_max in 0xa9 */
1975 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
1976
1977 /* init serdes_eq_min in 0xaa */
1978 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
1979
1980 /* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
1981 ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
1982 BIT(3), BIT(3));
1983
1984 /* RX port to half-rate */
1985 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
1986 BIT(nport * 2));
1987 }
1988
ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data * priv,struct ub960_rxport * rxport)1989 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
1990 struct ub960_rxport *rxport)
1991 {
1992 unsigned int nport = rxport->nport;
1993 bool first_time_power_up = true;
1994
1995 if (first_time_power_up) {
1996 u8 v;
1997
1998 /* AEQ init */
1999 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
2000
2001 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
2002 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
2003
2004 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
2005 }
2006
2007 /* enable serdes_eq_ctl2 */
2008 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
2009
2010 /* enable serdes_eq_ctl1 */
2011 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
2012
2013 /* enable serdes_eq_en */
2014 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
2015
2016 /* disable serdes_eq_override */
2017 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
2018
2019 /* disable serdes_gain_override */
2020 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2021 }
2022
ub960_init_rx_port_ub9702_fpd4(struct ub960_data * priv,struct ub960_rxport * rxport)2023 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2024 struct ub960_rxport *rxport)
2025 {
2026 unsigned int nport = rxport->nport;
2027 u8 bc_freq_val;
2028
2029 switch (rxport->rx_mode) {
2030 case RXPORT_MODE_RAW10:
2031 bc_freq_val = 0;
2032 break;
2033
2034 case RXPORT_MODE_RAW12_HF:
2035 bc_freq_val = 0;
2036 break;
2037
2038 case RXPORT_MODE_RAW12_LF:
2039 bc_freq_val = 0;
2040 break;
2041
2042 case RXPORT_MODE_CSI2_SYNC:
2043 bc_freq_val = 6;
2044 break;
2045
2046 case RXPORT_MODE_CSI2_NONSYNC:
2047 bc_freq_val = 2;
2048 break;
2049
2050 default:
2051 return;
2052 }
2053
2054 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2055 bc_freq_val);
2056
2057 /* FPD4 Sync Mode */
2058 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2059
2060 /* add serdes_eq_offset of 4 */
2061 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2062
2063 /* FPD4 serdes_start_eq in 0x27: assign default */
2064 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2065 /* FPD4 serdes_end_eq in 0x28: assign default */
2066 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2067
2068 /* set serdes_driver_mode into FPD IV mode */
2069 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2070 /* set FPD PBC drv into FPD IV mode */
2071 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2072
2073 /* set serdes_system_init to 0x2f */
2074 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2075 /* set serdes_system_rst in reset mode */
2076 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2077
2078 /* RX port to 7.55G mode */
2079 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2080 0 << (nport * 2));
2081
2082 ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2083 }
2084
ub960_init_rx_port_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)2085 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2086 struct ub960_rxport *rxport)
2087 {
2088 unsigned int nport = rxport->nport;
2089
2090 if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2091 ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2092 else /* RXPORT_CDR_FPD4 */
2093 ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2094
2095 switch (rxport->rx_mode) {
2096 case RXPORT_MODE_RAW10:
2097 /*
2098 * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2099 * 0b10 : 8-bit processing using upper 8 bits
2100 */
2101 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2102 0x3 << 6, 0x2 << 6);
2103
2104 break;
2105
2106 case RXPORT_MODE_RAW12_HF:
2107 case RXPORT_MODE_RAW12_LF:
2108 /* Not implemented */
2109 return;
2110
2111 case RXPORT_MODE_CSI2_SYNC:
2112 case RXPORT_MODE_CSI2_NONSYNC:
2113
2114 break;
2115 }
2116
2117 /* LV_POLARITY & FV_POLARITY */
2118 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2119 rxport->lv_fv_pol);
2120
2121 /* Enable all interrupt sources from this port */
2122 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2123 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2124
2125 /* Enable I2C_PASS_THROUGH */
2126 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2127 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2128 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2129
2130 /* Enable I2C communication to the serializer via the alias addr */
2131 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2132 rxport->ser.alias << 1);
2133
2134 /* Enable RX port */
2135 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2136
2137 if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2138 /* unreset 960 AEQ */
2139 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2140 }
2141 }
2142
ub960_init_rx_ports(struct ub960_data * priv)2143 static int ub960_init_rx_ports(struct ub960_data *priv)
2144 {
2145 unsigned int nport;
2146
2147 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2148 struct ub960_rxport *rxport = priv->rxports[nport];
2149
2150 if (!rxport)
2151 continue;
2152
2153 if (priv->hw_data->is_ub9702)
2154 ub960_init_rx_port_ub9702(priv, rxport);
2155 else
2156 ub960_init_rx_port_ub960(priv, rxport);
2157 }
2158
2159 return 0;
2160 }
2161
ub960_rxport_handle_events(struct ub960_data * priv,u8 nport)2162 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2163 {
2164 struct device *dev = &priv->client->dev;
2165 u8 rx_port_sts1;
2166 u8 rx_port_sts2;
2167 u8 csi_rx_sts;
2168 u8 bcc_sts;
2169 int ret = 0;
2170
2171 /* Read interrupts (also clears most of them) */
2172 if (!ret)
2173 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2174 &rx_port_sts1);
2175 if (!ret)
2176 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2177 &rx_port_sts2);
2178 if (!ret)
2179 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2180 &csi_rx_sts);
2181 if (!ret)
2182 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2183 &bcc_sts);
2184
2185 if (ret)
2186 return;
2187
2188 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2189 u16 v;
2190
2191 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2192 &v);
2193 if (!ret)
2194 dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2195 }
2196
2197 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2198 dev_err(dev, "rx%u BCC CRC error\n", nport);
2199
2200 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2201 dev_err(dev, "rx%u BCC SEQ error\n", nport);
2202
2203 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2204 dev_err(dev, "rx%u line length unstable\n", nport);
2205
2206 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2207 dev_err(dev, "rx%u FPD3 encode error\n", nport);
2208
2209 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2210 dev_err(dev, "rx%u buffer error\n", nport);
2211
2212 if (csi_rx_sts)
2213 dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2214
2215 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2216 dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2217
2218 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2219 dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2220
2221 if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2222 dev_err(dev, "rx%u CSI checksum error\n", nport);
2223
2224 if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2225 dev_err(dev, "rx%u CSI length error\n", nport);
2226
2227 if (bcc_sts)
2228 dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2229
2230 if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2231 dev_err(dev, "rx%u BCC response error", nport);
2232
2233 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2234 dev_err(dev, "rx%u BCC slave timeout", nport);
2235
2236 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2237 dev_err(dev, "rx%u BCC slave error", nport);
2238
2239 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2240 dev_err(dev, "rx%u BCC master timeout", nport);
2241
2242 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2243 dev_err(dev, "rx%u BCC master error", nport);
2244
2245 if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2246 dev_err(dev, "rx%u BCC sequence error", nport);
2247
2248 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2249 u16 v;
2250
2251 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2252 if (!ret)
2253 dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2254 }
2255
2256 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2257 u16 v;
2258
2259 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2260 &v);
2261 if (!ret)
2262 dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2263 }
2264
2265 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2266 dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2267 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2268 "locked" :
2269 "unlocked",
2270 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2271 "passed" :
2272 "not passed",
2273 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2274 "no clock" :
2275 "clock ok",
2276 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2277 "stable freq" :
2278 "unstable freq");
2279 }
2280 }
2281
2282 /* -----------------------------------------------------------------------------
2283 * V4L2
2284 */
2285
2286 /*
2287 * The current implementation only supports a simple VC mapping, where all VCs
2288 * from a one RX port will be mapped to the same VC. Also, the hardware
2289 * dictates that all streams from an RX port must go to a single TX port.
2290 *
2291 * This function decides the target VC numbers for each RX port with a simple
2292 * algorithm, so that for each TX port, we get VC numbers starting from 0,
2293 * and counting up.
2294 *
2295 * E.g. if all four RX ports are in use, of which the first two go to the
2296 * first TX port and the secont two go to the second TX port, we would get
2297 * the following VCs for the four RX ports: 0, 1, 0, 1.
2298 *
2299 * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2300 * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2301 * probably needs to be somehow configurable. Device tree?
2302 */
ub960_get_vc_maps(struct ub960_data * priv,struct v4l2_subdev_state * state,u8 * vc)2303 static void ub960_get_vc_maps(struct ub960_data *priv,
2304 struct v4l2_subdev_state *state, u8 *vc)
2305 {
2306 u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2307 struct v4l2_subdev_route *route;
2308 u8 handled_mask = 0;
2309
2310 for_each_active_route(&state->routing, route) {
2311 unsigned int rx, tx;
2312
2313 rx = ub960_pad_to_port(priv, route->sink_pad);
2314 if (BIT(rx) & handled_mask)
2315 continue;
2316
2317 tx = ub960_pad_to_port(priv, route->source_pad);
2318
2319 vc[rx] = cur_vc[tx]++;
2320 handled_mask |= BIT(rx);
2321 }
2322 }
2323
ub960_enable_tx_port(struct ub960_data * priv,unsigned int nport)2324 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2325 {
2326 struct device *dev = &priv->client->dev;
2327
2328 dev_dbg(dev, "enable TX port %u\n", nport);
2329
2330 return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2331 UB960_TR_CSI_CTL_CSI_ENABLE,
2332 UB960_TR_CSI_CTL_CSI_ENABLE);
2333 }
2334
ub960_disable_tx_port(struct ub960_data * priv,unsigned int nport)2335 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2336 {
2337 struct device *dev = &priv->client->dev;
2338
2339 dev_dbg(dev, "disable TX port %u\n", nport);
2340
2341 ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2342 UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2343 }
2344
ub960_enable_rx_port(struct ub960_data * priv,unsigned int nport)2345 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2346 {
2347 struct device *dev = &priv->client->dev;
2348
2349 dev_dbg(dev, "enable RX port %u\n", nport);
2350
2351 /* Enable forwarding */
2352 return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2353 UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2354 }
2355
ub960_disable_rx_port(struct ub960_data * priv,unsigned int nport)2356 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2357 {
2358 struct device *dev = &priv->client->dev;
2359
2360 dev_dbg(dev, "disable RX port %u\n", nport);
2361
2362 /* Disable forwarding */
2363 ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2364 UB960_SR_FWD_CTL1_PORT_DIS(nport),
2365 UB960_SR_FWD_CTL1_PORT_DIS(nport));
2366 }
2367
2368 /*
2369 * The driver only supports using a single VC for each source. This function
2370 * checks that each source only provides streams using a single VC.
2371 */
ub960_validate_stream_vcs(struct ub960_data * priv)2372 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2373 {
2374 unsigned int nport;
2375 unsigned int i;
2376
2377 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2378 struct ub960_rxport *rxport = priv->rxports[nport];
2379 struct v4l2_mbus_frame_desc desc;
2380 int ret;
2381 u8 vc;
2382
2383 if (!rxport)
2384 continue;
2385
2386 ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2387 rxport->source.pad, &desc);
2388 if (ret)
2389 return ret;
2390
2391 if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2392 continue;
2393
2394 if (desc.num_entries == 0)
2395 continue;
2396
2397 vc = desc.entry[0].bus.csi2.vc;
2398
2399 for (i = 1; i < desc.num_entries; i++) {
2400 if (vc == desc.entry[i].bus.csi2.vc)
2401 continue;
2402
2403 dev_err(&priv->client->dev,
2404 "rx%u: source with multiple virtual-channels is not supported\n",
2405 nport);
2406 return -ENODEV;
2407 }
2408 }
2409
2410 return 0;
2411 }
2412
ub960_configure_ports_for_streaming(struct ub960_data * priv,struct v4l2_subdev_state * state)2413 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2414 struct v4l2_subdev_state *state)
2415 {
2416 u8 fwd_ctl;
2417 struct {
2418 u32 num_streams;
2419 u8 pixel_dt;
2420 u8 meta_dt;
2421 u32 meta_lines;
2422 u32 tx_port;
2423 } rx_data[UB960_MAX_RX_NPORTS] = {};
2424 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2425 struct v4l2_subdev_route *route;
2426 unsigned int nport;
2427 int ret;
2428
2429 ret = ub960_validate_stream_vcs(priv);
2430 if (ret)
2431 return ret;
2432
2433 ub960_get_vc_maps(priv, state, vc_map);
2434
2435 for_each_active_route(&state->routing, route) {
2436 struct ub960_rxport *rxport;
2437 struct ub960_txport *txport;
2438 struct v4l2_mbus_framefmt *fmt;
2439 const struct ub960_format_info *ub960_fmt;
2440 unsigned int nport;
2441
2442 nport = ub960_pad_to_port(priv, route->sink_pad);
2443
2444 rxport = priv->rxports[nport];
2445 if (!rxport)
2446 return -EINVAL;
2447
2448 txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2449 if (!txport)
2450 return -EINVAL;
2451
2452 rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2453
2454 rx_data[nport].num_streams++;
2455
2456 /* For the rest, we are only interested in parallel busses */
2457 if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2458 rxport->rx_mode == RXPORT_MODE_CSI2_NONSYNC)
2459 continue;
2460
2461 if (rx_data[nport].num_streams > 2)
2462 return -EPIPE;
2463
2464 fmt = v4l2_subdev_state_get_format(state, route->sink_pad,
2465 route->sink_stream);
2466 if (!fmt)
2467 return -EPIPE;
2468
2469 ub960_fmt = ub960_find_format(fmt->code);
2470 if (!ub960_fmt)
2471 return -EPIPE;
2472
2473 if (ub960_fmt->meta) {
2474 if (fmt->height > 3) {
2475 dev_err(&priv->client->dev,
2476 "rx%u: unsupported metadata height %u\n",
2477 nport, fmt->height);
2478 return -EPIPE;
2479 }
2480
2481 rx_data[nport].meta_dt = ub960_fmt->datatype;
2482 rx_data[nport].meta_lines = fmt->height;
2483 } else {
2484 rx_data[nport].pixel_dt = ub960_fmt->datatype;
2485 }
2486 }
2487
2488 /* Configure RX ports */
2489
2490 /*
2491 * Keep all port forwardings disabled by default. Forwarding will be
2492 * enabled in ub960_enable_rx_port.
2493 */
2494 fwd_ctl = GENMASK(7, 4);
2495
2496 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2497 struct ub960_rxport *rxport = priv->rxports[nport];
2498 u8 vc = vc_map[nport];
2499
2500 if (rx_data[nport].num_streams == 0)
2501 continue;
2502
2503 switch (rxport->rx_mode) {
2504 case RXPORT_MODE_RAW10:
2505 ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2506 rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2507
2508 ub960_rxport_write(priv, rxport->nport,
2509 UB960_RR_RAW_EMBED_DTYPE,
2510 (rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2511 rx_data[nport].meta_dt);
2512
2513 break;
2514
2515 case RXPORT_MODE_RAW12_HF:
2516 case RXPORT_MODE_RAW12_LF:
2517 /* Not implemented */
2518 break;
2519
2520 case RXPORT_MODE_CSI2_SYNC:
2521 case RXPORT_MODE_CSI2_NONSYNC:
2522 if (!priv->hw_data->is_ub9702) {
2523 /* Map all VCs from this port to the same VC */
2524 ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2525 (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2526 (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2527 (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2528 (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2529 } else {
2530 unsigned int i;
2531
2532 /* Map all VCs from this port to VC(nport) */
2533 for (i = 0; i < 8; i++)
2534 ub960_rxport_write(priv, nport,
2535 UB960_RR_VC_ID_MAP(i),
2536 (nport << 4) | nport);
2537 }
2538
2539 break;
2540 }
2541
2542 if (rx_data[nport].tx_port == 1)
2543 fwd_ctl |= BIT(nport); /* forward to TX1 */
2544 else
2545 fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2546 }
2547
2548 ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2549
2550 return 0;
2551 }
2552
ub960_update_streaming_status(struct ub960_data * priv)2553 static void ub960_update_streaming_status(struct ub960_data *priv)
2554 {
2555 unsigned int i;
2556
2557 for (i = 0; i < UB960_MAX_NPORTS; i++) {
2558 if (priv->stream_enable_mask[i])
2559 break;
2560 }
2561
2562 priv->streaming = i < UB960_MAX_NPORTS;
2563 }
2564
ub960_enable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2565 static int ub960_enable_streams(struct v4l2_subdev *sd,
2566 struct v4l2_subdev_state *state, u32 source_pad,
2567 u64 source_streams_mask)
2568 {
2569 struct ub960_data *priv = sd_to_ub960(sd);
2570 struct device *dev = &priv->client->dev;
2571 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2572 struct v4l2_subdev_route *route;
2573 unsigned int failed_port;
2574 unsigned int nport;
2575 int ret;
2576
2577 if (!priv->streaming) {
2578 dev_dbg(dev, "Prepare for streaming\n");
2579 ret = ub960_configure_ports_for_streaming(priv, state);
2580 if (ret)
2581 return ret;
2582 }
2583
2584 /* Enable TX port if not yet enabled */
2585 if (!priv->stream_enable_mask[source_pad]) {
2586 ret = ub960_enable_tx_port(priv,
2587 ub960_pad_to_port(priv, source_pad));
2588 if (ret)
2589 return ret;
2590 }
2591
2592 priv->stream_enable_mask[source_pad] |= source_streams_mask;
2593
2594 /* Collect sink streams per pad which we need to enable */
2595 for_each_active_route(&state->routing, route) {
2596 if (route->source_pad != source_pad)
2597 continue;
2598
2599 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2600 continue;
2601
2602 nport = ub960_pad_to_port(priv, route->sink_pad);
2603
2604 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2605 }
2606
2607 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2608 if (!sink_streams[nport])
2609 continue;
2610
2611 /* Enable the RX port if not yet enabled */
2612 if (!priv->stream_enable_mask[nport]) {
2613 ret = ub960_enable_rx_port(priv, nport);
2614 if (ret) {
2615 failed_port = nport;
2616 goto err;
2617 }
2618 }
2619
2620 priv->stream_enable_mask[nport] |= sink_streams[nport];
2621
2622 dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2623 sink_streams[nport]);
2624
2625 ret = v4l2_subdev_enable_streams(
2626 priv->rxports[nport]->source.sd,
2627 priv->rxports[nport]->source.pad,
2628 sink_streams[nport]);
2629 if (ret) {
2630 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2631
2632 if (!priv->stream_enable_mask[nport])
2633 ub960_disable_rx_port(priv, nport);
2634
2635 failed_port = nport;
2636 goto err;
2637 }
2638 }
2639
2640 priv->streaming = true;
2641
2642 return 0;
2643
2644 err:
2645 for (nport = 0; nport < failed_port; nport++) {
2646 if (!sink_streams[nport])
2647 continue;
2648
2649 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2650 sink_streams[nport]);
2651
2652 ret = v4l2_subdev_disable_streams(
2653 priv->rxports[nport]->source.sd,
2654 priv->rxports[nport]->source.pad,
2655 sink_streams[nport]);
2656 if (ret)
2657 dev_err(dev, "Failed to disable streams: %d\n", ret);
2658
2659 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2660
2661 /* Disable RX port if no active streams */
2662 if (!priv->stream_enable_mask[nport])
2663 ub960_disable_rx_port(priv, nport);
2664 }
2665
2666 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2667
2668 if (!priv->stream_enable_mask[source_pad])
2669 ub960_disable_tx_port(priv,
2670 ub960_pad_to_port(priv, source_pad));
2671
2672 ub960_update_streaming_status(priv);
2673
2674 return ret;
2675 }
2676
ub960_disable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2677 static int ub960_disable_streams(struct v4l2_subdev *sd,
2678 struct v4l2_subdev_state *state,
2679 u32 source_pad, u64 source_streams_mask)
2680 {
2681 struct ub960_data *priv = sd_to_ub960(sd);
2682 struct device *dev = &priv->client->dev;
2683 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2684 struct v4l2_subdev_route *route;
2685 unsigned int nport;
2686 int ret;
2687
2688 /* Collect sink streams per pad which we need to disable */
2689 for_each_active_route(&state->routing, route) {
2690 if (route->source_pad != source_pad)
2691 continue;
2692
2693 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2694 continue;
2695
2696 nport = ub960_pad_to_port(priv, route->sink_pad);
2697
2698 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2699 }
2700
2701 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2702 if (!sink_streams[nport])
2703 continue;
2704
2705 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2706 sink_streams[nport]);
2707
2708 ret = v4l2_subdev_disable_streams(
2709 priv->rxports[nport]->source.sd,
2710 priv->rxports[nport]->source.pad,
2711 sink_streams[nport]);
2712 if (ret)
2713 dev_err(dev, "Failed to disable streams: %d\n", ret);
2714
2715 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2716
2717 /* Disable RX port if no active streams */
2718 if (!priv->stream_enable_mask[nport])
2719 ub960_disable_rx_port(priv, nport);
2720 }
2721
2722 /* Disable TX port if no active streams */
2723
2724 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2725
2726 if (!priv->stream_enable_mask[source_pad])
2727 ub960_disable_tx_port(priv,
2728 ub960_pad_to_port(priv, source_pad));
2729
2730 ub960_update_streaming_status(priv);
2731
2732 return 0;
2733 }
2734
_ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_krouting * routing)2735 static int _ub960_set_routing(struct v4l2_subdev *sd,
2736 struct v4l2_subdev_state *state,
2737 struct v4l2_subdev_krouting *routing)
2738 {
2739 static const struct v4l2_mbus_framefmt format = {
2740 .width = 640,
2741 .height = 480,
2742 .code = MEDIA_BUS_FMT_UYVY8_1X16,
2743 .field = V4L2_FIELD_NONE,
2744 .colorspace = V4L2_COLORSPACE_SRGB,
2745 .ycbcr_enc = V4L2_YCBCR_ENC_601,
2746 .quantization = V4L2_QUANTIZATION_LIM_RANGE,
2747 .xfer_func = V4L2_XFER_FUNC_SRGB,
2748 };
2749 int ret;
2750
2751 /*
2752 * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2753 * frame desc is made dynamically allocated.
2754 */
2755
2756 if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2757 return -E2BIG;
2758
2759 ret = v4l2_subdev_routing_validate(sd, routing,
2760 V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2761 V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2762 if (ret)
2763 return ret;
2764
2765 ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2766 if (ret)
2767 return ret;
2768
2769 return 0;
2770 }
2771
ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,enum v4l2_subdev_format_whence which,struct v4l2_subdev_krouting * routing)2772 static int ub960_set_routing(struct v4l2_subdev *sd,
2773 struct v4l2_subdev_state *state,
2774 enum v4l2_subdev_format_whence which,
2775 struct v4l2_subdev_krouting *routing)
2776 {
2777 struct ub960_data *priv = sd_to_ub960(sd);
2778
2779 if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2780 return -EBUSY;
2781
2782 return _ub960_set_routing(sd, state, routing);
2783 }
2784
ub960_get_frame_desc(struct v4l2_subdev * sd,unsigned int pad,struct v4l2_mbus_frame_desc * fd)2785 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2786 struct v4l2_mbus_frame_desc *fd)
2787 {
2788 struct ub960_data *priv = sd_to_ub960(sd);
2789 struct v4l2_subdev_route *route;
2790 struct v4l2_subdev_state *state;
2791 int ret = 0;
2792 struct device *dev = &priv->client->dev;
2793 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2794
2795 if (!ub960_pad_is_source(priv, pad))
2796 return -EINVAL;
2797
2798 fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2799
2800 state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2801
2802 ub960_get_vc_maps(priv, state, vc_map);
2803
2804 for_each_active_route(&state->routing, route) {
2805 struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2806 struct v4l2_mbus_frame_desc source_fd;
2807 unsigned int nport;
2808 unsigned int i;
2809
2810 if (route->source_pad != pad)
2811 continue;
2812
2813 nport = ub960_pad_to_port(priv, route->sink_pad);
2814
2815 ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2816 get_frame_desc,
2817 priv->rxports[nport]->source.pad,
2818 &source_fd);
2819 if (ret) {
2820 dev_err(dev,
2821 "Failed to get source frame desc for pad %u\n",
2822 route->sink_pad);
2823 goto out_unlock;
2824 }
2825
2826 for (i = 0; i < source_fd.num_entries; i++) {
2827 if (source_fd.entry[i].stream == route->sink_stream) {
2828 source_entry = &source_fd.entry[i];
2829 break;
2830 }
2831 }
2832
2833 if (!source_entry) {
2834 dev_err(dev,
2835 "Failed to find stream from source frame desc\n");
2836 ret = -EPIPE;
2837 goto out_unlock;
2838 }
2839
2840 fd->entry[fd->num_entries].stream = route->source_stream;
2841 fd->entry[fd->num_entries].flags = source_entry->flags;
2842 fd->entry[fd->num_entries].length = source_entry->length;
2843 fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2844
2845 fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2846
2847 if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2848 fd->entry[fd->num_entries].bus.csi2.dt =
2849 source_entry->bus.csi2.dt;
2850 } else {
2851 const struct ub960_format_info *ub960_fmt;
2852 struct v4l2_mbus_framefmt *fmt;
2853
2854 fmt = v4l2_subdev_state_get_format(state, pad,
2855 route->source_stream);
2856
2857 if (!fmt) {
2858 ret = -EINVAL;
2859 goto out_unlock;
2860 }
2861
2862 ub960_fmt = ub960_find_format(fmt->code);
2863 if (!ub960_fmt) {
2864 dev_err(dev, "Unable to find format\n");
2865 ret = -EINVAL;
2866 goto out_unlock;
2867 }
2868
2869 fd->entry[fd->num_entries].bus.csi2.dt =
2870 ub960_fmt->datatype;
2871 }
2872
2873 fd->num_entries++;
2874 }
2875
2876 out_unlock:
2877 v4l2_subdev_unlock_state(state);
2878
2879 return ret;
2880 }
2881
ub960_set_fmt(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_format * format)2882 static int ub960_set_fmt(struct v4l2_subdev *sd,
2883 struct v4l2_subdev_state *state,
2884 struct v4l2_subdev_format *format)
2885 {
2886 struct ub960_data *priv = sd_to_ub960(sd);
2887 struct v4l2_mbus_framefmt *fmt;
2888
2889 if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2890 return -EBUSY;
2891
2892 /* No transcoding, source and sink formats must match. */
2893 if (ub960_pad_is_source(priv, format->pad))
2894 return v4l2_subdev_get_fmt(sd, state, format);
2895
2896 /*
2897 * Default to the first format if the requested media bus code isn't
2898 * supported.
2899 */
2900 if (!ub960_find_format(format->format.code))
2901 format->format.code = ub960_formats[0].code;
2902
2903 fmt = v4l2_subdev_state_get_format(state, format->pad, format->stream);
2904 if (!fmt)
2905 return -EINVAL;
2906
2907 *fmt = format->format;
2908
2909 fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2910 format->stream);
2911 if (!fmt)
2912 return -EINVAL;
2913
2914 *fmt = format->format;
2915
2916 return 0;
2917 }
2918
ub960_init_state(struct v4l2_subdev * sd,struct v4l2_subdev_state * state)2919 static int ub960_init_state(struct v4l2_subdev *sd,
2920 struct v4l2_subdev_state *state)
2921 {
2922 struct ub960_data *priv = sd_to_ub960(sd);
2923
2924 struct v4l2_subdev_route routes[] = {
2925 {
2926 .sink_pad = 0,
2927 .sink_stream = 0,
2928 .source_pad = priv->hw_data->num_rxports,
2929 .source_stream = 0,
2930 .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2931 },
2932 };
2933
2934 struct v4l2_subdev_krouting routing = {
2935 .num_routes = ARRAY_SIZE(routes),
2936 .routes = routes,
2937 };
2938
2939 return _ub960_set_routing(sd, state, &routing);
2940 }
2941
2942 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2943 .enable_streams = ub960_enable_streams,
2944 .disable_streams = ub960_disable_streams,
2945
2946 .set_routing = ub960_set_routing,
2947 .get_frame_desc = ub960_get_frame_desc,
2948
2949 .get_fmt = v4l2_subdev_get_fmt,
2950 .set_fmt = ub960_set_fmt,
2951 };
2952
ub960_log_status_ub960_sp_eq(struct ub960_data * priv,unsigned int nport)2953 static void ub960_log_status_ub960_sp_eq(struct ub960_data *priv,
2954 unsigned int nport)
2955 {
2956 struct device *dev = &priv->client->dev;
2957 u8 eq_level;
2958 s8 strobe_pos;
2959 u8 v = 0;
2960
2961 /* Strobe */
2962
2963 ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
2964
2965 dev_info(dev, "\t%s strobe\n",
2966 (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
2967 "Manual");
2968
2969 if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
2970 ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
2971
2972 dev_info(dev, "\tStrobe range [%d, %d]\n",
2973 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
2974 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
2975 }
2976
2977 ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
2978
2979 dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
2980
2981 /* EQ */
2982
2983 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
2984
2985 dev_info(dev, "\t%s EQ\n",
2986 (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
2987 "Adaptive");
2988
2989 if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
2990 ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
2991
2992 dev_info(dev, "\tEQ range [%u, %u]\n",
2993 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
2994 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
2995 }
2996
2997 if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
2998 dev_info(dev, "\tEQ level %u\n", eq_level);
2999 }
3000
ub960_log_status(struct v4l2_subdev * sd)3001 static int ub960_log_status(struct v4l2_subdev *sd)
3002 {
3003 struct ub960_data *priv = sd_to_ub960(sd);
3004 struct device *dev = &priv->client->dev;
3005 struct v4l2_subdev_state *state;
3006 unsigned int nport;
3007 unsigned int i;
3008 u16 v16 = 0;
3009 u8 v = 0;
3010 u8 id[UB960_SR_FPD3_RX_ID_LEN];
3011
3012 state = v4l2_subdev_lock_and_get_active_state(sd);
3013
3014 for (i = 0; i < sizeof(id); i++)
3015 ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
3016
3017 dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
3018
3019 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3020 struct ub960_txport *txport = priv->txports[nport];
3021
3022 dev_info(dev, "TX %u\n", nport);
3023
3024 if (!txport) {
3025 dev_info(dev, "\tNot initialized\n");
3026 continue;
3027 }
3028
3029 ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
3030 dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
3031 v & (u8)BIT(0));
3032
3033 ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
3034 dev_info(dev, "\tframe counter %u\n", v16);
3035
3036 ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
3037 dev_info(dev, "\tframe error counter %u\n", v16);
3038
3039 ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
3040 dev_info(dev, "\tline counter %u\n", v16);
3041
3042 ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
3043 dev_info(dev, "\tline error counter %u\n", v16);
3044 }
3045
3046 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3047 struct ub960_rxport *rxport = priv->rxports[nport];
3048 unsigned int i;
3049
3050 dev_info(dev, "RX %u\n", nport);
3051
3052 if (!rxport) {
3053 dev_info(dev, "\tNot initialized\n");
3054 continue;
3055 }
3056
3057 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
3058
3059 if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3060 dev_info(dev, "\tLocked\n");
3061 else
3062 dev_info(dev, "\tNot locked\n");
3063
3064 dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3065 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3066 dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3067
3068 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3069 dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
3070
3071 ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3072 dev_info(dev, "\tparity errors %u\n", v16);
3073
3074 ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3075 dev_info(dev, "\tlines per frame %u\n", v16);
3076
3077 ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3078 dev_info(dev, "\tbytes per line %u\n", v16);
3079
3080 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3081 dev_info(dev, "\tcsi_err_counter %u\n", v);
3082
3083 if (!priv->hw_data->is_ub9702)
3084 ub960_log_status_ub960_sp_eq(priv, nport);
3085
3086 /* GPIOs */
3087 for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3088 u8 ctl_reg;
3089 u8 ctl_shift;
3090
3091 ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3092 ctl_shift = (i % 2) * 4;
3093
3094 ub960_rxport_read(priv, nport, ctl_reg, &v);
3095
3096 dev_info(dev, "\tGPIO%u: mode %u\n", i,
3097 (v >> ctl_shift) & 0xf);
3098 }
3099 }
3100
3101 v4l2_subdev_unlock_state(state);
3102
3103 return 0;
3104 }
3105
3106 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3107 .log_status = ub960_log_status,
3108 .subscribe_event = v4l2_ctrl_subdev_subscribe_event,
3109 .unsubscribe_event = v4l2_event_subdev_unsubscribe,
3110 };
3111
3112 static const struct v4l2_subdev_internal_ops ub960_internal_ops = {
3113 .init_state = ub960_init_state,
3114 };
3115
3116 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3117 .core = &ub960_subdev_core_ops,
3118 .pad = &ub960_pad_ops,
3119 };
3120
3121 static const struct media_entity_operations ub960_entity_ops = {
3122 .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3123 .link_validate = v4l2_subdev_link_validate,
3124 .has_pad_interdep = v4l2_subdev_has_pad_interdep,
3125 };
3126
3127 /* -----------------------------------------------------------------------------
3128 * Core
3129 */
3130
ub960_handle_events(int irq,void * arg)3131 static irqreturn_t ub960_handle_events(int irq, void *arg)
3132 {
3133 struct ub960_data *priv = arg;
3134 unsigned int i;
3135 u8 int_sts;
3136 u8 fwd_sts;
3137 int ret;
3138
3139 ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3140 if (ret || !int_sts)
3141 return IRQ_NONE;
3142
3143 dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3144
3145 ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3146 if (ret)
3147 return IRQ_NONE;
3148
3149 dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3150
3151 for (i = 0; i < priv->hw_data->num_txports; i++) {
3152 if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3153 ub960_csi_handle_events(priv, i);
3154 }
3155
3156 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3157 if (!priv->rxports[i])
3158 continue;
3159
3160 if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3161 ub960_rxport_handle_events(priv, i);
3162 }
3163
3164 return IRQ_HANDLED;
3165 }
3166
ub960_handler_work(struct work_struct * work)3167 static void ub960_handler_work(struct work_struct *work)
3168 {
3169 struct delayed_work *dwork = to_delayed_work(work);
3170 struct ub960_data *priv =
3171 container_of(dwork, struct ub960_data, poll_work);
3172
3173 ub960_handle_events(0, priv);
3174
3175 schedule_delayed_work(&priv->poll_work,
3176 msecs_to_jiffies(UB960_POLL_TIME_MS));
3177 }
3178
ub960_txport_free_ports(struct ub960_data * priv)3179 static void ub960_txport_free_ports(struct ub960_data *priv)
3180 {
3181 unsigned int nport;
3182
3183 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3184 struct ub960_txport *txport = priv->txports[nport];
3185
3186 if (!txport)
3187 continue;
3188
3189 kfree(txport);
3190 priv->txports[nport] = NULL;
3191 }
3192 }
3193
ub960_rxport_free_ports(struct ub960_data * priv)3194 static void ub960_rxport_free_ports(struct ub960_data *priv)
3195 {
3196 unsigned int nport;
3197
3198 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3199 struct ub960_rxport *rxport = priv->rxports[nport];
3200
3201 if (!rxport)
3202 continue;
3203
3204 fwnode_handle_put(rxport->source.ep_fwnode);
3205 fwnode_handle_put(rxport->ser.fwnode);
3206
3207 kfree(rxport);
3208 priv->rxports[nport] = NULL;
3209 }
3210 }
3211
3212 static int
ub960_parse_dt_rxport_link_properties(struct ub960_data * priv,struct fwnode_handle * link_fwnode,struct ub960_rxport * rxport)3213 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3214 struct fwnode_handle *link_fwnode,
3215 struct ub960_rxport *rxport)
3216 {
3217 struct device *dev = &priv->client->dev;
3218 unsigned int nport = rxport->nport;
3219 u32 rx_mode;
3220 u32 cdr_mode;
3221 s32 strobe_pos;
3222 u32 eq_level;
3223 u32 ser_i2c_alias;
3224 int ret;
3225
3226 cdr_mode = RXPORT_CDR_FPD3;
3227
3228 ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3229 if (ret < 0 && ret != -EINVAL) {
3230 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3231 "ti,cdr-mode", ret);
3232 return ret;
3233 }
3234
3235 if (cdr_mode > RXPORT_CDR_LAST) {
3236 dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3237 return -EINVAL;
3238 }
3239
3240 if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3241 dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3242 return -EINVAL;
3243 }
3244
3245 rxport->cdr_mode = cdr_mode;
3246
3247 ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3248 if (ret < 0) {
3249 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3250 "ti,rx-mode", ret);
3251 return ret;
3252 }
3253
3254 if (rx_mode > RXPORT_MODE_LAST) {
3255 dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3256 return -EINVAL;
3257 }
3258
3259 switch (rx_mode) {
3260 case RXPORT_MODE_RAW12_HF:
3261 case RXPORT_MODE_RAW12_LF:
3262 dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3263 rx_mode);
3264 return -EINVAL;
3265 default:
3266 break;
3267 }
3268
3269 rxport->rx_mode = rx_mode;
3270
3271 /* EQ & Strobe related */
3272
3273 /* Defaults */
3274 rxport->eq.manual_eq = false;
3275 rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3276 rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3277
3278 ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3279 &strobe_pos);
3280 if (ret) {
3281 if (ret != -EINVAL) {
3282 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3283 "ti,strobe-pos", ret);
3284 return ret;
3285 }
3286 } else {
3287 if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3288 strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3289 dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3290 nport, strobe_pos);
3291 return -EINVAL;
3292 }
3293
3294 /* NOTE: ignored unless global manual strobe pos is also set */
3295 rxport->eq.strobe_pos = strobe_pos;
3296 if (!priv->strobe.manual)
3297 dev_warn(dev,
3298 "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3299 nport);
3300 }
3301
3302 ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3303 if (ret) {
3304 if (ret != -EINVAL) {
3305 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3306 "ti,eq-level", ret);
3307 return ret;
3308 }
3309 } else {
3310 if (eq_level > UB960_MAX_EQ_LEVEL) {
3311 dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3312 nport, eq_level);
3313 return -EINVAL;
3314 }
3315
3316 rxport->eq.manual_eq = true;
3317 rxport->eq.manual.eq_level = eq_level;
3318 }
3319
3320 ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3321 &ser_i2c_alias);
3322 if (ret) {
3323 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3324 "i2c-alias", ret);
3325 return ret;
3326 }
3327 rxport->ser.alias = ser_i2c_alias;
3328
3329 rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3330 if (!rxport->ser.fwnode) {
3331 dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3332 return -EINVAL;
3333 }
3334
3335 return 0;
3336 }
3337
ub960_parse_dt_rxport_ep_properties(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,struct ub960_rxport * rxport)3338 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3339 struct fwnode_handle *ep_fwnode,
3340 struct ub960_rxport *rxport)
3341 {
3342 struct device *dev = &priv->client->dev;
3343 struct v4l2_fwnode_endpoint vep = {};
3344 unsigned int nport = rxport->nport;
3345 bool hsync_hi;
3346 bool vsync_hi;
3347 int ret;
3348
3349 rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3350 if (!rxport->source.ep_fwnode) {
3351 dev_err(dev, "rx%u: no remote endpoint\n", nport);
3352 return -ENODEV;
3353 }
3354
3355 /* We currently have properties only for RAW modes */
3356
3357 switch (rxport->rx_mode) {
3358 case RXPORT_MODE_RAW10:
3359 case RXPORT_MODE_RAW12_HF:
3360 case RXPORT_MODE_RAW12_LF:
3361 break;
3362 default:
3363 return 0;
3364 }
3365
3366 vep.bus_type = V4L2_MBUS_PARALLEL;
3367 ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3368 if (ret) {
3369 dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3370 goto err_put_source_ep_fwnode;
3371 }
3372
3373 hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3374 vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3375
3376 /* LineValid and FrameValid are inverse to the h/vsync active */
3377 rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3378 (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3379
3380 return 0;
3381
3382 err_put_source_ep_fwnode:
3383 fwnode_handle_put(rxport->source.ep_fwnode);
3384 return ret;
3385 }
3386
ub960_parse_dt_rxport(struct ub960_data * priv,unsigned int nport,struct fwnode_handle * link_fwnode,struct fwnode_handle * ep_fwnode)3387 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3388 struct fwnode_handle *link_fwnode,
3389 struct fwnode_handle *ep_fwnode)
3390 {
3391 static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3392 "vpoc0", "vpoc1", "vpoc2", "vpoc3"
3393 };
3394 struct device *dev = &priv->client->dev;
3395 struct ub960_rxport *rxport;
3396 int ret;
3397
3398 rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3399 if (!rxport)
3400 return -ENOMEM;
3401
3402 priv->rxports[nport] = rxport;
3403
3404 rxport->nport = nport;
3405 rxport->priv = priv;
3406
3407 ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3408 if (ret)
3409 goto err_free_rxport;
3410
3411 rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3412 if (IS_ERR(rxport->vpoc)) {
3413 ret = PTR_ERR(rxport->vpoc);
3414 if (ret == -ENODEV) {
3415 rxport->vpoc = NULL;
3416 } else {
3417 dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3418 nport, ret);
3419 goto err_put_remote_fwnode;
3420 }
3421 }
3422
3423 ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3424 if (ret)
3425 goto err_put_remote_fwnode;
3426
3427 return 0;
3428
3429 err_put_remote_fwnode:
3430 fwnode_handle_put(rxport->ser.fwnode);
3431 err_free_rxport:
3432 priv->rxports[nport] = NULL;
3433 kfree(rxport);
3434 return ret;
3435 }
3436
3437 static struct fwnode_handle *
ub960_fwnode_get_link_by_regs(struct fwnode_handle * links_fwnode,unsigned int nport)3438 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3439 unsigned int nport)
3440 {
3441 struct fwnode_handle *link_fwnode;
3442 int ret;
3443
3444 fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3445 u32 link_num;
3446
3447 if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3448 continue;
3449
3450 ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3451 if (ret) {
3452 fwnode_handle_put(link_fwnode);
3453 return NULL;
3454 }
3455
3456 if (nport == link_num)
3457 return link_fwnode;
3458 }
3459
3460 return NULL;
3461 }
3462
ub960_parse_dt_rxports(struct ub960_data * priv)3463 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3464 {
3465 struct device *dev = &priv->client->dev;
3466 struct fwnode_handle *links_fwnode;
3467 unsigned int nport;
3468 int ret;
3469
3470 links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3471 if (!links_fwnode) {
3472 dev_err(dev, "'links' node missing\n");
3473 return -ENODEV;
3474 }
3475
3476 /* Defaults, recommended by TI */
3477 priv->strobe.min = 2;
3478 priv->strobe.max = 3;
3479
3480 priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3481
3482 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3483 struct fwnode_handle *link_fwnode;
3484 struct fwnode_handle *ep_fwnode;
3485
3486 link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3487 if (!link_fwnode)
3488 continue;
3489
3490 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3491 nport, 0, 0);
3492 if (!ep_fwnode) {
3493 fwnode_handle_put(link_fwnode);
3494 continue;
3495 }
3496
3497 ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3498 ep_fwnode);
3499
3500 fwnode_handle_put(link_fwnode);
3501 fwnode_handle_put(ep_fwnode);
3502
3503 if (ret) {
3504 dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3505 goto err_put_links;
3506 }
3507 }
3508
3509 fwnode_handle_put(links_fwnode);
3510
3511 return 0;
3512
3513 err_put_links:
3514 fwnode_handle_put(links_fwnode);
3515
3516 return ret;
3517 }
3518
ub960_parse_dt_txports(struct ub960_data * priv)3519 static int ub960_parse_dt_txports(struct ub960_data *priv)
3520 {
3521 struct device *dev = &priv->client->dev;
3522 u32 nport;
3523 int ret;
3524
3525 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3526 unsigned int port = nport + priv->hw_data->num_rxports;
3527 struct fwnode_handle *ep_fwnode;
3528
3529 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3530 port, 0, 0);
3531 if (!ep_fwnode)
3532 continue;
3533
3534 ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3535
3536 fwnode_handle_put(ep_fwnode);
3537
3538 if (ret)
3539 break;
3540 }
3541
3542 return 0;
3543 }
3544
ub960_parse_dt(struct ub960_data * priv)3545 static int ub960_parse_dt(struct ub960_data *priv)
3546 {
3547 int ret;
3548
3549 ret = ub960_parse_dt_rxports(priv);
3550 if (ret)
3551 return ret;
3552
3553 ret = ub960_parse_dt_txports(priv);
3554 if (ret)
3555 goto err_free_rxports;
3556
3557 return 0;
3558
3559 err_free_rxports:
3560 ub960_rxport_free_ports(priv);
3561
3562 return ret;
3563 }
3564
ub960_notify_bound(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3565 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3566 struct v4l2_subdev *subdev,
3567 struct v4l2_async_connection *asd)
3568 {
3569 struct ub960_data *priv = sd_to_ub960(notifier->sd);
3570 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3571 struct device *dev = &priv->client->dev;
3572 u8 nport = rxport->nport;
3573 unsigned int i;
3574 int ret;
3575
3576 ret = media_entity_get_fwnode_pad(&subdev->entity,
3577 rxport->source.ep_fwnode,
3578 MEDIA_PAD_FL_SOURCE);
3579 if (ret < 0) {
3580 dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3581 return ret;
3582 }
3583
3584 rxport->source.sd = subdev;
3585 rxport->source.pad = ret;
3586
3587 ret = media_create_pad_link(&rxport->source.sd->entity,
3588 rxport->source.pad, &priv->sd.entity, nport,
3589 MEDIA_LNK_FL_ENABLED |
3590 MEDIA_LNK_FL_IMMUTABLE);
3591 if (ret) {
3592 dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3593 rxport->source.sd->name, rxport->source.pad,
3594 priv->sd.name, nport);
3595 return ret;
3596 }
3597
3598 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3599 if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3600 dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3601 return 0;
3602 }
3603 }
3604
3605 return 0;
3606 }
3607
ub960_notify_unbind(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3608 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3609 struct v4l2_subdev *subdev,
3610 struct v4l2_async_connection *asd)
3611 {
3612 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3613
3614 rxport->source.sd = NULL;
3615 }
3616
3617 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3618 .bound = ub960_notify_bound,
3619 .unbind = ub960_notify_unbind,
3620 };
3621
ub960_v4l2_notifier_register(struct ub960_data * priv)3622 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3623 {
3624 struct device *dev = &priv->client->dev;
3625 unsigned int i;
3626 int ret;
3627
3628 v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
3629
3630 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3631 struct ub960_rxport *rxport = priv->rxports[i];
3632 struct ub960_asd *asd;
3633
3634 if (!rxport)
3635 continue;
3636
3637 asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3638 rxport->source.ep_fwnode,
3639 struct ub960_asd);
3640 if (IS_ERR(asd)) {
3641 dev_err(dev, "Failed to add subdev for source %u: %pe",
3642 i, asd);
3643 v4l2_async_nf_cleanup(&priv->notifier);
3644 return PTR_ERR(asd);
3645 }
3646
3647 asd->rxport = rxport;
3648 }
3649
3650 priv->notifier.ops = &ub960_notify_ops;
3651
3652 ret = v4l2_async_nf_register(&priv->notifier);
3653 if (ret) {
3654 dev_err(dev, "Failed to register subdev_notifier");
3655 v4l2_async_nf_cleanup(&priv->notifier);
3656 return ret;
3657 }
3658
3659 return 0;
3660 }
3661
ub960_v4l2_notifier_unregister(struct ub960_data * priv)3662 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3663 {
3664 v4l2_async_nf_unregister(&priv->notifier);
3665 v4l2_async_nf_cleanup(&priv->notifier);
3666 }
3667
ub960_create_subdev(struct ub960_data * priv)3668 static int ub960_create_subdev(struct ub960_data *priv)
3669 {
3670 struct device *dev = &priv->client->dev;
3671 unsigned int i;
3672 int ret;
3673
3674 v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3675 priv->sd.internal_ops = &ub960_internal_ops;
3676
3677 v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3678 priv->sd.ctrl_handler = &priv->ctrl_handler;
3679
3680 v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3681 ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3682 priv->tx_link_freq);
3683
3684 if (priv->ctrl_handler.error) {
3685 ret = priv->ctrl_handler.error;
3686 goto err_free_ctrl;
3687 }
3688
3689 priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3690 V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
3691 priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3692 priv->sd.entity.ops = &ub960_entity_ops;
3693
3694 for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3695 priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3696 MEDIA_PAD_FL_SINK :
3697 MEDIA_PAD_FL_SOURCE;
3698 }
3699
3700 ret = media_entity_pads_init(&priv->sd.entity,
3701 priv->hw_data->num_rxports +
3702 priv->hw_data->num_txports,
3703 priv->pads);
3704 if (ret)
3705 goto err_free_ctrl;
3706
3707 priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3708
3709 ret = v4l2_subdev_init_finalize(&priv->sd);
3710 if (ret)
3711 goto err_entity_cleanup;
3712
3713 ret = ub960_v4l2_notifier_register(priv);
3714 if (ret) {
3715 dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3716 goto err_subdev_cleanup;
3717 }
3718
3719 ret = v4l2_async_register_subdev(&priv->sd);
3720 if (ret) {
3721 dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3722 goto err_unreg_notif;
3723 }
3724
3725 return 0;
3726
3727 err_unreg_notif:
3728 ub960_v4l2_notifier_unregister(priv);
3729 err_subdev_cleanup:
3730 v4l2_subdev_cleanup(&priv->sd);
3731 err_entity_cleanup:
3732 media_entity_cleanup(&priv->sd.entity);
3733 err_free_ctrl:
3734 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3735
3736 return ret;
3737 }
3738
ub960_destroy_subdev(struct ub960_data * priv)3739 static void ub960_destroy_subdev(struct ub960_data *priv)
3740 {
3741 ub960_v4l2_notifier_unregister(priv);
3742 v4l2_async_unregister_subdev(&priv->sd);
3743
3744 v4l2_subdev_cleanup(&priv->sd);
3745
3746 media_entity_cleanup(&priv->sd.entity);
3747 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3748 }
3749
3750 static const struct regmap_config ub960_regmap_config = {
3751 .name = "ds90ub960",
3752
3753 .reg_bits = 8,
3754 .val_bits = 8,
3755
3756 .max_register = 0xff,
3757
3758 /*
3759 * We do locking in the driver to cover the TX/RX port selection and the
3760 * indirect register access.
3761 */
3762 .disable_locking = true,
3763 };
3764
ub960_reset(struct ub960_data * priv,bool reset_regs)3765 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3766 {
3767 struct device *dev = &priv->client->dev;
3768 unsigned int v;
3769 int ret;
3770 u8 bit;
3771
3772 bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3773 UB960_SR_RESET_DIGITAL_RESET0;
3774
3775 ub960_write(priv, UB960_SR_RESET, bit);
3776
3777 mutex_lock(&priv->reg_lock);
3778
3779 ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3780 (v & bit) == 0, 2000, 100000);
3781
3782 mutex_unlock(&priv->reg_lock);
3783
3784 if (ret)
3785 dev_err(dev, "reset failed: %d\n", ret);
3786 }
3787
ub960_get_hw_resources(struct ub960_data * priv)3788 static int ub960_get_hw_resources(struct ub960_data *priv)
3789 {
3790 struct device *dev = &priv->client->dev;
3791
3792 priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3793 if (IS_ERR(priv->regmap))
3794 return PTR_ERR(priv->regmap);
3795
3796 priv->vddio = devm_regulator_get(dev, "vddio");
3797 if (IS_ERR(priv->vddio))
3798 return dev_err_probe(dev, PTR_ERR(priv->vddio),
3799 "cannot get VDDIO regulator\n");
3800
3801 /* get power-down pin from DT */
3802 priv->pd_gpio =
3803 devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3804 if (IS_ERR(priv->pd_gpio))
3805 return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3806 "Cannot get powerdown GPIO\n");
3807
3808 priv->refclk = devm_clk_get(dev, "refclk");
3809 if (IS_ERR(priv->refclk))
3810 return dev_err_probe(dev, PTR_ERR(priv->refclk),
3811 "Cannot get REFCLK\n");
3812
3813 return 0;
3814 }
3815
ub960_enable_core_hw(struct ub960_data * priv)3816 static int ub960_enable_core_hw(struct ub960_data *priv)
3817 {
3818 struct device *dev = &priv->client->dev;
3819 u8 rev_mask;
3820 int ret;
3821 u8 dev_sts;
3822 u8 refclk_freq;
3823
3824 ret = regulator_enable(priv->vddio);
3825 if (ret)
3826 return dev_err_probe(dev, ret,
3827 "failed to enable VDDIO regulator\n");
3828
3829 ret = clk_prepare_enable(priv->refclk);
3830 if (ret) {
3831 dev_err_probe(dev, ret, "Failed to enable refclk\n");
3832 goto err_disable_vddio;
3833 }
3834
3835 if (priv->pd_gpio) {
3836 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3837 /* wait min 2 ms for reset to complete */
3838 fsleep(2000);
3839 gpiod_set_value_cansleep(priv->pd_gpio, 0);
3840 /* wait min 2 ms for power up to finish */
3841 fsleep(2000);
3842 }
3843
3844 ub960_reset(priv, true);
3845
3846 /* Runtime check register accessibility */
3847 ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3848 if (ret) {
3849 dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3850 goto err_pd_gpio;
3851 }
3852
3853 dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3854 rev_mask);
3855
3856 ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3857 if (ret)
3858 goto err_pd_gpio;
3859
3860 if (priv->hw_data->is_ub9702)
3861 ret = ub960_read(priv, UB9702_SR_REFCLK_FREQ, &refclk_freq);
3862 else
3863 ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3864 if (ret)
3865 goto err_pd_gpio;
3866
3867 dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3868 !!(dev_sts & BIT(4)), refclk_freq,
3869 clk_get_rate(priv->refclk) / 1000000);
3870
3871 /* Disable all RX ports by default */
3872 ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3873 if (ret)
3874 goto err_pd_gpio;
3875
3876 /* release GPIO lock */
3877 if (priv->hw_data->is_ub9702) {
3878 ret = ub960_update_bits(priv, UB960_SR_RESET,
3879 UB960_SR_RESET_GPIO_LOCK_RELEASE,
3880 UB960_SR_RESET_GPIO_LOCK_RELEASE);
3881 if (ret)
3882 goto err_pd_gpio;
3883 }
3884
3885 return 0;
3886
3887 err_pd_gpio:
3888 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3889 clk_disable_unprepare(priv->refclk);
3890 err_disable_vddio:
3891 regulator_disable(priv->vddio);
3892
3893 return ret;
3894 }
3895
ub960_disable_core_hw(struct ub960_data * priv)3896 static void ub960_disable_core_hw(struct ub960_data *priv)
3897 {
3898 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3899 clk_disable_unprepare(priv->refclk);
3900 regulator_disable(priv->vddio);
3901 }
3902
ub960_probe(struct i2c_client * client)3903 static int ub960_probe(struct i2c_client *client)
3904 {
3905 struct device *dev = &client->dev;
3906 struct ub960_data *priv;
3907 unsigned int port_lock_mask;
3908 unsigned int port_mask;
3909 unsigned int nport;
3910 int ret;
3911
3912 priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3913 if (!priv)
3914 return -ENOMEM;
3915
3916 priv->client = client;
3917
3918 priv->hw_data = device_get_match_data(dev);
3919
3920 mutex_init(&priv->reg_lock);
3921
3922 INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3923
3924 /*
3925 * Initialize these to invalid values so that the first reg writes will
3926 * configure the target.
3927 */
3928 priv->reg_current.indirect_target = 0xff;
3929 priv->reg_current.rxport = 0xff;
3930 priv->reg_current.txport = 0xff;
3931
3932 ret = ub960_get_hw_resources(priv);
3933 if (ret)
3934 goto err_mutex_destroy;
3935
3936 ret = ub960_enable_core_hw(priv);
3937 if (ret)
3938 goto err_mutex_destroy;
3939
3940 ret = ub960_parse_dt(priv);
3941 if (ret)
3942 goto err_disable_core_hw;
3943
3944 ret = ub960_init_tx_ports(priv);
3945 if (ret)
3946 goto err_free_ports;
3947
3948 ret = ub960_rxport_enable_vpocs(priv);
3949 if (ret)
3950 goto err_free_ports;
3951
3952 ret = ub960_init_rx_ports(priv);
3953 if (ret)
3954 goto err_disable_vpocs;
3955
3956 ub960_reset(priv, false);
3957
3958 port_mask = 0;
3959
3960 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3961 struct ub960_rxport *rxport = priv->rxports[nport];
3962
3963 if (!rxport)
3964 continue;
3965
3966 port_mask |= BIT(nport);
3967 }
3968
3969 ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
3970 if (ret)
3971 goto err_disable_vpocs;
3972
3973 if (port_mask != port_lock_mask) {
3974 ret = -EIO;
3975 dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
3976 goto err_disable_vpocs;
3977 }
3978
3979 /*
3980 * Clear any errors caused by switching the RX port settings while
3981 * probing.
3982 */
3983 ub960_clear_rx_errors(priv);
3984
3985 ret = ub960_init_atr(priv);
3986 if (ret)
3987 goto err_disable_vpocs;
3988
3989 ret = ub960_rxport_add_serializers(priv);
3990 if (ret)
3991 goto err_uninit_atr;
3992
3993 ret = ub960_create_subdev(priv);
3994 if (ret)
3995 goto err_free_sers;
3996
3997 if (client->irq)
3998 dev_warn(dev, "irq support not implemented, using polling\n");
3999
4000 schedule_delayed_work(&priv->poll_work,
4001 msecs_to_jiffies(UB960_POLL_TIME_MS));
4002
4003 return 0;
4004
4005 err_free_sers:
4006 ub960_rxport_remove_serializers(priv);
4007 err_uninit_atr:
4008 ub960_uninit_atr(priv);
4009 err_disable_vpocs:
4010 ub960_rxport_disable_vpocs(priv);
4011 err_free_ports:
4012 ub960_rxport_free_ports(priv);
4013 ub960_txport_free_ports(priv);
4014 err_disable_core_hw:
4015 ub960_disable_core_hw(priv);
4016 err_mutex_destroy:
4017 mutex_destroy(&priv->reg_lock);
4018 return ret;
4019 }
4020
ub960_remove(struct i2c_client * client)4021 static void ub960_remove(struct i2c_client *client)
4022 {
4023 struct v4l2_subdev *sd = i2c_get_clientdata(client);
4024 struct ub960_data *priv = sd_to_ub960(sd);
4025
4026 cancel_delayed_work_sync(&priv->poll_work);
4027
4028 ub960_destroy_subdev(priv);
4029 ub960_rxport_remove_serializers(priv);
4030 ub960_uninit_atr(priv);
4031 ub960_rxport_disable_vpocs(priv);
4032 ub960_rxport_free_ports(priv);
4033 ub960_txport_free_ports(priv);
4034 ub960_disable_core_hw(priv);
4035 mutex_destroy(&priv->reg_lock);
4036 }
4037
4038 static const struct ub960_hw_data ds90ub960_hw = {
4039 .model = "ub960",
4040 .num_rxports = 4,
4041 .num_txports = 2,
4042 };
4043
4044 static const struct ub960_hw_data ds90ub9702_hw = {
4045 .model = "ub9702",
4046 .num_rxports = 4,
4047 .num_txports = 2,
4048 .is_ub9702 = true,
4049 .is_fpdlink4 = true,
4050 };
4051
4052 static const struct i2c_device_id ub960_id[] = {
4053 { "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4054 { "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4055 {}
4056 };
4057 MODULE_DEVICE_TABLE(i2c, ub960_id);
4058
4059 static const struct of_device_id ub960_dt_ids[] = {
4060 { .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4061 { .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4062 {}
4063 };
4064 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4065
4066 static struct i2c_driver ds90ub960_driver = {
4067 .probe = ub960_probe,
4068 .remove = ub960_remove,
4069 .id_table = ub960_id,
4070 .driver = {
4071 .name = "ds90ub960",
4072 .of_match_table = ub960_dt_ids,
4073 },
4074 };
4075 module_i2c_driver(ds90ub960_driver);
4076
4077 MODULE_LICENSE("GPL");
4078 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4079 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4080 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4081 MODULE_IMPORT_NS(I2C_ATR);
4082