1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
4 *
5 * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6 * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
7 */
8
9 /*
10 * (Possible) TODOs:
11 *
12 * - PM for serializer and remote peripherals. We need to manage:
13 * - VPOC
14 * - Power domain? Regulator? Somehow any remote device should be able to
15 * cause the VPOC to be turned on.
16 * - Link between the deserializer and the serializer
17 * - Related to VPOC management. We probably always want to turn on the VPOC
18 * and then enable the link.
19 * - Serializer's services: i2c, gpios, power
20 * - The serializer needs to resume before the remote peripherals can
21 * e.g. use the i2c.
22 * - How to handle gpios? Reserving a gpio essentially keeps the provider
23 * (serializer) always powered on.
24 * - Do we need a new bus for the FPD-Link? At the moment the serializers
25 * are children of the same i2c-adapter where the deserializer resides.
26 * - i2c-atr could be made embeddable instead of allocatable.
27 */
28
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/units.h>
47 #include <linux/workqueue.h>
48
49 #include <media/i2c/ds90ub9xx.h>
50 #include <media/mipi-csi2.h>
51 #include <media/v4l2-ctrls.h>
52 #include <media/v4l2-fwnode.h>
53 #include <media/v4l2-subdev.h>
54
55 #define MHZ(v) ((u32)((v) * HZ_PER_MHZ))
56
57 /*
58 * If this is defined, the i2c addresses from UB960_DEBUG_I2C_RX_ID to
59 * UB960_DEBUG_I2C_RX_ID + 3 can be used to access the paged RX port registers
60 * directly.
61 *
62 * Only for debug purposes.
63 */
64 /* #define UB960_DEBUG_I2C_RX_ID 0x40 */
65
66 #define UB960_POLL_TIME_MS 500
67
68 #define UB960_MAX_RX_NPORTS 4
69 #define UB960_MAX_TX_NPORTS 2
70 #define UB960_MAX_NPORTS (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
71
72 #define UB960_MAX_PORT_ALIASES 8
73
74 #define UB960_NUM_BC_GPIOS 4
75
76 /*
77 * Register map
78 *
79 * 0x00-0x32 Shared (UB960_SR)
80 * 0x33-0x3a CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
81 * 0x4c Shared (UB960_SR)
82 * 0x4d-0x7f FPD-Link RX, per-port paged (UB960_RR)
83 * 0xb0-0xbf Shared (UB960_SR)
84 * 0xd0-0xdf FPD-Link RX, per-port paged (UB960_RR)
85 * 0xf0-0xf5 Shared (UB960_SR)
86 * 0xf8-0xfb Shared (UB960_SR)
87 * All others Reserved
88 *
89 * Register prefixes:
90 * UB960_SR_* = Shared register
91 * UB960_RR_* = FPD-Link RX, per-port paged register
92 * UB960_TR_* = CSI-2 TX, per-port paged register
93 * UB960_XR_* = Reserved register
94 * UB960_IR_* = Indirect register
95 */
96
97 #define UB960_SR_I2C_DEV_ID 0x00
98 #define UB960_SR_RESET 0x01
99 #define UB960_SR_RESET_DIGITAL_RESET1 BIT(1)
100 #define UB960_SR_RESET_DIGITAL_RESET0 BIT(0)
101 #define UB960_SR_RESET_GPIO_LOCK_RELEASE BIT(5)
102
103 #define UB960_SR_GEN_CONFIG 0x02
104 #define UB960_SR_REV_MASK 0x03
105 #define UB960_SR_DEVICE_STS 0x04
106 #define UB960_SR_PAR_ERR_THOLD_HI 0x05
107 #define UB960_SR_PAR_ERR_THOLD_LO 0x06
108 #define UB960_SR_BCC_WDOG_CTL 0x07
109 #define UB960_SR_I2C_CTL1 0x08
110 #define UB960_SR_I2C_CTL2 0x09
111 #define UB960_SR_SCL_HIGH_TIME 0x0a
112 #define UB960_SR_SCL_LOW_TIME 0x0b
113 #define UB960_SR_RX_PORT_CTL 0x0c
114 #define UB960_SR_IO_CTL 0x0d
115 #define UB960_SR_GPIO_PIN_STS 0x0e
116 #define UB960_SR_GPIO_INPUT_CTL 0x0f
117 #define UB960_SR_GPIO_PIN_CTL(n) (0x10 + (n)) /* n < UB960_NUM_GPIOS */
118 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL 5
119 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT 2
120 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN BIT(0)
121
122 #define UB960_SR_FS_CTL 0x18
123 #define UB960_SR_FS_HIGH_TIME_1 0x19
124 #define UB960_SR_FS_HIGH_TIME_0 0x1a
125 #define UB960_SR_FS_LOW_TIME_1 0x1b
126 #define UB960_SR_FS_LOW_TIME_0 0x1c
127 #define UB960_SR_MAX_FRM_HI 0x1d
128 #define UB960_SR_MAX_FRM_LO 0x1e
129 #define UB960_SR_CSI_PLL_CTL 0x1f
130
131 #define UB960_SR_FWD_CTL1 0x20
132 #define UB960_SR_FWD_CTL1_PORT_DIS(n) BIT((n) + 4)
133
134 #define UB960_SR_FWD_CTL2 0x21
135 #define UB960_SR_FWD_STS 0x22
136
137 #define UB960_SR_INTERRUPT_CTL 0x23
138 #define UB960_SR_INTERRUPT_CTL_INT_EN BIT(7)
139 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0 BIT(4)
140 #define UB960_SR_INTERRUPT_CTL_IE_RX(n) BIT((n)) /* rxport[n] IRQ */
141
142 #define UB960_SR_INTERRUPT_STS 0x24
143 #define UB960_SR_INTERRUPT_STS_INT BIT(7)
144 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n) BIT(4 + (n)) /* txport[n] IRQ */
145 #define UB960_SR_INTERRUPT_STS_IS_RX(n) BIT((n)) /* rxport[n] IRQ */
146
147 #define UB960_SR_TS_CONFIG 0x25
148 #define UB960_SR_TS_CONTROL 0x26
149 #define UB960_SR_TS_LINE_HI 0x27
150 #define UB960_SR_TS_LINE_LO 0x28
151 #define UB960_SR_TS_STATUS 0x29
152 #define UB960_SR_TIMESTAMP_P0_HI 0x2a
153 #define UB960_SR_TIMESTAMP_P0_LO 0x2b
154 #define UB960_SR_TIMESTAMP_P1_HI 0x2c
155 #define UB960_SR_TIMESTAMP_P1_LO 0x2d
156
157 #define UB960_SR_CSI_PORT_SEL 0x32
158
159 #define UB960_TR_CSI_CTL 0x33
160 #define UB960_TR_CSI_CTL_CSI_CAL_EN BIT(6)
161 #define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK BIT(1)
162 #define UB960_TR_CSI_CTL_CSI_ENABLE BIT(0)
163
164 #define UB960_TR_CSI_CTL2 0x34
165 #define UB960_TR_CSI_STS 0x35
166 #define UB960_TR_CSI_TX_ICR 0x36
167
168 #define UB960_TR_CSI_TX_ISR 0x37
169 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR BIT(3)
170 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR BIT(1)
171
172 #define UB960_TR_CSI_TEST_CTL 0x38
173 #define UB960_TR_CSI_TEST_PATT_HI 0x39
174 #define UB960_TR_CSI_TEST_PATT_LO 0x3a
175
176 #define UB960_XR_SFILTER_CFG 0x41
177 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT 4
178 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT 0
179
180 #define UB960_XR_AEQ_CTL1 0x42
181 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK BIT(6)
182 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING BIT(5)
183 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY BIT(4)
184 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK \
185 (UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK | \
186 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
187 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
188 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN BIT(0)
189
190 #define UB960_XR_AEQ_ERR_THOLD 0x43
191
192 #define UB960_RR_BCC_ERR_CTL 0x46
193 #define UB960_RR_BCC_STATUS 0x47
194 #define UB960_RR_BCC_STATUS_SEQ_ERROR BIT(5)
195 #define UB960_RR_BCC_STATUS_MASTER_ERR BIT(4)
196 #define UB960_RR_BCC_STATUS_MASTER_TO BIT(3)
197 #define UB960_RR_BCC_STATUS_SLAVE_ERR BIT(2)
198 #define UB960_RR_BCC_STATUS_SLAVE_TO BIT(1)
199 #define UB960_RR_BCC_STATUS_RESP_ERR BIT(0)
200 #define UB960_RR_BCC_STATUS_ERROR_MASK \
201 (UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
202 UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR | \
203 UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
204
205 #define UB960_RR_FPD3_CAP 0x4a
206 #define UB960_RR_RAW_EMBED_DTYPE 0x4b
207 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT 6
208
209 #define UB960_SR_FPD3_PORT_SEL 0x4c
210
211 #define UB960_RR_RX_PORT_STS1 0x4d
212 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR BIT(5)
213 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG BIT(4)
214 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR BIT(3)
215 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR BIT(2)
216 #define UB960_RR_RX_PORT_STS1_PORT_PASS BIT(1)
217 #define UB960_RR_RX_PORT_STS1_LOCK_STS BIT(0)
218 #define UB960_RR_RX_PORT_STS1_ERROR_MASK \
219 (UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
220 UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
221 UB960_RR_RX_PORT_STS1_PARITY_ERROR)
222
223 #define UB960_RR_RX_PORT_STS2 0x4e
224 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
225 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG BIT(6)
226 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
227 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR BIT(4)
228 #define UB960_RR_RX_PORT_STS2_CSI_ERROR BIT(3)
229 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE BIT(2)
230 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT BIT(1)
231 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG BIT(0)
232 #define UB960_RR_RX_PORT_STS2_ERROR_MASK \
233 UB960_RR_RX_PORT_STS2_BUFFER_ERROR
234
235 #define UB960_RR_RX_FREQ_HIGH 0x4f
236 #define UB960_RR_RX_FREQ_LOW 0x50
237 #define UB960_RR_SENSOR_STS_0 0x51
238 #define UB960_RR_SENSOR_STS_1 0x52
239 #define UB960_RR_SENSOR_STS_2 0x53
240 #define UB960_RR_SENSOR_STS_3 0x54
241 #define UB960_RR_RX_PAR_ERR_HI 0x55
242 #define UB960_RR_RX_PAR_ERR_LO 0x56
243 #define UB960_RR_BIST_ERR_COUNT 0x57
244
245 #define UB960_RR_BCC_CONFIG 0x58
246 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH BIT(6)
247 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK GENMASK(2, 0)
248
249 #define UB960_RR_DATAPATH_CTL1 0x59
250 #define UB960_RR_DATAPATH_CTL2 0x5a
251 #define UB960_RR_SER_ID 0x5b
252 #define UB960_RR_SER_ALIAS_ID 0x5c
253
254 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
255 #define UB960_RR_SLAVE_ID(n) (0x5d + (n))
256 #define UB960_RR_SLAVE_ALIAS(n) (0x65 + (n))
257
258 #define UB960_RR_PORT_CONFIG 0x6d
259 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK GENMASK(1, 0)
260
261 #define UB960_RR_BC_GPIO_CTL(n) (0x6e + (n)) /* n < 2 */
262 #define UB960_RR_RAW10_ID 0x70
263 #define UB960_RR_RAW10_ID_VC_SHIFT 6
264 #define UB960_RR_RAW10_ID_DT_SHIFT 0
265
266 #define UB960_RR_RAW12_ID 0x71
267 #define UB960_RR_CSI_VC_MAP 0x72
268 #define UB960_RR_CSI_VC_MAP_SHIFT(x) ((x) * 2)
269
270 #define UB960_RR_LINE_COUNT_HI 0x73
271 #define UB960_RR_LINE_COUNT_LO 0x74
272 #define UB960_RR_LINE_LEN_1 0x75
273 #define UB960_RR_LINE_LEN_0 0x76
274 #define UB960_RR_FREQ_DET_CTL 0x77
275 #define UB960_RR_MAILBOX_1 0x78
276 #define UB960_RR_MAILBOX_2 0x79
277
278 #define UB960_RR_CSI_RX_STS 0x7a
279 #define UB960_RR_CSI_RX_STS_LENGTH_ERR BIT(3)
280 #define UB960_RR_CSI_RX_STS_CKSUM_ERR BIT(2)
281 #define UB960_RR_CSI_RX_STS_ECC2_ERR BIT(1)
282 #define UB960_RR_CSI_RX_STS_ECC1_ERR BIT(0)
283 #define UB960_RR_CSI_RX_STS_ERROR_MASK \
284 (UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
285 UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
286
287 #define UB960_RR_CSI_ERR_COUNTER 0x7b
288 #define UB960_RR_PORT_CONFIG2 0x7c
289 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
290 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
291
292 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW BIT(1)
293 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW BIT(0)
294
295 #define UB960_RR_PORT_PASS_CTL 0x7d
296 #define UB960_RR_SEN_INT_RISE_CTL 0x7e
297 #define UB960_RR_SEN_INT_FALL_CTL 0x7f
298
299 #define UB960_SR_CSI_FRAME_COUNT_HI(n) (0x90 + 8 * (n))
300 #define UB960_SR_CSI_FRAME_COUNT_LO(n) (0x91 + 8 * (n))
301 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n) (0x92 + 8 * (n))
302 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n) (0x93 + 8 * (n))
303 #define UB960_SR_CSI_LINE_COUNT_HI(n) (0x94 + 8 * (n))
304 #define UB960_SR_CSI_LINE_COUNT_LO(n) (0x95 + 8 * (n))
305 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n) (0x96 + 8 * (n))
306 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n) (0x97 + 8 * (n))
307
308 #define UB960_XR_REFCLK_FREQ 0xa5 /* UB960 */
309
310 #define UB960_RR_VC_ID_MAP(x) (0xa0 + (x)) /* UB9702 */
311
312 #define UB960_SR_IND_ACC_CTL 0xb0
313 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC BIT(1)
314
315 #define UB960_SR_IND_ACC_ADDR 0xb1
316 #define UB960_SR_IND_ACC_DATA 0xb2
317 #define UB960_SR_BIST_CONTROL 0xb3
318 #define UB960_SR_MODE_IDX_STS 0xb8
319 #define UB960_SR_LINK_ERROR_COUNT 0xb9
320 #define UB960_SR_FPD3_ENC_CTL 0xba
321 #define UB960_SR_FV_MIN_TIME 0xbc
322 #define UB960_SR_GPIO_PD_CTL 0xbe
323
324 #define UB960_SR_FPD_RATE_CFG 0xc2 /* UB9702 */
325 #define UB960_SR_CSI_PLL_DIV 0xc9 /* UB9702 */
326
327 #define UB960_RR_PORT_DEBUG 0xd0
328 #define UB960_RR_AEQ_CTL2 0xd2
329 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR BIT(2)
330
331 #define UB960_RR_AEQ_STATUS 0xd3
332 #define UB960_RR_AEQ_STATUS_STATUS_2 GENMASK(5, 3)
333 #define UB960_RR_AEQ_STATUS_STATUS_1 GENMASK(2, 0)
334
335 #define UB960_RR_AEQ_BYPASS 0xd4
336 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT 5
337 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK GENMASK(7, 5)
338 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT 1
339 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK GENMASK(3, 1)
340 #define UB960_RR_AEQ_BYPASS_ENABLE BIT(0)
341
342 #define UB960_RR_AEQ_MIN_MAX 0xd5
343 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT 4
344 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT 0
345
346 #define UB960_RR_SFILTER_STS_0 0xd6
347 #define UB960_RR_SFILTER_STS_1 0xd7
348 #define UB960_RR_PORT_ICR_HI 0xd8
349 #define UB960_RR_PORT_ICR_LO 0xd9
350 #define UB960_RR_PORT_ISR_HI 0xda
351 #define UB960_RR_PORT_ISR_LO 0xdb
352 #define UB960_RR_FC_GPIO_STS 0xdc
353 #define UB960_RR_FC_GPIO_ICR 0xdd
354 #define UB960_RR_SEN_INT_RISE_STS 0xde
355 #define UB960_RR_SEN_INT_FALL_STS 0xdf
356
357 #define UB960_RR_CHANNEL_MODE 0xe4 /* UB9702 */
358
359 #define UB960_SR_FPD3_RX_ID(n) (0xf0 + (n))
360 #define UB960_SR_FPD3_RX_ID_LEN 6
361
362 #define UB960_SR_I2C_RX_ID(n) (0xf8 + (n))
363
364 #define UB9702_SR_REFCLK_FREQ 0x3d
365
366 /* Indirect register blocks */
367 #define UB960_IND_TARGET_PAT_GEN 0x00
368 #define UB960_IND_TARGET_RX_ANA(n) (0x01 + (n))
369 #define UB960_IND_TARGET_CSI_ANA 0x07
370
371 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
372
373 #define UB960_IR_PGEN_CTL 0x01
374 #define UB960_IR_PGEN_CTL_PGEN_ENABLE BIT(0)
375
376 #define UB960_IR_PGEN_CFG 0x02
377 #define UB960_IR_PGEN_CSI_DI 0x03
378 #define UB960_IR_PGEN_LINE_SIZE1 0x04
379 #define UB960_IR_PGEN_LINE_SIZE0 0x05
380 #define UB960_IR_PGEN_BAR_SIZE1 0x06
381 #define UB960_IR_PGEN_BAR_SIZE0 0x07
382 #define UB960_IR_PGEN_ACT_LPF1 0x08
383 #define UB960_IR_PGEN_ACT_LPF0 0x09
384 #define UB960_IR_PGEN_TOT_LPF1 0x0a
385 #define UB960_IR_PGEN_TOT_LPF0 0x0b
386 #define UB960_IR_PGEN_LINE_PD1 0x0c
387 #define UB960_IR_PGEN_LINE_PD0 0x0d
388 #define UB960_IR_PGEN_VBP 0x0e
389 #define UB960_IR_PGEN_VFP 0x0f
390 #define UB960_IR_PGEN_COLOR(n) (0x10 + (n)) /* n < 15 */
391
392 #define UB960_IR_RX_ANA_STROBE_SET_CLK 0x08
393 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY BIT(3)
394 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK GENMASK(2, 0)
395
396 #define UB960_IR_RX_ANA_STROBE_SET_DATA 0x09
397 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY BIT(3)
398 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK GENMASK(2, 0)
399
400 /* EQ related */
401
402 #define UB960_MIN_AEQ_STROBE_POS -7
403 #define UB960_MAX_AEQ_STROBE_POS 7
404
405 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
406
407 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
408 #define UB960_MAX_MANUAL_STROBE_POS (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
409 #define UB960_NUM_MANUAL_STROBE_POS (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
410
411 #define UB960_MIN_EQ_LEVEL 0
412 #define UB960_MAX_EQ_LEVEL 14
413 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
414
415 struct ub960_hw_data {
416 const char *model;
417 u8 num_rxports;
418 u8 num_txports;
419 bool is_ub9702;
420 bool is_fpdlink4;
421 };
422
423 enum ub960_rxport_mode {
424 RXPORT_MODE_RAW10 = 0,
425 RXPORT_MODE_RAW12_HF = 1,
426 RXPORT_MODE_RAW12_LF = 2,
427 RXPORT_MODE_CSI2_SYNC = 3,
428 RXPORT_MODE_CSI2_NONSYNC = 4,
429 RXPORT_MODE_LAST = RXPORT_MODE_CSI2_NONSYNC,
430 };
431
432 enum ub960_rxport_cdr {
433 RXPORT_CDR_FPD3 = 0,
434 RXPORT_CDR_FPD4 = 1,
435 RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
436 };
437
438 struct ub960_rxport {
439 struct ub960_data *priv;
440 u8 nport; /* RX port number, and index in priv->rxport[] */
441
442 struct {
443 struct v4l2_subdev *sd;
444 u16 pad;
445 struct fwnode_handle *ep_fwnode;
446 } source;
447
448 /* Serializer */
449 struct {
450 struct fwnode_handle *fwnode;
451 struct i2c_client *client;
452 unsigned short alias; /* I2C alias (lower 7 bits) */
453 struct ds90ub9xx_platform_data pdata;
454 } ser;
455
456 enum ub960_rxport_mode rx_mode;
457 enum ub960_rxport_cdr cdr_mode;
458
459 u8 lv_fv_pol; /* LV and FV polarities */
460
461 struct regulator *vpoc;
462
463 /* EQ settings */
464 struct {
465 bool manual_eq;
466
467 s8 strobe_pos;
468
469 union {
470 struct {
471 u8 eq_level_min;
472 u8 eq_level_max;
473 } aeq;
474
475 struct {
476 u8 eq_level;
477 } manual;
478 };
479 } eq;
480
481 const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
482 };
483
484 struct ub960_asd {
485 struct v4l2_async_connection base;
486 struct ub960_rxport *rxport;
487 };
488
to_ub960_asd(struct v4l2_async_connection * asd)489 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
490 {
491 return container_of(asd, struct ub960_asd, base);
492 }
493
494 struct ub960_txport {
495 struct ub960_data *priv;
496 u8 nport; /* TX port number, and index in priv->txport[] */
497
498 u32 num_data_lanes;
499 bool non_continous_clk;
500 };
501
502 struct ub960_data {
503 const struct ub960_hw_data *hw_data;
504 struct i2c_client *client; /* for shared local registers */
505 struct regmap *regmap;
506
507 /* lock for register access */
508 struct mutex reg_lock;
509
510 struct clk *refclk;
511
512 struct regulator *vddio;
513
514 struct gpio_desc *pd_gpio;
515 struct delayed_work poll_work;
516 struct ub960_rxport *rxports[UB960_MAX_RX_NPORTS];
517 struct ub960_txport *txports[UB960_MAX_TX_NPORTS];
518
519 struct v4l2_subdev sd;
520 struct media_pad pads[UB960_MAX_NPORTS];
521
522 struct v4l2_ctrl_handler ctrl_handler;
523 struct v4l2_async_notifier notifier;
524
525 u32 tx_data_rate; /* Nominal data rate (Gb/s) */
526 s64 tx_link_freq[1];
527
528 struct i2c_atr *atr;
529
530 struct {
531 u8 rxport;
532 u8 txport;
533 u8 indirect_target;
534 } reg_current;
535
536 bool streaming;
537
538 u8 stored_fwd_ctl;
539
540 u64 stream_enable_mask[UB960_MAX_NPORTS];
541
542 /* These are common to all ports */
543 struct {
544 bool manual;
545
546 s8 min;
547 s8 max;
548 } strobe;
549 };
550
sd_to_ub960(struct v4l2_subdev * sd)551 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
552 {
553 return container_of(sd, struct ub960_data, sd);
554 }
555
ub960_pad_is_sink(struct ub960_data * priv,u32 pad)556 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
557 {
558 return pad < priv->hw_data->num_rxports;
559 }
560
ub960_pad_is_source(struct ub960_data * priv,u32 pad)561 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
562 {
563 return pad >= priv->hw_data->num_rxports;
564 }
565
ub960_pad_to_port(struct ub960_data * priv,u32 pad)566 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
567 {
568 if (ub960_pad_is_sink(priv, pad))
569 return pad;
570 else
571 return pad - priv->hw_data->num_rxports;
572 }
573
574 struct ub960_format_info {
575 u32 code;
576 u32 bpp;
577 u8 datatype;
578 bool meta;
579 };
580
581 static const struct ub960_format_info ub960_formats[] = {
582 { .code = MEDIA_BUS_FMT_RGB888_1X24, .bpp = 24, .datatype = MIPI_CSI2_DT_RGB888, },
583
584 { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
585 { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
586 { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
587 { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
588
589 { .code = MEDIA_BUS_FMT_SBGGR8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, },
590 { .code = MEDIA_BUS_FMT_SGBRG8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, },
591 { .code = MEDIA_BUS_FMT_SGRBG8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, },
592 { .code = MEDIA_BUS_FMT_SRGGB8_1X8, .bpp = 8, .datatype = MIPI_CSI2_DT_RAW8, },
593
594 { .code = MEDIA_BUS_FMT_SBGGR10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, },
595 { .code = MEDIA_BUS_FMT_SGBRG10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, },
596 { .code = MEDIA_BUS_FMT_SGRBG10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, },
597 { .code = MEDIA_BUS_FMT_SRGGB10_1X10, .bpp = 10, .datatype = MIPI_CSI2_DT_RAW10, },
598
599 { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
600 { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
601 { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
602 { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
603 };
604
ub960_find_format(u32 code)605 static const struct ub960_format_info *ub960_find_format(u32 code)
606 {
607 unsigned int i;
608
609 for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
610 if (ub960_formats[i].code == code)
611 return &ub960_formats[i];
612 }
613
614 return NULL;
615 }
616
617 /* -----------------------------------------------------------------------------
618 * Basic device access
619 */
620
ub960_read(struct ub960_data * priv,u8 reg,u8 * val)621 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
622 {
623 struct device *dev = &priv->client->dev;
624 unsigned int v;
625 int ret;
626
627 mutex_lock(&priv->reg_lock);
628
629 ret = regmap_read(priv->regmap, reg, &v);
630 if (ret) {
631 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
632 __func__, reg, ret);
633 goto out_unlock;
634 }
635
636 *val = v;
637
638 out_unlock:
639 mutex_unlock(&priv->reg_lock);
640
641 return ret;
642 }
643
ub960_write(struct ub960_data * priv,u8 reg,u8 val)644 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
645 {
646 struct device *dev = &priv->client->dev;
647 int ret;
648
649 mutex_lock(&priv->reg_lock);
650
651 ret = regmap_write(priv->regmap, reg, val);
652 if (ret)
653 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
654 __func__, reg, ret);
655
656 mutex_unlock(&priv->reg_lock);
657
658 return ret;
659 }
660
ub960_update_bits(struct ub960_data * priv,u8 reg,u8 mask,u8 val)661 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
662 {
663 struct device *dev = &priv->client->dev;
664 int ret;
665
666 mutex_lock(&priv->reg_lock);
667
668 ret = regmap_update_bits(priv->regmap, reg, mask, val);
669 if (ret)
670 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
671 __func__, reg, ret);
672
673 mutex_unlock(&priv->reg_lock);
674
675 return ret;
676 }
677
ub960_read16(struct ub960_data * priv,u8 reg,u16 * val)678 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
679 {
680 struct device *dev = &priv->client->dev;
681 __be16 __v;
682 int ret;
683
684 mutex_lock(&priv->reg_lock);
685
686 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
687 if (ret) {
688 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
689 __func__, reg, ret);
690 goto out_unlock;
691 }
692
693 *val = be16_to_cpu(__v);
694
695 out_unlock:
696 mutex_unlock(&priv->reg_lock);
697
698 return ret;
699 }
700
ub960_rxport_select(struct ub960_data * priv,u8 nport)701 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
702 {
703 struct device *dev = &priv->client->dev;
704 int ret;
705
706 lockdep_assert_held(&priv->reg_lock);
707
708 if (priv->reg_current.rxport == nport)
709 return 0;
710
711 ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
712 (nport << 4) | BIT(nport));
713 if (ret) {
714 dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
715 nport, ret);
716 return ret;
717 }
718
719 priv->reg_current.rxport = nport;
720
721 return 0;
722 }
723
ub960_rxport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)724 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
725 {
726 struct device *dev = &priv->client->dev;
727 unsigned int v;
728 int ret;
729
730 mutex_lock(&priv->reg_lock);
731
732 ret = ub960_rxport_select(priv, nport);
733 if (ret)
734 goto out_unlock;
735
736 ret = regmap_read(priv->regmap, reg, &v);
737 if (ret) {
738 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
739 __func__, reg, ret);
740 goto out_unlock;
741 }
742
743 *val = v;
744
745 out_unlock:
746 mutex_unlock(&priv->reg_lock);
747
748 return ret;
749 }
750
ub960_rxport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)751 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
752 {
753 struct device *dev = &priv->client->dev;
754 int ret;
755
756 mutex_lock(&priv->reg_lock);
757
758 ret = ub960_rxport_select(priv, nport);
759 if (ret)
760 goto out_unlock;
761
762 ret = regmap_write(priv->regmap, reg, val);
763 if (ret)
764 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
765 __func__, reg, ret);
766
767 out_unlock:
768 mutex_unlock(&priv->reg_lock);
769
770 return ret;
771 }
772
ub960_rxport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)773 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
774 u8 mask, u8 val)
775 {
776 struct device *dev = &priv->client->dev;
777 int ret;
778
779 mutex_lock(&priv->reg_lock);
780
781 ret = ub960_rxport_select(priv, nport);
782 if (ret)
783 goto out_unlock;
784
785 ret = regmap_update_bits(priv->regmap, reg, mask, val);
786 if (ret)
787 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
788 __func__, reg, ret);
789
790 out_unlock:
791 mutex_unlock(&priv->reg_lock);
792
793 return ret;
794 }
795
ub960_rxport_read16(struct ub960_data * priv,u8 nport,u8 reg,u16 * val)796 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
797 u16 *val)
798 {
799 struct device *dev = &priv->client->dev;
800 __be16 __v;
801 int ret;
802
803 mutex_lock(&priv->reg_lock);
804
805 ret = ub960_rxport_select(priv, nport);
806 if (ret)
807 goto out_unlock;
808
809 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
810 if (ret) {
811 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
812 __func__, reg, ret);
813 goto out_unlock;
814 }
815
816 *val = be16_to_cpu(__v);
817
818 out_unlock:
819 mutex_unlock(&priv->reg_lock);
820
821 return ret;
822 }
823
ub960_txport_select(struct ub960_data * priv,u8 nport)824 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
825 {
826 struct device *dev = &priv->client->dev;
827 int ret;
828
829 lockdep_assert_held(&priv->reg_lock);
830
831 if (priv->reg_current.txport == nport)
832 return 0;
833
834 ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
835 (nport << 4) | BIT(nport));
836 if (ret) {
837 dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
838 nport, ret);
839 return ret;
840 }
841
842 priv->reg_current.txport = nport;
843
844 return 0;
845 }
846
ub960_txport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)847 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
848 {
849 struct device *dev = &priv->client->dev;
850 unsigned int v;
851 int ret;
852
853 mutex_lock(&priv->reg_lock);
854
855 ret = ub960_txport_select(priv, nport);
856 if (ret)
857 goto out_unlock;
858
859 ret = regmap_read(priv->regmap, reg, &v);
860 if (ret) {
861 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
862 __func__, reg, ret);
863 goto out_unlock;
864 }
865
866 *val = v;
867
868 out_unlock:
869 mutex_unlock(&priv->reg_lock);
870
871 return ret;
872 }
873
ub960_txport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)874 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
875 {
876 struct device *dev = &priv->client->dev;
877 int ret;
878
879 mutex_lock(&priv->reg_lock);
880
881 ret = ub960_txport_select(priv, nport);
882 if (ret)
883 goto out_unlock;
884
885 ret = regmap_write(priv->regmap, reg, val);
886 if (ret)
887 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
888 __func__, reg, ret);
889
890 out_unlock:
891 mutex_unlock(&priv->reg_lock);
892
893 return ret;
894 }
895
ub960_txport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)896 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
897 u8 mask, u8 val)
898 {
899 struct device *dev = &priv->client->dev;
900 int ret;
901
902 mutex_lock(&priv->reg_lock);
903
904 ret = ub960_txport_select(priv, nport);
905 if (ret)
906 goto out_unlock;
907
908 ret = regmap_update_bits(priv->regmap, reg, mask, val);
909 if (ret)
910 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
911 __func__, reg, ret);
912
913 out_unlock:
914 mutex_unlock(&priv->reg_lock);
915
916 return ret;
917 }
918
ub960_select_ind_reg_block(struct ub960_data * priv,u8 block)919 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
920 {
921 struct device *dev = &priv->client->dev;
922 int ret;
923
924 lockdep_assert_held(&priv->reg_lock);
925
926 if (priv->reg_current.indirect_target == block)
927 return 0;
928
929 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
930 if (ret) {
931 dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
932 __func__, block, ret);
933 return ret;
934 }
935
936 priv->reg_current.indirect_target = block;
937
938 return 0;
939 }
940
ub960_read_ind(struct ub960_data * priv,u8 block,u8 reg,u8 * val)941 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
942 {
943 struct device *dev = &priv->client->dev;
944 unsigned int v;
945 int ret;
946
947 mutex_lock(&priv->reg_lock);
948
949 ret = ub960_select_ind_reg_block(priv, block);
950 if (ret)
951 goto out_unlock;
952
953 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
954 if (ret) {
955 dev_err(dev,
956 "Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
957 block, reg, ret);
958 goto out_unlock;
959 }
960
961 ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
962 if (ret) {
963 dev_err(dev,
964 "Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
965 block, reg, ret);
966 goto out_unlock;
967 }
968
969 *val = v;
970
971 out_unlock:
972 mutex_unlock(&priv->reg_lock);
973
974 return ret;
975 }
976
ub960_write_ind(struct ub960_data * priv,u8 block,u8 reg,u8 val)977 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
978 {
979 struct device *dev = &priv->client->dev;
980 int ret;
981
982 mutex_lock(&priv->reg_lock);
983
984 ret = ub960_select_ind_reg_block(priv, block);
985 if (ret)
986 goto out_unlock;
987
988 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
989 if (ret) {
990 dev_err(dev,
991 "Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
992 block, reg, ret);
993 goto out_unlock;
994 }
995
996 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
997 if (ret) {
998 dev_err(dev,
999 "Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
1000 block, reg, ret);
1001 goto out_unlock;
1002 }
1003
1004 out_unlock:
1005 mutex_unlock(&priv->reg_lock);
1006
1007 return ret;
1008 }
1009
ub960_ind_update_bits(struct ub960_data * priv,u8 block,u8 reg,u8 mask,u8 val)1010 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
1011 u8 mask, u8 val)
1012 {
1013 struct device *dev = &priv->client->dev;
1014 int ret;
1015
1016 mutex_lock(&priv->reg_lock);
1017
1018 ret = ub960_select_ind_reg_block(priv, block);
1019 if (ret)
1020 goto out_unlock;
1021
1022 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
1023 if (ret) {
1024 dev_err(dev,
1025 "Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1026 block, reg, ret);
1027 goto out_unlock;
1028 }
1029
1030 ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1031 val);
1032 if (ret) {
1033 dev_err(dev,
1034 "Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1035 block, reg, ret);
1036 goto out_unlock;
1037 }
1038
1039 out_unlock:
1040 mutex_unlock(&priv->reg_lock);
1041
1042 return ret;
1043 }
1044
1045 /* -----------------------------------------------------------------------------
1046 * I2C-ATR (address translator)
1047 */
1048
ub960_atr_attach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client,u16 alias)1049 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1050 const struct i2c_client *client, u16 alias)
1051 {
1052 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1053 struct ub960_rxport *rxport = priv->rxports[chan_id];
1054 struct device *dev = &priv->client->dev;
1055 unsigned int reg_idx;
1056
1057 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1058 if (!rxport->aliased_clients[reg_idx])
1059 break;
1060 }
1061
1062 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1063 dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1064 return -EADDRNOTAVAIL;
1065 }
1066
1067 rxport->aliased_clients[reg_idx] = client;
1068
1069 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1070 client->addr << 1);
1071 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1072 alias << 1);
1073
1074 dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1075 rxport->nport, client->addr, alias, reg_idx);
1076
1077 return 0;
1078 }
1079
ub960_atr_detach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client)1080 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1081 const struct i2c_client *client)
1082 {
1083 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1084 struct ub960_rxport *rxport = priv->rxports[chan_id];
1085 struct device *dev = &priv->client->dev;
1086 unsigned int reg_idx;
1087
1088 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1089 if (rxport->aliased_clients[reg_idx] == client)
1090 break;
1091 }
1092
1093 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1094 dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1095 rxport->nport, client->addr);
1096 return;
1097 }
1098
1099 rxport->aliased_clients[reg_idx] = NULL;
1100
1101 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1102
1103 dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1104 client->addr, reg_idx);
1105 }
1106
1107 static const struct i2c_atr_ops ub960_atr_ops = {
1108 .attach_client = ub960_atr_attach_client,
1109 .detach_client = ub960_atr_detach_client,
1110 };
1111
ub960_init_atr(struct ub960_data * priv)1112 static int ub960_init_atr(struct ub960_data *priv)
1113 {
1114 struct device *dev = &priv->client->dev;
1115 struct i2c_adapter *parent_adap = priv->client->adapter;
1116
1117 priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1118 priv->hw_data->num_rxports);
1119 if (IS_ERR(priv->atr))
1120 return PTR_ERR(priv->atr);
1121
1122 i2c_atr_set_driver_data(priv->atr, priv);
1123
1124 return 0;
1125 }
1126
ub960_uninit_atr(struct ub960_data * priv)1127 static void ub960_uninit_atr(struct ub960_data *priv)
1128 {
1129 i2c_atr_delete(priv->atr);
1130 priv->atr = NULL;
1131 }
1132
1133 /* -----------------------------------------------------------------------------
1134 * TX ports
1135 */
1136
ub960_parse_dt_txport(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,u8 nport)1137 static int ub960_parse_dt_txport(struct ub960_data *priv,
1138 struct fwnode_handle *ep_fwnode,
1139 u8 nport)
1140 {
1141 struct device *dev = &priv->client->dev;
1142 struct v4l2_fwnode_endpoint vep = {};
1143 struct ub960_txport *txport;
1144 int ret;
1145
1146 txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1147 if (!txport)
1148 return -ENOMEM;
1149
1150 txport->priv = priv;
1151 txport->nport = nport;
1152
1153 vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1154 ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1155 if (ret) {
1156 dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1157 goto err_free_txport;
1158 }
1159
1160 txport->non_continous_clk = vep.bus.mipi_csi2.flags &
1161 V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK;
1162
1163 txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1164
1165 if (vep.nr_of_link_frequencies != 1) {
1166 ret = -EINVAL;
1167 goto err_free_vep;
1168 }
1169
1170 priv->tx_link_freq[0] = vep.link_frequencies[0];
1171 priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1172
1173 if (priv->tx_data_rate != MHZ(1600) &&
1174 priv->tx_data_rate != MHZ(1200) &&
1175 priv->tx_data_rate != MHZ(800) &&
1176 priv->tx_data_rate != MHZ(400)) {
1177 dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1178 ret = -EINVAL;
1179 goto err_free_vep;
1180 }
1181
1182 v4l2_fwnode_endpoint_free(&vep);
1183
1184 priv->txports[nport] = txport;
1185
1186 return 0;
1187
1188 err_free_vep:
1189 v4l2_fwnode_endpoint_free(&vep);
1190 err_free_txport:
1191 kfree(txport);
1192
1193 return ret;
1194 }
1195
ub960_csi_handle_events(struct ub960_data * priv,u8 nport)1196 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1197 {
1198 struct device *dev = &priv->client->dev;
1199 u8 csi_tx_isr;
1200 int ret;
1201
1202 ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1203 if (ret)
1204 return;
1205
1206 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1207 dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1208
1209 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1210 dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1211 }
1212
1213 /* -----------------------------------------------------------------------------
1214 * RX ports
1215 */
1216
ub960_rxport_enable_vpocs(struct ub960_data * priv)1217 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1218 {
1219 unsigned int nport;
1220 int ret;
1221
1222 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1223 struct ub960_rxport *rxport = priv->rxports[nport];
1224
1225 if (!rxport || !rxport->vpoc)
1226 continue;
1227
1228 ret = regulator_enable(rxport->vpoc);
1229 if (ret)
1230 goto err_disable_vpocs;
1231 }
1232
1233 return 0;
1234
1235 err_disable_vpocs:
1236 while (nport--) {
1237 struct ub960_rxport *rxport = priv->rxports[nport];
1238
1239 if (!rxport || !rxport->vpoc)
1240 continue;
1241
1242 regulator_disable(rxport->vpoc);
1243 }
1244
1245 return ret;
1246 }
1247
ub960_rxport_disable_vpocs(struct ub960_data * priv)1248 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1249 {
1250 unsigned int nport;
1251
1252 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1253 struct ub960_rxport *rxport = priv->rxports[nport];
1254
1255 if (!rxport || !rxport->vpoc)
1256 continue;
1257
1258 regulator_disable(rxport->vpoc);
1259 }
1260 }
1261
ub960_rxport_clear_errors(struct ub960_data * priv,unsigned int nport)1262 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1263 unsigned int nport)
1264 {
1265 u8 v;
1266
1267 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1268 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1269 ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1270 ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1271
1272 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1273 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1274
1275 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1276 }
1277
ub960_clear_rx_errors(struct ub960_data * priv)1278 static void ub960_clear_rx_errors(struct ub960_data *priv)
1279 {
1280 unsigned int nport;
1281
1282 for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1283 ub960_rxport_clear_errors(priv, nport);
1284 }
1285
ub960_rxport_get_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 * strobe_pos)1286 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1287 unsigned int nport, s8 *strobe_pos)
1288 {
1289 u8 v;
1290 u8 clk_delay, data_delay;
1291 int ret;
1292
1293 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1294 UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1295
1296 clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1297 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1298
1299 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1300 UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1301
1302 data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1303 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1304
1305 ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1306 if (ret)
1307 return ret;
1308
1309 clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1310
1311 ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1312 if (ret)
1313 return ret;
1314
1315 data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1316
1317 *strobe_pos = data_delay - clk_delay;
1318
1319 return 0;
1320 }
1321
ub960_rxport_set_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 strobe_pos)1322 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1323 unsigned int nport, s8 strobe_pos)
1324 {
1325 u8 clk_delay, data_delay;
1326
1327 clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1328 data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1329
1330 if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1331 clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1332 else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1333 data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1334 else if (strobe_pos < 0)
1335 clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1336 else if (strobe_pos > 0)
1337 data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1338
1339 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1340 UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1341
1342 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1343 UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1344 }
1345
ub960_rxport_set_strobe_range(struct ub960_data * priv,s8 strobe_min,s8 strobe_max)1346 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1347 s8 strobe_min, s8 strobe_max)
1348 {
1349 /* Convert the signed strobe pos to positive zero based value */
1350 strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1351 strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1352
1353 ub960_write(priv, UB960_XR_SFILTER_CFG,
1354 ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1355 ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1356 }
1357
ub960_rxport_get_eq_level(struct ub960_data * priv,unsigned int nport,u8 * eq_level)1358 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1359 unsigned int nport, u8 *eq_level)
1360 {
1361 int ret;
1362 u8 v;
1363
1364 ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1365 if (ret)
1366 return ret;
1367
1368 *eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1369 (v & UB960_RR_AEQ_STATUS_STATUS_2);
1370
1371 return 0;
1372 }
1373
ub960_rxport_set_eq_level(struct ub960_data * priv,unsigned int nport,u8 eq_level)1374 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1375 unsigned int nport, u8 eq_level)
1376 {
1377 u8 eq_stage_1_select_value, eq_stage_2_select_value;
1378 const unsigned int eq_stage_max = 7;
1379 u8 v;
1380
1381 if (eq_level <= eq_stage_max) {
1382 eq_stage_1_select_value = eq_level;
1383 eq_stage_2_select_value = 0;
1384 } else {
1385 eq_stage_1_select_value = eq_stage_max;
1386 eq_stage_2_select_value = eq_level - eq_stage_max;
1387 }
1388
1389 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1390
1391 v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1392 UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1393 v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1394 v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1395 v |= UB960_RR_AEQ_BYPASS_ENABLE;
1396
1397 ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1398 }
1399
ub960_rxport_set_eq_range(struct ub960_data * priv,unsigned int nport,u8 eq_min,u8 eq_max)1400 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1401 unsigned int nport, u8 eq_min, u8 eq_max)
1402 {
1403 ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1404 (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1405 (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1406
1407 /* Enable AEQ min setting */
1408 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1409 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1410 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1411 }
1412
ub960_rxport_config_eq(struct ub960_data * priv,unsigned int nport)1413 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1414 {
1415 struct ub960_rxport *rxport = priv->rxports[nport];
1416
1417 /* We also set common settings here. Should be moved elsewhere. */
1418
1419 if (priv->strobe.manual) {
1420 /* Disable AEQ_SFILTER_EN */
1421 ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1422 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1423 } else {
1424 /* Enable SFILTER and error control */
1425 ub960_write(priv, UB960_XR_AEQ_CTL1,
1426 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1427 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1428
1429 /* Set AEQ strobe range */
1430 ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1431 priv->strobe.max);
1432 }
1433
1434 /* The rest are port specific */
1435
1436 if (priv->strobe.manual)
1437 ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1438 else
1439 ub960_rxport_set_strobe_pos(priv, nport, 0);
1440
1441 if (rxport->eq.manual_eq) {
1442 ub960_rxport_set_eq_level(priv, nport,
1443 rxport->eq.manual.eq_level);
1444
1445 /* Enable AEQ Bypass */
1446 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1447 UB960_RR_AEQ_BYPASS_ENABLE,
1448 UB960_RR_AEQ_BYPASS_ENABLE);
1449 } else {
1450 ub960_rxport_set_eq_range(priv, nport,
1451 rxport->eq.aeq.eq_level_min,
1452 rxport->eq.aeq.eq_level_max);
1453
1454 /* Disable AEQ Bypass */
1455 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1456 UB960_RR_AEQ_BYPASS_ENABLE, 0);
1457 }
1458 }
1459
ub960_rxport_link_ok(struct ub960_data * priv,unsigned int nport,bool * ok)1460 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1461 bool *ok)
1462 {
1463 u8 rx_port_sts1, rx_port_sts2;
1464 u16 parity_errors;
1465 u8 csi_rx_sts;
1466 u8 csi_err_cnt;
1467 u8 bcc_sts;
1468 int ret;
1469 bool errors;
1470
1471 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1472 &rx_port_sts1);
1473 if (ret)
1474 return ret;
1475
1476 if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1477 *ok = false;
1478 return 0;
1479 }
1480
1481 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1482 &rx_port_sts2);
1483 if (ret)
1484 return ret;
1485
1486 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1487 if (ret)
1488 return ret;
1489
1490 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1491 &csi_err_cnt);
1492 if (ret)
1493 return ret;
1494
1495 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1496 if (ret)
1497 return ret;
1498
1499 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1500 &parity_errors);
1501 if (ret)
1502 return ret;
1503
1504 errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1505 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1506 (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1507 (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1508 parity_errors;
1509
1510 *ok = !errors;
1511
1512 return 0;
1513 }
1514
1515 /*
1516 * Wait for the RX ports to lock, have no errors and have stable strobe position
1517 * and EQ level.
1518 */
ub960_rxport_wait_locks(struct ub960_data * priv,unsigned long port_mask,unsigned int * lock_mask)1519 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1520 unsigned long port_mask,
1521 unsigned int *lock_mask)
1522 {
1523 struct device *dev = &priv->client->dev;
1524 unsigned long timeout;
1525 unsigned int link_ok_mask;
1526 unsigned int missing;
1527 unsigned int loops;
1528 u8 nport;
1529 int ret;
1530
1531 if (port_mask == 0) {
1532 if (lock_mask)
1533 *lock_mask = 0;
1534 return 0;
1535 }
1536
1537 if (port_mask >= BIT(priv->hw_data->num_rxports))
1538 return -EINVAL;
1539
1540 timeout = jiffies + msecs_to_jiffies(1000);
1541 loops = 0;
1542 link_ok_mask = 0;
1543
1544 while (time_before(jiffies, timeout)) {
1545 missing = 0;
1546
1547 for_each_set_bit(nport, &port_mask,
1548 priv->hw_data->num_rxports) {
1549 struct ub960_rxport *rxport = priv->rxports[nport];
1550 bool ok;
1551
1552 if (!rxport)
1553 continue;
1554
1555 ret = ub960_rxport_link_ok(priv, nport, &ok);
1556 if (ret)
1557 return ret;
1558
1559 /*
1560 * We want the link to be ok for two consecutive loops,
1561 * as a link could get established just before our test
1562 * and drop soon after.
1563 */
1564 if (!ok || !(link_ok_mask & BIT(nport)))
1565 missing++;
1566
1567 if (ok)
1568 link_ok_mask |= BIT(nport);
1569 else
1570 link_ok_mask &= ~BIT(nport);
1571 }
1572
1573 loops++;
1574
1575 if (missing == 0)
1576 break;
1577
1578 /*
1579 * The sleep time of 10 ms was found by testing to give a lock
1580 * with a few iterations. It can be decreased if on some setups
1581 * the lock can be achieved much faster.
1582 */
1583 fsleep(10 * USEC_PER_MSEC);
1584 }
1585
1586 if (lock_mask)
1587 *lock_mask = link_ok_mask;
1588
1589 dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1590 for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1591 struct ub960_rxport *rxport = priv->rxports[nport];
1592 s8 strobe_pos, eq_level;
1593 u16 v;
1594
1595 if (!rxport)
1596 continue;
1597
1598 if (!(link_ok_mask & BIT(nport))) {
1599 dev_dbg(dev, "\trx%u: not locked\n", nport);
1600 continue;
1601 }
1602
1603 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1604
1605 if (priv->hw_data->is_ub9702) {
1606 dev_dbg(dev, "\trx%u: locked, freq %llu Hz\n",
1607 nport, ((u64)v * HZ_PER_MHZ) >> 8);
1608 } else {
1609 ret = ub960_rxport_get_strobe_pos(priv, nport,
1610 &strobe_pos);
1611 if (ret)
1612 return ret;
1613
1614 ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1615 if (ret)
1616 return ret;
1617
1618 dev_dbg(dev,
1619 "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1620 nport, strobe_pos, eq_level,
1621 ((u64)v * HZ_PER_MHZ) >> 8);
1622 }
1623 }
1624
1625 return 0;
1626 }
1627
ub960_calc_bc_clk_rate_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1628 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1629 struct ub960_rxport *rxport)
1630 {
1631 unsigned int mult;
1632 unsigned int div;
1633
1634 switch (rxport->rx_mode) {
1635 case RXPORT_MODE_RAW10:
1636 case RXPORT_MODE_RAW12_HF:
1637 case RXPORT_MODE_RAW12_LF:
1638 mult = 1;
1639 div = 10;
1640 break;
1641
1642 case RXPORT_MODE_CSI2_SYNC:
1643 mult = 2;
1644 div = 1;
1645 break;
1646
1647 case RXPORT_MODE_CSI2_NONSYNC:
1648 mult = 2;
1649 div = 5;
1650 break;
1651
1652 default:
1653 return 0;
1654 }
1655
1656 return clk_get_rate(priv->refclk) * mult / div;
1657 }
1658
ub960_calc_bc_clk_rate_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)1659 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1660 struct ub960_rxport *rxport)
1661 {
1662 switch (rxport->rx_mode) {
1663 case RXPORT_MODE_RAW10:
1664 case RXPORT_MODE_RAW12_HF:
1665 case RXPORT_MODE_RAW12_LF:
1666 return 2359400;
1667
1668 case RXPORT_MODE_CSI2_SYNC:
1669 return 47187500;
1670
1671 case RXPORT_MODE_CSI2_NONSYNC:
1672 return 9437500;
1673
1674 default:
1675 return 0;
1676 }
1677 }
1678
ub960_rxport_add_serializer(struct ub960_data * priv,u8 nport)1679 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1680 {
1681 struct ub960_rxport *rxport = priv->rxports[nport];
1682 struct device *dev = &priv->client->dev;
1683 struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1684 struct i2c_board_info ser_info = {
1685 .of_node = to_of_node(rxport->ser.fwnode),
1686 .fwnode = rxport->ser.fwnode,
1687 .platform_data = ser_pdata,
1688 };
1689
1690 ser_pdata->port = nport;
1691 ser_pdata->atr = priv->atr;
1692 if (priv->hw_data->is_ub9702)
1693 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1694 else
1695 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1696
1697 /*
1698 * The serializer is added under the same i2c adapter as the
1699 * deserializer. This is not quite right, as the serializer is behind
1700 * the FPD-Link.
1701 */
1702 ser_info.addr = rxport->ser.alias;
1703 rxport->ser.client =
1704 i2c_new_client_device(priv->client->adapter, &ser_info);
1705 if (IS_ERR(rxport->ser.client)) {
1706 dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1707 ser_info.type);
1708 return PTR_ERR(rxport->ser.client);
1709 }
1710
1711 dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1712 nport, rxport->ser.client->addr,
1713 rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1714
1715 return 0;
1716 }
1717
ub960_rxport_remove_serializer(struct ub960_data * priv,u8 nport)1718 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1719 {
1720 struct ub960_rxport *rxport = priv->rxports[nport];
1721
1722 i2c_unregister_device(rxport->ser.client);
1723 rxport->ser.client = NULL;
1724 }
1725
1726 /* Add serializer i2c devices for all initialized ports */
ub960_rxport_add_serializers(struct ub960_data * priv)1727 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1728 {
1729 unsigned int nport;
1730 int ret;
1731
1732 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1733 struct ub960_rxport *rxport = priv->rxports[nport];
1734
1735 if (!rxport)
1736 continue;
1737
1738 ret = ub960_rxport_add_serializer(priv, nport);
1739 if (ret)
1740 goto err_remove_sers;
1741 }
1742
1743 return 0;
1744
1745 err_remove_sers:
1746 while (nport--) {
1747 struct ub960_rxport *rxport = priv->rxports[nport];
1748
1749 if (!rxport)
1750 continue;
1751
1752 ub960_rxport_remove_serializer(priv, nport);
1753 }
1754
1755 return ret;
1756 }
1757
ub960_rxport_remove_serializers(struct ub960_data * priv)1758 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1759 {
1760 unsigned int nport;
1761
1762 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1763 struct ub960_rxport *rxport = priv->rxports[nport];
1764
1765 if (!rxport)
1766 continue;
1767
1768 ub960_rxport_remove_serializer(priv, nport);
1769 }
1770 }
1771
ub960_init_tx_port(struct ub960_data * priv,struct ub960_txport * txport)1772 static void ub960_init_tx_port(struct ub960_data *priv,
1773 struct ub960_txport *txport)
1774 {
1775 unsigned int nport = txport->nport;
1776 u8 csi_ctl = 0;
1777
1778 /*
1779 * From the datasheet: "initial CSI Skew-Calibration
1780 * sequence [...] should be set when operating at 1.6 Gbps"
1781 */
1782 if (priv->tx_data_rate == MHZ(1600))
1783 csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1784
1785 csi_ctl |= (4 - txport->num_data_lanes) << 4;
1786
1787 if (!txport->non_continous_clk)
1788 csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK;
1789
1790 ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1791 }
1792
ub960_init_tx_ports(struct ub960_data * priv)1793 static int ub960_init_tx_ports(struct ub960_data *priv)
1794 {
1795 unsigned int nport;
1796 u8 speed_select;
1797 u8 pll_div;
1798
1799 /* TX ports */
1800
1801 switch (priv->tx_data_rate) {
1802 case MHZ(1600):
1803 default:
1804 speed_select = 0;
1805 pll_div = 0x10;
1806 break;
1807 case MHZ(1200):
1808 speed_select = 1;
1809 pll_div = 0x18;
1810 break;
1811 case MHZ(800):
1812 speed_select = 2;
1813 pll_div = 0x10;
1814 break;
1815 case MHZ(400):
1816 speed_select = 3;
1817 pll_div = 0x10;
1818 break;
1819 }
1820
1821 ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1822
1823 if (priv->hw_data->is_ub9702) {
1824 ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1825
1826 switch (priv->tx_data_rate) {
1827 case MHZ(1600):
1828 default:
1829 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1830 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1831 break;
1832 case MHZ(800):
1833 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1834 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1835 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1836 break;
1837 case MHZ(400):
1838 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1839 break;
1840 }
1841 }
1842
1843 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1844 struct ub960_txport *txport = priv->txports[nport];
1845
1846 if (!txport)
1847 continue;
1848
1849 ub960_init_tx_port(priv, txport);
1850 }
1851
1852 return 0;
1853 }
1854
ub960_init_rx_port_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1855 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1856 struct ub960_rxport *rxport)
1857 {
1858 unsigned int nport = rxport->nport;
1859 u32 bc_freq_val;
1860
1861 /*
1862 * Back channel frequency select.
1863 * Override FREQ_SELECT from the strap.
1864 * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1865 * 2 - 10 Mbps
1866 * 6 - 50 Mbps (DS90UB953-Q1)
1867 *
1868 * Note that changing this setting will result in some errors on the back
1869 * channel for a short period of time.
1870 */
1871
1872 switch (rxport->rx_mode) {
1873 case RXPORT_MODE_RAW10:
1874 case RXPORT_MODE_RAW12_HF:
1875 case RXPORT_MODE_RAW12_LF:
1876 bc_freq_val = 0;
1877 break;
1878
1879 case RXPORT_MODE_CSI2_NONSYNC:
1880 bc_freq_val = 2;
1881 break;
1882
1883 case RXPORT_MODE_CSI2_SYNC:
1884 bc_freq_val = 6;
1885 break;
1886
1887 default:
1888 return;
1889 }
1890
1891 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1892 UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1893 bc_freq_val);
1894
1895 switch (rxport->rx_mode) {
1896 case RXPORT_MODE_RAW10:
1897 /* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1898 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1899 UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1900 0x3);
1901
1902 /*
1903 * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1904 */
1905 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1906 UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1907 0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1908
1909 break;
1910
1911 case RXPORT_MODE_RAW12_HF:
1912 case RXPORT_MODE_RAW12_LF:
1913 /* Not implemented */
1914 return;
1915
1916 case RXPORT_MODE_CSI2_SYNC:
1917 case RXPORT_MODE_CSI2_NONSYNC:
1918 /* CSI-2 Mode (DS90UB953-Q1 compatible) */
1919 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1920 0x0);
1921
1922 break;
1923 }
1924
1925 /* LV_POLARITY & FV_POLARITY */
1926 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1927 rxport->lv_fv_pol);
1928
1929 /* Enable all interrupt sources from this port */
1930 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1931 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1932
1933 /* Enable I2C_PASS_THROUGH */
1934 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1935 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1936 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1937
1938 /* Enable I2C communication to the serializer via the alias addr */
1939 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1940 rxport->ser.alias << 1);
1941
1942 /* Configure EQ related settings */
1943 ub960_rxport_config_eq(priv, nport);
1944
1945 /* Enable RX port */
1946 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1947 }
1948
ub960_init_rx_port_ub9702_fpd3(struct ub960_data * priv,struct ub960_rxport * rxport)1949 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1950 struct ub960_rxport *rxport)
1951 {
1952 unsigned int nport = rxport->nport;
1953 u8 bc_freq_val;
1954 u8 fpd_func_mode;
1955
1956 switch (rxport->rx_mode) {
1957 case RXPORT_MODE_RAW10:
1958 bc_freq_val = 0;
1959 fpd_func_mode = 5;
1960 break;
1961
1962 case RXPORT_MODE_RAW12_HF:
1963 bc_freq_val = 0;
1964 fpd_func_mode = 4;
1965 break;
1966
1967 case RXPORT_MODE_RAW12_LF:
1968 bc_freq_val = 0;
1969 fpd_func_mode = 6;
1970 break;
1971
1972 case RXPORT_MODE_CSI2_SYNC:
1973 bc_freq_val = 6;
1974 fpd_func_mode = 2;
1975 break;
1976
1977 case RXPORT_MODE_CSI2_NONSYNC:
1978 bc_freq_val = 2;
1979 fpd_func_mode = 2;
1980 break;
1981
1982 default:
1983 return;
1984 }
1985
1986 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1987 bc_freq_val);
1988 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1989
1990 /* set serdes_eq_mode = 1 */
1991 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1992
1993 /* enable serdes driver */
1994 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1995
1996 /* set serdes_eq_offset=4 */
1997 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1998
1999 /* init default serdes_eq_max in 0xa9 */
2000 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
2001
2002 /* init serdes_eq_min in 0xaa */
2003 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
2004
2005 /* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
2006 ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
2007 BIT(3), BIT(3));
2008
2009 /* RX port to half-rate */
2010 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2011 BIT(nport * 2));
2012 }
2013
ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data * priv,struct ub960_rxport * rxport)2014 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
2015 struct ub960_rxport *rxport)
2016 {
2017 unsigned int nport = rxport->nport;
2018 bool first_time_power_up = true;
2019
2020 if (first_time_power_up) {
2021 u8 v;
2022
2023 /* AEQ init */
2024 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
2025
2026 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
2027 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
2028
2029 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
2030 }
2031
2032 /* enable serdes_eq_ctl2 */
2033 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
2034
2035 /* enable serdes_eq_ctl1 */
2036 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
2037
2038 /* enable serdes_eq_en */
2039 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
2040
2041 /* disable serdes_eq_override */
2042 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
2043
2044 /* disable serdes_gain_override */
2045 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2046 }
2047
ub960_init_rx_port_ub9702_fpd4(struct ub960_data * priv,struct ub960_rxport * rxport)2048 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2049 struct ub960_rxport *rxport)
2050 {
2051 unsigned int nport = rxport->nport;
2052 u8 bc_freq_val;
2053
2054 switch (rxport->rx_mode) {
2055 case RXPORT_MODE_RAW10:
2056 bc_freq_val = 0;
2057 break;
2058
2059 case RXPORT_MODE_RAW12_HF:
2060 bc_freq_val = 0;
2061 break;
2062
2063 case RXPORT_MODE_RAW12_LF:
2064 bc_freq_val = 0;
2065 break;
2066
2067 case RXPORT_MODE_CSI2_SYNC:
2068 bc_freq_val = 6;
2069 break;
2070
2071 case RXPORT_MODE_CSI2_NONSYNC:
2072 bc_freq_val = 2;
2073 break;
2074
2075 default:
2076 return;
2077 }
2078
2079 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2080 bc_freq_val);
2081
2082 /* FPD4 Sync Mode */
2083 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2084
2085 /* add serdes_eq_offset of 4 */
2086 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2087
2088 /* FPD4 serdes_start_eq in 0x27: assign default */
2089 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2090 /* FPD4 serdes_end_eq in 0x28: assign default */
2091 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2092
2093 /* set serdes_driver_mode into FPD IV mode */
2094 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2095 /* set FPD PBC drv into FPD IV mode */
2096 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2097
2098 /* set serdes_system_init to 0x2f */
2099 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2100 /* set serdes_system_rst in reset mode */
2101 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2102
2103 /* RX port to 7.55G mode */
2104 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2105 0 << (nport * 2));
2106
2107 ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2108 }
2109
ub960_init_rx_port_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)2110 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2111 struct ub960_rxport *rxport)
2112 {
2113 unsigned int nport = rxport->nport;
2114
2115 if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2116 ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2117 else /* RXPORT_CDR_FPD4 */
2118 ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2119
2120 switch (rxport->rx_mode) {
2121 case RXPORT_MODE_RAW10:
2122 /*
2123 * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2124 * 0b10 : 8-bit processing using upper 8 bits
2125 */
2126 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2127 0x3 << 6, 0x2 << 6);
2128
2129 break;
2130
2131 case RXPORT_MODE_RAW12_HF:
2132 case RXPORT_MODE_RAW12_LF:
2133 /* Not implemented */
2134 return;
2135
2136 case RXPORT_MODE_CSI2_SYNC:
2137 case RXPORT_MODE_CSI2_NONSYNC:
2138
2139 break;
2140 }
2141
2142 /* LV_POLARITY & FV_POLARITY */
2143 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2144 rxport->lv_fv_pol);
2145
2146 /* Enable all interrupt sources from this port */
2147 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2148 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2149
2150 /* Enable I2C_PASS_THROUGH */
2151 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2152 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2153 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2154
2155 /* Enable I2C communication to the serializer via the alias addr */
2156 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2157 rxport->ser.alias << 1);
2158
2159 /* Enable RX port */
2160 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2161
2162 if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2163 /* unreset 960 AEQ */
2164 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2165 }
2166 }
2167
ub960_init_rx_ports(struct ub960_data * priv)2168 static int ub960_init_rx_ports(struct ub960_data *priv)
2169 {
2170 unsigned int nport;
2171
2172 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2173 struct ub960_rxport *rxport = priv->rxports[nport];
2174
2175 if (!rxport)
2176 continue;
2177
2178 if (priv->hw_data->is_ub9702)
2179 ub960_init_rx_port_ub9702(priv, rxport);
2180 else
2181 ub960_init_rx_port_ub960(priv, rxport);
2182 }
2183
2184 return 0;
2185 }
2186
ub960_rxport_handle_events(struct ub960_data * priv,u8 nport)2187 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2188 {
2189 struct device *dev = &priv->client->dev;
2190 u8 rx_port_sts1;
2191 u8 rx_port_sts2;
2192 u8 csi_rx_sts;
2193 u8 bcc_sts;
2194 int ret = 0;
2195
2196 /* Read interrupts (also clears most of them) */
2197 if (!ret)
2198 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2199 &rx_port_sts1);
2200 if (!ret)
2201 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2202 &rx_port_sts2);
2203 if (!ret)
2204 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2205 &csi_rx_sts);
2206 if (!ret)
2207 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2208 &bcc_sts);
2209
2210 if (ret)
2211 return;
2212
2213 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2214 u16 v;
2215
2216 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2217 &v);
2218 if (!ret)
2219 dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2220 }
2221
2222 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2223 dev_err(dev, "rx%u BCC CRC error\n", nport);
2224
2225 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2226 dev_err(dev, "rx%u BCC SEQ error\n", nport);
2227
2228 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2229 dev_err(dev, "rx%u line length unstable\n", nport);
2230
2231 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2232 dev_err(dev, "rx%u FPD3 encode error\n", nport);
2233
2234 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2235 dev_err(dev, "rx%u buffer error\n", nport);
2236
2237 if (csi_rx_sts)
2238 dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2239
2240 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2241 dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2242
2243 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2244 dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2245
2246 if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2247 dev_err(dev, "rx%u CSI checksum error\n", nport);
2248
2249 if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2250 dev_err(dev, "rx%u CSI length error\n", nport);
2251
2252 if (bcc_sts)
2253 dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2254
2255 if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2256 dev_err(dev, "rx%u BCC response error", nport);
2257
2258 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2259 dev_err(dev, "rx%u BCC slave timeout", nport);
2260
2261 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2262 dev_err(dev, "rx%u BCC slave error", nport);
2263
2264 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2265 dev_err(dev, "rx%u BCC master timeout", nport);
2266
2267 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2268 dev_err(dev, "rx%u BCC master error", nport);
2269
2270 if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2271 dev_err(dev, "rx%u BCC sequence error", nport);
2272
2273 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2274 u16 v;
2275
2276 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2277 if (!ret)
2278 dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2279 }
2280
2281 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2282 u16 v;
2283
2284 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2285 &v);
2286 if (!ret)
2287 dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2288 }
2289
2290 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2291 dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2292 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2293 "locked" :
2294 "unlocked",
2295 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2296 "passed" :
2297 "not passed",
2298 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2299 "no clock" :
2300 "clock ok",
2301 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2302 "stable freq" :
2303 "unstable freq");
2304 }
2305 }
2306
2307 /* -----------------------------------------------------------------------------
2308 * V4L2
2309 */
2310
2311 /*
2312 * The current implementation only supports a simple VC mapping, where all VCs
2313 * from a one RX port will be mapped to the same VC. Also, the hardware
2314 * dictates that all streams from an RX port must go to a single TX port.
2315 *
2316 * This function decides the target VC numbers for each RX port with a simple
2317 * algorithm, so that for each TX port, we get VC numbers starting from 0,
2318 * and counting up.
2319 *
2320 * E.g. if all four RX ports are in use, of which the first two go to the
2321 * first TX port and the secont two go to the second TX port, we would get
2322 * the following VCs for the four RX ports: 0, 1, 0, 1.
2323 *
2324 * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2325 * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2326 * probably needs to be somehow configurable. Device tree?
2327 */
ub960_get_vc_maps(struct ub960_data * priv,struct v4l2_subdev_state * state,u8 * vc)2328 static void ub960_get_vc_maps(struct ub960_data *priv,
2329 struct v4l2_subdev_state *state, u8 *vc)
2330 {
2331 u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2332 struct v4l2_subdev_route *route;
2333 u8 handled_mask = 0;
2334
2335 for_each_active_route(&state->routing, route) {
2336 unsigned int rx, tx;
2337
2338 rx = ub960_pad_to_port(priv, route->sink_pad);
2339 if (BIT(rx) & handled_mask)
2340 continue;
2341
2342 tx = ub960_pad_to_port(priv, route->source_pad);
2343
2344 vc[rx] = cur_vc[tx]++;
2345 handled_mask |= BIT(rx);
2346 }
2347 }
2348
ub960_enable_tx_port(struct ub960_data * priv,unsigned int nport)2349 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2350 {
2351 struct device *dev = &priv->client->dev;
2352
2353 dev_dbg(dev, "enable TX port %u\n", nport);
2354
2355 return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2356 UB960_TR_CSI_CTL_CSI_ENABLE,
2357 UB960_TR_CSI_CTL_CSI_ENABLE);
2358 }
2359
ub960_disable_tx_port(struct ub960_data * priv,unsigned int nport)2360 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2361 {
2362 struct device *dev = &priv->client->dev;
2363
2364 dev_dbg(dev, "disable TX port %u\n", nport);
2365
2366 ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2367 UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2368 }
2369
ub960_enable_rx_port(struct ub960_data * priv,unsigned int nport)2370 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2371 {
2372 struct device *dev = &priv->client->dev;
2373
2374 dev_dbg(dev, "enable RX port %u\n", nport);
2375
2376 /* Enable forwarding */
2377 return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2378 UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2379 }
2380
ub960_disable_rx_port(struct ub960_data * priv,unsigned int nport)2381 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2382 {
2383 struct device *dev = &priv->client->dev;
2384
2385 dev_dbg(dev, "disable RX port %u\n", nport);
2386
2387 /* Disable forwarding */
2388 ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2389 UB960_SR_FWD_CTL1_PORT_DIS(nport),
2390 UB960_SR_FWD_CTL1_PORT_DIS(nport));
2391 }
2392
2393 /*
2394 * The driver only supports using a single VC for each source. This function
2395 * checks that each source only provides streams using a single VC.
2396 */
ub960_validate_stream_vcs(struct ub960_data * priv)2397 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2398 {
2399 unsigned int nport;
2400 unsigned int i;
2401
2402 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2403 struct ub960_rxport *rxport = priv->rxports[nport];
2404 struct v4l2_mbus_frame_desc desc;
2405 int ret;
2406 u8 vc;
2407
2408 if (!rxport)
2409 continue;
2410
2411 ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2412 rxport->source.pad, &desc);
2413 if (ret)
2414 return ret;
2415
2416 if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2417 continue;
2418
2419 if (desc.num_entries == 0)
2420 continue;
2421
2422 vc = desc.entry[0].bus.csi2.vc;
2423
2424 for (i = 1; i < desc.num_entries; i++) {
2425 if (vc == desc.entry[i].bus.csi2.vc)
2426 continue;
2427
2428 dev_err(&priv->client->dev,
2429 "rx%u: source with multiple virtual-channels is not supported\n",
2430 nport);
2431 return -ENODEV;
2432 }
2433 }
2434
2435 return 0;
2436 }
2437
ub960_configure_ports_for_streaming(struct ub960_data * priv,struct v4l2_subdev_state * state)2438 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2439 struct v4l2_subdev_state *state)
2440 {
2441 u8 fwd_ctl;
2442 struct {
2443 u32 num_streams;
2444 u8 pixel_dt;
2445 u8 meta_dt;
2446 u32 meta_lines;
2447 u32 tx_port;
2448 } rx_data[UB960_MAX_RX_NPORTS] = {};
2449 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2450 struct v4l2_subdev_route *route;
2451 int ret;
2452
2453 ret = ub960_validate_stream_vcs(priv);
2454 if (ret)
2455 return ret;
2456
2457 ub960_get_vc_maps(priv, state, vc_map);
2458
2459 for_each_active_route(&state->routing, route) {
2460 struct ub960_rxport *rxport;
2461 struct ub960_txport *txport;
2462 struct v4l2_mbus_framefmt *fmt;
2463 const struct ub960_format_info *ub960_fmt;
2464 unsigned int nport;
2465
2466 nport = ub960_pad_to_port(priv, route->sink_pad);
2467
2468 rxport = priv->rxports[nport];
2469 if (!rxport)
2470 return -EINVAL;
2471
2472 txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2473 if (!txport)
2474 return -EINVAL;
2475
2476 rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2477
2478 rx_data[nport].num_streams++;
2479
2480 /* For the rest, we are only interested in parallel busses */
2481 if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2482 rxport->rx_mode == RXPORT_MODE_CSI2_NONSYNC)
2483 continue;
2484
2485 if (rx_data[nport].num_streams > 2)
2486 return -EPIPE;
2487
2488 fmt = v4l2_subdev_state_get_format(state, route->sink_pad,
2489 route->sink_stream);
2490 if (!fmt)
2491 return -EPIPE;
2492
2493 ub960_fmt = ub960_find_format(fmt->code);
2494 if (!ub960_fmt)
2495 return -EPIPE;
2496
2497 if (ub960_fmt->meta) {
2498 if (fmt->height > 3) {
2499 dev_err(&priv->client->dev,
2500 "rx%u: unsupported metadata height %u\n",
2501 nport, fmt->height);
2502 return -EPIPE;
2503 }
2504
2505 rx_data[nport].meta_dt = ub960_fmt->datatype;
2506 rx_data[nport].meta_lines = fmt->height;
2507 } else {
2508 rx_data[nport].pixel_dt = ub960_fmt->datatype;
2509 }
2510 }
2511
2512 /* Configure RX ports */
2513
2514 /*
2515 * Keep all port forwardings disabled by default. Forwarding will be
2516 * enabled in ub960_enable_rx_port.
2517 */
2518 fwd_ctl = GENMASK(7, 4);
2519
2520 for (unsigned int nport = 0; nport < priv->hw_data->num_rxports;
2521 nport++) {
2522 struct ub960_rxport *rxport = priv->rxports[nport];
2523 u8 vc = vc_map[nport];
2524
2525 if (rx_data[nport].num_streams == 0)
2526 continue;
2527
2528 switch (rxport->rx_mode) {
2529 case RXPORT_MODE_RAW10:
2530 ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2531 rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2532
2533 ub960_rxport_write(priv, rxport->nport,
2534 UB960_RR_RAW_EMBED_DTYPE,
2535 (rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2536 rx_data[nport].meta_dt);
2537
2538 break;
2539
2540 case RXPORT_MODE_RAW12_HF:
2541 case RXPORT_MODE_RAW12_LF:
2542 /* Not implemented */
2543 break;
2544
2545 case RXPORT_MODE_CSI2_SYNC:
2546 case RXPORT_MODE_CSI2_NONSYNC:
2547 if (!priv->hw_data->is_ub9702) {
2548 /* Map all VCs from this port to the same VC */
2549 ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2550 (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2551 (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2552 (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2553 (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2554 } else {
2555 unsigned int i;
2556
2557 /* Map all VCs from this port to VC(nport) */
2558 for (i = 0; i < 8; i++)
2559 ub960_rxport_write(priv, nport,
2560 UB960_RR_VC_ID_MAP(i),
2561 (nport << 4) | nport);
2562 }
2563
2564 break;
2565 }
2566
2567 if (rx_data[nport].tx_port == 1)
2568 fwd_ctl |= BIT(nport); /* forward to TX1 */
2569 else
2570 fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2571 }
2572
2573 ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2574
2575 return 0;
2576 }
2577
ub960_update_streaming_status(struct ub960_data * priv)2578 static void ub960_update_streaming_status(struct ub960_data *priv)
2579 {
2580 unsigned int i;
2581
2582 for (i = 0; i < UB960_MAX_NPORTS; i++) {
2583 if (priv->stream_enable_mask[i])
2584 break;
2585 }
2586
2587 priv->streaming = i < UB960_MAX_NPORTS;
2588 }
2589
ub960_enable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2590 static int ub960_enable_streams(struct v4l2_subdev *sd,
2591 struct v4l2_subdev_state *state, u32 source_pad,
2592 u64 source_streams_mask)
2593 {
2594 struct ub960_data *priv = sd_to_ub960(sd);
2595 struct device *dev = &priv->client->dev;
2596 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2597 struct v4l2_subdev_route *route;
2598 unsigned int failed_port;
2599 unsigned int nport;
2600 int ret;
2601
2602 if (!priv->streaming) {
2603 dev_dbg(dev, "Prepare for streaming\n");
2604 ret = ub960_configure_ports_for_streaming(priv, state);
2605 if (ret)
2606 return ret;
2607 }
2608
2609 /* Enable TX port if not yet enabled */
2610 if (!priv->stream_enable_mask[source_pad]) {
2611 ret = ub960_enable_tx_port(priv,
2612 ub960_pad_to_port(priv, source_pad));
2613 if (ret)
2614 return ret;
2615 }
2616
2617 priv->stream_enable_mask[source_pad] |= source_streams_mask;
2618
2619 /* Collect sink streams per pad which we need to enable */
2620 for_each_active_route(&state->routing, route) {
2621 if (route->source_pad != source_pad)
2622 continue;
2623
2624 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2625 continue;
2626
2627 nport = ub960_pad_to_port(priv, route->sink_pad);
2628
2629 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2630 }
2631
2632 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2633 if (!sink_streams[nport])
2634 continue;
2635
2636 /* Enable the RX port if not yet enabled */
2637 if (!priv->stream_enable_mask[nport]) {
2638 ret = ub960_enable_rx_port(priv, nport);
2639 if (ret) {
2640 failed_port = nport;
2641 goto err;
2642 }
2643 }
2644
2645 priv->stream_enable_mask[nport] |= sink_streams[nport];
2646
2647 dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2648 sink_streams[nport]);
2649
2650 ret = v4l2_subdev_enable_streams(
2651 priv->rxports[nport]->source.sd,
2652 priv->rxports[nport]->source.pad,
2653 sink_streams[nport]);
2654 if (ret) {
2655 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2656
2657 if (!priv->stream_enable_mask[nport])
2658 ub960_disable_rx_port(priv, nport);
2659
2660 failed_port = nport;
2661 goto err;
2662 }
2663 }
2664
2665 priv->streaming = true;
2666
2667 return 0;
2668
2669 err:
2670 for (nport = 0; nport < failed_port; nport++) {
2671 if (!sink_streams[nport])
2672 continue;
2673
2674 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2675 sink_streams[nport]);
2676
2677 ret = v4l2_subdev_disable_streams(
2678 priv->rxports[nport]->source.sd,
2679 priv->rxports[nport]->source.pad,
2680 sink_streams[nport]);
2681 if (ret)
2682 dev_err(dev, "Failed to disable streams: %d\n", ret);
2683
2684 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2685
2686 /* Disable RX port if no active streams */
2687 if (!priv->stream_enable_mask[nport])
2688 ub960_disable_rx_port(priv, nport);
2689 }
2690
2691 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2692
2693 if (!priv->stream_enable_mask[source_pad])
2694 ub960_disable_tx_port(priv,
2695 ub960_pad_to_port(priv, source_pad));
2696
2697 ub960_update_streaming_status(priv);
2698
2699 return ret;
2700 }
2701
ub960_disable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2702 static int ub960_disable_streams(struct v4l2_subdev *sd,
2703 struct v4l2_subdev_state *state,
2704 u32 source_pad, u64 source_streams_mask)
2705 {
2706 struct ub960_data *priv = sd_to_ub960(sd);
2707 struct device *dev = &priv->client->dev;
2708 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2709 struct v4l2_subdev_route *route;
2710 unsigned int nport;
2711 int ret;
2712
2713 /* Collect sink streams per pad which we need to disable */
2714 for_each_active_route(&state->routing, route) {
2715 if (route->source_pad != source_pad)
2716 continue;
2717
2718 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2719 continue;
2720
2721 nport = ub960_pad_to_port(priv, route->sink_pad);
2722
2723 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2724 }
2725
2726 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2727 if (!sink_streams[nport])
2728 continue;
2729
2730 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2731 sink_streams[nport]);
2732
2733 ret = v4l2_subdev_disable_streams(
2734 priv->rxports[nport]->source.sd,
2735 priv->rxports[nport]->source.pad,
2736 sink_streams[nport]);
2737 if (ret)
2738 dev_err(dev, "Failed to disable streams: %d\n", ret);
2739
2740 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2741
2742 /* Disable RX port if no active streams */
2743 if (!priv->stream_enable_mask[nport])
2744 ub960_disable_rx_port(priv, nport);
2745 }
2746
2747 /* Disable TX port if no active streams */
2748
2749 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2750
2751 if (!priv->stream_enable_mask[source_pad])
2752 ub960_disable_tx_port(priv,
2753 ub960_pad_to_port(priv, source_pad));
2754
2755 ub960_update_streaming_status(priv);
2756
2757 return 0;
2758 }
2759
_ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_krouting * routing)2760 static int _ub960_set_routing(struct v4l2_subdev *sd,
2761 struct v4l2_subdev_state *state,
2762 struct v4l2_subdev_krouting *routing)
2763 {
2764 static const struct v4l2_mbus_framefmt format = {
2765 .width = 640,
2766 .height = 480,
2767 .code = MEDIA_BUS_FMT_UYVY8_1X16,
2768 .field = V4L2_FIELD_NONE,
2769 .colorspace = V4L2_COLORSPACE_SRGB,
2770 .ycbcr_enc = V4L2_YCBCR_ENC_601,
2771 .quantization = V4L2_QUANTIZATION_LIM_RANGE,
2772 .xfer_func = V4L2_XFER_FUNC_SRGB,
2773 };
2774 int ret;
2775
2776 /*
2777 * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2778 * frame desc is made dynamically allocated.
2779 */
2780
2781 if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2782 return -E2BIG;
2783
2784 ret = v4l2_subdev_routing_validate(sd, routing,
2785 V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2786 V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2787 if (ret)
2788 return ret;
2789
2790 ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2791 if (ret)
2792 return ret;
2793
2794 return 0;
2795 }
2796
ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,enum v4l2_subdev_format_whence which,struct v4l2_subdev_krouting * routing)2797 static int ub960_set_routing(struct v4l2_subdev *sd,
2798 struct v4l2_subdev_state *state,
2799 enum v4l2_subdev_format_whence which,
2800 struct v4l2_subdev_krouting *routing)
2801 {
2802 struct ub960_data *priv = sd_to_ub960(sd);
2803
2804 if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2805 return -EBUSY;
2806
2807 return _ub960_set_routing(sd, state, routing);
2808 }
2809
ub960_get_frame_desc(struct v4l2_subdev * sd,unsigned int pad,struct v4l2_mbus_frame_desc * fd)2810 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2811 struct v4l2_mbus_frame_desc *fd)
2812 {
2813 struct ub960_data *priv = sd_to_ub960(sd);
2814 struct v4l2_subdev_route *route;
2815 struct v4l2_subdev_state *state;
2816 int ret = 0;
2817 struct device *dev = &priv->client->dev;
2818 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2819
2820 if (!ub960_pad_is_source(priv, pad))
2821 return -EINVAL;
2822
2823 fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2824
2825 state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2826
2827 ub960_get_vc_maps(priv, state, vc_map);
2828
2829 for_each_active_route(&state->routing, route) {
2830 struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2831 struct v4l2_mbus_frame_desc source_fd;
2832 unsigned int nport;
2833 unsigned int i;
2834
2835 if (route->source_pad != pad)
2836 continue;
2837
2838 nport = ub960_pad_to_port(priv, route->sink_pad);
2839
2840 ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2841 get_frame_desc,
2842 priv->rxports[nport]->source.pad,
2843 &source_fd);
2844 if (ret) {
2845 dev_err(dev,
2846 "Failed to get source frame desc for pad %u\n",
2847 route->sink_pad);
2848 goto out_unlock;
2849 }
2850
2851 for (i = 0; i < source_fd.num_entries; i++) {
2852 if (source_fd.entry[i].stream == route->sink_stream) {
2853 source_entry = &source_fd.entry[i];
2854 break;
2855 }
2856 }
2857
2858 if (!source_entry) {
2859 dev_err(dev,
2860 "Failed to find stream from source frame desc\n");
2861 ret = -EPIPE;
2862 goto out_unlock;
2863 }
2864
2865 fd->entry[fd->num_entries].stream = route->source_stream;
2866 fd->entry[fd->num_entries].flags = source_entry->flags;
2867 fd->entry[fd->num_entries].length = source_entry->length;
2868 fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2869
2870 fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2871
2872 if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2873 fd->entry[fd->num_entries].bus.csi2.dt =
2874 source_entry->bus.csi2.dt;
2875 } else {
2876 const struct ub960_format_info *ub960_fmt;
2877 struct v4l2_mbus_framefmt *fmt;
2878
2879 fmt = v4l2_subdev_state_get_format(state, pad,
2880 route->source_stream);
2881
2882 if (!fmt) {
2883 ret = -EINVAL;
2884 goto out_unlock;
2885 }
2886
2887 ub960_fmt = ub960_find_format(fmt->code);
2888 if (!ub960_fmt) {
2889 dev_err(dev, "Unable to find format\n");
2890 ret = -EINVAL;
2891 goto out_unlock;
2892 }
2893
2894 fd->entry[fd->num_entries].bus.csi2.dt =
2895 ub960_fmt->datatype;
2896 }
2897
2898 fd->num_entries++;
2899 }
2900
2901 out_unlock:
2902 v4l2_subdev_unlock_state(state);
2903
2904 return ret;
2905 }
2906
ub960_set_fmt(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_format * format)2907 static int ub960_set_fmt(struct v4l2_subdev *sd,
2908 struct v4l2_subdev_state *state,
2909 struct v4l2_subdev_format *format)
2910 {
2911 struct ub960_data *priv = sd_to_ub960(sd);
2912 struct v4l2_mbus_framefmt *fmt;
2913
2914 if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2915 return -EBUSY;
2916
2917 /* No transcoding, source and sink formats must match. */
2918 if (ub960_pad_is_source(priv, format->pad))
2919 return v4l2_subdev_get_fmt(sd, state, format);
2920
2921 /*
2922 * Default to the first format if the requested media bus code isn't
2923 * supported.
2924 */
2925 if (!ub960_find_format(format->format.code))
2926 format->format.code = ub960_formats[0].code;
2927
2928 fmt = v4l2_subdev_state_get_format(state, format->pad, format->stream);
2929 if (!fmt)
2930 return -EINVAL;
2931
2932 *fmt = format->format;
2933
2934 fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2935 format->stream);
2936 if (!fmt)
2937 return -EINVAL;
2938
2939 *fmt = format->format;
2940
2941 return 0;
2942 }
2943
ub960_init_state(struct v4l2_subdev * sd,struct v4l2_subdev_state * state)2944 static int ub960_init_state(struct v4l2_subdev *sd,
2945 struct v4l2_subdev_state *state)
2946 {
2947 struct ub960_data *priv = sd_to_ub960(sd);
2948
2949 struct v4l2_subdev_route routes[] = {
2950 {
2951 .sink_pad = 0,
2952 .sink_stream = 0,
2953 .source_pad = priv->hw_data->num_rxports,
2954 .source_stream = 0,
2955 .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2956 },
2957 };
2958
2959 struct v4l2_subdev_krouting routing = {
2960 .num_routes = ARRAY_SIZE(routes),
2961 .routes = routes,
2962 };
2963
2964 return _ub960_set_routing(sd, state, &routing);
2965 }
2966
2967 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2968 .enable_streams = ub960_enable_streams,
2969 .disable_streams = ub960_disable_streams,
2970
2971 .set_routing = ub960_set_routing,
2972 .get_frame_desc = ub960_get_frame_desc,
2973
2974 .get_fmt = v4l2_subdev_get_fmt,
2975 .set_fmt = ub960_set_fmt,
2976 };
2977
ub960_log_status_ub960_sp_eq(struct ub960_data * priv,unsigned int nport)2978 static void ub960_log_status_ub960_sp_eq(struct ub960_data *priv,
2979 unsigned int nport)
2980 {
2981 struct device *dev = &priv->client->dev;
2982 u8 eq_level;
2983 s8 strobe_pos;
2984 int ret;
2985 u8 v;
2986
2987 /* Strobe */
2988
2989 ret = ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
2990 if (ret)
2991 return;
2992
2993 dev_info(dev, "\t%s strobe\n",
2994 (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
2995 "Manual");
2996
2997 if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
2998 ret = ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
2999 if (ret)
3000 return;
3001
3002 dev_info(dev, "\tStrobe range [%d, %d]\n",
3003 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
3004 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
3005 }
3006
3007 ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
3008 if (ret)
3009 return;
3010
3011 dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
3012
3013 /* EQ */
3014
3015 ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
3016 if (ret)
3017 return;
3018
3019 dev_info(dev, "\t%s EQ\n",
3020 (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
3021 "Adaptive");
3022
3023 if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
3024 ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
3025 if (ret)
3026 return;
3027
3028 dev_info(dev, "\tEQ range [%u, %u]\n",
3029 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
3030 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
3031 }
3032
3033 if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
3034 dev_info(dev, "\tEQ level %u\n", eq_level);
3035 }
3036
ub960_log_status(struct v4l2_subdev * sd)3037 static int ub960_log_status(struct v4l2_subdev *sd)
3038 {
3039 struct ub960_data *priv = sd_to_ub960(sd);
3040 struct device *dev = &priv->client->dev;
3041 struct v4l2_subdev_state *state;
3042 unsigned int nport;
3043 u16 v16 = 0;
3044 u8 v = 0;
3045 u8 id[UB960_SR_FPD3_RX_ID_LEN];
3046
3047 state = v4l2_subdev_lock_and_get_active_state(sd);
3048
3049 for (unsigned int i = 0; i < sizeof(id); i++)
3050 ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
3051
3052 dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
3053
3054 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3055 struct ub960_txport *txport = priv->txports[nport];
3056
3057 dev_info(dev, "TX %u\n", nport);
3058
3059 if (!txport) {
3060 dev_info(dev, "\tNot initialized\n");
3061 continue;
3062 }
3063
3064 ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
3065 dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
3066 v & (u8)BIT(0));
3067
3068 ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
3069 dev_info(dev, "\tframe counter %u\n", v16);
3070
3071 ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
3072 dev_info(dev, "\tframe error counter %u\n", v16);
3073
3074 ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
3075 dev_info(dev, "\tline counter %u\n", v16);
3076
3077 ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
3078 dev_info(dev, "\tline error counter %u\n", v16);
3079 }
3080
3081 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3082 struct ub960_rxport *rxport = priv->rxports[nport];
3083
3084 dev_info(dev, "RX %u\n", nport);
3085
3086 if (!rxport) {
3087 dev_info(dev, "\tNot initialized\n");
3088 continue;
3089 }
3090
3091 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
3092
3093 if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3094 dev_info(dev, "\tLocked\n");
3095 else
3096 dev_info(dev, "\tNot locked\n");
3097
3098 dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3099 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3100 dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3101
3102 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3103 dev_info(dev, "\tlink freq %llu Hz\n", ((u64)v16 * HZ_PER_MHZ) >> 8);
3104
3105 ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3106 dev_info(dev, "\tparity errors %u\n", v16);
3107
3108 ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3109 dev_info(dev, "\tlines per frame %u\n", v16);
3110
3111 ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3112 dev_info(dev, "\tbytes per line %u\n", v16);
3113
3114 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3115 dev_info(dev, "\tcsi_err_counter %u\n", v);
3116
3117 if (!priv->hw_data->is_ub9702)
3118 ub960_log_status_ub960_sp_eq(priv, nport);
3119
3120 /* GPIOs */
3121 for (unsigned int i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3122 u8 ctl_reg;
3123 u8 ctl_shift;
3124
3125 ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3126 ctl_shift = (i % 2) * 4;
3127
3128 ub960_rxport_read(priv, nport, ctl_reg, &v);
3129
3130 dev_info(dev, "\tGPIO%u: mode %u\n", i,
3131 (v >> ctl_shift) & 0xf);
3132 }
3133 }
3134
3135 v4l2_subdev_unlock_state(state);
3136
3137 return 0;
3138 }
3139
3140 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3141 .log_status = ub960_log_status,
3142 };
3143
3144 static const struct v4l2_subdev_internal_ops ub960_internal_ops = {
3145 .init_state = ub960_init_state,
3146 };
3147
3148 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3149 .core = &ub960_subdev_core_ops,
3150 .pad = &ub960_pad_ops,
3151 };
3152
3153 static const struct media_entity_operations ub960_entity_ops = {
3154 .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3155 .link_validate = v4l2_subdev_link_validate,
3156 .has_pad_interdep = v4l2_subdev_has_pad_interdep,
3157 };
3158
3159 /* -----------------------------------------------------------------------------
3160 * Core
3161 */
3162
ub960_handle_events(int irq,void * arg)3163 static irqreturn_t ub960_handle_events(int irq, void *arg)
3164 {
3165 struct ub960_data *priv = arg;
3166 unsigned int i;
3167 u8 int_sts;
3168 u8 fwd_sts;
3169 int ret;
3170
3171 ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3172 if (ret || !int_sts)
3173 return IRQ_NONE;
3174
3175 dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3176
3177 ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3178 if (ret)
3179 return IRQ_NONE;
3180
3181 dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3182
3183 for (i = 0; i < priv->hw_data->num_txports; i++) {
3184 if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3185 ub960_csi_handle_events(priv, i);
3186 }
3187
3188 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3189 if (!priv->rxports[i])
3190 continue;
3191
3192 if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3193 ub960_rxport_handle_events(priv, i);
3194 }
3195
3196 return IRQ_HANDLED;
3197 }
3198
ub960_handler_work(struct work_struct * work)3199 static void ub960_handler_work(struct work_struct *work)
3200 {
3201 struct delayed_work *dwork = to_delayed_work(work);
3202 struct ub960_data *priv =
3203 container_of(dwork, struct ub960_data, poll_work);
3204
3205 ub960_handle_events(0, priv);
3206
3207 schedule_delayed_work(&priv->poll_work,
3208 msecs_to_jiffies(UB960_POLL_TIME_MS));
3209 }
3210
ub960_txport_free_ports(struct ub960_data * priv)3211 static void ub960_txport_free_ports(struct ub960_data *priv)
3212 {
3213 unsigned int nport;
3214
3215 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3216 struct ub960_txport *txport = priv->txports[nport];
3217
3218 if (!txport)
3219 continue;
3220
3221 kfree(txport);
3222 priv->txports[nport] = NULL;
3223 }
3224 }
3225
ub960_rxport_free_ports(struct ub960_data * priv)3226 static void ub960_rxport_free_ports(struct ub960_data *priv)
3227 {
3228 unsigned int nport;
3229
3230 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3231 struct ub960_rxport *rxport = priv->rxports[nport];
3232
3233 if (!rxport)
3234 continue;
3235
3236 fwnode_handle_put(rxport->source.ep_fwnode);
3237 fwnode_handle_put(rxport->ser.fwnode);
3238
3239 kfree(rxport);
3240 priv->rxports[nport] = NULL;
3241 }
3242 }
3243
3244 static int
ub960_parse_dt_rxport_link_properties(struct ub960_data * priv,struct fwnode_handle * link_fwnode,struct ub960_rxport * rxport)3245 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3246 struct fwnode_handle *link_fwnode,
3247 struct ub960_rxport *rxport)
3248 {
3249 struct device *dev = &priv->client->dev;
3250 unsigned int nport = rxport->nport;
3251 u32 rx_mode;
3252 u32 cdr_mode;
3253 s32 strobe_pos;
3254 u32 eq_level;
3255 u32 ser_i2c_alias;
3256 int ret;
3257
3258 cdr_mode = RXPORT_CDR_FPD3;
3259
3260 ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3261 if (ret < 0 && ret != -EINVAL) {
3262 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3263 "ti,cdr-mode", ret);
3264 return ret;
3265 }
3266
3267 if (cdr_mode > RXPORT_CDR_LAST) {
3268 dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3269 return -EINVAL;
3270 }
3271
3272 if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3273 dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3274 return -EINVAL;
3275 }
3276
3277 rxport->cdr_mode = cdr_mode;
3278
3279 ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3280 if (ret < 0) {
3281 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3282 "ti,rx-mode", ret);
3283 return ret;
3284 }
3285
3286 if (rx_mode > RXPORT_MODE_LAST) {
3287 dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3288 return -EINVAL;
3289 }
3290
3291 switch (rx_mode) {
3292 case RXPORT_MODE_RAW12_HF:
3293 case RXPORT_MODE_RAW12_LF:
3294 dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3295 rx_mode);
3296 return -EINVAL;
3297 default:
3298 break;
3299 }
3300
3301 rxport->rx_mode = rx_mode;
3302
3303 /* EQ & Strobe related */
3304
3305 /* Defaults */
3306 rxport->eq.manual_eq = false;
3307 rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3308 rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3309
3310 ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3311 &strobe_pos);
3312 if (ret) {
3313 if (ret != -EINVAL) {
3314 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3315 "ti,strobe-pos", ret);
3316 return ret;
3317 }
3318 } else {
3319 if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3320 strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3321 dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3322 nport, strobe_pos);
3323 return -EINVAL;
3324 }
3325
3326 /* NOTE: ignored unless global manual strobe pos is also set */
3327 rxport->eq.strobe_pos = strobe_pos;
3328 if (!priv->strobe.manual)
3329 dev_warn(dev,
3330 "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3331 nport);
3332 }
3333
3334 ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3335 if (ret) {
3336 if (ret != -EINVAL) {
3337 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3338 "ti,eq-level", ret);
3339 return ret;
3340 }
3341 } else {
3342 if (eq_level > UB960_MAX_EQ_LEVEL) {
3343 dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3344 nport, eq_level);
3345 return -EINVAL;
3346 }
3347
3348 rxport->eq.manual_eq = true;
3349 rxport->eq.manual.eq_level = eq_level;
3350 }
3351
3352 ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3353 &ser_i2c_alias);
3354 if (ret) {
3355 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3356 "i2c-alias", ret);
3357 return ret;
3358 }
3359 rxport->ser.alias = ser_i2c_alias;
3360
3361 rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3362 if (!rxport->ser.fwnode) {
3363 dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3364 return -EINVAL;
3365 }
3366
3367 return 0;
3368 }
3369
ub960_parse_dt_rxport_ep_properties(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,struct ub960_rxport * rxport)3370 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3371 struct fwnode_handle *ep_fwnode,
3372 struct ub960_rxport *rxport)
3373 {
3374 struct device *dev = &priv->client->dev;
3375 struct v4l2_fwnode_endpoint vep = {};
3376 unsigned int nport = rxport->nport;
3377 bool hsync_hi;
3378 bool vsync_hi;
3379 int ret;
3380
3381 rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3382 if (!rxport->source.ep_fwnode) {
3383 dev_err(dev, "rx%u: no remote endpoint\n", nport);
3384 return -ENODEV;
3385 }
3386
3387 /* We currently have properties only for RAW modes */
3388
3389 switch (rxport->rx_mode) {
3390 case RXPORT_MODE_RAW10:
3391 case RXPORT_MODE_RAW12_HF:
3392 case RXPORT_MODE_RAW12_LF:
3393 break;
3394 default:
3395 return 0;
3396 }
3397
3398 vep.bus_type = V4L2_MBUS_PARALLEL;
3399 ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3400 if (ret) {
3401 dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3402 goto err_put_source_ep_fwnode;
3403 }
3404
3405 hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3406 vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3407
3408 /* LineValid and FrameValid are inverse to the h/vsync active */
3409 rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3410 (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3411
3412 return 0;
3413
3414 err_put_source_ep_fwnode:
3415 fwnode_handle_put(rxport->source.ep_fwnode);
3416 return ret;
3417 }
3418
ub960_parse_dt_rxport(struct ub960_data * priv,unsigned int nport,struct fwnode_handle * link_fwnode,struct fwnode_handle * ep_fwnode)3419 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3420 struct fwnode_handle *link_fwnode,
3421 struct fwnode_handle *ep_fwnode)
3422 {
3423 static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3424 "vpoc0", "vpoc1", "vpoc2", "vpoc3"
3425 };
3426 struct device *dev = &priv->client->dev;
3427 struct ub960_rxport *rxport;
3428 int ret;
3429
3430 rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3431 if (!rxport)
3432 return -ENOMEM;
3433
3434 priv->rxports[nport] = rxport;
3435
3436 rxport->nport = nport;
3437 rxport->priv = priv;
3438
3439 ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3440 if (ret)
3441 goto err_free_rxport;
3442
3443 rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3444 if (IS_ERR(rxport->vpoc)) {
3445 ret = PTR_ERR(rxport->vpoc);
3446 if (ret == -ENODEV) {
3447 rxport->vpoc = NULL;
3448 } else {
3449 dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3450 nport, ret);
3451 goto err_put_remote_fwnode;
3452 }
3453 }
3454
3455 ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3456 if (ret)
3457 goto err_put_remote_fwnode;
3458
3459 return 0;
3460
3461 err_put_remote_fwnode:
3462 fwnode_handle_put(rxport->ser.fwnode);
3463 err_free_rxport:
3464 priv->rxports[nport] = NULL;
3465 kfree(rxport);
3466 return ret;
3467 }
3468
3469 static struct fwnode_handle *
ub960_fwnode_get_link_by_regs(struct fwnode_handle * links_fwnode,unsigned int nport)3470 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3471 unsigned int nport)
3472 {
3473 struct fwnode_handle *link_fwnode;
3474 int ret;
3475
3476 fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3477 u32 link_num;
3478
3479 if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3480 continue;
3481
3482 ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3483 if (ret) {
3484 fwnode_handle_put(link_fwnode);
3485 return NULL;
3486 }
3487
3488 if (nport == link_num)
3489 return link_fwnode;
3490 }
3491
3492 return NULL;
3493 }
3494
ub960_parse_dt_rxports(struct ub960_data * priv)3495 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3496 {
3497 struct device *dev = &priv->client->dev;
3498 struct fwnode_handle *links_fwnode;
3499 unsigned int nport;
3500 int ret;
3501
3502 links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3503 if (!links_fwnode) {
3504 dev_err(dev, "'links' node missing\n");
3505 return -ENODEV;
3506 }
3507
3508 /* Defaults, recommended by TI */
3509 priv->strobe.min = 2;
3510 priv->strobe.max = 3;
3511
3512 priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3513
3514 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3515 struct fwnode_handle *link_fwnode;
3516 struct fwnode_handle *ep_fwnode;
3517
3518 link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3519 if (!link_fwnode)
3520 continue;
3521
3522 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3523 nport, 0, 0);
3524 if (!ep_fwnode) {
3525 fwnode_handle_put(link_fwnode);
3526 continue;
3527 }
3528
3529 ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3530 ep_fwnode);
3531
3532 fwnode_handle_put(link_fwnode);
3533 fwnode_handle_put(ep_fwnode);
3534
3535 if (ret) {
3536 dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3537 goto err_put_links;
3538 }
3539 }
3540
3541 fwnode_handle_put(links_fwnode);
3542
3543 return 0;
3544
3545 err_put_links:
3546 fwnode_handle_put(links_fwnode);
3547
3548 return ret;
3549 }
3550
ub960_parse_dt_txports(struct ub960_data * priv)3551 static int ub960_parse_dt_txports(struct ub960_data *priv)
3552 {
3553 struct device *dev = &priv->client->dev;
3554 u32 nport;
3555 int ret;
3556
3557 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3558 unsigned int port = nport + priv->hw_data->num_rxports;
3559 struct fwnode_handle *ep_fwnode;
3560
3561 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3562 port, 0, 0);
3563 if (!ep_fwnode)
3564 continue;
3565
3566 ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3567
3568 fwnode_handle_put(ep_fwnode);
3569
3570 if (ret)
3571 break;
3572 }
3573
3574 return 0;
3575 }
3576
ub960_parse_dt(struct ub960_data * priv)3577 static int ub960_parse_dt(struct ub960_data *priv)
3578 {
3579 int ret;
3580
3581 ret = ub960_parse_dt_rxports(priv);
3582 if (ret)
3583 return ret;
3584
3585 ret = ub960_parse_dt_txports(priv);
3586 if (ret)
3587 goto err_free_rxports;
3588
3589 return 0;
3590
3591 err_free_rxports:
3592 ub960_rxport_free_ports(priv);
3593
3594 return ret;
3595 }
3596
ub960_notify_bound(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3597 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3598 struct v4l2_subdev *subdev,
3599 struct v4l2_async_connection *asd)
3600 {
3601 struct ub960_data *priv = sd_to_ub960(notifier->sd);
3602 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3603 struct device *dev = &priv->client->dev;
3604 u8 nport = rxport->nport;
3605 unsigned int i;
3606 int ret;
3607
3608 ret = media_entity_get_fwnode_pad(&subdev->entity,
3609 rxport->source.ep_fwnode,
3610 MEDIA_PAD_FL_SOURCE);
3611 if (ret < 0) {
3612 dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3613 return ret;
3614 }
3615
3616 rxport->source.sd = subdev;
3617 rxport->source.pad = ret;
3618
3619 ret = media_create_pad_link(&rxport->source.sd->entity,
3620 rxport->source.pad, &priv->sd.entity, nport,
3621 MEDIA_LNK_FL_ENABLED |
3622 MEDIA_LNK_FL_IMMUTABLE);
3623 if (ret) {
3624 dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3625 rxport->source.sd->name, rxport->source.pad,
3626 priv->sd.name, nport);
3627 return ret;
3628 }
3629
3630 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3631 if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3632 dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3633 return 0;
3634 }
3635 }
3636
3637 return 0;
3638 }
3639
ub960_notify_unbind(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3640 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3641 struct v4l2_subdev *subdev,
3642 struct v4l2_async_connection *asd)
3643 {
3644 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3645
3646 rxport->source.sd = NULL;
3647 }
3648
3649 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3650 .bound = ub960_notify_bound,
3651 .unbind = ub960_notify_unbind,
3652 };
3653
ub960_v4l2_notifier_register(struct ub960_data * priv)3654 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3655 {
3656 struct device *dev = &priv->client->dev;
3657 unsigned int i;
3658 int ret;
3659
3660 v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
3661
3662 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3663 struct ub960_rxport *rxport = priv->rxports[i];
3664 struct ub960_asd *asd;
3665
3666 if (!rxport)
3667 continue;
3668
3669 asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3670 rxport->source.ep_fwnode,
3671 struct ub960_asd);
3672 if (IS_ERR(asd)) {
3673 dev_err(dev, "Failed to add subdev for source %u: %pe",
3674 i, asd);
3675 v4l2_async_nf_cleanup(&priv->notifier);
3676 return PTR_ERR(asd);
3677 }
3678
3679 asd->rxport = rxport;
3680 }
3681
3682 priv->notifier.ops = &ub960_notify_ops;
3683
3684 ret = v4l2_async_nf_register(&priv->notifier);
3685 if (ret) {
3686 dev_err(dev, "Failed to register subdev_notifier");
3687 v4l2_async_nf_cleanup(&priv->notifier);
3688 return ret;
3689 }
3690
3691 return 0;
3692 }
3693
ub960_v4l2_notifier_unregister(struct ub960_data * priv)3694 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3695 {
3696 v4l2_async_nf_unregister(&priv->notifier);
3697 v4l2_async_nf_cleanup(&priv->notifier);
3698 }
3699
ub960_create_subdev(struct ub960_data * priv)3700 static int ub960_create_subdev(struct ub960_data *priv)
3701 {
3702 struct device *dev = &priv->client->dev;
3703 unsigned int i;
3704 int ret;
3705
3706 v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3707 priv->sd.internal_ops = &ub960_internal_ops;
3708
3709 v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3710 priv->sd.ctrl_handler = &priv->ctrl_handler;
3711
3712 v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3713 ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3714 priv->tx_link_freq);
3715
3716 if (priv->ctrl_handler.error) {
3717 ret = priv->ctrl_handler.error;
3718 goto err_free_ctrl;
3719 }
3720
3721 priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3722 V4L2_SUBDEV_FL_STREAMS;
3723 priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3724 priv->sd.entity.ops = &ub960_entity_ops;
3725
3726 for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3727 priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3728 MEDIA_PAD_FL_SINK :
3729 MEDIA_PAD_FL_SOURCE;
3730 }
3731
3732 ret = media_entity_pads_init(&priv->sd.entity,
3733 priv->hw_data->num_rxports +
3734 priv->hw_data->num_txports,
3735 priv->pads);
3736 if (ret)
3737 goto err_free_ctrl;
3738
3739 priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3740
3741 ret = v4l2_subdev_init_finalize(&priv->sd);
3742 if (ret)
3743 goto err_entity_cleanup;
3744
3745 ret = ub960_v4l2_notifier_register(priv);
3746 if (ret) {
3747 dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3748 goto err_subdev_cleanup;
3749 }
3750
3751 ret = v4l2_async_register_subdev(&priv->sd);
3752 if (ret) {
3753 dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3754 goto err_unreg_notif;
3755 }
3756
3757 return 0;
3758
3759 err_unreg_notif:
3760 ub960_v4l2_notifier_unregister(priv);
3761 err_subdev_cleanup:
3762 v4l2_subdev_cleanup(&priv->sd);
3763 err_entity_cleanup:
3764 media_entity_cleanup(&priv->sd.entity);
3765 err_free_ctrl:
3766 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3767
3768 return ret;
3769 }
3770
ub960_destroy_subdev(struct ub960_data * priv)3771 static void ub960_destroy_subdev(struct ub960_data *priv)
3772 {
3773 ub960_v4l2_notifier_unregister(priv);
3774 v4l2_async_unregister_subdev(&priv->sd);
3775
3776 v4l2_subdev_cleanup(&priv->sd);
3777
3778 media_entity_cleanup(&priv->sd.entity);
3779 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3780 }
3781
3782 static const struct regmap_config ub960_regmap_config = {
3783 .name = "ds90ub960",
3784
3785 .reg_bits = 8,
3786 .val_bits = 8,
3787
3788 .max_register = 0xff,
3789
3790 /*
3791 * We do locking in the driver to cover the TX/RX port selection and the
3792 * indirect register access.
3793 */
3794 .disable_locking = true,
3795 };
3796
ub960_reset(struct ub960_data * priv,bool reset_regs)3797 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3798 {
3799 struct device *dev = &priv->client->dev;
3800 unsigned int v;
3801 int ret;
3802 u8 bit;
3803
3804 bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3805 UB960_SR_RESET_DIGITAL_RESET0;
3806
3807 ub960_write(priv, UB960_SR_RESET, bit);
3808
3809 mutex_lock(&priv->reg_lock);
3810
3811 ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3812 (v & bit) == 0, 2000, 100000);
3813
3814 mutex_unlock(&priv->reg_lock);
3815
3816 if (ret)
3817 dev_err(dev, "reset failed: %d\n", ret);
3818 }
3819
ub960_get_hw_resources(struct ub960_data * priv)3820 static int ub960_get_hw_resources(struct ub960_data *priv)
3821 {
3822 struct device *dev = &priv->client->dev;
3823
3824 priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3825 if (IS_ERR(priv->regmap))
3826 return PTR_ERR(priv->regmap);
3827
3828 priv->vddio = devm_regulator_get(dev, "vddio");
3829 if (IS_ERR(priv->vddio))
3830 return dev_err_probe(dev, PTR_ERR(priv->vddio),
3831 "cannot get VDDIO regulator\n");
3832
3833 /* get power-down pin from DT */
3834 priv->pd_gpio =
3835 devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3836 if (IS_ERR(priv->pd_gpio))
3837 return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3838 "Cannot get powerdown GPIO\n");
3839
3840 priv->refclk = devm_clk_get(dev, "refclk");
3841 if (IS_ERR(priv->refclk))
3842 return dev_err_probe(dev, PTR_ERR(priv->refclk),
3843 "Cannot get REFCLK\n");
3844
3845 return 0;
3846 }
3847
ub960_enable_core_hw(struct ub960_data * priv)3848 static int ub960_enable_core_hw(struct ub960_data *priv)
3849 {
3850 struct device *dev = &priv->client->dev;
3851 u8 rev_mask;
3852 int ret;
3853 u8 dev_sts;
3854 u8 refclk_freq;
3855
3856 ret = regulator_enable(priv->vddio);
3857 if (ret)
3858 return dev_err_probe(dev, ret,
3859 "failed to enable VDDIO regulator\n");
3860
3861 ret = clk_prepare_enable(priv->refclk);
3862 if (ret) {
3863 dev_err_probe(dev, ret, "Failed to enable refclk\n");
3864 goto err_disable_vddio;
3865 }
3866
3867 if (priv->pd_gpio) {
3868 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3869 /* wait min 2 ms for reset to complete */
3870 fsleep(2000);
3871 gpiod_set_value_cansleep(priv->pd_gpio, 0);
3872 /* wait min 2 ms for power up to finish */
3873 fsleep(2000);
3874 }
3875
3876 ub960_reset(priv, true);
3877
3878 /* Runtime check register accessibility */
3879 ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3880 if (ret) {
3881 dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3882 goto err_pd_gpio;
3883 }
3884
3885 dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3886 rev_mask);
3887
3888 ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3889 if (ret)
3890 goto err_pd_gpio;
3891
3892 if (priv->hw_data->is_ub9702)
3893 ret = ub960_read(priv, UB9702_SR_REFCLK_FREQ, &refclk_freq);
3894 else
3895 ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3896 if (ret)
3897 goto err_pd_gpio;
3898
3899 dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3900 !!(dev_sts & BIT(4)), refclk_freq,
3901 clk_get_rate(priv->refclk) / HZ_PER_MHZ);
3902
3903 /* Disable all RX ports by default */
3904 ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3905 if (ret)
3906 goto err_pd_gpio;
3907
3908 /* release GPIO lock */
3909 if (priv->hw_data->is_ub9702) {
3910 ret = ub960_update_bits(priv, UB960_SR_RESET,
3911 UB960_SR_RESET_GPIO_LOCK_RELEASE,
3912 UB960_SR_RESET_GPIO_LOCK_RELEASE);
3913 if (ret)
3914 goto err_pd_gpio;
3915 }
3916
3917 return 0;
3918
3919 err_pd_gpio:
3920 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3921 clk_disable_unprepare(priv->refclk);
3922 err_disable_vddio:
3923 regulator_disable(priv->vddio);
3924
3925 return ret;
3926 }
3927
ub960_disable_core_hw(struct ub960_data * priv)3928 static void ub960_disable_core_hw(struct ub960_data *priv)
3929 {
3930 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3931 clk_disable_unprepare(priv->refclk);
3932 regulator_disable(priv->vddio);
3933 }
3934
ub960_probe(struct i2c_client * client)3935 static int ub960_probe(struct i2c_client *client)
3936 {
3937 struct device *dev = &client->dev;
3938 struct ub960_data *priv;
3939 unsigned int port_lock_mask;
3940 unsigned int port_mask;
3941 unsigned int nport;
3942 int ret;
3943
3944 priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3945 if (!priv)
3946 return -ENOMEM;
3947
3948 priv->client = client;
3949
3950 priv->hw_data = device_get_match_data(dev);
3951
3952 mutex_init(&priv->reg_lock);
3953
3954 INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3955
3956 /*
3957 * Initialize these to invalid values so that the first reg writes will
3958 * configure the target.
3959 */
3960 priv->reg_current.indirect_target = 0xff;
3961 priv->reg_current.rxport = 0xff;
3962 priv->reg_current.txport = 0xff;
3963
3964 ret = ub960_get_hw_resources(priv);
3965 if (ret)
3966 goto err_mutex_destroy;
3967
3968 ret = ub960_enable_core_hw(priv);
3969 if (ret)
3970 goto err_mutex_destroy;
3971
3972 ret = ub960_parse_dt(priv);
3973 if (ret)
3974 goto err_disable_core_hw;
3975
3976 ret = ub960_init_tx_ports(priv);
3977 if (ret)
3978 goto err_free_ports;
3979
3980 ret = ub960_rxport_enable_vpocs(priv);
3981 if (ret)
3982 goto err_free_ports;
3983
3984 ret = ub960_init_rx_ports(priv);
3985 if (ret)
3986 goto err_disable_vpocs;
3987
3988 ub960_reset(priv, false);
3989
3990 port_mask = 0;
3991
3992 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3993 struct ub960_rxport *rxport = priv->rxports[nport];
3994
3995 if (!rxport)
3996 continue;
3997
3998 port_mask |= BIT(nport);
3999 }
4000
4001 ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
4002 if (ret)
4003 goto err_disable_vpocs;
4004
4005 if (port_mask != port_lock_mask) {
4006 ret = -EIO;
4007 dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
4008 goto err_disable_vpocs;
4009 }
4010
4011 /*
4012 * Clear any errors caused by switching the RX port settings while
4013 * probing.
4014 */
4015 ub960_clear_rx_errors(priv);
4016
4017 ret = ub960_init_atr(priv);
4018 if (ret)
4019 goto err_disable_vpocs;
4020
4021 ret = ub960_rxport_add_serializers(priv);
4022 if (ret)
4023 goto err_uninit_atr;
4024
4025 ret = ub960_create_subdev(priv);
4026 if (ret)
4027 goto err_free_sers;
4028
4029 if (client->irq)
4030 dev_warn(dev, "irq support not implemented, using polling\n");
4031
4032 schedule_delayed_work(&priv->poll_work,
4033 msecs_to_jiffies(UB960_POLL_TIME_MS));
4034
4035 #ifdef UB960_DEBUG_I2C_RX_ID
4036 for (unsigned int i = 0; i < priv->hw_data->num_rxports; i++)
4037 ub960_write(priv, UB960_SR_I2C_RX_ID(i),
4038 (UB960_DEBUG_I2C_RX_ID + i) << 1);
4039 #endif
4040
4041 return 0;
4042
4043 err_free_sers:
4044 ub960_rxport_remove_serializers(priv);
4045 err_uninit_atr:
4046 ub960_uninit_atr(priv);
4047 err_disable_vpocs:
4048 ub960_rxport_disable_vpocs(priv);
4049 err_free_ports:
4050 ub960_rxport_free_ports(priv);
4051 ub960_txport_free_ports(priv);
4052 err_disable_core_hw:
4053 ub960_disable_core_hw(priv);
4054 err_mutex_destroy:
4055 mutex_destroy(&priv->reg_lock);
4056 return ret;
4057 }
4058
ub960_remove(struct i2c_client * client)4059 static void ub960_remove(struct i2c_client *client)
4060 {
4061 struct v4l2_subdev *sd = i2c_get_clientdata(client);
4062 struct ub960_data *priv = sd_to_ub960(sd);
4063
4064 cancel_delayed_work_sync(&priv->poll_work);
4065
4066 ub960_destroy_subdev(priv);
4067 ub960_rxport_remove_serializers(priv);
4068 ub960_uninit_atr(priv);
4069 ub960_rxport_disable_vpocs(priv);
4070 ub960_rxport_free_ports(priv);
4071 ub960_txport_free_ports(priv);
4072 ub960_disable_core_hw(priv);
4073 mutex_destroy(&priv->reg_lock);
4074 }
4075
4076 static const struct ub960_hw_data ds90ub960_hw = {
4077 .model = "ub960",
4078 .num_rxports = 4,
4079 .num_txports = 2,
4080 };
4081
4082 static const struct ub960_hw_data ds90ub9702_hw = {
4083 .model = "ub9702",
4084 .num_rxports = 4,
4085 .num_txports = 2,
4086 .is_ub9702 = true,
4087 .is_fpdlink4 = true,
4088 };
4089
4090 static const struct i2c_device_id ub960_id[] = {
4091 { "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4092 { "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4093 {}
4094 };
4095 MODULE_DEVICE_TABLE(i2c, ub960_id);
4096
4097 static const struct of_device_id ub960_dt_ids[] = {
4098 { .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4099 { .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4100 {}
4101 };
4102 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4103
4104 static struct i2c_driver ds90ub960_driver = {
4105 .probe = ub960_probe,
4106 .remove = ub960_remove,
4107 .id_table = ub960_id,
4108 .driver = {
4109 .name = "ds90ub960",
4110 .of_match_table = ub960_dt_ids,
4111 },
4112 };
4113 module_i2c_driver(ds90ub960_driver);
4114
4115 MODULE_LICENSE("GPL");
4116 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4117 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4118 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4119 MODULE_IMPORT_NS("I2C_ATR");
4120