1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
4 *
5 * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6 * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
7 */
8
9 /*
10 * (Possible) TODOs:
11 *
12 * - PM for serializer and remote peripherals. We need to manage:
13 * - VPOC
14 * - Power domain? Regulator? Somehow any remote device should be able to
15 * cause the VPOC to be turned on.
16 * - Link between the deserializer and the serializer
17 * - Related to VPOC management. We probably always want to turn on the VPOC
18 * and then enable the link.
19 * - Serializer's services: i2c, gpios, power
20 * - The serializer needs to resume before the remote peripherals can
21 * e.g. use the i2c.
22 * - How to handle gpios? Reserving a gpio essentially keeps the provider
23 * (serializer) always powered on.
24 * - Do we need a new bus for the FPD-Link? At the moment the serializers
25 * are children of the same i2c-adapter where the deserializer resides.
26 * - i2c-atr could be made embeddable instead of allocatable.
27 */
28
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/workqueue.h>
47
48 #include <media/i2c/ds90ub9xx.h>
49 #include <media/mipi-csi2.h>
50 #include <media/v4l2-ctrls.h>
51 #include <media/v4l2-event.h>
52 #include <media/v4l2-fwnode.h>
53 #include <media/v4l2-subdev.h>
54
55 #define MHZ(v) ((u32)((v) * 1000000U))
56
57 #define UB960_POLL_TIME_MS 500
58
59 #define UB960_MAX_RX_NPORTS 4
60 #define UB960_MAX_TX_NPORTS 2
61 #define UB960_MAX_NPORTS (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
62
63 #define UB960_MAX_PORT_ALIASES 8
64
65 #define UB960_NUM_BC_GPIOS 4
66
67 /*
68 * Register map
69 *
70 * 0x00-0x32 Shared (UB960_SR)
71 * 0x33-0x3a CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
72 * 0x4c Shared (UB960_SR)
73 * 0x4d-0x7f FPD-Link RX, per-port paged (UB960_RR)
74 * 0xb0-0xbf Shared (UB960_SR)
75 * 0xd0-0xdf FPD-Link RX, per-port paged (UB960_RR)
76 * 0xf0-0xf5 Shared (UB960_SR)
77 * 0xf8-0xfb Shared (UB960_SR)
78 * All others Reserved
79 *
80 * Register prefixes:
81 * UB960_SR_* = Shared register
82 * UB960_RR_* = FPD-Link RX, per-port paged register
83 * UB960_TR_* = CSI-2 TX, per-port paged register
84 * UB960_XR_* = Reserved register
85 * UB960_IR_* = Indirect register
86 */
87
88 #define UB960_SR_I2C_DEV_ID 0x00
89 #define UB960_SR_RESET 0x01
90 #define UB960_SR_RESET_DIGITAL_RESET1 BIT(1)
91 #define UB960_SR_RESET_DIGITAL_RESET0 BIT(0)
92 #define UB960_SR_RESET_GPIO_LOCK_RELEASE BIT(5)
93
94 #define UB960_SR_GEN_CONFIG 0x02
95 #define UB960_SR_REV_MASK 0x03
96 #define UB960_SR_DEVICE_STS 0x04
97 #define UB960_SR_PAR_ERR_THOLD_HI 0x05
98 #define UB960_SR_PAR_ERR_THOLD_LO 0x06
99 #define UB960_SR_BCC_WDOG_CTL 0x07
100 #define UB960_SR_I2C_CTL1 0x08
101 #define UB960_SR_I2C_CTL2 0x09
102 #define UB960_SR_SCL_HIGH_TIME 0x0a
103 #define UB960_SR_SCL_LOW_TIME 0x0b
104 #define UB960_SR_RX_PORT_CTL 0x0c
105 #define UB960_SR_IO_CTL 0x0d
106 #define UB960_SR_GPIO_PIN_STS 0x0e
107 #define UB960_SR_GPIO_INPUT_CTL 0x0f
108 #define UB960_SR_GPIO_PIN_CTL(n) (0x10 + (n)) /* n < UB960_NUM_GPIOS */
109 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL 5
110 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT 2
111 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN BIT(0)
112
113 #define UB960_SR_FS_CTL 0x18
114 #define UB960_SR_FS_HIGH_TIME_1 0x19
115 #define UB960_SR_FS_HIGH_TIME_0 0x1a
116 #define UB960_SR_FS_LOW_TIME_1 0x1b
117 #define UB960_SR_FS_LOW_TIME_0 0x1c
118 #define UB960_SR_MAX_FRM_HI 0x1d
119 #define UB960_SR_MAX_FRM_LO 0x1e
120 #define UB960_SR_CSI_PLL_CTL 0x1f
121
122 #define UB960_SR_FWD_CTL1 0x20
123 #define UB960_SR_FWD_CTL1_PORT_DIS(n) BIT((n) + 4)
124
125 #define UB960_SR_FWD_CTL2 0x21
126 #define UB960_SR_FWD_STS 0x22
127
128 #define UB960_SR_INTERRUPT_CTL 0x23
129 #define UB960_SR_INTERRUPT_CTL_INT_EN BIT(7)
130 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0 BIT(4)
131 #define UB960_SR_INTERRUPT_CTL_IE_RX(n) BIT((n)) /* rxport[n] IRQ */
132
133 #define UB960_SR_INTERRUPT_STS 0x24
134 #define UB960_SR_INTERRUPT_STS_INT BIT(7)
135 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n) BIT(4 + (n)) /* txport[n] IRQ */
136 #define UB960_SR_INTERRUPT_STS_IS_RX(n) BIT((n)) /* rxport[n] IRQ */
137
138 #define UB960_SR_TS_CONFIG 0x25
139 #define UB960_SR_TS_CONTROL 0x26
140 #define UB960_SR_TS_LINE_HI 0x27
141 #define UB960_SR_TS_LINE_LO 0x28
142 #define UB960_SR_TS_STATUS 0x29
143 #define UB960_SR_TIMESTAMP_P0_HI 0x2a
144 #define UB960_SR_TIMESTAMP_P0_LO 0x2b
145 #define UB960_SR_TIMESTAMP_P1_HI 0x2c
146 #define UB960_SR_TIMESTAMP_P1_LO 0x2d
147
148 #define UB960_SR_CSI_PORT_SEL 0x32
149
150 #define UB960_TR_CSI_CTL 0x33
151 #define UB960_TR_CSI_CTL_CSI_CAL_EN BIT(6)
152 #define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK BIT(1)
153 #define UB960_TR_CSI_CTL_CSI_ENABLE BIT(0)
154
155 #define UB960_TR_CSI_CTL2 0x34
156 #define UB960_TR_CSI_STS 0x35
157 #define UB960_TR_CSI_TX_ICR 0x36
158
159 #define UB960_TR_CSI_TX_ISR 0x37
160 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR BIT(3)
161 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR BIT(1)
162
163 #define UB960_TR_CSI_TEST_CTL 0x38
164 #define UB960_TR_CSI_TEST_PATT_HI 0x39
165 #define UB960_TR_CSI_TEST_PATT_LO 0x3a
166
167 #define UB960_XR_SFILTER_CFG 0x41
168 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT 4
169 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT 0
170
171 #define UB960_XR_AEQ_CTL1 0x42
172 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK BIT(6)
173 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING BIT(5)
174 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY BIT(4)
175 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK \
176 (UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK | \
177 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
178 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
179 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN BIT(0)
180
181 #define UB960_XR_AEQ_ERR_THOLD 0x43
182
183 #define UB960_RR_BCC_ERR_CTL 0x46
184 #define UB960_RR_BCC_STATUS 0x47
185 #define UB960_RR_BCC_STATUS_SEQ_ERROR BIT(5)
186 #define UB960_RR_BCC_STATUS_MASTER_ERR BIT(4)
187 #define UB960_RR_BCC_STATUS_MASTER_TO BIT(3)
188 #define UB960_RR_BCC_STATUS_SLAVE_ERR BIT(2)
189 #define UB960_RR_BCC_STATUS_SLAVE_TO BIT(1)
190 #define UB960_RR_BCC_STATUS_RESP_ERR BIT(0)
191 #define UB960_RR_BCC_STATUS_ERROR_MASK \
192 (UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
193 UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR | \
194 UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
195
196 #define UB960_RR_FPD3_CAP 0x4a
197 #define UB960_RR_RAW_EMBED_DTYPE 0x4b
198 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT 6
199
200 #define UB960_SR_FPD3_PORT_SEL 0x4c
201
202 #define UB960_RR_RX_PORT_STS1 0x4d
203 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR BIT(5)
204 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG BIT(4)
205 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR BIT(3)
206 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR BIT(2)
207 #define UB960_RR_RX_PORT_STS1_PORT_PASS BIT(1)
208 #define UB960_RR_RX_PORT_STS1_LOCK_STS BIT(0)
209 #define UB960_RR_RX_PORT_STS1_ERROR_MASK \
210 (UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
211 UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
212 UB960_RR_RX_PORT_STS1_PARITY_ERROR)
213
214 #define UB960_RR_RX_PORT_STS2 0x4e
215 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
216 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG BIT(6)
217 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
218 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR BIT(4)
219 #define UB960_RR_RX_PORT_STS2_CSI_ERROR BIT(3)
220 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE BIT(2)
221 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT BIT(1)
222 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG BIT(0)
223 #define UB960_RR_RX_PORT_STS2_ERROR_MASK \
224 UB960_RR_RX_PORT_STS2_BUFFER_ERROR
225
226 #define UB960_RR_RX_FREQ_HIGH 0x4f
227 #define UB960_RR_RX_FREQ_LOW 0x50
228 #define UB960_RR_SENSOR_STS_0 0x51
229 #define UB960_RR_SENSOR_STS_1 0x52
230 #define UB960_RR_SENSOR_STS_2 0x53
231 #define UB960_RR_SENSOR_STS_3 0x54
232 #define UB960_RR_RX_PAR_ERR_HI 0x55
233 #define UB960_RR_RX_PAR_ERR_LO 0x56
234 #define UB960_RR_BIST_ERR_COUNT 0x57
235
236 #define UB960_RR_BCC_CONFIG 0x58
237 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH BIT(6)
238 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK GENMASK(2, 0)
239
240 #define UB960_RR_DATAPATH_CTL1 0x59
241 #define UB960_RR_DATAPATH_CTL2 0x5a
242 #define UB960_RR_SER_ID 0x5b
243 #define UB960_RR_SER_ALIAS_ID 0x5c
244
245 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
246 #define UB960_RR_SLAVE_ID(n) (0x5d + (n))
247 #define UB960_RR_SLAVE_ALIAS(n) (0x65 + (n))
248
249 #define UB960_RR_PORT_CONFIG 0x6d
250 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK GENMASK(1, 0)
251
252 #define UB960_RR_BC_GPIO_CTL(n) (0x6e + (n)) /* n < 2 */
253 #define UB960_RR_RAW10_ID 0x70
254 #define UB960_RR_RAW10_ID_VC_SHIFT 6
255 #define UB960_RR_RAW10_ID_DT_SHIFT 0
256
257 #define UB960_RR_RAW12_ID 0x71
258 #define UB960_RR_CSI_VC_MAP 0x72
259 #define UB960_RR_CSI_VC_MAP_SHIFT(x) ((x) * 2)
260
261 #define UB960_RR_LINE_COUNT_HI 0x73
262 #define UB960_RR_LINE_COUNT_LO 0x74
263 #define UB960_RR_LINE_LEN_1 0x75
264 #define UB960_RR_LINE_LEN_0 0x76
265 #define UB960_RR_FREQ_DET_CTL 0x77
266 #define UB960_RR_MAILBOX_1 0x78
267 #define UB960_RR_MAILBOX_2 0x79
268
269 #define UB960_RR_CSI_RX_STS 0x7a
270 #define UB960_RR_CSI_RX_STS_LENGTH_ERR BIT(3)
271 #define UB960_RR_CSI_RX_STS_CKSUM_ERR BIT(2)
272 #define UB960_RR_CSI_RX_STS_ECC2_ERR BIT(1)
273 #define UB960_RR_CSI_RX_STS_ECC1_ERR BIT(0)
274 #define UB960_RR_CSI_RX_STS_ERROR_MASK \
275 (UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
276 UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
277
278 #define UB960_RR_CSI_ERR_COUNTER 0x7b
279 #define UB960_RR_PORT_CONFIG2 0x7c
280 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
281 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
282
283 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW BIT(1)
284 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW BIT(0)
285
286 #define UB960_RR_PORT_PASS_CTL 0x7d
287 #define UB960_RR_SEN_INT_RISE_CTL 0x7e
288 #define UB960_RR_SEN_INT_FALL_CTL 0x7f
289
290 #define UB960_SR_CSI_FRAME_COUNT_HI(n) (0x90 + 8 * (n))
291 #define UB960_SR_CSI_FRAME_COUNT_LO(n) (0x91 + 8 * (n))
292 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n) (0x92 + 8 * (n))
293 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n) (0x93 + 8 * (n))
294 #define UB960_SR_CSI_LINE_COUNT_HI(n) (0x94 + 8 * (n))
295 #define UB960_SR_CSI_LINE_COUNT_LO(n) (0x95 + 8 * (n))
296 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n) (0x96 + 8 * (n))
297 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n) (0x97 + 8 * (n))
298
299 #define UB960_XR_REFCLK_FREQ 0xa5 /* UB960 */
300
301 #define UB960_RR_VC_ID_MAP(x) (0xa0 + (x)) /* UB9702 */
302
303 #define UB960_SR_IND_ACC_CTL 0xb0
304 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC BIT(1)
305
306 #define UB960_SR_IND_ACC_ADDR 0xb1
307 #define UB960_SR_IND_ACC_DATA 0xb2
308 #define UB960_SR_BIST_CONTROL 0xb3
309 #define UB960_SR_MODE_IDX_STS 0xb8
310 #define UB960_SR_LINK_ERROR_COUNT 0xb9
311 #define UB960_SR_FPD3_ENC_CTL 0xba
312 #define UB960_SR_FV_MIN_TIME 0xbc
313 #define UB960_SR_GPIO_PD_CTL 0xbe
314
315 #define UB960_SR_FPD_RATE_CFG 0xc2 /* UB9702 */
316 #define UB960_SR_CSI_PLL_DIV 0xc9 /* UB9702 */
317
318 #define UB960_RR_PORT_DEBUG 0xd0
319 #define UB960_RR_AEQ_CTL2 0xd2
320 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR BIT(2)
321
322 #define UB960_RR_AEQ_STATUS 0xd3
323 #define UB960_RR_AEQ_STATUS_STATUS_2 GENMASK(5, 3)
324 #define UB960_RR_AEQ_STATUS_STATUS_1 GENMASK(2, 0)
325
326 #define UB960_RR_AEQ_BYPASS 0xd4
327 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT 5
328 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK GENMASK(7, 5)
329 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT 1
330 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK GENMASK(3, 1)
331 #define UB960_RR_AEQ_BYPASS_ENABLE BIT(0)
332
333 #define UB960_RR_AEQ_MIN_MAX 0xd5
334 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT 4
335 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT 0
336
337 #define UB960_RR_SFILTER_STS_0 0xd6
338 #define UB960_RR_SFILTER_STS_1 0xd7
339 #define UB960_RR_PORT_ICR_HI 0xd8
340 #define UB960_RR_PORT_ICR_LO 0xd9
341 #define UB960_RR_PORT_ISR_HI 0xda
342 #define UB960_RR_PORT_ISR_LO 0xdb
343 #define UB960_RR_FC_GPIO_STS 0xdc
344 #define UB960_RR_FC_GPIO_ICR 0xdd
345 #define UB960_RR_SEN_INT_RISE_STS 0xde
346 #define UB960_RR_SEN_INT_FALL_STS 0xdf
347
348 #define UB960_RR_CHANNEL_MODE 0xe4 /* UB9702 */
349
350 #define UB960_SR_FPD3_RX_ID(n) (0xf0 + (n))
351 #define UB960_SR_FPD3_RX_ID_LEN 6
352
353 #define UB960_SR_I2C_RX_ID(n) (0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
354
355 /* Indirect register blocks */
356 #define UB960_IND_TARGET_PAT_GEN 0x00
357 #define UB960_IND_TARGET_RX_ANA(n) (0x01 + (n))
358 #define UB960_IND_TARGET_CSI_CSIPLL_REG_1 0x92 /* UB9702 */
359 #define UB960_IND_TARGET_CSI_ANA 0x07
360
361 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
362
363 #define UB960_IR_PGEN_CTL 0x01
364 #define UB960_IR_PGEN_CTL_PGEN_ENABLE BIT(0)
365
366 #define UB960_IR_PGEN_CFG 0x02
367 #define UB960_IR_PGEN_CSI_DI 0x03
368 #define UB960_IR_PGEN_LINE_SIZE1 0x04
369 #define UB960_IR_PGEN_LINE_SIZE0 0x05
370 #define UB960_IR_PGEN_BAR_SIZE1 0x06
371 #define UB960_IR_PGEN_BAR_SIZE0 0x07
372 #define UB960_IR_PGEN_ACT_LPF1 0x08
373 #define UB960_IR_PGEN_ACT_LPF0 0x09
374 #define UB960_IR_PGEN_TOT_LPF1 0x0a
375 #define UB960_IR_PGEN_TOT_LPF0 0x0b
376 #define UB960_IR_PGEN_LINE_PD1 0x0c
377 #define UB960_IR_PGEN_LINE_PD0 0x0d
378 #define UB960_IR_PGEN_VBP 0x0e
379 #define UB960_IR_PGEN_VFP 0x0f
380 #define UB960_IR_PGEN_COLOR(n) (0x10 + (n)) /* n < 15 */
381
382 #define UB960_IR_RX_ANA_STROBE_SET_CLK 0x08
383 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY BIT(3)
384 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK GENMASK(2, 0)
385
386 #define UB960_IR_RX_ANA_STROBE_SET_DATA 0x09
387 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY BIT(3)
388 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK GENMASK(2, 0)
389
390 /* EQ related */
391
392 #define UB960_MIN_AEQ_STROBE_POS -7
393 #define UB960_MAX_AEQ_STROBE_POS 7
394
395 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
396
397 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
398 #define UB960_MAX_MANUAL_STROBE_POS (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
399 #define UB960_NUM_MANUAL_STROBE_POS (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
400
401 #define UB960_MIN_EQ_LEVEL 0
402 #define UB960_MAX_EQ_LEVEL 14
403 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
404
405 struct ub960_hw_data {
406 const char *model;
407 u8 num_rxports;
408 u8 num_txports;
409 bool is_ub9702;
410 bool is_fpdlink4;
411 };
412
413 enum ub960_rxport_mode {
414 RXPORT_MODE_RAW10 = 0,
415 RXPORT_MODE_RAW12_HF = 1,
416 RXPORT_MODE_RAW12_LF = 2,
417 RXPORT_MODE_CSI2_SYNC = 3,
418 RXPORT_MODE_CSI2_NONSYNC = 4,
419 RXPORT_MODE_LAST = RXPORT_MODE_CSI2_NONSYNC,
420 };
421
422 enum ub960_rxport_cdr {
423 RXPORT_CDR_FPD3 = 0,
424 RXPORT_CDR_FPD4 = 1,
425 RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
426 };
427
428 struct ub960_rxport {
429 struct ub960_data *priv;
430 u8 nport; /* RX port number, and index in priv->rxport[] */
431
432 struct {
433 struct v4l2_subdev *sd;
434 u16 pad;
435 struct fwnode_handle *ep_fwnode;
436 } source;
437
438 /* Serializer */
439 struct {
440 struct fwnode_handle *fwnode;
441 struct i2c_client *client;
442 unsigned short alias; /* I2C alias (lower 7 bits) */
443 struct ds90ub9xx_platform_data pdata;
444 } ser;
445
446 enum ub960_rxport_mode rx_mode;
447 enum ub960_rxport_cdr cdr_mode;
448
449 u8 lv_fv_pol; /* LV and FV polarities */
450
451 struct regulator *vpoc;
452
453 /* EQ settings */
454 struct {
455 bool manual_eq;
456
457 s8 strobe_pos;
458
459 union {
460 struct {
461 u8 eq_level_min;
462 u8 eq_level_max;
463 } aeq;
464
465 struct {
466 u8 eq_level;
467 } manual;
468 };
469 } eq;
470
471 const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
472 };
473
474 struct ub960_asd {
475 struct v4l2_async_connection base;
476 struct ub960_rxport *rxport;
477 };
478
to_ub960_asd(struct v4l2_async_connection * asd)479 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
480 {
481 return container_of(asd, struct ub960_asd, base);
482 }
483
484 struct ub960_txport {
485 struct ub960_data *priv;
486 u8 nport; /* TX port number, and index in priv->txport[] */
487
488 u32 num_data_lanes;
489 bool non_continous_clk;
490 };
491
492 struct ub960_data {
493 const struct ub960_hw_data *hw_data;
494 struct i2c_client *client; /* for shared local registers */
495 struct regmap *regmap;
496
497 /* lock for register access */
498 struct mutex reg_lock;
499
500 struct clk *refclk;
501
502 struct regulator *vddio;
503
504 struct gpio_desc *pd_gpio;
505 struct delayed_work poll_work;
506 struct ub960_rxport *rxports[UB960_MAX_RX_NPORTS];
507 struct ub960_txport *txports[UB960_MAX_TX_NPORTS];
508
509 struct v4l2_subdev sd;
510 struct media_pad pads[UB960_MAX_NPORTS];
511
512 struct v4l2_ctrl_handler ctrl_handler;
513 struct v4l2_async_notifier notifier;
514
515 u32 tx_data_rate; /* Nominal data rate (Gb/s) */
516 s64 tx_link_freq[1];
517
518 struct i2c_atr *atr;
519
520 struct {
521 u8 rxport;
522 u8 txport;
523 u8 indirect_target;
524 } reg_current;
525
526 bool streaming;
527
528 u8 stored_fwd_ctl;
529
530 u64 stream_enable_mask[UB960_MAX_NPORTS];
531
532 /* These are common to all ports */
533 struct {
534 bool manual;
535
536 s8 min;
537 s8 max;
538 } strobe;
539 };
540
sd_to_ub960(struct v4l2_subdev * sd)541 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
542 {
543 return container_of(sd, struct ub960_data, sd);
544 }
545
ub960_pad_is_sink(struct ub960_data * priv,u32 pad)546 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
547 {
548 return pad < priv->hw_data->num_rxports;
549 }
550
ub960_pad_is_source(struct ub960_data * priv,u32 pad)551 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
552 {
553 return pad >= priv->hw_data->num_rxports;
554 }
555
ub960_pad_to_port(struct ub960_data * priv,u32 pad)556 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
557 {
558 if (ub960_pad_is_sink(priv, pad))
559 return pad;
560 else
561 return pad - priv->hw_data->num_rxports;
562 }
563
564 struct ub960_format_info {
565 u32 code;
566 u32 bpp;
567 u8 datatype;
568 bool meta;
569 };
570
571 static const struct ub960_format_info ub960_formats[] = {
572 { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
573 { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
574 { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
575 { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
576
577 { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
578 { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
579 { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
580 { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
581 };
582
ub960_find_format(u32 code)583 static const struct ub960_format_info *ub960_find_format(u32 code)
584 {
585 unsigned int i;
586
587 for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
588 if (ub960_formats[i].code == code)
589 return &ub960_formats[i];
590 }
591
592 return NULL;
593 }
594
595 /* -----------------------------------------------------------------------------
596 * Basic device access
597 */
598
ub960_read(struct ub960_data * priv,u8 reg,u8 * val)599 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
600 {
601 struct device *dev = &priv->client->dev;
602 unsigned int v;
603 int ret;
604
605 mutex_lock(&priv->reg_lock);
606
607 ret = regmap_read(priv->regmap, reg, &v);
608 if (ret) {
609 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
610 __func__, reg, ret);
611 goto out_unlock;
612 }
613
614 *val = v;
615
616 out_unlock:
617 mutex_unlock(&priv->reg_lock);
618
619 return ret;
620 }
621
ub960_write(struct ub960_data * priv,u8 reg,u8 val)622 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
623 {
624 struct device *dev = &priv->client->dev;
625 int ret;
626
627 mutex_lock(&priv->reg_lock);
628
629 ret = regmap_write(priv->regmap, reg, val);
630 if (ret)
631 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
632 __func__, reg, ret);
633
634 mutex_unlock(&priv->reg_lock);
635
636 return ret;
637 }
638
ub960_update_bits(struct ub960_data * priv,u8 reg,u8 mask,u8 val)639 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
640 {
641 struct device *dev = &priv->client->dev;
642 int ret;
643
644 mutex_lock(&priv->reg_lock);
645
646 ret = regmap_update_bits(priv->regmap, reg, mask, val);
647 if (ret)
648 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
649 __func__, reg, ret);
650
651 mutex_unlock(&priv->reg_lock);
652
653 return ret;
654 }
655
ub960_read16(struct ub960_data * priv,u8 reg,u16 * val)656 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
657 {
658 struct device *dev = &priv->client->dev;
659 __be16 __v;
660 int ret;
661
662 mutex_lock(&priv->reg_lock);
663
664 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
665 if (ret) {
666 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
667 __func__, reg, ret);
668 goto out_unlock;
669 }
670
671 *val = be16_to_cpu(__v);
672
673 out_unlock:
674 mutex_unlock(&priv->reg_lock);
675
676 return ret;
677 }
678
ub960_rxport_select(struct ub960_data * priv,u8 nport)679 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
680 {
681 struct device *dev = &priv->client->dev;
682 int ret;
683
684 lockdep_assert_held(&priv->reg_lock);
685
686 if (priv->reg_current.rxport == nport)
687 return 0;
688
689 ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
690 (nport << 4) | BIT(nport));
691 if (ret) {
692 dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
693 nport, ret);
694 return ret;
695 }
696
697 priv->reg_current.rxport = nport;
698
699 return 0;
700 }
701
ub960_rxport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)702 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
703 {
704 struct device *dev = &priv->client->dev;
705 unsigned int v;
706 int ret;
707
708 mutex_lock(&priv->reg_lock);
709
710 ret = ub960_rxport_select(priv, nport);
711 if (ret)
712 goto out_unlock;
713
714 ret = regmap_read(priv->regmap, reg, &v);
715 if (ret) {
716 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
717 __func__, reg, ret);
718 goto out_unlock;
719 }
720
721 *val = v;
722
723 out_unlock:
724 mutex_unlock(&priv->reg_lock);
725
726 return ret;
727 }
728
ub960_rxport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)729 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
730 {
731 struct device *dev = &priv->client->dev;
732 int ret;
733
734 mutex_lock(&priv->reg_lock);
735
736 ret = ub960_rxport_select(priv, nport);
737 if (ret)
738 goto out_unlock;
739
740 ret = regmap_write(priv->regmap, reg, val);
741 if (ret)
742 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
743 __func__, reg, ret);
744
745 out_unlock:
746 mutex_unlock(&priv->reg_lock);
747
748 return ret;
749 }
750
ub960_rxport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)751 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
752 u8 mask, u8 val)
753 {
754 struct device *dev = &priv->client->dev;
755 int ret;
756
757 mutex_lock(&priv->reg_lock);
758
759 ret = ub960_rxport_select(priv, nport);
760 if (ret)
761 goto out_unlock;
762
763 ret = regmap_update_bits(priv->regmap, reg, mask, val);
764 if (ret)
765 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
766 __func__, reg, ret);
767
768 out_unlock:
769 mutex_unlock(&priv->reg_lock);
770
771 return ret;
772 }
773
ub960_rxport_read16(struct ub960_data * priv,u8 nport,u8 reg,u16 * val)774 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
775 u16 *val)
776 {
777 struct device *dev = &priv->client->dev;
778 __be16 __v;
779 int ret;
780
781 mutex_lock(&priv->reg_lock);
782
783 ret = ub960_rxport_select(priv, nport);
784 if (ret)
785 goto out_unlock;
786
787 ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
788 if (ret) {
789 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
790 __func__, reg, ret);
791 goto out_unlock;
792 }
793
794 *val = be16_to_cpu(__v);
795
796 out_unlock:
797 mutex_unlock(&priv->reg_lock);
798
799 return ret;
800 }
801
ub960_txport_select(struct ub960_data * priv,u8 nport)802 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
803 {
804 struct device *dev = &priv->client->dev;
805 int ret;
806
807 lockdep_assert_held(&priv->reg_lock);
808
809 if (priv->reg_current.txport == nport)
810 return 0;
811
812 ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
813 (nport << 4) | BIT(nport));
814 if (ret) {
815 dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
816 nport, ret);
817 return ret;
818 }
819
820 priv->reg_current.txport = nport;
821
822 return 0;
823 }
824
ub960_txport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)825 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
826 {
827 struct device *dev = &priv->client->dev;
828 unsigned int v;
829 int ret;
830
831 mutex_lock(&priv->reg_lock);
832
833 ret = ub960_txport_select(priv, nport);
834 if (ret)
835 goto out_unlock;
836
837 ret = regmap_read(priv->regmap, reg, &v);
838 if (ret) {
839 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
840 __func__, reg, ret);
841 goto out_unlock;
842 }
843
844 *val = v;
845
846 out_unlock:
847 mutex_unlock(&priv->reg_lock);
848
849 return ret;
850 }
851
ub960_txport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)852 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
853 {
854 struct device *dev = &priv->client->dev;
855 int ret;
856
857 mutex_lock(&priv->reg_lock);
858
859 ret = ub960_txport_select(priv, nport);
860 if (ret)
861 goto out_unlock;
862
863 ret = regmap_write(priv->regmap, reg, val);
864 if (ret)
865 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
866 __func__, reg, ret);
867
868 out_unlock:
869 mutex_unlock(&priv->reg_lock);
870
871 return ret;
872 }
873
ub960_txport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)874 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
875 u8 mask, u8 val)
876 {
877 struct device *dev = &priv->client->dev;
878 int ret;
879
880 mutex_lock(&priv->reg_lock);
881
882 ret = ub960_txport_select(priv, nport);
883 if (ret)
884 goto out_unlock;
885
886 ret = regmap_update_bits(priv->regmap, reg, mask, val);
887 if (ret)
888 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
889 __func__, reg, ret);
890
891 out_unlock:
892 mutex_unlock(&priv->reg_lock);
893
894 return ret;
895 }
896
ub960_select_ind_reg_block(struct ub960_data * priv,u8 block)897 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
898 {
899 struct device *dev = &priv->client->dev;
900 int ret;
901
902 lockdep_assert_held(&priv->reg_lock);
903
904 if (priv->reg_current.indirect_target == block)
905 return 0;
906
907 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
908 if (ret) {
909 dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
910 __func__, block, ret);
911 return ret;
912 }
913
914 priv->reg_current.indirect_target = block;
915
916 return 0;
917 }
918
ub960_read_ind(struct ub960_data * priv,u8 block,u8 reg,u8 * val)919 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
920 {
921 struct device *dev = &priv->client->dev;
922 unsigned int v;
923 int ret;
924
925 mutex_lock(&priv->reg_lock);
926
927 ret = ub960_select_ind_reg_block(priv, block);
928 if (ret)
929 goto out_unlock;
930
931 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
932 if (ret) {
933 dev_err(dev,
934 "Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
935 block, reg, ret);
936 goto out_unlock;
937 }
938
939 ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
940 if (ret) {
941 dev_err(dev,
942 "Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
943 block, reg, ret);
944 goto out_unlock;
945 }
946
947 *val = v;
948
949 out_unlock:
950 mutex_unlock(&priv->reg_lock);
951
952 return ret;
953 }
954
ub960_write_ind(struct ub960_data * priv,u8 block,u8 reg,u8 val)955 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
956 {
957 struct device *dev = &priv->client->dev;
958 int ret;
959
960 mutex_lock(&priv->reg_lock);
961
962 ret = ub960_select_ind_reg_block(priv, block);
963 if (ret)
964 goto out_unlock;
965
966 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
967 if (ret) {
968 dev_err(dev,
969 "Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
970 block, reg, ret);
971 goto out_unlock;
972 }
973
974 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
975 if (ret) {
976 dev_err(dev,
977 "Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
978 block, reg, ret);
979 goto out_unlock;
980 }
981
982 out_unlock:
983 mutex_unlock(&priv->reg_lock);
984
985 return ret;
986 }
987
ub960_ind_update_bits(struct ub960_data * priv,u8 block,u8 reg,u8 mask,u8 val)988 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
989 u8 mask, u8 val)
990 {
991 struct device *dev = &priv->client->dev;
992 int ret;
993
994 mutex_lock(&priv->reg_lock);
995
996 ret = ub960_select_ind_reg_block(priv, block);
997 if (ret)
998 goto out_unlock;
999
1000 ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
1001 if (ret) {
1002 dev_err(dev,
1003 "Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1004 block, reg, ret);
1005 goto out_unlock;
1006 }
1007
1008 ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1009 val);
1010 if (ret) {
1011 dev_err(dev,
1012 "Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1013 block, reg, ret);
1014 goto out_unlock;
1015 }
1016
1017 out_unlock:
1018 mutex_unlock(&priv->reg_lock);
1019
1020 return ret;
1021 }
1022
1023 /* -----------------------------------------------------------------------------
1024 * I2C-ATR (address translator)
1025 */
1026
ub960_atr_attach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client,u16 alias)1027 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1028 const struct i2c_client *client, u16 alias)
1029 {
1030 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1031 struct ub960_rxport *rxport = priv->rxports[chan_id];
1032 struct device *dev = &priv->client->dev;
1033 unsigned int reg_idx;
1034
1035 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1036 if (!rxport->aliased_clients[reg_idx])
1037 break;
1038 }
1039
1040 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1041 dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1042 return -EADDRNOTAVAIL;
1043 }
1044
1045 rxport->aliased_clients[reg_idx] = client;
1046
1047 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1048 client->addr << 1);
1049 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1050 alias << 1);
1051
1052 dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1053 rxport->nport, client->addr, alias, reg_idx);
1054
1055 return 0;
1056 }
1057
ub960_atr_detach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client)1058 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1059 const struct i2c_client *client)
1060 {
1061 struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1062 struct ub960_rxport *rxport = priv->rxports[chan_id];
1063 struct device *dev = &priv->client->dev;
1064 unsigned int reg_idx;
1065
1066 for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1067 if (rxport->aliased_clients[reg_idx] == client)
1068 break;
1069 }
1070
1071 if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1072 dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1073 rxport->nport, client->addr);
1074 return;
1075 }
1076
1077 rxport->aliased_clients[reg_idx] = NULL;
1078
1079 ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1080
1081 dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1082 client->addr, reg_idx);
1083 }
1084
1085 static const struct i2c_atr_ops ub960_atr_ops = {
1086 .attach_client = ub960_atr_attach_client,
1087 .detach_client = ub960_atr_detach_client,
1088 };
1089
ub960_init_atr(struct ub960_data * priv)1090 static int ub960_init_atr(struct ub960_data *priv)
1091 {
1092 struct device *dev = &priv->client->dev;
1093 struct i2c_adapter *parent_adap = priv->client->adapter;
1094
1095 priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1096 priv->hw_data->num_rxports);
1097 if (IS_ERR(priv->atr))
1098 return PTR_ERR(priv->atr);
1099
1100 i2c_atr_set_driver_data(priv->atr, priv);
1101
1102 return 0;
1103 }
1104
ub960_uninit_atr(struct ub960_data * priv)1105 static void ub960_uninit_atr(struct ub960_data *priv)
1106 {
1107 i2c_atr_delete(priv->atr);
1108 priv->atr = NULL;
1109 }
1110
1111 /* -----------------------------------------------------------------------------
1112 * TX ports
1113 */
1114
ub960_parse_dt_txport(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,u8 nport)1115 static int ub960_parse_dt_txport(struct ub960_data *priv,
1116 struct fwnode_handle *ep_fwnode,
1117 u8 nport)
1118 {
1119 struct device *dev = &priv->client->dev;
1120 struct v4l2_fwnode_endpoint vep = {};
1121 struct ub960_txport *txport;
1122 int ret;
1123
1124 txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1125 if (!txport)
1126 return -ENOMEM;
1127
1128 txport->priv = priv;
1129 txport->nport = nport;
1130
1131 vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1132 ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1133 if (ret) {
1134 dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1135 goto err_free_txport;
1136 }
1137
1138 txport->non_continous_clk = vep.bus.mipi_csi2.flags &
1139 V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK;
1140
1141 txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1142
1143 if (vep.nr_of_link_frequencies != 1) {
1144 ret = -EINVAL;
1145 goto err_free_vep;
1146 }
1147
1148 priv->tx_link_freq[0] = vep.link_frequencies[0];
1149 priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1150
1151 if (priv->tx_data_rate != MHZ(1600) &&
1152 priv->tx_data_rate != MHZ(1200) &&
1153 priv->tx_data_rate != MHZ(800) &&
1154 priv->tx_data_rate != MHZ(400)) {
1155 dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1156 ret = -EINVAL;
1157 goto err_free_vep;
1158 }
1159
1160 v4l2_fwnode_endpoint_free(&vep);
1161
1162 priv->txports[nport] = txport;
1163
1164 return 0;
1165
1166 err_free_vep:
1167 v4l2_fwnode_endpoint_free(&vep);
1168 err_free_txport:
1169 kfree(txport);
1170
1171 return ret;
1172 }
1173
ub960_csi_handle_events(struct ub960_data * priv,u8 nport)1174 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1175 {
1176 struct device *dev = &priv->client->dev;
1177 u8 csi_tx_isr;
1178 int ret;
1179
1180 ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1181 if (ret)
1182 return;
1183
1184 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1185 dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1186
1187 if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1188 dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1189 }
1190
1191 /* -----------------------------------------------------------------------------
1192 * RX ports
1193 */
1194
ub960_rxport_enable_vpocs(struct ub960_data * priv)1195 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1196 {
1197 unsigned int nport;
1198 int ret;
1199
1200 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1201 struct ub960_rxport *rxport = priv->rxports[nport];
1202
1203 if (!rxport || !rxport->vpoc)
1204 continue;
1205
1206 ret = regulator_enable(rxport->vpoc);
1207 if (ret)
1208 goto err_disable_vpocs;
1209 }
1210
1211 return 0;
1212
1213 err_disable_vpocs:
1214 while (nport--) {
1215 struct ub960_rxport *rxport = priv->rxports[nport];
1216
1217 if (!rxport || !rxport->vpoc)
1218 continue;
1219
1220 regulator_disable(rxport->vpoc);
1221 }
1222
1223 return ret;
1224 }
1225
ub960_rxport_disable_vpocs(struct ub960_data * priv)1226 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1227 {
1228 unsigned int nport;
1229
1230 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1231 struct ub960_rxport *rxport = priv->rxports[nport];
1232
1233 if (!rxport || !rxport->vpoc)
1234 continue;
1235
1236 regulator_disable(rxport->vpoc);
1237 }
1238 }
1239
ub960_rxport_clear_errors(struct ub960_data * priv,unsigned int nport)1240 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1241 unsigned int nport)
1242 {
1243 u8 v;
1244
1245 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1246 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1247 ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1248 ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1249
1250 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1251 ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1252
1253 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1254 }
1255
ub960_clear_rx_errors(struct ub960_data * priv)1256 static void ub960_clear_rx_errors(struct ub960_data *priv)
1257 {
1258 unsigned int nport;
1259
1260 for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1261 ub960_rxport_clear_errors(priv, nport);
1262 }
1263
ub960_rxport_get_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 * strobe_pos)1264 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1265 unsigned int nport, s8 *strobe_pos)
1266 {
1267 u8 v;
1268 u8 clk_delay, data_delay;
1269 int ret;
1270
1271 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1272 UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1273
1274 clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1275 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1276
1277 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1278 UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1279
1280 data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1281 0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1282
1283 ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1284 if (ret)
1285 return ret;
1286
1287 clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1288
1289 ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1290 if (ret)
1291 return ret;
1292
1293 data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1294
1295 *strobe_pos = data_delay - clk_delay;
1296
1297 return 0;
1298 }
1299
ub960_rxport_set_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 strobe_pos)1300 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1301 unsigned int nport, s8 strobe_pos)
1302 {
1303 u8 clk_delay, data_delay;
1304
1305 clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1306 data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1307
1308 if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1309 clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1310 else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1311 data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1312 else if (strobe_pos < 0)
1313 clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1314 else if (strobe_pos > 0)
1315 data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1316
1317 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1318 UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1319
1320 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1321 UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1322 }
1323
ub960_rxport_set_strobe_range(struct ub960_data * priv,s8 strobe_min,s8 strobe_max)1324 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1325 s8 strobe_min, s8 strobe_max)
1326 {
1327 /* Convert the signed strobe pos to positive zero based value */
1328 strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1329 strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1330
1331 ub960_write(priv, UB960_XR_SFILTER_CFG,
1332 ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1333 ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1334 }
1335
ub960_rxport_get_eq_level(struct ub960_data * priv,unsigned int nport,u8 * eq_level)1336 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1337 unsigned int nport, u8 *eq_level)
1338 {
1339 int ret;
1340 u8 v;
1341
1342 ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1343 if (ret)
1344 return ret;
1345
1346 *eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1347 (v & UB960_RR_AEQ_STATUS_STATUS_2);
1348
1349 return 0;
1350 }
1351
ub960_rxport_set_eq_level(struct ub960_data * priv,unsigned int nport,u8 eq_level)1352 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1353 unsigned int nport, u8 eq_level)
1354 {
1355 u8 eq_stage_1_select_value, eq_stage_2_select_value;
1356 const unsigned int eq_stage_max = 7;
1357 u8 v;
1358
1359 if (eq_level <= eq_stage_max) {
1360 eq_stage_1_select_value = eq_level;
1361 eq_stage_2_select_value = 0;
1362 } else {
1363 eq_stage_1_select_value = eq_stage_max;
1364 eq_stage_2_select_value = eq_level - eq_stage_max;
1365 }
1366
1367 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1368
1369 v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1370 UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1371 v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1372 v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1373 v |= UB960_RR_AEQ_BYPASS_ENABLE;
1374
1375 ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1376 }
1377
ub960_rxport_set_eq_range(struct ub960_data * priv,unsigned int nport,u8 eq_min,u8 eq_max)1378 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1379 unsigned int nport, u8 eq_min, u8 eq_max)
1380 {
1381 ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1382 (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1383 (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1384
1385 /* Enable AEQ min setting */
1386 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1387 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1388 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1389 }
1390
ub960_rxport_config_eq(struct ub960_data * priv,unsigned int nport)1391 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1392 {
1393 struct ub960_rxport *rxport = priv->rxports[nport];
1394
1395 /* We also set common settings here. Should be moved elsewhere. */
1396
1397 if (priv->strobe.manual) {
1398 /* Disable AEQ_SFILTER_EN */
1399 ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1400 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1401 } else {
1402 /* Enable SFILTER and error control */
1403 ub960_write(priv, UB960_XR_AEQ_CTL1,
1404 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1405 UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1406
1407 /* Set AEQ strobe range */
1408 ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1409 priv->strobe.max);
1410 }
1411
1412 /* The rest are port specific */
1413
1414 if (priv->strobe.manual)
1415 ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1416 else
1417 ub960_rxport_set_strobe_pos(priv, nport, 0);
1418
1419 if (rxport->eq.manual_eq) {
1420 ub960_rxport_set_eq_level(priv, nport,
1421 rxport->eq.manual.eq_level);
1422
1423 /* Enable AEQ Bypass */
1424 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1425 UB960_RR_AEQ_BYPASS_ENABLE,
1426 UB960_RR_AEQ_BYPASS_ENABLE);
1427 } else {
1428 ub960_rxport_set_eq_range(priv, nport,
1429 rxport->eq.aeq.eq_level_min,
1430 rxport->eq.aeq.eq_level_max);
1431
1432 /* Disable AEQ Bypass */
1433 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1434 UB960_RR_AEQ_BYPASS_ENABLE, 0);
1435 }
1436 }
1437
ub960_rxport_link_ok(struct ub960_data * priv,unsigned int nport,bool * ok)1438 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1439 bool *ok)
1440 {
1441 u8 rx_port_sts1, rx_port_sts2;
1442 u16 parity_errors;
1443 u8 csi_rx_sts;
1444 u8 csi_err_cnt;
1445 u8 bcc_sts;
1446 int ret;
1447 bool errors;
1448
1449 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1450 &rx_port_sts1);
1451 if (ret)
1452 return ret;
1453
1454 if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1455 *ok = false;
1456 return 0;
1457 }
1458
1459 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1460 &rx_port_sts2);
1461 if (ret)
1462 return ret;
1463
1464 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1465 if (ret)
1466 return ret;
1467
1468 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1469 &csi_err_cnt);
1470 if (ret)
1471 return ret;
1472
1473 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1474 if (ret)
1475 return ret;
1476
1477 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1478 &parity_errors);
1479 if (ret)
1480 return ret;
1481
1482 errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1483 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1484 (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1485 (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1486 parity_errors;
1487
1488 *ok = !errors;
1489
1490 return 0;
1491 }
1492
1493 /*
1494 * Wait for the RX ports to lock, have no errors and have stable strobe position
1495 * and EQ level.
1496 */
ub960_rxport_wait_locks(struct ub960_data * priv,unsigned long port_mask,unsigned int * lock_mask)1497 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1498 unsigned long port_mask,
1499 unsigned int *lock_mask)
1500 {
1501 struct device *dev = &priv->client->dev;
1502 unsigned long timeout;
1503 unsigned int link_ok_mask;
1504 unsigned int missing;
1505 unsigned int loops;
1506 u8 nport;
1507 int ret;
1508
1509 if (port_mask == 0) {
1510 if (lock_mask)
1511 *lock_mask = 0;
1512 return 0;
1513 }
1514
1515 if (port_mask >= BIT(priv->hw_data->num_rxports))
1516 return -EINVAL;
1517
1518 timeout = jiffies + msecs_to_jiffies(1000);
1519 loops = 0;
1520 link_ok_mask = 0;
1521
1522 while (time_before(jiffies, timeout)) {
1523 missing = 0;
1524
1525 for_each_set_bit(nport, &port_mask,
1526 priv->hw_data->num_rxports) {
1527 struct ub960_rxport *rxport = priv->rxports[nport];
1528 bool ok;
1529
1530 if (!rxport)
1531 continue;
1532
1533 ret = ub960_rxport_link_ok(priv, nport, &ok);
1534 if (ret)
1535 return ret;
1536
1537 /*
1538 * We want the link to be ok for two consecutive loops,
1539 * as a link could get established just before our test
1540 * and drop soon after.
1541 */
1542 if (!ok || !(link_ok_mask & BIT(nport)))
1543 missing++;
1544
1545 if (ok)
1546 link_ok_mask |= BIT(nport);
1547 else
1548 link_ok_mask &= ~BIT(nport);
1549 }
1550
1551 loops++;
1552
1553 if (missing == 0)
1554 break;
1555
1556 msleep(50);
1557 }
1558
1559 if (lock_mask)
1560 *lock_mask = link_ok_mask;
1561
1562 dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1563 for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1564 struct ub960_rxport *rxport = priv->rxports[nport];
1565 s8 strobe_pos, eq_level;
1566 u16 v;
1567
1568 if (!rxport)
1569 continue;
1570
1571 if (!(link_ok_mask & BIT(nport))) {
1572 dev_dbg(dev, "\trx%u: not locked\n", nport);
1573 continue;
1574 }
1575
1576 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1577
1578 ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
1579 if (ret)
1580 return ret;
1581
1582 ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1583 if (ret)
1584 return ret;
1585
1586 dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1587 nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
1588 }
1589
1590 return 0;
1591 }
1592
ub960_calc_bc_clk_rate_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1593 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1594 struct ub960_rxport *rxport)
1595 {
1596 unsigned int mult;
1597 unsigned int div;
1598
1599 switch (rxport->rx_mode) {
1600 case RXPORT_MODE_RAW10:
1601 case RXPORT_MODE_RAW12_HF:
1602 case RXPORT_MODE_RAW12_LF:
1603 mult = 1;
1604 div = 10;
1605 break;
1606
1607 case RXPORT_MODE_CSI2_SYNC:
1608 mult = 2;
1609 div = 1;
1610 break;
1611
1612 case RXPORT_MODE_CSI2_NONSYNC:
1613 mult = 2;
1614 div = 5;
1615 break;
1616
1617 default:
1618 return 0;
1619 }
1620
1621 return clk_get_rate(priv->refclk) * mult / div;
1622 }
1623
ub960_calc_bc_clk_rate_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)1624 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1625 struct ub960_rxport *rxport)
1626 {
1627 switch (rxport->rx_mode) {
1628 case RXPORT_MODE_RAW10:
1629 case RXPORT_MODE_RAW12_HF:
1630 case RXPORT_MODE_RAW12_LF:
1631 return 2359400;
1632
1633 case RXPORT_MODE_CSI2_SYNC:
1634 return 47187500;
1635
1636 case RXPORT_MODE_CSI2_NONSYNC:
1637 return 9437500;
1638
1639 default:
1640 return 0;
1641 }
1642 }
1643
ub960_rxport_add_serializer(struct ub960_data * priv,u8 nport)1644 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1645 {
1646 struct ub960_rxport *rxport = priv->rxports[nport];
1647 struct device *dev = &priv->client->dev;
1648 struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1649 struct i2c_board_info ser_info = {
1650 .of_node = to_of_node(rxport->ser.fwnode),
1651 .fwnode = rxport->ser.fwnode,
1652 .platform_data = ser_pdata,
1653 };
1654
1655 ser_pdata->port = nport;
1656 ser_pdata->atr = priv->atr;
1657 if (priv->hw_data->is_ub9702)
1658 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1659 else
1660 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1661
1662 /*
1663 * The serializer is added under the same i2c adapter as the
1664 * deserializer. This is not quite right, as the serializer is behind
1665 * the FPD-Link.
1666 */
1667 ser_info.addr = rxport->ser.alias;
1668 rxport->ser.client =
1669 i2c_new_client_device(priv->client->adapter, &ser_info);
1670 if (IS_ERR(rxport->ser.client)) {
1671 dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1672 ser_info.type);
1673 return PTR_ERR(rxport->ser.client);
1674 }
1675
1676 dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1677 nport, rxport->ser.client->addr,
1678 rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1679
1680 return 0;
1681 }
1682
ub960_rxport_remove_serializer(struct ub960_data * priv,u8 nport)1683 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1684 {
1685 struct ub960_rxport *rxport = priv->rxports[nport];
1686
1687 i2c_unregister_device(rxport->ser.client);
1688 rxport->ser.client = NULL;
1689 }
1690
1691 /* Add serializer i2c devices for all initialized ports */
ub960_rxport_add_serializers(struct ub960_data * priv)1692 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1693 {
1694 unsigned int nport;
1695 int ret;
1696
1697 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1698 struct ub960_rxport *rxport = priv->rxports[nport];
1699
1700 if (!rxport)
1701 continue;
1702
1703 ret = ub960_rxport_add_serializer(priv, nport);
1704 if (ret)
1705 goto err_remove_sers;
1706 }
1707
1708 return 0;
1709
1710 err_remove_sers:
1711 while (nport--) {
1712 struct ub960_rxport *rxport = priv->rxports[nport];
1713
1714 if (!rxport)
1715 continue;
1716
1717 ub960_rxport_remove_serializer(priv, nport);
1718 }
1719
1720 return ret;
1721 }
1722
ub960_rxport_remove_serializers(struct ub960_data * priv)1723 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1724 {
1725 unsigned int nport;
1726
1727 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1728 struct ub960_rxport *rxport = priv->rxports[nport];
1729
1730 if (!rxport)
1731 continue;
1732
1733 ub960_rxport_remove_serializer(priv, nport);
1734 }
1735 }
1736
ub960_init_tx_port(struct ub960_data * priv,struct ub960_txport * txport)1737 static void ub960_init_tx_port(struct ub960_data *priv,
1738 struct ub960_txport *txport)
1739 {
1740 unsigned int nport = txport->nport;
1741 u8 csi_ctl = 0;
1742
1743 /*
1744 * From the datasheet: "initial CSI Skew-Calibration
1745 * sequence [...] should be set when operating at 1.6 Gbps"
1746 */
1747 if (priv->tx_data_rate == MHZ(1600))
1748 csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1749
1750 csi_ctl |= (4 - txport->num_data_lanes) << 4;
1751
1752 if (!txport->non_continous_clk)
1753 csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK;
1754
1755 ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1756 }
1757
ub960_init_tx_ports(struct ub960_data * priv)1758 static int ub960_init_tx_ports(struct ub960_data *priv)
1759 {
1760 unsigned int nport;
1761 u8 speed_select;
1762 u8 pll_div;
1763
1764 /* TX ports */
1765
1766 switch (priv->tx_data_rate) {
1767 case MHZ(1600):
1768 default:
1769 speed_select = 0;
1770 pll_div = 0x10;
1771 break;
1772 case MHZ(1200):
1773 speed_select = 1;
1774 pll_div = 0x18;
1775 break;
1776 case MHZ(800):
1777 speed_select = 2;
1778 pll_div = 0x10;
1779 break;
1780 case MHZ(400):
1781 speed_select = 3;
1782 pll_div = 0x10;
1783 break;
1784 }
1785
1786 ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1787
1788 if (priv->hw_data->is_ub9702) {
1789 ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1790
1791 switch (priv->tx_data_rate) {
1792 case MHZ(1600):
1793 default:
1794 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1795 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1796 break;
1797 case MHZ(800):
1798 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1799 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1800 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1801 break;
1802 case MHZ(400):
1803 ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1804 break;
1805 }
1806 }
1807
1808 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1809 struct ub960_txport *txport = priv->txports[nport];
1810
1811 if (!txport)
1812 continue;
1813
1814 ub960_init_tx_port(priv, txport);
1815 }
1816
1817 return 0;
1818 }
1819
ub960_init_rx_port_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1820 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1821 struct ub960_rxport *rxport)
1822 {
1823 unsigned int nport = rxport->nport;
1824 u32 bc_freq_val;
1825
1826 /*
1827 * Back channel frequency select.
1828 * Override FREQ_SELECT from the strap.
1829 * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1830 * 2 - 10 Mbps
1831 * 6 - 50 Mbps (DS90UB953-Q1)
1832 *
1833 * Note that changing this setting will result in some errors on the back
1834 * channel for a short period of time.
1835 */
1836
1837 switch (rxport->rx_mode) {
1838 case RXPORT_MODE_RAW10:
1839 case RXPORT_MODE_RAW12_HF:
1840 case RXPORT_MODE_RAW12_LF:
1841 bc_freq_val = 0;
1842 break;
1843
1844 case RXPORT_MODE_CSI2_NONSYNC:
1845 bc_freq_val = 2;
1846 break;
1847
1848 case RXPORT_MODE_CSI2_SYNC:
1849 bc_freq_val = 6;
1850 break;
1851
1852 default:
1853 return;
1854 }
1855
1856 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1857 UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1858 bc_freq_val);
1859
1860 switch (rxport->rx_mode) {
1861 case RXPORT_MODE_RAW10:
1862 /* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1863 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1864 UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1865 0x3);
1866
1867 /*
1868 * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1869 */
1870 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1871 UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1872 0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1873
1874 break;
1875
1876 case RXPORT_MODE_RAW12_HF:
1877 case RXPORT_MODE_RAW12_LF:
1878 /* Not implemented */
1879 return;
1880
1881 case RXPORT_MODE_CSI2_SYNC:
1882 case RXPORT_MODE_CSI2_NONSYNC:
1883 /* CSI-2 Mode (DS90UB953-Q1 compatible) */
1884 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1885 0x0);
1886
1887 break;
1888 }
1889
1890 /* LV_POLARITY & FV_POLARITY */
1891 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1892 rxport->lv_fv_pol);
1893
1894 /* Enable all interrupt sources from this port */
1895 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1896 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1897
1898 /* Enable I2C_PASS_THROUGH */
1899 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1900 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1901 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1902
1903 /* Enable I2C communication to the serializer via the alias addr */
1904 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1905 rxport->ser.alias << 1);
1906
1907 /* Configure EQ related settings */
1908 ub960_rxport_config_eq(priv, nport);
1909
1910 /* Enable RX port */
1911 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1912 }
1913
ub960_init_rx_port_ub9702_fpd3(struct ub960_data * priv,struct ub960_rxport * rxport)1914 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1915 struct ub960_rxport *rxport)
1916 {
1917 unsigned int nport = rxport->nport;
1918 u8 bc_freq_val;
1919 u8 fpd_func_mode;
1920
1921 switch (rxport->rx_mode) {
1922 case RXPORT_MODE_RAW10:
1923 bc_freq_val = 0;
1924 fpd_func_mode = 5;
1925 break;
1926
1927 case RXPORT_MODE_RAW12_HF:
1928 bc_freq_val = 0;
1929 fpd_func_mode = 4;
1930 break;
1931
1932 case RXPORT_MODE_RAW12_LF:
1933 bc_freq_val = 0;
1934 fpd_func_mode = 6;
1935 break;
1936
1937 case RXPORT_MODE_CSI2_SYNC:
1938 bc_freq_val = 6;
1939 fpd_func_mode = 2;
1940 break;
1941
1942 case RXPORT_MODE_CSI2_NONSYNC:
1943 bc_freq_val = 2;
1944 fpd_func_mode = 2;
1945 break;
1946
1947 default:
1948 return;
1949 }
1950
1951 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1952 bc_freq_val);
1953 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1954
1955 /* set serdes_eq_mode = 1 */
1956 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1957
1958 /* enable serdes driver */
1959 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1960
1961 /* set serdes_eq_offset=4 */
1962 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1963
1964 /* init default serdes_eq_max in 0xa9 */
1965 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
1966
1967 /* init serdes_eq_min in 0xaa */
1968 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
1969
1970 /* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
1971 ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
1972 BIT(3), BIT(3));
1973
1974 /* RX port to half-rate */
1975 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
1976 BIT(nport * 2));
1977 }
1978
ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data * priv,struct ub960_rxport * rxport)1979 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
1980 struct ub960_rxport *rxport)
1981 {
1982 unsigned int nport = rxport->nport;
1983 bool first_time_power_up = true;
1984
1985 if (first_time_power_up) {
1986 u8 v;
1987
1988 /* AEQ init */
1989 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
1990
1991 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
1992 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
1993
1994 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
1995 }
1996
1997 /* enable serdes_eq_ctl2 */
1998 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
1999
2000 /* enable serdes_eq_ctl1 */
2001 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
2002
2003 /* enable serdes_eq_en */
2004 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
2005
2006 /* disable serdes_eq_override */
2007 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
2008
2009 /* disable serdes_gain_override */
2010 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2011 }
2012
ub960_init_rx_port_ub9702_fpd4(struct ub960_data * priv,struct ub960_rxport * rxport)2013 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2014 struct ub960_rxport *rxport)
2015 {
2016 unsigned int nport = rxport->nport;
2017 u8 bc_freq_val;
2018
2019 switch (rxport->rx_mode) {
2020 case RXPORT_MODE_RAW10:
2021 bc_freq_val = 0;
2022 break;
2023
2024 case RXPORT_MODE_RAW12_HF:
2025 bc_freq_val = 0;
2026 break;
2027
2028 case RXPORT_MODE_RAW12_LF:
2029 bc_freq_val = 0;
2030 break;
2031
2032 case RXPORT_MODE_CSI2_SYNC:
2033 bc_freq_val = 6;
2034 break;
2035
2036 case RXPORT_MODE_CSI2_NONSYNC:
2037 bc_freq_val = 2;
2038 break;
2039
2040 default:
2041 return;
2042 }
2043
2044 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2045 bc_freq_val);
2046
2047 /* FPD4 Sync Mode */
2048 ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2049
2050 /* add serdes_eq_offset of 4 */
2051 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2052
2053 /* FPD4 serdes_start_eq in 0x27: assign default */
2054 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2055 /* FPD4 serdes_end_eq in 0x28: assign default */
2056 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2057
2058 /* set serdes_driver_mode into FPD IV mode */
2059 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2060 /* set FPD PBC drv into FPD IV mode */
2061 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2062
2063 /* set serdes_system_init to 0x2f */
2064 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2065 /* set serdes_system_rst in reset mode */
2066 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2067
2068 /* RX port to 7.55G mode */
2069 ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2070 0 << (nport * 2));
2071
2072 ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2073 }
2074
ub960_init_rx_port_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)2075 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2076 struct ub960_rxport *rxport)
2077 {
2078 unsigned int nport = rxport->nport;
2079
2080 if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2081 ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2082 else /* RXPORT_CDR_FPD4 */
2083 ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2084
2085 switch (rxport->rx_mode) {
2086 case RXPORT_MODE_RAW10:
2087 /*
2088 * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2089 * 0b10 : 8-bit processing using upper 8 bits
2090 */
2091 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2092 0x3 << 6, 0x2 << 6);
2093
2094 break;
2095
2096 case RXPORT_MODE_RAW12_HF:
2097 case RXPORT_MODE_RAW12_LF:
2098 /* Not implemented */
2099 return;
2100
2101 case RXPORT_MODE_CSI2_SYNC:
2102 case RXPORT_MODE_CSI2_NONSYNC:
2103
2104 break;
2105 }
2106
2107 /* LV_POLARITY & FV_POLARITY */
2108 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2109 rxport->lv_fv_pol);
2110
2111 /* Enable all interrupt sources from this port */
2112 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2113 ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2114
2115 /* Enable I2C_PASS_THROUGH */
2116 ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2117 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2118 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2119
2120 /* Enable I2C communication to the serializer via the alias addr */
2121 ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2122 rxport->ser.alias << 1);
2123
2124 /* Enable RX port */
2125 ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2126
2127 if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2128 /* unreset 960 AEQ */
2129 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2130 }
2131 }
2132
ub960_init_rx_ports(struct ub960_data * priv)2133 static int ub960_init_rx_ports(struct ub960_data *priv)
2134 {
2135 unsigned int nport;
2136
2137 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2138 struct ub960_rxport *rxport = priv->rxports[nport];
2139
2140 if (!rxport)
2141 continue;
2142
2143 if (priv->hw_data->is_ub9702)
2144 ub960_init_rx_port_ub9702(priv, rxport);
2145 else
2146 ub960_init_rx_port_ub960(priv, rxport);
2147 }
2148
2149 return 0;
2150 }
2151
ub960_rxport_handle_events(struct ub960_data * priv,u8 nport)2152 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2153 {
2154 struct device *dev = &priv->client->dev;
2155 u8 rx_port_sts1;
2156 u8 rx_port_sts2;
2157 u8 csi_rx_sts;
2158 u8 bcc_sts;
2159 int ret = 0;
2160
2161 /* Read interrupts (also clears most of them) */
2162 if (!ret)
2163 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2164 &rx_port_sts1);
2165 if (!ret)
2166 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2167 &rx_port_sts2);
2168 if (!ret)
2169 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2170 &csi_rx_sts);
2171 if (!ret)
2172 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2173 &bcc_sts);
2174
2175 if (ret)
2176 return;
2177
2178 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2179 u16 v;
2180
2181 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2182 &v);
2183 if (!ret)
2184 dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2185 }
2186
2187 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2188 dev_err(dev, "rx%u BCC CRC error\n", nport);
2189
2190 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2191 dev_err(dev, "rx%u BCC SEQ error\n", nport);
2192
2193 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2194 dev_err(dev, "rx%u line length unstable\n", nport);
2195
2196 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2197 dev_err(dev, "rx%u FPD3 encode error\n", nport);
2198
2199 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2200 dev_err(dev, "rx%u buffer error\n", nport);
2201
2202 if (csi_rx_sts)
2203 dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2204
2205 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2206 dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2207
2208 if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2209 dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2210
2211 if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2212 dev_err(dev, "rx%u CSI checksum error\n", nport);
2213
2214 if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2215 dev_err(dev, "rx%u CSI length error\n", nport);
2216
2217 if (bcc_sts)
2218 dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2219
2220 if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2221 dev_err(dev, "rx%u BCC response error", nport);
2222
2223 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2224 dev_err(dev, "rx%u BCC slave timeout", nport);
2225
2226 if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2227 dev_err(dev, "rx%u BCC slave error", nport);
2228
2229 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2230 dev_err(dev, "rx%u BCC master timeout", nport);
2231
2232 if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2233 dev_err(dev, "rx%u BCC master error", nport);
2234
2235 if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2236 dev_err(dev, "rx%u BCC sequence error", nport);
2237
2238 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2239 u16 v;
2240
2241 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2242 if (!ret)
2243 dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2244 }
2245
2246 if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2247 u16 v;
2248
2249 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2250 &v);
2251 if (!ret)
2252 dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2253 }
2254
2255 if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2256 dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2257 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2258 "locked" :
2259 "unlocked",
2260 (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2261 "passed" :
2262 "not passed",
2263 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2264 "no clock" :
2265 "clock ok",
2266 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2267 "stable freq" :
2268 "unstable freq");
2269 }
2270 }
2271
2272 /* -----------------------------------------------------------------------------
2273 * V4L2
2274 */
2275
2276 /*
2277 * The current implementation only supports a simple VC mapping, where all VCs
2278 * from a one RX port will be mapped to the same VC. Also, the hardware
2279 * dictates that all streams from an RX port must go to a single TX port.
2280 *
2281 * This function decides the target VC numbers for each RX port with a simple
2282 * algorithm, so that for each TX port, we get VC numbers starting from 0,
2283 * and counting up.
2284 *
2285 * E.g. if all four RX ports are in use, of which the first two go to the
2286 * first TX port and the secont two go to the second TX port, we would get
2287 * the following VCs for the four RX ports: 0, 1, 0, 1.
2288 *
2289 * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2290 * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2291 * probably needs to be somehow configurable. Device tree?
2292 */
ub960_get_vc_maps(struct ub960_data * priv,struct v4l2_subdev_state * state,u8 * vc)2293 static void ub960_get_vc_maps(struct ub960_data *priv,
2294 struct v4l2_subdev_state *state, u8 *vc)
2295 {
2296 u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2297 struct v4l2_subdev_route *route;
2298 u8 handled_mask = 0;
2299
2300 for_each_active_route(&state->routing, route) {
2301 unsigned int rx, tx;
2302
2303 rx = ub960_pad_to_port(priv, route->sink_pad);
2304 if (BIT(rx) & handled_mask)
2305 continue;
2306
2307 tx = ub960_pad_to_port(priv, route->source_pad);
2308
2309 vc[rx] = cur_vc[tx]++;
2310 handled_mask |= BIT(rx);
2311 }
2312 }
2313
ub960_enable_tx_port(struct ub960_data * priv,unsigned int nport)2314 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2315 {
2316 struct device *dev = &priv->client->dev;
2317
2318 dev_dbg(dev, "enable TX port %u\n", nport);
2319
2320 return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2321 UB960_TR_CSI_CTL_CSI_ENABLE,
2322 UB960_TR_CSI_CTL_CSI_ENABLE);
2323 }
2324
ub960_disable_tx_port(struct ub960_data * priv,unsigned int nport)2325 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2326 {
2327 struct device *dev = &priv->client->dev;
2328
2329 dev_dbg(dev, "disable TX port %u\n", nport);
2330
2331 ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2332 UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2333 }
2334
ub960_enable_rx_port(struct ub960_data * priv,unsigned int nport)2335 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2336 {
2337 struct device *dev = &priv->client->dev;
2338
2339 dev_dbg(dev, "enable RX port %u\n", nport);
2340
2341 /* Enable forwarding */
2342 return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2343 UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2344 }
2345
ub960_disable_rx_port(struct ub960_data * priv,unsigned int nport)2346 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2347 {
2348 struct device *dev = &priv->client->dev;
2349
2350 dev_dbg(dev, "disable RX port %u\n", nport);
2351
2352 /* Disable forwarding */
2353 ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2354 UB960_SR_FWD_CTL1_PORT_DIS(nport),
2355 UB960_SR_FWD_CTL1_PORT_DIS(nport));
2356 }
2357
2358 /*
2359 * The driver only supports using a single VC for each source. This function
2360 * checks that each source only provides streams using a single VC.
2361 */
ub960_validate_stream_vcs(struct ub960_data * priv)2362 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2363 {
2364 unsigned int nport;
2365 unsigned int i;
2366
2367 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2368 struct ub960_rxport *rxport = priv->rxports[nport];
2369 struct v4l2_mbus_frame_desc desc;
2370 int ret;
2371 u8 vc;
2372
2373 if (!rxport)
2374 continue;
2375
2376 ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2377 rxport->source.pad, &desc);
2378 if (ret)
2379 return ret;
2380
2381 if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2382 continue;
2383
2384 if (desc.num_entries == 0)
2385 continue;
2386
2387 vc = desc.entry[0].bus.csi2.vc;
2388
2389 for (i = 1; i < desc.num_entries; i++) {
2390 if (vc == desc.entry[i].bus.csi2.vc)
2391 continue;
2392
2393 dev_err(&priv->client->dev,
2394 "rx%u: source with multiple virtual-channels is not supported\n",
2395 nport);
2396 return -ENODEV;
2397 }
2398 }
2399
2400 return 0;
2401 }
2402
ub960_configure_ports_for_streaming(struct ub960_data * priv,struct v4l2_subdev_state * state)2403 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2404 struct v4l2_subdev_state *state)
2405 {
2406 u8 fwd_ctl;
2407 struct {
2408 u32 num_streams;
2409 u8 pixel_dt;
2410 u8 meta_dt;
2411 u32 meta_lines;
2412 u32 tx_port;
2413 } rx_data[UB960_MAX_RX_NPORTS] = {};
2414 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2415 struct v4l2_subdev_route *route;
2416 unsigned int nport;
2417 int ret;
2418
2419 ret = ub960_validate_stream_vcs(priv);
2420 if (ret)
2421 return ret;
2422
2423 ub960_get_vc_maps(priv, state, vc_map);
2424
2425 for_each_active_route(&state->routing, route) {
2426 struct ub960_rxport *rxport;
2427 struct ub960_txport *txport;
2428 struct v4l2_mbus_framefmt *fmt;
2429 const struct ub960_format_info *ub960_fmt;
2430 unsigned int nport;
2431
2432 nport = ub960_pad_to_port(priv, route->sink_pad);
2433
2434 rxport = priv->rxports[nport];
2435 if (!rxport)
2436 return -EINVAL;
2437
2438 txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2439 if (!txport)
2440 return -EINVAL;
2441
2442 rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2443
2444 rx_data[nport].num_streams++;
2445
2446 /* For the rest, we are only interested in parallel busses */
2447 if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2448 rxport->rx_mode == RXPORT_MODE_CSI2_NONSYNC)
2449 continue;
2450
2451 if (rx_data[nport].num_streams > 2)
2452 return -EPIPE;
2453
2454 fmt = v4l2_subdev_state_get_format(state, route->sink_pad,
2455 route->sink_stream);
2456 if (!fmt)
2457 return -EPIPE;
2458
2459 ub960_fmt = ub960_find_format(fmt->code);
2460 if (!ub960_fmt)
2461 return -EPIPE;
2462
2463 if (ub960_fmt->meta) {
2464 if (fmt->height > 3) {
2465 dev_err(&priv->client->dev,
2466 "rx%u: unsupported metadata height %u\n",
2467 nport, fmt->height);
2468 return -EPIPE;
2469 }
2470
2471 rx_data[nport].meta_dt = ub960_fmt->datatype;
2472 rx_data[nport].meta_lines = fmt->height;
2473 } else {
2474 rx_data[nport].pixel_dt = ub960_fmt->datatype;
2475 }
2476 }
2477
2478 /* Configure RX ports */
2479
2480 /*
2481 * Keep all port forwardings disabled by default. Forwarding will be
2482 * enabled in ub960_enable_rx_port.
2483 */
2484 fwd_ctl = GENMASK(7, 4);
2485
2486 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2487 struct ub960_rxport *rxport = priv->rxports[nport];
2488 u8 vc = vc_map[nport];
2489
2490 if (rx_data[nport].num_streams == 0)
2491 continue;
2492
2493 switch (rxport->rx_mode) {
2494 case RXPORT_MODE_RAW10:
2495 ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2496 rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2497
2498 ub960_rxport_write(priv, rxport->nport,
2499 UB960_RR_RAW_EMBED_DTYPE,
2500 (rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2501 rx_data[nport].meta_dt);
2502
2503 break;
2504
2505 case RXPORT_MODE_RAW12_HF:
2506 case RXPORT_MODE_RAW12_LF:
2507 /* Not implemented */
2508 break;
2509
2510 case RXPORT_MODE_CSI2_SYNC:
2511 case RXPORT_MODE_CSI2_NONSYNC:
2512 if (!priv->hw_data->is_ub9702) {
2513 /* Map all VCs from this port to the same VC */
2514 ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2515 (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2516 (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2517 (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2518 (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2519 } else {
2520 unsigned int i;
2521
2522 /* Map all VCs from this port to VC(nport) */
2523 for (i = 0; i < 8; i++)
2524 ub960_rxport_write(priv, nport,
2525 UB960_RR_VC_ID_MAP(i),
2526 nport);
2527 }
2528
2529 break;
2530 }
2531
2532 if (rx_data[nport].tx_port == 1)
2533 fwd_ctl |= BIT(nport); /* forward to TX1 */
2534 else
2535 fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2536 }
2537
2538 ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2539
2540 return 0;
2541 }
2542
ub960_update_streaming_status(struct ub960_data * priv)2543 static void ub960_update_streaming_status(struct ub960_data *priv)
2544 {
2545 unsigned int i;
2546
2547 for (i = 0; i < UB960_MAX_NPORTS; i++) {
2548 if (priv->stream_enable_mask[i])
2549 break;
2550 }
2551
2552 priv->streaming = i < UB960_MAX_NPORTS;
2553 }
2554
ub960_enable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2555 static int ub960_enable_streams(struct v4l2_subdev *sd,
2556 struct v4l2_subdev_state *state, u32 source_pad,
2557 u64 source_streams_mask)
2558 {
2559 struct ub960_data *priv = sd_to_ub960(sd);
2560 struct device *dev = &priv->client->dev;
2561 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2562 struct v4l2_subdev_route *route;
2563 unsigned int failed_port;
2564 unsigned int nport;
2565 int ret;
2566
2567 if (!priv->streaming) {
2568 dev_dbg(dev, "Prepare for streaming\n");
2569 ret = ub960_configure_ports_for_streaming(priv, state);
2570 if (ret)
2571 return ret;
2572 }
2573
2574 /* Enable TX port if not yet enabled */
2575 if (!priv->stream_enable_mask[source_pad]) {
2576 ret = ub960_enable_tx_port(priv,
2577 ub960_pad_to_port(priv, source_pad));
2578 if (ret)
2579 return ret;
2580 }
2581
2582 priv->stream_enable_mask[source_pad] |= source_streams_mask;
2583
2584 /* Collect sink streams per pad which we need to enable */
2585 for_each_active_route(&state->routing, route) {
2586 if (route->source_pad != source_pad)
2587 continue;
2588
2589 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2590 continue;
2591
2592 nport = ub960_pad_to_port(priv, route->sink_pad);
2593
2594 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2595 }
2596
2597 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2598 if (!sink_streams[nport])
2599 continue;
2600
2601 /* Enable the RX port if not yet enabled */
2602 if (!priv->stream_enable_mask[nport]) {
2603 ret = ub960_enable_rx_port(priv, nport);
2604 if (ret) {
2605 failed_port = nport;
2606 goto err;
2607 }
2608 }
2609
2610 priv->stream_enable_mask[nport] |= sink_streams[nport];
2611
2612 dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2613 sink_streams[nport]);
2614
2615 ret = v4l2_subdev_enable_streams(
2616 priv->rxports[nport]->source.sd,
2617 priv->rxports[nport]->source.pad,
2618 sink_streams[nport]);
2619 if (ret) {
2620 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2621
2622 if (!priv->stream_enable_mask[nport])
2623 ub960_disable_rx_port(priv, nport);
2624
2625 failed_port = nport;
2626 goto err;
2627 }
2628 }
2629
2630 priv->streaming = true;
2631
2632 return 0;
2633
2634 err:
2635 for (nport = 0; nport < failed_port; nport++) {
2636 if (!sink_streams[nport])
2637 continue;
2638
2639 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2640 sink_streams[nport]);
2641
2642 ret = v4l2_subdev_disable_streams(
2643 priv->rxports[nport]->source.sd,
2644 priv->rxports[nport]->source.pad,
2645 sink_streams[nport]);
2646 if (ret)
2647 dev_err(dev, "Failed to disable streams: %d\n", ret);
2648
2649 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2650
2651 /* Disable RX port if no active streams */
2652 if (!priv->stream_enable_mask[nport])
2653 ub960_disable_rx_port(priv, nport);
2654 }
2655
2656 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2657
2658 if (!priv->stream_enable_mask[source_pad])
2659 ub960_disable_tx_port(priv,
2660 ub960_pad_to_port(priv, source_pad));
2661
2662 ub960_update_streaming_status(priv);
2663
2664 return ret;
2665 }
2666
ub960_disable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2667 static int ub960_disable_streams(struct v4l2_subdev *sd,
2668 struct v4l2_subdev_state *state,
2669 u32 source_pad, u64 source_streams_mask)
2670 {
2671 struct ub960_data *priv = sd_to_ub960(sd);
2672 struct device *dev = &priv->client->dev;
2673 u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2674 struct v4l2_subdev_route *route;
2675 unsigned int nport;
2676 int ret;
2677
2678 /* Collect sink streams per pad which we need to disable */
2679 for_each_active_route(&state->routing, route) {
2680 if (route->source_pad != source_pad)
2681 continue;
2682
2683 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2684 continue;
2685
2686 nport = ub960_pad_to_port(priv, route->sink_pad);
2687
2688 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2689 }
2690
2691 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2692 if (!sink_streams[nport])
2693 continue;
2694
2695 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2696 sink_streams[nport]);
2697
2698 ret = v4l2_subdev_disable_streams(
2699 priv->rxports[nport]->source.sd,
2700 priv->rxports[nport]->source.pad,
2701 sink_streams[nport]);
2702 if (ret)
2703 dev_err(dev, "Failed to disable streams: %d\n", ret);
2704
2705 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2706
2707 /* Disable RX port if no active streams */
2708 if (!priv->stream_enable_mask[nport])
2709 ub960_disable_rx_port(priv, nport);
2710 }
2711
2712 /* Disable TX port if no active streams */
2713
2714 priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2715
2716 if (!priv->stream_enable_mask[source_pad])
2717 ub960_disable_tx_port(priv,
2718 ub960_pad_to_port(priv, source_pad));
2719
2720 ub960_update_streaming_status(priv);
2721
2722 return 0;
2723 }
2724
_ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_krouting * routing)2725 static int _ub960_set_routing(struct v4l2_subdev *sd,
2726 struct v4l2_subdev_state *state,
2727 struct v4l2_subdev_krouting *routing)
2728 {
2729 static const struct v4l2_mbus_framefmt format = {
2730 .width = 640,
2731 .height = 480,
2732 .code = MEDIA_BUS_FMT_UYVY8_1X16,
2733 .field = V4L2_FIELD_NONE,
2734 .colorspace = V4L2_COLORSPACE_SRGB,
2735 .ycbcr_enc = V4L2_YCBCR_ENC_601,
2736 .quantization = V4L2_QUANTIZATION_LIM_RANGE,
2737 .xfer_func = V4L2_XFER_FUNC_SRGB,
2738 };
2739 int ret;
2740
2741 /*
2742 * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2743 * frame desc is made dynamically allocated.
2744 */
2745
2746 if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2747 return -E2BIG;
2748
2749 ret = v4l2_subdev_routing_validate(sd, routing,
2750 V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2751 V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2752 if (ret)
2753 return ret;
2754
2755 ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2756 if (ret)
2757 return ret;
2758
2759 return 0;
2760 }
2761
ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,enum v4l2_subdev_format_whence which,struct v4l2_subdev_krouting * routing)2762 static int ub960_set_routing(struct v4l2_subdev *sd,
2763 struct v4l2_subdev_state *state,
2764 enum v4l2_subdev_format_whence which,
2765 struct v4l2_subdev_krouting *routing)
2766 {
2767 struct ub960_data *priv = sd_to_ub960(sd);
2768
2769 if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2770 return -EBUSY;
2771
2772 return _ub960_set_routing(sd, state, routing);
2773 }
2774
ub960_get_frame_desc(struct v4l2_subdev * sd,unsigned int pad,struct v4l2_mbus_frame_desc * fd)2775 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2776 struct v4l2_mbus_frame_desc *fd)
2777 {
2778 struct ub960_data *priv = sd_to_ub960(sd);
2779 struct v4l2_subdev_route *route;
2780 struct v4l2_subdev_state *state;
2781 int ret = 0;
2782 struct device *dev = &priv->client->dev;
2783 u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2784
2785 if (!ub960_pad_is_source(priv, pad))
2786 return -EINVAL;
2787
2788 fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2789
2790 state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2791
2792 ub960_get_vc_maps(priv, state, vc_map);
2793
2794 for_each_active_route(&state->routing, route) {
2795 struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2796 struct v4l2_mbus_frame_desc source_fd;
2797 unsigned int nport;
2798 unsigned int i;
2799
2800 if (route->source_pad != pad)
2801 continue;
2802
2803 nport = ub960_pad_to_port(priv, route->sink_pad);
2804
2805 ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2806 get_frame_desc,
2807 priv->rxports[nport]->source.pad,
2808 &source_fd);
2809 if (ret) {
2810 dev_err(dev,
2811 "Failed to get source frame desc for pad %u\n",
2812 route->sink_pad);
2813 goto out_unlock;
2814 }
2815
2816 for (i = 0; i < source_fd.num_entries; i++) {
2817 if (source_fd.entry[i].stream == route->sink_stream) {
2818 source_entry = &source_fd.entry[i];
2819 break;
2820 }
2821 }
2822
2823 if (!source_entry) {
2824 dev_err(dev,
2825 "Failed to find stream from source frame desc\n");
2826 ret = -EPIPE;
2827 goto out_unlock;
2828 }
2829
2830 fd->entry[fd->num_entries].stream = route->source_stream;
2831 fd->entry[fd->num_entries].flags = source_entry->flags;
2832 fd->entry[fd->num_entries].length = source_entry->length;
2833 fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2834
2835 fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2836
2837 if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2838 fd->entry[fd->num_entries].bus.csi2.dt =
2839 source_entry->bus.csi2.dt;
2840 } else {
2841 const struct ub960_format_info *ub960_fmt;
2842 struct v4l2_mbus_framefmt *fmt;
2843
2844 fmt = v4l2_subdev_state_get_format(state, pad,
2845 route->source_stream);
2846
2847 if (!fmt) {
2848 ret = -EINVAL;
2849 goto out_unlock;
2850 }
2851
2852 ub960_fmt = ub960_find_format(fmt->code);
2853 if (!ub960_fmt) {
2854 dev_err(dev, "Unable to find format\n");
2855 ret = -EINVAL;
2856 goto out_unlock;
2857 }
2858
2859 fd->entry[fd->num_entries].bus.csi2.dt =
2860 ub960_fmt->datatype;
2861 }
2862
2863 fd->num_entries++;
2864 }
2865
2866 out_unlock:
2867 v4l2_subdev_unlock_state(state);
2868
2869 return ret;
2870 }
2871
ub960_set_fmt(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_format * format)2872 static int ub960_set_fmt(struct v4l2_subdev *sd,
2873 struct v4l2_subdev_state *state,
2874 struct v4l2_subdev_format *format)
2875 {
2876 struct ub960_data *priv = sd_to_ub960(sd);
2877 struct v4l2_mbus_framefmt *fmt;
2878
2879 if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2880 return -EBUSY;
2881
2882 /* No transcoding, source and sink formats must match. */
2883 if (ub960_pad_is_source(priv, format->pad))
2884 return v4l2_subdev_get_fmt(sd, state, format);
2885
2886 /*
2887 * Default to the first format if the requested media bus code isn't
2888 * supported.
2889 */
2890 if (!ub960_find_format(format->format.code))
2891 format->format.code = ub960_formats[0].code;
2892
2893 fmt = v4l2_subdev_state_get_format(state, format->pad, format->stream);
2894 if (!fmt)
2895 return -EINVAL;
2896
2897 *fmt = format->format;
2898
2899 fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2900 format->stream);
2901 if (!fmt)
2902 return -EINVAL;
2903
2904 *fmt = format->format;
2905
2906 return 0;
2907 }
2908
ub960_init_state(struct v4l2_subdev * sd,struct v4l2_subdev_state * state)2909 static int ub960_init_state(struct v4l2_subdev *sd,
2910 struct v4l2_subdev_state *state)
2911 {
2912 struct ub960_data *priv = sd_to_ub960(sd);
2913
2914 struct v4l2_subdev_route routes[] = {
2915 {
2916 .sink_pad = 0,
2917 .sink_stream = 0,
2918 .source_pad = priv->hw_data->num_rxports,
2919 .source_stream = 0,
2920 .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2921 },
2922 };
2923
2924 struct v4l2_subdev_krouting routing = {
2925 .num_routes = ARRAY_SIZE(routes),
2926 .routes = routes,
2927 };
2928
2929 return _ub960_set_routing(sd, state, &routing);
2930 }
2931
2932 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2933 .enable_streams = ub960_enable_streams,
2934 .disable_streams = ub960_disable_streams,
2935
2936 .set_routing = ub960_set_routing,
2937 .get_frame_desc = ub960_get_frame_desc,
2938
2939 .get_fmt = v4l2_subdev_get_fmt,
2940 .set_fmt = ub960_set_fmt,
2941 };
2942
ub960_log_status(struct v4l2_subdev * sd)2943 static int ub960_log_status(struct v4l2_subdev *sd)
2944 {
2945 struct ub960_data *priv = sd_to_ub960(sd);
2946 struct device *dev = &priv->client->dev;
2947 struct v4l2_subdev_state *state;
2948 unsigned int nport;
2949 unsigned int i;
2950 u16 v16 = 0;
2951 u8 v = 0;
2952 u8 id[UB960_SR_FPD3_RX_ID_LEN];
2953
2954 state = v4l2_subdev_lock_and_get_active_state(sd);
2955
2956 for (i = 0; i < sizeof(id); i++)
2957 ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
2958
2959 dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
2960
2961 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
2962 struct ub960_txport *txport = priv->txports[nport];
2963
2964 dev_info(dev, "TX %u\n", nport);
2965
2966 if (!txport) {
2967 dev_info(dev, "\tNot initialized\n");
2968 continue;
2969 }
2970
2971 ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
2972 dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
2973 v & (u8)BIT(0));
2974
2975 ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
2976 dev_info(dev, "\tframe counter %u\n", v16);
2977
2978 ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
2979 dev_info(dev, "\tframe error counter %u\n", v16);
2980
2981 ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
2982 dev_info(dev, "\tline counter %u\n", v16);
2983
2984 ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
2985 dev_info(dev, "\tline error counter %u\n", v16);
2986 }
2987
2988 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2989 struct ub960_rxport *rxport = priv->rxports[nport];
2990 u8 eq_level;
2991 s8 strobe_pos;
2992 unsigned int i;
2993
2994 dev_info(dev, "RX %u\n", nport);
2995
2996 if (!rxport) {
2997 dev_info(dev, "\tNot initialized\n");
2998 continue;
2999 }
3000
3001 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
3002
3003 if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3004 dev_info(dev, "\tLocked\n");
3005 else
3006 dev_info(dev, "\tNot locked\n");
3007
3008 dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3009 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3010 dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3011
3012 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3013 dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
3014
3015 ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3016 dev_info(dev, "\tparity errors %u\n", v16);
3017
3018 ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3019 dev_info(dev, "\tlines per frame %u\n", v16);
3020
3021 ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3022 dev_info(dev, "\tbytes per line %u\n", v16);
3023
3024 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3025 dev_info(dev, "\tcsi_err_counter %u\n", v);
3026
3027 /* Strobe */
3028
3029 ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
3030
3031 dev_info(dev, "\t%s strobe\n",
3032 (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
3033 "Manual");
3034
3035 if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
3036 ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
3037
3038 dev_info(dev, "\tStrobe range [%d, %d]\n",
3039 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
3040 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
3041 }
3042
3043 ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
3044
3045 dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
3046
3047 /* EQ */
3048
3049 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
3050
3051 dev_info(dev, "\t%s EQ\n",
3052 (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
3053 "Adaptive");
3054
3055 if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
3056 ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
3057
3058 dev_info(dev, "\tEQ range [%u, %u]\n",
3059 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
3060 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
3061 }
3062
3063 if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
3064 dev_info(dev, "\tEQ level %u\n", eq_level);
3065
3066 /* GPIOs */
3067 for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3068 u8 ctl_reg;
3069 u8 ctl_shift;
3070
3071 ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3072 ctl_shift = (i % 2) * 4;
3073
3074 ub960_rxport_read(priv, nport, ctl_reg, &v);
3075
3076 dev_info(dev, "\tGPIO%u: mode %u\n", i,
3077 (v >> ctl_shift) & 0xf);
3078 }
3079 }
3080
3081 v4l2_subdev_unlock_state(state);
3082
3083 return 0;
3084 }
3085
3086 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3087 .log_status = ub960_log_status,
3088 .subscribe_event = v4l2_ctrl_subdev_subscribe_event,
3089 .unsubscribe_event = v4l2_event_subdev_unsubscribe,
3090 };
3091
3092 static const struct v4l2_subdev_internal_ops ub960_internal_ops = {
3093 .init_state = ub960_init_state,
3094 };
3095
3096 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3097 .core = &ub960_subdev_core_ops,
3098 .pad = &ub960_pad_ops,
3099 };
3100
3101 static const struct media_entity_operations ub960_entity_ops = {
3102 .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3103 .link_validate = v4l2_subdev_link_validate,
3104 .has_pad_interdep = v4l2_subdev_has_pad_interdep,
3105 };
3106
3107 /* -----------------------------------------------------------------------------
3108 * Core
3109 */
3110
ub960_handle_events(int irq,void * arg)3111 static irqreturn_t ub960_handle_events(int irq, void *arg)
3112 {
3113 struct ub960_data *priv = arg;
3114 unsigned int i;
3115 u8 int_sts;
3116 u8 fwd_sts;
3117 int ret;
3118
3119 ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3120 if (ret || !int_sts)
3121 return IRQ_NONE;
3122
3123 dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3124
3125 ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3126 if (ret)
3127 return IRQ_NONE;
3128
3129 dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3130
3131 for (i = 0; i < priv->hw_data->num_txports; i++) {
3132 if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3133 ub960_csi_handle_events(priv, i);
3134 }
3135
3136 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3137 if (!priv->rxports[i])
3138 continue;
3139
3140 if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3141 ub960_rxport_handle_events(priv, i);
3142 }
3143
3144 return IRQ_HANDLED;
3145 }
3146
ub960_handler_work(struct work_struct * work)3147 static void ub960_handler_work(struct work_struct *work)
3148 {
3149 struct delayed_work *dwork = to_delayed_work(work);
3150 struct ub960_data *priv =
3151 container_of(dwork, struct ub960_data, poll_work);
3152
3153 ub960_handle_events(0, priv);
3154
3155 schedule_delayed_work(&priv->poll_work,
3156 msecs_to_jiffies(UB960_POLL_TIME_MS));
3157 }
3158
ub960_txport_free_ports(struct ub960_data * priv)3159 static void ub960_txport_free_ports(struct ub960_data *priv)
3160 {
3161 unsigned int nport;
3162
3163 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3164 struct ub960_txport *txport = priv->txports[nport];
3165
3166 if (!txport)
3167 continue;
3168
3169 kfree(txport);
3170 priv->txports[nport] = NULL;
3171 }
3172 }
3173
ub960_rxport_free_ports(struct ub960_data * priv)3174 static void ub960_rxport_free_ports(struct ub960_data *priv)
3175 {
3176 unsigned int nport;
3177
3178 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3179 struct ub960_rxport *rxport = priv->rxports[nport];
3180
3181 if (!rxport)
3182 continue;
3183
3184 fwnode_handle_put(rxport->source.ep_fwnode);
3185 fwnode_handle_put(rxport->ser.fwnode);
3186
3187 kfree(rxport);
3188 priv->rxports[nport] = NULL;
3189 }
3190 }
3191
3192 static int
ub960_parse_dt_rxport_link_properties(struct ub960_data * priv,struct fwnode_handle * link_fwnode,struct ub960_rxport * rxport)3193 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3194 struct fwnode_handle *link_fwnode,
3195 struct ub960_rxport *rxport)
3196 {
3197 struct device *dev = &priv->client->dev;
3198 unsigned int nport = rxport->nport;
3199 u32 rx_mode;
3200 u32 cdr_mode;
3201 s32 strobe_pos;
3202 u32 eq_level;
3203 u32 ser_i2c_alias;
3204 int ret;
3205
3206 cdr_mode = RXPORT_CDR_FPD3;
3207
3208 ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3209 if (ret < 0 && ret != -EINVAL) {
3210 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3211 "ti,cdr-mode", ret);
3212 return ret;
3213 }
3214
3215 if (cdr_mode > RXPORT_CDR_LAST) {
3216 dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3217 return -EINVAL;
3218 }
3219
3220 if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3221 dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3222 return -EINVAL;
3223 }
3224
3225 rxport->cdr_mode = cdr_mode;
3226
3227 ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3228 if (ret < 0) {
3229 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3230 "ti,rx-mode", ret);
3231 return ret;
3232 }
3233
3234 if (rx_mode > RXPORT_MODE_LAST) {
3235 dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3236 return -EINVAL;
3237 }
3238
3239 switch (rx_mode) {
3240 case RXPORT_MODE_RAW12_HF:
3241 case RXPORT_MODE_RAW12_LF:
3242 dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3243 rx_mode);
3244 return -EINVAL;
3245 default:
3246 break;
3247 }
3248
3249 rxport->rx_mode = rx_mode;
3250
3251 /* EQ & Strobe related */
3252
3253 /* Defaults */
3254 rxport->eq.manual_eq = false;
3255 rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3256 rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3257
3258 ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3259 &strobe_pos);
3260 if (ret) {
3261 if (ret != -EINVAL) {
3262 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3263 "ti,strobe-pos", ret);
3264 return ret;
3265 }
3266 } else {
3267 if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3268 strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3269 dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3270 nport, strobe_pos);
3271 return -EINVAL;
3272 }
3273
3274 /* NOTE: ignored unless global manual strobe pos is also set */
3275 rxport->eq.strobe_pos = strobe_pos;
3276 if (!priv->strobe.manual)
3277 dev_warn(dev,
3278 "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3279 nport);
3280 }
3281
3282 ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3283 if (ret) {
3284 if (ret != -EINVAL) {
3285 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3286 "ti,eq-level", ret);
3287 return ret;
3288 }
3289 } else {
3290 if (eq_level > UB960_MAX_EQ_LEVEL) {
3291 dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3292 nport, eq_level);
3293 return -EINVAL;
3294 }
3295
3296 rxport->eq.manual_eq = true;
3297 rxport->eq.manual.eq_level = eq_level;
3298 }
3299
3300 ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3301 &ser_i2c_alias);
3302 if (ret) {
3303 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3304 "i2c-alias", ret);
3305 return ret;
3306 }
3307 rxport->ser.alias = ser_i2c_alias;
3308
3309 rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3310 if (!rxport->ser.fwnode) {
3311 dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3312 return -EINVAL;
3313 }
3314
3315 return 0;
3316 }
3317
ub960_parse_dt_rxport_ep_properties(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,struct ub960_rxport * rxport)3318 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3319 struct fwnode_handle *ep_fwnode,
3320 struct ub960_rxport *rxport)
3321 {
3322 struct device *dev = &priv->client->dev;
3323 struct v4l2_fwnode_endpoint vep = {};
3324 unsigned int nport = rxport->nport;
3325 bool hsync_hi;
3326 bool vsync_hi;
3327 int ret;
3328
3329 rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3330 if (!rxport->source.ep_fwnode) {
3331 dev_err(dev, "rx%u: no remote endpoint\n", nport);
3332 return -ENODEV;
3333 }
3334
3335 /* We currently have properties only for RAW modes */
3336
3337 switch (rxport->rx_mode) {
3338 case RXPORT_MODE_RAW10:
3339 case RXPORT_MODE_RAW12_HF:
3340 case RXPORT_MODE_RAW12_LF:
3341 break;
3342 default:
3343 return 0;
3344 }
3345
3346 vep.bus_type = V4L2_MBUS_PARALLEL;
3347 ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3348 if (ret) {
3349 dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3350 goto err_put_source_ep_fwnode;
3351 }
3352
3353 hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3354 vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3355
3356 /* LineValid and FrameValid are inverse to the h/vsync active */
3357 rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3358 (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3359
3360 return 0;
3361
3362 err_put_source_ep_fwnode:
3363 fwnode_handle_put(rxport->source.ep_fwnode);
3364 return ret;
3365 }
3366
ub960_parse_dt_rxport(struct ub960_data * priv,unsigned int nport,struct fwnode_handle * link_fwnode,struct fwnode_handle * ep_fwnode)3367 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3368 struct fwnode_handle *link_fwnode,
3369 struct fwnode_handle *ep_fwnode)
3370 {
3371 static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3372 "vpoc0", "vpoc1", "vpoc2", "vpoc3"
3373 };
3374 struct device *dev = &priv->client->dev;
3375 struct ub960_rxport *rxport;
3376 int ret;
3377
3378 rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3379 if (!rxport)
3380 return -ENOMEM;
3381
3382 priv->rxports[nport] = rxport;
3383
3384 rxport->nport = nport;
3385 rxport->priv = priv;
3386
3387 ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3388 if (ret)
3389 goto err_free_rxport;
3390
3391 rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3392 if (IS_ERR(rxport->vpoc)) {
3393 ret = PTR_ERR(rxport->vpoc);
3394 if (ret == -ENODEV) {
3395 rxport->vpoc = NULL;
3396 } else {
3397 dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3398 nport, ret);
3399 goto err_put_remote_fwnode;
3400 }
3401 }
3402
3403 ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3404 if (ret)
3405 goto err_put_remote_fwnode;
3406
3407 return 0;
3408
3409 err_put_remote_fwnode:
3410 fwnode_handle_put(rxport->ser.fwnode);
3411 err_free_rxport:
3412 priv->rxports[nport] = NULL;
3413 kfree(rxport);
3414 return ret;
3415 }
3416
3417 static struct fwnode_handle *
ub960_fwnode_get_link_by_regs(struct fwnode_handle * links_fwnode,unsigned int nport)3418 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3419 unsigned int nport)
3420 {
3421 struct fwnode_handle *link_fwnode;
3422 int ret;
3423
3424 fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3425 u32 link_num;
3426
3427 if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3428 continue;
3429
3430 ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3431 if (ret) {
3432 fwnode_handle_put(link_fwnode);
3433 return NULL;
3434 }
3435
3436 if (nport == link_num)
3437 return link_fwnode;
3438 }
3439
3440 return NULL;
3441 }
3442
ub960_parse_dt_rxports(struct ub960_data * priv)3443 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3444 {
3445 struct device *dev = &priv->client->dev;
3446 struct fwnode_handle *links_fwnode;
3447 unsigned int nport;
3448 int ret;
3449
3450 links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3451 if (!links_fwnode) {
3452 dev_err(dev, "'links' node missing\n");
3453 return -ENODEV;
3454 }
3455
3456 /* Defaults, recommended by TI */
3457 priv->strobe.min = 2;
3458 priv->strobe.max = 3;
3459
3460 priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3461
3462 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3463 struct fwnode_handle *link_fwnode;
3464 struct fwnode_handle *ep_fwnode;
3465
3466 link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3467 if (!link_fwnode)
3468 continue;
3469
3470 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3471 nport, 0, 0);
3472 if (!ep_fwnode) {
3473 fwnode_handle_put(link_fwnode);
3474 continue;
3475 }
3476
3477 ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3478 ep_fwnode);
3479
3480 fwnode_handle_put(link_fwnode);
3481 fwnode_handle_put(ep_fwnode);
3482
3483 if (ret) {
3484 dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3485 goto err_put_links;
3486 }
3487 }
3488
3489 fwnode_handle_put(links_fwnode);
3490
3491 return 0;
3492
3493 err_put_links:
3494 fwnode_handle_put(links_fwnode);
3495
3496 return ret;
3497 }
3498
ub960_parse_dt_txports(struct ub960_data * priv)3499 static int ub960_parse_dt_txports(struct ub960_data *priv)
3500 {
3501 struct device *dev = &priv->client->dev;
3502 u32 nport;
3503 int ret;
3504
3505 for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3506 unsigned int port = nport + priv->hw_data->num_rxports;
3507 struct fwnode_handle *ep_fwnode;
3508
3509 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3510 port, 0, 0);
3511 if (!ep_fwnode)
3512 continue;
3513
3514 ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3515
3516 fwnode_handle_put(ep_fwnode);
3517
3518 if (ret)
3519 break;
3520 }
3521
3522 return 0;
3523 }
3524
ub960_parse_dt(struct ub960_data * priv)3525 static int ub960_parse_dt(struct ub960_data *priv)
3526 {
3527 int ret;
3528
3529 ret = ub960_parse_dt_rxports(priv);
3530 if (ret)
3531 return ret;
3532
3533 ret = ub960_parse_dt_txports(priv);
3534 if (ret)
3535 goto err_free_rxports;
3536
3537 return 0;
3538
3539 err_free_rxports:
3540 ub960_rxport_free_ports(priv);
3541
3542 return ret;
3543 }
3544
ub960_notify_bound(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3545 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3546 struct v4l2_subdev *subdev,
3547 struct v4l2_async_connection *asd)
3548 {
3549 struct ub960_data *priv = sd_to_ub960(notifier->sd);
3550 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3551 struct device *dev = &priv->client->dev;
3552 u8 nport = rxport->nport;
3553 unsigned int i;
3554 int ret;
3555
3556 ret = media_entity_get_fwnode_pad(&subdev->entity,
3557 rxport->source.ep_fwnode,
3558 MEDIA_PAD_FL_SOURCE);
3559 if (ret < 0) {
3560 dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3561 return ret;
3562 }
3563
3564 rxport->source.sd = subdev;
3565 rxport->source.pad = ret;
3566
3567 ret = media_create_pad_link(&rxport->source.sd->entity,
3568 rxport->source.pad, &priv->sd.entity, nport,
3569 MEDIA_LNK_FL_ENABLED |
3570 MEDIA_LNK_FL_IMMUTABLE);
3571 if (ret) {
3572 dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3573 rxport->source.sd->name, rxport->source.pad,
3574 priv->sd.name, nport);
3575 return ret;
3576 }
3577
3578 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3579 if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3580 dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3581 return 0;
3582 }
3583 }
3584
3585 return 0;
3586 }
3587
ub960_notify_unbind(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3588 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3589 struct v4l2_subdev *subdev,
3590 struct v4l2_async_connection *asd)
3591 {
3592 struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3593
3594 rxport->source.sd = NULL;
3595 }
3596
3597 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3598 .bound = ub960_notify_bound,
3599 .unbind = ub960_notify_unbind,
3600 };
3601
ub960_v4l2_notifier_register(struct ub960_data * priv)3602 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3603 {
3604 struct device *dev = &priv->client->dev;
3605 unsigned int i;
3606 int ret;
3607
3608 v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
3609
3610 for (i = 0; i < priv->hw_data->num_rxports; i++) {
3611 struct ub960_rxport *rxport = priv->rxports[i];
3612 struct ub960_asd *asd;
3613
3614 if (!rxport)
3615 continue;
3616
3617 asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3618 rxport->source.ep_fwnode,
3619 struct ub960_asd);
3620 if (IS_ERR(asd)) {
3621 dev_err(dev, "Failed to add subdev for source %u: %pe",
3622 i, asd);
3623 v4l2_async_nf_cleanup(&priv->notifier);
3624 return PTR_ERR(asd);
3625 }
3626
3627 asd->rxport = rxport;
3628 }
3629
3630 priv->notifier.ops = &ub960_notify_ops;
3631
3632 ret = v4l2_async_nf_register(&priv->notifier);
3633 if (ret) {
3634 dev_err(dev, "Failed to register subdev_notifier");
3635 v4l2_async_nf_cleanup(&priv->notifier);
3636 return ret;
3637 }
3638
3639 return 0;
3640 }
3641
ub960_v4l2_notifier_unregister(struct ub960_data * priv)3642 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3643 {
3644 v4l2_async_nf_unregister(&priv->notifier);
3645 v4l2_async_nf_cleanup(&priv->notifier);
3646 }
3647
ub960_create_subdev(struct ub960_data * priv)3648 static int ub960_create_subdev(struct ub960_data *priv)
3649 {
3650 struct device *dev = &priv->client->dev;
3651 unsigned int i;
3652 int ret;
3653
3654 v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3655 priv->sd.internal_ops = &ub960_internal_ops;
3656
3657 v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3658 priv->sd.ctrl_handler = &priv->ctrl_handler;
3659
3660 v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3661 ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3662 priv->tx_link_freq);
3663
3664 if (priv->ctrl_handler.error) {
3665 ret = priv->ctrl_handler.error;
3666 goto err_free_ctrl;
3667 }
3668
3669 priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3670 V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
3671 priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3672 priv->sd.entity.ops = &ub960_entity_ops;
3673
3674 for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3675 priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3676 MEDIA_PAD_FL_SINK :
3677 MEDIA_PAD_FL_SOURCE;
3678 }
3679
3680 ret = media_entity_pads_init(&priv->sd.entity,
3681 priv->hw_data->num_rxports +
3682 priv->hw_data->num_txports,
3683 priv->pads);
3684 if (ret)
3685 goto err_free_ctrl;
3686
3687 priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3688
3689 ret = v4l2_subdev_init_finalize(&priv->sd);
3690 if (ret)
3691 goto err_entity_cleanup;
3692
3693 ret = ub960_v4l2_notifier_register(priv);
3694 if (ret) {
3695 dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3696 goto err_subdev_cleanup;
3697 }
3698
3699 ret = v4l2_async_register_subdev(&priv->sd);
3700 if (ret) {
3701 dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3702 goto err_unreg_notif;
3703 }
3704
3705 return 0;
3706
3707 err_unreg_notif:
3708 ub960_v4l2_notifier_unregister(priv);
3709 err_subdev_cleanup:
3710 v4l2_subdev_cleanup(&priv->sd);
3711 err_entity_cleanup:
3712 media_entity_cleanup(&priv->sd.entity);
3713 err_free_ctrl:
3714 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3715
3716 return ret;
3717 }
3718
ub960_destroy_subdev(struct ub960_data * priv)3719 static void ub960_destroy_subdev(struct ub960_data *priv)
3720 {
3721 ub960_v4l2_notifier_unregister(priv);
3722 v4l2_async_unregister_subdev(&priv->sd);
3723
3724 v4l2_subdev_cleanup(&priv->sd);
3725
3726 media_entity_cleanup(&priv->sd.entity);
3727 v4l2_ctrl_handler_free(&priv->ctrl_handler);
3728 }
3729
3730 static const struct regmap_config ub960_regmap_config = {
3731 .name = "ds90ub960",
3732
3733 .reg_bits = 8,
3734 .val_bits = 8,
3735
3736 .max_register = 0xff,
3737
3738 /*
3739 * We do locking in the driver to cover the TX/RX port selection and the
3740 * indirect register access.
3741 */
3742 .disable_locking = true,
3743 };
3744
ub960_reset(struct ub960_data * priv,bool reset_regs)3745 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3746 {
3747 struct device *dev = &priv->client->dev;
3748 unsigned int v;
3749 int ret;
3750 u8 bit;
3751
3752 bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3753 UB960_SR_RESET_DIGITAL_RESET0;
3754
3755 ub960_write(priv, UB960_SR_RESET, bit);
3756
3757 mutex_lock(&priv->reg_lock);
3758
3759 ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3760 (v & bit) == 0, 2000, 100000);
3761
3762 mutex_unlock(&priv->reg_lock);
3763
3764 if (ret)
3765 dev_err(dev, "reset failed: %d\n", ret);
3766 }
3767
ub960_get_hw_resources(struct ub960_data * priv)3768 static int ub960_get_hw_resources(struct ub960_data *priv)
3769 {
3770 struct device *dev = &priv->client->dev;
3771
3772 priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3773 if (IS_ERR(priv->regmap))
3774 return PTR_ERR(priv->regmap);
3775
3776 priv->vddio = devm_regulator_get(dev, "vddio");
3777 if (IS_ERR(priv->vddio))
3778 return dev_err_probe(dev, PTR_ERR(priv->vddio),
3779 "cannot get VDDIO regulator\n");
3780
3781 /* get power-down pin from DT */
3782 priv->pd_gpio =
3783 devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3784 if (IS_ERR(priv->pd_gpio))
3785 return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3786 "Cannot get powerdown GPIO\n");
3787
3788 priv->refclk = devm_clk_get(dev, "refclk");
3789 if (IS_ERR(priv->refclk))
3790 return dev_err_probe(dev, PTR_ERR(priv->refclk),
3791 "Cannot get REFCLK\n");
3792
3793 return 0;
3794 }
3795
ub960_enable_core_hw(struct ub960_data * priv)3796 static int ub960_enable_core_hw(struct ub960_data *priv)
3797 {
3798 struct device *dev = &priv->client->dev;
3799 u8 rev_mask;
3800 int ret;
3801 u8 dev_sts;
3802 u8 refclk_freq;
3803
3804 ret = regulator_enable(priv->vddio);
3805 if (ret)
3806 return dev_err_probe(dev, ret,
3807 "failed to enable VDDIO regulator\n");
3808
3809 ret = clk_prepare_enable(priv->refclk);
3810 if (ret) {
3811 dev_err_probe(dev, ret, "Failed to enable refclk\n");
3812 goto err_disable_vddio;
3813 }
3814
3815 if (priv->pd_gpio) {
3816 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3817 /* wait min 2 ms for reset to complete */
3818 fsleep(2000);
3819 gpiod_set_value_cansleep(priv->pd_gpio, 0);
3820 /* wait min 2 ms for power up to finish */
3821 fsleep(2000);
3822 }
3823
3824 ub960_reset(priv, true);
3825
3826 /* Runtime check register accessibility */
3827 ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3828 if (ret) {
3829 dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3830 goto err_pd_gpio;
3831 }
3832
3833 dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3834 rev_mask);
3835
3836 ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3837 if (ret)
3838 goto err_pd_gpio;
3839
3840 ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3841 if (ret)
3842 goto err_pd_gpio;
3843
3844 dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3845 !!(dev_sts & BIT(4)), refclk_freq,
3846 clk_get_rate(priv->refclk) / 1000000);
3847
3848 /* Disable all RX ports by default */
3849 ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3850 if (ret)
3851 goto err_pd_gpio;
3852
3853 /* release GPIO lock */
3854 if (priv->hw_data->is_ub9702) {
3855 ret = ub960_update_bits(priv, UB960_SR_RESET,
3856 UB960_SR_RESET_GPIO_LOCK_RELEASE,
3857 UB960_SR_RESET_GPIO_LOCK_RELEASE);
3858 if (ret)
3859 goto err_pd_gpio;
3860 }
3861
3862 return 0;
3863
3864 err_pd_gpio:
3865 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3866 clk_disable_unprepare(priv->refclk);
3867 err_disable_vddio:
3868 regulator_disable(priv->vddio);
3869
3870 return ret;
3871 }
3872
ub960_disable_core_hw(struct ub960_data * priv)3873 static void ub960_disable_core_hw(struct ub960_data *priv)
3874 {
3875 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3876 clk_disable_unprepare(priv->refclk);
3877 regulator_disable(priv->vddio);
3878 }
3879
ub960_probe(struct i2c_client * client)3880 static int ub960_probe(struct i2c_client *client)
3881 {
3882 struct device *dev = &client->dev;
3883 struct ub960_data *priv;
3884 unsigned int port_lock_mask;
3885 unsigned int port_mask;
3886 unsigned int nport;
3887 int ret;
3888
3889 priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3890 if (!priv)
3891 return -ENOMEM;
3892
3893 priv->client = client;
3894
3895 priv->hw_data = device_get_match_data(dev);
3896
3897 mutex_init(&priv->reg_lock);
3898
3899 INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3900
3901 /*
3902 * Initialize these to invalid values so that the first reg writes will
3903 * configure the target.
3904 */
3905 priv->reg_current.indirect_target = 0xff;
3906 priv->reg_current.rxport = 0xff;
3907 priv->reg_current.txport = 0xff;
3908
3909 ret = ub960_get_hw_resources(priv);
3910 if (ret)
3911 goto err_mutex_destroy;
3912
3913 ret = ub960_enable_core_hw(priv);
3914 if (ret)
3915 goto err_mutex_destroy;
3916
3917 ret = ub960_parse_dt(priv);
3918 if (ret)
3919 goto err_disable_core_hw;
3920
3921 ret = ub960_init_tx_ports(priv);
3922 if (ret)
3923 goto err_free_ports;
3924
3925 ret = ub960_rxport_enable_vpocs(priv);
3926 if (ret)
3927 goto err_free_ports;
3928
3929 ret = ub960_init_rx_ports(priv);
3930 if (ret)
3931 goto err_disable_vpocs;
3932
3933 ub960_reset(priv, false);
3934
3935 port_mask = 0;
3936
3937 for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3938 struct ub960_rxport *rxport = priv->rxports[nport];
3939
3940 if (!rxport)
3941 continue;
3942
3943 port_mask |= BIT(nport);
3944 }
3945
3946 ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
3947 if (ret)
3948 goto err_disable_vpocs;
3949
3950 if (port_mask != port_lock_mask) {
3951 ret = -EIO;
3952 dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
3953 goto err_disable_vpocs;
3954 }
3955
3956 /*
3957 * Clear any errors caused by switching the RX port settings while
3958 * probing.
3959 */
3960 ub960_clear_rx_errors(priv);
3961
3962 ret = ub960_init_atr(priv);
3963 if (ret)
3964 goto err_disable_vpocs;
3965
3966 ret = ub960_rxport_add_serializers(priv);
3967 if (ret)
3968 goto err_uninit_atr;
3969
3970 ret = ub960_create_subdev(priv);
3971 if (ret)
3972 goto err_free_sers;
3973
3974 if (client->irq)
3975 dev_warn(dev, "irq support not implemented, using polling\n");
3976
3977 schedule_delayed_work(&priv->poll_work,
3978 msecs_to_jiffies(UB960_POLL_TIME_MS));
3979
3980 return 0;
3981
3982 err_free_sers:
3983 ub960_rxport_remove_serializers(priv);
3984 err_uninit_atr:
3985 ub960_uninit_atr(priv);
3986 err_disable_vpocs:
3987 ub960_rxport_disable_vpocs(priv);
3988 err_free_ports:
3989 ub960_rxport_free_ports(priv);
3990 ub960_txport_free_ports(priv);
3991 err_disable_core_hw:
3992 ub960_disable_core_hw(priv);
3993 err_mutex_destroy:
3994 mutex_destroy(&priv->reg_lock);
3995 return ret;
3996 }
3997
ub960_remove(struct i2c_client * client)3998 static void ub960_remove(struct i2c_client *client)
3999 {
4000 struct v4l2_subdev *sd = i2c_get_clientdata(client);
4001 struct ub960_data *priv = sd_to_ub960(sd);
4002
4003 cancel_delayed_work_sync(&priv->poll_work);
4004
4005 ub960_destroy_subdev(priv);
4006 ub960_rxport_remove_serializers(priv);
4007 ub960_uninit_atr(priv);
4008 ub960_rxport_disable_vpocs(priv);
4009 ub960_rxport_free_ports(priv);
4010 ub960_txport_free_ports(priv);
4011 ub960_disable_core_hw(priv);
4012 mutex_destroy(&priv->reg_lock);
4013 }
4014
4015 static const struct ub960_hw_data ds90ub960_hw = {
4016 .model = "ub960",
4017 .num_rxports = 4,
4018 .num_txports = 2,
4019 };
4020
4021 static const struct ub960_hw_data ds90ub9702_hw = {
4022 .model = "ub9702",
4023 .num_rxports = 4,
4024 .num_txports = 2,
4025 .is_ub9702 = true,
4026 .is_fpdlink4 = true,
4027 };
4028
4029 static const struct i2c_device_id ub960_id[] = {
4030 { "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4031 { "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4032 {}
4033 };
4034 MODULE_DEVICE_TABLE(i2c, ub960_id);
4035
4036 static const struct of_device_id ub960_dt_ids[] = {
4037 { .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4038 { .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4039 {}
4040 };
4041 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4042
4043 static struct i2c_driver ds90ub960_driver = {
4044 .probe = ub960_probe,
4045 .remove = ub960_remove,
4046 .id_table = ub960_id,
4047 .driver = {
4048 .name = "ds90ub960",
4049 .of_match_table = ub960_dt_ids,
4050 },
4051 };
4052 module_i2c_driver(ds90ub960_driver);
4053
4054 MODULE_LICENSE("GPL");
4055 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4056 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4057 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4058 MODULE_IMPORT_NS(I2C_ATR);
4059