xref: /linux/drivers/media/i2c/ds90ub960.c (revision cdd30ebb1b9f36159d66f088b61aee264e649d7a)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
4  *
5  * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6  * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
7  */
8 
9 /*
10  * (Possible) TODOs:
11  *
12  * - PM for serializer and remote peripherals. We need to manage:
13  *   - VPOC
14  *     - Power domain? Regulator? Somehow any remote device should be able to
15  *       cause the VPOC to be turned on.
16  *   - Link between the deserializer and the serializer
17  *     - Related to VPOC management. We probably always want to turn on the VPOC
18  *       and then enable the link.
19  *   - Serializer's services: i2c, gpios, power
20  *     - The serializer needs to resume before the remote peripherals can
21  *       e.g. use the i2c.
22  *     - How to handle gpios? Reserving a gpio essentially keeps the provider
23  *       (serializer) always powered on.
24  * - Do we need a new bus for the FPD-Link? At the moment the serializers
25  *   are children of the same i2c-adapter where the deserializer resides.
26  * - i2c-atr could be made embeddable instead of allocatable.
27  */
28 
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/workqueue.h>
47 
48 #include <media/i2c/ds90ub9xx.h>
49 #include <media/mipi-csi2.h>
50 #include <media/v4l2-ctrls.h>
51 #include <media/v4l2-fwnode.h>
52 #include <media/v4l2-subdev.h>
53 
54 #define MHZ(v) ((u32)((v) * 1000000U))
55 
56 #define UB960_POLL_TIME_MS	500
57 
58 #define UB960_MAX_RX_NPORTS	4
59 #define UB960_MAX_TX_NPORTS	2
60 #define UB960_MAX_NPORTS	(UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
61 
62 #define UB960_MAX_PORT_ALIASES	8
63 
64 #define UB960_NUM_BC_GPIOS		4
65 
66 /*
67  * Register map
68  *
69  * 0x00-0x32   Shared (UB960_SR)
70  * 0x33-0x3a   CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
71  * 0x4c        Shared (UB960_SR)
72  * 0x4d-0x7f   FPD-Link RX, per-port paged (UB960_RR)
73  * 0xb0-0xbf   Shared (UB960_SR)
74  * 0xd0-0xdf   FPD-Link RX, per-port paged (UB960_RR)
75  * 0xf0-0xf5   Shared (UB960_SR)
76  * 0xf8-0xfb   Shared (UB960_SR)
77  * All others  Reserved
78  *
79  * Register prefixes:
80  * UB960_SR_* = Shared register
81  * UB960_RR_* = FPD-Link RX, per-port paged register
82  * UB960_TR_* = CSI-2 TX, per-port paged register
83  * UB960_XR_* = Reserved register
84  * UB960_IR_* = Indirect register
85  */
86 
87 #define UB960_SR_I2C_DEV_ID			0x00
88 #define UB960_SR_RESET				0x01
89 #define UB960_SR_RESET_DIGITAL_RESET1		BIT(1)
90 #define UB960_SR_RESET_DIGITAL_RESET0		BIT(0)
91 #define UB960_SR_RESET_GPIO_LOCK_RELEASE	BIT(5)
92 
93 #define UB960_SR_GEN_CONFIG			0x02
94 #define UB960_SR_REV_MASK			0x03
95 #define UB960_SR_DEVICE_STS			0x04
96 #define UB960_SR_PAR_ERR_THOLD_HI		0x05
97 #define UB960_SR_PAR_ERR_THOLD_LO		0x06
98 #define UB960_SR_BCC_WDOG_CTL			0x07
99 #define UB960_SR_I2C_CTL1			0x08
100 #define UB960_SR_I2C_CTL2			0x09
101 #define UB960_SR_SCL_HIGH_TIME			0x0a
102 #define UB960_SR_SCL_LOW_TIME			0x0b
103 #define UB960_SR_RX_PORT_CTL			0x0c
104 #define UB960_SR_IO_CTL				0x0d
105 #define UB960_SR_GPIO_PIN_STS			0x0e
106 #define UB960_SR_GPIO_INPUT_CTL			0x0f
107 #define UB960_SR_GPIO_PIN_CTL(n)		(0x10 + (n)) /* n < UB960_NUM_GPIOS */
108 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL		5
109 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT	2
110 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN		BIT(0)
111 
112 #define UB960_SR_FS_CTL				0x18
113 #define UB960_SR_FS_HIGH_TIME_1			0x19
114 #define UB960_SR_FS_HIGH_TIME_0			0x1a
115 #define UB960_SR_FS_LOW_TIME_1			0x1b
116 #define UB960_SR_FS_LOW_TIME_0			0x1c
117 #define UB960_SR_MAX_FRM_HI			0x1d
118 #define UB960_SR_MAX_FRM_LO			0x1e
119 #define UB960_SR_CSI_PLL_CTL			0x1f
120 
121 #define UB960_SR_FWD_CTL1			0x20
122 #define UB960_SR_FWD_CTL1_PORT_DIS(n)		BIT((n) + 4)
123 
124 #define UB960_SR_FWD_CTL2			0x21
125 #define UB960_SR_FWD_STS			0x22
126 
127 #define UB960_SR_INTERRUPT_CTL			0x23
128 #define UB960_SR_INTERRUPT_CTL_INT_EN		BIT(7)
129 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0	BIT(4)
130 #define UB960_SR_INTERRUPT_CTL_IE_RX(n)		BIT((n)) /* rxport[n] IRQ */
131 
132 #define UB960_SR_INTERRUPT_STS			0x24
133 #define UB960_SR_INTERRUPT_STS_INT		BIT(7)
134 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n)	BIT(4 + (n)) /* txport[n] IRQ */
135 #define UB960_SR_INTERRUPT_STS_IS_RX(n)		BIT((n)) /* rxport[n] IRQ */
136 
137 #define UB960_SR_TS_CONFIG			0x25
138 #define UB960_SR_TS_CONTROL			0x26
139 #define UB960_SR_TS_LINE_HI			0x27
140 #define UB960_SR_TS_LINE_LO			0x28
141 #define UB960_SR_TS_STATUS			0x29
142 #define UB960_SR_TIMESTAMP_P0_HI		0x2a
143 #define UB960_SR_TIMESTAMP_P0_LO		0x2b
144 #define UB960_SR_TIMESTAMP_P1_HI		0x2c
145 #define UB960_SR_TIMESTAMP_P1_LO		0x2d
146 
147 #define UB960_SR_CSI_PORT_SEL			0x32
148 
149 #define UB960_TR_CSI_CTL			0x33
150 #define UB960_TR_CSI_CTL_CSI_CAL_EN		BIT(6)
151 #define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK	BIT(1)
152 #define UB960_TR_CSI_CTL_CSI_ENABLE		BIT(0)
153 
154 #define UB960_TR_CSI_CTL2			0x34
155 #define UB960_TR_CSI_STS			0x35
156 #define UB960_TR_CSI_TX_ICR			0x36
157 
158 #define UB960_TR_CSI_TX_ISR			0x37
159 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR	BIT(3)
160 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR	BIT(1)
161 
162 #define UB960_TR_CSI_TEST_CTL			0x38
163 #define UB960_TR_CSI_TEST_PATT_HI		0x39
164 #define UB960_TR_CSI_TEST_PATT_LO		0x3a
165 
166 #define UB960_XR_SFILTER_CFG			0x41
167 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT	4
168 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT	0
169 
170 #define UB960_XR_AEQ_CTL1			0x42
171 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK	BIT(6)
172 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING	BIT(5)
173 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY	BIT(4)
174 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK        \
175 	(UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK |  \
176 	 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
177 	 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
178 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN	BIT(0)
179 
180 #define UB960_XR_AEQ_ERR_THOLD			0x43
181 
182 #define UB960_RR_BCC_ERR_CTL			0x46
183 #define UB960_RR_BCC_STATUS			0x47
184 #define UB960_RR_BCC_STATUS_SEQ_ERROR		BIT(5)
185 #define UB960_RR_BCC_STATUS_MASTER_ERR		BIT(4)
186 #define UB960_RR_BCC_STATUS_MASTER_TO		BIT(3)
187 #define UB960_RR_BCC_STATUS_SLAVE_ERR		BIT(2)
188 #define UB960_RR_BCC_STATUS_SLAVE_TO		BIT(1)
189 #define UB960_RR_BCC_STATUS_RESP_ERR		BIT(0)
190 #define UB960_RR_BCC_STATUS_ERROR_MASK                                    \
191 	(UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
192 	 UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR |  \
193 	 UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
194 
195 #define UB960_RR_FPD3_CAP			0x4a
196 #define UB960_RR_RAW_EMBED_DTYPE		0x4b
197 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT	6
198 
199 #define UB960_SR_FPD3_PORT_SEL			0x4c
200 
201 #define UB960_RR_RX_PORT_STS1			0x4d
202 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR	BIT(5)
203 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG	BIT(4)
204 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR	BIT(3)
205 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR	BIT(2)
206 #define UB960_RR_RX_PORT_STS1_PORT_PASS		BIT(1)
207 #define UB960_RR_RX_PORT_STS1_LOCK_STS		BIT(0)
208 #define UB960_RR_RX_PORT_STS1_ERROR_MASK       \
209 	(UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
210 	 UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
211 	 UB960_RR_RX_PORT_STS1_PARITY_ERROR)
212 
213 #define UB960_RR_RX_PORT_STS2			0x4e
214 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE	BIT(7)
215 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG	BIT(6)
216 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR	BIT(5)
217 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR	BIT(4)
218 #define UB960_RR_RX_PORT_STS2_CSI_ERROR		BIT(3)
219 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE	BIT(2)
220 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT	BIT(1)
221 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG	BIT(0)
222 #define UB960_RR_RX_PORT_STS2_ERROR_MASK       \
223 	UB960_RR_RX_PORT_STS2_BUFFER_ERROR
224 
225 #define UB960_RR_RX_FREQ_HIGH			0x4f
226 #define UB960_RR_RX_FREQ_LOW			0x50
227 #define UB960_RR_SENSOR_STS_0			0x51
228 #define UB960_RR_SENSOR_STS_1			0x52
229 #define UB960_RR_SENSOR_STS_2			0x53
230 #define UB960_RR_SENSOR_STS_3			0x54
231 #define UB960_RR_RX_PAR_ERR_HI			0x55
232 #define UB960_RR_RX_PAR_ERR_LO			0x56
233 #define UB960_RR_BIST_ERR_COUNT			0x57
234 
235 #define UB960_RR_BCC_CONFIG			0x58
236 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH	BIT(6)
237 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK	GENMASK(2, 0)
238 
239 #define UB960_RR_DATAPATH_CTL1			0x59
240 #define UB960_RR_DATAPATH_CTL2			0x5a
241 #define UB960_RR_SER_ID				0x5b
242 #define UB960_RR_SER_ALIAS_ID			0x5c
243 
244 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
245 #define UB960_RR_SLAVE_ID(n)			(0x5d + (n))
246 #define UB960_RR_SLAVE_ALIAS(n)			(0x65 + (n))
247 
248 #define UB960_RR_PORT_CONFIG			0x6d
249 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK	GENMASK(1, 0)
250 
251 #define UB960_RR_BC_GPIO_CTL(n)			(0x6e + (n)) /* n < 2 */
252 #define UB960_RR_RAW10_ID			0x70
253 #define UB960_RR_RAW10_ID_VC_SHIFT		6
254 #define UB960_RR_RAW10_ID_DT_SHIFT		0
255 
256 #define UB960_RR_RAW12_ID			0x71
257 #define UB960_RR_CSI_VC_MAP			0x72
258 #define UB960_RR_CSI_VC_MAP_SHIFT(x)		((x) * 2)
259 
260 #define UB960_RR_LINE_COUNT_HI			0x73
261 #define UB960_RR_LINE_COUNT_LO			0x74
262 #define UB960_RR_LINE_LEN_1			0x75
263 #define UB960_RR_LINE_LEN_0			0x76
264 #define UB960_RR_FREQ_DET_CTL			0x77
265 #define UB960_RR_MAILBOX_1			0x78
266 #define UB960_RR_MAILBOX_2			0x79
267 
268 #define UB960_RR_CSI_RX_STS			0x7a
269 #define UB960_RR_CSI_RX_STS_LENGTH_ERR		BIT(3)
270 #define UB960_RR_CSI_RX_STS_CKSUM_ERR		BIT(2)
271 #define UB960_RR_CSI_RX_STS_ECC2_ERR		BIT(1)
272 #define UB960_RR_CSI_RX_STS_ECC1_ERR		BIT(0)
273 #define UB960_RR_CSI_RX_STS_ERROR_MASK                                    \
274 	(UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
275 	 UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
276 
277 #define UB960_RR_CSI_ERR_COUNTER		0x7b
278 #define UB960_RR_PORT_CONFIG2			0x7c
279 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
280 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
281 
282 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW	BIT(1)
283 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW	BIT(0)
284 
285 #define UB960_RR_PORT_PASS_CTL			0x7d
286 #define UB960_RR_SEN_INT_RISE_CTL		0x7e
287 #define UB960_RR_SEN_INT_FALL_CTL		0x7f
288 
289 #define UB960_SR_CSI_FRAME_COUNT_HI(n)		(0x90 + 8 * (n))
290 #define UB960_SR_CSI_FRAME_COUNT_LO(n)		(0x91 + 8 * (n))
291 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n)	(0x92 + 8 * (n))
292 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n)	(0x93 + 8 * (n))
293 #define UB960_SR_CSI_LINE_COUNT_HI(n)		(0x94 + 8 * (n))
294 #define UB960_SR_CSI_LINE_COUNT_LO(n)		(0x95 + 8 * (n))
295 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n)	(0x96 + 8 * (n))
296 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n)	(0x97 + 8 * (n))
297 
298 #define UB960_XR_REFCLK_FREQ			0xa5	/* UB960 */
299 
300 #define UB960_RR_VC_ID_MAP(x)			(0xa0 + (x)) /* UB9702 */
301 
302 #define UB960_SR_IND_ACC_CTL			0xb0
303 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC	BIT(1)
304 
305 #define UB960_SR_IND_ACC_ADDR			0xb1
306 #define UB960_SR_IND_ACC_DATA			0xb2
307 #define UB960_SR_BIST_CONTROL			0xb3
308 #define UB960_SR_MODE_IDX_STS			0xb8
309 #define UB960_SR_LINK_ERROR_COUNT		0xb9
310 #define UB960_SR_FPD3_ENC_CTL			0xba
311 #define UB960_SR_FV_MIN_TIME			0xbc
312 #define UB960_SR_GPIO_PD_CTL			0xbe
313 
314 #define UB960_SR_FPD_RATE_CFG			0xc2	/* UB9702 */
315 #define UB960_SR_CSI_PLL_DIV			0xc9	/* UB9702 */
316 
317 #define UB960_RR_PORT_DEBUG			0xd0
318 #define UB960_RR_AEQ_CTL2			0xd2
319 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR		BIT(2)
320 
321 #define UB960_RR_AEQ_STATUS			0xd3
322 #define UB960_RR_AEQ_STATUS_STATUS_2		GENMASK(5, 3)
323 #define UB960_RR_AEQ_STATUS_STATUS_1		GENMASK(2, 0)
324 
325 #define UB960_RR_AEQ_BYPASS			0xd4
326 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT	5
327 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK	GENMASK(7, 5)
328 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT	1
329 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK	GENMASK(3, 1)
330 #define UB960_RR_AEQ_BYPASS_ENABLE			BIT(0)
331 
332 #define UB960_RR_AEQ_MIN_MAX			0xd5
333 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT	4
334 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT	0
335 
336 #define UB960_RR_SFILTER_STS_0			0xd6
337 #define UB960_RR_SFILTER_STS_1			0xd7
338 #define UB960_RR_PORT_ICR_HI			0xd8
339 #define UB960_RR_PORT_ICR_LO			0xd9
340 #define UB960_RR_PORT_ISR_HI			0xda
341 #define UB960_RR_PORT_ISR_LO			0xdb
342 #define UB960_RR_FC_GPIO_STS			0xdc
343 #define UB960_RR_FC_GPIO_ICR			0xdd
344 #define UB960_RR_SEN_INT_RISE_STS		0xde
345 #define UB960_RR_SEN_INT_FALL_STS		0xdf
346 
347 #define UB960_RR_CHANNEL_MODE			0xe4	/* UB9702 */
348 
349 #define UB960_SR_FPD3_RX_ID(n)			(0xf0 + (n))
350 #define UB960_SR_FPD3_RX_ID_LEN			6
351 
352 #define UB960_SR_I2C_RX_ID(n)			(0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
353 
354 /* Indirect register blocks */
355 #define UB960_IND_TARGET_PAT_GEN		0x00
356 #define UB960_IND_TARGET_RX_ANA(n)		(0x01 + (n))
357 #define UB960_IND_TARGET_CSI_CSIPLL_REG_1	0x92	/* UB9702 */
358 #define UB960_IND_TARGET_CSI_ANA		0x07
359 
360 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
361 
362 #define UB960_IR_PGEN_CTL			0x01
363 #define UB960_IR_PGEN_CTL_PGEN_ENABLE		BIT(0)
364 
365 #define UB960_IR_PGEN_CFG			0x02
366 #define UB960_IR_PGEN_CSI_DI			0x03
367 #define UB960_IR_PGEN_LINE_SIZE1		0x04
368 #define UB960_IR_PGEN_LINE_SIZE0		0x05
369 #define UB960_IR_PGEN_BAR_SIZE1			0x06
370 #define UB960_IR_PGEN_BAR_SIZE0			0x07
371 #define UB960_IR_PGEN_ACT_LPF1			0x08
372 #define UB960_IR_PGEN_ACT_LPF0			0x09
373 #define UB960_IR_PGEN_TOT_LPF1			0x0a
374 #define UB960_IR_PGEN_TOT_LPF0			0x0b
375 #define UB960_IR_PGEN_LINE_PD1			0x0c
376 #define UB960_IR_PGEN_LINE_PD0			0x0d
377 #define UB960_IR_PGEN_VBP			0x0e
378 #define UB960_IR_PGEN_VFP			0x0f
379 #define UB960_IR_PGEN_COLOR(n)			(0x10 + (n)) /* n < 15 */
380 
381 #define UB960_IR_RX_ANA_STROBE_SET_CLK		0x08
382 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY	BIT(3)
383 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK	GENMASK(2, 0)
384 
385 #define UB960_IR_RX_ANA_STROBE_SET_DATA		0x09
386 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY	BIT(3)
387 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK	GENMASK(2, 0)
388 
389 /* EQ related */
390 
391 #define UB960_MIN_AEQ_STROBE_POS -7
392 #define UB960_MAX_AEQ_STROBE_POS  7
393 
394 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
395 
396 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
397 #define UB960_MAX_MANUAL_STROBE_POS  (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
398 #define UB960_NUM_MANUAL_STROBE_POS  (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
399 
400 #define UB960_MIN_EQ_LEVEL  0
401 #define UB960_MAX_EQ_LEVEL  14
402 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
403 
404 struct ub960_hw_data {
405 	const char *model;
406 	u8 num_rxports;
407 	u8 num_txports;
408 	bool is_ub9702;
409 	bool is_fpdlink4;
410 };
411 
412 enum ub960_rxport_mode {
413 	RXPORT_MODE_RAW10 = 0,
414 	RXPORT_MODE_RAW12_HF = 1,
415 	RXPORT_MODE_RAW12_LF = 2,
416 	RXPORT_MODE_CSI2_SYNC = 3,
417 	RXPORT_MODE_CSI2_NONSYNC = 4,
418 	RXPORT_MODE_LAST = RXPORT_MODE_CSI2_NONSYNC,
419 };
420 
421 enum ub960_rxport_cdr {
422 	RXPORT_CDR_FPD3 = 0,
423 	RXPORT_CDR_FPD4 = 1,
424 	RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
425 };
426 
427 struct ub960_rxport {
428 	struct ub960_data      *priv;
429 	u8                      nport;	/* RX port number, and index in priv->rxport[] */
430 
431 	struct {
432 		struct v4l2_subdev *sd;
433 		u16 pad;
434 		struct fwnode_handle *ep_fwnode;
435 	} source;
436 
437 	/* Serializer */
438 	struct {
439 		struct fwnode_handle *fwnode;
440 		struct i2c_client *client;
441 		unsigned short alias; /* I2C alias (lower 7 bits) */
442 		struct ds90ub9xx_platform_data pdata;
443 	} ser;
444 
445 	enum ub960_rxport_mode  rx_mode;
446 	enum ub960_rxport_cdr	cdr_mode;
447 
448 	u8			lv_fv_pol;	/* LV and FV polarities */
449 
450 	struct regulator	*vpoc;
451 
452 	/* EQ settings */
453 	struct {
454 		bool manual_eq;
455 
456 		s8 strobe_pos;
457 
458 		union {
459 			struct {
460 				u8 eq_level_min;
461 				u8 eq_level_max;
462 			} aeq;
463 
464 			struct {
465 				u8 eq_level;
466 			} manual;
467 		};
468 	} eq;
469 
470 	const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
471 };
472 
473 struct ub960_asd {
474 	struct v4l2_async_connection base;
475 	struct ub960_rxport *rxport;
476 };
477 
to_ub960_asd(struct v4l2_async_connection * asd)478 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
479 {
480 	return container_of(asd, struct ub960_asd, base);
481 }
482 
483 struct ub960_txport {
484 	struct ub960_data      *priv;
485 	u8                      nport;	/* TX port number, and index in priv->txport[] */
486 
487 	u32 num_data_lanes;
488 	bool non_continous_clk;
489 };
490 
491 struct ub960_data {
492 	const struct ub960_hw_data	*hw_data;
493 	struct i2c_client	*client; /* for shared local registers */
494 	struct regmap		*regmap;
495 
496 	/* lock for register access */
497 	struct mutex		reg_lock;
498 
499 	struct clk		*refclk;
500 
501 	struct regulator	*vddio;
502 
503 	struct gpio_desc	*pd_gpio;
504 	struct delayed_work	poll_work;
505 	struct ub960_rxport	*rxports[UB960_MAX_RX_NPORTS];
506 	struct ub960_txport	*txports[UB960_MAX_TX_NPORTS];
507 
508 	struct v4l2_subdev	sd;
509 	struct media_pad	pads[UB960_MAX_NPORTS];
510 
511 	struct v4l2_ctrl_handler   ctrl_handler;
512 	struct v4l2_async_notifier notifier;
513 
514 	u32 tx_data_rate;		/* Nominal data rate (Gb/s) */
515 	s64 tx_link_freq[1];
516 
517 	struct i2c_atr *atr;
518 
519 	struct {
520 		u8 rxport;
521 		u8 txport;
522 		u8 indirect_target;
523 	} reg_current;
524 
525 	bool streaming;
526 
527 	u8 stored_fwd_ctl;
528 
529 	u64 stream_enable_mask[UB960_MAX_NPORTS];
530 
531 	/* These are common to all ports */
532 	struct {
533 		bool manual;
534 
535 		s8 min;
536 		s8 max;
537 	} strobe;
538 };
539 
sd_to_ub960(struct v4l2_subdev * sd)540 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
541 {
542 	return container_of(sd, struct ub960_data, sd);
543 }
544 
ub960_pad_is_sink(struct ub960_data * priv,u32 pad)545 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
546 {
547 	return pad < priv->hw_data->num_rxports;
548 }
549 
ub960_pad_is_source(struct ub960_data * priv,u32 pad)550 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
551 {
552 	return pad >= priv->hw_data->num_rxports;
553 }
554 
ub960_pad_to_port(struct ub960_data * priv,u32 pad)555 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
556 {
557 	if (ub960_pad_is_sink(priv, pad))
558 		return pad;
559 	else
560 		return pad - priv->hw_data->num_rxports;
561 }
562 
563 struct ub960_format_info {
564 	u32 code;
565 	u32 bpp;
566 	u8 datatype;
567 	bool meta;
568 };
569 
570 static const struct ub960_format_info ub960_formats[] = {
571 	{ .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
572 	{ .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
573 	{ .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
574 	{ .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
575 
576 	{ .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
577 	{ .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
578 	{ .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
579 	{ .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
580 };
581 
ub960_find_format(u32 code)582 static const struct ub960_format_info *ub960_find_format(u32 code)
583 {
584 	unsigned int i;
585 
586 	for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
587 		if (ub960_formats[i].code == code)
588 			return &ub960_formats[i];
589 	}
590 
591 	return NULL;
592 }
593 
594 /* -----------------------------------------------------------------------------
595  * Basic device access
596  */
597 
ub960_read(struct ub960_data * priv,u8 reg,u8 * val)598 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
599 {
600 	struct device *dev = &priv->client->dev;
601 	unsigned int v;
602 	int ret;
603 
604 	mutex_lock(&priv->reg_lock);
605 
606 	ret = regmap_read(priv->regmap, reg, &v);
607 	if (ret) {
608 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
609 			__func__, reg, ret);
610 		goto out_unlock;
611 	}
612 
613 	*val = v;
614 
615 out_unlock:
616 	mutex_unlock(&priv->reg_lock);
617 
618 	return ret;
619 }
620 
ub960_write(struct ub960_data * priv,u8 reg,u8 val)621 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
622 {
623 	struct device *dev = &priv->client->dev;
624 	int ret;
625 
626 	mutex_lock(&priv->reg_lock);
627 
628 	ret = regmap_write(priv->regmap, reg, val);
629 	if (ret)
630 		dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
631 			__func__, reg, ret);
632 
633 	mutex_unlock(&priv->reg_lock);
634 
635 	return ret;
636 }
637 
ub960_update_bits(struct ub960_data * priv,u8 reg,u8 mask,u8 val)638 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
639 {
640 	struct device *dev = &priv->client->dev;
641 	int ret;
642 
643 	mutex_lock(&priv->reg_lock);
644 
645 	ret = regmap_update_bits(priv->regmap, reg, mask, val);
646 	if (ret)
647 		dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
648 			__func__, reg, ret);
649 
650 	mutex_unlock(&priv->reg_lock);
651 
652 	return ret;
653 }
654 
ub960_read16(struct ub960_data * priv,u8 reg,u16 * val)655 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
656 {
657 	struct device *dev = &priv->client->dev;
658 	__be16 __v;
659 	int ret;
660 
661 	mutex_lock(&priv->reg_lock);
662 
663 	ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
664 	if (ret) {
665 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
666 			__func__, reg, ret);
667 		goto out_unlock;
668 	}
669 
670 	*val = be16_to_cpu(__v);
671 
672 out_unlock:
673 	mutex_unlock(&priv->reg_lock);
674 
675 	return ret;
676 }
677 
ub960_rxport_select(struct ub960_data * priv,u8 nport)678 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
679 {
680 	struct device *dev = &priv->client->dev;
681 	int ret;
682 
683 	lockdep_assert_held(&priv->reg_lock);
684 
685 	if (priv->reg_current.rxport == nport)
686 		return 0;
687 
688 	ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
689 			   (nport << 4) | BIT(nport));
690 	if (ret) {
691 		dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
692 			nport, ret);
693 		return ret;
694 	}
695 
696 	priv->reg_current.rxport = nport;
697 
698 	return 0;
699 }
700 
ub960_rxport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)701 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
702 {
703 	struct device *dev = &priv->client->dev;
704 	unsigned int v;
705 	int ret;
706 
707 	mutex_lock(&priv->reg_lock);
708 
709 	ret = ub960_rxport_select(priv, nport);
710 	if (ret)
711 		goto out_unlock;
712 
713 	ret = regmap_read(priv->regmap, reg, &v);
714 	if (ret) {
715 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
716 			__func__, reg, ret);
717 		goto out_unlock;
718 	}
719 
720 	*val = v;
721 
722 out_unlock:
723 	mutex_unlock(&priv->reg_lock);
724 
725 	return ret;
726 }
727 
ub960_rxport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)728 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
729 {
730 	struct device *dev = &priv->client->dev;
731 	int ret;
732 
733 	mutex_lock(&priv->reg_lock);
734 
735 	ret = ub960_rxport_select(priv, nport);
736 	if (ret)
737 		goto out_unlock;
738 
739 	ret = regmap_write(priv->regmap, reg, val);
740 	if (ret)
741 		dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
742 			__func__, reg, ret);
743 
744 out_unlock:
745 	mutex_unlock(&priv->reg_lock);
746 
747 	return ret;
748 }
749 
ub960_rxport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)750 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
751 				    u8 mask, u8 val)
752 {
753 	struct device *dev = &priv->client->dev;
754 	int ret;
755 
756 	mutex_lock(&priv->reg_lock);
757 
758 	ret = ub960_rxport_select(priv, nport);
759 	if (ret)
760 		goto out_unlock;
761 
762 	ret = regmap_update_bits(priv->regmap, reg, mask, val);
763 	if (ret)
764 		dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
765 			__func__, reg, ret);
766 
767 out_unlock:
768 	mutex_unlock(&priv->reg_lock);
769 
770 	return ret;
771 }
772 
ub960_rxport_read16(struct ub960_data * priv,u8 nport,u8 reg,u16 * val)773 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
774 			       u16 *val)
775 {
776 	struct device *dev = &priv->client->dev;
777 	__be16 __v;
778 	int ret;
779 
780 	mutex_lock(&priv->reg_lock);
781 
782 	ret = ub960_rxport_select(priv, nport);
783 	if (ret)
784 		goto out_unlock;
785 
786 	ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
787 	if (ret) {
788 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
789 			__func__, reg, ret);
790 		goto out_unlock;
791 	}
792 
793 	*val = be16_to_cpu(__v);
794 
795 out_unlock:
796 	mutex_unlock(&priv->reg_lock);
797 
798 	return ret;
799 }
800 
ub960_txport_select(struct ub960_data * priv,u8 nport)801 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
802 {
803 	struct device *dev = &priv->client->dev;
804 	int ret;
805 
806 	lockdep_assert_held(&priv->reg_lock);
807 
808 	if (priv->reg_current.txport == nport)
809 		return 0;
810 
811 	ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
812 			   (nport << 4) | BIT(nport));
813 	if (ret) {
814 		dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
815 			nport, ret);
816 		return ret;
817 	}
818 
819 	priv->reg_current.txport = nport;
820 
821 	return 0;
822 }
823 
ub960_txport_read(struct ub960_data * priv,u8 nport,u8 reg,u8 * val)824 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
825 {
826 	struct device *dev = &priv->client->dev;
827 	unsigned int v;
828 	int ret;
829 
830 	mutex_lock(&priv->reg_lock);
831 
832 	ret = ub960_txport_select(priv, nport);
833 	if (ret)
834 		goto out_unlock;
835 
836 	ret = regmap_read(priv->regmap, reg, &v);
837 	if (ret) {
838 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
839 			__func__, reg, ret);
840 		goto out_unlock;
841 	}
842 
843 	*val = v;
844 
845 out_unlock:
846 	mutex_unlock(&priv->reg_lock);
847 
848 	return ret;
849 }
850 
ub960_txport_write(struct ub960_data * priv,u8 nport,u8 reg,u8 val)851 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
852 {
853 	struct device *dev = &priv->client->dev;
854 	int ret;
855 
856 	mutex_lock(&priv->reg_lock);
857 
858 	ret = ub960_txport_select(priv, nport);
859 	if (ret)
860 		goto out_unlock;
861 
862 	ret = regmap_write(priv->regmap, reg, val);
863 	if (ret)
864 		dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
865 			__func__, reg, ret);
866 
867 out_unlock:
868 	mutex_unlock(&priv->reg_lock);
869 
870 	return ret;
871 }
872 
ub960_txport_update_bits(struct ub960_data * priv,u8 nport,u8 reg,u8 mask,u8 val)873 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
874 				    u8 mask, u8 val)
875 {
876 	struct device *dev = &priv->client->dev;
877 	int ret;
878 
879 	mutex_lock(&priv->reg_lock);
880 
881 	ret = ub960_txport_select(priv, nport);
882 	if (ret)
883 		goto out_unlock;
884 
885 	ret = regmap_update_bits(priv->regmap, reg, mask, val);
886 	if (ret)
887 		dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
888 			__func__, reg, ret);
889 
890 out_unlock:
891 	mutex_unlock(&priv->reg_lock);
892 
893 	return ret;
894 }
895 
ub960_select_ind_reg_block(struct ub960_data * priv,u8 block)896 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
897 {
898 	struct device *dev = &priv->client->dev;
899 	int ret;
900 
901 	lockdep_assert_held(&priv->reg_lock);
902 
903 	if (priv->reg_current.indirect_target == block)
904 		return 0;
905 
906 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
907 	if (ret) {
908 		dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
909 			__func__, block, ret);
910 		return ret;
911 	}
912 
913 	priv->reg_current.indirect_target = block;
914 
915 	return 0;
916 }
917 
ub960_read_ind(struct ub960_data * priv,u8 block,u8 reg,u8 * val)918 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
919 {
920 	struct device *dev = &priv->client->dev;
921 	unsigned int v;
922 	int ret;
923 
924 	mutex_lock(&priv->reg_lock);
925 
926 	ret = ub960_select_ind_reg_block(priv, block);
927 	if (ret)
928 		goto out_unlock;
929 
930 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
931 	if (ret) {
932 		dev_err(dev,
933 			"Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
934 			block, reg, ret);
935 		goto out_unlock;
936 	}
937 
938 	ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
939 	if (ret) {
940 		dev_err(dev,
941 			"Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
942 			block, reg, ret);
943 		goto out_unlock;
944 	}
945 
946 	*val = v;
947 
948 out_unlock:
949 	mutex_unlock(&priv->reg_lock);
950 
951 	return ret;
952 }
953 
ub960_write_ind(struct ub960_data * priv,u8 block,u8 reg,u8 val)954 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
955 {
956 	struct device *dev = &priv->client->dev;
957 	int ret;
958 
959 	mutex_lock(&priv->reg_lock);
960 
961 	ret = ub960_select_ind_reg_block(priv, block);
962 	if (ret)
963 		goto out_unlock;
964 
965 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
966 	if (ret) {
967 		dev_err(dev,
968 			"Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
969 			block, reg, ret);
970 		goto out_unlock;
971 	}
972 
973 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
974 	if (ret) {
975 		dev_err(dev,
976 			"Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
977 			block, reg, ret);
978 		goto out_unlock;
979 	}
980 
981 out_unlock:
982 	mutex_unlock(&priv->reg_lock);
983 
984 	return ret;
985 }
986 
ub960_ind_update_bits(struct ub960_data * priv,u8 block,u8 reg,u8 mask,u8 val)987 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
988 				 u8 mask, u8 val)
989 {
990 	struct device *dev = &priv->client->dev;
991 	int ret;
992 
993 	mutex_lock(&priv->reg_lock);
994 
995 	ret = ub960_select_ind_reg_block(priv, block);
996 	if (ret)
997 		goto out_unlock;
998 
999 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
1000 	if (ret) {
1001 		dev_err(dev,
1002 			"Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1003 			block, reg, ret);
1004 		goto out_unlock;
1005 	}
1006 
1007 	ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1008 				 val);
1009 	if (ret) {
1010 		dev_err(dev,
1011 			"Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1012 			block, reg, ret);
1013 		goto out_unlock;
1014 	}
1015 
1016 out_unlock:
1017 	mutex_unlock(&priv->reg_lock);
1018 
1019 	return ret;
1020 }
1021 
1022 /* -----------------------------------------------------------------------------
1023  * I2C-ATR (address translator)
1024  */
1025 
ub960_atr_attach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client,u16 alias)1026 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1027 				   const struct i2c_client *client, u16 alias)
1028 {
1029 	struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1030 	struct ub960_rxport *rxport = priv->rxports[chan_id];
1031 	struct device *dev = &priv->client->dev;
1032 	unsigned int reg_idx;
1033 
1034 	for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1035 		if (!rxport->aliased_clients[reg_idx])
1036 			break;
1037 	}
1038 
1039 	if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1040 		dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1041 		return -EADDRNOTAVAIL;
1042 	}
1043 
1044 	rxport->aliased_clients[reg_idx] = client;
1045 
1046 	ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1047 			   client->addr << 1);
1048 	ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1049 			   alias << 1);
1050 
1051 	dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1052 		rxport->nport, client->addr, alias, reg_idx);
1053 
1054 	return 0;
1055 }
1056 
ub960_atr_detach_client(struct i2c_atr * atr,u32 chan_id,const struct i2c_client * client)1057 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1058 				    const struct i2c_client *client)
1059 {
1060 	struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1061 	struct ub960_rxport *rxport = priv->rxports[chan_id];
1062 	struct device *dev = &priv->client->dev;
1063 	unsigned int reg_idx;
1064 
1065 	for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1066 		if (rxport->aliased_clients[reg_idx] == client)
1067 			break;
1068 	}
1069 
1070 	if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1071 		dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1072 			rxport->nport, client->addr);
1073 		return;
1074 	}
1075 
1076 	rxport->aliased_clients[reg_idx] = NULL;
1077 
1078 	ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1079 
1080 	dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1081 		client->addr, reg_idx);
1082 }
1083 
1084 static const struct i2c_atr_ops ub960_atr_ops = {
1085 	.attach_client = ub960_atr_attach_client,
1086 	.detach_client = ub960_atr_detach_client,
1087 };
1088 
ub960_init_atr(struct ub960_data * priv)1089 static int ub960_init_atr(struct ub960_data *priv)
1090 {
1091 	struct device *dev = &priv->client->dev;
1092 	struct i2c_adapter *parent_adap = priv->client->adapter;
1093 
1094 	priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1095 				priv->hw_data->num_rxports);
1096 	if (IS_ERR(priv->atr))
1097 		return PTR_ERR(priv->atr);
1098 
1099 	i2c_atr_set_driver_data(priv->atr, priv);
1100 
1101 	return 0;
1102 }
1103 
ub960_uninit_atr(struct ub960_data * priv)1104 static void ub960_uninit_atr(struct ub960_data *priv)
1105 {
1106 	i2c_atr_delete(priv->atr);
1107 	priv->atr = NULL;
1108 }
1109 
1110 /* -----------------------------------------------------------------------------
1111  * TX ports
1112  */
1113 
ub960_parse_dt_txport(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,u8 nport)1114 static int ub960_parse_dt_txport(struct ub960_data *priv,
1115 				 struct fwnode_handle *ep_fwnode,
1116 				 u8 nport)
1117 {
1118 	struct device *dev = &priv->client->dev;
1119 	struct v4l2_fwnode_endpoint vep = {};
1120 	struct ub960_txport *txport;
1121 	int ret;
1122 
1123 	txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1124 	if (!txport)
1125 		return -ENOMEM;
1126 
1127 	txport->priv = priv;
1128 	txport->nport = nport;
1129 
1130 	vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1131 	ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1132 	if (ret) {
1133 		dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1134 		goto err_free_txport;
1135 	}
1136 
1137 	txport->non_continous_clk = vep.bus.mipi_csi2.flags &
1138 				    V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK;
1139 
1140 	txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1141 
1142 	if (vep.nr_of_link_frequencies != 1) {
1143 		ret = -EINVAL;
1144 		goto err_free_vep;
1145 	}
1146 
1147 	priv->tx_link_freq[0] = vep.link_frequencies[0];
1148 	priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1149 
1150 	if (priv->tx_data_rate != MHZ(1600) &&
1151 	    priv->tx_data_rate != MHZ(1200) &&
1152 	    priv->tx_data_rate != MHZ(800) &&
1153 	    priv->tx_data_rate != MHZ(400)) {
1154 		dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1155 		ret = -EINVAL;
1156 		goto err_free_vep;
1157 	}
1158 
1159 	v4l2_fwnode_endpoint_free(&vep);
1160 
1161 	priv->txports[nport] = txport;
1162 
1163 	return 0;
1164 
1165 err_free_vep:
1166 	v4l2_fwnode_endpoint_free(&vep);
1167 err_free_txport:
1168 	kfree(txport);
1169 
1170 	return ret;
1171 }
1172 
ub960_csi_handle_events(struct ub960_data * priv,u8 nport)1173 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1174 {
1175 	struct device *dev = &priv->client->dev;
1176 	u8 csi_tx_isr;
1177 	int ret;
1178 
1179 	ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1180 	if (ret)
1181 		return;
1182 
1183 	if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1184 		dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1185 
1186 	if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1187 		dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1188 }
1189 
1190 /* -----------------------------------------------------------------------------
1191  * RX ports
1192  */
1193 
ub960_rxport_enable_vpocs(struct ub960_data * priv)1194 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1195 {
1196 	unsigned int nport;
1197 	int ret;
1198 
1199 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1200 		struct ub960_rxport *rxport = priv->rxports[nport];
1201 
1202 		if (!rxport || !rxport->vpoc)
1203 			continue;
1204 
1205 		ret = regulator_enable(rxport->vpoc);
1206 		if (ret)
1207 			goto err_disable_vpocs;
1208 	}
1209 
1210 	return 0;
1211 
1212 err_disable_vpocs:
1213 	while (nport--) {
1214 		struct ub960_rxport *rxport = priv->rxports[nport];
1215 
1216 		if (!rxport || !rxport->vpoc)
1217 			continue;
1218 
1219 		regulator_disable(rxport->vpoc);
1220 	}
1221 
1222 	return ret;
1223 }
1224 
ub960_rxport_disable_vpocs(struct ub960_data * priv)1225 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1226 {
1227 	unsigned int nport;
1228 
1229 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1230 		struct ub960_rxport *rxport = priv->rxports[nport];
1231 
1232 		if (!rxport || !rxport->vpoc)
1233 			continue;
1234 
1235 		regulator_disable(rxport->vpoc);
1236 	}
1237 }
1238 
ub960_rxport_clear_errors(struct ub960_data * priv,unsigned int nport)1239 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1240 				      unsigned int nport)
1241 {
1242 	u8 v;
1243 
1244 	ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1245 	ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1246 	ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1247 	ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1248 
1249 	ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1250 	ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1251 
1252 	ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1253 }
1254 
ub960_clear_rx_errors(struct ub960_data * priv)1255 static void ub960_clear_rx_errors(struct ub960_data *priv)
1256 {
1257 	unsigned int nport;
1258 
1259 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1260 		ub960_rxport_clear_errors(priv, nport);
1261 }
1262 
ub960_rxport_get_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 * strobe_pos)1263 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1264 				       unsigned int nport, s8 *strobe_pos)
1265 {
1266 	u8 v;
1267 	u8 clk_delay, data_delay;
1268 	int ret;
1269 
1270 	ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1271 		       UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1272 
1273 	clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1274 			    0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1275 
1276 	ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1277 		       UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1278 
1279 	data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1280 			     0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1281 
1282 	ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1283 	if (ret)
1284 		return ret;
1285 
1286 	clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1287 
1288 	ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1289 	if (ret)
1290 		return ret;
1291 
1292 	data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1293 
1294 	*strobe_pos = data_delay - clk_delay;
1295 
1296 	return 0;
1297 }
1298 
ub960_rxport_set_strobe_pos(struct ub960_data * priv,unsigned int nport,s8 strobe_pos)1299 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1300 					unsigned int nport, s8 strobe_pos)
1301 {
1302 	u8 clk_delay, data_delay;
1303 
1304 	clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1305 	data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1306 
1307 	if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1308 		clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1309 	else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1310 		data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1311 	else if (strobe_pos < 0)
1312 		clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1313 	else if (strobe_pos > 0)
1314 		data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1315 
1316 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1317 			UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1318 
1319 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1320 			UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1321 }
1322 
ub960_rxport_set_strobe_range(struct ub960_data * priv,s8 strobe_min,s8 strobe_max)1323 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1324 					  s8 strobe_min, s8 strobe_max)
1325 {
1326 	/* Convert the signed strobe pos to positive zero based value */
1327 	strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1328 	strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1329 
1330 	ub960_write(priv, UB960_XR_SFILTER_CFG,
1331 		    ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1332 		    ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1333 }
1334 
ub960_rxport_get_eq_level(struct ub960_data * priv,unsigned int nport,u8 * eq_level)1335 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1336 				     unsigned int nport, u8 *eq_level)
1337 {
1338 	int ret;
1339 	u8 v;
1340 
1341 	ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1342 	if (ret)
1343 		return ret;
1344 
1345 	*eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1346 		    (v & UB960_RR_AEQ_STATUS_STATUS_2);
1347 
1348 	return 0;
1349 }
1350 
ub960_rxport_set_eq_level(struct ub960_data * priv,unsigned int nport,u8 eq_level)1351 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1352 				      unsigned int nport, u8 eq_level)
1353 {
1354 	u8 eq_stage_1_select_value, eq_stage_2_select_value;
1355 	const unsigned int eq_stage_max = 7;
1356 	u8 v;
1357 
1358 	if (eq_level <= eq_stage_max) {
1359 		eq_stage_1_select_value = eq_level;
1360 		eq_stage_2_select_value = 0;
1361 	} else {
1362 		eq_stage_1_select_value = eq_stage_max;
1363 		eq_stage_2_select_value = eq_level - eq_stage_max;
1364 	}
1365 
1366 	ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1367 
1368 	v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1369 	       UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1370 	v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1371 	v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1372 	v |= UB960_RR_AEQ_BYPASS_ENABLE;
1373 
1374 	ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1375 }
1376 
ub960_rxport_set_eq_range(struct ub960_data * priv,unsigned int nport,u8 eq_min,u8 eq_max)1377 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1378 				      unsigned int nport, u8 eq_min, u8 eq_max)
1379 {
1380 	ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1381 			   (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1382 			   (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1383 
1384 	/* Enable AEQ min setting */
1385 	ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1386 				 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1387 				 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1388 }
1389 
ub960_rxport_config_eq(struct ub960_data * priv,unsigned int nport)1390 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1391 {
1392 	struct ub960_rxport *rxport = priv->rxports[nport];
1393 
1394 	/* We also set common settings here. Should be moved elsewhere. */
1395 
1396 	if (priv->strobe.manual) {
1397 		/* Disable AEQ_SFILTER_EN */
1398 		ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1399 				  UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1400 	} else {
1401 		/* Enable SFILTER and error control */
1402 		ub960_write(priv, UB960_XR_AEQ_CTL1,
1403 			    UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1404 				    UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1405 
1406 		/* Set AEQ strobe range */
1407 		ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1408 					      priv->strobe.max);
1409 	}
1410 
1411 	/* The rest are port specific */
1412 
1413 	if (priv->strobe.manual)
1414 		ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1415 	else
1416 		ub960_rxport_set_strobe_pos(priv, nport, 0);
1417 
1418 	if (rxport->eq.manual_eq) {
1419 		ub960_rxport_set_eq_level(priv, nport,
1420 					  rxport->eq.manual.eq_level);
1421 
1422 		/* Enable AEQ Bypass */
1423 		ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1424 					 UB960_RR_AEQ_BYPASS_ENABLE,
1425 					 UB960_RR_AEQ_BYPASS_ENABLE);
1426 	} else {
1427 		ub960_rxport_set_eq_range(priv, nport,
1428 					  rxport->eq.aeq.eq_level_min,
1429 					  rxport->eq.aeq.eq_level_max);
1430 
1431 		/* Disable AEQ Bypass */
1432 		ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1433 					 UB960_RR_AEQ_BYPASS_ENABLE, 0);
1434 	}
1435 }
1436 
ub960_rxport_link_ok(struct ub960_data * priv,unsigned int nport,bool * ok)1437 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1438 				bool *ok)
1439 {
1440 	u8 rx_port_sts1, rx_port_sts2;
1441 	u16 parity_errors;
1442 	u8 csi_rx_sts;
1443 	u8 csi_err_cnt;
1444 	u8 bcc_sts;
1445 	int ret;
1446 	bool errors;
1447 
1448 	ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1449 				&rx_port_sts1);
1450 	if (ret)
1451 		return ret;
1452 
1453 	if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1454 		*ok = false;
1455 		return 0;
1456 	}
1457 
1458 	ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1459 				&rx_port_sts2);
1460 	if (ret)
1461 		return ret;
1462 
1463 	ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1464 	if (ret)
1465 		return ret;
1466 
1467 	ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1468 				&csi_err_cnt);
1469 	if (ret)
1470 		return ret;
1471 
1472 	ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1473 	if (ret)
1474 		return ret;
1475 
1476 	ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1477 				  &parity_errors);
1478 	if (ret)
1479 		return ret;
1480 
1481 	errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1482 		 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1483 		 (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1484 		 (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1485 		 parity_errors;
1486 
1487 	*ok = !errors;
1488 
1489 	return 0;
1490 }
1491 
1492 /*
1493  * Wait for the RX ports to lock, have no errors and have stable strobe position
1494  * and EQ level.
1495  */
ub960_rxport_wait_locks(struct ub960_data * priv,unsigned long port_mask,unsigned int * lock_mask)1496 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1497 				   unsigned long port_mask,
1498 				   unsigned int *lock_mask)
1499 {
1500 	struct device *dev = &priv->client->dev;
1501 	unsigned long timeout;
1502 	unsigned int link_ok_mask;
1503 	unsigned int missing;
1504 	unsigned int loops;
1505 	u8 nport;
1506 	int ret;
1507 
1508 	if (port_mask == 0) {
1509 		if (lock_mask)
1510 			*lock_mask = 0;
1511 		return 0;
1512 	}
1513 
1514 	if (port_mask >= BIT(priv->hw_data->num_rxports))
1515 		return -EINVAL;
1516 
1517 	timeout = jiffies + msecs_to_jiffies(1000);
1518 	loops = 0;
1519 	link_ok_mask = 0;
1520 
1521 	while (time_before(jiffies, timeout)) {
1522 		missing = 0;
1523 
1524 		for_each_set_bit(nport, &port_mask,
1525 				 priv->hw_data->num_rxports) {
1526 			struct ub960_rxport *rxport = priv->rxports[nport];
1527 			bool ok;
1528 
1529 			if (!rxport)
1530 				continue;
1531 
1532 			ret = ub960_rxport_link_ok(priv, nport, &ok);
1533 			if (ret)
1534 				return ret;
1535 
1536 			/*
1537 			 * We want the link to be ok for two consecutive loops,
1538 			 * as a link could get established just before our test
1539 			 * and drop soon after.
1540 			 */
1541 			if (!ok || !(link_ok_mask & BIT(nport)))
1542 				missing++;
1543 
1544 			if (ok)
1545 				link_ok_mask |= BIT(nport);
1546 			else
1547 				link_ok_mask &= ~BIT(nport);
1548 		}
1549 
1550 		loops++;
1551 
1552 		if (missing == 0)
1553 			break;
1554 
1555 		msleep(50);
1556 	}
1557 
1558 	if (lock_mask)
1559 		*lock_mask = link_ok_mask;
1560 
1561 	dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1562 	for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1563 		struct ub960_rxport *rxport = priv->rxports[nport];
1564 		s8 strobe_pos, eq_level;
1565 		u16 v;
1566 
1567 		if (!rxport)
1568 			continue;
1569 
1570 		if (!(link_ok_mask & BIT(nport))) {
1571 			dev_dbg(dev, "\trx%u: not locked\n", nport);
1572 			continue;
1573 		}
1574 
1575 		ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1576 
1577 		ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
1578 		if (ret)
1579 			return ret;
1580 
1581 		ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1582 		if (ret)
1583 			return ret;
1584 
1585 		dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1586 			nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
1587 	}
1588 
1589 	return 0;
1590 }
1591 
ub960_calc_bc_clk_rate_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1592 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1593 						  struct ub960_rxport *rxport)
1594 {
1595 	unsigned int mult;
1596 	unsigned int div;
1597 
1598 	switch (rxport->rx_mode) {
1599 	case RXPORT_MODE_RAW10:
1600 	case RXPORT_MODE_RAW12_HF:
1601 	case RXPORT_MODE_RAW12_LF:
1602 		mult = 1;
1603 		div = 10;
1604 		break;
1605 
1606 	case RXPORT_MODE_CSI2_SYNC:
1607 		mult = 2;
1608 		div = 1;
1609 		break;
1610 
1611 	case RXPORT_MODE_CSI2_NONSYNC:
1612 		mult = 2;
1613 		div = 5;
1614 		break;
1615 
1616 	default:
1617 		return 0;
1618 	}
1619 
1620 	return clk_get_rate(priv->refclk) * mult / div;
1621 }
1622 
ub960_calc_bc_clk_rate_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)1623 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1624 						   struct ub960_rxport *rxport)
1625 {
1626 	switch (rxport->rx_mode) {
1627 	case RXPORT_MODE_RAW10:
1628 	case RXPORT_MODE_RAW12_HF:
1629 	case RXPORT_MODE_RAW12_LF:
1630 		return 2359400;
1631 
1632 	case RXPORT_MODE_CSI2_SYNC:
1633 		return 47187500;
1634 
1635 	case RXPORT_MODE_CSI2_NONSYNC:
1636 		return 9437500;
1637 
1638 	default:
1639 		return 0;
1640 	}
1641 }
1642 
ub960_rxport_add_serializer(struct ub960_data * priv,u8 nport)1643 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1644 {
1645 	struct ub960_rxport *rxport = priv->rxports[nport];
1646 	struct device *dev = &priv->client->dev;
1647 	struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1648 	struct i2c_board_info ser_info = {
1649 		.of_node = to_of_node(rxport->ser.fwnode),
1650 		.fwnode = rxport->ser.fwnode,
1651 		.platform_data = ser_pdata,
1652 	};
1653 
1654 	ser_pdata->port = nport;
1655 	ser_pdata->atr = priv->atr;
1656 	if (priv->hw_data->is_ub9702)
1657 		ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1658 	else
1659 		ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1660 
1661 	/*
1662 	 * The serializer is added under the same i2c adapter as the
1663 	 * deserializer. This is not quite right, as the serializer is behind
1664 	 * the FPD-Link.
1665 	 */
1666 	ser_info.addr = rxport->ser.alias;
1667 	rxport->ser.client =
1668 		i2c_new_client_device(priv->client->adapter, &ser_info);
1669 	if (IS_ERR(rxport->ser.client)) {
1670 		dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1671 			ser_info.type);
1672 		return PTR_ERR(rxport->ser.client);
1673 	}
1674 
1675 	dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1676 		nport, rxport->ser.client->addr,
1677 		rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1678 
1679 	return 0;
1680 }
1681 
ub960_rxport_remove_serializer(struct ub960_data * priv,u8 nport)1682 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1683 {
1684 	struct ub960_rxport *rxport = priv->rxports[nport];
1685 
1686 	i2c_unregister_device(rxport->ser.client);
1687 	rxport->ser.client = NULL;
1688 }
1689 
1690 /* Add serializer i2c devices for all initialized ports */
ub960_rxport_add_serializers(struct ub960_data * priv)1691 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1692 {
1693 	unsigned int nport;
1694 	int ret;
1695 
1696 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1697 		struct ub960_rxport *rxport = priv->rxports[nport];
1698 
1699 		if (!rxport)
1700 			continue;
1701 
1702 		ret = ub960_rxport_add_serializer(priv, nport);
1703 		if (ret)
1704 			goto err_remove_sers;
1705 	}
1706 
1707 	return 0;
1708 
1709 err_remove_sers:
1710 	while (nport--) {
1711 		struct ub960_rxport *rxport = priv->rxports[nport];
1712 
1713 		if (!rxport)
1714 			continue;
1715 
1716 		ub960_rxport_remove_serializer(priv, nport);
1717 	}
1718 
1719 	return ret;
1720 }
1721 
ub960_rxport_remove_serializers(struct ub960_data * priv)1722 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1723 {
1724 	unsigned int nport;
1725 
1726 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1727 		struct ub960_rxport *rxport = priv->rxports[nport];
1728 
1729 		if (!rxport)
1730 			continue;
1731 
1732 		ub960_rxport_remove_serializer(priv, nport);
1733 	}
1734 }
1735 
ub960_init_tx_port(struct ub960_data * priv,struct ub960_txport * txport)1736 static void ub960_init_tx_port(struct ub960_data *priv,
1737 			       struct ub960_txport *txport)
1738 {
1739 	unsigned int nport = txport->nport;
1740 	u8 csi_ctl = 0;
1741 
1742 	/*
1743 	 * From the datasheet: "initial CSI Skew-Calibration
1744 	 * sequence [...] should be set when operating at 1.6 Gbps"
1745 	 */
1746 	if (priv->tx_data_rate == MHZ(1600))
1747 		csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1748 
1749 	csi_ctl |= (4 - txport->num_data_lanes) << 4;
1750 
1751 	if (!txport->non_continous_clk)
1752 		csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK;
1753 
1754 	ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1755 }
1756 
ub960_init_tx_ports(struct ub960_data * priv)1757 static int ub960_init_tx_ports(struct ub960_data *priv)
1758 {
1759 	unsigned int nport;
1760 	u8 speed_select;
1761 	u8 pll_div;
1762 
1763 	/* TX ports */
1764 
1765 	switch (priv->tx_data_rate) {
1766 	case MHZ(1600):
1767 	default:
1768 		speed_select = 0;
1769 		pll_div = 0x10;
1770 		break;
1771 	case MHZ(1200):
1772 		speed_select = 1;
1773 		pll_div = 0x18;
1774 		break;
1775 	case MHZ(800):
1776 		speed_select = 2;
1777 		pll_div = 0x10;
1778 		break;
1779 	case MHZ(400):
1780 		speed_select = 3;
1781 		pll_div = 0x10;
1782 		break;
1783 	}
1784 
1785 	ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1786 
1787 	if (priv->hw_data->is_ub9702) {
1788 		ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1789 
1790 		switch (priv->tx_data_rate) {
1791 		case MHZ(1600):
1792 		default:
1793 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1794 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1795 			break;
1796 		case MHZ(800):
1797 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1798 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1799 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1800 			break;
1801 		case MHZ(400):
1802 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1803 			break;
1804 		}
1805 	}
1806 
1807 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1808 		struct ub960_txport *txport = priv->txports[nport];
1809 
1810 		if (!txport)
1811 			continue;
1812 
1813 		ub960_init_tx_port(priv, txport);
1814 	}
1815 
1816 	return 0;
1817 }
1818 
ub960_init_rx_port_ub960(struct ub960_data * priv,struct ub960_rxport * rxport)1819 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1820 				     struct ub960_rxport *rxport)
1821 {
1822 	unsigned int nport = rxport->nport;
1823 	u32 bc_freq_val;
1824 
1825 	/*
1826 	 * Back channel frequency select.
1827 	 * Override FREQ_SELECT from the strap.
1828 	 * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1829 	 * 2 - 10 Mbps
1830 	 * 6 - 50 Mbps (DS90UB953-Q1)
1831 	 *
1832 	 * Note that changing this setting will result in some errors on the back
1833 	 * channel for a short period of time.
1834 	 */
1835 
1836 	switch (rxport->rx_mode) {
1837 	case RXPORT_MODE_RAW10:
1838 	case RXPORT_MODE_RAW12_HF:
1839 	case RXPORT_MODE_RAW12_LF:
1840 		bc_freq_val = 0;
1841 		break;
1842 
1843 	case RXPORT_MODE_CSI2_NONSYNC:
1844 		bc_freq_val = 2;
1845 		break;
1846 
1847 	case RXPORT_MODE_CSI2_SYNC:
1848 		bc_freq_val = 6;
1849 		break;
1850 
1851 	default:
1852 		return;
1853 	}
1854 
1855 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1856 				 UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1857 				 bc_freq_val);
1858 
1859 	switch (rxport->rx_mode) {
1860 	case RXPORT_MODE_RAW10:
1861 		/* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1862 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1863 					 UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1864 					 0x3);
1865 
1866 		/*
1867 		 * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1868 		 */
1869 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1870 			UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1871 			0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1872 
1873 		break;
1874 
1875 	case RXPORT_MODE_RAW12_HF:
1876 	case RXPORT_MODE_RAW12_LF:
1877 		/* Not implemented */
1878 		return;
1879 
1880 	case RXPORT_MODE_CSI2_SYNC:
1881 	case RXPORT_MODE_CSI2_NONSYNC:
1882 		/* CSI-2 Mode (DS90UB953-Q1 compatible) */
1883 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1884 					 0x0);
1885 
1886 		break;
1887 	}
1888 
1889 	/* LV_POLARITY & FV_POLARITY */
1890 	ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1891 				 rxport->lv_fv_pol);
1892 
1893 	/* Enable all interrupt sources from this port */
1894 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1895 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1896 
1897 	/* Enable I2C_PASS_THROUGH */
1898 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1899 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1900 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1901 
1902 	/* Enable I2C communication to the serializer via the alias addr */
1903 	ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1904 			   rxport->ser.alias << 1);
1905 
1906 	/* Configure EQ related settings */
1907 	ub960_rxport_config_eq(priv, nport);
1908 
1909 	/* Enable RX port */
1910 	ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1911 }
1912 
ub960_init_rx_port_ub9702_fpd3(struct ub960_data * priv,struct ub960_rxport * rxport)1913 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1914 					   struct ub960_rxport *rxport)
1915 {
1916 	unsigned int nport = rxport->nport;
1917 	u8 bc_freq_val;
1918 	u8 fpd_func_mode;
1919 
1920 	switch (rxport->rx_mode) {
1921 	case RXPORT_MODE_RAW10:
1922 		bc_freq_val = 0;
1923 		fpd_func_mode = 5;
1924 		break;
1925 
1926 	case RXPORT_MODE_RAW12_HF:
1927 		bc_freq_val = 0;
1928 		fpd_func_mode = 4;
1929 		break;
1930 
1931 	case RXPORT_MODE_RAW12_LF:
1932 		bc_freq_val = 0;
1933 		fpd_func_mode = 6;
1934 		break;
1935 
1936 	case RXPORT_MODE_CSI2_SYNC:
1937 		bc_freq_val = 6;
1938 		fpd_func_mode = 2;
1939 		break;
1940 
1941 	case RXPORT_MODE_CSI2_NONSYNC:
1942 		bc_freq_val = 2;
1943 		fpd_func_mode = 2;
1944 		break;
1945 
1946 	default:
1947 		return;
1948 	}
1949 
1950 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1951 				 bc_freq_val);
1952 	ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1953 
1954 	/* set serdes_eq_mode = 1 */
1955 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1956 
1957 	/* enable serdes driver */
1958 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1959 
1960 	/* set serdes_eq_offset=4 */
1961 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1962 
1963 	/* init default serdes_eq_max in 0xa9 */
1964 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
1965 
1966 	/* init serdes_eq_min in 0xaa */
1967 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
1968 
1969 	/* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
1970 	ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
1971 			      BIT(3), BIT(3));
1972 
1973 	/* RX port to half-rate */
1974 	ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
1975 			  BIT(nport * 2));
1976 }
1977 
ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data * priv,struct ub960_rxport * rxport)1978 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
1979 					       struct ub960_rxport *rxport)
1980 {
1981 	unsigned int nport = rxport->nport;
1982 	bool first_time_power_up = true;
1983 
1984 	if (first_time_power_up) {
1985 		u8 v;
1986 
1987 		/* AEQ init */
1988 		ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
1989 
1990 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
1991 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
1992 
1993 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
1994 	}
1995 
1996 	/* enable serdes_eq_ctl2 */
1997 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
1998 
1999 	/* enable serdes_eq_ctl1 */
2000 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
2001 
2002 	/* enable serdes_eq_en */
2003 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
2004 
2005 	/* disable serdes_eq_override */
2006 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
2007 
2008 	/* disable serdes_gain_override */
2009 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2010 }
2011 
ub960_init_rx_port_ub9702_fpd4(struct ub960_data * priv,struct ub960_rxport * rxport)2012 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2013 					   struct ub960_rxport *rxport)
2014 {
2015 	unsigned int nport = rxport->nport;
2016 	u8 bc_freq_val;
2017 
2018 	switch (rxport->rx_mode) {
2019 	case RXPORT_MODE_RAW10:
2020 		bc_freq_val = 0;
2021 		break;
2022 
2023 	case RXPORT_MODE_RAW12_HF:
2024 		bc_freq_val = 0;
2025 		break;
2026 
2027 	case RXPORT_MODE_RAW12_LF:
2028 		bc_freq_val = 0;
2029 		break;
2030 
2031 	case RXPORT_MODE_CSI2_SYNC:
2032 		bc_freq_val = 6;
2033 		break;
2034 
2035 	case RXPORT_MODE_CSI2_NONSYNC:
2036 		bc_freq_val = 2;
2037 		break;
2038 
2039 	default:
2040 		return;
2041 	}
2042 
2043 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2044 				 bc_freq_val);
2045 
2046 	/* FPD4 Sync Mode */
2047 	ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2048 
2049 	/* add serdes_eq_offset of 4 */
2050 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2051 
2052 	/* FPD4 serdes_start_eq in 0x27: assign default */
2053 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2054 	/* FPD4 serdes_end_eq in 0x28: assign default */
2055 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2056 
2057 	/* set serdes_driver_mode into FPD IV mode */
2058 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2059 	/* set FPD PBC drv into FPD IV mode */
2060 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2061 
2062 	/* set serdes_system_init to 0x2f */
2063 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2064 	/* set serdes_system_rst in reset mode */
2065 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2066 
2067 	/* RX port to 7.55G mode */
2068 	ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2069 			  0 << (nport * 2));
2070 
2071 	ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2072 }
2073 
ub960_init_rx_port_ub9702(struct ub960_data * priv,struct ub960_rxport * rxport)2074 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2075 				      struct ub960_rxport *rxport)
2076 {
2077 	unsigned int nport = rxport->nport;
2078 
2079 	if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2080 		ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2081 	else /* RXPORT_CDR_FPD4 */
2082 		ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2083 
2084 	switch (rxport->rx_mode) {
2085 	case RXPORT_MODE_RAW10:
2086 		/*
2087 		 * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2088 		 * 0b10 : 8-bit processing using upper 8 bits
2089 		 */
2090 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2091 					 0x3 << 6, 0x2 << 6);
2092 
2093 		break;
2094 
2095 	case RXPORT_MODE_RAW12_HF:
2096 	case RXPORT_MODE_RAW12_LF:
2097 		/* Not implemented */
2098 		return;
2099 
2100 	case RXPORT_MODE_CSI2_SYNC:
2101 	case RXPORT_MODE_CSI2_NONSYNC:
2102 
2103 		break;
2104 	}
2105 
2106 	/* LV_POLARITY & FV_POLARITY */
2107 	ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2108 				 rxport->lv_fv_pol);
2109 
2110 	/* Enable all interrupt sources from this port */
2111 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2112 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2113 
2114 	/* Enable I2C_PASS_THROUGH */
2115 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2116 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2117 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2118 
2119 	/* Enable I2C communication to the serializer via the alias addr */
2120 	ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2121 			   rxport->ser.alias << 1);
2122 
2123 	/* Enable RX port */
2124 	ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2125 
2126 	if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2127 		/* unreset 960 AEQ */
2128 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2129 	}
2130 }
2131 
ub960_init_rx_ports(struct ub960_data * priv)2132 static int ub960_init_rx_ports(struct ub960_data *priv)
2133 {
2134 	unsigned int nport;
2135 
2136 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2137 		struct ub960_rxport *rxport = priv->rxports[nport];
2138 
2139 		if (!rxport)
2140 			continue;
2141 
2142 		if (priv->hw_data->is_ub9702)
2143 			ub960_init_rx_port_ub9702(priv, rxport);
2144 		else
2145 			ub960_init_rx_port_ub960(priv, rxport);
2146 	}
2147 
2148 	return 0;
2149 }
2150 
ub960_rxport_handle_events(struct ub960_data * priv,u8 nport)2151 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2152 {
2153 	struct device *dev = &priv->client->dev;
2154 	u8 rx_port_sts1;
2155 	u8 rx_port_sts2;
2156 	u8 csi_rx_sts;
2157 	u8 bcc_sts;
2158 	int ret = 0;
2159 
2160 	/* Read interrupts (also clears most of them) */
2161 	if (!ret)
2162 		ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2163 					&rx_port_sts1);
2164 	if (!ret)
2165 		ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2166 					&rx_port_sts2);
2167 	if (!ret)
2168 		ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2169 					&csi_rx_sts);
2170 	if (!ret)
2171 		ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2172 					&bcc_sts);
2173 
2174 	if (ret)
2175 		return;
2176 
2177 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2178 		u16 v;
2179 
2180 		ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2181 					  &v);
2182 		if (!ret)
2183 			dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2184 	}
2185 
2186 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2187 		dev_err(dev, "rx%u BCC CRC error\n", nport);
2188 
2189 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2190 		dev_err(dev, "rx%u BCC SEQ error\n", nport);
2191 
2192 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2193 		dev_err(dev, "rx%u line length unstable\n", nport);
2194 
2195 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2196 		dev_err(dev, "rx%u FPD3 encode error\n", nport);
2197 
2198 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2199 		dev_err(dev, "rx%u buffer error\n", nport);
2200 
2201 	if (csi_rx_sts)
2202 		dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2203 
2204 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2205 		dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2206 
2207 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2208 		dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2209 
2210 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2211 		dev_err(dev, "rx%u CSI checksum error\n", nport);
2212 
2213 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2214 		dev_err(dev, "rx%u CSI length error\n", nport);
2215 
2216 	if (bcc_sts)
2217 		dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2218 
2219 	if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2220 		dev_err(dev, "rx%u BCC response error", nport);
2221 
2222 	if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2223 		dev_err(dev, "rx%u BCC slave timeout", nport);
2224 
2225 	if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2226 		dev_err(dev, "rx%u BCC slave error", nport);
2227 
2228 	if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2229 		dev_err(dev, "rx%u BCC master timeout", nport);
2230 
2231 	if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2232 		dev_err(dev, "rx%u BCC master error", nport);
2233 
2234 	if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2235 		dev_err(dev, "rx%u BCC sequence error", nport);
2236 
2237 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2238 		u16 v;
2239 
2240 		ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2241 		if (!ret)
2242 			dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2243 	}
2244 
2245 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2246 		u16 v;
2247 
2248 		ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2249 					  &v);
2250 		if (!ret)
2251 			dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2252 	}
2253 
2254 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2255 		dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2256 			(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2257 				"locked" :
2258 				"unlocked",
2259 			(rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2260 				"passed" :
2261 				"not passed",
2262 			(rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2263 				"no clock" :
2264 				"clock ok",
2265 			(rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2266 				"stable freq" :
2267 				"unstable freq");
2268 	}
2269 }
2270 
2271 /* -----------------------------------------------------------------------------
2272  * V4L2
2273  */
2274 
2275 /*
2276  * The current implementation only supports a simple VC mapping, where all VCs
2277  * from a one RX port will be mapped to the same VC. Also, the hardware
2278  * dictates that all streams from an RX port must go to a single TX port.
2279  *
2280  * This function decides the target VC numbers for each RX port with a simple
2281  * algorithm, so that for each TX port, we get VC numbers starting from 0,
2282  * and counting up.
2283  *
2284  * E.g. if all four RX ports are in use, of which the first two go to the
2285  * first TX port and the secont two go to the second TX port, we would get
2286  * the following VCs for the four RX ports: 0, 1, 0, 1.
2287  *
2288  * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2289  * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2290  * probably needs to be somehow configurable. Device tree?
2291  */
ub960_get_vc_maps(struct ub960_data * priv,struct v4l2_subdev_state * state,u8 * vc)2292 static void ub960_get_vc_maps(struct ub960_data *priv,
2293 			      struct v4l2_subdev_state *state, u8 *vc)
2294 {
2295 	u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2296 	struct v4l2_subdev_route *route;
2297 	u8 handled_mask = 0;
2298 
2299 	for_each_active_route(&state->routing, route) {
2300 		unsigned int rx, tx;
2301 
2302 		rx = ub960_pad_to_port(priv, route->sink_pad);
2303 		if (BIT(rx) & handled_mask)
2304 			continue;
2305 
2306 		tx = ub960_pad_to_port(priv, route->source_pad);
2307 
2308 		vc[rx] = cur_vc[tx]++;
2309 		handled_mask |= BIT(rx);
2310 	}
2311 }
2312 
ub960_enable_tx_port(struct ub960_data * priv,unsigned int nport)2313 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2314 {
2315 	struct device *dev = &priv->client->dev;
2316 
2317 	dev_dbg(dev, "enable TX port %u\n", nport);
2318 
2319 	return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2320 					UB960_TR_CSI_CTL_CSI_ENABLE,
2321 					UB960_TR_CSI_CTL_CSI_ENABLE);
2322 }
2323 
ub960_disable_tx_port(struct ub960_data * priv,unsigned int nport)2324 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2325 {
2326 	struct device *dev = &priv->client->dev;
2327 
2328 	dev_dbg(dev, "disable TX port %u\n", nport);
2329 
2330 	ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2331 				 UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2332 }
2333 
ub960_enable_rx_port(struct ub960_data * priv,unsigned int nport)2334 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2335 {
2336 	struct device *dev = &priv->client->dev;
2337 
2338 	dev_dbg(dev, "enable RX port %u\n", nport);
2339 
2340 	/* Enable forwarding */
2341 	return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2342 				 UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2343 }
2344 
ub960_disable_rx_port(struct ub960_data * priv,unsigned int nport)2345 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2346 {
2347 	struct device *dev = &priv->client->dev;
2348 
2349 	dev_dbg(dev, "disable RX port %u\n", nport);
2350 
2351 	/* Disable forwarding */
2352 	ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2353 			  UB960_SR_FWD_CTL1_PORT_DIS(nport),
2354 			  UB960_SR_FWD_CTL1_PORT_DIS(nport));
2355 }
2356 
2357 /*
2358  * The driver only supports using a single VC for each source. This function
2359  * checks that each source only provides streams using a single VC.
2360  */
ub960_validate_stream_vcs(struct ub960_data * priv)2361 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2362 {
2363 	unsigned int nport;
2364 	unsigned int i;
2365 
2366 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2367 		struct ub960_rxport *rxport = priv->rxports[nport];
2368 		struct v4l2_mbus_frame_desc desc;
2369 		int ret;
2370 		u8 vc;
2371 
2372 		if (!rxport)
2373 			continue;
2374 
2375 		ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2376 				       rxport->source.pad, &desc);
2377 		if (ret)
2378 			return ret;
2379 
2380 		if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2381 			continue;
2382 
2383 		if (desc.num_entries == 0)
2384 			continue;
2385 
2386 		vc = desc.entry[0].bus.csi2.vc;
2387 
2388 		for (i = 1; i < desc.num_entries; i++) {
2389 			if (vc == desc.entry[i].bus.csi2.vc)
2390 				continue;
2391 
2392 			dev_err(&priv->client->dev,
2393 				"rx%u: source with multiple virtual-channels is not supported\n",
2394 				nport);
2395 			return -ENODEV;
2396 		}
2397 	}
2398 
2399 	return 0;
2400 }
2401 
ub960_configure_ports_for_streaming(struct ub960_data * priv,struct v4l2_subdev_state * state)2402 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2403 					       struct v4l2_subdev_state *state)
2404 {
2405 	u8 fwd_ctl;
2406 	struct {
2407 		u32 num_streams;
2408 		u8 pixel_dt;
2409 		u8 meta_dt;
2410 		u32 meta_lines;
2411 		u32 tx_port;
2412 	} rx_data[UB960_MAX_RX_NPORTS] = {};
2413 	u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2414 	struct v4l2_subdev_route *route;
2415 	unsigned int nport;
2416 	int ret;
2417 
2418 	ret = ub960_validate_stream_vcs(priv);
2419 	if (ret)
2420 		return ret;
2421 
2422 	ub960_get_vc_maps(priv, state, vc_map);
2423 
2424 	for_each_active_route(&state->routing, route) {
2425 		struct ub960_rxport *rxport;
2426 		struct ub960_txport *txport;
2427 		struct v4l2_mbus_framefmt *fmt;
2428 		const struct ub960_format_info *ub960_fmt;
2429 		unsigned int nport;
2430 
2431 		nport = ub960_pad_to_port(priv, route->sink_pad);
2432 
2433 		rxport = priv->rxports[nport];
2434 		if (!rxport)
2435 			return -EINVAL;
2436 
2437 		txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2438 		if (!txport)
2439 			return -EINVAL;
2440 
2441 		rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2442 
2443 		rx_data[nport].num_streams++;
2444 
2445 		/* For the rest, we are only interested in parallel busses */
2446 		if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2447 		    rxport->rx_mode == RXPORT_MODE_CSI2_NONSYNC)
2448 			continue;
2449 
2450 		if (rx_data[nport].num_streams > 2)
2451 			return -EPIPE;
2452 
2453 		fmt = v4l2_subdev_state_get_format(state, route->sink_pad,
2454 						   route->sink_stream);
2455 		if (!fmt)
2456 			return -EPIPE;
2457 
2458 		ub960_fmt = ub960_find_format(fmt->code);
2459 		if (!ub960_fmt)
2460 			return -EPIPE;
2461 
2462 		if (ub960_fmt->meta) {
2463 			if (fmt->height > 3) {
2464 				dev_err(&priv->client->dev,
2465 					"rx%u: unsupported metadata height %u\n",
2466 					nport, fmt->height);
2467 				return -EPIPE;
2468 			}
2469 
2470 			rx_data[nport].meta_dt = ub960_fmt->datatype;
2471 			rx_data[nport].meta_lines = fmt->height;
2472 		} else {
2473 			rx_data[nport].pixel_dt = ub960_fmt->datatype;
2474 		}
2475 	}
2476 
2477 	/* Configure RX ports */
2478 
2479 	/*
2480 	 * Keep all port forwardings disabled by default. Forwarding will be
2481 	 * enabled in ub960_enable_rx_port.
2482 	 */
2483 	fwd_ctl = GENMASK(7, 4);
2484 
2485 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2486 		struct ub960_rxport *rxport = priv->rxports[nport];
2487 		u8 vc = vc_map[nport];
2488 
2489 		if (rx_data[nport].num_streams == 0)
2490 			continue;
2491 
2492 		switch (rxport->rx_mode) {
2493 		case RXPORT_MODE_RAW10:
2494 			ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2495 				rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2496 
2497 			ub960_rxport_write(priv, rxport->nport,
2498 				UB960_RR_RAW_EMBED_DTYPE,
2499 				(rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2500 					rx_data[nport].meta_dt);
2501 
2502 			break;
2503 
2504 		case RXPORT_MODE_RAW12_HF:
2505 		case RXPORT_MODE_RAW12_LF:
2506 			/* Not implemented */
2507 			break;
2508 
2509 		case RXPORT_MODE_CSI2_SYNC:
2510 		case RXPORT_MODE_CSI2_NONSYNC:
2511 			if (!priv->hw_data->is_ub9702) {
2512 				/* Map all VCs from this port to the same VC */
2513 				ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2514 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2515 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2516 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2517 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2518 			} else {
2519 				unsigned int i;
2520 
2521 				/* Map all VCs from this port to VC(nport) */
2522 				for (i = 0; i < 8; i++)
2523 					ub960_rxport_write(priv, nport,
2524 							   UB960_RR_VC_ID_MAP(i),
2525 							   nport);
2526 			}
2527 
2528 			break;
2529 		}
2530 
2531 		if (rx_data[nport].tx_port == 1)
2532 			fwd_ctl |= BIT(nport); /* forward to TX1 */
2533 		else
2534 			fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2535 	}
2536 
2537 	ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2538 
2539 	return 0;
2540 }
2541 
ub960_update_streaming_status(struct ub960_data * priv)2542 static void ub960_update_streaming_status(struct ub960_data *priv)
2543 {
2544 	unsigned int i;
2545 
2546 	for (i = 0; i < UB960_MAX_NPORTS; i++) {
2547 		if (priv->stream_enable_mask[i])
2548 			break;
2549 	}
2550 
2551 	priv->streaming = i < UB960_MAX_NPORTS;
2552 }
2553 
ub960_enable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2554 static int ub960_enable_streams(struct v4l2_subdev *sd,
2555 				struct v4l2_subdev_state *state, u32 source_pad,
2556 				u64 source_streams_mask)
2557 {
2558 	struct ub960_data *priv = sd_to_ub960(sd);
2559 	struct device *dev = &priv->client->dev;
2560 	u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2561 	struct v4l2_subdev_route *route;
2562 	unsigned int failed_port;
2563 	unsigned int nport;
2564 	int ret;
2565 
2566 	if (!priv->streaming) {
2567 		dev_dbg(dev, "Prepare for streaming\n");
2568 		ret = ub960_configure_ports_for_streaming(priv, state);
2569 		if (ret)
2570 			return ret;
2571 	}
2572 
2573 	/* Enable TX port if not yet enabled */
2574 	if (!priv->stream_enable_mask[source_pad]) {
2575 		ret = ub960_enable_tx_port(priv,
2576 					   ub960_pad_to_port(priv, source_pad));
2577 		if (ret)
2578 			return ret;
2579 	}
2580 
2581 	priv->stream_enable_mask[source_pad] |= source_streams_mask;
2582 
2583 	/* Collect sink streams per pad which we need to enable */
2584 	for_each_active_route(&state->routing, route) {
2585 		if (route->source_pad != source_pad)
2586 			continue;
2587 
2588 		if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2589 			continue;
2590 
2591 		nport = ub960_pad_to_port(priv, route->sink_pad);
2592 
2593 		sink_streams[nport] |= BIT_ULL(route->sink_stream);
2594 	}
2595 
2596 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2597 		if (!sink_streams[nport])
2598 			continue;
2599 
2600 		/* Enable the RX port if not yet enabled */
2601 		if (!priv->stream_enable_mask[nport]) {
2602 			ret = ub960_enable_rx_port(priv, nport);
2603 			if (ret) {
2604 				failed_port = nport;
2605 				goto err;
2606 			}
2607 		}
2608 
2609 		priv->stream_enable_mask[nport] |= sink_streams[nport];
2610 
2611 		dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2612 			sink_streams[nport]);
2613 
2614 		ret = v4l2_subdev_enable_streams(
2615 			priv->rxports[nport]->source.sd,
2616 			priv->rxports[nport]->source.pad,
2617 			sink_streams[nport]);
2618 		if (ret) {
2619 			priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2620 
2621 			if (!priv->stream_enable_mask[nport])
2622 				ub960_disable_rx_port(priv, nport);
2623 
2624 			failed_port = nport;
2625 			goto err;
2626 		}
2627 	}
2628 
2629 	priv->streaming = true;
2630 
2631 	return 0;
2632 
2633 err:
2634 	for (nport = 0; nport < failed_port; nport++) {
2635 		if (!sink_streams[nport])
2636 			continue;
2637 
2638 		dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2639 			sink_streams[nport]);
2640 
2641 		ret = v4l2_subdev_disable_streams(
2642 			priv->rxports[nport]->source.sd,
2643 			priv->rxports[nport]->source.pad,
2644 			sink_streams[nport]);
2645 		if (ret)
2646 			dev_err(dev, "Failed to disable streams: %d\n", ret);
2647 
2648 		priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2649 
2650 		/* Disable RX port if no active streams */
2651 		if (!priv->stream_enable_mask[nport])
2652 			ub960_disable_rx_port(priv, nport);
2653 	}
2654 
2655 	priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2656 
2657 	if (!priv->stream_enable_mask[source_pad])
2658 		ub960_disable_tx_port(priv,
2659 				      ub960_pad_to_port(priv, source_pad));
2660 
2661 	ub960_update_streaming_status(priv);
2662 
2663 	return ret;
2664 }
2665 
ub960_disable_streams(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,u32 source_pad,u64 source_streams_mask)2666 static int ub960_disable_streams(struct v4l2_subdev *sd,
2667 				 struct v4l2_subdev_state *state,
2668 				 u32 source_pad, u64 source_streams_mask)
2669 {
2670 	struct ub960_data *priv = sd_to_ub960(sd);
2671 	struct device *dev = &priv->client->dev;
2672 	u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2673 	struct v4l2_subdev_route *route;
2674 	unsigned int nport;
2675 	int ret;
2676 
2677 	/* Collect sink streams per pad which we need to disable */
2678 	for_each_active_route(&state->routing, route) {
2679 		if (route->source_pad != source_pad)
2680 			continue;
2681 
2682 		if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2683 			continue;
2684 
2685 		nport = ub960_pad_to_port(priv, route->sink_pad);
2686 
2687 		sink_streams[nport] |= BIT_ULL(route->sink_stream);
2688 	}
2689 
2690 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2691 		if (!sink_streams[nport])
2692 			continue;
2693 
2694 		dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2695 			sink_streams[nport]);
2696 
2697 		ret = v4l2_subdev_disable_streams(
2698 			priv->rxports[nport]->source.sd,
2699 			priv->rxports[nport]->source.pad,
2700 			sink_streams[nport]);
2701 		if (ret)
2702 			dev_err(dev, "Failed to disable streams: %d\n", ret);
2703 
2704 		priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2705 
2706 		/* Disable RX port if no active streams */
2707 		if (!priv->stream_enable_mask[nport])
2708 			ub960_disable_rx_port(priv, nport);
2709 	}
2710 
2711 	/* Disable TX port if no active streams */
2712 
2713 	priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2714 
2715 	if (!priv->stream_enable_mask[source_pad])
2716 		ub960_disable_tx_port(priv,
2717 				      ub960_pad_to_port(priv, source_pad));
2718 
2719 	ub960_update_streaming_status(priv);
2720 
2721 	return 0;
2722 }
2723 
_ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_krouting * routing)2724 static int _ub960_set_routing(struct v4l2_subdev *sd,
2725 			      struct v4l2_subdev_state *state,
2726 			      struct v4l2_subdev_krouting *routing)
2727 {
2728 	static const struct v4l2_mbus_framefmt format = {
2729 		.width = 640,
2730 		.height = 480,
2731 		.code = MEDIA_BUS_FMT_UYVY8_1X16,
2732 		.field = V4L2_FIELD_NONE,
2733 		.colorspace = V4L2_COLORSPACE_SRGB,
2734 		.ycbcr_enc = V4L2_YCBCR_ENC_601,
2735 		.quantization = V4L2_QUANTIZATION_LIM_RANGE,
2736 		.xfer_func = V4L2_XFER_FUNC_SRGB,
2737 	};
2738 	int ret;
2739 
2740 	/*
2741 	 * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2742 	 * frame desc is made dynamically allocated.
2743 	 */
2744 
2745 	if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2746 		return -E2BIG;
2747 
2748 	ret = v4l2_subdev_routing_validate(sd, routing,
2749 					   V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2750 					   V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2751 	if (ret)
2752 		return ret;
2753 
2754 	ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2755 	if (ret)
2756 		return ret;
2757 
2758 	return 0;
2759 }
2760 
ub960_set_routing(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,enum v4l2_subdev_format_whence which,struct v4l2_subdev_krouting * routing)2761 static int ub960_set_routing(struct v4l2_subdev *sd,
2762 			     struct v4l2_subdev_state *state,
2763 			     enum v4l2_subdev_format_whence which,
2764 			     struct v4l2_subdev_krouting *routing)
2765 {
2766 	struct ub960_data *priv = sd_to_ub960(sd);
2767 
2768 	if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2769 		return -EBUSY;
2770 
2771 	return _ub960_set_routing(sd, state, routing);
2772 }
2773 
ub960_get_frame_desc(struct v4l2_subdev * sd,unsigned int pad,struct v4l2_mbus_frame_desc * fd)2774 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2775 				struct v4l2_mbus_frame_desc *fd)
2776 {
2777 	struct ub960_data *priv = sd_to_ub960(sd);
2778 	struct v4l2_subdev_route *route;
2779 	struct v4l2_subdev_state *state;
2780 	int ret = 0;
2781 	struct device *dev = &priv->client->dev;
2782 	u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2783 
2784 	if (!ub960_pad_is_source(priv, pad))
2785 		return -EINVAL;
2786 
2787 	fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2788 
2789 	state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2790 
2791 	ub960_get_vc_maps(priv, state, vc_map);
2792 
2793 	for_each_active_route(&state->routing, route) {
2794 		struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2795 		struct v4l2_mbus_frame_desc source_fd;
2796 		unsigned int nport;
2797 		unsigned int i;
2798 
2799 		if (route->source_pad != pad)
2800 			continue;
2801 
2802 		nport = ub960_pad_to_port(priv, route->sink_pad);
2803 
2804 		ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2805 				       get_frame_desc,
2806 				       priv->rxports[nport]->source.pad,
2807 				       &source_fd);
2808 		if (ret) {
2809 			dev_err(dev,
2810 				"Failed to get source frame desc for pad %u\n",
2811 				route->sink_pad);
2812 			goto out_unlock;
2813 		}
2814 
2815 		for (i = 0; i < source_fd.num_entries; i++) {
2816 			if (source_fd.entry[i].stream == route->sink_stream) {
2817 				source_entry = &source_fd.entry[i];
2818 				break;
2819 			}
2820 		}
2821 
2822 		if (!source_entry) {
2823 			dev_err(dev,
2824 				"Failed to find stream from source frame desc\n");
2825 			ret = -EPIPE;
2826 			goto out_unlock;
2827 		}
2828 
2829 		fd->entry[fd->num_entries].stream = route->source_stream;
2830 		fd->entry[fd->num_entries].flags = source_entry->flags;
2831 		fd->entry[fd->num_entries].length = source_entry->length;
2832 		fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2833 
2834 		fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2835 
2836 		if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2837 			fd->entry[fd->num_entries].bus.csi2.dt =
2838 				source_entry->bus.csi2.dt;
2839 		} else {
2840 			const struct ub960_format_info *ub960_fmt;
2841 			struct v4l2_mbus_framefmt *fmt;
2842 
2843 			fmt = v4l2_subdev_state_get_format(state, pad,
2844 							   route->source_stream);
2845 
2846 			if (!fmt) {
2847 				ret = -EINVAL;
2848 				goto out_unlock;
2849 			}
2850 
2851 			ub960_fmt = ub960_find_format(fmt->code);
2852 			if (!ub960_fmt) {
2853 				dev_err(dev, "Unable to find format\n");
2854 				ret = -EINVAL;
2855 				goto out_unlock;
2856 			}
2857 
2858 			fd->entry[fd->num_entries].bus.csi2.dt =
2859 				ub960_fmt->datatype;
2860 		}
2861 
2862 		fd->num_entries++;
2863 	}
2864 
2865 out_unlock:
2866 	v4l2_subdev_unlock_state(state);
2867 
2868 	return ret;
2869 }
2870 
ub960_set_fmt(struct v4l2_subdev * sd,struct v4l2_subdev_state * state,struct v4l2_subdev_format * format)2871 static int ub960_set_fmt(struct v4l2_subdev *sd,
2872 			 struct v4l2_subdev_state *state,
2873 			 struct v4l2_subdev_format *format)
2874 {
2875 	struct ub960_data *priv = sd_to_ub960(sd);
2876 	struct v4l2_mbus_framefmt *fmt;
2877 
2878 	if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2879 		return -EBUSY;
2880 
2881 	/* No transcoding, source and sink formats must match. */
2882 	if (ub960_pad_is_source(priv, format->pad))
2883 		return v4l2_subdev_get_fmt(sd, state, format);
2884 
2885 	/*
2886 	 * Default to the first format if the requested media bus code isn't
2887 	 * supported.
2888 	 */
2889 	if (!ub960_find_format(format->format.code))
2890 		format->format.code = ub960_formats[0].code;
2891 
2892 	fmt = v4l2_subdev_state_get_format(state, format->pad, format->stream);
2893 	if (!fmt)
2894 		return -EINVAL;
2895 
2896 	*fmt = format->format;
2897 
2898 	fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2899 							   format->stream);
2900 	if (!fmt)
2901 		return -EINVAL;
2902 
2903 	*fmt = format->format;
2904 
2905 	return 0;
2906 }
2907 
ub960_init_state(struct v4l2_subdev * sd,struct v4l2_subdev_state * state)2908 static int ub960_init_state(struct v4l2_subdev *sd,
2909 			    struct v4l2_subdev_state *state)
2910 {
2911 	struct ub960_data *priv = sd_to_ub960(sd);
2912 
2913 	struct v4l2_subdev_route routes[] = {
2914 		{
2915 			.sink_pad = 0,
2916 			.sink_stream = 0,
2917 			.source_pad = priv->hw_data->num_rxports,
2918 			.source_stream = 0,
2919 			.flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2920 		},
2921 	};
2922 
2923 	struct v4l2_subdev_krouting routing = {
2924 		.num_routes = ARRAY_SIZE(routes),
2925 		.routes = routes,
2926 	};
2927 
2928 	return _ub960_set_routing(sd, state, &routing);
2929 }
2930 
2931 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2932 	.enable_streams = ub960_enable_streams,
2933 	.disable_streams = ub960_disable_streams,
2934 
2935 	.set_routing = ub960_set_routing,
2936 	.get_frame_desc = ub960_get_frame_desc,
2937 
2938 	.get_fmt = v4l2_subdev_get_fmt,
2939 	.set_fmt = ub960_set_fmt,
2940 };
2941 
ub960_log_status(struct v4l2_subdev * sd)2942 static int ub960_log_status(struct v4l2_subdev *sd)
2943 {
2944 	struct ub960_data *priv = sd_to_ub960(sd);
2945 	struct device *dev = &priv->client->dev;
2946 	struct v4l2_subdev_state *state;
2947 	unsigned int nport;
2948 	unsigned int i;
2949 	u16 v16 = 0;
2950 	u8 v = 0;
2951 	u8 id[UB960_SR_FPD3_RX_ID_LEN];
2952 
2953 	state = v4l2_subdev_lock_and_get_active_state(sd);
2954 
2955 	for (i = 0; i < sizeof(id); i++)
2956 		ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
2957 
2958 	dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
2959 
2960 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
2961 		struct ub960_txport *txport = priv->txports[nport];
2962 
2963 		dev_info(dev, "TX %u\n", nport);
2964 
2965 		if (!txport) {
2966 			dev_info(dev, "\tNot initialized\n");
2967 			continue;
2968 		}
2969 
2970 		ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
2971 		dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
2972 			 v & (u8)BIT(0));
2973 
2974 		ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
2975 		dev_info(dev, "\tframe counter %u\n", v16);
2976 
2977 		ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
2978 		dev_info(dev, "\tframe error counter %u\n", v16);
2979 
2980 		ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
2981 		dev_info(dev, "\tline counter %u\n", v16);
2982 
2983 		ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
2984 		dev_info(dev, "\tline error counter %u\n", v16);
2985 	}
2986 
2987 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2988 		struct ub960_rxport *rxport = priv->rxports[nport];
2989 		u8 eq_level;
2990 		s8 strobe_pos;
2991 		unsigned int i;
2992 
2993 		dev_info(dev, "RX %u\n", nport);
2994 
2995 		if (!rxport) {
2996 			dev_info(dev, "\tNot initialized\n");
2997 			continue;
2998 		}
2999 
3000 		ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
3001 
3002 		if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3003 			dev_info(dev, "\tLocked\n");
3004 		else
3005 			dev_info(dev, "\tNot locked\n");
3006 
3007 		dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3008 		ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3009 		dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3010 
3011 		ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3012 		dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
3013 
3014 		ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3015 		dev_info(dev, "\tparity errors %u\n", v16);
3016 
3017 		ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3018 		dev_info(dev, "\tlines per frame %u\n", v16);
3019 
3020 		ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3021 		dev_info(dev, "\tbytes per line %u\n", v16);
3022 
3023 		ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3024 		dev_info(dev, "\tcsi_err_counter %u\n", v);
3025 
3026 		/* Strobe */
3027 
3028 		ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
3029 
3030 		dev_info(dev, "\t%s strobe\n",
3031 			 (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
3032 								  "Manual");
3033 
3034 		if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
3035 			ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
3036 
3037 			dev_info(dev, "\tStrobe range [%d, %d]\n",
3038 				 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
3039 				 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
3040 		}
3041 
3042 		ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
3043 
3044 		dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
3045 
3046 		/* EQ */
3047 
3048 		ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
3049 
3050 		dev_info(dev, "\t%s EQ\n",
3051 			 (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
3052 							    "Adaptive");
3053 
3054 		if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
3055 			ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
3056 
3057 			dev_info(dev, "\tEQ range [%u, %u]\n",
3058 				 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
3059 				 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
3060 		}
3061 
3062 		if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
3063 			dev_info(dev, "\tEQ level %u\n", eq_level);
3064 
3065 		/* GPIOs */
3066 		for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3067 			u8 ctl_reg;
3068 			u8 ctl_shift;
3069 
3070 			ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3071 			ctl_shift = (i % 2) * 4;
3072 
3073 			ub960_rxport_read(priv, nport, ctl_reg, &v);
3074 
3075 			dev_info(dev, "\tGPIO%u: mode %u\n", i,
3076 				 (v >> ctl_shift) & 0xf);
3077 		}
3078 	}
3079 
3080 	v4l2_subdev_unlock_state(state);
3081 
3082 	return 0;
3083 }
3084 
3085 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3086 	.log_status = ub960_log_status,
3087 };
3088 
3089 static const struct v4l2_subdev_internal_ops ub960_internal_ops = {
3090 	.init_state = ub960_init_state,
3091 };
3092 
3093 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3094 	.core = &ub960_subdev_core_ops,
3095 	.pad = &ub960_pad_ops,
3096 };
3097 
3098 static const struct media_entity_operations ub960_entity_ops = {
3099 	.get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3100 	.link_validate = v4l2_subdev_link_validate,
3101 	.has_pad_interdep = v4l2_subdev_has_pad_interdep,
3102 };
3103 
3104 /* -----------------------------------------------------------------------------
3105  * Core
3106  */
3107 
ub960_handle_events(int irq,void * arg)3108 static irqreturn_t ub960_handle_events(int irq, void *arg)
3109 {
3110 	struct ub960_data *priv = arg;
3111 	unsigned int i;
3112 	u8 int_sts;
3113 	u8 fwd_sts;
3114 	int ret;
3115 
3116 	ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3117 	if (ret || !int_sts)
3118 		return IRQ_NONE;
3119 
3120 	dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3121 
3122 	ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3123 	if (ret)
3124 		return IRQ_NONE;
3125 
3126 	dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3127 
3128 	for (i = 0; i < priv->hw_data->num_txports; i++) {
3129 		if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3130 			ub960_csi_handle_events(priv, i);
3131 	}
3132 
3133 	for (i = 0; i < priv->hw_data->num_rxports; i++) {
3134 		if (!priv->rxports[i])
3135 			continue;
3136 
3137 		if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3138 			ub960_rxport_handle_events(priv, i);
3139 	}
3140 
3141 	return IRQ_HANDLED;
3142 }
3143 
ub960_handler_work(struct work_struct * work)3144 static void ub960_handler_work(struct work_struct *work)
3145 {
3146 	struct delayed_work *dwork = to_delayed_work(work);
3147 	struct ub960_data *priv =
3148 		container_of(dwork, struct ub960_data, poll_work);
3149 
3150 	ub960_handle_events(0, priv);
3151 
3152 	schedule_delayed_work(&priv->poll_work,
3153 			      msecs_to_jiffies(UB960_POLL_TIME_MS));
3154 }
3155 
ub960_txport_free_ports(struct ub960_data * priv)3156 static void ub960_txport_free_ports(struct ub960_data *priv)
3157 {
3158 	unsigned int nport;
3159 
3160 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3161 		struct ub960_txport *txport = priv->txports[nport];
3162 
3163 		if (!txport)
3164 			continue;
3165 
3166 		kfree(txport);
3167 		priv->txports[nport] = NULL;
3168 	}
3169 }
3170 
ub960_rxport_free_ports(struct ub960_data * priv)3171 static void ub960_rxport_free_ports(struct ub960_data *priv)
3172 {
3173 	unsigned int nport;
3174 
3175 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3176 		struct ub960_rxport *rxport = priv->rxports[nport];
3177 
3178 		if (!rxport)
3179 			continue;
3180 
3181 		fwnode_handle_put(rxport->source.ep_fwnode);
3182 		fwnode_handle_put(rxport->ser.fwnode);
3183 
3184 		kfree(rxport);
3185 		priv->rxports[nport] = NULL;
3186 	}
3187 }
3188 
3189 static int
ub960_parse_dt_rxport_link_properties(struct ub960_data * priv,struct fwnode_handle * link_fwnode,struct ub960_rxport * rxport)3190 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3191 				      struct fwnode_handle *link_fwnode,
3192 				      struct ub960_rxport *rxport)
3193 {
3194 	struct device *dev = &priv->client->dev;
3195 	unsigned int nport = rxport->nport;
3196 	u32 rx_mode;
3197 	u32 cdr_mode;
3198 	s32 strobe_pos;
3199 	u32 eq_level;
3200 	u32 ser_i2c_alias;
3201 	int ret;
3202 
3203 	cdr_mode = RXPORT_CDR_FPD3;
3204 
3205 	ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3206 	if (ret < 0 && ret != -EINVAL) {
3207 		dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3208 			"ti,cdr-mode", ret);
3209 		return ret;
3210 	}
3211 
3212 	if (cdr_mode > RXPORT_CDR_LAST) {
3213 		dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3214 		return -EINVAL;
3215 	}
3216 
3217 	if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3218 		dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3219 		return -EINVAL;
3220 	}
3221 
3222 	rxport->cdr_mode = cdr_mode;
3223 
3224 	ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3225 	if (ret < 0) {
3226 		dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3227 			"ti,rx-mode", ret);
3228 		return ret;
3229 	}
3230 
3231 	if (rx_mode > RXPORT_MODE_LAST) {
3232 		dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3233 		return -EINVAL;
3234 	}
3235 
3236 	switch (rx_mode) {
3237 	case RXPORT_MODE_RAW12_HF:
3238 	case RXPORT_MODE_RAW12_LF:
3239 		dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3240 			rx_mode);
3241 		return -EINVAL;
3242 	default:
3243 		break;
3244 	}
3245 
3246 	rxport->rx_mode = rx_mode;
3247 
3248 	/* EQ & Strobe related */
3249 
3250 	/* Defaults */
3251 	rxport->eq.manual_eq = false;
3252 	rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3253 	rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3254 
3255 	ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3256 				       &strobe_pos);
3257 	if (ret) {
3258 		if (ret != -EINVAL) {
3259 			dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3260 				"ti,strobe-pos", ret);
3261 			return ret;
3262 		}
3263 	} else {
3264 		if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3265 		    strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3266 			dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3267 				nport, strobe_pos);
3268 			return -EINVAL;
3269 		}
3270 
3271 		/* NOTE: ignored unless global manual strobe pos is also set */
3272 		rxport->eq.strobe_pos = strobe_pos;
3273 		if (!priv->strobe.manual)
3274 			dev_warn(dev,
3275 				 "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3276 				 nport);
3277 	}
3278 
3279 	ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3280 	if (ret) {
3281 		if (ret != -EINVAL) {
3282 			dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3283 				"ti,eq-level", ret);
3284 			return ret;
3285 		}
3286 	} else {
3287 		if (eq_level > UB960_MAX_EQ_LEVEL) {
3288 			dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3289 				nport, eq_level);
3290 			return -EINVAL;
3291 		}
3292 
3293 		rxport->eq.manual_eq = true;
3294 		rxport->eq.manual.eq_level = eq_level;
3295 	}
3296 
3297 	ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3298 				       &ser_i2c_alias);
3299 	if (ret) {
3300 		dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3301 			"i2c-alias", ret);
3302 		return ret;
3303 	}
3304 	rxport->ser.alias = ser_i2c_alias;
3305 
3306 	rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3307 	if (!rxport->ser.fwnode) {
3308 		dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3309 		return -EINVAL;
3310 	}
3311 
3312 	return 0;
3313 }
3314 
ub960_parse_dt_rxport_ep_properties(struct ub960_data * priv,struct fwnode_handle * ep_fwnode,struct ub960_rxport * rxport)3315 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3316 					       struct fwnode_handle *ep_fwnode,
3317 					       struct ub960_rxport *rxport)
3318 {
3319 	struct device *dev = &priv->client->dev;
3320 	struct v4l2_fwnode_endpoint vep = {};
3321 	unsigned int nport = rxport->nport;
3322 	bool hsync_hi;
3323 	bool vsync_hi;
3324 	int ret;
3325 
3326 	rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3327 	if (!rxport->source.ep_fwnode) {
3328 		dev_err(dev, "rx%u: no remote endpoint\n", nport);
3329 		return -ENODEV;
3330 	}
3331 
3332 	/* We currently have properties only for RAW modes */
3333 
3334 	switch (rxport->rx_mode) {
3335 	case RXPORT_MODE_RAW10:
3336 	case RXPORT_MODE_RAW12_HF:
3337 	case RXPORT_MODE_RAW12_LF:
3338 		break;
3339 	default:
3340 		return 0;
3341 	}
3342 
3343 	vep.bus_type = V4L2_MBUS_PARALLEL;
3344 	ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3345 	if (ret) {
3346 		dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3347 		goto err_put_source_ep_fwnode;
3348 	}
3349 
3350 	hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3351 	vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3352 
3353 	/* LineValid and FrameValid are inverse to the h/vsync active */
3354 	rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3355 			    (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3356 
3357 	return 0;
3358 
3359 err_put_source_ep_fwnode:
3360 	fwnode_handle_put(rxport->source.ep_fwnode);
3361 	return ret;
3362 }
3363 
ub960_parse_dt_rxport(struct ub960_data * priv,unsigned int nport,struct fwnode_handle * link_fwnode,struct fwnode_handle * ep_fwnode)3364 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3365 				 struct fwnode_handle *link_fwnode,
3366 				 struct fwnode_handle *ep_fwnode)
3367 {
3368 	static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3369 		"vpoc0", "vpoc1", "vpoc2", "vpoc3"
3370 	};
3371 	struct device *dev = &priv->client->dev;
3372 	struct ub960_rxport *rxport;
3373 	int ret;
3374 
3375 	rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3376 	if (!rxport)
3377 		return -ENOMEM;
3378 
3379 	priv->rxports[nport] = rxport;
3380 
3381 	rxport->nport = nport;
3382 	rxport->priv = priv;
3383 
3384 	ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3385 	if (ret)
3386 		goto err_free_rxport;
3387 
3388 	rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3389 	if (IS_ERR(rxport->vpoc)) {
3390 		ret = PTR_ERR(rxport->vpoc);
3391 		if (ret == -ENODEV) {
3392 			rxport->vpoc = NULL;
3393 		} else {
3394 			dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3395 				nport, ret);
3396 			goto err_put_remote_fwnode;
3397 		}
3398 	}
3399 
3400 	ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3401 	if (ret)
3402 		goto err_put_remote_fwnode;
3403 
3404 	return 0;
3405 
3406 err_put_remote_fwnode:
3407 	fwnode_handle_put(rxport->ser.fwnode);
3408 err_free_rxport:
3409 	priv->rxports[nport] = NULL;
3410 	kfree(rxport);
3411 	return ret;
3412 }
3413 
3414 static struct fwnode_handle *
ub960_fwnode_get_link_by_regs(struct fwnode_handle * links_fwnode,unsigned int nport)3415 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3416 			      unsigned int nport)
3417 {
3418 	struct fwnode_handle *link_fwnode;
3419 	int ret;
3420 
3421 	fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3422 		u32 link_num;
3423 
3424 		if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3425 			continue;
3426 
3427 		ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3428 		if (ret) {
3429 			fwnode_handle_put(link_fwnode);
3430 			return NULL;
3431 		}
3432 
3433 		if (nport == link_num)
3434 			return link_fwnode;
3435 	}
3436 
3437 	return NULL;
3438 }
3439 
ub960_parse_dt_rxports(struct ub960_data * priv)3440 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3441 {
3442 	struct device *dev = &priv->client->dev;
3443 	struct fwnode_handle *links_fwnode;
3444 	unsigned int nport;
3445 	int ret;
3446 
3447 	links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3448 	if (!links_fwnode) {
3449 		dev_err(dev, "'links' node missing\n");
3450 		return -ENODEV;
3451 	}
3452 
3453 	/* Defaults, recommended by TI */
3454 	priv->strobe.min = 2;
3455 	priv->strobe.max = 3;
3456 
3457 	priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3458 
3459 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3460 		struct fwnode_handle *link_fwnode;
3461 		struct fwnode_handle *ep_fwnode;
3462 
3463 		link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3464 		if (!link_fwnode)
3465 			continue;
3466 
3467 		ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3468 							    nport, 0, 0);
3469 		if (!ep_fwnode) {
3470 			fwnode_handle_put(link_fwnode);
3471 			continue;
3472 		}
3473 
3474 		ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3475 					    ep_fwnode);
3476 
3477 		fwnode_handle_put(link_fwnode);
3478 		fwnode_handle_put(ep_fwnode);
3479 
3480 		if (ret) {
3481 			dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3482 			goto err_put_links;
3483 		}
3484 	}
3485 
3486 	fwnode_handle_put(links_fwnode);
3487 
3488 	return 0;
3489 
3490 err_put_links:
3491 	fwnode_handle_put(links_fwnode);
3492 
3493 	return ret;
3494 }
3495 
ub960_parse_dt_txports(struct ub960_data * priv)3496 static int ub960_parse_dt_txports(struct ub960_data *priv)
3497 {
3498 	struct device *dev = &priv->client->dev;
3499 	u32 nport;
3500 	int ret;
3501 
3502 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3503 		unsigned int port = nport + priv->hw_data->num_rxports;
3504 		struct fwnode_handle *ep_fwnode;
3505 
3506 		ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3507 							    port, 0, 0);
3508 		if (!ep_fwnode)
3509 			continue;
3510 
3511 		ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3512 
3513 		fwnode_handle_put(ep_fwnode);
3514 
3515 		if (ret)
3516 			break;
3517 	}
3518 
3519 	return 0;
3520 }
3521 
ub960_parse_dt(struct ub960_data * priv)3522 static int ub960_parse_dt(struct ub960_data *priv)
3523 {
3524 	int ret;
3525 
3526 	ret = ub960_parse_dt_rxports(priv);
3527 	if (ret)
3528 		return ret;
3529 
3530 	ret = ub960_parse_dt_txports(priv);
3531 	if (ret)
3532 		goto err_free_rxports;
3533 
3534 	return 0;
3535 
3536 err_free_rxports:
3537 	ub960_rxport_free_ports(priv);
3538 
3539 	return ret;
3540 }
3541 
ub960_notify_bound(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3542 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3543 			      struct v4l2_subdev *subdev,
3544 			      struct v4l2_async_connection *asd)
3545 {
3546 	struct ub960_data *priv = sd_to_ub960(notifier->sd);
3547 	struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3548 	struct device *dev = &priv->client->dev;
3549 	u8 nport = rxport->nport;
3550 	unsigned int i;
3551 	int ret;
3552 
3553 	ret = media_entity_get_fwnode_pad(&subdev->entity,
3554 					  rxport->source.ep_fwnode,
3555 					  MEDIA_PAD_FL_SOURCE);
3556 	if (ret < 0) {
3557 		dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3558 		return ret;
3559 	}
3560 
3561 	rxport->source.sd = subdev;
3562 	rxport->source.pad = ret;
3563 
3564 	ret = media_create_pad_link(&rxport->source.sd->entity,
3565 				    rxport->source.pad, &priv->sd.entity, nport,
3566 				    MEDIA_LNK_FL_ENABLED |
3567 					    MEDIA_LNK_FL_IMMUTABLE);
3568 	if (ret) {
3569 		dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3570 			rxport->source.sd->name, rxport->source.pad,
3571 			priv->sd.name, nport);
3572 		return ret;
3573 	}
3574 
3575 	for (i = 0; i < priv->hw_data->num_rxports; i++) {
3576 		if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3577 			dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3578 			return 0;
3579 		}
3580 	}
3581 
3582 	return 0;
3583 }
3584 
ub960_notify_unbind(struct v4l2_async_notifier * notifier,struct v4l2_subdev * subdev,struct v4l2_async_connection * asd)3585 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3586 				struct v4l2_subdev *subdev,
3587 				struct v4l2_async_connection *asd)
3588 {
3589 	struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3590 
3591 	rxport->source.sd = NULL;
3592 }
3593 
3594 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3595 	.bound = ub960_notify_bound,
3596 	.unbind = ub960_notify_unbind,
3597 };
3598 
ub960_v4l2_notifier_register(struct ub960_data * priv)3599 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3600 {
3601 	struct device *dev = &priv->client->dev;
3602 	unsigned int i;
3603 	int ret;
3604 
3605 	v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
3606 
3607 	for (i = 0; i < priv->hw_data->num_rxports; i++) {
3608 		struct ub960_rxport *rxport = priv->rxports[i];
3609 		struct ub960_asd *asd;
3610 
3611 		if (!rxport)
3612 			continue;
3613 
3614 		asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3615 					       rxport->source.ep_fwnode,
3616 					       struct ub960_asd);
3617 		if (IS_ERR(asd)) {
3618 			dev_err(dev, "Failed to add subdev for source %u: %pe",
3619 				i, asd);
3620 			v4l2_async_nf_cleanup(&priv->notifier);
3621 			return PTR_ERR(asd);
3622 		}
3623 
3624 		asd->rxport = rxport;
3625 	}
3626 
3627 	priv->notifier.ops = &ub960_notify_ops;
3628 
3629 	ret = v4l2_async_nf_register(&priv->notifier);
3630 	if (ret) {
3631 		dev_err(dev, "Failed to register subdev_notifier");
3632 		v4l2_async_nf_cleanup(&priv->notifier);
3633 		return ret;
3634 	}
3635 
3636 	return 0;
3637 }
3638 
ub960_v4l2_notifier_unregister(struct ub960_data * priv)3639 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3640 {
3641 	v4l2_async_nf_unregister(&priv->notifier);
3642 	v4l2_async_nf_cleanup(&priv->notifier);
3643 }
3644 
ub960_create_subdev(struct ub960_data * priv)3645 static int ub960_create_subdev(struct ub960_data *priv)
3646 {
3647 	struct device *dev = &priv->client->dev;
3648 	unsigned int i;
3649 	int ret;
3650 
3651 	v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3652 	priv->sd.internal_ops = &ub960_internal_ops;
3653 
3654 	v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3655 	priv->sd.ctrl_handler = &priv->ctrl_handler;
3656 
3657 	v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3658 			       ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3659 			       priv->tx_link_freq);
3660 
3661 	if (priv->ctrl_handler.error) {
3662 		ret = priv->ctrl_handler.error;
3663 		goto err_free_ctrl;
3664 	}
3665 
3666 	priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3667 			  V4L2_SUBDEV_FL_STREAMS;
3668 	priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3669 	priv->sd.entity.ops = &ub960_entity_ops;
3670 
3671 	for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3672 		priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3673 					      MEDIA_PAD_FL_SINK :
3674 					      MEDIA_PAD_FL_SOURCE;
3675 	}
3676 
3677 	ret = media_entity_pads_init(&priv->sd.entity,
3678 				     priv->hw_data->num_rxports +
3679 					     priv->hw_data->num_txports,
3680 				     priv->pads);
3681 	if (ret)
3682 		goto err_free_ctrl;
3683 
3684 	priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3685 
3686 	ret = v4l2_subdev_init_finalize(&priv->sd);
3687 	if (ret)
3688 		goto err_entity_cleanup;
3689 
3690 	ret = ub960_v4l2_notifier_register(priv);
3691 	if (ret) {
3692 		dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3693 		goto err_subdev_cleanup;
3694 	}
3695 
3696 	ret = v4l2_async_register_subdev(&priv->sd);
3697 	if (ret) {
3698 		dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3699 		goto err_unreg_notif;
3700 	}
3701 
3702 	return 0;
3703 
3704 err_unreg_notif:
3705 	ub960_v4l2_notifier_unregister(priv);
3706 err_subdev_cleanup:
3707 	v4l2_subdev_cleanup(&priv->sd);
3708 err_entity_cleanup:
3709 	media_entity_cleanup(&priv->sd.entity);
3710 err_free_ctrl:
3711 	v4l2_ctrl_handler_free(&priv->ctrl_handler);
3712 
3713 	return ret;
3714 }
3715 
ub960_destroy_subdev(struct ub960_data * priv)3716 static void ub960_destroy_subdev(struct ub960_data *priv)
3717 {
3718 	ub960_v4l2_notifier_unregister(priv);
3719 	v4l2_async_unregister_subdev(&priv->sd);
3720 
3721 	v4l2_subdev_cleanup(&priv->sd);
3722 
3723 	media_entity_cleanup(&priv->sd.entity);
3724 	v4l2_ctrl_handler_free(&priv->ctrl_handler);
3725 }
3726 
3727 static const struct regmap_config ub960_regmap_config = {
3728 	.name = "ds90ub960",
3729 
3730 	.reg_bits = 8,
3731 	.val_bits = 8,
3732 
3733 	.max_register = 0xff,
3734 
3735 	/*
3736 	 * We do locking in the driver to cover the TX/RX port selection and the
3737 	 * indirect register access.
3738 	 */
3739 	.disable_locking = true,
3740 };
3741 
ub960_reset(struct ub960_data * priv,bool reset_regs)3742 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3743 {
3744 	struct device *dev = &priv->client->dev;
3745 	unsigned int v;
3746 	int ret;
3747 	u8 bit;
3748 
3749 	bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3750 			   UB960_SR_RESET_DIGITAL_RESET0;
3751 
3752 	ub960_write(priv, UB960_SR_RESET, bit);
3753 
3754 	mutex_lock(&priv->reg_lock);
3755 
3756 	ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3757 				       (v & bit) == 0, 2000, 100000);
3758 
3759 	mutex_unlock(&priv->reg_lock);
3760 
3761 	if (ret)
3762 		dev_err(dev, "reset failed: %d\n", ret);
3763 }
3764 
ub960_get_hw_resources(struct ub960_data * priv)3765 static int ub960_get_hw_resources(struct ub960_data *priv)
3766 {
3767 	struct device *dev = &priv->client->dev;
3768 
3769 	priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3770 	if (IS_ERR(priv->regmap))
3771 		return PTR_ERR(priv->regmap);
3772 
3773 	priv->vddio = devm_regulator_get(dev, "vddio");
3774 	if (IS_ERR(priv->vddio))
3775 		return dev_err_probe(dev, PTR_ERR(priv->vddio),
3776 				     "cannot get VDDIO regulator\n");
3777 
3778 	/* get power-down pin from DT */
3779 	priv->pd_gpio =
3780 		devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3781 	if (IS_ERR(priv->pd_gpio))
3782 		return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3783 				     "Cannot get powerdown GPIO\n");
3784 
3785 	priv->refclk = devm_clk_get(dev, "refclk");
3786 	if (IS_ERR(priv->refclk))
3787 		return dev_err_probe(dev, PTR_ERR(priv->refclk),
3788 				     "Cannot get REFCLK\n");
3789 
3790 	return 0;
3791 }
3792 
ub960_enable_core_hw(struct ub960_data * priv)3793 static int ub960_enable_core_hw(struct ub960_data *priv)
3794 {
3795 	struct device *dev = &priv->client->dev;
3796 	u8 rev_mask;
3797 	int ret;
3798 	u8 dev_sts;
3799 	u8 refclk_freq;
3800 
3801 	ret = regulator_enable(priv->vddio);
3802 	if (ret)
3803 		return dev_err_probe(dev, ret,
3804 				     "failed to enable VDDIO regulator\n");
3805 
3806 	ret = clk_prepare_enable(priv->refclk);
3807 	if (ret) {
3808 		dev_err_probe(dev, ret, "Failed to enable refclk\n");
3809 		goto err_disable_vddio;
3810 	}
3811 
3812 	if (priv->pd_gpio) {
3813 		gpiod_set_value_cansleep(priv->pd_gpio, 1);
3814 		/* wait min 2 ms for reset to complete */
3815 		fsleep(2000);
3816 		gpiod_set_value_cansleep(priv->pd_gpio, 0);
3817 		/* wait min 2 ms for power up to finish */
3818 		fsleep(2000);
3819 	}
3820 
3821 	ub960_reset(priv, true);
3822 
3823 	/* Runtime check register accessibility */
3824 	ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3825 	if (ret) {
3826 		dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3827 		goto err_pd_gpio;
3828 	}
3829 
3830 	dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3831 		rev_mask);
3832 
3833 	ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3834 	if (ret)
3835 		goto err_pd_gpio;
3836 
3837 	ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3838 	if (ret)
3839 		goto err_pd_gpio;
3840 
3841 	dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3842 		!!(dev_sts & BIT(4)), refclk_freq,
3843 		clk_get_rate(priv->refclk) / 1000000);
3844 
3845 	/* Disable all RX ports by default */
3846 	ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3847 	if (ret)
3848 		goto err_pd_gpio;
3849 
3850 	/* release GPIO lock */
3851 	if (priv->hw_data->is_ub9702) {
3852 		ret = ub960_update_bits(priv, UB960_SR_RESET,
3853 					UB960_SR_RESET_GPIO_LOCK_RELEASE,
3854 					UB960_SR_RESET_GPIO_LOCK_RELEASE);
3855 		if (ret)
3856 			goto err_pd_gpio;
3857 	}
3858 
3859 	return 0;
3860 
3861 err_pd_gpio:
3862 	gpiod_set_value_cansleep(priv->pd_gpio, 1);
3863 	clk_disable_unprepare(priv->refclk);
3864 err_disable_vddio:
3865 	regulator_disable(priv->vddio);
3866 
3867 	return ret;
3868 }
3869 
ub960_disable_core_hw(struct ub960_data * priv)3870 static void ub960_disable_core_hw(struct ub960_data *priv)
3871 {
3872 	gpiod_set_value_cansleep(priv->pd_gpio, 1);
3873 	clk_disable_unprepare(priv->refclk);
3874 	regulator_disable(priv->vddio);
3875 }
3876 
ub960_probe(struct i2c_client * client)3877 static int ub960_probe(struct i2c_client *client)
3878 {
3879 	struct device *dev = &client->dev;
3880 	struct ub960_data *priv;
3881 	unsigned int port_lock_mask;
3882 	unsigned int port_mask;
3883 	unsigned int nport;
3884 	int ret;
3885 
3886 	priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3887 	if (!priv)
3888 		return -ENOMEM;
3889 
3890 	priv->client = client;
3891 
3892 	priv->hw_data = device_get_match_data(dev);
3893 
3894 	mutex_init(&priv->reg_lock);
3895 
3896 	INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3897 
3898 	/*
3899 	 * Initialize these to invalid values so that the first reg writes will
3900 	 * configure the target.
3901 	 */
3902 	priv->reg_current.indirect_target = 0xff;
3903 	priv->reg_current.rxport = 0xff;
3904 	priv->reg_current.txport = 0xff;
3905 
3906 	ret = ub960_get_hw_resources(priv);
3907 	if (ret)
3908 		goto err_mutex_destroy;
3909 
3910 	ret = ub960_enable_core_hw(priv);
3911 	if (ret)
3912 		goto err_mutex_destroy;
3913 
3914 	ret = ub960_parse_dt(priv);
3915 	if (ret)
3916 		goto err_disable_core_hw;
3917 
3918 	ret = ub960_init_tx_ports(priv);
3919 	if (ret)
3920 		goto err_free_ports;
3921 
3922 	ret = ub960_rxport_enable_vpocs(priv);
3923 	if (ret)
3924 		goto err_free_ports;
3925 
3926 	ret = ub960_init_rx_ports(priv);
3927 	if (ret)
3928 		goto err_disable_vpocs;
3929 
3930 	ub960_reset(priv, false);
3931 
3932 	port_mask = 0;
3933 
3934 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3935 		struct ub960_rxport *rxport = priv->rxports[nport];
3936 
3937 		if (!rxport)
3938 			continue;
3939 
3940 		port_mask |= BIT(nport);
3941 	}
3942 
3943 	ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
3944 	if (ret)
3945 		goto err_disable_vpocs;
3946 
3947 	if (port_mask != port_lock_mask) {
3948 		ret = -EIO;
3949 		dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
3950 		goto err_disable_vpocs;
3951 	}
3952 
3953 	/*
3954 	 * Clear any errors caused by switching the RX port settings while
3955 	 * probing.
3956 	 */
3957 	ub960_clear_rx_errors(priv);
3958 
3959 	ret = ub960_init_atr(priv);
3960 	if (ret)
3961 		goto err_disable_vpocs;
3962 
3963 	ret = ub960_rxport_add_serializers(priv);
3964 	if (ret)
3965 		goto err_uninit_atr;
3966 
3967 	ret = ub960_create_subdev(priv);
3968 	if (ret)
3969 		goto err_free_sers;
3970 
3971 	if (client->irq)
3972 		dev_warn(dev, "irq support not implemented, using polling\n");
3973 
3974 	schedule_delayed_work(&priv->poll_work,
3975 			      msecs_to_jiffies(UB960_POLL_TIME_MS));
3976 
3977 	return 0;
3978 
3979 err_free_sers:
3980 	ub960_rxport_remove_serializers(priv);
3981 err_uninit_atr:
3982 	ub960_uninit_atr(priv);
3983 err_disable_vpocs:
3984 	ub960_rxport_disable_vpocs(priv);
3985 err_free_ports:
3986 	ub960_rxport_free_ports(priv);
3987 	ub960_txport_free_ports(priv);
3988 err_disable_core_hw:
3989 	ub960_disable_core_hw(priv);
3990 err_mutex_destroy:
3991 	mutex_destroy(&priv->reg_lock);
3992 	return ret;
3993 }
3994 
ub960_remove(struct i2c_client * client)3995 static void ub960_remove(struct i2c_client *client)
3996 {
3997 	struct v4l2_subdev *sd = i2c_get_clientdata(client);
3998 	struct ub960_data *priv = sd_to_ub960(sd);
3999 
4000 	cancel_delayed_work_sync(&priv->poll_work);
4001 
4002 	ub960_destroy_subdev(priv);
4003 	ub960_rxport_remove_serializers(priv);
4004 	ub960_uninit_atr(priv);
4005 	ub960_rxport_disable_vpocs(priv);
4006 	ub960_rxport_free_ports(priv);
4007 	ub960_txport_free_ports(priv);
4008 	ub960_disable_core_hw(priv);
4009 	mutex_destroy(&priv->reg_lock);
4010 }
4011 
4012 static const struct ub960_hw_data ds90ub960_hw = {
4013 	.model = "ub960",
4014 	.num_rxports = 4,
4015 	.num_txports = 2,
4016 };
4017 
4018 static const struct ub960_hw_data ds90ub9702_hw = {
4019 	.model = "ub9702",
4020 	.num_rxports = 4,
4021 	.num_txports = 2,
4022 	.is_ub9702 = true,
4023 	.is_fpdlink4 = true,
4024 };
4025 
4026 static const struct i2c_device_id ub960_id[] = {
4027 	{ "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4028 	{ "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4029 	{}
4030 };
4031 MODULE_DEVICE_TABLE(i2c, ub960_id);
4032 
4033 static const struct of_device_id ub960_dt_ids[] = {
4034 	{ .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4035 	{ .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4036 	{}
4037 };
4038 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4039 
4040 static struct i2c_driver ds90ub960_driver = {
4041 	.probe		= ub960_probe,
4042 	.remove		= ub960_remove,
4043 	.id_table	= ub960_id,
4044 	.driver = {
4045 		.name	= "ds90ub960",
4046 		.of_match_table = ub960_dt_ids,
4047 	},
4048 };
4049 module_i2c_driver(ds90ub960_driver);
4050 
4051 MODULE_LICENSE("GPL");
4052 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4053 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4054 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4055 MODULE_IMPORT_NS("I2C_ATR");
4056