GNU Linux-libre 6.7.9-gnu
[releases.git] / drivers / media / i2c / ds90ub960.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
4  *
5  * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6  * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
7  */
8
9 /*
10  * (Possible) TODOs:
11  *
12  * - PM for serializer and remote peripherals. We need to manage:
13  *   - VPOC
14  *     - Power domain? Regulator? Somehow any remote device should be able to
15  *       cause the VPOC to be turned on.
16  *   - Link between the deserializer and the serializer
17  *     - Related to VPOC management. We probably always want to turn on the VPOC
18  *       and then enable the link.
19  *   - Serializer's services: i2c, gpios, power
20  *     - The serializer needs to resume before the remote peripherals can
21  *       e.g. use the i2c.
22  *     - How to handle gpios? Reserving a gpio essentially keeps the provider
23  *       (serializer) always powered on.
24  * - Do we need a new bus for the FPD-Link? At the moment the serializers
25  *   are children of the same i2c-adapter where the deserializer resides.
26  * - i2c-atr could be made embeddable instead of allocatable.
27  */
28
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/workqueue.h>
47
48 #include <media/i2c/ds90ub9xx.h>
49 #include <media/mipi-csi2.h>
50 #include <media/v4l2-ctrls.h>
51 #include <media/v4l2-event.h>
52 #include <media/v4l2-fwnode.h>
53 #include <media/v4l2-subdev.h>
54
55 #define MHZ(v) ((u32)((v) * 1000000U))
56
57 #define UB960_POLL_TIME_MS      500
58
59 #define UB960_MAX_RX_NPORTS     4
60 #define UB960_MAX_TX_NPORTS     2
61 #define UB960_MAX_NPORTS        (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
62
63 #define UB960_MAX_PORT_ALIASES  8
64
65 #define UB960_NUM_BC_GPIOS              4
66
67 /*
68  * Register map
69  *
70  * 0x00-0x32   Shared (UB960_SR)
71  * 0x33-0x3a   CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
72  * 0x4c        Shared (UB960_SR)
73  * 0x4d-0x7f   FPD-Link RX, per-port paged (UB960_RR)
74  * 0xb0-0xbf   Shared (UB960_SR)
75  * 0xd0-0xdf   FPD-Link RX, per-port paged (UB960_RR)
76  * 0xf0-0xf5   Shared (UB960_SR)
77  * 0xf8-0xfb   Shared (UB960_SR)
78  * All others  Reserved
79  *
80  * Register prefixes:
81  * UB960_SR_* = Shared register
82  * UB960_RR_* = FPD-Link RX, per-port paged register
83  * UB960_TR_* = CSI-2 TX, per-port paged register
84  * UB960_XR_* = Reserved register
85  * UB960_IR_* = Indirect register
86  */
87
88 #define UB960_SR_I2C_DEV_ID                     0x00
89 #define UB960_SR_RESET                          0x01
90 #define UB960_SR_RESET_DIGITAL_RESET1           BIT(1)
91 #define UB960_SR_RESET_DIGITAL_RESET0           BIT(0)
92 #define UB960_SR_RESET_GPIO_LOCK_RELEASE        BIT(5)
93
94 #define UB960_SR_GEN_CONFIG                     0x02
95 #define UB960_SR_REV_MASK                       0x03
96 #define UB960_SR_DEVICE_STS                     0x04
97 #define UB960_SR_PAR_ERR_THOLD_HI               0x05
98 #define UB960_SR_PAR_ERR_THOLD_LO               0x06
99 #define UB960_SR_BCC_WDOG_CTL                   0x07
100 #define UB960_SR_I2C_CTL1                       0x08
101 #define UB960_SR_I2C_CTL2                       0x09
102 #define UB960_SR_SCL_HIGH_TIME                  0x0a
103 #define UB960_SR_SCL_LOW_TIME                   0x0b
104 #define UB960_SR_RX_PORT_CTL                    0x0c
105 #define UB960_SR_IO_CTL                         0x0d
106 #define UB960_SR_GPIO_PIN_STS                   0x0e
107 #define UB960_SR_GPIO_INPUT_CTL                 0x0f
108 #define UB960_SR_GPIO_PIN_CTL(n)                (0x10 + (n)) /* n < UB960_NUM_GPIOS */
109 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL              5
110 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT        2
111 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN               BIT(0)
112
113 #define UB960_SR_FS_CTL                         0x18
114 #define UB960_SR_FS_HIGH_TIME_1                 0x19
115 #define UB960_SR_FS_HIGH_TIME_0                 0x1a
116 #define UB960_SR_FS_LOW_TIME_1                  0x1b
117 #define UB960_SR_FS_LOW_TIME_0                  0x1c
118 #define UB960_SR_MAX_FRM_HI                     0x1d
119 #define UB960_SR_MAX_FRM_LO                     0x1e
120 #define UB960_SR_CSI_PLL_CTL                    0x1f
121
122 #define UB960_SR_FWD_CTL1                       0x20
123 #define UB960_SR_FWD_CTL1_PORT_DIS(n)           BIT((n) + 4)
124
125 #define UB960_SR_FWD_CTL2                       0x21
126 #define UB960_SR_FWD_STS                        0x22
127
128 #define UB960_SR_INTERRUPT_CTL                  0x23
129 #define UB960_SR_INTERRUPT_CTL_INT_EN           BIT(7)
130 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0       BIT(4)
131 #define UB960_SR_INTERRUPT_CTL_IE_RX(n)         BIT((n)) /* rxport[n] IRQ */
132
133 #define UB960_SR_INTERRUPT_STS                  0x24
134 #define UB960_SR_INTERRUPT_STS_INT              BIT(7)
135 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n)     BIT(4 + (n)) /* txport[n] IRQ */
136 #define UB960_SR_INTERRUPT_STS_IS_RX(n)         BIT((n)) /* rxport[n] IRQ */
137
138 #define UB960_SR_TS_CONFIG                      0x25
139 #define UB960_SR_TS_CONTROL                     0x26
140 #define UB960_SR_TS_LINE_HI                     0x27
141 #define UB960_SR_TS_LINE_LO                     0x28
142 #define UB960_SR_TS_STATUS                      0x29
143 #define UB960_SR_TIMESTAMP_P0_HI                0x2a
144 #define UB960_SR_TIMESTAMP_P0_LO                0x2b
145 #define UB960_SR_TIMESTAMP_P1_HI                0x2c
146 #define UB960_SR_TIMESTAMP_P1_LO                0x2d
147
148 #define UB960_SR_CSI_PORT_SEL                   0x32
149
150 #define UB960_TR_CSI_CTL                        0x33
151 #define UB960_TR_CSI_CTL_CSI_CAL_EN             BIT(6)
152 #define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK        BIT(1)
153 #define UB960_TR_CSI_CTL_CSI_ENABLE             BIT(0)
154
155 #define UB960_TR_CSI_CTL2                       0x34
156 #define UB960_TR_CSI_STS                        0x35
157 #define UB960_TR_CSI_TX_ICR                     0x36
158
159 #define UB960_TR_CSI_TX_ISR                     0x37
160 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR   BIT(3)
161 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR   BIT(1)
162
163 #define UB960_TR_CSI_TEST_CTL                   0x38
164 #define UB960_TR_CSI_TEST_PATT_HI               0x39
165 #define UB960_TR_CSI_TEST_PATT_LO               0x3a
166
167 #define UB960_XR_SFILTER_CFG                    0x41
168 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT  4
169 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT  0
170
171 #define UB960_XR_AEQ_CTL1                       0x42
172 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK   BIT(6)
173 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING  BIT(5)
174 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY    BIT(4)
175 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK        \
176         (UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK |  \
177          UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
178          UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
179 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN        BIT(0)
180
181 #define UB960_XR_AEQ_ERR_THOLD                  0x43
182
183 #define UB960_RR_BCC_ERR_CTL                    0x46
184 #define UB960_RR_BCC_STATUS                     0x47
185 #define UB960_RR_BCC_STATUS_SEQ_ERROR           BIT(5)
186 #define UB960_RR_BCC_STATUS_MASTER_ERR          BIT(4)
187 #define UB960_RR_BCC_STATUS_MASTER_TO           BIT(3)
188 #define UB960_RR_BCC_STATUS_SLAVE_ERR           BIT(2)
189 #define UB960_RR_BCC_STATUS_SLAVE_TO            BIT(1)
190 #define UB960_RR_BCC_STATUS_RESP_ERR            BIT(0)
191 #define UB960_RR_BCC_STATUS_ERROR_MASK                                    \
192         (UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
193          UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR |  \
194          UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
195
196 #define UB960_RR_FPD3_CAP                       0x4a
197 #define UB960_RR_RAW_EMBED_DTYPE                0x4b
198 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT    6
199
200 #define UB960_SR_FPD3_PORT_SEL                  0x4c
201
202 #define UB960_RR_RX_PORT_STS1                   0x4d
203 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR     BIT(5)
204 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG      BIT(4)
205 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR     BIT(3)
206 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR      BIT(2)
207 #define UB960_RR_RX_PORT_STS1_PORT_PASS         BIT(1)
208 #define UB960_RR_RX_PORT_STS1_LOCK_STS          BIT(0)
209 #define UB960_RR_RX_PORT_STS1_ERROR_MASK       \
210         (UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
211          UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
212          UB960_RR_RX_PORT_STS1_PARITY_ERROR)
213
214 #define UB960_RR_RX_PORT_STS2                   0x4e
215 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
216 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG      BIT(6)
217 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
218 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR      BIT(4)
219 #define UB960_RR_RX_PORT_STS2_CSI_ERROR         BIT(3)
220 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE       BIT(2)
221 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT       BIT(1)
222 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG      BIT(0)
223 #define UB960_RR_RX_PORT_STS2_ERROR_MASK       \
224         UB960_RR_RX_PORT_STS2_BUFFER_ERROR
225
226 #define UB960_RR_RX_FREQ_HIGH                   0x4f
227 #define UB960_RR_RX_FREQ_LOW                    0x50
228 #define UB960_RR_SENSOR_STS_0                   0x51
229 #define UB960_RR_SENSOR_STS_1                   0x52
230 #define UB960_RR_SENSOR_STS_2                   0x53
231 #define UB960_RR_SENSOR_STS_3                   0x54
232 #define UB960_RR_RX_PAR_ERR_HI                  0x55
233 #define UB960_RR_RX_PAR_ERR_LO                  0x56
234 #define UB960_RR_BIST_ERR_COUNT                 0x57
235
236 #define UB960_RR_BCC_CONFIG                     0x58
237 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH    BIT(6)
238 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK    GENMASK(2, 0)
239
240 #define UB960_RR_DATAPATH_CTL1                  0x59
241 #define UB960_RR_DATAPATH_CTL2                  0x5a
242 #define UB960_RR_SER_ID                         0x5b
243 #define UB960_RR_SER_ALIAS_ID                   0x5c
244
245 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
246 #define UB960_RR_SLAVE_ID(n)                    (0x5d + (n))
247 #define UB960_RR_SLAVE_ALIAS(n)                 (0x65 + (n))
248
249 #define UB960_RR_PORT_CONFIG                    0x6d
250 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK     GENMASK(1, 0)
251
252 #define UB960_RR_BC_GPIO_CTL(n)                 (0x6e + (n)) /* n < 2 */
253 #define UB960_RR_RAW10_ID                       0x70
254 #define UB960_RR_RAW10_ID_VC_SHIFT              6
255 #define UB960_RR_RAW10_ID_DT_SHIFT              0
256
257 #define UB960_RR_RAW12_ID                       0x71
258 #define UB960_RR_CSI_VC_MAP                     0x72
259 #define UB960_RR_CSI_VC_MAP_SHIFT(x)            ((x) * 2)
260
261 #define UB960_RR_LINE_COUNT_HI                  0x73
262 #define UB960_RR_LINE_COUNT_LO                  0x74
263 #define UB960_RR_LINE_LEN_1                     0x75
264 #define UB960_RR_LINE_LEN_0                     0x76
265 #define UB960_RR_FREQ_DET_CTL                   0x77
266 #define UB960_RR_MAILBOX_1                      0x78
267 #define UB960_RR_MAILBOX_2                      0x79
268
269 #define UB960_RR_CSI_RX_STS                     0x7a
270 #define UB960_RR_CSI_RX_STS_LENGTH_ERR          BIT(3)
271 #define UB960_RR_CSI_RX_STS_CKSUM_ERR           BIT(2)
272 #define UB960_RR_CSI_RX_STS_ECC2_ERR            BIT(1)
273 #define UB960_RR_CSI_RX_STS_ECC1_ERR            BIT(0)
274 #define UB960_RR_CSI_RX_STS_ERROR_MASK                                    \
275         (UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
276          UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
277
278 #define UB960_RR_CSI_ERR_COUNTER                0x7b
279 #define UB960_RR_PORT_CONFIG2                   0x7c
280 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
281 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
282
283 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW        BIT(1)
284 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW        BIT(0)
285
286 #define UB960_RR_PORT_PASS_CTL                  0x7d
287 #define UB960_RR_SEN_INT_RISE_CTL               0x7e
288 #define UB960_RR_SEN_INT_FALL_CTL               0x7f
289
290 #define UB960_SR_CSI_FRAME_COUNT_HI(n)          (0x90 + 8 * (n))
291 #define UB960_SR_CSI_FRAME_COUNT_LO(n)          (0x91 + 8 * (n))
292 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n)      (0x92 + 8 * (n))
293 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n)      (0x93 + 8 * (n))
294 #define UB960_SR_CSI_LINE_COUNT_HI(n)           (0x94 + 8 * (n))
295 #define UB960_SR_CSI_LINE_COUNT_LO(n)           (0x95 + 8 * (n))
296 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n)       (0x96 + 8 * (n))
297 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n)       (0x97 + 8 * (n))
298
299 #define UB960_XR_REFCLK_FREQ                    0xa5    /* UB960 */
300
301 #define UB960_RR_VC_ID_MAP(x)                   (0xa0 + (x)) /* UB9702 */
302
303 #define UB960_SR_IND_ACC_CTL                    0xb0
304 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC        BIT(1)
305
306 #define UB960_SR_IND_ACC_ADDR                   0xb1
307 #define UB960_SR_IND_ACC_DATA                   0xb2
308 #define UB960_SR_BIST_CONTROL                   0xb3
309 #define UB960_SR_MODE_IDX_STS                   0xb8
310 #define UB960_SR_LINK_ERROR_COUNT               0xb9
311 #define UB960_SR_FPD3_ENC_CTL                   0xba
312 #define UB960_SR_FV_MIN_TIME                    0xbc
313 #define UB960_SR_GPIO_PD_CTL                    0xbe
314
315 #define UB960_SR_FPD_RATE_CFG                   0xc2    /* UB9702 */
316 #define UB960_SR_CSI_PLL_DIV                    0xc9    /* UB9702 */
317
318 #define UB960_RR_PORT_DEBUG                     0xd0
319 #define UB960_RR_AEQ_CTL2                       0xd2
320 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR         BIT(2)
321
322 #define UB960_RR_AEQ_STATUS                     0xd3
323 #define UB960_RR_AEQ_STATUS_STATUS_2            GENMASK(5, 3)
324 #define UB960_RR_AEQ_STATUS_STATUS_1            GENMASK(2, 0)
325
326 #define UB960_RR_AEQ_BYPASS                     0xd4
327 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT       5
328 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK        GENMASK(7, 5)
329 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT       1
330 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK        GENMASK(3, 1)
331 #define UB960_RR_AEQ_BYPASS_ENABLE                      BIT(0)
332
333 #define UB960_RR_AEQ_MIN_MAX                    0xd5
334 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT      4
335 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT    0
336
337 #define UB960_RR_SFILTER_STS_0                  0xd6
338 #define UB960_RR_SFILTER_STS_1                  0xd7
339 #define UB960_RR_PORT_ICR_HI                    0xd8
340 #define UB960_RR_PORT_ICR_LO                    0xd9
341 #define UB960_RR_PORT_ISR_HI                    0xda
342 #define UB960_RR_PORT_ISR_LO                    0xdb
343 #define UB960_RR_FC_GPIO_STS                    0xdc
344 #define UB960_RR_FC_GPIO_ICR                    0xdd
345 #define UB960_RR_SEN_INT_RISE_STS               0xde
346 #define UB960_RR_SEN_INT_FALL_STS               0xdf
347
348 #define UB960_RR_CHANNEL_MODE                   0xe4    /* UB9702 */
349
350 #define UB960_SR_FPD3_RX_ID(n)                  (0xf0 + (n))
351 #define UB960_SR_FPD3_RX_ID_LEN                 6
352
353 #define UB960_SR_I2C_RX_ID(n)                   (0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
354
355 /* Indirect register blocks */
356 #define UB960_IND_TARGET_PAT_GEN                0x00
357 #define UB960_IND_TARGET_RX_ANA(n)              (0x01 + (n))
358 #define UB960_IND_TARGET_CSI_CSIPLL_REG_1       0x92    /* UB9702 */
359 #define UB960_IND_TARGET_CSI_ANA                0x07
360
361 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
362
363 #define UB960_IR_PGEN_CTL                       0x01
364 #define UB960_IR_PGEN_CTL_PGEN_ENABLE           BIT(0)
365
366 #define UB960_IR_PGEN_CFG                       0x02
367 #define UB960_IR_PGEN_CSI_DI                    0x03
368 #define UB960_IR_PGEN_LINE_SIZE1                0x04
369 #define UB960_IR_PGEN_LINE_SIZE0                0x05
370 #define UB960_IR_PGEN_BAR_SIZE1                 0x06
371 #define UB960_IR_PGEN_BAR_SIZE0                 0x07
372 #define UB960_IR_PGEN_ACT_LPF1                  0x08
373 #define UB960_IR_PGEN_ACT_LPF0                  0x09
374 #define UB960_IR_PGEN_TOT_LPF1                  0x0a
375 #define UB960_IR_PGEN_TOT_LPF0                  0x0b
376 #define UB960_IR_PGEN_LINE_PD1                  0x0c
377 #define UB960_IR_PGEN_LINE_PD0                  0x0d
378 #define UB960_IR_PGEN_VBP                       0x0e
379 #define UB960_IR_PGEN_VFP                       0x0f
380 #define UB960_IR_PGEN_COLOR(n)                  (0x10 + (n)) /* n < 15 */
381
382 #define UB960_IR_RX_ANA_STROBE_SET_CLK          0x08
383 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY   BIT(3)
384 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK       GENMASK(2, 0)
385
386 #define UB960_IR_RX_ANA_STROBE_SET_DATA         0x09
387 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY  BIT(3)
388 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK      GENMASK(2, 0)
389
390 /* EQ related */
391
392 #define UB960_MIN_AEQ_STROBE_POS -7
393 #define UB960_MAX_AEQ_STROBE_POS  7
394
395 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
396
397 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
398 #define UB960_MAX_MANUAL_STROBE_POS  (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
399 #define UB960_NUM_MANUAL_STROBE_POS  (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
400
401 #define UB960_MIN_EQ_LEVEL  0
402 #define UB960_MAX_EQ_LEVEL  14
403 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
404
405 struct ub960_hw_data {
406         const char *model;
407         u8 num_rxports;
408         u8 num_txports;
409         bool is_ub9702;
410         bool is_fpdlink4;
411 };
412
413 enum ub960_rxport_mode {
414         RXPORT_MODE_RAW10 = 0,
415         RXPORT_MODE_RAW12_HF = 1,
416         RXPORT_MODE_RAW12_LF = 2,
417         RXPORT_MODE_CSI2_SYNC = 3,
418         RXPORT_MODE_CSI2_NONSYNC = 4,
419         RXPORT_MODE_LAST = RXPORT_MODE_CSI2_NONSYNC,
420 };
421
422 enum ub960_rxport_cdr {
423         RXPORT_CDR_FPD3 = 0,
424         RXPORT_CDR_FPD4 = 1,
425         RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
426 };
427
428 struct ub960_rxport {
429         struct ub960_data      *priv;
430         u8                      nport;  /* RX port number, and index in priv->rxport[] */
431
432         struct {
433                 struct v4l2_subdev *sd;
434                 u16 pad;
435                 struct fwnode_handle *ep_fwnode;
436         } source;
437
438         /* Serializer */
439         struct {
440                 struct fwnode_handle *fwnode;
441                 struct i2c_client *client;
442                 unsigned short alias; /* I2C alias (lower 7 bits) */
443                 struct ds90ub9xx_platform_data pdata;
444         } ser;
445
446         enum ub960_rxport_mode  rx_mode;
447         enum ub960_rxport_cdr   cdr_mode;
448
449         u8                      lv_fv_pol;      /* LV and FV polarities */
450
451         struct regulator        *vpoc;
452
453         /* EQ settings */
454         struct {
455                 bool manual_eq;
456
457                 s8 strobe_pos;
458
459                 union {
460                         struct {
461                                 u8 eq_level_min;
462                                 u8 eq_level_max;
463                         } aeq;
464
465                         struct {
466                                 u8 eq_level;
467                         } manual;
468                 };
469         } eq;
470
471         const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
472 };
473
474 struct ub960_asd {
475         struct v4l2_async_connection base;
476         struct ub960_rxport *rxport;
477 };
478
479 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
480 {
481         return container_of(asd, struct ub960_asd, base);
482 }
483
484 struct ub960_txport {
485         struct ub960_data      *priv;
486         u8                      nport;  /* TX port number, and index in priv->txport[] */
487
488         u32 num_data_lanes;
489         bool non_continous_clk;
490 };
491
492 struct ub960_data {
493         const struct ub960_hw_data      *hw_data;
494         struct i2c_client       *client; /* for shared local registers */
495         struct regmap           *regmap;
496
497         /* lock for register access */
498         struct mutex            reg_lock;
499
500         struct clk              *refclk;
501
502         struct regulator        *vddio;
503
504         struct gpio_desc        *pd_gpio;
505         struct delayed_work     poll_work;
506         struct ub960_rxport     *rxports[UB960_MAX_RX_NPORTS];
507         struct ub960_txport     *txports[UB960_MAX_TX_NPORTS];
508
509         struct v4l2_subdev      sd;
510         struct media_pad        pads[UB960_MAX_NPORTS];
511
512         struct v4l2_ctrl_handler   ctrl_handler;
513         struct v4l2_async_notifier notifier;
514
515         u32 tx_data_rate;               /* Nominal data rate (Gb/s) */
516         s64 tx_link_freq[1];
517
518         struct i2c_atr *atr;
519
520         struct {
521                 u8 rxport;
522                 u8 txport;
523                 u8 indirect_target;
524         } reg_current;
525
526         bool streaming;
527
528         u8 stored_fwd_ctl;
529
530         u64 stream_enable_mask[UB960_MAX_NPORTS];
531
532         /* These are common to all ports */
533         struct {
534                 bool manual;
535
536                 s8 min;
537                 s8 max;
538         } strobe;
539 };
540
541 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
542 {
543         return container_of(sd, struct ub960_data, sd);
544 }
545
546 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
547 {
548         return pad < priv->hw_data->num_rxports;
549 }
550
551 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
552 {
553         return pad >= priv->hw_data->num_rxports;
554 }
555
556 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
557 {
558         if (ub960_pad_is_sink(priv, pad))
559                 return pad;
560         else
561                 return pad - priv->hw_data->num_rxports;
562 }
563
564 struct ub960_format_info {
565         u32 code;
566         u32 bpp;
567         u8 datatype;
568         bool meta;
569 };
570
571 static const struct ub960_format_info ub960_formats[] = {
572         { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
573         { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
574         { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
575         { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
576
577         { .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
578         { .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
579         { .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
580         { .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
581 };
582
583 static const struct ub960_format_info *ub960_find_format(u32 code)
584 {
585         unsigned int i;
586
587         for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
588                 if (ub960_formats[i].code == code)
589                         return &ub960_formats[i];
590         }
591
592         return NULL;
593 }
594
595 /* -----------------------------------------------------------------------------
596  * Basic device access
597  */
598
599 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
600 {
601         struct device *dev = &priv->client->dev;
602         unsigned int v;
603         int ret;
604
605         mutex_lock(&priv->reg_lock);
606
607         ret = regmap_read(priv->regmap, reg, &v);
608         if (ret) {
609                 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
610                         __func__, reg, ret);
611                 goto out_unlock;
612         }
613
614         *val = v;
615
616 out_unlock:
617         mutex_unlock(&priv->reg_lock);
618
619         return ret;
620 }
621
622 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
623 {
624         struct device *dev = &priv->client->dev;
625         int ret;
626
627         mutex_lock(&priv->reg_lock);
628
629         ret = regmap_write(priv->regmap, reg, val);
630         if (ret)
631                 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
632                         __func__, reg, ret);
633
634         mutex_unlock(&priv->reg_lock);
635
636         return ret;
637 }
638
639 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
640 {
641         struct device *dev = &priv->client->dev;
642         int ret;
643
644         mutex_lock(&priv->reg_lock);
645
646         ret = regmap_update_bits(priv->regmap, reg, mask, val);
647         if (ret)
648                 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
649                         __func__, reg, ret);
650
651         mutex_unlock(&priv->reg_lock);
652
653         return ret;
654 }
655
656 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
657 {
658         struct device *dev = &priv->client->dev;
659         __be16 __v;
660         int ret;
661
662         mutex_lock(&priv->reg_lock);
663
664         ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
665         if (ret) {
666                 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
667                         __func__, reg, ret);
668                 goto out_unlock;
669         }
670
671         *val = be16_to_cpu(__v);
672
673 out_unlock:
674         mutex_unlock(&priv->reg_lock);
675
676         return ret;
677 }
678
679 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
680 {
681         struct device *dev = &priv->client->dev;
682         int ret;
683
684         lockdep_assert_held(&priv->reg_lock);
685
686         if (priv->reg_current.rxport == nport)
687                 return 0;
688
689         ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
690                            (nport << 4) | BIT(nport));
691         if (ret) {
692                 dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
693                         nport, ret);
694                 return ret;
695         }
696
697         priv->reg_current.rxport = nport;
698
699         return 0;
700 }
701
702 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
703 {
704         struct device *dev = &priv->client->dev;
705         unsigned int v;
706         int ret;
707
708         mutex_lock(&priv->reg_lock);
709
710         ret = ub960_rxport_select(priv, nport);
711         if (ret)
712                 goto out_unlock;
713
714         ret = regmap_read(priv->regmap, reg, &v);
715         if (ret) {
716                 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
717                         __func__, reg, ret);
718                 goto out_unlock;
719         }
720
721         *val = v;
722
723 out_unlock:
724         mutex_unlock(&priv->reg_lock);
725
726         return ret;
727 }
728
729 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
730 {
731         struct device *dev = &priv->client->dev;
732         int ret;
733
734         mutex_lock(&priv->reg_lock);
735
736         ret = ub960_rxport_select(priv, nport);
737         if (ret)
738                 goto out_unlock;
739
740         ret = regmap_write(priv->regmap, reg, val);
741         if (ret)
742                 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
743                         __func__, reg, ret);
744
745 out_unlock:
746         mutex_unlock(&priv->reg_lock);
747
748         return ret;
749 }
750
751 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
752                                     u8 mask, u8 val)
753 {
754         struct device *dev = &priv->client->dev;
755         int ret;
756
757         mutex_lock(&priv->reg_lock);
758
759         ret = ub960_rxport_select(priv, nport);
760         if (ret)
761                 goto out_unlock;
762
763         ret = regmap_update_bits(priv->regmap, reg, mask, val);
764         if (ret)
765                 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
766                         __func__, reg, ret);
767
768 out_unlock:
769         mutex_unlock(&priv->reg_lock);
770
771         return ret;
772 }
773
774 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
775                                u16 *val)
776 {
777         struct device *dev = &priv->client->dev;
778         __be16 __v;
779         int ret;
780
781         mutex_lock(&priv->reg_lock);
782
783         ret = ub960_rxport_select(priv, nport);
784         if (ret)
785                 goto out_unlock;
786
787         ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
788         if (ret) {
789                 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
790                         __func__, reg, ret);
791                 goto out_unlock;
792         }
793
794         *val = be16_to_cpu(__v);
795
796 out_unlock:
797         mutex_unlock(&priv->reg_lock);
798
799         return ret;
800 }
801
802 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
803 {
804         struct device *dev = &priv->client->dev;
805         int ret;
806
807         lockdep_assert_held(&priv->reg_lock);
808
809         if (priv->reg_current.txport == nport)
810                 return 0;
811
812         ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
813                            (nport << 4) | BIT(nport));
814         if (ret) {
815                 dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
816                         nport, ret);
817                 return ret;
818         }
819
820         priv->reg_current.txport = nport;
821
822         return 0;
823 }
824
825 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
826 {
827         struct device *dev = &priv->client->dev;
828         unsigned int v;
829         int ret;
830
831         mutex_lock(&priv->reg_lock);
832
833         ret = ub960_txport_select(priv, nport);
834         if (ret)
835                 goto out_unlock;
836
837         ret = regmap_read(priv->regmap, reg, &v);
838         if (ret) {
839                 dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
840                         __func__, reg, ret);
841                 goto out_unlock;
842         }
843
844         *val = v;
845
846 out_unlock:
847         mutex_unlock(&priv->reg_lock);
848
849         return ret;
850 }
851
852 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
853 {
854         struct device *dev = &priv->client->dev;
855         int ret;
856
857         mutex_lock(&priv->reg_lock);
858
859         ret = ub960_txport_select(priv, nport);
860         if (ret)
861                 goto out_unlock;
862
863         ret = regmap_write(priv->regmap, reg, val);
864         if (ret)
865                 dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
866                         __func__, reg, ret);
867
868 out_unlock:
869         mutex_unlock(&priv->reg_lock);
870
871         return ret;
872 }
873
874 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
875                                     u8 mask, u8 val)
876 {
877         struct device *dev = &priv->client->dev;
878         int ret;
879
880         mutex_lock(&priv->reg_lock);
881
882         ret = ub960_txport_select(priv, nport);
883         if (ret)
884                 goto out_unlock;
885
886         ret = regmap_update_bits(priv->regmap, reg, mask, val);
887         if (ret)
888                 dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
889                         __func__, reg, ret);
890
891 out_unlock:
892         mutex_unlock(&priv->reg_lock);
893
894         return ret;
895 }
896
897 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
898 {
899         struct device *dev = &priv->client->dev;
900         int ret;
901
902         lockdep_assert_held(&priv->reg_lock);
903
904         if (priv->reg_current.indirect_target == block)
905                 return 0;
906
907         ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
908         if (ret) {
909                 dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
910                         __func__, block, ret);
911                 return ret;
912         }
913
914         priv->reg_current.indirect_target = block;
915
916         return 0;
917 }
918
919 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
920 {
921         struct device *dev = &priv->client->dev;
922         unsigned int v;
923         int ret;
924
925         mutex_lock(&priv->reg_lock);
926
927         ret = ub960_select_ind_reg_block(priv, block);
928         if (ret)
929                 goto out_unlock;
930
931         ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
932         if (ret) {
933                 dev_err(dev,
934                         "Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
935                         block, reg, ret);
936                 goto out_unlock;
937         }
938
939         ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
940         if (ret) {
941                 dev_err(dev,
942                         "Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
943                         block, reg, ret);
944                 goto out_unlock;
945         }
946
947         *val = v;
948
949 out_unlock:
950         mutex_unlock(&priv->reg_lock);
951
952         return ret;
953 }
954
955 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
956 {
957         struct device *dev = &priv->client->dev;
958         int ret;
959
960         mutex_lock(&priv->reg_lock);
961
962         ret = ub960_select_ind_reg_block(priv, block);
963         if (ret)
964                 goto out_unlock;
965
966         ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
967         if (ret) {
968                 dev_err(dev,
969                         "Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
970                         block, reg, ret);
971                 goto out_unlock;
972         }
973
974         ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
975         if (ret) {
976                 dev_err(dev,
977                         "Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
978                         block, reg, ret);
979                 goto out_unlock;
980         }
981
982 out_unlock:
983         mutex_unlock(&priv->reg_lock);
984
985         return ret;
986 }
987
988 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
989                                  u8 mask, u8 val)
990 {
991         struct device *dev = &priv->client->dev;
992         int ret;
993
994         mutex_lock(&priv->reg_lock);
995
996         ret = ub960_select_ind_reg_block(priv, block);
997         if (ret)
998                 goto out_unlock;
999
1000         ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
1001         if (ret) {
1002                 dev_err(dev,
1003                         "Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1004                         block, reg, ret);
1005                 goto out_unlock;
1006         }
1007
1008         ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1009                                  val);
1010         if (ret) {
1011                 dev_err(dev,
1012                         "Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1013                         block, reg, ret);
1014                 goto out_unlock;
1015         }
1016
1017 out_unlock:
1018         mutex_unlock(&priv->reg_lock);
1019
1020         return ret;
1021 }
1022
1023 /* -----------------------------------------------------------------------------
1024  * I2C-ATR (address translator)
1025  */
1026
1027 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1028                                    const struct i2c_client *client, u16 alias)
1029 {
1030         struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1031         struct ub960_rxport *rxport = priv->rxports[chan_id];
1032         struct device *dev = &priv->client->dev;
1033         unsigned int reg_idx;
1034
1035         for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1036                 if (!rxport->aliased_clients[reg_idx])
1037                         break;
1038         }
1039
1040         if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1041                 dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1042                 return -EADDRNOTAVAIL;
1043         }
1044
1045         rxport->aliased_clients[reg_idx] = client;
1046
1047         ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1048                            client->addr << 1);
1049         ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1050                            alias << 1);
1051
1052         dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1053                 rxport->nport, client->addr, alias, reg_idx);
1054
1055         return 0;
1056 }
1057
1058 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1059                                     const struct i2c_client *client)
1060 {
1061         struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1062         struct ub960_rxport *rxport = priv->rxports[chan_id];
1063         struct device *dev = &priv->client->dev;
1064         unsigned int reg_idx;
1065
1066         for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1067                 if (rxport->aliased_clients[reg_idx] == client)
1068                         break;
1069         }
1070
1071         if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1072                 dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1073                         rxport->nport, client->addr);
1074                 return;
1075         }
1076
1077         rxport->aliased_clients[reg_idx] = NULL;
1078
1079         ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1080
1081         dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1082                 client->addr, reg_idx);
1083 }
1084
1085 static const struct i2c_atr_ops ub960_atr_ops = {
1086         .attach_client = ub960_atr_attach_client,
1087         .detach_client = ub960_atr_detach_client,
1088 };
1089
1090 static int ub960_init_atr(struct ub960_data *priv)
1091 {
1092         struct device *dev = &priv->client->dev;
1093         struct i2c_adapter *parent_adap = priv->client->adapter;
1094
1095         priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1096                                 priv->hw_data->num_rxports);
1097         if (IS_ERR(priv->atr))
1098                 return PTR_ERR(priv->atr);
1099
1100         i2c_atr_set_driver_data(priv->atr, priv);
1101
1102         return 0;
1103 }
1104
1105 static void ub960_uninit_atr(struct ub960_data *priv)
1106 {
1107         i2c_atr_delete(priv->atr);
1108         priv->atr = NULL;
1109 }
1110
1111 /* -----------------------------------------------------------------------------
1112  * TX ports
1113  */
1114
1115 static int ub960_parse_dt_txport(struct ub960_data *priv,
1116                                  struct fwnode_handle *ep_fwnode,
1117                                  u8 nport)
1118 {
1119         struct device *dev = &priv->client->dev;
1120         struct v4l2_fwnode_endpoint vep = {};
1121         struct ub960_txport *txport;
1122         int ret;
1123
1124         txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1125         if (!txport)
1126                 return -ENOMEM;
1127
1128         txport->priv = priv;
1129         txport->nport = nport;
1130
1131         vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1132         ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1133         if (ret) {
1134                 dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1135                 goto err_free_txport;
1136         }
1137
1138         txport->non_continous_clk = vep.bus.mipi_csi2.flags &
1139                                     V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK;
1140
1141         txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1142
1143         if (vep.nr_of_link_frequencies != 1) {
1144                 ret = -EINVAL;
1145                 goto err_free_vep;
1146         }
1147
1148         priv->tx_link_freq[0] = vep.link_frequencies[0];
1149         priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1150
1151         if (priv->tx_data_rate != MHZ(1600) &&
1152             priv->tx_data_rate != MHZ(1200) &&
1153             priv->tx_data_rate != MHZ(800) &&
1154             priv->tx_data_rate != MHZ(400)) {
1155                 dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1156                 ret = -EINVAL;
1157                 goto err_free_vep;
1158         }
1159
1160         v4l2_fwnode_endpoint_free(&vep);
1161
1162         priv->txports[nport] = txport;
1163
1164         return 0;
1165
1166 err_free_vep:
1167         v4l2_fwnode_endpoint_free(&vep);
1168 err_free_txport:
1169         kfree(txport);
1170
1171         return ret;
1172 }
1173
1174 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1175 {
1176         struct device *dev = &priv->client->dev;
1177         u8 csi_tx_isr;
1178         int ret;
1179
1180         ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1181         if (ret)
1182                 return;
1183
1184         if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1185                 dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1186
1187         if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1188                 dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1189 }
1190
1191 /* -----------------------------------------------------------------------------
1192  * RX ports
1193  */
1194
1195 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1196 {
1197         unsigned int nport;
1198         int ret;
1199
1200         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1201                 struct ub960_rxport *rxport = priv->rxports[nport];
1202
1203                 if (!rxport || !rxport->vpoc)
1204                         continue;
1205
1206                 ret = regulator_enable(rxport->vpoc);
1207                 if (ret)
1208                         goto err_disable_vpocs;
1209         }
1210
1211         return 0;
1212
1213 err_disable_vpocs:
1214         while (nport--) {
1215                 struct ub960_rxport *rxport = priv->rxports[nport];
1216
1217                 if (!rxport || !rxport->vpoc)
1218                         continue;
1219
1220                 regulator_disable(rxport->vpoc);
1221         }
1222
1223         return ret;
1224 }
1225
1226 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1227 {
1228         unsigned int nport;
1229
1230         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1231                 struct ub960_rxport *rxport = priv->rxports[nport];
1232
1233                 if (!rxport || !rxport->vpoc)
1234                         continue;
1235
1236                 regulator_disable(rxport->vpoc);
1237         }
1238 }
1239
1240 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1241                                       unsigned int nport)
1242 {
1243         u8 v;
1244
1245         ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1246         ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1247         ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1248         ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1249
1250         ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1251         ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1252
1253         ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1254 }
1255
1256 static void ub960_clear_rx_errors(struct ub960_data *priv)
1257 {
1258         unsigned int nport;
1259
1260         for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1261                 ub960_rxport_clear_errors(priv, nport);
1262 }
1263
1264 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1265                                        unsigned int nport, s8 *strobe_pos)
1266 {
1267         u8 v;
1268         u8 clk_delay, data_delay;
1269         int ret;
1270
1271         ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1272                        UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1273
1274         clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1275                             0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1276
1277         ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1278                        UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1279
1280         data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1281                              0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1282
1283         ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1284         if (ret)
1285                 return ret;
1286
1287         clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1288
1289         ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1290         if (ret)
1291                 return ret;
1292
1293         data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1294
1295         *strobe_pos = data_delay - clk_delay;
1296
1297         return 0;
1298 }
1299
1300 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1301                                         unsigned int nport, s8 strobe_pos)
1302 {
1303         u8 clk_delay, data_delay;
1304
1305         clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1306         data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1307
1308         if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1309                 clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1310         else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1311                 data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1312         else if (strobe_pos < 0)
1313                 clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1314         else if (strobe_pos > 0)
1315                 data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1316
1317         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1318                         UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1319
1320         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1321                         UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1322 }
1323
1324 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1325                                           s8 strobe_min, s8 strobe_max)
1326 {
1327         /* Convert the signed strobe pos to positive zero based value */
1328         strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1329         strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1330
1331         ub960_write(priv, UB960_XR_SFILTER_CFG,
1332                     ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1333                     ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1334 }
1335
1336 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1337                                      unsigned int nport, u8 *eq_level)
1338 {
1339         int ret;
1340         u8 v;
1341
1342         ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1343         if (ret)
1344                 return ret;
1345
1346         *eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1347                     (v & UB960_RR_AEQ_STATUS_STATUS_2);
1348
1349         return 0;
1350 }
1351
1352 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1353                                       unsigned int nport, u8 eq_level)
1354 {
1355         u8 eq_stage_1_select_value, eq_stage_2_select_value;
1356         const unsigned int eq_stage_max = 7;
1357         u8 v;
1358
1359         if (eq_level <= eq_stage_max) {
1360                 eq_stage_1_select_value = eq_level;
1361                 eq_stage_2_select_value = 0;
1362         } else {
1363                 eq_stage_1_select_value = eq_stage_max;
1364                 eq_stage_2_select_value = eq_level - eq_stage_max;
1365         }
1366
1367         ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1368
1369         v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1370                UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1371         v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1372         v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1373         v |= UB960_RR_AEQ_BYPASS_ENABLE;
1374
1375         ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1376 }
1377
1378 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1379                                       unsigned int nport, u8 eq_min, u8 eq_max)
1380 {
1381         ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1382                            (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1383                            (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1384
1385         /* Enable AEQ min setting */
1386         ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1387                                  UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1388                                  UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1389 }
1390
1391 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1392 {
1393         struct ub960_rxport *rxport = priv->rxports[nport];
1394
1395         /* We also set common settings here. Should be moved elsewhere. */
1396
1397         if (priv->strobe.manual) {
1398                 /* Disable AEQ_SFILTER_EN */
1399                 ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1400                                   UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1401         } else {
1402                 /* Enable SFILTER and error control */
1403                 ub960_write(priv, UB960_XR_AEQ_CTL1,
1404                             UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1405                                     UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1406
1407                 /* Set AEQ strobe range */
1408                 ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1409                                               priv->strobe.max);
1410         }
1411
1412         /* The rest are port specific */
1413
1414         if (priv->strobe.manual)
1415                 ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1416         else
1417                 ub960_rxport_set_strobe_pos(priv, nport, 0);
1418
1419         if (rxport->eq.manual_eq) {
1420                 ub960_rxport_set_eq_level(priv, nport,
1421                                           rxport->eq.manual.eq_level);
1422
1423                 /* Enable AEQ Bypass */
1424                 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1425                                          UB960_RR_AEQ_BYPASS_ENABLE,
1426                                          UB960_RR_AEQ_BYPASS_ENABLE);
1427         } else {
1428                 ub960_rxport_set_eq_range(priv, nport,
1429                                           rxport->eq.aeq.eq_level_min,
1430                                           rxport->eq.aeq.eq_level_max);
1431
1432                 /* Disable AEQ Bypass */
1433                 ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1434                                          UB960_RR_AEQ_BYPASS_ENABLE, 0);
1435         }
1436 }
1437
1438 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1439                                 bool *ok)
1440 {
1441         u8 rx_port_sts1, rx_port_sts2;
1442         u16 parity_errors;
1443         u8 csi_rx_sts;
1444         u8 csi_err_cnt;
1445         u8 bcc_sts;
1446         int ret;
1447         bool errors;
1448
1449         ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1450                                 &rx_port_sts1);
1451         if (ret)
1452                 return ret;
1453
1454         if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1455                 *ok = false;
1456                 return 0;
1457         }
1458
1459         ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1460                                 &rx_port_sts2);
1461         if (ret)
1462                 return ret;
1463
1464         ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1465         if (ret)
1466                 return ret;
1467
1468         ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1469                                 &csi_err_cnt);
1470         if (ret)
1471                 return ret;
1472
1473         ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1474         if (ret)
1475                 return ret;
1476
1477         ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1478                                   &parity_errors);
1479         if (ret)
1480                 return ret;
1481
1482         errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1483                  (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1484                  (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1485                  (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1486                  parity_errors;
1487
1488         *ok = !errors;
1489
1490         return 0;
1491 }
1492
1493 /*
1494  * Wait for the RX ports to lock, have no errors and have stable strobe position
1495  * and EQ level.
1496  */
1497 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1498                                    unsigned long port_mask,
1499                                    unsigned int *lock_mask)
1500 {
1501         struct device *dev = &priv->client->dev;
1502         unsigned long timeout;
1503         unsigned int link_ok_mask;
1504         unsigned int missing;
1505         unsigned int loops;
1506         u8 nport;
1507         int ret;
1508
1509         if (port_mask == 0) {
1510                 if (lock_mask)
1511                         *lock_mask = 0;
1512                 return 0;
1513         }
1514
1515         if (port_mask >= BIT(priv->hw_data->num_rxports))
1516                 return -EINVAL;
1517
1518         timeout = jiffies + msecs_to_jiffies(1000);
1519         loops = 0;
1520         link_ok_mask = 0;
1521
1522         while (time_before(jiffies, timeout)) {
1523                 missing = 0;
1524
1525                 for_each_set_bit(nport, &port_mask,
1526                                  priv->hw_data->num_rxports) {
1527                         struct ub960_rxport *rxport = priv->rxports[nport];
1528                         bool ok;
1529
1530                         if (!rxport)
1531                                 continue;
1532
1533                         ret = ub960_rxport_link_ok(priv, nport, &ok);
1534                         if (ret)
1535                                 return ret;
1536
1537                         /*
1538                          * We want the link to be ok for two consecutive loops,
1539                          * as a link could get established just before our test
1540                          * and drop soon after.
1541                          */
1542                         if (!ok || !(link_ok_mask & BIT(nport)))
1543                                 missing++;
1544
1545                         if (ok)
1546                                 link_ok_mask |= BIT(nport);
1547                         else
1548                                 link_ok_mask &= ~BIT(nport);
1549                 }
1550
1551                 loops++;
1552
1553                 if (missing == 0)
1554                         break;
1555
1556                 msleep(50);
1557         }
1558
1559         if (lock_mask)
1560                 *lock_mask = link_ok_mask;
1561
1562         dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1563         for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1564                 struct ub960_rxport *rxport = priv->rxports[nport];
1565                 s8 strobe_pos, eq_level;
1566                 u16 v;
1567
1568                 if (!rxport)
1569                         continue;
1570
1571                 if (!(link_ok_mask & BIT(nport))) {
1572                         dev_dbg(dev, "\trx%u: not locked\n", nport);
1573                         continue;
1574                 }
1575
1576                 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1577
1578                 ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
1579                 if (ret)
1580                         return ret;
1581
1582                 ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1583                 if (ret)
1584                         return ret;
1585
1586                 dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1587                         nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
1588         }
1589
1590         return 0;
1591 }
1592
1593 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1594                                                   struct ub960_rxport *rxport)
1595 {
1596         unsigned int mult;
1597         unsigned int div;
1598
1599         switch (rxport->rx_mode) {
1600         case RXPORT_MODE_RAW10:
1601         case RXPORT_MODE_RAW12_HF:
1602         case RXPORT_MODE_RAW12_LF:
1603                 mult = 1;
1604                 div = 10;
1605                 break;
1606
1607         case RXPORT_MODE_CSI2_SYNC:
1608                 mult = 2;
1609                 div = 1;
1610                 break;
1611
1612         case RXPORT_MODE_CSI2_NONSYNC:
1613                 mult = 2;
1614                 div = 5;
1615                 break;
1616
1617         default:
1618                 return 0;
1619         }
1620
1621         return clk_get_rate(priv->refclk) * mult / div;
1622 }
1623
1624 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1625                                                    struct ub960_rxport *rxport)
1626 {
1627         switch (rxport->rx_mode) {
1628         case RXPORT_MODE_RAW10:
1629         case RXPORT_MODE_RAW12_HF:
1630         case RXPORT_MODE_RAW12_LF:
1631                 return 2359400;
1632
1633         case RXPORT_MODE_CSI2_SYNC:
1634                 return 47187500;
1635
1636         case RXPORT_MODE_CSI2_NONSYNC:
1637                 return 9437500;
1638
1639         default:
1640                 return 0;
1641         }
1642 }
1643
1644 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1645 {
1646         struct ub960_rxport *rxport = priv->rxports[nport];
1647         struct device *dev = &priv->client->dev;
1648         struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1649         struct i2c_board_info ser_info = {
1650                 .of_node = to_of_node(rxport->ser.fwnode),
1651                 .fwnode = rxport->ser.fwnode,
1652                 .platform_data = ser_pdata,
1653         };
1654
1655         ser_pdata->port = nport;
1656         ser_pdata->atr = priv->atr;
1657         if (priv->hw_data->is_ub9702)
1658                 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1659         else
1660                 ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1661
1662         /*
1663          * The serializer is added under the same i2c adapter as the
1664          * deserializer. This is not quite right, as the serializer is behind
1665          * the FPD-Link.
1666          */
1667         ser_info.addr = rxport->ser.alias;
1668         rxport->ser.client =
1669                 i2c_new_client_device(priv->client->adapter, &ser_info);
1670         if (IS_ERR(rxport->ser.client)) {
1671                 dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1672                         ser_info.type);
1673                 return PTR_ERR(rxport->ser.client);
1674         }
1675
1676         dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1677                 nport, rxport->ser.client->addr,
1678                 rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1679
1680         return 0;
1681 }
1682
1683 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1684 {
1685         struct ub960_rxport *rxport = priv->rxports[nport];
1686
1687         i2c_unregister_device(rxport->ser.client);
1688         rxport->ser.client = NULL;
1689 }
1690
1691 /* Add serializer i2c devices for all initialized ports */
1692 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1693 {
1694         unsigned int nport;
1695         int ret;
1696
1697         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1698                 struct ub960_rxport *rxport = priv->rxports[nport];
1699
1700                 if (!rxport)
1701                         continue;
1702
1703                 ret = ub960_rxport_add_serializer(priv, nport);
1704                 if (ret)
1705                         goto err_remove_sers;
1706         }
1707
1708         return 0;
1709
1710 err_remove_sers:
1711         while (nport--) {
1712                 struct ub960_rxport *rxport = priv->rxports[nport];
1713
1714                 if (!rxport)
1715                         continue;
1716
1717                 ub960_rxport_remove_serializer(priv, nport);
1718         }
1719
1720         return ret;
1721 }
1722
1723 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1724 {
1725         unsigned int nport;
1726
1727         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1728                 struct ub960_rxport *rxport = priv->rxports[nport];
1729
1730                 if (!rxport)
1731                         continue;
1732
1733                 ub960_rxport_remove_serializer(priv, nport);
1734         }
1735 }
1736
1737 static void ub960_init_tx_port(struct ub960_data *priv,
1738                                struct ub960_txport *txport)
1739 {
1740         unsigned int nport = txport->nport;
1741         u8 csi_ctl = 0;
1742
1743         /*
1744          * From the datasheet: "initial CSI Skew-Calibration
1745          * sequence [...] should be set when operating at 1.6 Gbps"
1746          */
1747         if (priv->tx_data_rate == MHZ(1600))
1748                 csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1749
1750         csi_ctl |= (4 - txport->num_data_lanes) << 4;
1751
1752         if (!txport->non_continous_clk)
1753                 csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK;
1754
1755         ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1756 }
1757
1758 static int ub960_init_tx_ports(struct ub960_data *priv)
1759 {
1760         unsigned int nport;
1761         u8 speed_select;
1762         u8 pll_div;
1763
1764         /* TX ports */
1765
1766         switch (priv->tx_data_rate) {
1767         case MHZ(1600):
1768         default:
1769                 speed_select = 0;
1770                 pll_div = 0x10;
1771                 break;
1772         case MHZ(1200):
1773                 speed_select = 1;
1774                 pll_div = 0x18;
1775                 break;
1776         case MHZ(800):
1777                 speed_select = 2;
1778                 pll_div = 0x10;
1779                 break;
1780         case MHZ(400):
1781                 speed_select = 3;
1782                 pll_div = 0x10;
1783                 break;
1784         }
1785
1786         ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1787
1788         if (priv->hw_data->is_ub9702) {
1789                 ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1790
1791                 switch (priv->tx_data_rate) {
1792                 case MHZ(1600):
1793                 default:
1794                         ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1795                         ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1796                         break;
1797                 case MHZ(800):
1798                         ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1799                         ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1800                         ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1801                         break;
1802                 case MHZ(400):
1803                         ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1804                         break;
1805                 }
1806         }
1807
1808         for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1809                 struct ub960_txport *txport = priv->txports[nport];
1810
1811                 if (!txport)
1812                         continue;
1813
1814                 ub960_init_tx_port(priv, txport);
1815         }
1816
1817         return 0;
1818 }
1819
1820 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1821                                      struct ub960_rxport *rxport)
1822 {
1823         unsigned int nport = rxport->nport;
1824         u32 bc_freq_val;
1825
1826         /*
1827          * Back channel frequency select.
1828          * Override FREQ_SELECT from the strap.
1829          * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1830          * 2 - 10 Mbps
1831          * 6 - 50 Mbps (DS90UB953-Q1)
1832          *
1833          * Note that changing this setting will result in some errors on the back
1834          * channel for a short period of time.
1835          */
1836
1837         switch (rxport->rx_mode) {
1838         case RXPORT_MODE_RAW10:
1839         case RXPORT_MODE_RAW12_HF:
1840         case RXPORT_MODE_RAW12_LF:
1841                 bc_freq_val = 0;
1842                 break;
1843
1844         case RXPORT_MODE_CSI2_NONSYNC:
1845                 bc_freq_val = 2;
1846                 break;
1847
1848         case RXPORT_MODE_CSI2_SYNC:
1849                 bc_freq_val = 6;
1850                 break;
1851
1852         default:
1853                 return;
1854         }
1855
1856         ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1857                                  UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1858                                  bc_freq_val);
1859
1860         switch (rxport->rx_mode) {
1861         case RXPORT_MODE_RAW10:
1862                 /* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1863                 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1864                                          UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1865                                          0x3);
1866
1867                 /*
1868                  * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1869                  */
1870                 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1871                         UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1872                         0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1873
1874                 break;
1875
1876         case RXPORT_MODE_RAW12_HF:
1877         case RXPORT_MODE_RAW12_LF:
1878                 /* Not implemented */
1879                 return;
1880
1881         case RXPORT_MODE_CSI2_SYNC:
1882         case RXPORT_MODE_CSI2_NONSYNC:
1883                 /* CSI-2 Mode (DS90UB953-Q1 compatible) */
1884                 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1885                                          0x0);
1886
1887                 break;
1888         }
1889
1890         /* LV_POLARITY & FV_POLARITY */
1891         ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1892                                  rxport->lv_fv_pol);
1893
1894         /* Enable all interrupt sources from this port */
1895         ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1896         ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1897
1898         /* Enable I2C_PASS_THROUGH */
1899         ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1900                                  UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1901                                  UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1902
1903         /* Enable I2C communication to the serializer via the alias addr */
1904         ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1905                            rxport->ser.alias << 1);
1906
1907         /* Configure EQ related settings */
1908         ub960_rxport_config_eq(priv, nport);
1909
1910         /* Enable RX port */
1911         ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1912 }
1913
1914 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1915                                            struct ub960_rxport *rxport)
1916 {
1917         unsigned int nport = rxport->nport;
1918         u8 bc_freq_val;
1919         u8 fpd_func_mode;
1920
1921         switch (rxport->rx_mode) {
1922         case RXPORT_MODE_RAW10:
1923                 bc_freq_val = 0;
1924                 fpd_func_mode = 5;
1925                 break;
1926
1927         case RXPORT_MODE_RAW12_HF:
1928                 bc_freq_val = 0;
1929                 fpd_func_mode = 4;
1930                 break;
1931
1932         case RXPORT_MODE_RAW12_LF:
1933                 bc_freq_val = 0;
1934                 fpd_func_mode = 6;
1935                 break;
1936
1937         case RXPORT_MODE_CSI2_SYNC:
1938                 bc_freq_val = 6;
1939                 fpd_func_mode = 2;
1940                 break;
1941
1942         case RXPORT_MODE_CSI2_NONSYNC:
1943                 bc_freq_val = 2;
1944                 fpd_func_mode = 2;
1945                 break;
1946
1947         default:
1948                 return;
1949         }
1950
1951         ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1952                                  bc_freq_val);
1953         ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1954
1955         /* set serdes_eq_mode = 1 */
1956         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1957
1958         /* enable serdes driver */
1959         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1960
1961         /* set serdes_eq_offset=4 */
1962         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1963
1964         /* init default serdes_eq_max in 0xa9 */
1965         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
1966
1967         /* init serdes_eq_min in 0xaa */
1968         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
1969
1970         /* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
1971         ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
1972                               BIT(3), BIT(3));
1973
1974         /* RX port to half-rate */
1975         ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
1976                           BIT(nport * 2));
1977 }
1978
1979 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
1980                                                struct ub960_rxport *rxport)
1981 {
1982         unsigned int nport = rxport->nport;
1983         bool first_time_power_up = true;
1984
1985         if (first_time_power_up) {
1986                 u8 v;
1987
1988                 /* AEQ init */
1989                 ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
1990
1991                 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
1992                 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
1993
1994                 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
1995         }
1996
1997         /* enable serdes_eq_ctl2 */
1998         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
1999
2000         /* enable serdes_eq_ctl1 */
2001         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
2002
2003         /* enable serdes_eq_en */
2004         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
2005
2006         /* disable serdes_eq_override */
2007         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
2008
2009         /* disable serdes_gain_override */
2010         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2011 }
2012
2013 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2014                                            struct ub960_rxport *rxport)
2015 {
2016         unsigned int nport = rxport->nport;
2017         u8 bc_freq_val;
2018
2019         switch (rxport->rx_mode) {
2020         case RXPORT_MODE_RAW10:
2021                 bc_freq_val = 0;
2022                 break;
2023
2024         case RXPORT_MODE_RAW12_HF:
2025                 bc_freq_val = 0;
2026                 break;
2027
2028         case RXPORT_MODE_RAW12_LF:
2029                 bc_freq_val = 0;
2030                 break;
2031
2032         case RXPORT_MODE_CSI2_SYNC:
2033                 bc_freq_val = 6;
2034                 break;
2035
2036         case RXPORT_MODE_CSI2_NONSYNC:
2037                 bc_freq_val = 2;
2038                 break;
2039
2040         default:
2041                 return;
2042         }
2043
2044         ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2045                                  bc_freq_val);
2046
2047         /* FPD4 Sync Mode */
2048         ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2049
2050         /* add serdes_eq_offset of 4 */
2051         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2052
2053         /* FPD4 serdes_start_eq in 0x27: assign default */
2054         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2055         /* FPD4 serdes_end_eq in 0x28: assign default */
2056         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2057
2058         /* set serdes_driver_mode into FPD IV mode */
2059         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2060         /* set FPD PBC drv into FPD IV mode */
2061         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2062
2063         /* set serdes_system_init to 0x2f */
2064         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2065         /* set serdes_system_rst in reset mode */
2066         ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2067
2068         /* RX port to 7.55G mode */
2069         ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2070                           0 << (nport * 2));
2071
2072         ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2073 }
2074
2075 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2076                                       struct ub960_rxport *rxport)
2077 {
2078         unsigned int nport = rxport->nport;
2079
2080         if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2081                 ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2082         else /* RXPORT_CDR_FPD4 */
2083                 ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2084
2085         switch (rxport->rx_mode) {
2086         case RXPORT_MODE_RAW10:
2087                 /*
2088                  * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2089                  * 0b10 : 8-bit processing using upper 8 bits
2090                  */
2091                 ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2092                                          0x3 << 6, 0x2 << 6);
2093
2094                 break;
2095
2096         case RXPORT_MODE_RAW12_HF:
2097         case RXPORT_MODE_RAW12_LF:
2098                 /* Not implemented */
2099                 return;
2100
2101         case RXPORT_MODE_CSI2_SYNC:
2102         case RXPORT_MODE_CSI2_NONSYNC:
2103
2104                 break;
2105         }
2106
2107         /* LV_POLARITY & FV_POLARITY */
2108         ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2109                                  rxport->lv_fv_pol);
2110
2111         /* Enable all interrupt sources from this port */
2112         ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2113         ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2114
2115         /* Enable I2C_PASS_THROUGH */
2116         ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2117                                  UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2118                                  UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2119
2120         /* Enable I2C communication to the serializer via the alias addr */
2121         ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2122                            rxport->ser.alias << 1);
2123
2124         /* Enable RX port */
2125         ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2126
2127         if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2128                 /* unreset 960 AEQ */
2129                 ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2130         }
2131 }
2132
2133 static int ub960_init_rx_ports(struct ub960_data *priv)
2134 {
2135         unsigned int nport;
2136
2137         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2138                 struct ub960_rxport *rxport = priv->rxports[nport];
2139
2140                 if (!rxport)
2141                         continue;
2142
2143                 if (priv->hw_data->is_ub9702)
2144                         ub960_init_rx_port_ub9702(priv, rxport);
2145                 else
2146                         ub960_init_rx_port_ub960(priv, rxport);
2147         }
2148
2149         return 0;
2150 }
2151
2152 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2153 {
2154         struct device *dev = &priv->client->dev;
2155         u8 rx_port_sts1;
2156         u8 rx_port_sts2;
2157         u8 csi_rx_sts;
2158         u8 bcc_sts;
2159         int ret = 0;
2160
2161         /* Read interrupts (also clears most of them) */
2162         if (!ret)
2163                 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2164                                         &rx_port_sts1);
2165         if (!ret)
2166                 ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2167                                         &rx_port_sts2);
2168         if (!ret)
2169                 ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2170                                         &csi_rx_sts);
2171         if (!ret)
2172                 ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2173                                         &bcc_sts);
2174
2175         if (ret)
2176                 return;
2177
2178         if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2179                 u16 v;
2180
2181                 ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2182                                           &v);
2183                 if (!ret)
2184                         dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2185         }
2186
2187         if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2188                 dev_err(dev, "rx%u BCC CRC error\n", nport);
2189
2190         if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2191                 dev_err(dev, "rx%u BCC SEQ error\n", nport);
2192
2193         if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2194                 dev_err(dev, "rx%u line length unstable\n", nport);
2195
2196         if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2197                 dev_err(dev, "rx%u FPD3 encode error\n", nport);
2198
2199         if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2200                 dev_err(dev, "rx%u buffer error\n", nport);
2201
2202         if (csi_rx_sts)
2203                 dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2204
2205         if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2206                 dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2207
2208         if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2209                 dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2210
2211         if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2212                 dev_err(dev, "rx%u CSI checksum error\n", nport);
2213
2214         if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2215                 dev_err(dev, "rx%u CSI length error\n", nport);
2216
2217         if (bcc_sts)
2218                 dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2219
2220         if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2221                 dev_err(dev, "rx%u BCC response error", nport);
2222
2223         if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2224                 dev_err(dev, "rx%u BCC slave timeout", nport);
2225
2226         if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2227                 dev_err(dev, "rx%u BCC slave error", nport);
2228
2229         if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2230                 dev_err(dev, "rx%u BCC master timeout", nport);
2231
2232         if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2233                 dev_err(dev, "rx%u BCC master error", nport);
2234
2235         if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2236                 dev_err(dev, "rx%u BCC sequence error", nport);
2237
2238         if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2239                 u16 v;
2240
2241                 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2242                 if (!ret)
2243                         dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2244         }
2245
2246         if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2247                 u16 v;
2248
2249                 ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2250                                           &v);
2251                 if (!ret)
2252                         dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2253         }
2254
2255         if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2256                 dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2257                         (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2258                                 "locked" :
2259                                 "unlocked",
2260                         (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2261                                 "passed" :
2262                                 "not passed",
2263                         (rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2264                                 "no clock" :
2265                                 "clock ok",
2266                         (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2267                                 "stable freq" :
2268                                 "unstable freq");
2269         }
2270 }
2271
2272 /* -----------------------------------------------------------------------------
2273  * V4L2
2274  */
2275
2276 /*
2277  * The current implementation only supports a simple VC mapping, where all VCs
2278  * from a one RX port will be mapped to the same VC. Also, the hardware
2279  * dictates that all streams from an RX port must go to a single TX port.
2280  *
2281  * This function decides the target VC numbers for each RX port with a simple
2282  * algorithm, so that for each TX port, we get VC numbers starting from 0,
2283  * and counting up.
2284  *
2285  * E.g. if all four RX ports are in use, of which the first two go to the
2286  * first TX port and the secont two go to the second TX port, we would get
2287  * the following VCs for the four RX ports: 0, 1, 0, 1.
2288  *
2289  * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2290  * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2291  * probably needs to be somehow configurable. Device tree?
2292  */
2293 static void ub960_get_vc_maps(struct ub960_data *priv,
2294                               struct v4l2_subdev_state *state, u8 *vc)
2295 {
2296         u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2297         struct v4l2_subdev_route *route;
2298         u8 handled_mask = 0;
2299
2300         for_each_active_route(&state->routing, route) {
2301                 unsigned int rx, tx;
2302
2303                 rx = ub960_pad_to_port(priv, route->sink_pad);
2304                 if (BIT(rx) & handled_mask)
2305                         continue;
2306
2307                 tx = ub960_pad_to_port(priv, route->source_pad);
2308
2309                 vc[rx] = cur_vc[tx]++;
2310                 handled_mask |= BIT(rx);
2311         }
2312 }
2313
2314 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2315 {
2316         struct device *dev = &priv->client->dev;
2317
2318         dev_dbg(dev, "enable TX port %u\n", nport);
2319
2320         return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2321                                         UB960_TR_CSI_CTL_CSI_ENABLE,
2322                                         UB960_TR_CSI_CTL_CSI_ENABLE);
2323 }
2324
2325 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2326 {
2327         struct device *dev = &priv->client->dev;
2328
2329         dev_dbg(dev, "disable TX port %u\n", nport);
2330
2331         ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2332                                  UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2333 }
2334
2335 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2336 {
2337         struct device *dev = &priv->client->dev;
2338
2339         dev_dbg(dev, "enable RX port %u\n", nport);
2340
2341         /* Enable forwarding */
2342         return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2343                                  UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2344 }
2345
2346 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2347 {
2348         struct device *dev = &priv->client->dev;
2349
2350         dev_dbg(dev, "disable RX port %u\n", nport);
2351
2352         /* Disable forwarding */
2353         ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2354                           UB960_SR_FWD_CTL1_PORT_DIS(nport),
2355                           UB960_SR_FWD_CTL1_PORT_DIS(nport));
2356 }
2357
2358 /*
2359  * The driver only supports using a single VC for each source. This function
2360  * checks that each source only provides streams using a single VC.
2361  */
2362 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2363 {
2364         unsigned int nport;
2365         unsigned int i;
2366
2367         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2368                 struct ub960_rxport *rxport = priv->rxports[nport];
2369                 struct v4l2_mbus_frame_desc desc;
2370                 int ret;
2371                 u8 vc;
2372
2373                 if (!rxport)
2374                         continue;
2375
2376                 ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2377                                        rxport->source.pad, &desc);
2378                 if (ret)
2379                         return ret;
2380
2381                 if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2382                         continue;
2383
2384                 if (desc.num_entries == 0)
2385                         continue;
2386
2387                 vc = desc.entry[0].bus.csi2.vc;
2388
2389                 for (i = 1; i < desc.num_entries; i++) {
2390                         if (vc == desc.entry[i].bus.csi2.vc)
2391                                 continue;
2392
2393                         dev_err(&priv->client->dev,
2394                                 "rx%u: source with multiple virtual-channels is not supported\n",
2395                                 nport);
2396                         return -ENODEV;
2397                 }
2398         }
2399
2400         return 0;
2401 }
2402
2403 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2404                                                struct v4l2_subdev_state *state)
2405 {
2406         u8 fwd_ctl;
2407         struct {
2408                 u32 num_streams;
2409                 u8 pixel_dt;
2410                 u8 meta_dt;
2411                 u32 meta_lines;
2412                 u32 tx_port;
2413         } rx_data[UB960_MAX_RX_NPORTS] = {};
2414         u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2415         struct v4l2_subdev_route *route;
2416         unsigned int nport;
2417         int ret;
2418
2419         ret = ub960_validate_stream_vcs(priv);
2420         if (ret)
2421                 return ret;
2422
2423         ub960_get_vc_maps(priv, state, vc_map);
2424
2425         for_each_active_route(&state->routing, route) {
2426                 struct ub960_rxport *rxport;
2427                 struct ub960_txport *txport;
2428                 struct v4l2_mbus_framefmt *fmt;
2429                 const struct ub960_format_info *ub960_fmt;
2430                 unsigned int nport;
2431
2432                 nport = ub960_pad_to_port(priv, route->sink_pad);
2433
2434                 rxport = priv->rxports[nport];
2435                 if (!rxport)
2436                         return -EINVAL;
2437
2438                 txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2439                 if (!txport)
2440                         return -EINVAL;
2441
2442                 rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2443
2444                 rx_data[nport].num_streams++;
2445
2446                 /* For the rest, we are only interested in parallel busses */
2447                 if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2448                     rxport->rx_mode == RXPORT_MODE_CSI2_NONSYNC)
2449                         continue;
2450
2451                 if (rx_data[nport].num_streams > 2)
2452                         return -EPIPE;
2453
2454                 fmt = v4l2_subdev_state_get_stream_format(state,
2455                                                           route->sink_pad,
2456                                                           route->sink_stream);
2457                 if (!fmt)
2458                         return -EPIPE;
2459
2460                 ub960_fmt = ub960_find_format(fmt->code);
2461                 if (!ub960_fmt)
2462                         return -EPIPE;
2463
2464                 if (ub960_fmt->meta) {
2465                         if (fmt->height > 3) {
2466                                 dev_err(&priv->client->dev,
2467                                         "rx%u: unsupported metadata height %u\n",
2468                                         nport, fmt->height);
2469                                 return -EPIPE;
2470                         }
2471
2472                         rx_data[nport].meta_dt = ub960_fmt->datatype;
2473                         rx_data[nport].meta_lines = fmt->height;
2474                 } else {
2475                         rx_data[nport].pixel_dt = ub960_fmt->datatype;
2476                 }
2477         }
2478
2479         /* Configure RX ports */
2480
2481         /*
2482          * Keep all port forwardings disabled by default. Forwarding will be
2483          * enabled in ub960_enable_rx_port.
2484          */
2485         fwd_ctl = GENMASK(7, 4);
2486
2487         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2488                 struct ub960_rxport *rxport = priv->rxports[nport];
2489                 u8 vc = vc_map[nport];
2490
2491                 if (rx_data[nport].num_streams == 0)
2492                         continue;
2493
2494                 switch (rxport->rx_mode) {
2495                 case RXPORT_MODE_RAW10:
2496                         ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2497                                 rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2498
2499                         ub960_rxport_write(priv, rxport->nport,
2500                                 UB960_RR_RAW_EMBED_DTYPE,
2501                                 (rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2502                                         rx_data[nport].meta_dt);
2503
2504                         break;
2505
2506                 case RXPORT_MODE_RAW12_HF:
2507                 case RXPORT_MODE_RAW12_LF:
2508                         /* Not implemented */
2509                         break;
2510
2511                 case RXPORT_MODE_CSI2_SYNC:
2512                 case RXPORT_MODE_CSI2_NONSYNC:
2513                         if (!priv->hw_data->is_ub9702) {
2514                                 /* Map all VCs from this port to the same VC */
2515                                 ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2516                                                    (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2517                                                    (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2518                                                    (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2519                                                    (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2520                         } else {
2521                                 unsigned int i;
2522
2523                                 /* Map all VCs from this port to VC(nport) */
2524                                 for (i = 0; i < 8; i++)
2525                                         ub960_rxport_write(priv, nport,
2526                                                            UB960_RR_VC_ID_MAP(i),
2527                                                            nport);
2528                         }
2529
2530                         break;
2531                 }
2532
2533                 if (rx_data[nport].tx_port == 1)
2534                         fwd_ctl |= BIT(nport); /* forward to TX1 */
2535                 else
2536                         fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2537         }
2538
2539         ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2540
2541         return 0;
2542 }
2543
2544 static void ub960_update_streaming_status(struct ub960_data *priv)
2545 {
2546         unsigned int i;
2547
2548         for (i = 0; i < UB960_MAX_NPORTS; i++) {
2549                 if (priv->stream_enable_mask[i])
2550                         break;
2551         }
2552
2553         priv->streaming = i < UB960_MAX_NPORTS;
2554 }
2555
2556 static int ub960_enable_streams(struct v4l2_subdev *sd,
2557                                 struct v4l2_subdev_state *state, u32 source_pad,
2558                                 u64 source_streams_mask)
2559 {
2560         struct ub960_data *priv = sd_to_ub960(sd);
2561         struct device *dev = &priv->client->dev;
2562         u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2563         struct v4l2_subdev_route *route;
2564         unsigned int failed_port;
2565         unsigned int nport;
2566         int ret;
2567
2568         if (!priv->streaming) {
2569                 dev_dbg(dev, "Prepare for streaming\n");
2570                 ret = ub960_configure_ports_for_streaming(priv, state);
2571                 if (ret)
2572                         return ret;
2573         }
2574
2575         /* Enable TX port if not yet enabled */
2576         if (!priv->stream_enable_mask[source_pad]) {
2577                 ret = ub960_enable_tx_port(priv,
2578                                            ub960_pad_to_port(priv, source_pad));
2579                 if (ret)
2580                         return ret;
2581         }
2582
2583         priv->stream_enable_mask[source_pad] |= source_streams_mask;
2584
2585         /* Collect sink streams per pad which we need to enable */
2586         for_each_active_route(&state->routing, route) {
2587                 if (route->source_pad != source_pad)
2588                         continue;
2589
2590                 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2591                         continue;
2592
2593                 nport = ub960_pad_to_port(priv, route->sink_pad);
2594
2595                 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2596         }
2597
2598         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2599                 if (!sink_streams[nport])
2600                         continue;
2601
2602                 /* Enable the RX port if not yet enabled */
2603                 if (!priv->stream_enable_mask[nport]) {
2604                         ret = ub960_enable_rx_port(priv, nport);
2605                         if (ret) {
2606                                 failed_port = nport;
2607                                 goto err;
2608                         }
2609                 }
2610
2611                 priv->stream_enable_mask[nport] |= sink_streams[nport];
2612
2613                 dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2614                         sink_streams[nport]);
2615
2616                 ret = v4l2_subdev_enable_streams(
2617                         priv->rxports[nport]->source.sd,
2618                         priv->rxports[nport]->source.pad,
2619                         sink_streams[nport]);
2620                 if (ret) {
2621                         priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2622
2623                         if (!priv->stream_enable_mask[nport])
2624                                 ub960_disable_rx_port(priv, nport);
2625
2626                         failed_port = nport;
2627                         goto err;
2628                 }
2629         }
2630
2631         priv->streaming = true;
2632
2633         return 0;
2634
2635 err:
2636         for (nport = 0; nport < failed_port; nport++) {
2637                 if (!sink_streams[nport])
2638                         continue;
2639
2640                 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2641                         sink_streams[nport]);
2642
2643                 ret = v4l2_subdev_disable_streams(
2644                         priv->rxports[nport]->source.sd,
2645                         priv->rxports[nport]->source.pad,
2646                         sink_streams[nport]);
2647                 if (ret)
2648                         dev_err(dev, "Failed to disable streams: %d\n", ret);
2649
2650                 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2651
2652                 /* Disable RX port if no active streams */
2653                 if (!priv->stream_enable_mask[nport])
2654                         ub960_disable_rx_port(priv, nport);
2655         }
2656
2657         priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2658
2659         if (!priv->stream_enable_mask[source_pad])
2660                 ub960_disable_tx_port(priv,
2661                                       ub960_pad_to_port(priv, source_pad));
2662
2663         ub960_update_streaming_status(priv);
2664
2665         return ret;
2666 }
2667
2668 static int ub960_disable_streams(struct v4l2_subdev *sd,
2669                                  struct v4l2_subdev_state *state,
2670                                  u32 source_pad, u64 source_streams_mask)
2671 {
2672         struct ub960_data *priv = sd_to_ub960(sd);
2673         struct device *dev = &priv->client->dev;
2674         u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2675         struct v4l2_subdev_route *route;
2676         unsigned int nport;
2677         int ret;
2678
2679         /* Collect sink streams per pad which we need to disable */
2680         for_each_active_route(&state->routing, route) {
2681                 if (route->source_pad != source_pad)
2682                         continue;
2683
2684                 if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2685                         continue;
2686
2687                 nport = ub960_pad_to_port(priv, route->sink_pad);
2688
2689                 sink_streams[nport] |= BIT_ULL(route->sink_stream);
2690         }
2691
2692         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2693                 if (!sink_streams[nport])
2694                         continue;
2695
2696                 dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2697                         sink_streams[nport]);
2698
2699                 ret = v4l2_subdev_disable_streams(
2700                         priv->rxports[nport]->source.sd,
2701                         priv->rxports[nport]->source.pad,
2702                         sink_streams[nport]);
2703                 if (ret)
2704                         dev_err(dev, "Failed to disable streams: %d\n", ret);
2705
2706                 priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2707
2708                 /* Disable RX port if no active streams */
2709                 if (!priv->stream_enable_mask[nport])
2710                         ub960_disable_rx_port(priv, nport);
2711         }
2712
2713         /* Disable TX port if no active streams */
2714
2715         priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2716
2717         if (!priv->stream_enable_mask[source_pad])
2718                 ub960_disable_tx_port(priv,
2719                                       ub960_pad_to_port(priv, source_pad));
2720
2721         ub960_update_streaming_status(priv);
2722
2723         return 0;
2724 }
2725
2726 static int _ub960_set_routing(struct v4l2_subdev *sd,
2727                               struct v4l2_subdev_state *state,
2728                               struct v4l2_subdev_krouting *routing)
2729 {
2730         static const struct v4l2_mbus_framefmt format = {
2731                 .width = 640,
2732                 .height = 480,
2733                 .code = MEDIA_BUS_FMT_UYVY8_1X16,
2734                 .field = V4L2_FIELD_NONE,
2735                 .colorspace = V4L2_COLORSPACE_SRGB,
2736                 .ycbcr_enc = V4L2_YCBCR_ENC_601,
2737                 .quantization = V4L2_QUANTIZATION_LIM_RANGE,
2738                 .xfer_func = V4L2_XFER_FUNC_SRGB,
2739         };
2740         int ret;
2741
2742         /*
2743          * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2744          * frame desc is made dynamically allocated.
2745          */
2746
2747         if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2748                 return -E2BIG;
2749
2750         ret = v4l2_subdev_routing_validate(sd, routing,
2751                                            V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2752                                            V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2753         if (ret)
2754                 return ret;
2755
2756         ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2757         if (ret)
2758                 return ret;
2759
2760         return 0;
2761 }
2762
2763 static int ub960_set_routing(struct v4l2_subdev *sd,
2764                              struct v4l2_subdev_state *state,
2765                              enum v4l2_subdev_format_whence which,
2766                              struct v4l2_subdev_krouting *routing)
2767 {
2768         struct ub960_data *priv = sd_to_ub960(sd);
2769
2770         if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2771                 return -EBUSY;
2772
2773         return _ub960_set_routing(sd, state, routing);
2774 }
2775
2776 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2777                                 struct v4l2_mbus_frame_desc *fd)
2778 {
2779         struct ub960_data *priv = sd_to_ub960(sd);
2780         struct v4l2_subdev_route *route;
2781         struct v4l2_subdev_state *state;
2782         int ret = 0;
2783         struct device *dev = &priv->client->dev;
2784         u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2785
2786         if (!ub960_pad_is_source(priv, pad))
2787                 return -EINVAL;
2788
2789         fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2790
2791         state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2792
2793         ub960_get_vc_maps(priv, state, vc_map);
2794
2795         for_each_active_route(&state->routing, route) {
2796                 struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2797                 struct v4l2_mbus_frame_desc source_fd;
2798                 unsigned int nport;
2799                 unsigned int i;
2800
2801                 if (route->source_pad != pad)
2802                         continue;
2803
2804                 nport = ub960_pad_to_port(priv, route->sink_pad);
2805
2806                 ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2807                                        get_frame_desc,
2808                                        priv->rxports[nport]->source.pad,
2809                                        &source_fd);
2810                 if (ret) {
2811                         dev_err(dev,
2812                                 "Failed to get source frame desc for pad %u\n",
2813                                 route->sink_pad);
2814                         goto out_unlock;
2815                 }
2816
2817                 for (i = 0; i < source_fd.num_entries; i++) {
2818                         if (source_fd.entry[i].stream == route->sink_stream) {
2819                                 source_entry = &source_fd.entry[i];
2820                                 break;
2821                         }
2822                 }
2823
2824                 if (!source_entry) {
2825                         dev_err(dev,
2826                                 "Failed to find stream from source frame desc\n");
2827                         ret = -EPIPE;
2828                         goto out_unlock;
2829                 }
2830
2831                 fd->entry[fd->num_entries].stream = route->source_stream;
2832                 fd->entry[fd->num_entries].flags = source_entry->flags;
2833                 fd->entry[fd->num_entries].length = source_entry->length;
2834                 fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2835
2836                 fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2837
2838                 if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2839                         fd->entry[fd->num_entries].bus.csi2.dt =
2840                                 source_entry->bus.csi2.dt;
2841                 } else {
2842                         const struct ub960_format_info *ub960_fmt;
2843                         struct v4l2_mbus_framefmt *fmt;
2844
2845                         fmt = v4l2_subdev_state_get_stream_format(state, pad,
2846                                                                   route->source_stream);
2847
2848                         if (!fmt) {
2849                                 ret = -EINVAL;
2850                                 goto out_unlock;
2851                         }
2852
2853                         ub960_fmt = ub960_find_format(fmt->code);
2854                         if (!ub960_fmt) {
2855                                 dev_err(dev, "Unable to find format\n");
2856                                 ret = -EINVAL;
2857                                 goto out_unlock;
2858                         }
2859
2860                         fd->entry[fd->num_entries].bus.csi2.dt =
2861                                 ub960_fmt->datatype;
2862                 }
2863
2864                 fd->num_entries++;
2865         }
2866
2867 out_unlock:
2868         v4l2_subdev_unlock_state(state);
2869
2870         return ret;
2871 }
2872
2873 static int ub960_set_fmt(struct v4l2_subdev *sd,
2874                          struct v4l2_subdev_state *state,
2875                          struct v4l2_subdev_format *format)
2876 {
2877         struct ub960_data *priv = sd_to_ub960(sd);
2878         struct v4l2_mbus_framefmt *fmt;
2879
2880         if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2881                 return -EBUSY;
2882
2883         /* No transcoding, source and sink formats must match. */
2884         if (ub960_pad_is_source(priv, format->pad))
2885                 return v4l2_subdev_get_fmt(sd, state, format);
2886
2887         /*
2888          * Default to the first format if the requested media bus code isn't
2889          * supported.
2890          */
2891         if (!ub960_find_format(format->format.code))
2892                 format->format.code = ub960_formats[0].code;
2893
2894         fmt = v4l2_subdev_state_get_stream_format(state, format->pad,
2895                                                   format->stream);
2896         if (!fmt)
2897                 return -EINVAL;
2898
2899         *fmt = format->format;
2900
2901         fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2902                                                            format->stream);
2903         if (!fmt)
2904                 return -EINVAL;
2905
2906         *fmt = format->format;
2907
2908         return 0;
2909 }
2910
2911 static int ub960_init_cfg(struct v4l2_subdev *sd,
2912                           struct v4l2_subdev_state *state)
2913 {
2914         struct ub960_data *priv = sd_to_ub960(sd);
2915
2916         struct v4l2_subdev_route routes[] = {
2917                 {
2918                         .sink_pad = 0,
2919                         .sink_stream = 0,
2920                         .source_pad = priv->hw_data->num_rxports,
2921                         .source_stream = 0,
2922                         .flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2923                 },
2924         };
2925
2926         struct v4l2_subdev_krouting routing = {
2927                 .num_routes = ARRAY_SIZE(routes),
2928                 .routes = routes,
2929         };
2930
2931         return _ub960_set_routing(sd, state, &routing);
2932 }
2933
2934 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2935         .enable_streams = ub960_enable_streams,
2936         .disable_streams = ub960_disable_streams,
2937
2938         .set_routing = ub960_set_routing,
2939         .get_frame_desc = ub960_get_frame_desc,
2940
2941         .get_fmt = v4l2_subdev_get_fmt,
2942         .set_fmt = ub960_set_fmt,
2943
2944         .init_cfg = ub960_init_cfg,
2945 };
2946
2947 static int ub960_log_status(struct v4l2_subdev *sd)
2948 {
2949         struct ub960_data *priv = sd_to_ub960(sd);
2950         struct device *dev = &priv->client->dev;
2951         struct v4l2_subdev_state *state;
2952         unsigned int nport;
2953         unsigned int i;
2954         u16 v16 = 0;
2955         u8 v = 0;
2956         u8 id[UB960_SR_FPD3_RX_ID_LEN];
2957
2958         state = v4l2_subdev_lock_and_get_active_state(sd);
2959
2960         for (i = 0; i < sizeof(id); i++)
2961                 ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
2962
2963         dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
2964
2965         for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
2966                 struct ub960_txport *txport = priv->txports[nport];
2967
2968                 dev_info(dev, "TX %u\n", nport);
2969
2970                 if (!txport) {
2971                         dev_info(dev, "\tNot initialized\n");
2972                         continue;
2973                 }
2974
2975                 ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
2976                 dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
2977                          v & (u8)BIT(0));
2978
2979                 ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
2980                 dev_info(dev, "\tframe counter %u\n", v16);
2981
2982                 ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
2983                 dev_info(dev, "\tframe error counter %u\n", v16);
2984
2985                 ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
2986                 dev_info(dev, "\tline counter %u\n", v16);
2987
2988                 ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
2989                 dev_info(dev, "\tline error counter %u\n", v16);
2990         }
2991
2992         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2993                 struct ub960_rxport *rxport = priv->rxports[nport];
2994                 u8 eq_level;
2995                 s8 strobe_pos;
2996                 unsigned int i;
2997
2998                 dev_info(dev, "RX %u\n", nport);
2999
3000                 if (!rxport) {
3001                         dev_info(dev, "\tNot initialized\n");
3002                         continue;
3003                 }
3004
3005                 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
3006
3007                 if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3008                         dev_info(dev, "\tLocked\n");
3009                 else
3010                         dev_info(dev, "\tNot locked\n");
3011
3012                 dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3013                 ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3014                 dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3015
3016                 ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3017                 dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
3018
3019                 ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3020                 dev_info(dev, "\tparity errors %u\n", v16);
3021
3022                 ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3023                 dev_info(dev, "\tlines per frame %u\n", v16);
3024
3025                 ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3026                 dev_info(dev, "\tbytes per line %u\n", v16);
3027
3028                 ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3029                 dev_info(dev, "\tcsi_err_counter %u\n", v);
3030
3031                 /* Strobe */
3032
3033                 ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
3034
3035                 dev_info(dev, "\t%s strobe\n",
3036                          (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
3037                                                                   "Manual");
3038
3039                 if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
3040                         ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
3041
3042                         dev_info(dev, "\tStrobe range [%d, %d]\n",
3043                                  ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
3044                                  ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
3045                 }
3046
3047                 ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
3048
3049                 dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
3050
3051                 /* EQ */
3052
3053                 ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
3054
3055                 dev_info(dev, "\t%s EQ\n",
3056                          (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
3057                                                             "Adaptive");
3058
3059                 if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
3060                         ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
3061
3062                         dev_info(dev, "\tEQ range [%u, %u]\n",
3063                                  (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
3064                                  (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
3065                 }
3066
3067                 if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
3068                         dev_info(dev, "\tEQ level %u\n", eq_level);
3069
3070                 /* GPIOs */
3071                 for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3072                         u8 ctl_reg;
3073                         u8 ctl_shift;
3074
3075                         ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3076                         ctl_shift = (i % 2) * 4;
3077
3078                         ub960_rxport_read(priv, nport, ctl_reg, &v);
3079
3080                         dev_info(dev, "\tGPIO%u: mode %u\n", i,
3081                                  (v >> ctl_shift) & 0xf);
3082                 }
3083         }
3084
3085         v4l2_subdev_unlock_state(state);
3086
3087         return 0;
3088 }
3089
3090 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3091         .log_status = ub960_log_status,
3092         .subscribe_event = v4l2_ctrl_subdev_subscribe_event,
3093         .unsubscribe_event = v4l2_event_subdev_unsubscribe,
3094 };
3095
3096 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3097         .core = &ub960_subdev_core_ops,
3098         .pad = &ub960_pad_ops,
3099 };
3100
3101 static const struct media_entity_operations ub960_entity_ops = {
3102         .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3103         .link_validate = v4l2_subdev_link_validate,
3104         .has_pad_interdep = v4l2_subdev_has_pad_interdep,
3105 };
3106
3107 /* -----------------------------------------------------------------------------
3108  * Core
3109  */
3110
3111 static irqreturn_t ub960_handle_events(int irq, void *arg)
3112 {
3113         struct ub960_data *priv = arg;
3114         unsigned int i;
3115         u8 int_sts;
3116         u8 fwd_sts;
3117         int ret;
3118
3119         ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3120         if (ret || !int_sts)
3121                 return IRQ_NONE;
3122
3123         dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3124
3125         ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3126         if (ret)
3127                 return IRQ_NONE;
3128
3129         dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3130
3131         for (i = 0; i < priv->hw_data->num_txports; i++) {
3132                 if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3133                         ub960_csi_handle_events(priv, i);
3134         }
3135
3136         for (i = 0; i < priv->hw_data->num_rxports; i++) {
3137                 if (!priv->rxports[i])
3138                         continue;
3139
3140                 if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3141                         ub960_rxport_handle_events(priv, i);
3142         }
3143
3144         return IRQ_HANDLED;
3145 }
3146
3147 static void ub960_handler_work(struct work_struct *work)
3148 {
3149         struct delayed_work *dwork = to_delayed_work(work);
3150         struct ub960_data *priv =
3151                 container_of(dwork, struct ub960_data, poll_work);
3152
3153         ub960_handle_events(0, priv);
3154
3155         schedule_delayed_work(&priv->poll_work,
3156                               msecs_to_jiffies(UB960_POLL_TIME_MS));
3157 }
3158
3159 static void ub960_txport_free_ports(struct ub960_data *priv)
3160 {
3161         unsigned int nport;
3162
3163         for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3164                 struct ub960_txport *txport = priv->txports[nport];
3165
3166                 if (!txport)
3167                         continue;
3168
3169                 kfree(txport);
3170                 priv->txports[nport] = NULL;
3171         }
3172 }
3173
3174 static void ub960_rxport_free_ports(struct ub960_data *priv)
3175 {
3176         unsigned int nport;
3177
3178         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3179                 struct ub960_rxport *rxport = priv->rxports[nport];
3180
3181                 if (!rxport)
3182                         continue;
3183
3184                 fwnode_handle_put(rxport->source.ep_fwnode);
3185                 fwnode_handle_put(rxport->ser.fwnode);
3186
3187                 kfree(rxport);
3188                 priv->rxports[nport] = NULL;
3189         }
3190 }
3191
3192 static int
3193 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3194                                       struct fwnode_handle *link_fwnode,
3195                                       struct ub960_rxport *rxport)
3196 {
3197         struct device *dev = &priv->client->dev;
3198         unsigned int nport = rxport->nport;
3199         u32 rx_mode;
3200         u32 cdr_mode;
3201         s32 strobe_pos;
3202         u32 eq_level;
3203         u32 ser_i2c_alias;
3204         int ret;
3205
3206         cdr_mode = RXPORT_CDR_FPD3;
3207
3208         ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3209         if (ret < 0 && ret != -EINVAL) {
3210                 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3211                         "ti,cdr-mode", ret);
3212                 return ret;
3213         }
3214
3215         if (cdr_mode > RXPORT_CDR_LAST) {
3216                 dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3217                 return -EINVAL;
3218         }
3219
3220         if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3221                 dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3222                 return -EINVAL;
3223         }
3224
3225         rxport->cdr_mode = cdr_mode;
3226
3227         ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3228         if (ret < 0) {
3229                 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3230                         "ti,rx-mode", ret);
3231                 return ret;
3232         }
3233
3234         if (rx_mode > RXPORT_MODE_LAST) {
3235                 dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3236                 return -EINVAL;
3237         }
3238
3239         switch (rx_mode) {
3240         case RXPORT_MODE_RAW12_HF:
3241         case RXPORT_MODE_RAW12_LF:
3242                 dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3243                         rx_mode);
3244                 return -EINVAL;
3245         default:
3246                 break;
3247         }
3248
3249         rxport->rx_mode = rx_mode;
3250
3251         /* EQ & Strobe related */
3252
3253         /* Defaults */
3254         rxport->eq.manual_eq = false;
3255         rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3256         rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3257
3258         ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3259                                        &strobe_pos);
3260         if (ret) {
3261                 if (ret != -EINVAL) {
3262                         dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3263                                 "ti,strobe-pos", ret);
3264                         return ret;
3265                 }
3266         } else {
3267                 if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3268                     strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3269                         dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3270                                 nport, strobe_pos);
3271                         return -EINVAL;
3272                 }
3273
3274                 /* NOTE: ignored unless global manual strobe pos is also set */
3275                 rxport->eq.strobe_pos = strobe_pos;
3276                 if (!priv->strobe.manual)
3277                         dev_warn(dev,
3278                                  "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3279                                  nport);
3280         }
3281
3282         ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3283         if (ret) {
3284                 if (ret != -EINVAL) {
3285                         dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3286                                 "ti,eq-level", ret);
3287                         return ret;
3288                 }
3289         } else {
3290                 if (eq_level > UB960_MAX_EQ_LEVEL) {
3291                         dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3292                                 nport, eq_level);
3293                         return -EINVAL;
3294                 }
3295
3296                 rxport->eq.manual_eq = true;
3297                 rxport->eq.manual.eq_level = eq_level;
3298         }
3299
3300         ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3301                                        &ser_i2c_alias);
3302         if (ret) {
3303                 dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3304                         "i2c-alias", ret);
3305                 return ret;
3306         }
3307         rxport->ser.alias = ser_i2c_alias;
3308
3309         rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3310         if (!rxport->ser.fwnode) {
3311                 dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3312                 return -EINVAL;
3313         }
3314
3315         return 0;
3316 }
3317
3318 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3319                                                struct fwnode_handle *ep_fwnode,
3320                                                struct ub960_rxport *rxport)
3321 {
3322         struct device *dev = &priv->client->dev;
3323         struct v4l2_fwnode_endpoint vep = {};
3324         unsigned int nport = rxport->nport;
3325         bool hsync_hi;
3326         bool vsync_hi;
3327         int ret;
3328
3329         rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3330         if (!rxport->source.ep_fwnode) {
3331                 dev_err(dev, "rx%u: no remote endpoint\n", nport);
3332                 return -ENODEV;
3333         }
3334
3335         /* We currently have properties only for RAW modes */
3336
3337         switch (rxport->rx_mode) {
3338         case RXPORT_MODE_RAW10:
3339         case RXPORT_MODE_RAW12_HF:
3340         case RXPORT_MODE_RAW12_LF:
3341                 break;
3342         default:
3343                 return 0;
3344         }
3345
3346         vep.bus_type = V4L2_MBUS_PARALLEL;
3347         ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3348         if (ret) {
3349                 dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3350                 goto err_put_source_ep_fwnode;
3351         }
3352
3353         hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3354         vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3355
3356         /* LineValid and FrameValid are inverse to the h/vsync active */
3357         rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3358                             (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3359
3360         return 0;
3361
3362 err_put_source_ep_fwnode:
3363         fwnode_handle_put(rxport->source.ep_fwnode);
3364         return ret;
3365 }
3366
3367 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3368                                  struct fwnode_handle *link_fwnode,
3369                                  struct fwnode_handle *ep_fwnode)
3370 {
3371         static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3372                 "vpoc0", "vpoc1", "vpoc2", "vpoc3"
3373         };
3374         struct device *dev = &priv->client->dev;
3375         struct ub960_rxport *rxport;
3376         int ret;
3377
3378         rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3379         if (!rxport)
3380                 return -ENOMEM;
3381
3382         priv->rxports[nport] = rxport;
3383
3384         rxport->nport = nport;
3385         rxport->priv = priv;
3386
3387         ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3388         if (ret)
3389                 goto err_free_rxport;
3390
3391         rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3392         if (IS_ERR(rxport->vpoc)) {
3393                 ret = PTR_ERR(rxport->vpoc);
3394                 if (ret == -ENODEV) {
3395                         rxport->vpoc = NULL;
3396                 } else {
3397                         dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3398                                 nport, ret);
3399                         goto err_put_remote_fwnode;
3400                 }
3401         }
3402
3403         ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3404         if (ret)
3405                 goto err_put_remote_fwnode;
3406
3407         return 0;
3408
3409 err_put_remote_fwnode:
3410         fwnode_handle_put(rxport->ser.fwnode);
3411 err_free_rxport:
3412         priv->rxports[nport] = NULL;
3413         kfree(rxport);
3414         return ret;
3415 }
3416
3417 static struct fwnode_handle *
3418 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3419                               unsigned int nport)
3420 {
3421         struct fwnode_handle *link_fwnode;
3422         int ret;
3423
3424         fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3425                 u32 link_num;
3426
3427                 if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3428                         continue;
3429
3430                 ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3431                 if (ret) {
3432                         fwnode_handle_put(link_fwnode);
3433                         return NULL;
3434                 }
3435
3436                 if (nport == link_num)
3437                         return link_fwnode;
3438         }
3439
3440         return NULL;
3441 }
3442
3443 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3444 {
3445         struct device *dev = &priv->client->dev;
3446         struct fwnode_handle *links_fwnode;
3447         unsigned int nport;
3448         int ret;
3449
3450         links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3451         if (!links_fwnode) {
3452                 dev_err(dev, "'links' node missing\n");
3453                 return -ENODEV;
3454         }
3455
3456         /* Defaults, recommended by TI */
3457         priv->strobe.min = 2;
3458         priv->strobe.max = 3;
3459
3460         priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3461
3462         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3463                 struct fwnode_handle *link_fwnode;
3464                 struct fwnode_handle *ep_fwnode;
3465
3466                 link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3467                 if (!link_fwnode)
3468                         continue;
3469
3470                 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3471                                                             nport, 0, 0);
3472                 if (!ep_fwnode) {
3473                         fwnode_handle_put(link_fwnode);
3474                         continue;
3475                 }
3476
3477                 ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3478                                             ep_fwnode);
3479
3480                 fwnode_handle_put(link_fwnode);
3481                 fwnode_handle_put(ep_fwnode);
3482
3483                 if (ret) {
3484                         dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3485                         goto err_put_links;
3486                 }
3487         }
3488
3489         fwnode_handle_put(links_fwnode);
3490
3491         return 0;
3492
3493 err_put_links:
3494         fwnode_handle_put(links_fwnode);
3495
3496         return ret;
3497 }
3498
3499 static int ub960_parse_dt_txports(struct ub960_data *priv)
3500 {
3501         struct device *dev = &priv->client->dev;
3502         u32 nport;
3503         int ret;
3504
3505         for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3506                 unsigned int port = nport + priv->hw_data->num_rxports;
3507                 struct fwnode_handle *ep_fwnode;
3508
3509                 ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3510                                                             port, 0, 0);
3511                 if (!ep_fwnode)
3512                         continue;
3513
3514                 ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3515
3516                 fwnode_handle_put(ep_fwnode);
3517
3518                 if (ret)
3519                         break;
3520         }
3521
3522         return 0;
3523 }
3524
3525 static int ub960_parse_dt(struct ub960_data *priv)
3526 {
3527         int ret;
3528
3529         ret = ub960_parse_dt_rxports(priv);
3530         if (ret)
3531                 return ret;
3532
3533         ret = ub960_parse_dt_txports(priv);
3534         if (ret)
3535                 goto err_free_rxports;
3536
3537         return 0;
3538
3539 err_free_rxports:
3540         ub960_rxport_free_ports(priv);
3541
3542         return ret;
3543 }
3544
3545 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3546                               struct v4l2_subdev *subdev,
3547                               struct v4l2_async_connection *asd)
3548 {
3549         struct ub960_data *priv = sd_to_ub960(notifier->sd);
3550         struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3551         struct device *dev = &priv->client->dev;
3552         u8 nport = rxport->nport;
3553         unsigned int i;
3554         int ret;
3555
3556         ret = media_entity_get_fwnode_pad(&subdev->entity,
3557                                           rxport->source.ep_fwnode,
3558                                           MEDIA_PAD_FL_SOURCE);
3559         if (ret < 0) {
3560                 dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3561                 return ret;
3562         }
3563
3564         rxport->source.sd = subdev;
3565         rxport->source.pad = ret;
3566
3567         ret = media_create_pad_link(&rxport->source.sd->entity,
3568                                     rxport->source.pad, &priv->sd.entity, nport,
3569                                     MEDIA_LNK_FL_ENABLED |
3570                                             MEDIA_LNK_FL_IMMUTABLE);
3571         if (ret) {
3572                 dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3573                         rxport->source.sd->name, rxport->source.pad,
3574                         priv->sd.name, nport);
3575                 return ret;
3576         }
3577
3578         for (i = 0; i < priv->hw_data->num_rxports; i++) {
3579                 if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3580                         dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3581                         return 0;
3582                 }
3583         }
3584
3585         return 0;
3586 }
3587
3588 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3589                                 struct v4l2_subdev *subdev,
3590                                 struct v4l2_async_connection *asd)
3591 {
3592         struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3593
3594         rxport->source.sd = NULL;
3595 }
3596
3597 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3598         .bound = ub960_notify_bound,
3599         .unbind = ub960_notify_unbind,
3600 };
3601
3602 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3603 {
3604         struct device *dev = &priv->client->dev;
3605         unsigned int i;
3606         int ret;
3607
3608         v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
3609
3610         for (i = 0; i < priv->hw_data->num_rxports; i++) {
3611                 struct ub960_rxport *rxport = priv->rxports[i];
3612                 struct ub960_asd *asd;
3613
3614                 if (!rxport)
3615                         continue;
3616
3617                 asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3618                                                rxport->source.ep_fwnode,
3619                                                struct ub960_asd);
3620                 if (IS_ERR(asd)) {
3621                         dev_err(dev, "Failed to add subdev for source %u: %pe",
3622                                 i, asd);
3623                         v4l2_async_nf_cleanup(&priv->notifier);
3624                         return PTR_ERR(asd);
3625                 }
3626
3627                 asd->rxport = rxport;
3628         }
3629
3630         priv->notifier.ops = &ub960_notify_ops;
3631
3632         ret = v4l2_async_nf_register(&priv->notifier);
3633         if (ret) {
3634                 dev_err(dev, "Failed to register subdev_notifier");
3635                 v4l2_async_nf_cleanup(&priv->notifier);
3636                 return ret;
3637         }
3638
3639         return 0;
3640 }
3641
3642 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3643 {
3644         v4l2_async_nf_unregister(&priv->notifier);
3645         v4l2_async_nf_cleanup(&priv->notifier);
3646 }
3647
3648 static int ub960_create_subdev(struct ub960_data *priv)
3649 {
3650         struct device *dev = &priv->client->dev;
3651         unsigned int i;
3652         int ret;
3653
3654         v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3655
3656         v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3657         priv->sd.ctrl_handler = &priv->ctrl_handler;
3658
3659         v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3660                                ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3661                                priv->tx_link_freq);
3662
3663         if (priv->ctrl_handler.error) {
3664                 ret = priv->ctrl_handler.error;
3665                 goto err_free_ctrl;
3666         }
3667
3668         priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3669                           V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
3670         priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3671         priv->sd.entity.ops = &ub960_entity_ops;
3672
3673         for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3674                 priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3675                                               MEDIA_PAD_FL_SINK :
3676                                               MEDIA_PAD_FL_SOURCE;
3677         }
3678
3679         ret = media_entity_pads_init(&priv->sd.entity,
3680                                      priv->hw_data->num_rxports +
3681                                              priv->hw_data->num_txports,
3682                                      priv->pads);
3683         if (ret)
3684                 goto err_free_ctrl;
3685
3686         priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3687
3688         ret = v4l2_subdev_init_finalize(&priv->sd);
3689         if (ret)
3690                 goto err_entity_cleanup;
3691
3692         ret = ub960_v4l2_notifier_register(priv);
3693         if (ret) {
3694                 dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3695                 goto err_subdev_cleanup;
3696         }
3697
3698         ret = v4l2_async_register_subdev(&priv->sd);
3699         if (ret) {
3700                 dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3701                 goto err_unreg_notif;
3702         }
3703
3704         return 0;
3705
3706 err_unreg_notif:
3707         ub960_v4l2_notifier_unregister(priv);
3708 err_subdev_cleanup:
3709         v4l2_subdev_cleanup(&priv->sd);
3710 err_entity_cleanup:
3711         media_entity_cleanup(&priv->sd.entity);
3712 err_free_ctrl:
3713         v4l2_ctrl_handler_free(&priv->ctrl_handler);
3714
3715         return ret;
3716 }
3717
3718 static void ub960_destroy_subdev(struct ub960_data *priv)
3719 {
3720         ub960_v4l2_notifier_unregister(priv);
3721         v4l2_async_unregister_subdev(&priv->sd);
3722
3723         v4l2_subdev_cleanup(&priv->sd);
3724
3725         media_entity_cleanup(&priv->sd.entity);
3726         v4l2_ctrl_handler_free(&priv->ctrl_handler);
3727 }
3728
3729 static const struct regmap_config ub960_regmap_config = {
3730         .name = "ds90ub960",
3731
3732         .reg_bits = 8,
3733         .val_bits = 8,
3734
3735         .max_register = 0xff,
3736
3737         /*
3738          * We do locking in the driver to cover the TX/RX port selection and the
3739          * indirect register access.
3740          */
3741         .disable_locking = true,
3742 };
3743
3744 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3745 {
3746         struct device *dev = &priv->client->dev;
3747         unsigned int v;
3748         int ret;
3749         u8 bit;
3750
3751         bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3752                            UB960_SR_RESET_DIGITAL_RESET0;
3753
3754         ub960_write(priv, UB960_SR_RESET, bit);
3755
3756         mutex_lock(&priv->reg_lock);
3757
3758         ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3759                                        (v & bit) == 0, 2000, 100000);
3760
3761         mutex_unlock(&priv->reg_lock);
3762
3763         if (ret)
3764                 dev_err(dev, "reset failed: %d\n", ret);
3765 }
3766
3767 static int ub960_get_hw_resources(struct ub960_data *priv)
3768 {
3769         struct device *dev = &priv->client->dev;
3770
3771         priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3772         if (IS_ERR(priv->regmap))
3773                 return PTR_ERR(priv->regmap);
3774
3775         priv->vddio = devm_regulator_get(dev, "vddio");
3776         if (IS_ERR(priv->vddio))
3777                 return dev_err_probe(dev, PTR_ERR(priv->vddio),
3778                                      "cannot get VDDIO regulator\n");
3779
3780         /* get power-down pin from DT */
3781         priv->pd_gpio =
3782                 devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3783         if (IS_ERR(priv->pd_gpio))
3784                 return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3785                                      "Cannot get powerdown GPIO\n");
3786
3787         priv->refclk = devm_clk_get(dev, "refclk");
3788         if (IS_ERR(priv->refclk))
3789                 return dev_err_probe(dev, PTR_ERR(priv->refclk),
3790                                      "Cannot get REFCLK\n");
3791
3792         return 0;
3793 }
3794
3795 static int ub960_enable_core_hw(struct ub960_data *priv)
3796 {
3797         struct device *dev = &priv->client->dev;
3798         u8 rev_mask;
3799         int ret;
3800         u8 dev_sts;
3801         u8 refclk_freq;
3802
3803         ret = regulator_enable(priv->vddio);
3804         if (ret)
3805                 return dev_err_probe(dev, ret,
3806                                      "failed to enable VDDIO regulator\n");
3807
3808         ret = clk_prepare_enable(priv->refclk);
3809         if (ret) {
3810                 dev_err_probe(dev, ret, "Failed to enable refclk\n");
3811                 goto err_disable_vddio;
3812         }
3813
3814         if (priv->pd_gpio) {
3815                 gpiod_set_value_cansleep(priv->pd_gpio, 1);
3816                 /* wait min 2 ms for reset to complete */
3817                 fsleep(2000);
3818                 gpiod_set_value_cansleep(priv->pd_gpio, 0);
3819                 /* wait min 2 ms for power up to finish */
3820                 fsleep(2000);
3821         }
3822
3823         ub960_reset(priv, true);
3824
3825         /* Runtime check register accessibility */
3826         ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3827         if (ret) {
3828                 dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3829                 goto err_pd_gpio;
3830         }
3831
3832         dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3833                 rev_mask);
3834
3835         ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3836         if (ret)
3837                 goto err_pd_gpio;
3838
3839         ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3840         if (ret)
3841                 goto err_pd_gpio;
3842
3843         dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3844                 !!(dev_sts & BIT(4)), refclk_freq,
3845                 clk_get_rate(priv->refclk) / 1000000);
3846
3847         /* Disable all RX ports by default */
3848         ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3849         if (ret)
3850                 goto err_pd_gpio;
3851
3852         /* release GPIO lock */
3853         if (priv->hw_data->is_ub9702) {
3854                 ret = ub960_update_bits(priv, UB960_SR_RESET,
3855                                         UB960_SR_RESET_GPIO_LOCK_RELEASE,
3856                                         UB960_SR_RESET_GPIO_LOCK_RELEASE);
3857                 if (ret)
3858                         goto err_pd_gpio;
3859         }
3860
3861         return 0;
3862
3863 err_pd_gpio:
3864         gpiod_set_value_cansleep(priv->pd_gpio, 1);
3865         clk_disable_unprepare(priv->refclk);
3866 err_disable_vddio:
3867         regulator_disable(priv->vddio);
3868
3869         return ret;
3870 }
3871
3872 static void ub960_disable_core_hw(struct ub960_data *priv)
3873 {
3874         gpiod_set_value_cansleep(priv->pd_gpio, 1);
3875         clk_disable_unprepare(priv->refclk);
3876         regulator_disable(priv->vddio);
3877 }
3878
3879 static int ub960_probe(struct i2c_client *client)
3880 {
3881         struct device *dev = &client->dev;
3882         struct ub960_data *priv;
3883         unsigned int port_lock_mask;
3884         unsigned int port_mask;
3885         unsigned int nport;
3886         int ret;
3887
3888         priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3889         if (!priv)
3890                 return -ENOMEM;
3891
3892         priv->client = client;
3893
3894         priv->hw_data = device_get_match_data(dev);
3895
3896         mutex_init(&priv->reg_lock);
3897
3898         INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3899
3900         /*
3901          * Initialize these to invalid values so that the first reg writes will
3902          * configure the target.
3903          */
3904         priv->reg_current.indirect_target = 0xff;
3905         priv->reg_current.rxport = 0xff;
3906         priv->reg_current.txport = 0xff;
3907
3908         ret = ub960_get_hw_resources(priv);
3909         if (ret)
3910                 goto err_mutex_destroy;
3911
3912         ret = ub960_enable_core_hw(priv);
3913         if (ret)
3914                 goto err_mutex_destroy;
3915
3916         ret = ub960_parse_dt(priv);
3917         if (ret)
3918                 goto err_disable_core_hw;
3919
3920         ret = ub960_init_tx_ports(priv);
3921         if (ret)
3922                 goto err_free_ports;
3923
3924         ret = ub960_rxport_enable_vpocs(priv);
3925         if (ret)
3926                 goto err_free_ports;
3927
3928         ret = ub960_init_rx_ports(priv);
3929         if (ret)
3930                 goto err_disable_vpocs;
3931
3932         ub960_reset(priv, false);
3933
3934         port_mask = 0;
3935
3936         for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3937                 struct ub960_rxport *rxport = priv->rxports[nport];
3938
3939                 if (!rxport)
3940                         continue;
3941
3942                 port_mask |= BIT(nport);
3943         }
3944
3945         ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
3946         if (ret)
3947                 goto err_disable_vpocs;
3948
3949         if (port_mask != port_lock_mask) {
3950                 ret = -EIO;
3951                 dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
3952                 goto err_disable_vpocs;
3953         }
3954
3955         /*
3956          * Clear any errors caused by switching the RX port settings while
3957          * probing.
3958          */
3959         ub960_clear_rx_errors(priv);
3960
3961         ret = ub960_init_atr(priv);
3962         if (ret)
3963                 goto err_disable_vpocs;
3964
3965         ret = ub960_rxport_add_serializers(priv);
3966         if (ret)
3967                 goto err_uninit_atr;
3968
3969         ret = ub960_create_subdev(priv);
3970         if (ret)
3971                 goto err_free_sers;
3972
3973         if (client->irq)
3974                 dev_warn(dev, "irq support not implemented, using polling\n");
3975
3976         schedule_delayed_work(&priv->poll_work,
3977                               msecs_to_jiffies(UB960_POLL_TIME_MS));
3978
3979         return 0;
3980
3981 err_free_sers:
3982         ub960_rxport_remove_serializers(priv);
3983 err_uninit_atr:
3984         ub960_uninit_atr(priv);
3985 err_disable_vpocs:
3986         ub960_rxport_disable_vpocs(priv);
3987 err_free_ports:
3988         ub960_rxport_free_ports(priv);
3989         ub960_txport_free_ports(priv);
3990 err_disable_core_hw:
3991         ub960_disable_core_hw(priv);
3992 err_mutex_destroy:
3993         mutex_destroy(&priv->reg_lock);
3994         return ret;
3995 }
3996
3997 static void ub960_remove(struct i2c_client *client)
3998 {
3999         struct v4l2_subdev *sd = i2c_get_clientdata(client);
4000         struct ub960_data *priv = sd_to_ub960(sd);
4001
4002         cancel_delayed_work_sync(&priv->poll_work);
4003
4004         ub960_destroy_subdev(priv);
4005         ub960_rxport_remove_serializers(priv);
4006         ub960_uninit_atr(priv);
4007         ub960_rxport_disable_vpocs(priv);
4008         ub960_rxport_free_ports(priv);
4009         ub960_txport_free_ports(priv);
4010         ub960_disable_core_hw(priv);
4011         mutex_destroy(&priv->reg_lock);
4012 }
4013
4014 static const struct ub960_hw_data ds90ub960_hw = {
4015         .model = "ub960",
4016         .num_rxports = 4,
4017         .num_txports = 2,
4018 };
4019
4020 static const struct ub960_hw_data ds90ub9702_hw = {
4021         .model = "ub9702",
4022         .num_rxports = 4,
4023         .num_txports = 2,
4024         .is_ub9702 = true,
4025         .is_fpdlink4 = true,
4026 };
4027
4028 static const struct i2c_device_id ub960_id[] = {
4029         { "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4030         { "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4031         {}
4032 };
4033 MODULE_DEVICE_TABLE(i2c, ub960_id);
4034
4035 static const struct of_device_id ub960_dt_ids[] = {
4036         { .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4037         { .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4038         {}
4039 };
4040 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4041
4042 static struct i2c_driver ds90ub960_driver = {
4043         .probe          = ub960_probe,
4044         .remove         = ub960_remove,
4045         .id_table       = ub960_id,
4046         .driver = {
4047                 .name   = "ds90ub960",
4048                 .of_match_table = ub960_dt_ids,
4049         },
4050 };
4051 module_i2c_driver(ds90ub960_driver);
4052
4053 MODULE_LICENSE("GPL");
4054 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4055 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4056 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4057 MODULE_IMPORT_NS(I2C_ATR);