GNU Linux-libre 4.9.301-gnu1
[releases.git] / drivers / media / platform / sti / c8sectpfe / c8sectpfe-core.c
1 /*
2  * c8sectpfe-core.c - C8SECTPFE STi DVB driver
3  *
4  * Copyright (c) STMicroelectronics 2015
5  *
6  *   Author:Peter Bennett <peter.bennett@st.com>
7  *          Peter Griffin <peter.griffin@linaro.org>
8  *
9  *      This program is free software; you can redistribute it and/or
10  *      modify it under the terms of the GNU General Public License as
11  *      published by the Free Software Foundation; either version 2 of
12  *      the License, or (at your option) any later version.
13  */
14 #include <linux/atomic.h>
15 #include <linux/clk.h>
16 #include <linux/completion.h>
17 #include <linux/delay.h>
18 #include <linux/device.h>
19 #include <linux/dma-mapping.h>
20 #include <linux/dvb/dmx.h>
21 #include <linux/dvb/frontend.h>
22 #include <linux/errno.h>
23 #include <linux/firmware.h>
24 #include <linux/init.h>
25 #include <linux/interrupt.h>
26 #include <linux/io.h>
27 #include <linux/module.h>
28 #include <linux/of_gpio.h>
29 #include <linux/of_platform.h>
30 #include <linux/platform_device.h>
31 #include <linux/usb.h>
32 #include <linux/slab.h>
33 #include <linux/time.h>
34 #include <linux/version.h>
35 #include <linux/wait.h>
36 #include <linux/pinctrl/pinctrl.h>
37
38 #include "c8sectpfe-core.h"
39 #include "c8sectpfe-common.h"
40 #include "c8sectpfe-debugfs.h"
41 #include "dmxdev.h"
42 #include "dvb_demux.h"
43 #include "dvb_frontend.h"
44 #include "dvb_net.h"
45
46 #define FIRMWARE_MEMDMA "/*(DEBLOBBED)*/"
47 /*(DEBLOBBED)*/
48
49 #define PID_TABLE_SIZE 1024
50 #define POLL_MSECS 50
51
52 static int load_c8sectpfe_fw(struct c8sectpfei *fei);
53
54 #define TS_PKT_SIZE 188
55 #define HEADER_SIZE (4)
56 #define PACKET_SIZE (TS_PKT_SIZE+HEADER_SIZE)
57
58 #define FEI_ALIGNMENT (32)
59 /* hw requires minimum of 8*PACKET_SIZE and padded to 8byte boundary */
60 #define FEI_BUFFER_SIZE (8*PACKET_SIZE*340)
61
62 #define FIFO_LEN 1024
63
64 static void c8sectpfe_timer_interrupt(unsigned long ac8sectpfei)
65 {
66         struct c8sectpfei *fei = (struct c8sectpfei *)ac8sectpfei;
67         struct channel_info *channel;
68         int chan_num;
69
70         /* iterate through input block channels */
71         for (chan_num = 0; chan_num < fei->tsin_count; chan_num++) {
72                 channel = fei->channel_data[chan_num];
73
74                 /* is this descriptor initialised and TP enabled */
75                 if (channel->irec && readl(channel->irec + DMA_PRDS_TPENABLE))
76                         tasklet_schedule(&channel->tsklet);
77         }
78
79         fei->timer.expires = jiffies +  msecs_to_jiffies(POLL_MSECS);
80         add_timer(&fei->timer);
81 }
82
83 static void channel_swdemux_tsklet(unsigned long data)
84 {
85         struct channel_info *channel = (struct channel_info *)data;
86         struct c8sectpfei *fei;
87         unsigned long wp, rp;
88         int pos, num_packets, n, size;
89         u8 *buf;
90
91         if (unlikely(!channel || !channel->irec))
92                 return;
93
94         fei = channel->fei;
95
96         wp = readl(channel->irec + DMA_PRDS_BUSWP_TP(0));
97         rp = readl(channel->irec + DMA_PRDS_BUSRP_TP(0));
98
99         pos = rp - channel->back_buffer_busaddr;
100
101         /* has it wrapped */
102         if (wp < rp)
103                 wp = channel->back_buffer_busaddr + FEI_BUFFER_SIZE;
104
105         size = wp - rp;
106         num_packets = size / PACKET_SIZE;
107
108         /* manage cache so data is visible to CPU */
109         dma_sync_single_for_cpu(fei->dev,
110                                 rp,
111                                 size,
112                                 DMA_FROM_DEVICE);
113
114         buf = (u8 *) channel->back_buffer_aligned;
115
116         dev_dbg(fei->dev,
117                 "chan=%d channel=%p num_packets = %d, buf = %p, pos = 0x%x\n\t"
118                 "rp=0x%lx, wp=0x%lx\n",
119                 channel->tsin_id, channel, num_packets, buf, pos, rp, wp);
120
121         for (n = 0; n < num_packets; n++) {
122                 dvb_dmx_swfilter_packets(
123                         &fei->c8sectpfe[0]->
124                                 demux[channel->demux_mapping].dvb_demux,
125                         &buf[pos], 1);
126
127                 pos += PACKET_SIZE;
128         }
129
130         /* advance the read pointer */
131         if (wp == (channel->back_buffer_busaddr + FEI_BUFFER_SIZE))
132                 writel(channel->back_buffer_busaddr, channel->irec +
133                         DMA_PRDS_BUSRP_TP(0));
134         else
135                 writel(wp, channel->irec + DMA_PRDS_BUSRP_TP(0));
136 }
137
138 static int c8sectpfe_start_feed(struct dvb_demux_feed *dvbdmxfeed)
139 {
140         struct dvb_demux *demux = dvbdmxfeed->demux;
141         struct stdemux *stdemux = (struct stdemux *)demux->priv;
142         struct c8sectpfei *fei = stdemux->c8sectpfei;
143         struct channel_info *channel;
144         u32 tmp;
145         unsigned long *bitmap;
146         int ret;
147
148         switch (dvbdmxfeed->type) {
149         case DMX_TYPE_TS:
150                 break;
151         case DMX_TYPE_SEC:
152                 break;
153         default:
154                 dev_err(fei->dev, "%s:%d Error bailing\n"
155                         , __func__, __LINE__);
156                 return -EINVAL;
157         }
158
159         if (dvbdmxfeed->type == DMX_TYPE_TS) {
160                 switch (dvbdmxfeed->pes_type) {
161                 case DMX_PES_VIDEO:
162                 case DMX_PES_AUDIO:
163                 case DMX_PES_TELETEXT:
164                 case DMX_PES_PCR:
165                 case DMX_PES_OTHER:
166                         break;
167                 default:
168                         dev_err(fei->dev, "%s:%d Error bailing\n"
169                                 , __func__, __LINE__);
170                         return -EINVAL;
171                 }
172         }
173
174         if (!atomic_read(&fei->fw_loaded)) {
175                 ret = load_c8sectpfe_fw(fei);
176                 if (ret)
177                         return ret;
178         }
179
180         mutex_lock(&fei->lock);
181
182         channel = fei->channel_data[stdemux->tsin_index];
183
184         bitmap = (unsigned long *) channel->pid_buffer_aligned;
185
186         /* 8192 is a special PID */
187         if (dvbdmxfeed->pid == 8192) {
188                 tmp = readl(fei->io + C8SECTPFE_IB_PID_SET(channel->tsin_id));
189                 tmp &= ~C8SECTPFE_PID_ENABLE;
190                 writel(tmp, fei->io + C8SECTPFE_IB_PID_SET(channel->tsin_id));
191
192         } else {
193                 bitmap_set(bitmap, dvbdmxfeed->pid, 1);
194         }
195
196         /* manage cache so PID bitmap is visible to HW */
197         dma_sync_single_for_device(fei->dev,
198                                         channel->pid_buffer_busaddr,
199                                         PID_TABLE_SIZE,
200                                         DMA_TO_DEVICE);
201
202         channel->active = 1;
203
204         if (fei->global_feed_count == 0) {
205                 fei->timer.expires = jiffies +
206                         msecs_to_jiffies(msecs_to_jiffies(POLL_MSECS));
207
208                 add_timer(&fei->timer);
209         }
210
211         if (stdemux->running_feed_count == 0) {
212
213                 dev_dbg(fei->dev, "Starting channel=%p\n", channel);
214
215                 tasklet_init(&channel->tsklet, channel_swdemux_tsklet,
216                              (unsigned long) channel);
217
218                 /* Reset the internal inputblock sram pointers */
219                 writel(channel->fifo,
220                         fei->io + C8SECTPFE_IB_BUFF_STRT(channel->tsin_id));
221                 writel(channel->fifo + FIFO_LEN - 1,
222                         fei->io + C8SECTPFE_IB_BUFF_END(channel->tsin_id));
223
224                 writel(channel->fifo,
225                         fei->io + C8SECTPFE_IB_READ_PNT(channel->tsin_id));
226                 writel(channel->fifo,
227                         fei->io + C8SECTPFE_IB_WRT_PNT(channel->tsin_id));
228
229
230                 /* reset read / write memdma ptrs for this channel */
231                 writel(channel->back_buffer_busaddr, channel->irec +
232                         DMA_PRDS_BUSBASE_TP(0));
233
234                 tmp = channel->back_buffer_busaddr + FEI_BUFFER_SIZE - 1;
235                 writel(tmp, channel->irec + DMA_PRDS_BUSTOP_TP(0));
236
237                 writel(channel->back_buffer_busaddr, channel->irec +
238                         DMA_PRDS_BUSWP_TP(0));
239
240                 /* Issue a reset and enable InputBlock */
241                 writel(C8SECTPFE_SYS_ENABLE | C8SECTPFE_SYS_RESET
242                         , fei->io + C8SECTPFE_IB_SYS(channel->tsin_id));
243
244                 /* and enable the tp */
245                 writel(0x1, channel->irec + DMA_PRDS_TPENABLE);
246
247                 dev_dbg(fei->dev, "%s:%d Starting DMA feed on stdemux=%p\n"
248                         , __func__, __LINE__, stdemux);
249         }
250
251         stdemux->running_feed_count++;
252         fei->global_feed_count++;
253
254         mutex_unlock(&fei->lock);
255
256         return 0;
257 }
258
259 static int c8sectpfe_stop_feed(struct dvb_demux_feed *dvbdmxfeed)
260 {
261
262         struct dvb_demux *demux = dvbdmxfeed->demux;
263         struct stdemux *stdemux = (struct stdemux *)demux->priv;
264         struct c8sectpfei *fei = stdemux->c8sectpfei;
265         struct channel_info *channel;
266         int idlereq;
267         u32 tmp;
268         int ret;
269         unsigned long *bitmap;
270
271         if (!atomic_read(&fei->fw_loaded)) {
272                 ret = load_c8sectpfe_fw(fei);
273                 if (ret)
274                         return ret;
275         }
276
277         mutex_lock(&fei->lock);
278
279         channel = fei->channel_data[stdemux->tsin_index];
280
281         bitmap = (unsigned long *) channel->pid_buffer_aligned;
282
283         if (dvbdmxfeed->pid == 8192) {
284                 tmp = readl(fei->io + C8SECTPFE_IB_PID_SET(channel->tsin_id));
285                 tmp |= C8SECTPFE_PID_ENABLE;
286                 writel(tmp, fei->io + C8SECTPFE_IB_PID_SET(channel->tsin_id));
287         } else {
288                 bitmap_clear(bitmap, dvbdmxfeed->pid, 1);
289         }
290
291         /* manage cache so data is visible to HW */
292         dma_sync_single_for_device(fei->dev,
293                                         channel->pid_buffer_busaddr,
294                                         PID_TABLE_SIZE,
295                                         DMA_TO_DEVICE);
296
297         if (--stdemux->running_feed_count == 0) {
298
299                 channel = fei->channel_data[stdemux->tsin_index];
300
301                 /* TP re-configuration on page 168 of functional spec */
302
303                 /* disable IB (prevents more TS data going to memdma) */
304                 writel(0, fei->io + C8SECTPFE_IB_SYS(channel->tsin_id));
305
306                 /* disable this channels descriptor */
307                 writel(0,  channel->irec + DMA_PRDS_TPENABLE);
308
309                 tasklet_disable(&channel->tsklet);
310
311                 /* now request memdma channel goes idle */
312                 idlereq = (1 << channel->tsin_id) | IDLEREQ;
313                 writel(idlereq, fei->io + DMA_IDLE_REQ);
314
315                 /* wait for idle irq handler to signal completion */
316                 ret = wait_for_completion_timeout(&channel->idle_completion,
317                                                 msecs_to_jiffies(100));
318
319                 if (ret == 0)
320                         dev_warn(fei->dev,
321                                 "Timeout waiting for idle irq on tsin%d\n",
322                                 channel->tsin_id);
323
324                 reinit_completion(&channel->idle_completion);
325
326                 /* reset read / write ptrs for this channel */
327
328                 writel(channel->back_buffer_busaddr,
329                         channel->irec + DMA_PRDS_BUSBASE_TP(0));
330
331                 tmp = channel->back_buffer_busaddr + FEI_BUFFER_SIZE - 1;
332                 writel(tmp, channel->irec + DMA_PRDS_BUSTOP_TP(0));
333
334                 writel(channel->back_buffer_busaddr,
335                         channel->irec + DMA_PRDS_BUSWP_TP(0));
336
337                 dev_dbg(fei->dev,
338                         "%s:%d stopping DMA feed on stdemux=%p channel=%d\n",
339                         __func__, __LINE__, stdemux, channel->tsin_id);
340
341                 /* turn off all PIDS in the bitmap */
342                 memset((void *)channel->pid_buffer_aligned
343                         , 0x00, PID_TABLE_SIZE);
344
345                 /* manage cache so data is visible to HW */
346                 dma_sync_single_for_device(fei->dev,
347                                         channel->pid_buffer_busaddr,
348                                         PID_TABLE_SIZE,
349                                         DMA_TO_DEVICE);
350
351                 channel->active = 0;
352         }
353
354         if (--fei->global_feed_count == 0) {
355                 dev_dbg(fei->dev, "%s:%d global_feed_count=%d\n"
356                         , __func__, __LINE__, fei->global_feed_count);
357
358                 del_timer(&fei->timer);
359         }
360
361         mutex_unlock(&fei->lock);
362
363         return 0;
364 }
365
366 static struct channel_info *find_channel(struct c8sectpfei *fei, int tsin_num)
367 {
368         int i;
369
370         for (i = 0; i < C8SECTPFE_MAX_TSIN_CHAN; i++) {
371                 if (!fei->channel_data[i])
372                         continue;
373
374                 if (fei->channel_data[i]->tsin_id == tsin_num)
375                         return fei->channel_data[i];
376         }
377
378         return NULL;
379 }
380
381 static void c8sectpfe_getconfig(struct c8sectpfei *fei)
382 {
383         struct c8sectpfe_hw *hw = &fei->hw_stats;
384
385         hw->num_ib = readl(fei->io + SYS_CFG_NUM_IB);
386         hw->num_mib = readl(fei->io + SYS_CFG_NUM_MIB);
387         hw->num_swts = readl(fei->io + SYS_CFG_NUM_SWTS);
388         hw->num_tsout = readl(fei->io + SYS_CFG_NUM_TSOUT);
389         hw->num_ccsc = readl(fei->io + SYS_CFG_NUM_CCSC);
390         hw->num_ram = readl(fei->io + SYS_CFG_NUM_RAM);
391         hw->num_tp = readl(fei->io + SYS_CFG_NUM_TP);
392
393         dev_info(fei->dev, "C8SECTPFE hw supports the following:\n");
394         dev_info(fei->dev, "Input Blocks: %d\n", hw->num_ib);
395         dev_info(fei->dev, "Merged Input Blocks: %d\n", hw->num_mib);
396         dev_info(fei->dev, "Software Transport Stream Inputs: %d\n"
397                                 , hw->num_swts);
398         dev_info(fei->dev, "Transport Stream Output: %d\n", hw->num_tsout);
399         dev_info(fei->dev, "Cable Card Converter: %d\n", hw->num_ccsc);
400         dev_info(fei->dev, "RAMs supported by C8SECTPFE: %d\n", hw->num_ram);
401         dev_info(fei->dev, "Tango TPs supported by C8SECTPFE: %d\n"
402                         , hw->num_tp);
403 }
404
405 static irqreturn_t c8sectpfe_idle_irq_handler(int irq, void *priv)
406 {
407         struct c8sectpfei *fei = priv;
408         struct channel_info *chan;
409         int bit;
410         unsigned long tmp = readl(fei->io + DMA_IDLE_REQ);
411
412         /* page 168 of functional spec: Clear the idle request
413            by writing 0 to the C8SECTPFE_DMA_IDLE_REQ register. */
414
415         /* signal idle completion */
416         for_each_set_bit(bit, &tmp, fei->hw_stats.num_ib) {
417
418                 chan = find_channel(fei, bit);
419
420                 if (chan)
421                         complete(&chan->idle_completion);
422         }
423
424         writel(0, fei->io + DMA_IDLE_REQ);
425
426         return IRQ_HANDLED;
427 }
428
429
430 static void free_input_block(struct c8sectpfei *fei, struct channel_info *tsin)
431 {
432         if (!fei || !tsin)
433                 return;
434
435         if (tsin->back_buffer_busaddr)
436                 if (!dma_mapping_error(fei->dev, tsin->back_buffer_busaddr))
437                         dma_unmap_single(fei->dev, tsin->back_buffer_busaddr,
438                                 FEI_BUFFER_SIZE, DMA_BIDIRECTIONAL);
439
440         kfree(tsin->back_buffer_start);
441
442         if (tsin->pid_buffer_busaddr)
443                 if (!dma_mapping_error(fei->dev, tsin->pid_buffer_busaddr))
444                         dma_unmap_single(fei->dev, tsin->pid_buffer_busaddr,
445                                 PID_TABLE_SIZE, DMA_BIDIRECTIONAL);
446
447         kfree(tsin->pid_buffer_start);
448 }
449
450 #define MAX_NAME 20
451
452 static int configure_memdma_and_inputblock(struct c8sectpfei *fei,
453                                 struct channel_info *tsin)
454 {
455         int ret;
456         u32 tmp;
457         char tsin_pin_name[MAX_NAME];
458
459         if (!fei || !tsin)
460                 return -EINVAL;
461
462         dev_dbg(fei->dev, "%s:%d Configuring channel=%p tsin=%d\n"
463                 , __func__, __LINE__, tsin, tsin->tsin_id);
464
465         init_completion(&tsin->idle_completion);
466
467         tsin->back_buffer_start = kzalloc(FEI_BUFFER_SIZE +
468                                         FEI_ALIGNMENT, GFP_KERNEL);
469
470         if (!tsin->back_buffer_start) {
471                 ret = -ENOMEM;
472                 goto err_unmap;
473         }
474
475         /* Ensure backbuffer is 32byte aligned */
476         tsin->back_buffer_aligned = tsin->back_buffer_start
477                 + FEI_ALIGNMENT;
478
479         tsin->back_buffer_aligned = (void *)
480                 (((uintptr_t) tsin->back_buffer_aligned) & ~0x1F);
481
482         tsin->back_buffer_busaddr = dma_map_single(fei->dev,
483                                         (void *)tsin->back_buffer_aligned,
484                                         FEI_BUFFER_SIZE,
485                                         DMA_BIDIRECTIONAL);
486
487         if (dma_mapping_error(fei->dev, tsin->back_buffer_busaddr)) {
488                 dev_err(fei->dev, "failed to map back_buffer\n");
489                 ret = -EFAULT;
490                 goto err_unmap;
491         }
492
493         /*
494          * The pid buffer can be configured (in hw) for byte or bit
495          * per pid. By powers of deduction we conclude stih407 family
496          * is configured (at SoC design stage) for bit per pid.
497          */
498         tsin->pid_buffer_start = kzalloc(2048, GFP_KERNEL);
499
500         if (!tsin->pid_buffer_start) {
501                 ret = -ENOMEM;
502                 goto err_unmap;
503         }
504
505         /*
506          * PID buffer needs to be aligned to size of the pid table
507          * which at bit per pid is 1024 bytes (8192 pids / 8).
508          * PIDF_BASE register enforces this alignment when writing
509          * the register.
510          */
511
512         tsin->pid_buffer_aligned = tsin->pid_buffer_start +
513                 PID_TABLE_SIZE;
514
515         tsin->pid_buffer_aligned = (void *)
516                 (((uintptr_t) tsin->pid_buffer_aligned) & ~0x3ff);
517
518         tsin->pid_buffer_busaddr = dma_map_single(fei->dev,
519                                                 tsin->pid_buffer_aligned,
520                                                 PID_TABLE_SIZE,
521                                                 DMA_BIDIRECTIONAL);
522
523         if (dma_mapping_error(fei->dev, tsin->pid_buffer_busaddr)) {
524                 dev_err(fei->dev, "failed to map pid_bitmap\n");
525                 ret = -EFAULT;
526                 goto err_unmap;
527         }
528
529         /* manage cache so pid bitmap is visible to HW */
530         dma_sync_single_for_device(fei->dev,
531                                 tsin->pid_buffer_busaddr,
532                                 PID_TABLE_SIZE,
533                                 DMA_TO_DEVICE);
534
535         snprintf(tsin_pin_name, MAX_NAME, "tsin%d-%s", tsin->tsin_id,
536                 (tsin->serial_not_parallel ? "serial" : "parallel"));
537
538         tsin->pstate = pinctrl_lookup_state(fei->pinctrl, tsin_pin_name);
539         if (IS_ERR(tsin->pstate)) {
540                 dev_err(fei->dev, "%s: pinctrl_lookup_state couldn't find %s state\n"
541                         , __func__, tsin_pin_name);
542                 ret = PTR_ERR(tsin->pstate);
543                 goto err_unmap;
544         }
545
546         ret = pinctrl_select_state(fei->pinctrl, tsin->pstate);
547
548         if (ret) {
549                 dev_err(fei->dev, "%s: pinctrl_select_state failed\n"
550                         , __func__);
551                 goto err_unmap;
552         }
553
554         /* Enable this input block */
555         tmp = readl(fei->io + SYS_INPUT_CLKEN);
556         tmp |= BIT(tsin->tsin_id);
557         writel(tmp, fei->io + SYS_INPUT_CLKEN);
558
559         if (tsin->serial_not_parallel)
560                 tmp |= C8SECTPFE_SERIAL_NOT_PARALLEL;
561
562         if (tsin->invert_ts_clk)
563                 tmp |= C8SECTPFE_INVERT_TSCLK;
564
565         if (tsin->async_not_sync)
566                 tmp |= C8SECTPFE_ASYNC_NOT_SYNC;
567
568         tmp |= C8SECTPFE_ALIGN_BYTE_SOP | C8SECTPFE_BYTE_ENDIANNESS_MSB;
569
570         writel(tmp, fei->io + C8SECTPFE_IB_IP_FMT_CFG(tsin->tsin_id));
571
572         writel(C8SECTPFE_SYNC(0x9) |
573                 C8SECTPFE_DROP(0x9) |
574                 C8SECTPFE_TOKEN(0x47),
575                 fei->io + C8SECTPFE_IB_SYNCLCKDRP_CFG(tsin->tsin_id));
576
577         writel(TS_PKT_SIZE, fei->io + C8SECTPFE_IB_PKT_LEN(tsin->tsin_id));
578
579         /* Place the FIFO's at the end of the irec descriptors */
580
581         tsin->fifo = (tsin->tsin_id * FIFO_LEN);
582
583         writel(tsin->fifo, fei->io + C8SECTPFE_IB_BUFF_STRT(tsin->tsin_id));
584         writel(tsin->fifo + FIFO_LEN - 1,
585                 fei->io + C8SECTPFE_IB_BUFF_END(tsin->tsin_id));
586
587         writel(tsin->fifo, fei->io + C8SECTPFE_IB_READ_PNT(tsin->tsin_id));
588         writel(tsin->fifo, fei->io + C8SECTPFE_IB_WRT_PNT(tsin->tsin_id));
589
590         writel(tsin->pid_buffer_busaddr,
591                 fei->io + PIDF_BASE(tsin->tsin_id));
592
593         dev_dbg(fei->dev, "chan=%d PIDF_BASE=0x%x pid_bus_addr=%pad\n",
594                 tsin->tsin_id, readl(fei->io + PIDF_BASE(tsin->tsin_id)),
595                 &tsin->pid_buffer_busaddr);
596
597         /* Configure and enable HW PID filtering */
598
599         /*
600          * The PID value is created by assembling the first 8 bytes of
601          * the TS packet into a 64-bit word in big-endian format. A
602          * slice of that 64-bit word is taken from
603          * (PID_OFFSET+PID_NUM_BITS-1) to PID_OFFSET.
604          */
605         tmp = (C8SECTPFE_PID_ENABLE | C8SECTPFE_PID_NUMBITS(13)
606                 | C8SECTPFE_PID_OFFSET(40));
607
608         writel(tmp, fei->io + C8SECTPFE_IB_PID_SET(tsin->tsin_id));
609
610         dev_dbg(fei->dev, "chan=%d setting wp: %d, rp: %d, buf: %d-%d\n",
611                 tsin->tsin_id,
612                 readl(fei->io + C8SECTPFE_IB_WRT_PNT(tsin->tsin_id)),
613                 readl(fei->io + C8SECTPFE_IB_READ_PNT(tsin->tsin_id)),
614                 readl(fei->io + C8SECTPFE_IB_BUFF_STRT(tsin->tsin_id)),
615                 readl(fei->io + C8SECTPFE_IB_BUFF_END(tsin->tsin_id)));
616
617         /* Get base addpress of pointer record block from DMEM */
618         tsin->irec = fei->io + DMA_MEMDMA_OFFSET + DMA_DMEM_OFFSET +
619                         readl(fei->io + DMA_PTRREC_BASE);
620
621         /* fill out pointer record data structure */
622
623         /* advance pointer record block to our channel */
624         tsin->irec += (tsin->tsin_id * DMA_PRDS_SIZE);
625
626         writel(tsin->fifo, tsin->irec + DMA_PRDS_MEMBASE);
627
628         writel(tsin->fifo + FIFO_LEN - 1, tsin->irec + DMA_PRDS_MEMTOP);
629
630         writel((188 + 7)&~7, tsin->irec + DMA_PRDS_PKTSIZE);
631
632         writel(0x1, tsin->irec + DMA_PRDS_TPENABLE);
633
634         /* read/write pointers with physical bus address */
635
636         writel(tsin->back_buffer_busaddr, tsin->irec + DMA_PRDS_BUSBASE_TP(0));
637
638         tmp = tsin->back_buffer_busaddr + FEI_BUFFER_SIZE - 1;
639         writel(tmp, tsin->irec + DMA_PRDS_BUSTOP_TP(0));
640
641         writel(tsin->back_buffer_busaddr, tsin->irec + DMA_PRDS_BUSWP_TP(0));
642         writel(tsin->back_buffer_busaddr, tsin->irec + DMA_PRDS_BUSRP_TP(0));
643
644         /* initialize tasklet */
645         tasklet_init(&tsin->tsklet, channel_swdemux_tsklet,
646                 (unsigned long) tsin);
647
648         return 0;
649
650 err_unmap:
651         free_input_block(fei, tsin);
652         return ret;
653 }
654
655 static irqreturn_t c8sectpfe_error_irq_handler(int irq, void *priv)
656 {
657         struct c8sectpfei *fei = priv;
658
659         dev_err(fei->dev, "%s: error handling not yet implemented\n"
660                 , __func__);
661
662         /*
663          * TODO FIXME we should detect some error conditions here
664          * and ideally so something about them!
665          */
666
667         return IRQ_HANDLED;
668 }
669
670 static int c8sectpfe_probe(struct platform_device *pdev)
671 {
672         struct device *dev = &pdev->dev;
673         struct device_node *child, *np = dev->of_node;
674         struct c8sectpfei *fei;
675         struct resource *res;
676         int ret, index = 0;
677         struct channel_info *tsin;
678
679         /* Allocate the c8sectpfei structure */
680         fei = devm_kzalloc(dev, sizeof(struct c8sectpfei), GFP_KERNEL);
681         if (!fei)
682                 return -ENOMEM;
683
684         fei->dev = dev;
685
686         res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "c8sectpfe");
687         fei->io = devm_ioremap_resource(dev, res);
688         if (IS_ERR(fei->io))
689                 return PTR_ERR(fei->io);
690
691         res = platform_get_resource_byname(pdev, IORESOURCE_MEM,
692                                         "c8sectpfe-ram");
693         fei->sram = devm_ioremap_resource(dev, res);
694         if (IS_ERR(fei->sram))
695                 return PTR_ERR(fei->sram);
696
697         fei->sram_size = res->end - res->start;
698
699         fei->idle_irq = platform_get_irq_byname(pdev, "c8sectpfe-idle-irq");
700         if (fei->idle_irq < 0) {
701                 dev_err(dev, "Can't get c8sectpfe-idle-irq\n");
702                 return fei->idle_irq;
703         }
704
705         fei->error_irq = platform_get_irq_byname(pdev, "c8sectpfe-error-irq");
706         if (fei->error_irq < 0) {
707                 dev_err(dev, "Can't get c8sectpfe-error-irq\n");
708                 return fei->error_irq;
709         }
710
711         platform_set_drvdata(pdev, fei);
712
713         fei->c8sectpfeclk = devm_clk_get(dev, "c8sectpfe");
714         if (IS_ERR(fei->c8sectpfeclk)) {
715                 dev_err(dev, "c8sectpfe clk not found\n");
716                 return PTR_ERR(fei->c8sectpfeclk);
717         }
718
719         ret = clk_prepare_enable(fei->c8sectpfeclk);
720         if (ret) {
721                 dev_err(dev, "Failed to enable c8sectpfe clock\n");
722                 return ret;
723         }
724
725         /* to save power disable all IP's (on by default) */
726         writel(0, fei->io + SYS_INPUT_CLKEN);
727
728         /* Enable memdma clock */
729         writel(MEMDMAENABLE, fei->io + SYS_OTHER_CLKEN);
730
731         /* clear internal sram */
732         memset_io(fei->sram, 0x0, fei->sram_size);
733
734         c8sectpfe_getconfig(fei);
735
736         ret = devm_request_irq(dev, fei->idle_irq, c8sectpfe_idle_irq_handler,
737                         0, "c8sectpfe-idle-irq", fei);
738         if (ret) {
739                 dev_err(dev, "Can't register c8sectpfe-idle-irq IRQ.\n");
740                 goto err_clk_disable;
741         }
742
743         ret = devm_request_irq(dev, fei->error_irq,
744                                 c8sectpfe_error_irq_handler, 0,
745                                 "c8sectpfe-error-irq", fei);
746         if (ret) {
747                 dev_err(dev, "Can't register c8sectpfe-error-irq IRQ.\n");
748                 goto err_clk_disable;
749         }
750
751         fei->tsin_count = of_get_child_count(np);
752
753         if (fei->tsin_count > C8SECTPFE_MAX_TSIN_CHAN ||
754                 fei->tsin_count > fei->hw_stats.num_ib) {
755
756                 dev_err(dev, "More tsin declared than exist on SoC!\n");
757                 ret = -EINVAL;
758                 goto err_clk_disable;
759         }
760
761         fei->pinctrl = devm_pinctrl_get(dev);
762
763         if (IS_ERR(fei->pinctrl)) {
764                 dev_err(dev, "Error getting tsin pins\n");
765                 ret = PTR_ERR(fei->pinctrl);
766                 goto err_clk_disable;
767         }
768
769         for_each_child_of_node(np, child) {
770                 struct device_node *i2c_bus;
771
772                 fei->channel_data[index] = devm_kzalloc(dev,
773                                                 sizeof(struct channel_info),
774                                                 GFP_KERNEL);
775
776                 if (!fei->channel_data[index]) {
777                         ret = -ENOMEM;
778                         goto err_clk_disable;
779                 }
780
781                 tsin = fei->channel_data[index];
782
783                 tsin->fei = fei;
784
785                 ret = of_property_read_u32(child, "tsin-num", &tsin->tsin_id);
786                 if (ret) {
787                         dev_err(&pdev->dev, "No tsin_num found\n");
788                         goto err_clk_disable;
789                 }
790
791                 /* sanity check value */
792                 if (tsin->tsin_id > fei->hw_stats.num_ib) {
793                         dev_err(&pdev->dev,
794                                 "tsin-num %d specified greater than number\n\t"
795                                 "of input block hw in SoC! (%d)",
796                                 tsin->tsin_id, fei->hw_stats.num_ib);
797                         ret = -EINVAL;
798                         goto err_clk_disable;
799                 }
800
801                 tsin->invert_ts_clk = of_property_read_bool(child,
802                                                         "invert-ts-clk");
803
804                 tsin->serial_not_parallel = of_property_read_bool(child,
805                                                         "serial-not-parallel");
806
807                 tsin->async_not_sync = of_property_read_bool(child,
808                                                         "async-not-sync");
809
810                 ret = of_property_read_u32(child, "dvb-card",
811                                         &tsin->dvb_card);
812                 if (ret) {
813                         dev_err(&pdev->dev, "No dvb-card found\n");
814                         goto err_clk_disable;
815                 }
816
817                 i2c_bus = of_parse_phandle(child, "i2c-bus", 0);
818                 if (!i2c_bus) {
819                         dev_err(&pdev->dev, "No i2c-bus found\n");
820                         goto err_clk_disable;
821                 }
822                 tsin->i2c_adapter =
823                         of_find_i2c_adapter_by_node(i2c_bus);
824                 if (!tsin->i2c_adapter) {
825                         dev_err(&pdev->dev, "No i2c adapter found\n");
826                         of_node_put(i2c_bus);
827                         goto err_clk_disable;
828                 }
829                 of_node_put(i2c_bus);
830
831                 tsin->rst_gpio = of_get_named_gpio(child, "reset-gpios", 0);
832
833                 ret = gpio_is_valid(tsin->rst_gpio);
834                 if (!ret) {
835                         dev_err(dev,
836                                 "reset gpio for tsin%d not valid (gpio=%d)\n",
837                                 tsin->tsin_id, tsin->rst_gpio);
838                         goto err_clk_disable;
839                 }
840
841                 ret = devm_gpio_request_one(dev, tsin->rst_gpio,
842                                         GPIOF_OUT_INIT_LOW, "NIM reset");
843                 if (ret && ret != -EBUSY) {
844                         dev_err(dev, "Can't request tsin%d reset gpio\n"
845                                 , fei->channel_data[index]->tsin_id);
846                         goto err_clk_disable;
847                 }
848
849                 if (!ret) {
850                         /* toggle reset lines */
851                         gpio_direction_output(tsin->rst_gpio, 0);
852                         usleep_range(3500, 5000);
853                         gpio_direction_output(tsin->rst_gpio, 1);
854                         usleep_range(3000, 5000);
855                 }
856
857                 tsin->demux_mapping = index;
858
859                 dev_dbg(fei->dev,
860                         "channel=%p n=%d tsin_num=%d, invert-ts-clk=%d\n\t"
861                         "serial-not-parallel=%d pkt-clk-valid=%d dvb-card=%d\n",
862                         fei->channel_data[index], index,
863                         tsin->tsin_id, tsin->invert_ts_clk,
864                         tsin->serial_not_parallel, tsin->async_not_sync,
865                         tsin->dvb_card);
866
867                 index++;
868         }
869
870         /* Setup timer interrupt */
871         init_timer(&fei->timer);
872         fei->timer.function = c8sectpfe_timer_interrupt;
873         fei->timer.data = (unsigned long)fei;
874
875         mutex_init(&fei->lock);
876
877         /* Get the configuration information about the tuners */
878         ret = c8sectpfe_tuner_register_frontend(&fei->c8sectpfe[0],
879                                         (void *)fei,
880                                         c8sectpfe_start_feed,
881                                         c8sectpfe_stop_feed);
882         if (ret) {
883                 dev_err(dev, "c8sectpfe_tuner_register_frontend failed (%d)\n",
884                         ret);
885                 goto err_clk_disable;
886         }
887
888         c8sectpfe_debugfs_init(fei);
889
890         return 0;
891
892 err_clk_disable:
893         /* TODO uncomment when upstream has taken a reference on this clk */
894         /*clk_disable_unprepare(fei->c8sectpfeclk);*/
895         return ret;
896 }
897
898 static int c8sectpfe_remove(struct platform_device *pdev)
899 {
900         struct c8sectpfei *fei = platform_get_drvdata(pdev);
901         struct channel_info *channel;
902         int i;
903
904         wait_for_completion(&fei->fw_ack);
905
906         c8sectpfe_tuner_unregister_frontend(fei->c8sectpfe[0], fei);
907
908         /*
909          * Now loop through and un-configure each of the InputBlock resources
910          */
911         for (i = 0; i < fei->tsin_count; i++) {
912                 channel = fei->channel_data[i];
913                 free_input_block(fei, channel);
914         }
915
916         c8sectpfe_debugfs_exit(fei);
917
918         dev_info(fei->dev, "Stopping memdma SLIM core\n");
919         if (readl(fei->io + DMA_CPU_RUN))
920                 writel(0x0,  fei->io + DMA_CPU_RUN);
921
922         /* unclock all internal IP's */
923         if (readl(fei->io + SYS_INPUT_CLKEN))
924                 writel(0, fei->io + SYS_INPUT_CLKEN);
925
926         if (readl(fei->io + SYS_OTHER_CLKEN))
927                 writel(0, fei->io + SYS_OTHER_CLKEN);
928
929         /* TODO uncomment when upstream has taken a reference on this clk */
930         /*
931         if (fei->c8sectpfeclk)
932                 clk_disable_unprepare(fei->c8sectpfeclk);
933         */
934
935         return 0;
936 }
937
938
939 static int configure_channels(struct c8sectpfei *fei)
940 {
941         int index = 0, ret;
942         struct channel_info *tsin;
943         struct device_node *child, *np = fei->dev->of_node;
944
945         /* iterate round each tsin and configure memdma descriptor and IB hw */
946         for_each_child_of_node(np, child) {
947
948                 tsin = fei->channel_data[index];
949
950                 ret = configure_memdma_and_inputblock(fei,
951                                                 fei->channel_data[index]);
952
953                 if (ret) {
954                         dev_err(fei->dev,
955                                 "configure_memdma_and_inputblock failed\n");
956                         goto err_unmap;
957                 }
958                 index++;
959         }
960
961         return 0;
962
963 err_unmap:
964         for (index = 0; index < fei->tsin_count; index++) {
965                 tsin = fei->channel_data[index];
966                 free_input_block(fei, tsin);
967         }
968         return ret;
969 }
970
971 static int
972 c8sectpfe_elf_sanity_check(struct c8sectpfei *fei, const struct firmware *fw)
973 {
974         struct elf32_hdr *ehdr;
975         char class;
976
977         if (!fw) {
978                 dev_err(fei->dev, "failed to load %s\n", FIRMWARE_MEMDMA);
979                 return -EINVAL;
980         }
981
982         if (fw->size < sizeof(struct elf32_hdr)) {
983                 dev_err(fei->dev, "Image is too small\n");
984                 return -EINVAL;
985         }
986
987         ehdr = (struct elf32_hdr *)fw->data;
988
989         /* We only support ELF32 at this point */
990         class = ehdr->e_ident[EI_CLASS];
991         if (class != ELFCLASS32) {
992                 dev_err(fei->dev, "Unsupported class: %d\n", class);
993                 return -EINVAL;
994         }
995
996         if (ehdr->e_ident[EI_DATA] != ELFDATA2LSB) {
997                 dev_err(fei->dev, "Unsupported firmware endianness\n");
998                 return -EINVAL;
999         }
1000
1001         if (fw->size < ehdr->e_shoff + sizeof(struct elf32_shdr)) {
1002                 dev_err(fei->dev, "Image is too small\n");
1003                 return -EINVAL;
1004         }
1005
1006         if (memcmp(ehdr->e_ident, ELFMAG, SELFMAG)) {
1007                 dev_err(fei->dev, "Image is corrupted (bad magic)\n");
1008                 return -EINVAL;
1009         }
1010
1011         /* Check ELF magic */
1012         ehdr = (Elf32_Ehdr *)fw->data;
1013         if (ehdr->e_ident[EI_MAG0] != ELFMAG0 ||
1014             ehdr->e_ident[EI_MAG1] != ELFMAG1 ||
1015             ehdr->e_ident[EI_MAG2] != ELFMAG2 ||
1016             ehdr->e_ident[EI_MAG3] != ELFMAG3) {
1017                 dev_err(fei->dev, "Invalid ELF magic\n");
1018                 return -EINVAL;
1019         }
1020
1021         if (ehdr->e_type != ET_EXEC) {
1022                 dev_err(fei->dev, "Unsupported ELF header type\n");
1023                 return -EINVAL;
1024         }
1025
1026         if (ehdr->e_phoff > fw->size) {
1027                 dev_err(fei->dev, "Firmware size is too small\n");
1028                 return -EINVAL;
1029         }
1030
1031         return 0;
1032 }
1033
1034
1035 static void load_imem_segment(struct c8sectpfei *fei, Elf32_Phdr *phdr,
1036                         const struct firmware *fw, u8 __iomem *dest,
1037                         int seg_num)
1038 {
1039         const u8 *imem_src = fw->data + phdr->p_offset;
1040         int i;
1041
1042         /*
1043          * For IMEM segments, the segment contains 24-bit
1044          * instructions which must be padded to 32-bit
1045          * instructions before being written. The written
1046          * segment is padded with NOP instructions.
1047          */
1048
1049         dev_dbg(fei->dev,
1050                 "Loading IMEM segment %d 0x%08x\n\t"
1051                 " (0x%x bytes) -> 0x%p (0x%x bytes)\n", seg_num,
1052                 phdr->p_paddr, phdr->p_filesz,
1053                 dest, phdr->p_memsz + phdr->p_memsz / 3);
1054
1055         for (i = 0; i < phdr->p_filesz; i++) {
1056
1057                 writeb(readb((void __iomem *)imem_src), (void __iomem *)dest);
1058
1059                 /* Every 3 bytes, add an additional
1060                  * padding zero in destination */
1061                 if (i % 3 == 2) {
1062                         dest++;
1063                         writeb(0x00, (void __iomem *)dest);
1064                 }
1065
1066                 dest++;
1067                 imem_src++;
1068         }
1069 }
1070
1071 static void load_dmem_segment(struct c8sectpfei *fei, Elf32_Phdr *phdr,
1072                         const struct firmware *fw, u8 __iomem *dst, int seg_num)
1073 {
1074         /*
1075          * For DMEM segments copy the segment data from the ELF
1076          * file and pad segment with zeroes
1077          */
1078
1079         dev_dbg(fei->dev,
1080                 "Loading DMEM segment %d 0x%08x\n\t"
1081                 "(0x%x bytes) -> 0x%p (0x%x bytes)\n",
1082                 seg_num, phdr->p_paddr, phdr->p_filesz,
1083                 dst, phdr->p_memsz);
1084
1085         memcpy((void __force *)dst, (void *)fw->data + phdr->p_offset,
1086                 phdr->p_filesz);
1087
1088         memset((void __force *)dst + phdr->p_filesz, 0,
1089                 phdr->p_memsz - phdr->p_filesz);
1090 }
1091
1092 static int load_slim_core_fw(const struct firmware *fw, struct c8sectpfei *fei)
1093 {
1094         Elf32_Ehdr *ehdr;
1095         Elf32_Phdr *phdr;
1096         u8 __iomem *dst;
1097         int err = 0, i;
1098
1099         if (!fw || !fei)
1100                 return -EINVAL;
1101
1102         ehdr = (Elf32_Ehdr *)fw->data;
1103         phdr = (Elf32_Phdr *)(fw->data + ehdr->e_phoff);
1104
1105         /* go through the available ELF segments */
1106         for (i = 0; i < ehdr->e_phnum; i++, phdr++) {
1107
1108                 /* Only consider LOAD segments */
1109                 if (phdr->p_type != PT_LOAD)
1110                         continue;
1111
1112                 /*
1113                  * Check segment is contained within the fw->data buffer
1114                  */
1115                 if (phdr->p_offset + phdr->p_filesz > fw->size) {
1116                         dev_err(fei->dev,
1117                                 "Segment %d is outside of firmware file\n", i);
1118                         err = -EINVAL;
1119                         break;
1120                 }
1121
1122                 /*
1123                  * MEMDMA IMEM has executable flag set, otherwise load
1124                  * this segment into DMEM.
1125                  *
1126                  */
1127
1128                 if (phdr->p_flags & PF_X) {
1129                         dst = (u8 __iomem *) fei->io + DMA_MEMDMA_IMEM;
1130                         /*
1131                          * The Slim ELF file uses 32-bit word addressing for
1132                          * load offsets.
1133                          */
1134                         dst += (phdr->p_paddr & 0xFFFFF) * sizeof(unsigned int);
1135                         load_imem_segment(fei, phdr, fw, dst, i);
1136                 } else {
1137                         dst = (u8 __iomem *) fei->io + DMA_MEMDMA_DMEM;
1138                         /*
1139                          * The Slim ELF file uses 32-bit word addressing for
1140                          * load offsets.
1141                          */
1142                         dst += (phdr->p_paddr & 0xFFFFF) * sizeof(unsigned int);
1143                         load_dmem_segment(fei, phdr, fw, dst, i);
1144                 }
1145         }
1146
1147         release_firmware(fw);
1148         return err;
1149 }
1150
1151 static int load_c8sectpfe_fw(struct c8sectpfei *fei)
1152 {
1153         const struct firmware *fw;
1154         int err;
1155
1156         dev_info(fei->dev, "Loading firmware: %s\n", FIRMWARE_MEMDMA);
1157
1158         err = reject_firmware(&fw, FIRMWARE_MEMDMA, fei->dev);
1159         if (err)
1160                 return err;
1161
1162         err = c8sectpfe_elf_sanity_check(fei, fw);
1163         if (err) {
1164                 dev_err(fei->dev, "c8sectpfe_elf_sanity_check failed err=(%d)\n"
1165                         , err);
1166                 release_firmware(fw);
1167                 return err;
1168         }
1169
1170         err = load_slim_core_fw(fw, fei);
1171         if (err) {
1172                 dev_err(fei->dev, "load_slim_core_fw failed err=(%d)\n", err);
1173                 return err;
1174         }
1175
1176         /* now the firmware is loaded configure the input blocks */
1177         err = configure_channels(fei);
1178         if (err) {
1179                 dev_err(fei->dev, "configure_channels failed err=(%d)\n", err);
1180                 return err;
1181         }
1182
1183         /*
1184          * STBus target port can access IMEM and DMEM ports
1185          * without waiting for CPU
1186          */
1187         writel(0x1, fei->io + DMA_PER_STBUS_SYNC);
1188
1189         dev_info(fei->dev, "Boot the memdma SLIM core\n");
1190         writel(0x1,  fei->io + DMA_CPU_RUN);
1191
1192         atomic_set(&fei->fw_loaded, 1);
1193
1194         return 0;
1195 }
1196
1197 static const struct of_device_id c8sectpfe_match[] = {
1198         { .compatible = "st,stih407-c8sectpfe" },
1199         { /* sentinel */ },
1200 };
1201 MODULE_DEVICE_TABLE(of, c8sectpfe_match);
1202
1203 static struct platform_driver c8sectpfe_driver = {
1204         .driver = {
1205                 .name = "c8sectpfe",
1206                 .of_match_table = of_match_ptr(c8sectpfe_match),
1207         },
1208         .probe  = c8sectpfe_probe,
1209         .remove = c8sectpfe_remove,
1210 };
1211
1212 module_platform_driver(c8sectpfe_driver);
1213
1214 MODULE_AUTHOR("Peter Bennett <peter.bennett@st.com>");
1215 MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>");
1216 MODULE_DESCRIPTION("C8SECTPFE STi DVB Driver");
1217 MODULE_LICENSE("GPL");