1 // SPDX-License-Identifier: GPL-2.0
3 * ISP1362 HCD (Host Controller Driver) for USB.
5 * Copyright (C) 2005 Lothar Wassmann <LW@KARO-electronics.de>
7 * Derived from the SL811 HCD, rewritten for ISP116x.
8 * Copyright (C) 2005 Olav Kongas <ok@artecdesign.ee>
11 * Copyright (C) 2004 Psion Teklogix (for NetBook PRO)
12 * Copyright (C) 2004 David Brownell
16 * The ISP1362 chip requires a large delay (300ns and 462ns) between
17 * accesses to the address and data register.
18 * The following timing options exist:
20 * 1. Configure your memory controller to add such delays if it can (the best)
21 * 2. Implement platform-specific delay function possibly
22 * combined with configuring the memory controller; see
23 * include/linux/usb_isp1362.h for more info.
24 * 3. Use ndelay (easiest, poorest).
26 * Use the corresponding macros USE_PLATFORM_DELAY and USE_NDELAY in the
27 * platform specific section of isp1362.h to select the appropriate variant.
29 * Also note that according to the Philips "ISP1362 Errata" document
30 * Rev 1.00 from 27 May data corruption may occur when the #WR signal
31 * is reasserted (even with #CS deasserted) within 132ns after a
32 * write cycle to any controller register. If the hardware doesn't
33 * implement the recommended fix (gating the #WR with #CS) software
34 * must ensure that no further write cycle (not necessarily to the chip!)
35 * is issued by the CPU within this interval.
37 * For PXA25x this can be ensured by using VLIO with the maximum
38 * recovery time (MSCx = 0x7f8c) with a memory clock of 99.53 MHz.
44 * The PXA255 UDC apparently doesn't handle GET_STATUS, GET_CONFIG and
45 * GET_INTERFACE requests correctly when the SETUP and DATA stages of the
46 * requests are carried out in separate frames. This will delay any SETUP
47 * packets until the start of the next frame so that this situation is
48 * unlikely to occur (and makes usbtest happy running with a PXA255 target
51 #undef BUGGY_PXA2XX_UDC_USBTEST
58 /* This enables a memory test on the ISP1362 chip memory to make sure the
59 * chip access timing is correct.
61 #undef CHIP_BUFFER_TEST
63 #include <linux/module.h>
64 #include <linux/moduleparam.h>
65 #include <linux/kernel.h>
66 #include <linux/delay.h>
67 #include <linux/ioport.h>
68 #include <linux/sched.h>
69 #include <linux/slab.h>
70 #include <linux/errno.h>
71 #include <linux/list.h>
72 #include <linux/interrupt.h>
73 #include <linux/usb.h>
74 #include <linux/usb/isp1362.h>
75 #include <linux/usb/hcd.h>
76 #include <linux/platform_device.h>
79 #include <linux/bitmap.h>
80 #include <linux/prefetch.h>
81 #include <linux/debugfs.h>
82 #include <linux/seq_file.h>
85 #include <asm/byteorder.h>
86 #include <asm/unaligned.h>
90 module_param(dbg_level, int, 0644);
92 module_param(dbg_level, int, 0);
95 #include "../core/usb.h"
99 #define DRIVER_VERSION "2005-04-04"
100 #define DRIVER_DESC "ISP1362 USB Host Controller Driver"
102 MODULE_DESCRIPTION(DRIVER_DESC);
103 MODULE_LICENSE("GPL");
105 static const char hcd_name[] = "isp1362-hcd";
107 static void isp1362_hc_stop(struct usb_hcd *hcd);
108 static int isp1362_hc_start(struct usb_hcd *hcd);
110 /*-------------------------------------------------------------------------*/
113 * When called from the interrupthandler only isp1362_hcd->irqenb is modified,
114 * since the interrupt handler will write isp1362_hcd->irqenb to HCuPINT upon
116 * We don't need a 'disable' counterpart, since interrupts will be disabled
117 * only by the interrupt handler.
119 static inline void isp1362_enable_int(struct isp1362_hcd *isp1362_hcd, u16 mask)
121 if ((isp1362_hcd->irqenb | mask) == isp1362_hcd->irqenb)
123 if (mask & ~isp1362_hcd->irqenb)
124 isp1362_write_reg16(isp1362_hcd, HCuPINT, mask & ~isp1362_hcd->irqenb);
125 isp1362_hcd->irqenb |= mask;
126 if (isp1362_hcd->irq_active)
128 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
131 /*-------------------------------------------------------------------------*/
133 static inline struct isp1362_ep_queue *get_ptd_queue(struct isp1362_hcd *isp1362_hcd,
136 struct isp1362_ep_queue *epq = NULL;
138 if (offset < isp1362_hcd->istl_queue[1].buf_start)
139 epq = &isp1362_hcd->istl_queue[0];
140 else if (offset < isp1362_hcd->intl_queue.buf_start)
141 epq = &isp1362_hcd->istl_queue[1];
142 else if (offset < isp1362_hcd->atl_queue.buf_start)
143 epq = &isp1362_hcd->intl_queue;
144 else if (offset < isp1362_hcd->atl_queue.buf_start +
145 isp1362_hcd->atl_queue.buf_size)
146 epq = &isp1362_hcd->atl_queue;
149 DBG(1, "%s: PTD $%04x is on %s queue\n", __func__, offset, epq->name);
151 pr_warn("%s: invalid PTD $%04x\n", __func__, offset);
156 static inline int get_ptd_offset(struct isp1362_ep_queue *epq, u8 index)
160 if (index * epq->blk_size > epq->buf_size) {
161 pr_warn("%s: Bad %s index %d(%d)\n",
162 __func__, epq->name, index,
163 epq->buf_size / epq->blk_size);
166 offset = epq->buf_start + index * epq->blk_size;
167 DBG(3, "%s: %s PTD[%02x] # %04x\n", __func__, epq->name, index, offset);
172 /*-------------------------------------------------------------------------*/
174 static inline u16 max_transfer_size(struct isp1362_ep_queue *epq, size_t size,
177 u16 xfer_size = min_t(size_t, MAX_XFER_SIZE, size);
179 xfer_size = min_t(size_t, xfer_size, epq->buf_avail * epq->blk_size - PTD_HEADER_SIZE);
180 if (xfer_size < size && xfer_size % mps)
181 xfer_size -= xfer_size % mps;
186 static int claim_ptd_buffers(struct isp1362_ep_queue *epq,
187 struct isp1362_ep *ep, u16 len)
189 int ptd_offset = -EINVAL;
190 int num_ptds = ((len + PTD_HEADER_SIZE - 1) / epq->blk_size) + 1;
193 BUG_ON(len > epq->buf_size);
199 pr_err("%s: %s len %d/%d num_ptds %d buf_map %08lx skip_map %08lx\n", __func__,
200 epq->name, len, epq->blk_size, num_ptds, epq->buf_map, epq->skip_map);
201 BUG_ON(ep->num_ptds != 0);
203 found = bitmap_find_next_zero_area(&epq->buf_map, epq->buf_count, 0,
205 if (found >= epq->buf_count)
208 DBG(1, "%s: Found %d PTDs[%d] for %d/%d byte\n", __func__,
209 num_ptds, found, len, (int)(epq->blk_size - PTD_HEADER_SIZE));
210 ptd_offset = get_ptd_offset(epq, found);
211 WARN_ON(ptd_offset < 0);
212 ep->ptd_offset = ptd_offset;
213 ep->num_ptds += num_ptds;
214 epq->buf_avail -= num_ptds;
215 BUG_ON(epq->buf_avail > epq->buf_count);
216 ep->ptd_index = found;
217 bitmap_set(&epq->buf_map, found, num_ptds);
218 DBG(1, "%s: Done %s PTD[%d] $%04x, avail %d count %d claimed %d %08lx:%08lx\n",
219 __func__, epq->name, ep->ptd_index, ep->ptd_offset,
220 epq->buf_avail, epq->buf_count, num_ptds, epq->buf_map, epq->skip_map);
225 static inline void release_ptd_buffers(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
227 int last = ep->ptd_index + ep->num_ptds;
229 if (last > epq->buf_count)
230 pr_err("%s: ep %p req %d len %d %s PTD[%d] $%04x num_ptds %d buf_count %d buf_avail %d buf_map %08lx skip_map %08lx\n",
231 __func__, ep, ep->num_req, ep->length, epq->name, ep->ptd_index,
232 ep->ptd_offset, ep->num_ptds, epq->buf_count, epq->buf_avail,
233 epq->buf_map, epq->skip_map);
234 BUG_ON(last > epq->buf_count);
236 bitmap_clear(&epq->buf_map, ep->ptd_index, ep->num_ptds);
237 bitmap_set(&epq->skip_map, ep->ptd_index, ep->num_ptds);
238 epq->buf_avail += ep->num_ptds;
241 BUG_ON(epq->buf_avail > epq->buf_count);
242 BUG_ON(epq->ptd_count > epq->buf_count);
244 DBG(1, "%s: Done %s PTDs $%04x released %d avail %d count %d\n",
246 ep->ptd_offset, ep->num_ptds, epq->buf_avail, epq->buf_count);
247 DBG(1, "%s: buf_map %08lx skip_map %08lx\n", __func__,
248 epq->buf_map, epq->skip_map);
251 ep->ptd_offset = -EINVAL;
252 ep->ptd_index = -EINVAL;
255 /*-------------------------------------------------------------------------*/
260 static void prepare_ptd(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
261 struct isp1362_ep *ep, struct isp1362_ep_queue *epq,
268 size_t buf_len = urb->transfer_buffer_length - urb->actual_length;
270 DBG(3, "%s: %s ep %p\n", __func__, epq->name, ep);
274 ep->data = (unsigned char *)urb->transfer_buffer + urb->actual_length;
276 switch (ep->nextpid) {
278 toggle = usb_gettoggle(urb->dev, ep->epnum, 0);
280 if (usb_pipecontrol(urb->pipe)) {
281 len = min_t(size_t, ep->maxpacket, buf_len);
282 } else if (usb_pipeisoc(urb->pipe)) {
283 len = min_t(size_t, urb->iso_frame_desc[fno].length, MAX_XFER_SIZE);
284 ep->data = urb->transfer_buffer + urb->iso_frame_desc[fno].offset;
286 len = max_transfer_size(epq, buf_len, ep->maxpacket);
287 DBG(1, "%s: IN len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
291 toggle = usb_gettoggle(urb->dev, ep->epnum, 1);
293 if (usb_pipecontrol(urb->pipe))
294 len = min_t(size_t, ep->maxpacket, buf_len);
295 else if (usb_pipeisoc(urb->pipe))
296 len = min_t(size_t, urb->iso_frame_desc[0].length, MAX_XFER_SIZE);
298 len = max_transfer_size(epq, buf_len, ep->maxpacket);
300 pr_info("%s: Sending ZERO packet: %d\n", __func__,
301 urb->transfer_flags & URB_ZERO_PACKET);
302 DBG(1, "%s: OUT len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
308 len = sizeof(struct usb_ctrlrequest);
309 DBG(1, "%s: SETUP len %d\n", __func__, len);
310 ep->data = urb->setup_packet;
315 dir = (urb->transfer_buffer_length && usb_pipein(urb->pipe)) ?
316 PTD_DIR_OUT : PTD_DIR_IN;
317 DBG(1, "%s: ACK len %d\n", __func__, len);
320 toggle = dir = len = 0;
321 pr_err("%s@%d: ep->nextpid %02x\n", __func__, __LINE__, ep->nextpid);
329 ptd->count = PTD_CC_MSK | PTD_ACTIVE_MSK | PTD_TOGGLE(toggle);
330 ptd->mps = PTD_MPS(ep->maxpacket) | PTD_SPD(urb->dev->speed == USB_SPEED_LOW) |
332 ptd->len = PTD_LEN(len) | PTD_DIR(dir);
333 ptd->faddr = PTD_FA(usb_pipedevice(urb->pipe));
335 if (usb_pipeint(urb->pipe)) {
336 ptd->faddr |= PTD_SF_INT(ep->branch);
337 ptd->faddr |= PTD_PR(ep->interval ? __ffs(ep->interval) : 0);
339 if (usb_pipeisoc(urb->pipe))
340 ptd->faddr |= PTD_SF_ISO(fno);
342 DBG(1, "%s: Finished\n", __func__);
345 static void isp1362_write_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
346 struct isp1362_ep_queue *epq)
348 struct ptd *ptd = &ep->ptd;
349 int len = PTD_GET_DIR(ptd) == PTD_DIR_IN ? 0 : ep->length;
352 isp1362_write_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
354 isp1362_write_buffer(isp1362_hcd, ep->data,
355 ep->ptd_offset + PTD_HEADER_SIZE, len);
358 dump_ptd_out_data(ptd, ep->data);
361 static void isp1362_read_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
362 struct isp1362_ep_queue *epq)
364 struct ptd *ptd = &ep->ptd;
367 WARN_ON(list_empty(&ep->active));
368 BUG_ON(ep->ptd_offset < 0);
370 list_del_init(&ep->active);
371 DBG(1, "%s: ep %p removed from active list %p\n", __func__, ep, &epq->active);
374 isp1362_read_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
376 act_len = PTD_GET_COUNT(ptd);
377 if (PTD_GET_DIR(ptd) != PTD_DIR_IN || act_len == 0)
379 if (act_len > ep->length)
380 pr_err("%s: ep %p PTD $%04x act_len %d ep->length %d\n", __func__, ep,
381 ep->ptd_offset, act_len, ep->length);
382 BUG_ON(act_len > ep->length);
383 /* Only transfer the amount of data that has actually been overwritten
384 * in the chip buffer. We don't want any data that doesn't belong to the
385 * transfer to leak out of the chip to the callers transfer buffer!
388 isp1362_read_buffer(isp1362_hcd, ep->data,
389 ep->ptd_offset + PTD_HEADER_SIZE, act_len);
390 dump_ptd_in_data(ptd, ep->data);
394 * INT PTDs will stay in the chip until data is available.
395 * This function will remove a PTD from the chip when the URB is dequeued.
396 * Must be called with the spinlock held and IRQs disabled
398 static void remove_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
402 struct isp1362_ep_queue *epq;
404 DBG(1, "%s: ep %p PTD[%d] $%04x\n", __func__, ep, ep->ptd_index, ep->ptd_offset);
405 BUG_ON(ep->ptd_offset < 0);
407 epq = get_ptd_queue(isp1362_hcd, ep->ptd_offset);
410 /* put ep in remove_list for cleanup */
411 WARN_ON(!list_empty(&ep->remove_list));
412 list_add_tail(&ep->remove_list, &isp1362_hcd->remove_list);
413 /* let SOF interrupt handle the cleanup */
414 isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
416 index = ep->ptd_index;
418 /* ISO queues don't have SKIP registers */
421 DBG(1, "%s: Disabling PTD[%02x] $%04x %08lx|%08x\n", __func__,
422 index, ep->ptd_offset, epq->skip_map, 1 << index);
424 /* prevent further processing of PTD (will be effective after next SOF) */
425 epq->skip_map |= 1 << index;
426 if (epq == &isp1362_hcd->atl_queue) {
427 DBG(2, "%s: ATLSKIP = %08x -> %08lx\n", __func__,
428 isp1362_read_reg32(isp1362_hcd, HCATLSKIP), epq->skip_map);
429 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, epq->skip_map);
430 if (~epq->skip_map == 0)
431 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
432 } else if (epq == &isp1362_hcd->intl_queue) {
433 DBG(2, "%s: INTLSKIP = %08x -> %08lx\n", __func__,
434 isp1362_read_reg32(isp1362_hcd, HCINTLSKIP), epq->skip_map);
435 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, epq->skip_map);
436 if (~epq->skip_map == 0)
437 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
442 Take done or failed requests out of schedule. Give back
445 static void finish_request(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
446 struct urb *urb, int status)
447 __releases(isp1362_hcd->lock)
448 __acquires(isp1362_hcd->lock)
453 if (usb_pipecontrol(urb->pipe))
454 ep->nextpid = USB_PID_SETUP;
456 URB_DBG("%s: req %d FA %d ep%d%s %s: len %d/%d %s stat %d\n", __func__,
457 ep->num_req, usb_pipedevice(urb->pipe),
458 usb_pipeendpoint(urb->pipe),
459 !usb_pipein(urb->pipe) ? "out" : "in",
460 usb_pipecontrol(urb->pipe) ? "ctrl" :
461 usb_pipeint(urb->pipe) ? "int" :
462 usb_pipebulk(urb->pipe) ? "bulk" :
464 urb->actual_length, urb->transfer_buffer_length,
465 !(urb->transfer_flags & URB_SHORT_NOT_OK) ?
466 "short_ok" : "", urb->status);
469 usb_hcd_unlink_urb_from_ep(isp1362_hcd_to_hcd(isp1362_hcd), urb);
470 spin_unlock(&isp1362_hcd->lock);
471 usb_hcd_giveback_urb(isp1362_hcd_to_hcd(isp1362_hcd), urb, status);
472 spin_lock(&isp1362_hcd->lock);
474 /* take idle endpoints out of the schedule right away */
475 if (!list_empty(&ep->hep->urb_list))
478 /* async deschedule */
479 if (!list_empty(&ep->schedule)) {
480 list_del_init(&ep->schedule);
486 /* periodic deschedule */
487 DBG(1, "deschedule qh%d/%p branch %d load %d bandwidth %d -> %d\n", ep->interval,
488 ep, ep->branch, ep->load,
489 isp1362_hcd->load[ep->branch],
490 isp1362_hcd->load[ep->branch] - ep->load);
491 isp1362_hcd->load[ep->branch] -= ep->load;
492 ep->branch = PERIODIC_SIZE;
497 * Analyze transfer results, handle partial transfers and errors
499 static void postproc_ep(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
501 struct urb *urb = get_urb(ep);
502 struct usb_device *udev;
506 int urbstat = -EINPROGRESS;
509 DBG(2, "%s: ep %p req %d\n", __func__, ep, ep->num_req);
513 cc = PTD_GET_CC(ptd);
514 if (cc == PTD_NOTACCESSED) {
515 pr_err("%s: req %d PTD %p Untouched by ISP1362\n", __func__,
520 short_ok = !(urb->transfer_flags & URB_SHORT_NOT_OK);
521 len = urb->transfer_buffer_length - urb->actual_length;
523 /* Data underrun is special. For allowed underrun
524 we clear the error and continue as normal. For
525 forbidden underrun we finish the DATA stage
526 immediately while for control transfer,
527 we do a STATUS stage.
529 if (cc == PTD_DATAUNDERRUN) {
531 DBG(1, "%s: req %d Allowed data underrun short_%sok %d/%d/%d byte\n",
532 __func__, ep->num_req, short_ok ? "" : "not_",
533 PTD_GET_COUNT(ptd), ep->maxpacket, len);
537 DBG(1, "%s: req %d Data Underrun %s nextpid %02x short_%sok %d/%d/%d byte\n",
538 __func__, ep->num_req,
539 usb_pipein(urb->pipe) ? "IN" : "OUT", ep->nextpid,
540 short_ok ? "" : "not_",
541 PTD_GET_COUNT(ptd), ep->maxpacket, len);
542 /* save the data underrun error code for later and
543 * proceed with the status stage
545 urb->actual_length += PTD_GET_COUNT(ptd);
546 if (usb_pipecontrol(urb->pipe)) {
547 ep->nextpid = USB_PID_ACK;
548 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
550 if (urb->status == -EINPROGRESS)
551 urb->status = cc_to_error[PTD_DATAUNDERRUN];
553 usb_settoggle(udev, ep->epnum, ep->nextpid == USB_PID_OUT,
554 PTD_GET_TOGGLE(ptd));
555 urbstat = cc_to_error[PTD_DATAUNDERRUN];
561 if (cc != PTD_CC_NOERROR) {
562 if (++ep->error_count >= 3 || cc == PTD_CC_STALL || cc == PTD_DATAOVERRUN) {
563 urbstat = cc_to_error[cc];
564 DBG(1, "%s: req %d nextpid %02x, status %d, error %d, error_count %d\n",
565 __func__, ep->num_req, ep->nextpid, urbstat, cc,
571 switch (ep->nextpid) {
573 if (PTD_GET_COUNT(ptd) != ep->length)
574 pr_err("%s: count=%d len=%d\n", __func__,
575 PTD_GET_COUNT(ptd), ep->length);
576 BUG_ON(PTD_GET_COUNT(ptd) != ep->length);
577 urb->actual_length += ep->length;
578 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
579 usb_settoggle(udev, ep->epnum, 1, PTD_GET_TOGGLE(ptd));
580 if (urb->actual_length == urb->transfer_buffer_length) {
581 DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
582 ep->num_req, len, ep->maxpacket, urbstat);
583 if (usb_pipecontrol(urb->pipe)) {
584 DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
586 usb_pipein(urb->pipe) ? "IN" : "OUT");
587 ep->nextpid = USB_PID_ACK;
589 if (len % ep->maxpacket ||
590 !(urb->transfer_flags & URB_ZERO_PACKET)) {
592 DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
593 __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
594 urbstat, len, ep->maxpacket, urb->actual_length);
600 len = PTD_GET_COUNT(ptd);
601 BUG_ON(len > ep->length);
602 urb->actual_length += len;
603 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
604 usb_settoggle(udev, ep->epnum, 0, PTD_GET_TOGGLE(ptd));
605 /* if transfer completed or (allowed) data underrun */
606 if ((urb->transfer_buffer_length == urb->actual_length) ||
607 len % ep->maxpacket) {
608 DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
609 ep->num_req, len, ep->maxpacket, urbstat);
610 if (usb_pipecontrol(urb->pipe)) {
611 DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
613 usb_pipein(urb->pipe) ? "IN" : "OUT");
614 ep->nextpid = USB_PID_ACK;
617 DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
618 __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
619 urbstat, len, ep->maxpacket, urb->actual_length);
624 if (urb->transfer_buffer_length == urb->actual_length) {
625 ep->nextpid = USB_PID_ACK;
626 } else if (usb_pipeout(urb->pipe)) {
627 usb_settoggle(udev, 0, 1, 1);
628 ep->nextpid = USB_PID_OUT;
630 usb_settoggle(udev, 0, 0, 1);
631 ep->nextpid = USB_PID_IN;
635 DBG(3, "%s: req %d got ACK %d -> 0\n", __func__, ep->num_req,
637 WARN_ON(urbstat != -EINPROGRESS);
646 if (urbstat != -EINPROGRESS) {
647 DBG(2, "%s: Finishing ep %p req %d urb %p status %d\n", __func__,
648 ep, ep->num_req, urb, urbstat);
649 finish_request(isp1362_hcd, ep, urb, urbstat);
653 static void finish_unlinks(struct isp1362_hcd *isp1362_hcd)
655 struct isp1362_ep *ep;
656 struct isp1362_ep *tmp;
658 list_for_each_entry_safe(ep, tmp, &isp1362_hcd->remove_list, remove_list) {
659 struct isp1362_ep_queue *epq =
660 get_ptd_queue(isp1362_hcd, ep->ptd_offset);
661 int index = ep->ptd_index;
665 DBG(1, "%s: remove PTD[%d] $%04x\n", __func__, index, ep->ptd_offset);
666 BUG_ON(ep->num_ptds == 0);
667 release_ptd_buffers(epq, ep);
669 if (!list_empty(&ep->hep->urb_list)) {
670 struct urb *urb = get_urb(ep);
672 DBG(1, "%s: Finishing req %d ep %p from remove_list\n", __func__,
674 finish_request(isp1362_hcd, ep, urb, -ESHUTDOWN);
676 WARN_ON(list_empty(&ep->active));
677 if (!list_empty(&ep->active)) {
678 list_del_init(&ep->active);
679 DBG(1, "%s: ep %p removed from active list\n", __func__, ep);
681 list_del_init(&ep->remove_list);
682 DBG(1, "%s: ep %p removed from remove_list\n", __func__, ep);
684 DBG(1, "%s: Done\n", __func__);
687 static inline void enable_atl_transfers(struct isp1362_hcd *isp1362_hcd, int count)
690 if (count < isp1362_hcd->atl_queue.ptd_count)
691 isp1362_write_reg16(isp1362_hcd, HCATLDTC, count);
692 isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
693 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, isp1362_hcd->atl_queue.skip_map);
694 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
696 isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
699 static inline void enable_intl_transfers(struct isp1362_hcd *isp1362_hcd)
701 isp1362_enable_int(isp1362_hcd, HCuPINT_INTL);
702 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
703 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, isp1362_hcd->intl_queue.skip_map);
706 static inline void enable_istl_transfers(struct isp1362_hcd *isp1362_hcd, int flip)
708 isp1362_enable_int(isp1362_hcd, flip ? HCuPINT_ISTL1 : HCuPINT_ISTL0);
709 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, flip ?
710 HCBUFSTAT_ISTL1_FULL : HCBUFSTAT_ISTL0_FULL);
713 static int submit_req(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
714 struct isp1362_ep *ep, struct isp1362_ep_queue *epq)
718 prepare_ptd(isp1362_hcd, urb, ep, epq, 0);
719 index = claim_ptd_buffers(epq, ep, ep->length);
720 if (index == -ENOMEM) {
721 DBG(1, "%s: req %d No free %s PTD available: %d, %08lx:%08lx\n", __func__,
722 ep->num_req, epq->name, ep->num_ptds, epq->buf_map, epq->skip_map);
724 } else if (index == -EOVERFLOW) {
725 DBG(1, "%s: req %d Not enough space for %d byte %s PTD %d %08lx:%08lx\n",
726 __func__, ep->num_req, ep->length, epq->name, ep->num_ptds,
727 epq->buf_map, epq->skip_map);
731 list_add_tail(&ep->active, &epq->active);
732 DBG(1, "%s: ep %p req %d len %d added to active list %p\n", __func__,
733 ep, ep->num_req, ep->length, &epq->active);
734 DBG(1, "%s: Submitting %s PTD $%04x for ep %p req %d\n", __func__, epq->name,
735 ep->ptd_offset, ep, ep->num_req);
736 isp1362_write_ptd(isp1362_hcd, ep, epq);
737 __clear_bit(ep->ptd_index, &epq->skip_map);
742 static void start_atl_transfers(struct isp1362_hcd *isp1362_hcd)
745 struct isp1362_ep_queue *epq = &isp1362_hcd->atl_queue;
746 struct isp1362_ep *ep;
749 if (atomic_read(&epq->finishing)) {
750 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
754 list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
755 struct urb *urb = get_urb(ep);
758 if (!list_empty(&ep->active)) {
759 DBG(2, "%s: Skipping active %s ep %p\n", __func__, epq->name, ep);
763 DBG(1, "%s: Processing %s ep %p req %d\n", __func__, epq->name,
766 ret = submit_req(isp1362_hcd, urb, ep, epq);
767 if (ret == -ENOMEM) {
770 } else if (ret == -EOVERFLOW) {
774 #ifdef BUGGY_PXA2XX_UDC_USBTEST
775 defer = ep->nextpid == USB_PID_SETUP;
780 /* Avoid starving of endpoints */
781 if (isp1362_hcd->async.next != isp1362_hcd->async.prev) {
782 DBG(2, "%s: Cycling ASYNC schedule %d\n", __func__, ptd_count);
783 list_move(&isp1362_hcd->async, isp1362_hcd->async.next);
785 if (ptd_count || defer)
786 enable_atl_transfers(isp1362_hcd, defer ? 0 : ptd_count);
788 epq->ptd_count += ptd_count;
789 if (epq->ptd_count > epq->stat_maxptds) {
790 epq->stat_maxptds = epq->ptd_count;
791 DBG(0, "%s: max_ptds: %d\n", __func__, epq->stat_maxptds);
795 static void start_intl_transfers(struct isp1362_hcd *isp1362_hcd)
798 struct isp1362_ep_queue *epq = &isp1362_hcd->intl_queue;
799 struct isp1362_ep *ep;
801 if (atomic_read(&epq->finishing)) {
802 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
806 list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
807 struct urb *urb = get_urb(ep);
810 if (!list_empty(&ep->active)) {
811 DBG(1, "%s: Skipping active %s ep %p\n", __func__,
816 DBG(1, "%s: Processing %s ep %p req %d\n", __func__,
817 epq->name, ep, ep->num_req);
818 ret = submit_req(isp1362_hcd, urb, ep, epq);
821 else if (ret == -EOVERFLOW)
827 static int last_count;
829 if (ptd_count != last_count) {
830 DBG(0, "%s: ptd_count: %d\n", __func__, ptd_count);
831 last_count = ptd_count;
833 enable_intl_transfers(isp1362_hcd);
836 epq->ptd_count += ptd_count;
837 if (epq->ptd_count > epq->stat_maxptds)
838 epq->stat_maxptds = epq->ptd_count;
841 static inline int next_ptd(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
843 u16 ptd_offset = ep->ptd_offset;
844 int num_ptds = (ep->length + PTD_HEADER_SIZE + (epq->blk_size - 1)) / epq->blk_size;
846 DBG(2, "%s: PTD offset $%04x + %04x => %d * %04x -> $%04x\n", __func__, ptd_offset,
847 ep->length, num_ptds, epq->blk_size, ptd_offset + num_ptds * epq->blk_size);
849 ptd_offset += num_ptds * epq->blk_size;
850 if (ptd_offset < epq->buf_start + epq->buf_size)
856 static void start_iso_transfers(struct isp1362_hcd *isp1362_hcd)
859 int flip = isp1362_hcd->istl_flip;
860 struct isp1362_ep_queue *epq;
862 struct isp1362_ep *ep;
863 struct isp1362_ep *tmp;
864 u16 fno = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
867 epq = &isp1362_hcd->istl_queue[flip];
868 if (atomic_read(&epq->finishing)) {
869 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
873 if (!list_empty(&epq->active))
876 ptd_offset = epq->buf_start;
877 list_for_each_entry_safe(ep, tmp, &isp1362_hcd->isoc, schedule) {
878 struct urb *urb = get_urb(ep);
879 s16 diff = fno - (u16)urb->start_frame;
881 DBG(1, "%s: Processing %s ep %p\n", __func__, epq->name, ep);
883 if (diff > urb->number_of_packets) {
884 /* time frame for this URB has elapsed */
885 finish_request(isp1362_hcd, ep, urb, -EOVERFLOW);
887 } else if (diff < -1) {
888 /* URB is not due in this frame or the next one.
889 * Comparing with '-1' instead of '0' accounts for double
890 * buffering in the ISP1362 which enables us to queue the PTD
891 * one frame ahead of time
893 } else if (diff == -1) {
894 /* submit PTD's that are due in the next frame */
895 prepare_ptd(isp1362_hcd, urb, ep, epq, fno);
896 if (ptd_offset + PTD_HEADER_SIZE + ep->length >
897 epq->buf_start + epq->buf_size) {
898 pr_err("%s: Not enough ISO buffer space for %d byte PTD\n",
899 __func__, ep->length);
902 ep->ptd_offset = ptd_offset;
903 list_add_tail(&ep->active, &epq->active);
905 ptd_offset = next_ptd(epq, ep);
906 if (ptd_offset < 0) {
907 pr_warn("%s: req %d No more %s PTD buffers available\n",
908 __func__, ep->num_req, epq->name);
913 list_for_each_entry(ep, &epq->active, active) {
914 if (epq->active.next == &ep->active)
915 ep->ptd.mps |= PTD_LAST_MSK;
916 isp1362_write_ptd(isp1362_hcd, ep, epq);
921 enable_istl_transfers(isp1362_hcd, flip);
923 epq->ptd_count += ptd_count;
924 if (epq->ptd_count > epq->stat_maxptds)
925 epq->stat_maxptds = epq->ptd_count;
927 /* check, whether the second ISTL buffer may also be filled */
928 if (!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
929 (flip ? HCBUFSTAT_ISTL0_FULL : HCBUFSTAT_ISTL1_FULL))) {
937 static void finish_transfers(struct isp1362_hcd *isp1362_hcd, unsigned long done_map,
938 struct isp1362_ep_queue *epq)
940 struct isp1362_ep *ep;
941 struct isp1362_ep *tmp;
943 if (list_empty(&epq->active)) {
944 DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
948 DBG(1, "%s: Finishing %s transfers %08lx\n", __func__, epq->name, done_map);
950 atomic_inc(&epq->finishing);
951 list_for_each_entry_safe(ep, tmp, &epq->active, active) {
952 int index = ep->ptd_index;
954 DBG(1, "%s: Checking %s PTD[%02x] $%04x\n", __func__, epq->name,
955 index, ep->ptd_offset);
958 if (__test_and_clear_bit(index, &done_map)) {
959 isp1362_read_ptd(isp1362_hcd, ep, epq);
960 epq->free_ptd = index;
961 BUG_ON(ep->num_ptds == 0);
962 release_ptd_buffers(epq, ep);
964 DBG(1, "%s: ep %p req %d removed from active list\n", __func__,
966 if (!list_empty(&ep->remove_list)) {
967 list_del_init(&ep->remove_list);
968 DBG(1, "%s: ep %p removed from remove list\n", __func__, ep);
970 DBG(1, "%s: Postprocessing %s ep %p req %d\n", __func__, epq->name,
972 postproc_ep(isp1362_hcd, ep);
978 pr_warn("%s: done_map not clear: %08lx:%08lx\n",
979 __func__, done_map, epq->skip_map);
980 atomic_dec(&epq->finishing);
983 static void finish_iso_transfers(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep_queue *epq)
985 struct isp1362_ep *ep;
986 struct isp1362_ep *tmp;
988 if (list_empty(&epq->active)) {
989 DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
993 DBG(1, "%s: Finishing %s transfers\n", __func__, epq->name);
995 atomic_inc(&epq->finishing);
996 list_for_each_entry_safe(ep, tmp, &epq->active, active) {
997 DBG(1, "%s: Checking PTD $%04x\n", __func__, ep->ptd_offset);
999 isp1362_read_ptd(isp1362_hcd, ep, epq);
1000 DBG(1, "%s: Postprocessing %s ep %p\n", __func__, epq->name, ep);
1001 postproc_ep(isp1362_hcd, ep);
1003 WARN_ON(epq->blk_size != 0);
1004 atomic_dec(&epq->finishing);
1007 static irqreturn_t isp1362_irq(struct usb_hcd *hcd)
1010 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1014 spin_lock(&isp1362_hcd->lock);
1016 BUG_ON(isp1362_hcd->irq_active++);
1018 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1020 irqstat = isp1362_read_reg16(isp1362_hcd, HCuPINT);
1021 DBG(3, "%s: got IRQ %04x:%04x\n", __func__, irqstat, isp1362_hcd->irqenb);
1023 /* only handle interrupts that are currently enabled */
1024 irqstat &= isp1362_hcd->irqenb;
1025 isp1362_write_reg16(isp1362_hcd, HCuPINT, irqstat);
1028 if (irqstat & HCuPINT_SOF) {
1029 isp1362_hcd->irqenb &= ~HCuPINT_SOF;
1030 isp1362_hcd->irq_stat[ISP1362_INT_SOF]++;
1032 svc_mask &= ~HCuPINT_SOF;
1033 DBG(3, "%s: SOF\n", __func__);
1034 isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1035 if (!list_empty(&isp1362_hcd->remove_list))
1036 finish_unlinks(isp1362_hcd);
1037 if (!list_empty(&isp1362_hcd->async) && !(irqstat & HCuPINT_ATL)) {
1038 if (list_empty(&isp1362_hcd->atl_queue.active)) {
1039 start_atl_transfers(isp1362_hcd);
1041 isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
1042 isp1362_write_reg32(isp1362_hcd, HCATLSKIP,
1043 isp1362_hcd->atl_queue.skip_map);
1044 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1049 if (irqstat & HCuPINT_ISTL0) {
1050 isp1362_hcd->irq_stat[ISP1362_INT_ISTL0]++;
1052 svc_mask &= ~HCuPINT_ISTL0;
1053 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL0_FULL);
1054 DBG(1, "%s: ISTL0\n", __func__);
1055 WARN_ON((int)!!isp1362_hcd->istl_flip);
1056 WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1057 HCBUFSTAT_ISTL0_ACTIVE);
1058 WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1059 HCBUFSTAT_ISTL0_DONE));
1060 isp1362_hcd->irqenb &= ~HCuPINT_ISTL0;
1063 if (irqstat & HCuPINT_ISTL1) {
1064 isp1362_hcd->irq_stat[ISP1362_INT_ISTL1]++;
1066 svc_mask &= ~HCuPINT_ISTL1;
1067 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL1_FULL);
1068 DBG(1, "%s: ISTL1\n", __func__);
1069 WARN_ON(!(int)isp1362_hcd->istl_flip);
1070 WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1071 HCBUFSTAT_ISTL1_ACTIVE);
1072 WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1073 HCBUFSTAT_ISTL1_DONE));
1074 isp1362_hcd->irqenb &= ~HCuPINT_ISTL1;
1077 if (irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) {
1078 WARN_ON((irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) ==
1079 (HCuPINT_ISTL0 | HCuPINT_ISTL1));
1080 finish_iso_transfers(isp1362_hcd,
1081 &isp1362_hcd->istl_queue[isp1362_hcd->istl_flip]);
1082 start_iso_transfers(isp1362_hcd);
1083 isp1362_hcd->istl_flip = 1 - isp1362_hcd->istl_flip;
1086 if (irqstat & HCuPINT_INTL) {
1087 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1088 u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCINTLSKIP);
1089 isp1362_hcd->irq_stat[ISP1362_INT_INTL]++;
1091 DBG(2, "%s: INTL\n", __func__);
1093 svc_mask &= ~HCuPINT_INTL;
1095 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, skip_map | done_map);
1096 if (~(done_map | skip_map) == 0)
1097 /* All PTDs are finished, disable INTL processing entirely */
1098 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
1103 DBG(3, "%s: INTL done_map %08x\n", __func__, done_map);
1104 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1105 start_intl_transfers(isp1362_hcd);
1109 if (irqstat & HCuPINT_ATL) {
1110 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1111 u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCATLSKIP);
1112 isp1362_hcd->irq_stat[ISP1362_INT_ATL]++;
1114 DBG(2, "%s: ATL\n", __func__);
1116 svc_mask &= ~HCuPINT_ATL;
1118 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, skip_map | done_map);
1119 if (~(done_map | skip_map) == 0)
1120 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1122 DBG(3, "%s: ATL done_map %08x\n", __func__, done_map);
1123 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1124 start_atl_transfers(isp1362_hcd);
1129 if (irqstat & HCuPINT_OPR) {
1130 u32 intstat = isp1362_read_reg32(isp1362_hcd, HCINTSTAT);
1131 isp1362_hcd->irq_stat[ISP1362_INT_OPR]++;
1133 svc_mask &= ~HCuPINT_OPR;
1134 DBG(2, "%s: OPR %08x:%08x\n", __func__, intstat, isp1362_hcd->intenb);
1135 intstat &= isp1362_hcd->intenb;
1136 if (intstat & OHCI_INTR_UE) {
1137 pr_err("Unrecoverable error\n");
1138 /* FIXME: do here reset or cleanup or whatever */
1140 if (intstat & OHCI_INTR_RHSC) {
1141 isp1362_hcd->rhstatus = isp1362_read_reg32(isp1362_hcd, HCRHSTATUS);
1142 isp1362_hcd->rhport[0] = isp1362_read_reg32(isp1362_hcd, HCRHPORT1);
1143 isp1362_hcd->rhport[1] = isp1362_read_reg32(isp1362_hcd, HCRHPORT2);
1145 if (intstat & OHCI_INTR_RD) {
1146 pr_info("%s: RESUME DETECTED\n", __func__);
1147 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1148 usb_hcd_resume_root_hub(hcd);
1150 isp1362_write_reg32(isp1362_hcd, HCINTSTAT, intstat);
1151 irqstat &= ~HCuPINT_OPR;
1155 if (irqstat & HCuPINT_SUSP) {
1156 isp1362_hcd->irq_stat[ISP1362_INT_SUSP]++;
1158 svc_mask &= ~HCuPINT_SUSP;
1160 pr_info("%s: SUSPEND IRQ\n", __func__);
1163 if (irqstat & HCuPINT_CLKRDY) {
1164 isp1362_hcd->irq_stat[ISP1362_INT_CLKRDY]++;
1166 isp1362_hcd->irqenb &= ~HCuPINT_CLKRDY;
1167 svc_mask &= ~HCuPINT_CLKRDY;
1168 pr_info("%s: CLKRDY IRQ\n", __func__);
1172 pr_err("%s: Unserviced interrupt(s) %04x\n", __func__, svc_mask);
1174 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
1175 isp1362_hcd->irq_active--;
1176 spin_unlock(&isp1362_hcd->lock);
1178 return IRQ_RETVAL(handled);
1181 /*-------------------------------------------------------------------------*/
1183 #define MAX_PERIODIC_LOAD 900 /* out of 1000 usec */
1184 static int balance(struct isp1362_hcd *isp1362_hcd, u16 interval, u16 load)
1186 int i, branch = -ENOSPC;
1188 /* search for the least loaded schedule branch of that interval
1189 * which has enough bandwidth left unreserved.
1191 for (i = 0; i < interval; i++) {
1192 if (branch < 0 || isp1362_hcd->load[branch] > isp1362_hcd->load[i]) {
1195 for (j = i; j < PERIODIC_SIZE; j += interval) {
1196 if ((isp1362_hcd->load[j] + load) > MAX_PERIODIC_LOAD) {
1197 pr_err("%s: new load %d load[%02x] %d max %d\n", __func__,
1198 load, j, isp1362_hcd->load[j], MAX_PERIODIC_LOAD);
1202 if (j < PERIODIC_SIZE)
1210 /* NB! ALL the code above this point runs with isp1362_hcd->lock
1214 /*-------------------------------------------------------------------------*/
1216 static int isp1362_urb_enqueue(struct usb_hcd *hcd,
1220 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1221 struct usb_device *udev = urb->dev;
1222 unsigned int pipe = urb->pipe;
1223 int is_out = !usb_pipein(pipe);
1224 int type = usb_pipetype(pipe);
1225 int epnum = usb_pipeendpoint(pipe);
1226 struct usb_host_endpoint *hep = urb->ep;
1227 struct isp1362_ep *ep = NULL;
1228 unsigned long flags;
1231 DBG(3, "%s: urb %p\n", __func__, urb);
1233 if (type == PIPE_ISOCHRONOUS) {
1234 pr_err("Isochronous transfers not supported\n");
1238 URB_DBG("%s: FA %d ep%d%s %s: len %d %s%s\n", __func__,
1239 usb_pipedevice(pipe), epnum,
1240 is_out ? "out" : "in",
1241 usb_pipecontrol(pipe) ? "ctrl" :
1242 usb_pipeint(pipe) ? "int" :
1243 usb_pipebulk(pipe) ? "bulk" :
1245 urb->transfer_buffer_length,
1246 (urb->transfer_flags & URB_ZERO_PACKET) ? "ZERO_PACKET " : "",
1247 !(urb->transfer_flags & URB_SHORT_NOT_OK) ?
1250 /* avoid all allocations within spinlocks: request or endpoint */
1252 ep = kzalloc(sizeof *ep, mem_flags);
1256 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1258 /* don't submit to a dead or disabled port */
1259 if (!((isp1362_hcd->rhport[0] | isp1362_hcd->rhport[1]) &
1260 USB_PORT_STAT_ENABLE) ||
1261 !HC_IS_RUNNING(hcd->state)) {
1264 goto fail_not_linked;
1267 retval = usb_hcd_link_urb_to_ep(hcd, urb);
1270 goto fail_not_linked;
1276 INIT_LIST_HEAD(&ep->schedule);
1277 INIT_LIST_HEAD(&ep->active);
1278 INIT_LIST_HEAD(&ep->remove_list);
1279 ep->udev = usb_get_dev(udev);
1282 ep->maxpacket = usb_maxpacket(udev, urb->pipe);
1283 ep->ptd_offset = -EINVAL;
1284 ep->ptd_index = -EINVAL;
1285 usb_settoggle(udev, epnum, is_out, 0);
1287 if (type == PIPE_CONTROL)
1288 ep->nextpid = USB_PID_SETUP;
1290 ep->nextpid = USB_PID_OUT;
1292 ep->nextpid = USB_PID_IN;
1295 case PIPE_ISOCHRONOUS:
1296 case PIPE_INTERRUPT:
1297 if (urb->interval > PERIODIC_SIZE)
1298 urb->interval = PERIODIC_SIZE;
1299 ep->interval = urb->interval;
1300 ep->branch = PERIODIC_SIZE;
1301 ep->load = usb_calc_bus_time(udev->speed, !is_out,
1302 type == PIPE_ISOCHRONOUS,
1303 usb_maxpacket(udev, pipe)) / 1000;
1308 ep->num_req = isp1362_hcd->req_serial++;
1310 /* maybe put endpoint into schedule */
1314 if (list_empty(&ep->schedule)) {
1315 DBG(1, "%s: Adding ep %p req %d to async schedule\n",
1316 __func__, ep, ep->num_req);
1317 list_add_tail(&ep->schedule, &isp1362_hcd->async);
1320 case PIPE_ISOCHRONOUS:
1321 case PIPE_INTERRUPT:
1322 urb->interval = ep->interval;
1324 /* urb submitted for already existing EP */
1325 if (ep->branch < PERIODIC_SIZE)
1328 retval = balance(isp1362_hcd, ep->interval, ep->load);
1330 pr_err("%s: balance returned %d\n", __func__, retval);
1333 ep->branch = retval;
1335 isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1336 DBG(1, "%s: Current frame %04x branch %02x start_frame %04x(%04x)\n",
1337 __func__, isp1362_hcd->fmindex, ep->branch,
1338 ((isp1362_hcd->fmindex + PERIODIC_SIZE - 1) &
1339 ~(PERIODIC_SIZE - 1)) + ep->branch,
1340 (isp1362_hcd->fmindex & (PERIODIC_SIZE - 1)) + ep->branch);
1342 if (list_empty(&ep->schedule)) {
1343 if (type == PIPE_ISOCHRONOUS) {
1344 u16 frame = isp1362_hcd->fmindex;
1346 frame += max_t(u16, 8, ep->interval);
1347 frame &= ~(ep->interval - 1);
1348 frame |= ep->branch;
1349 if (frame_before(frame, isp1362_hcd->fmindex))
1350 frame += ep->interval;
1351 urb->start_frame = frame;
1353 DBG(1, "%s: Adding ep %p to isoc schedule\n", __func__, ep);
1354 list_add_tail(&ep->schedule, &isp1362_hcd->isoc);
1356 DBG(1, "%s: Adding ep %p to periodic schedule\n", __func__, ep);
1357 list_add_tail(&ep->schedule, &isp1362_hcd->periodic);
1360 DBG(1, "%s: ep %p already scheduled\n", __func__, ep);
1362 DBG(2, "%s: load %d bandwidth %d -> %d\n", __func__,
1363 ep->load / ep->interval, isp1362_hcd->load[ep->branch],
1364 isp1362_hcd->load[ep->branch] + ep->load);
1365 isp1362_hcd->load[ep->branch] += ep->load;
1369 ALIGNSTAT(isp1362_hcd, urb->transfer_buffer);
1374 start_atl_transfers(isp1362_hcd);
1376 case PIPE_INTERRUPT:
1377 start_intl_transfers(isp1362_hcd);
1379 case PIPE_ISOCHRONOUS:
1380 start_iso_transfers(isp1362_hcd);
1387 usb_hcd_unlink_urb_from_ep(hcd, urb);
1391 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1393 DBG(0, "%s: urb %p failed with %d\n", __func__, urb, retval);
1397 static int isp1362_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status)
1399 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1400 struct usb_host_endpoint *hep;
1401 unsigned long flags;
1402 struct isp1362_ep *ep;
1405 DBG(3, "%s: urb %p\n", __func__, urb);
1407 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1408 retval = usb_hcd_check_unlink_urb(hcd, urb, status);
1415 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1421 /* In front of queue? */
1422 if (ep->hep->urb_list.next == &urb->urb_list) {
1423 if (!list_empty(&ep->active)) {
1424 DBG(1, "%s: urb %p ep %p req %d active PTD[%d] $%04x\n", __func__,
1425 urb, ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1426 /* disable processing and queue PTD for removal */
1427 remove_ptd(isp1362_hcd, ep);
1432 DBG(1, "%s: Finishing ep %p req %d\n", __func__, ep,
1434 finish_request(isp1362_hcd, ep, urb, status);
1436 DBG(1, "%s: urb %p active; wait4irq\n", __func__, urb);
1438 pr_warn("%s: No EP in URB %p\n", __func__, urb);
1442 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1444 DBG(3, "%s: exit\n", __func__);
1449 static void isp1362_endpoint_disable(struct usb_hcd *hcd, struct usb_host_endpoint *hep)
1451 struct isp1362_ep *ep = hep->hcpriv;
1452 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1453 unsigned long flags;
1455 DBG(1, "%s: ep %p\n", __func__, ep);
1458 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1459 if (!list_empty(&hep->urb_list)) {
1460 if (!list_empty(&ep->active) && list_empty(&ep->remove_list)) {
1461 DBG(1, "%s: Removing ep %p req %d PTD[%d] $%04x\n", __func__,
1462 ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1463 remove_ptd(isp1362_hcd, ep);
1464 pr_info("%s: Waiting for Interrupt to clean up\n", __func__);
1467 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1468 /* Wait for interrupt to clear out active list */
1469 while (!list_empty(&ep->active))
1472 DBG(1, "%s: Freeing EP %p\n", __func__, ep);
1474 usb_put_dev(ep->udev);
1479 static int isp1362_get_frame(struct usb_hcd *hcd)
1481 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1483 unsigned long flags;
1485 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1486 fmnum = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1487 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1492 /*-------------------------------------------------------------------------*/
1494 /* Adapted from ohci-hub.c */
1495 static int isp1362_hub_status_data(struct usb_hcd *hcd, char *buf)
1497 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1498 int ports, i, changed = 0;
1499 unsigned long flags;
1501 if (!HC_IS_RUNNING(hcd->state))
1504 /* Report no status change now, if we are scheduled to be
1506 if (timer_pending(&hcd->rh_timer))
1509 ports = isp1362_hcd->rhdesca & RH_A_NDP;
1512 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1514 if (isp1362_hcd->rhstatus & (RH_HS_LPSC | RH_HS_OCIC))
1515 buf[0] = changed = 1;
1519 for (i = 0; i < ports; i++) {
1520 u32 status = isp1362_hcd->rhport[i];
1522 if (status & (RH_PS_CSC | RH_PS_PESC | RH_PS_PSSC |
1523 RH_PS_OCIC | RH_PS_PRSC)) {
1525 buf[0] |= 1 << (i + 1);
1529 if (!(status & RH_PS_CCS))
1532 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1536 static void isp1362_hub_descriptor(struct isp1362_hcd *isp1362_hcd,
1537 struct usb_hub_descriptor *desc)
1539 u32 reg = isp1362_hcd->rhdesca;
1541 DBG(3, "%s: enter\n", __func__);
1543 desc->bDescriptorType = USB_DT_HUB;
1544 desc->bDescLength = 9;
1545 desc->bHubContrCurrent = 0;
1546 desc->bNbrPorts = reg & 0x3;
1547 /* Power switching, device type, overcurrent. */
1548 desc->wHubCharacteristics = cpu_to_le16((reg >> 8) &
1552 DBG(0, "%s: hubcharacteristics = %02x\n", __func__,
1553 desc->wHubCharacteristics);
1554 desc->bPwrOn2PwrGood = (reg >> 24) & 0xff;
1555 /* ports removable, and legacy PortPwrCtrlMask */
1556 desc->u.hs.DeviceRemovable[0] = desc->bNbrPorts == 1 ? 1 << 1 : 3 << 1;
1557 desc->u.hs.DeviceRemovable[1] = ~0;
1559 DBG(3, "%s: exit\n", __func__);
1562 /* Adapted from ohci-hub.c */
1563 static int isp1362_hub_control(struct usb_hcd *hcd, u16 typeReq, u16 wValue,
1564 u16 wIndex, char *buf, u16 wLength)
1566 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1568 unsigned long flags;
1570 int ports = isp1362_hcd->rhdesca & RH_A_NDP;
1574 case ClearHubFeature:
1575 DBG(0, "ClearHubFeature: ");
1577 case C_HUB_OVER_CURRENT:
1578 DBG(0, "C_HUB_OVER_CURRENT\n");
1579 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1580 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_OCIC);
1581 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1583 case C_HUB_LOCAL_POWER:
1584 DBG(0, "C_HUB_LOCAL_POWER\n");
1591 DBG(0, "SetHubFeature: ");
1593 case C_HUB_OVER_CURRENT:
1594 case C_HUB_LOCAL_POWER:
1595 DBG(0, "C_HUB_OVER_CURRENT or C_HUB_LOCAL_POWER\n");
1601 case GetHubDescriptor:
1602 DBG(0, "GetHubDescriptor\n");
1603 isp1362_hub_descriptor(isp1362_hcd, (struct usb_hub_descriptor *)buf);
1606 DBG(0, "GetHubStatus\n");
1607 put_unaligned(cpu_to_le32(0), (__le32 *) buf);
1611 DBG(0, "GetPortStatus\n");
1613 if (!wIndex || wIndex > ports)
1615 tmp = isp1362_hcd->rhport[--wIndex];
1616 put_unaligned(cpu_to_le32(tmp), (__le32 *) buf);
1618 case ClearPortFeature:
1619 DBG(0, "ClearPortFeature: ");
1620 if (!wIndex || wIndex > ports)
1625 case USB_PORT_FEAT_ENABLE:
1626 DBG(0, "USB_PORT_FEAT_ENABLE\n");
1629 case USB_PORT_FEAT_C_ENABLE:
1630 DBG(0, "USB_PORT_FEAT_C_ENABLE\n");
1633 case USB_PORT_FEAT_SUSPEND:
1634 DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1637 case USB_PORT_FEAT_C_SUSPEND:
1638 DBG(0, "USB_PORT_FEAT_C_SUSPEND\n");
1641 case USB_PORT_FEAT_POWER:
1642 DBG(0, "USB_PORT_FEAT_POWER\n");
1646 case USB_PORT_FEAT_C_CONNECTION:
1647 DBG(0, "USB_PORT_FEAT_C_CONNECTION\n");
1650 case USB_PORT_FEAT_C_OVER_CURRENT:
1651 DBG(0, "USB_PORT_FEAT_C_OVER_CURRENT\n");
1654 case USB_PORT_FEAT_C_RESET:
1655 DBG(0, "USB_PORT_FEAT_C_RESET\n");
1662 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1663 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, tmp);
1664 isp1362_hcd->rhport[wIndex] =
1665 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1666 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1668 case SetPortFeature:
1669 DBG(0, "SetPortFeature: ");
1670 if (!wIndex || wIndex > ports)
1674 case USB_PORT_FEAT_SUSPEND:
1675 DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1676 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1677 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PSS);
1678 isp1362_hcd->rhport[wIndex] =
1679 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1680 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1682 case USB_PORT_FEAT_POWER:
1683 DBG(0, "USB_PORT_FEAT_POWER\n");
1684 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1685 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PPS);
1686 isp1362_hcd->rhport[wIndex] =
1687 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1688 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1690 case USB_PORT_FEAT_RESET:
1691 DBG(0, "USB_PORT_FEAT_RESET\n");
1692 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1694 t1 = jiffies + msecs_to_jiffies(USB_RESET_WIDTH);
1695 while (time_before(jiffies, t1)) {
1696 /* spin until any current reset finishes */
1698 tmp = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1699 if (!(tmp & RH_PS_PRS))
1703 if (!(tmp & RH_PS_CCS))
1705 /* Reset lasts 10ms (claims datasheet) */
1706 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, (RH_PS_PRS));
1708 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1710 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1713 isp1362_hcd->rhport[wIndex] = isp1362_read_reg32(isp1362_hcd,
1714 HCRHPORT1 + wIndex);
1715 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1724 /* "protocol stall" on error */
1725 DBG(0, "PROTOCOL STALL\n");
1733 static int isp1362_bus_suspend(struct usb_hcd *hcd)
1736 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1737 unsigned long flags;
1739 if (time_before(jiffies, isp1362_hcd->next_statechange))
1742 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1744 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1745 switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1746 case OHCI_USB_RESUME:
1747 DBG(0, "%s: resume/suspend?\n", __func__);
1748 isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1749 isp1362_hcd->hc_control |= OHCI_USB_RESET;
1750 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1752 case OHCI_USB_RESET:
1754 pr_warn("%s: needs reinit!\n", __func__);
1756 case OHCI_USB_SUSPEND:
1757 pr_warn("%s: already suspended?\n", __func__);
1760 DBG(0, "%s: suspend root hub\n", __func__);
1762 /* First stop any processing */
1763 hcd->state = HC_STATE_QUIESCING;
1764 if (!list_empty(&isp1362_hcd->atl_queue.active) ||
1765 !list_empty(&isp1362_hcd->intl_queue.active) ||
1766 !list_empty(&isp1362_hcd->istl_queue[0] .active) ||
1767 !list_empty(&isp1362_hcd->istl_queue[1] .active)) {
1770 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
1771 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
1772 isp1362_write_reg16(isp1362_hcd, HCBUFSTAT, 0);
1773 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1774 isp1362_write_reg32(isp1362_hcd, HCINTSTAT, OHCI_INTR_SF);
1776 DBG(0, "%s: stopping schedules ...\n", __func__);
1781 if (isp1362_read_reg32(isp1362_hcd, HCINTSTAT) & OHCI_INTR_SF)
1785 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ATL) {
1786 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1787 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1789 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_INTL) {
1790 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1791 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1793 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL0)
1794 finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[0]);
1795 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL1)
1796 finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[1]);
1798 DBG(0, "%s: HCINTSTAT: %08x\n", __func__,
1799 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1800 isp1362_write_reg32(isp1362_hcd, HCINTSTAT,
1801 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1804 isp1362_hcd->hc_control = OHCI_USB_SUSPEND;
1805 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1806 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1807 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1810 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1811 if ((isp1362_hcd->hc_control & OHCI_CTRL_HCFS) != OHCI_USB_SUSPEND) {
1812 pr_err("%s: controller won't suspend %08x\n", __func__,
1813 isp1362_hcd->hc_control);
1818 /* no resumes until devices finish suspending */
1819 isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(5);
1823 hcd->state = HC_STATE_SUSPENDED;
1824 DBG(0, "%s: HCD suspended: %08x\n", __func__,
1825 isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1827 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1831 static int isp1362_bus_resume(struct usb_hcd *hcd)
1833 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1835 unsigned long flags;
1836 int status = -EINPROGRESS;
1838 if (time_before(jiffies, isp1362_hcd->next_statechange))
1841 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1842 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1843 pr_info("%s: HCCONTROL: %08x\n", __func__, isp1362_hcd->hc_control);
1844 if (hcd->state == HC_STATE_RESUMING) {
1845 pr_warn("%s: duplicate resume\n", __func__);
1848 switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1849 case OHCI_USB_SUSPEND:
1850 DBG(0, "%s: resume root hub\n", __func__);
1851 isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1852 isp1362_hcd->hc_control |= OHCI_USB_RESUME;
1853 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1855 case OHCI_USB_RESUME:
1856 /* HCFS changes sometime after INTR_RD */
1857 DBG(0, "%s: remote wakeup\n", __func__);
1860 DBG(0, "%s: odd resume\n", __func__);
1862 hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1864 default: /* RESET, we lost power */
1865 DBG(0, "%s: root hub hardware reset\n", __func__);
1868 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1869 if (status == -EBUSY) {
1870 DBG(0, "%s: Restarting HC\n", __func__);
1871 isp1362_hc_stop(hcd);
1872 return isp1362_hc_start(hcd);
1874 if (status != -EINPROGRESS)
1876 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1877 port = isp1362_read_reg32(isp1362_hcd, HCRHDESCA) & RH_A_NDP;
1879 u32 stat = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + port);
1881 /* force global, not selective, resume */
1882 if (!(stat & RH_PS_PSS)) {
1883 DBG(0, "%s: Not Resuming RH port %d\n", __func__, port);
1886 DBG(0, "%s: Resuming RH port %d\n", __func__, port);
1887 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + port, RH_PS_POCI);
1889 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1891 /* Some controllers (lucent) need extra-long delays */
1892 hcd->state = HC_STATE_RESUMING;
1893 mdelay(20 /* usb 11.5.1.10 */ + 15);
1895 isp1362_hcd->hc_control = OHCI_USB_OPER;
1896 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1897 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1898 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1899 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1903 /* keep it alive for ~5x suspend + resume costs */
1904 isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(250);
1906 hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1907 hcd->state = HC_STATE_RUNNING;
1911 #define isp1362_bus_suspend NULL
1912 #define isp1362_bus_resume NULL
1915 /*-------------------------------------------------------------------------*/
1917 static void dump_irq(struct seq_file *s, char *label, u16 mask)
1919 seq_printf(s, "%-15s %04x%s%s%s%s%s%s\n", label, mask,
1920 mask & HCuPINT_CLKRDY ? " clkrdy" : "",
1921 mask & HCuPINT_SUSP ? " susp" : "",
1922 mask & HCuPINT_OPR ? " opr" : "",
1923 mask & HCuPINT_EOT ? " eot" : "",
1924 mask & HCuPINT_ATL ? " atl" : "",
1925 mask & HCuPINT_SOF ? " sof" : "");
1928 static void dump_int(struct seq_file *s, char *label, u32 mask)
1930 seq_printf(s, "%-15s %08x%s%s%s%s%s%s%s\n", label, mask,
1931 mask & OHCI_INTR_MIE ? " MIE" : "",
1932 mask & OHCI_INTR_RHSC ? " rhsc" : "",
1933 mask & OHCI_INTR_FNO ? " fno" : "",
1934 mask & OHCI_INTR_UE ? " ue" : "",
1935 mask & OHCI_INTR_RD ? " rd" : "",
1936 mask & OHCI_INTR_SF ? " sof" : "",
1937 mask & OHCI_INTR_SO ? " so" : "");
1940 static void dump_ctrl(struct seq_file *s, char *label, u32 mask)
1942 seq_printf(s, "%-15s %08x%s%s%s\n", label, mask,
1943 mask & OHCI_CTRL_RWC ? " rwc" : "",
1944 mask & OHCI_CTRL_RWE ? " rwe" : "",
1947 switch (mask & OHCI_CTRL_HCFS) {
1951 case OHCI_USB_RESET:
1954 case OHCI_USB_RESUME:
1957 case OHCI_USB_SUSPEND:
1967 static void dump_regs(struct seq_file *s, struct isp1362_hcd *isp1362_hcd)
1969 seq_printf(s, "HCREVISION [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCREVISION),
1970 isp1362_read_reg32(isp1362_hcd, HCREVISION));
1971 seq_printf(s, "HCCONTROL [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCONTROL),
1972 isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1973 seq_printf(s, "HCCMDSTAT [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCMDSTAT),
1974 isp1362_read_reg32(isp1362_hcd, HCCMDSTAT));
1975 seq_printf(s, "HCINTSTAT [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTSTAT),
1976 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1977 seq_printf(s, "HCINTENB [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTENB),
1978 isp1362_read_reg32(isp1362_hcd, HCINTENB));
1979 seq_printf(s, "HCFMINTVL [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMINTVL),
1980 isp1362_read_reg32(isp1362_hcd, HCFMINTVL));
1981 seq_printf(s, "HCFMREM [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMREM),
1982 isp1362_read_reg32(isp1362_hcd, HCFMREM));
1983 seq_printf(s, "HCFMNUM [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMNUM),
1984 isp1362_read_reg32(isp1362_hcd, HCFMNUM));
1985 seq_printf(s, "HCLSTHRESH [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCLSTHRESH),
1986 isp1362_read_reg32(isp1362_hcd, HCLSTHRESH));
1987 seq_printf(s, "HCRHDESCA [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCA),
1988 isp1362_read_reg32(isp1362_hcd, HCRHDESCA));
1989 seq_printf(s, "HCRHDESCB [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCB),
1990 isp1362_read_reg32(isp1362_hcd, HCRHDESCB));
1991 seq_printf(s, "HCRHSTATUS [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHSTATUS),
1992 isp1362_read_reg32(isp1362_hcd, HCRHSTATUS));
1993 seq_printf(s, "HCRHPORT1 [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT1),
1994 isp1362_read_reg32(isp1362_hcd, HCRHPORT1));
1995 seq_printf(s, "HCRHPORT2 [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT2),
1996 isp1362_read_reg32(isp1362_hcd, HCRHPORT2));
1997 seq_printf(s, "\n");
1998 seq_printf(s, "HCHWCFG [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCHWCFG),
1999 isp1362_read_reg16(isp1362_hcd, HCHWCFG));
2000 seq_printf(s, "HCDMACFG [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCDMACFG),
2001 isp1362_read_reg16(isp1362_hcd, HCDMACFG));
2002 seq_printf(s, "HCXFERCTR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCXFERCTR),
2003 isp1362_read_reg16(isp1362_hcd, HCXFERCTR));
2004 seq_printf(s, "HCuPINT [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINT),
2005 isp1362_read_reg16(isp1362_hcd, HCuPINT));
2006 seq_printf(s, "HCuPINTENB [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINTENB),
2007 isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2008 seq_printf(s, "HCCHIPID [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCCHIPID),
2009 isp1362_read_reg16(isp1362_hcd, HCCHIPID));
2010 seq_printf(s, "HCSCRATCH [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCSCRATCH),
2011 isp1362_read_reg16(isp1362_hcd, HCSCRATCH));
2012 seq_printf(s, "HCBUFSTAT [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCBUFSTAT),
2013 isp1362_read_reg16(isp1362_hcd, HCBUFSTAT));
2014 seq_printf(s, "HCDIRADDR [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCDIRADDR),
2015 isp1362_read_reg32(isp1362_hcd, HCDIRADDR));
2017 seq_printf(s, "HCDIRDATA [%02x] %04x\n", ISP1362_REG_NO(HCDIRDATA),
2018 isp1362_read_reg16(isp1362_hcd, HCDIRDATA));
2020 seq_printf(s, "HCISTLBUFSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLBUFSZ),
2021 isp1362_read_reg16(isp1362_hcd, HCISTLBUFSZ));
2022 seq_printf(s, "HCISTLRATE [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLRATE),
2023 isp1362_read_reg16(isp1362_hcd, HCISTLRATE));
2024 seq_printf(s, "\n");
2025 seq_printf(s, "HCINTLBUFSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBUFSZ),
2026 isp1362_read_reg16(isp1362_hcd, HCINTLBUFSZ));
2027 seq_printf(s, "HCINTLBLKSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBLKSZ),
2028 isp1362_read_reg16(isp1362_hcd, HCINTLBLKSZ));
2029 seq_printf(s, "HCINTLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLDONE),
2030 isp1362_read_reg32(isp1362_hcd, HCINTLDONE));
2031 seq_printf(s, "HCINTLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLSKIP),
2032 isp1362_read_reg32(isp1362_hcd, HCINTLSKIP));
2033 seq_printf(s, "HCINTLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLLAST),
2034 isp1362_read_reg32(isp1362_hcd, HCINTLLAST));
2035 seq_printf(s, "HCINTLCURR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLCURR),
2036 isp1362_read_reg16(isp1362_hcd, HCINTLCURR));
2037 seq_printf(s, "\n");
2038 seq_printf(s, "HCATLBUFSZ [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBUFSZ),
2039 isp1362_read_reg16(isp1362_hcd, HCATLBUFSZ));
2040 seq_printf(s, "HCATLBLKSZ [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBLKSZ),
2041 isp1362_read_reg16(isp1362_hcd, HCATLBLKSZ));
2043 seq_printf(s, "HCATLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDONE),
2044 isp1362_read_reg32(isp1362_hcd, HCATLDONE));
2046 seq_printf(s, "HCATLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLSKIP),
2047 isp1362_read_reg32(isp1362_hcd, HCATLSKIP));
2048 seq_printf(s, "HCATLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLLAST),
2049 isp1362_read_reg32(isp1362_hcd, HCATLLAST));
2050 seq_printf(s, "HCATLCURR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLCURR),
2051 isp1362_read_reg16(isp1362_hcd, HCATLCURR));
2052 seq_printf(s, "\n");
2053 seq_printf(s, "HCATLDTC [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTC),
2054 isp1362_read_reg16(isp1362_hcd, HCATLDTC));
2055 seq_printf(s, "HCATLDTCTO [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTCTO),
2056 isp1362_read_reg16(isp1362_hcd, HCATLDTCTO));
2059 static int isp1362_show(struct seq_file *s, void *unused)
2061 struct isp1362_hcd *isp1362_hcd = s->private;
2062 struct isp1362_ep *ep;
2065 seq_printf(s, "%s\n%s version %s\n",
2066 isp1362_hcd_to_hcd(isp1362_hcd)->product_desc, hcd_name, DRIVER_VERSION);
2068 /* collect statistics to help estimate potential win for
2069 * DMA engines that care about alignment (PXA)
2071 seq_printf(s, "alignment: 16b/%ld 8b/%ld 4b/%ld 2b/%ld 1b/%ld\n",
2072 isp1362_hcd->stat16, isp1362_hcd->stat8, isp1362_hcd->stat4,
2073 isp1362_hcd->stat2, isp1362_hcd->stat1);
2074 seq_printf(s, "max # ptds in ATL fifo: %d\n", isp1362_hcd->atl_queue.stat_maxptds);
2075 seq_printf(s, "max # ptds in INTL fifo: %d\n", isp1362_hcd->intl_queue.stat_maxptds);
2076 seq_printf(s, "max # ptds in ISTL fifo: %d\n",
2077 max(isp1362_hcd->istl_queue[0] .stat_maxptds,
2078 isp1362_hcd->istl_queue[1] .stat_maxptds));
2080 /* FIXME: don't show the following in suspended state */
2081 spin_lock_irq(&isp1362_hcd->lock);
2083 dump_irq(s, "hc_irq_enable", isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2084 dump_irq(s, "hc_irq_status", isp1362_read_reg16(isp1362_hcd, HCuPINT));
2085 dump_int(s, "ohci_int_enable", isp1362_read_reg32(isp1362_hcd, HCINTENB));
2086 dump_int(s, "ohci_int_status", isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
2087 dump_ctrl(s, "ohci_control", isp1362_read_reg32(isp1362_hcd, HCCONTROL));
2089 for (i = 0; i < NUM_ISP1362_IRQS; i++)
2090 if (isp1362_hcd->irq_stat[i])
2091 seq_printf(s, "%-15s: %d\n",
2092 ISP1362_INT_NAME(i), isp1362_hcd->irq_stat[i]);
2094 dump_regs(s, isp1362_hcd);
2095 list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
2098 seq_printf(s, "%p, ep%d%s, maxpacket %d:\n", ep, ep->epnum,
2101 switch (ep->nextpid) {
2118 s;}), ep->maxpacket) ;
2119 list_for_each_entry(urb, &ep->hep->urb_list, urb_list) {
2120 seq_printf(s, " urb%p, %d/%d\n", urb,
2122 urb->transfer_buffer_length);
2125 if (!list_empty(&isp1362_hcd->async))
2126 seq_printf(s, "\n");
2127 dump_ptd_queue(&isp1362_hcd->atl_queue);
2129 seq_printf(s, "periodic size= %d\n", PERIODIC_SIZE);
2131 list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
2132 seq_printf(s, "branch:%2d load:%3d PTD[%d] $%04x:\n", ep->branch,
2133 isp1362_hcd->load[ep->branch], ep->ptd_index, ep->ptd_offset);
2135 seq_printf(s, " %d/%p (%sdev%d ep%d%s max %d)\n",
2137 (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2138 ep->udev->devnum, ep->epnum,
2139 (ep->epnum == 0) ? "" :
2140 ((ep->nextpid == USB_PID_IN) ?
2141 "in" : "out"), ep->maxpacket);
2143 dump_ptd_queue(&isp1362_hcd->intl_queue);
2145 seq_printf(s, "ISO:\n");
2147 list_for_each_entry(ep, &isp1362_hcd->isoc, schedule) {
2148 seq_printf(s, " %d/%p (%sdev%d ep%d%s max %d)\n",
2150 (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2151 ep->udev->devnum, ep->epnum,
2152 (ep->epnum == 0) ? "" :
2153 ((ep->nextpid == USB_PID_IN) ?
2154 "in" : "out"), ep->maxpacket);
2157 spin_unlock_irq(&isp1362_hcd->lock);
2158 seq_printf(s, "\n");
2162 DEFINE_SHOW_ATTRIBUTE(isp1362);
2164 /* expect just one isp1362_hcd per system */
2165 static void create_debug_file(struct isp1362_hcd *isp1362_hcd)
2167 debugfs_create_file("isp1362", S_IRUGO, usb_debug_root, isp1362_hcd,
2171 static void remove_debug_file(struct isp1362_hcd *isp1362_hcd)
2173 debugfs_lookup_and_remove("isp1362", usb_debug_root);
2176 /*-------------------------------------------------------------------------*/
2178 static void __isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2182 isp1362_write_reg16(isp1362_hcd, HCSWRES, HCSWRES_MAGIC);
2183 isp1362_write_reg32(isp1362_hcd, HCCMDSTAT, OHCI_HCR);
2186 if (!(isp1362_read_reg32(isp1362_hcd, HCCMDSTAT) & OHCI_HCR))
2190 pr_err("Software reset timeout\n");
2193 static void isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2195 unsigned long flags;
2197 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2198 __isp1362_sw_reset(isp1362_hcd);
2199 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2202 static int isp1362_mem_config(struct usb_hcd *hcd)
2204 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2205 unsigned long flags;
2207 u16 istl_size = ISP1362_ISTL_BUFSIZE;
2208 u16 intl_blksize = ISP1362_INTL_BLKSIZE + PTD_HEADER_SIZE;
2209 u16 intl_size = ISP1362_INTL_BUFFERS * intl_blksize;
2210 u16 atl_blksize = ISP1362_ATL_BLKSIZE + PTD_HEADER_SIZE;
2211 u16 atl_buffers = (ISP1362_BUF_SIZE - (istl_size + intl_size)) / atl_blksize;
2215 WARN_ON(istl_size & 3);
2216 WARN_ON(atl_blksize & 3);
2217 WARN_ON(intl_blksize & 3);
2218 WARN_ON(atl_blksize < PTD_HEADER_SIZE);
2219 WARN_ON(intl_blksize < PTD_HEADER_SIZE);
2221 BUG_ON((unsigned)ISP1362_INTL_BUFFERS > 32);
2222 if (atl_buffers > 32)
2224 atl_size = atl_buffers * atl_blksize;
2225 total = atl_size + intl_size + istl_size;
2226 dev_info(hcd->self.controller, "ISP1362 Memory usage:\n");
2227 dev_info(hcd->self.controller, " ISTL: 2 * %4d: %4d @ $%04x:$%04x\n",
2228 istl_size / 2, istl_size, 0, istl_size / 2);
2229 dev_info(hcd->self.controller, " INTL: %4d * (%3zu+8): %4d @ $%04x\n",
2230 ISP1362_INTL_BUFFERS, intl_blksize - PTD_HEADER_SIZE,
2231 intl_size, istl_size);
2232 dev_info(hcd->self.controller, " ATL : %4d * (%3zu+8): %4d @ $%04x\n",
2233 atl_buffers, atl_blksize - PTD_HEADER_SIZE,
2234 atl_size, istl_size + intl_size);
2235 dev_info(hcd->self.controller, " USED/FREE: %4d %4d\n", total,
2236 ISP1362_BUF_SIZE - total);
2238 if (total > ISP1362_BUF_SIZE) {
2239 dev_err(hcd->self.controller, "%s: Memory requested: %d, available %d\n",
2240 __func__, total, ISP1362_BUF_SIZE);
2244 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2246 for (i = 0; i < 2; i++) {
2247 isp1362_hcd->istl_queue[i].buf_start = i * istl_size / 2,
2248 isp1362_hcd->istl_queue[i].buf_size = istl_size / 2;
2249 isp1362_hcd->istl_queue[i].blk_size = 4;
2250 INIT_LIST_HEAD(&isp1362_hcd->istl_queue[i].active);
2251 snprintf(isp1362_hcd->istl_queue[i].name,
2252 sizeof(isp1362_hcd->istl_queue[i].name), "ISTL%d", i);
2253 DBG(3, "%s: %5s buf $%04x %d\n", __func__,
2254 isp1362_hcd->istl_queue[i].name,
2255 isp1362_hcd->istl_queue[i].buf_start,
2256 isp1362_hcd->istl_queue[i].buf_size);
2258 isp1362_write_reg16(isp1362_hcd, HCISTLBUFSZ, istl_size / 2);
2260 isp1362_hcd->intl_queue.buf_start = istl_size;
2261 isp1362_hcd->intl_queue.buf_size = intl_size;
2262 isp1362_hcd->intl_queue.buf_count = ISP1362_INTL_BUFFERS;
2263 isp1362_hcd->intl_queue.blk_size = intl_blksize;
2264 isp1362_hcd->intl_queue.buf_avail = isp1362_hcd->intl_queue.buf_count;
2265 isp1362_hcd->intl_queue.skip_map = ~0;
2266 INIT_LIST_HEAD(&isp1362_hcd->intl_queue.active);
2268 isp1362_write_reg16(isp1362_hcd, HCINTLBUFSZ,
2269 isp1362_hcd->intl_queue.buf_size);
2270 isp1362_write_reg16(isp1362_hcd, HCINTLBLKSZ,
2271 isp1362_hcd->intl_queue.blk_size - PTD_HEADER_SIZE);
2272 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
2273 isp1362_write_reg32(isp1362_hcd, HCINTLLAST,
2274 1 << (ISP1362_INTL_BUFFERS - 1));
2276 isp1362_hcd->atl_queue.buf_start = istl_size + intl_size;
2277 isp1362_hcd->atl_queue.buf_size = atl_size;
2278 isp1362_hcd->atl_queue.buf_count = atl_buffers;
2279 isp1362_hcd->atl_queue.blk_size = atl_blksize;
2280 isp1362_hcd->atl_queue.buf_avail = isp1362_hcd->atl_queue.buf_count;
2281 isp1362_hcd->atl_queue.skip_map = ~0;
2282 INIT_LIST_HEAD(&isp1362_hcd->atl_queue.active);
2284 isp1362_write_reg16(isp1362_hcd, HCATLBUFSZ,
2285 isp1362_hcd->atl_queue.buf_size);
2286 isp1362_write_reg16(isp1362_hcd, HCATLBLKSZ,
2287 isp1362_hcd->atl_queue.blk_size - PTD_HEADER_SIZE);
2288 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
2289 isp1362_write_reg32(isp1362_hcd, HCATLLAST,
2290 1 << (atl_buffers - 1));
2292 snprintf(isp1362_hcd->atl_queue.name,
2293 sizeof(isp1362_hcd->atl_queue.name), "ATL");
2294 snprintf(isp1362_hcd->intl_queue.name,
2295 sizeof(isp1362_hcd->intl_queue.name), "INTL");
2296 DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2297 isp1362_hcd->intl_queue.name,
2298 isp1362_hcd->intl_queue.buf_start,
2299 ISP1362_INTL_BUFFERS, isp1362_hcd->intl_queue.blk_size,
2300 isp1362_hcd->intl_queue.buf_size);
2301 DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2302 isp1362_hcd->atl_queue.name,
2303 isp1362_hcd->atl_queue.buf_start,
2304 atl_buffers, isp1362_hcd->atl_queue.blk_size,
2305 isp1362_hcd->atl_queue.buf_size);
2307 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2312 static int isp1362_hc_reset(struct usb_hcd *hcd)
2315 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2317 unsigned long timeout = 100;
2318 unsigned long flags;
2321 pr_debug("%s:\n", __func__);
2323 if (isp1362_hcd->board && isp1362_hcd->board->reset) {
2324 isp1362_hcd->board->reset(hcd->self.controller, 1);
2326 if (isp1362_hcd->board->clock)
2327 isp1362_hcd->board->clock(hcd->self.controller, 1);
2328 isp1362_hcd->board->reset(hcd->self.controller, 0);
2330 isp1362_sw_reset(isp1362_hcd);
2332 /* chip has been reset. First we need to see a clock */
2333 t = jiffies + msecs_to_jiffies(timeout);
2334 while (!clkrdy && time_before_eq(jiffies, t)) {
2335 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2336 clkrdy = isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_CLKRDY;
2337 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2342 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2343 isp1362_write_reg16(isp1362_hcd, HCuPINT, HCuPINT_CLKRDY);
2344 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2346 pr_err("Clock not ready after %lums\n", timeout);
2352 static void isp1362_hc_stop(struct usb_hcd *hcd)
2354 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2355 unsigned long flags;
2358 pr_debug("%s:\n", __func__);
2360 del_timer_sync(&hcd->rh_timer);
2362 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2364 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2366 /* Switch off power for all ports */
2367 tmp = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2368 tmp &= ~(RH_A_NPS | RH_A_PSM);
2369 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, tmp);
2370 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2372 /* Reset the chip */
2373 if (isp1362_hcd->board && isp1362_hcd->board->reset)
2374 isp1362_hcd->board->reset(hcd->self.controller, 1);
2376 __isp1362_sw_reset(isp1362_hcd);
2378 if (isp1362_hcd->board && isp1362_hcd->board->clock)
2379 isp1362_hcd->board->clock(hcd->self.controller, 0);
2381 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2384 #ifdef CHIP_BUFFER_TEST
2385 static int isp1362_chip_test(struct isp1362_hcd *isp1362_hcd)
2389 unsigned long flags;
2391 ref = kmalloc(2 * ISP1362_BUF_SIZE, GFP_KERNEL);
2394 u16 *tst = &ref[ISP1362_BUF_SIZE / 2];
2396 for (offset = 0; offset < ISP1362_BUF_SIZE / 2; offset++) {
2397 ref[offset] = ~offset;
2398 tst[offset] = offset;
2401 for (offset = 0; offset < 4; offset++) {
2404 for (j = 0; j < 8; j++) {
2405 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2406 isp1362_write_buffer(isp1362_hcd, (u8 *)ref + offset, 0, j);
2407 isp1362_read_buffer(isp1362_hcd, (u8 *)tst + offset, 0, j);
2408 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2410 if (memcmp(ref, tst, j)) {
2412 pr_err("%s: memory check with %d byte offset %d failed\n",
2413 __func__, j, offset);
2414 dump_data((u8 *)ref + offset, j);
2415 dump_data((u8 *)tst + offset, j);
2420 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2421 isp1362_write_buffer(isp1362_hcd, ref, 0, ISP1362_BUF_SIZE);
2422 isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2423 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2425 if (memcmp(ref, tst, ISP1362_BUF_SIZE)) {
2427 pr_err("%s: memory check failed\n", __func__);
2428 dump_data((u8 *)tst, ISP1362_BUF_SIZE / 2);
2431 for (offset = 0; offset < 256; offset++) {
2436 memset(tst, 0, ISP1362_BUF_SIZE);
2437 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2438 isp1362_write_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2439 isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2440 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2441 if (memcmp(tst, tst + (ISP1362_BUF_SIZE / (2 * sizeof(*tst))),
2442 ISP1362_BUF_SIZE / 2)) {
2443 pr_err("%s: Failed to clear buffer\n", __func__);
2444 dump_data((u8 *)tst, ISP1362_BUF_SIZE);
2447 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2448 isp1362_write_buffer(isp1362_hcd, ref, offset * 2, PTD_HEADER_SIZE);
2449 isp1362_write_buffer(isp1362_hcd, ref + PTD_HEADER_SIZE / sizeof(*ref),
2450 offset * 2 + PTD_HEADER_SIZE, test_size);
2451 isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2452 PTD_HEADER_SIZE + test_size);
2453 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2454 if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2455 dump_data(((u8 *)ref) + offset, PTD_HEADER_SIZE + test_size);
2456 dump_data((u8 *)tst, PTD_HEADER_SIZE + test_size);
2457 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2458 isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2459 PTD_HEADER_SIZE + test_size);
2460 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2461 if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2463 pr_err("%s: memory check with offset %02x failed\n",
2467 pr_warn("%s: memory check with offset %02x ok after second read\n",
2477 static int isp1362_hc_start(struct usb_hcd *hcd)
2480 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2481 struct isp1362_platform_data *board = isp1362_hcd->board;
2484 unsigned long flags;
2486 pr_debug("%s:\n", __func__);
2488 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2489 chipid = isp1362_read_reg16(isp1362_hcd, HCCHIPID);
2490 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2492 if ((chipid & HCCHIPID_MASK) != HCCHIPID_MAGIC) {
2493 pr_err("%s: Invalid chip ID %04x\n", __func__, chipid);
2497 #ifdef CHIP_BUFFER_TEST
2498 ret = isp1362_chip_test(isp1362_hcd);
2502 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2503 /* clear interrupt status and disable all interrupt sources */
2504 isp1362_write_reg16(isp1362_hcd, HCuPINT, 0xff);
2505 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2508 hwcfg = HCHWCFG_INT_ENABLE | HCHWCFG_DBWIDTH(1);
2509 if (board->sel15Kres)
2510 hwcfg |= HCHWCFG_PULLDOWN_DS2 |
2511 ((MAX_ROOT_PORTS > 1) ? HCHWCFG_PULLDOWN_DS1 : 0);
2512 if (board->clknotstop)
2513 hwcfg |= HCHWCFG_CLKNOTSTOP;
2514 if (board->oc_enable)
2515 hwcfg |= HCHWCFG_ANALOG_OC;
2516 if (board->int_act_high)
2517 hwcfg |= HCHWCFG_INT_POL;
2518 if (board->int_edge_triggered)
2519 hwcfg |= HCHWCFG_INT_TRIGGER;
2520 if (board->dreq_act_high)
2521 hwcfg |= HCHWCFG_DREQ_POL;
2522 if (board->dack_act_high)
2523 hwcfg |= HCHWCFG_DACK_POL;
2524 isp1362_write_reg16(isp1362_hcd, HCHWCFG, hwcfg);
2525 isp1362_show_reg(isp1362_hcd, HCHWCFG);
2526 isp1362_write_reg16(isp1362_hcd, HCDMACFG, 0);
2527 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2529 ret = isp1362_mem_config(hcd);
2533 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2536 isp1362_hcd->rhdesca = 0;
2537 if (board->no_power_switching)
2538 isp1362_hcd->rhdesca |= RH_A_NPS;
2539 if (board->power_switching_mode)
2540 isp1362_hcd->rhdesca |= RH_A_PSM;
2542 isp1362_hcd->rhdesca |= (board->potpg << 24) & RH_A_POTPGT;
2544 isp1362_hcd->rhdesca |= (25 << 24) & RH_A_POTPGT;
2546 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca & ~RH_A_OCPM);
2547 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca | RH_A_OCPM);
2548 isp1362_hcd->rhdesca = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2550 isp1362_hcd->rhdescb = RH_B_PPCM;
2551 isp1362_write_reg32(isp1362_hcd, HCRHDESCB, isp1362_hcd->rhdescb);
2552 isp1362_hcd->rhdescb = isp1362_read_reg32(isp1362_hcd, HCRHDESCB);
2554 isp1362_read_reg32(isp1362_hcd, HCFMINTVL);
2555 isp1362_write_reg32(isp1362_hcd, HCFMINTVL, (FSMP(FI) << 16) | FI);
2556 isp1362_write_reg32(isp1362_hcd, HCLSTHRESH, LSTHRESH);
2558 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2560 isp1362_hcd->hc_control = OHCI_USB_OPER;
2561 hcd->state = HC_STATE_RUNNING;
2563 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2564 /* Set up interrupts */
2565 isp1362_hcd->intenb = OHCI_INTR_MIE | OHCI_INTR_RHSC | OHCI_INTR_UE;
2566 isp1362_hcd->intenb |= OHCI_INTR_RD;
2567 isp1362_hcd->irqenb = HCuPINT_OPR | HCuPINT_SUSP;
2568 isp1362_write_reg32(isp1362_hcd, HCINTENB, isp1362_hcd->intenb);
2569 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
2571 /* Go operational */
2572 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
2573 /* enable global power */
2574 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC | RH_HS_DRWE);
2576 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2581 /*-------------------------------------------------------------------------*/
2583 static const struct hc_driver isp1362_hc_driver = {
2584 .description = hcd_name,
2585 .product_desc = "ISP1362 Host Controller",
2586 .hcd_priv_size = sizeof(struct isp1362_hcd),
2589 .flags = HCD_USB11 | HCD_MEMORY,
2591 .reset = isp1362_hc_reset,
2592 .start = isp1362_hc_start,
2593 .stop = isp1362_hc_stop,
2595 .urb_enqueue = isp1362_urb_enqueue,
2596 .urb_dequeue = isp1362_urb_dequeue,
2597 .endpoint_disable = isp1362_endpoint_disable,
2599 .get_frame_number = isp1362_get_frame,
2601 .hub_status_data = isp1362_hub_status_data,
2602 .hub_control = isp1362_hub_control,
2603 .bus_suspend = isp1362_bus_suspend,
2604 .bus_resume = isp1362_bus_resume,
2607 /*-------------------------------------------------------------------------*/
2609 static void isp1362_remove(struct platform_device *pdev)
2611 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2612 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2614 remove_debug_file(isp1362_hcd);
2615 DBG(0, "%s: Removing HCD\n", __func__);
2616 usb_remove_hcd(hcd);
2617 DBG(0, "%s: put_hcd\n", __func__);
2619 DBG(0, "%s: Done\n", __func__);
2622 static int isp1362_probe(struct platform_device *pdev)
2624 struct usb_hcd *hcd;
2625 struct isp1362_hcd *isp1362_hcd;
2626 struct resource *data, *irq_res;
2627 void __iomem *addr_reg;
2628 void __iomem *data_reg;
2631 unsigned int irq_flags = 0;
2636 /* basic sanity checks first. board-specific init logic should
2637 * have initialized this the three resources and probably board
2638 * specific platform_data. we don't probe for IRQs, and do only
2639 * minimal sanity checking.
2641 if (pdev->num_resources < 3)
2644 irq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
2648 irq = irq_res->start;
2650 addr_reg = devm_platform_ioremap_resource(pdev, 1);
2651 if (IS_ERR(addr_reg))
2652 return PTR_ERR(addr_reg);
2654 data_reg = devm_platform_get_and_ioremap_resource(pdev, 0, &data);
2655 if (IS_ERR(data_reg))
2656 return PTR_ERR(data_reg);
2658 /* allocate and initialize hcd */
2659 hcd = usb_create_hcd(&isp1362_hc_driver, &pdev->dev, dev_name(&pdev->dev));
2663 hcd->rsrc_start = data->start;
2664 isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2665 isp1362_hcd->data_reg = data_reg;
2666 isp1362_hcd->addr_reg = addr_reg;
2668 isp1362_hcd->next_statechange = jiffies;
2669 spin_lock_init(&isp1362_hcd->lock);
2670 INIT_LIST_HEAD(&isp1362_hcd->async);
2671 INIT_LIST_HEAD(&isp1362_hcd->periodic);
2672 INIT_LIST_HEAD(&isp1362_hcd->isoc);
2673 INIT_LIST_HEAD(&isp1362_hcd->remove_list);
2674 isp1362_hcd->board = dev_get_platdata(&pdev->dev);
2675 #if USE_PLATFORM_DELAY
2676 if (!isp1362_hcd->board->delay) {
2677 dev_err(hcd->self.controller, "No platform delay function given\n");
2683 if (irq_res->flags & IORESOURCE_IRQ_HIGHEDGE)
2684 irq_flags |= IRQF_TRIGGER_RISING;
2685 if (irq_res->flags & IORESOURCE_IRQ_LOWEDGE)
2686 irq_flags |= IRQF_TRIGGER_FALLING;
2687 if (irq_res->flags & IORESOURCE_IRQ_HIGHLEVEL)
2688 irq_flags |= IRQF_TRIGGER_HIGH;
2689 if (irq_res->flags & IORESOURCE_IRQ_LOWLEVEL)
2690 irq_flags |= IRQF_TRIGGER_LOW;
2692 retval = usb_add_hcd(hcd, irq, irq_flags | IRQF_SHARED);
2695 device_wakeup_enable(hcd->self.controller);
2697 dev_info(&pdev->dev, "%s, irq %d\n", hcd->product_desc, irq);
2699 create_debug_file(isp1362_hcd);
2710 static int isp1362_suspend(struct platform_device *pdev, pm_message_t state)
2712 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2713 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2714 unsigned long flags;
2717 DBG(0, "%s: Suspending device\n", __func__);
2719 if (state.event == PM_EVENT_FREEZE) {
2720 DBG(0, "%s: Suspending root hub\n", __func__);
2721 retval = isp1362_bus_suspend(hcd);
2723 DBG(0, "%s: Suspending RH ports\n", __func__);
2724 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2725 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2726 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2729 pdev->dev.power.power_state = state;
2733 static int isp1362_resume(struct platform_device *pdev)
2735 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2736 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2737 unsigned long flags;
2739 DBG(0, "%s: Resuming\n", __func__);
2741 if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
2742 DBG(0, "%s: Resume RH ports\n", __func__);
2743 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2744 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC);
2745 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2749 pdev->dev.power.power_state = PMSG_ON;
2751 return isp1362_bus_resume(isp1362_hcd_to_hcd(isp1362_hcd));
2754 #define isp1362_suspend NULL
2755 #define isp1362_resume NULL
2758 static struct platform_driver isp1362_driver = {
2759 .probe = isp1362_probe,
2760 .remove_new = isp1362_remove,
2762 .suspend = isp1362_suspend,
2763 .resume = isp1362_resume,
2769 module_platform_driver(isp1362_driver);