Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * ISP1362 HCD (Host Controller Driver) for USB.
4 *
5 * Copyright (C) 2005 Lothar Wassmann <LW@KARO-electronics.de>
6 *
7 * Derived from the SL811 HCD, rewritten for ISP116x.
8 * Copyright (C) 2005 Olav Kongas <ok@artecdesign.ee>
9 *
10 * Portions:
11 * Copyright (C) 2004 Psion Teklogix (for NetBook PRO)
12 * Copyright (C) 2004 David Brownell
13 */
14
15/*
16 * The ISP1362 chip requires a large delay (300ns and 462ns) between
17 * accesses to the address and data register.
18 * The following timing options exist:
19 *
20 * 1. Configure your memory controller to add such delays if it can (the best)
21 * 2. Implement platform-specific delay function possibly
22 * combined with configuring the memory controller; see
23 * include/linux/usb_isp1362.h for more info.
24 * 3. Use ndelay (easiest, poorest).
25 *
26 * Use the corresponding macros USE_PLATFORM_DELAY and USE_NDELAY in the
27 * platform specific section of isp1362.h to select the appropriate variant.
28 *
29 * Also note that according to the Philips "ISP1362 Errata" document
30 * Rev 1.00 from 27 May data corruption may occur when the #WR signal
31 * is reasserted (even with #CS deasserted) within 132ns after a
32 * write cycle to any controller register. If the hardware doesn't
33 * implement the recommended fix (gating the #WR with #CS) software
34 * must ensure that no further write cycle (not necessarily to the chip!)
35 * is issued by the CPU within this interval.
36
37 * For PXA25x this can be ensured by using VLIO with the maximum
38 * recovery time (MSCx = 0x7f8c) with a memory clock of 99.53 MHz.
39 */
40
41#undef ISP1362_DEBUG
42
43/*
44 * The PXA255 UDC apparently doesn't handle GET_STATUS, GET_CONFIG and
45 * GET_INTERFACE requests correctly when the SETUP and DATA stages of the
46 * requests are carried out in separate frames. This will delay any SETUP
47 * packets until the start of the next frame so that this situation is
48 * unlikely to occur (and makes usbtest happy running with a PXA255 target
49 * device).
50 */
51#undef BUGGY_PXA2XX_UDC_USBTEST
52
53#undef PTD_TRACE
54#undef URB_TRACE
55#undef VERBOSE
56#undef REGISTERS
57
58/* This enables a memory test on the ISP1362 chip memory to make sure the
59 * chip access timing is correct.
60 */
61#undef CHIP_BUFFER_TEST
62
63#include <linux/module.h>
64#include <linux/moduleparam.h>
65#include <linux/kernel.h>
66#include <linux/delay.h>
67#include <linux/ioport.h>
68#include <linux/sched.h>
69#include <linux/slab.h>
70#include <linux/errno.h>
71#include <linux/list.h>
72#include <linux/interrupt.h>
73#include <linux/usb.h>
74#include <linux/usb/isp1362.h>
75#include <linux/usb/hcd.h>
76#include <linux/platform_device.h>
77#include <linux/pm.h>
78#include <linux/io.h>
79#include <linux/bitmap.h>
80#include <linux/prefetch.h>
81#include <linux/debugfs.h>
82#include <linux/seq_file.h>
83
84#include <asm/irq.h>
85#include <asm/byteorder.h>
86#include <asm/unaligned.h>
87
88static int dbg_level;
89#ifdef ISP1362_DEBUG
90module_param(dbg_level, int, 0644);
91#else
92module_param(dbg_level, int, 0);
93#endif
94
95#include "../core/usb.h"
96#include "isp1362.h"
97
98
99#define DRIVER_VERSION "2005-04-04"
100#define DRIVER_DESC "ISP1362 USB Host Controller Driver"
101
102MODULE_DESCRIPTION(DRIVER_DESC);
103MODULE_LICENSE("GPL");
104
105static const char hcd_name[] = "isp1362-hcd";
106
107static void isp1362_hc_stop(struct usb_hcd *hcd);
108static int isp1362_hc_start(struct usb_hcd *hcd);
109
110/*-------------------------------------------------------------------------*/
111
112/*
113 * When called from the interrupthandler only isp1362_hcd->irqenb is modified,
114 * since the interrupt handler will write isp1362_hcd->irqenb to HCuPINT upon
115 * completion.
116 * We don't need a 'disable' counterpart, since interrupts will be disabled
117 * only by the interrupt handler.
118 */
119static inline void isp1362_enable_int(struct isp1362_hcd *isp1362_hcd, u16 mask)
120{
121 if ((isp1362_hcd->irqenb | mask) == isp1362_hcd->irqenb)
122 return;
123 if (mask & ~isp1362_hcd->irqenb)
124 isp1362_write_reg16(isp1362_hcd, HCuPINT, mask & ~isp1362_hcd->irqenb);
125 isp1362_hcd->irqenb |= mask;
126 if (isp1362_hcd->irq_active)
127 return;
128 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
129}
130
131/*-------------------------------------------------------------------------*/
132
133static inline struct isp1362_ep_queue *get_ptd_queue(struct isp1362_hcd *isp1362_hcd,
134 u16 offset)
135{
136 struct isp1362_ep_queue *epq = NULL;
137
138 if (offset < isp1362_hcd->istl_queue[1].buf_start)
139 epq = &isp1362_hcd->istl_queue[0];
140 else if (offset < isp1362_hcd->intl_queue.buf_start)
141 epq = &isp1362_hcd->istl_queue[1];
142 else if (offset < isp1362_hcd->atl_queue.buf_start)
143 epq = &isp1362_hcd->intl_queue;
144 else if (offset < isp1362_hcd->atl_queue.buf_start +
145 isp1362_hcd->atl_queue.buf_size)
146 epq = &isp1362_hcd->atl_queue;
147
148 if (epq)
149 DBG(1, "%s: PTD $%04x is on %s queue\n", __func__, offset, epq->name);
150 else
151 pr_warn("%s: invalid PTD $%04x\n", __func__, offset);
152
153 return epq;
154}
155
156static inline int get_ptd_offset(struct isp1362_ep_queue *epq, u8 index)
157{
158 int offset;
159
160 if (index * epq->blk_size > epq->buf_size) {
161 pr_warn("%s: Bad %s index %d(%d)\n",
162 __func__, epq->name, index,
163 epq->buf_size / epq->blk_size);
164 return -EINVAL;
165 }
166 offset = epq->buf_start + index * epq->blk_size;
167 DBG(3, "%s: %s PTD[%02x] # %04x\n", __func__, epq->name, index, offset);
168
169 return offset;
170}
171
172/*-------------------------------------------------------------------------*/
173
174static inline u16 max_transfer_size(struct isp1362_ep_queue *epq, size_t size,
175 int mps)
176{
177 u16 xfer_size = min_t(size_t, MAX_XFER_SIZE, size);
178
179 xfer_size = min_t(size_t, xfer_size, epq->buf_avail * epq->blk_size - PTD_HEADER_SIZE);
180 if (xfer_size < size && xfer_size % mps)
181 xfer_size -= xfer_size % mps;
182
183 return xfer_size;
184}
185
186static int claim_ptd_buffers(struct isp1362_ep_queue *epq,
187 struct isp1362_ep *ep, u16 len)
188{
189 int ptd_offset = -EINVAL;
190 int num_ptds = ((len + PTD_HEADER_SIZE - 1) / epq->blk_size) + 1;
191 int found;
192
193 BUG_ON(len > epq->buf_size);
194
195 if (!epq->buf_avail)
196 return -ENOMEM;
197
198 if (ep->num_ptds)
199 pr_err("%s: %s len %d/%d num_ptds %d buf_map %08lx skip_map %08lx\n", __func__,
200 epq->name, len, epq->blk_size, num_ptds, epq->buf_map, epq->skip_map);
201 BUG_ON(ep->num_ptds != 0);
202
203 found = bitmap_find_next_zero_area(&epq->buf_map, epq->buf_count, 0,
204 num_ptds, 0);
205 if (found >= epq->buf_count)
206 return -EOVERFLOW;
207
208 DBG(1, "%s: Found %d PTDs[%d] for %d/%d byte\n", __func__,
209 num_ptds, found, len, (int)(epq->blk_size - PTD_HEADER_SIZE));
210 ptd_offset = get_ptd_offset(epq, found);
211 WARN_ON(ptd_offset < 0);
212 ep->ptd_offset = ptd_offset;
213 ep->num_ptds += num_ptds;
214 epq->buf_avail -= num_ptds;
215 BUG_ON(epq->buf_avail > epq->buf_count);
216 ep->ptd_index = found;
217 bitmap_set(&epq->buf_map, found, num_ptds);
218 DBG(1, "%s: Done %s PTD[%d] $%04x, avail %d count %d claimed %d %08lx:%08lx\n",
219 __func__, epq->name, ep->ptd_index, ep->ptd_offset,
220 epq->buf_avail, epq->buf_count, num_ptds, epq->buf_map, epq->skip_map);
221
222 return found;
223}
224
225static inline void release_ptd_buffers(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
226{
227 int last = ep->ptd_index + ep->num_ptds;
228
229 if (last > epq->buf_count)
230 pr_err("%s: ep %p req %d len %d %s PTD[%d] $%04x num_ptds %d buf_count %d buf_avail %d buf_map %08lx skip_map %08lx\n",
231 __func__, ep, ep->num_req, ep->length, epq->name, ep->ptd_index,
232 ep->ptd_offset, ep->num_ptds, epq->buf_count, epq->buf_avail,
233 epq->buf_map, epq->skip_map);
234 BUG_ON(last > epq->buf_count);
235
236 bitmap_clear(&epq->buf_map, ep->ptd_index, ep->num_ptds);
237 bitmap_set(&epq->skip_map, ep->ptd_index, ep->num_ptds);
238 epq->buf_avail += ep->num_ptds;
239 epq->ptd_count--;
240
241 BUG_ON(epq->buf_avail > epq->buf_count);
242 BUG_ON(epq->ptd_count > epq->buf_count);
243
244 DBG(1, "%s: Done %s PTDs $%04x released %d avail %d count %d\n",
245 __func__, epq->name,
246 ep->ptd_offset, ep->num_ptds, epq->buf_avail, epq->buf_count);
247 DBG(1, "%s: buf_map %08lx skip_map %08lx\n", __func__,
248 epq->buf_map, epq->skip_map);
249
250 ep->num_ptds = 0;
251 ep->ptd_offset = -EINVAL;
252 ep->ptd_index = -EINVAL;
253}
254
255/*-------------------------------------------------------------------------*/
256
257/*
258 Set up PTD's.
259*/
260static void prepare_ptd(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
261 struct isp1362_ep *ep, struct isp1362_ep_queue *epq,
262 u16 fno)
263{
264 struct ptd *ptd;
265 int toggle;
266 int dir;
267 u16 len;
268 size_t buf_len = urb->transfer_buffer_length - urb->actual_length;
269
270 DBG(3, "%s: %s ep %p\n", __func__, epq->name, ep);
271
272 ptd = &ep->ptd;
273
274 ep->data = (unsigned char *)urb->transfer_buffer + urb->actual_length;
275
276 switch (ep->nextpid) {
277 case USB_PID_IN:
278 toggle = usb_gettoggle(urb->dev, ep->epnum, 0);
279 dir = PTD_DIR_IN;
280 if (usb_pipecontrol(urb->pipe)) {
281 len = min_t(size_t, ep->maxpacket, buf_len);
282 } else if (usb_pipeisoc(urb->pipe)) {
283 len = min_t(size_t, urb->iso_frame_desc[fno].length, MAX_XFER_SIZE);
284 ep->data = urb->transfer_buffer + urb->iso_frame_desc[fno].offset;
285 } else
286 len = max_transfer_size(epq, buf_len, ep->maxpacket);
287 DBG(1, "%s: IN len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
288 (int)buf_len);
289 break;
290 case USB_PID_OUT:
291 toggle = usb_gettoggle(urb->dev, ep->epnum, 1);
292 dir = PTD_DIR_OUT;
293 if (usb_pipecontrol(urb->pipe))
294 len = min_t(size_t, ep->maxpacket, buf_len);
295 else if (usb_pipeisoc(urb->pipe))
296 len = min_t(size_t, urb->iso_frame_desc[0].length, MAX_XFER_SIZE);
297 else
298 len = max_transfer_size(epq, buf_len, ep->maxpacket);
299 if (len == 0)
300 pr_info("%s: Sending ZERO packet: %d\n", __func__,
301 urb->transfer_flags & URB_ZERO_PACKET);
302 DBG(1, "%s: OUT len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
303 (int)buf_len);
304 break;
305 case USB_PID_SETUP:
306 toggle = 0;
307 dir = PTD_DIR_SETUP;
308 len = sizeof(struct usb_ctrlrequest);
309 DBG(1, "%s: SETUP len %d\n", __func__, len);
310 ep->data = urb->setup_packet;
311 break;
312 case USB_PID_ACK:
313 toggle = 1;
314 len = 0;
315 dir = (urb->transfer_buffer_length && usb_pipein(urb->pipe)) ?
316 PTD_DIR_OUT : PTD_DIR_IN;
317 DBG(1, "%s: ACK len %d\n", __func__, len);
318 break;
319 default:
320 toggle = dir = len = 0;
321 pr_err("%s@%d: ep->nextpid %02x\n", __func__, __LINE__, ep->nextpid);
322 BUG_ON(1);
323 }
324
325 ep->length = len;
326 if (!len)
327 ep->data = NULL;
328
329 ptd->count = PTD_CC_MSK | PTD_ACTIVE_MSK | PTD_TOGGLE(toggle);
330 ptd->mps = PTD_MPS(ep->maxpacket) | PTD_SPD(urb->dev->speed == USB_SPEED_LOW) |
331 PTD_EP(ep->epnum);
332 ptd->len = PTD_LEN(len) | PTD_DIR(dir);
333 ptd->faddr = PTD_FA(usb_pipedevice(urb->pipe));
334
335 if (usb_pipeint(urb->pipe)) {
336 ptd->faddr |= PTD_SF_INT(ep->branch);
337 ptd->faddr |= PTD_PR(ep->interval ? __ffs(ep->interval) : 0);
338 }
339 if (usb_pipeisoc(urb->pipe))
340 ptd->faddr |= PTD_SF_ISO(fno);
341
342 DBG(1, "%s: Finished\n", __func__);
343}
344
345static void isp1362_write_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
346 struct isp1362_ep_queue *epq)
347{
348 struct ptd *ptd = &ep->ptd;
349 int len = PTD_GET_DIR(ptd) == PTD_DIR_IN ? 0 : ep->length;
350
351 prefetch(ptd);
352 isp1362_write_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
353 if (len)
354 isp1362_write_buffer(isp1362_hcd, ep->data,
355 ep->ptd_offset + PTD_HEADER_SIZE, len);
356
357 dump_ptd(ptd);
358 dump_ptd_out_data(ptd, ep->data);
359}
360
361static void isp1362_read_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
362 struct isp1362_ep_queue *epq)
363{
364 struct ptd *ptd = &ep->ptd;
365 int act_len;
366
367 WARN_ON(list_empty(&ep->active));
368 BUG_ON(ep->ptd_offset < 0);
369
370 list_del_init(&ep->active);
371 DBG(1, "%s: ep %p removed from active list %p\n", __func__, ep, &epq->active);
372
373 prefetchw(ptd);
374 isp1362_read_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
375 dump_ptd(ptd);
376 act_len = PTD_GET_COUNT(ptd);
377 if (PTD_GET_DIR(ptd) != PTD_DIR_IN || act_len == 0)
378 return;
379 if (act_len > ep->length)
380 pr_err("%s: ep %p PTD $%04x act_len %d ep->length %d\n", __func__, ep,
381 ep->ptd_offset, act_len, ep->length);
382 BUG_ON(act_len > ep->length);
383 /* Only transfer the amount of data that has actually been overwritten
384 * in the chip buffer. We don't want any data that doesn't belong to the
385 * transfer to leak out of the chip to the callers transfer buffer!
386 */
387 prefetchw(ep->data);
388 isp1362_read_buffer(isp1362_hcd, ep->data,
389 ep->ptd_offset + PTD_HEADER_SIZE, act_len);
390 dump_ptd_in_data(ptd, ep->data);
391}
392
393/*
394 * INT PTDs will stay in the chip until data is available.
395 * This function will remove a PTD from the chip when the URB is dequeued.
396 * Must be called with the spinlock held and IRQs disabled
397 */
398static void remove_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
399
400{
401 int index;
402 struct isp1362_ep_queue *epq;
403
404 DBG(1, "%s: ep %p PTD[%d] $%04x\n", __func__, ep, ep->ptd_index, ep->ptd_offset);
405 BUG_ON(ep->ptd_offset < 0);
406
407 epq = get_ptd_queue(isp1362_hcd, ep->ptd_offset);
408 BUG_ON(!epq);
409
410 /* put ep in remove_list for cleanup */
411 WARN_ON(!list_empty(&ep->remove_list));
412 list_add_tail(&ep->remove_list, &isp1362_hcd->remove_list);
413 /* let SOF interrupt handle the cleanup */
414 isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
415
416 index = ep->ptd_index;
417 if (index < 0)
418 /* ISO queues don't have SKIP registers */
419 return;
420
421 DBG(1, "%s: Disabling PTD[%02x] $%04x %08lx|%08x\n", __func__,
422 index, ep->ptd_offset, epq->skip_map, 1 << index);
423
424 /* prevent further processing of PTD (will be effective after next SOF) */
425 epq->skip_map |= 1 << index;
426 if (epq == &isp1362_hcd->atl_queue) {
427 DBG(2, "%s: ATLSKIP = %08x -> %08lx\n", __func__,
428 isp1362_read_reg32(isp1362_hcd, HCATLSKIP), epq->skip_map);
429 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, epq->skip_map);
430 if (~epq->skip_map == 0)
431 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
432 } else if (epq == &isp1362_hcd->intl_queue) {
433 DBG(2, "%s: INTLSKIP = %08x -> %08lx\n", __func__,
434 isp1362_read_reg32(isp1362_hcd, HCINTLSKIP), epq->skip_map);
435 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, epq->skip_map);
436 if (~epq->skip_map == 0)
437 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
438 }
439}
440
441/*
442 Take done or failed requests out of schedule. Give back
443 processed urbs.
444*/
445static void finish_request(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
446 struct urb *urb, int status)
447 __releases(isp1362_hcd->lock)
448 __acquires(isp1362_hcd->lock)
449{
450 urb->hcpriv = NULL;
451 ep->error_count = 0;
452
453 if (usb_pipecontrol(urb->pipe))
454 ep->nextpid = USB_PID_SETUP;
455
456 URB_DBG("%s: req %d FA %d ep%d%s %s: len %d/%d %s stat %d\n", __func__,
457 ep->num_req, usb_pipedevice(urb->pipe),
458 usb_pipeendpoint(urb->pipe),
459 !usb_pipein(urb->pipe) ? "out" : "in",
460 usb_pipecontrol(urb->pipe) ? "ctrl" :
461 usb_pipeint(urb->pipe) ? "int" :
462 usb_pipebulk(urb->pipe) ? "bulk" :
463 "iso",
464 urb->actual_length, urb->transfer_buffer_length,
465 !(urb->transfer_flags & URB_SHORT_NOT_OK) ?
466 "short_ok" : "", urb->status);
467
468
469 usb_hcd_unlink_urb_from_ep(isp1362_hcd_to_hcd(isp1362_hcd), urb);
470 spin_unlock(&isp1362_hcd->lock);
471 usb_hcd_giveback_urb(isp1362_hcd_to_hcd(isp1362_hcd), urb, status);
472 spin_lock(&isp1362_hcd->lock);
473
474 /* take idle endpoints out of the schedule right away */
475 if (!list_empty(&ep->hep->urb_list))
476 return;
477
478 /* async deschedule */
479 if (!list_empty(&ep->schedule)) {
480 list_del_init(&ep->schedule);
481 return;
482 }
483
484
485 if (ep->interval) {
486 /* periodic deschedule */
487 DBG(1, "deschedule qh%d/%p branch %d load %d bandwidth %d -> %d\n", ep->interval,
488 ep, ep->branch, ep->load,
489 isp1362_hcd->load[ep->branch],
490 isp1362_hcd->load[ep->branch] - ep->load);
491 isp1362_hcd->load[ep->branch] -= ep->load;
492 ep->branch = PERIODIC_SIZE;
493 }
494}
495
496/*
497 * Analyze transfer results, handle partial transfers and errors
498*/
499static void postproc_ep(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
500{
501 struct urb *urb = get_urb(ep);
502 struct usb_device *udev;
503 struct ptd *ptd;
504 int short_ok;
505 u16 len;
506 int urbstat = -EINPROGRESS;
507 u8 cc;
508
509 DBG(2, "%s: ep %p req %d\n", __func__, ep, ep->num_req);
510
511 udev = urb->dev;
512 ptd = &ep->ptd;
513 cc = PTD_GET_CC(ptd);
514 if (cc == PTD_NOTACCESSED) {
515 pr_err("%s: req %d PTD %p Untouched by ISP1362\n", __func__,
516 ep->num_req, ptd);
517 cc = PTD_DEVNOTRESP;
518 }
519
520 short_ok = !(urb->transfer_flags & URB_SHORT_NOT_OK);
521 len = urb->transfer_buffer_length - urb->actual_length;
522
523 /* Data underrun is special. For allowed underrun
524 we clear the error and continue as normal. For
525 forbidden underrun we finish the DATA stage
526 immediately while for control transfer,
527 we do a STATUS stage.
528 */
529 if (cc == PTD_DATAUNDERRUN) {
530 if (short_ok) {
531 DBG(1, "%s: req %d Allowed data underrun short_%sok %d/%d/%d byte\n",
532 __func__, ep->num_req, short_ok ? "" : "not_",
533 PTD_GET_COUNT(ptd), ep->maxpacket, len);
534 cc = PTD_CC_NOERROR;
535 urbstat = 0;
536 } else {
537 DBG(1, "%s: req %d Data Underrun %s nextpid %02x short_%sok %d/%d/%d byte\n",
538 __func__, ep->num_req,
539 usb_pipein(urb->pipe) ? "IN" : "OUT", ep->nextpid,
540 short_ok ? "" : "not_",
541 PTD_GET_COUNT(ptd), ep->maxpacket, len);
542 /* save the data underrun error code for later and
543 * proceed with the status stage
544 */
545 urb->actual_length += PTD_GET_COUNT(ptd);
546 if (usb_pipecontrol(urb->pipe)) {
547 ep->nextpid = USB_PID_ACK;
548 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
549
550 if (urb->status == -EINPROGRESS)
551 urb->status = cc_to_error[PTD_DATAUNDERRUN];
552 } else {
553 usb_settoggle(udev, ep->epnum, ep->nextpid == USB_PID_OUT,
554 PTD_GET_TOGGLE(ptd));
555 urbstat = cc_to_error[PTD_DATAUNDERRUN];
556 }
557 goto out;
558 }
559 }
560
561 if (cc != PTD_CC_NOERROR) {
562 if (++ep->error_count >= 3 || cc == PTD_CC_STALL || cc == PTD_DATAOVERRUN) {
563 urbstat = cc_to_error[cc];
564 DBG(1, "%s: req %d nextpid %02x, status %d, error %d, error_count %d\n",
565 __func__, ep->num_req, ep->nextpid, urbstat, cc,
566 ep->error_count);
567 }
568 goto out;
569 }
570
571 switch (ep->nextpid) {
572 case USB_PID_OUT:
573 if (PTD_GET_COUNT(ptd) != ep->length)
574 pr_err("%s: count=%d len=%d\n", __func__,
575 PTD_GET_COUNT(ptd), ep->length);
576 BUG_ON(PTD_GET_COUNT(ptd) != ep->length);
577 urb->actual_length += ep->length;
578 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
579 usb_settoggle(udev, ep->epnum, 1, PTD_GET_TOGGLE(ptd));
580 if (urb->actual_length == urb->transfer_buffer_length) {
581 DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
582 ep->num_req, len, ep->maxpacket, urbstat);
583 if (usb_pipecontrol(urb->pipe)) {
584 DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
585 ep->num_req,
586 usb_pipein(urb->pipe) ? "IN" : "OUT");
587 ep->nextpid = USB_PID_ACK;
588 } else {
589 if (len % ep->maxpacket ||
590 !(urb->transfer_flags & URB_ZERO_PACKET)) {
591 urbstat = 0;
592 DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
593 __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
594 urbstat, len, ep->maxpacket, urb->actual_length);
595 }
596 }
597 }
598 break;
599 case USB_PID_IN:
600 len = PTD_GET_COUNT(ptd);
601 BUG_ON(len > ep->length);
602 urb->actual_length += len;
603 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
604 usb_settoggle(udev, ep->epnum, 0, PTD_GET_TOGGLE(ptd));
605 /* if transfer completed or (allowed) data underrun */
606 if ((urb->transfer_buffer_length == urb->actual_length) ||
607 len % ep->maxpacket) {
608 DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
609 ep->num_req, len, ep->maxpacket, urbstat);
610 if (usb_pipecontrol(urb->pipe)) {
611 DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
612 ep->num_req,
613 usb_pipein(urb->pipe) ? "IN" : "OUT");
614 ep->nextpid = USB_PID_ACK;
615 } else {
616 urbstat = 0;
617 DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
618 __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
619 urbstat, len, ep->maxpacket, urb->actual_length);
620 }
621 }
622 break;
623 case USB_PID_SETUP:
624 if (urb->transfer_buffer_length == urb->actual_length) {
625 ep->nextpid = USB_PID_ACK;
626 } else if (usb_pipeout(urb->pipe)) {
627 usb_settoggle(udev, 0, 1, 1);
628 ep->nextpid = USB_PID_OUT;
629 } else {
630 usb_settoggle(udev, 0, 0, 1);
631 ep->nextpid = USB_PID_IN;
632 }
633 break;
634 case USB_PID_ACK:
635 DBG(3, "%s: req %d got ACK %d -> 0\n", __func__, ep->num_req,
636 urbstat);
637 WARN_ON(urbstat != -EINPROGRESS);
638 urbstat = 0;
639 ep->nextpid = 0;
640 break;
641 default:
642 BUG_ON(1);
643 }
644
645 out:
646 if (urbstat != -EINPROGRESS) {
647 DBG(2, "%s: Finishing ep %p req %d urb %p status %d\n", __func__,
648 ep, ep->num_req, urb, urbstat);
649 finish_request(isp1362_hcd, ep, urb, urbstat);
650 }
651}
652
653static void finish_unlinks(struct isp1362_hcd *isp1362_hcd)
654{
655 struct isp1362_ep *ep;
656 struct isp1362_ep *tmp;
657
658 list_for_each_entry_safe(ep, tmp, &isp1362_hcd->remove_list, remove_list) {
659 struct isp1362_ep_queue *epq =
660 get_ptd_queue(isp1362_hcd, ep->ptd_offset);
661 int index = ep->ptd_index;
662
663 BUG_ON(epq == NULL);
664 if (index >= 0) {
665 DBG(1, "%s: remove PTD[%d] $%04x\n", __func__, index, ep->ptd_offset);
666 BUG_ON(ep->num_ptds == 0);
667 release_ptd_buffers(epq, ep);
668 }
669 if (!list_empty(&ep->hep->urb_list)) {
670 struct urb *urb = get_urb(ep);
671
672 DBG(1, "%s: Finishing req %d ep %p from remove_list\n", __func__,
673 ep->num_req, ep);
674 finish_request(isp1362_hcd, ep, urb, -ESHUTDOWN);
675 }
676 WARN_ON(list_empty(&ep->active));
677 if (!list_empty(&ep->active)) {
678 list_del_init(&ep->active);
679 DBG(1, "%s: ep %p removed from active list\n", __func__, ep);
680 }
681 list_del_init(&ep->remove_list);
682 DBG(1, "%s: ep %p removed from remove_list\n", __func__, ep);
683 }
684 DBG(1, "%s: Done\n", __func__);
685}
686
687static inline void enable_atl_transfers(struct isp1362_hcd *isp1362_hcd, int count)
688{
689 if (count > 0) {
690 if (count < isp1362_hcd->atl_queue.ptd_count)
691 isp1362_write_reg16(isp1362_hcd, HCATLDTC, count);
692 isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
693 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, isp1362_hcd->atl_queue.skip_map);
694 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
695 } else
696 isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
697}
698
699static inline void enable_intl_transfers(struct isp1362_hcd *isp1362_hcd)
700{
701 isp1362_enable_int(isp1362_hcd, HCuPINT_INTL);
702 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
703 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, isp1362_hcd->intl_queue.skip_map);
704}
705
706static inline void enable_istl_transfers(struct isp1362_hcd *isp1362_hcd, int flip)
707{
708 isp1362_enable_int(isp1362_hcd, flip ? HCuPINT_ISTL1 : HCuPINT_ISTL0);
709 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, flip ?
710 HCBUFSTAT_ISTL1_FULL : HCBUFSTAT_ISTL0_FULL);
711}
712
713static int submit_req(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
714 struct isp1362_ep *ep, struct isp1362_ep_queue *epq)
715{
716 int index;
717
718 prepare_ptd(isp1362_hcd, urb, ep, epq, 0);
719 index = claim_ptd_buffers(epq, ep, ep->length);
720 if (index == -ENOMEM) {
721 DBG(1, "%s: req %d No free %s PTD available: %d, %08lx:%08lx\n", __func__,
722 ep->num_req, epq->name, ep->num_ptds, epq->buf_map, epq->skip_map);
723 return index;
724 } else if (index == -EOVERFLOW) {
725 DBG(1, "%s: req %d Not enough space for %d byte %s PTD %d %08lx:%08lx\n",
726 __func__, ep->num_req, ep->length, epq->name, ep->num_ptds,
727 epq->buf_map, epq->skip_map);
728 return index;
729 } else
730 BUG_ON(index < 0);
731 list_add_tail(&ep->active, &epq->active);
732 DBG(1, "%s: ep %p req %d len %d added to active list %p\n", __func__,
733 ep, ep->num_req, ep->length, &epq->active);
734 DBG(1, "%s: Submitting %s PTD $%04x for ep %p req %d\n", __func__, epq->name,
735 ep->ptd_offset, ep, ep->num_req);
736 isp1362_write_ptd(isp1362_hcd, ep, epq);
737 __clear_bit(ep->ptd_index, &epq->skip_map);
738
739 return 0;
740}
741
742static void start_atl_transfers(struct isp1362_hcd *isp1362_hcd)
743{
744 int ptd_count = 0;
745 struct isp1362_ep_queue *epq = &isp1362_hcd->atl_queue;
746 struct isp1362_ep *ep;
747 int defer = 0;
748
749 if (atomic_read(&epq->finishing)) {
750 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
751 return;
752 }
753
754 list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
755 struct urb *urb = get_urb(ep);
756 int ret;
757
758 if (!list_empty(&ep->active)) {
759 DBG(2, "%s: Skipping active %s ep %p\n", __func__, epq->name, ep);
760 continue;
761 }
762
763 DBG(1, "%s: Processing %s ep %p req %d\n", __func__, epq->name,
764 ep, ep->num_req);
765
766 ret = submit_req(isp1362_hcd, urb, ep, epq);
767 if (ret == -ENOMEM) {
768 defer = 1;
769 break;
770 } else if (ret == -EOVERFLOW) {
771 defer = 1;
772 continue;
773 }
774#ifdef BUGGY_PXA2XX_UDC_USBTEST
775 defer = ep->nextpid == USB_PID_SETUP;
776#endif
777 ptd_count++;
778 }
779
780 /* Avoid starving of endpoints */
781 if (isp1362_hcd->async.next != isp1362_hcd->async.prev) {
782 DBG(2, "%s: Cycling ASYNC schedule %d\n", __func__, ptd_count);
783 list_move(&isp1362_hcd->async, isp1362_hcd->async.next);
784 }
785 if (ptd_count || defer)
786 enable_atl_transfers(isp1362_hcd, defer ? 0 : ptd_count);
787
788 epq->ptd_count += ptd_count;
789 if (epq->ptd_count > epq->stat_maxptds) {
790 epq->stat_maxptds = epq->ptd_count;
791 DBG(0, "%s: max_ptds: %d\n", __func__, epq->stat_maxptds);
792 }
793}
794
795static void start_intl_transfers(struct isp1362_hcd *isp1362_hcd)
796{
797 int ptd_count = 0;
798 struct isp1362_ep_queue *epq = &isp1362_hcd->intl_queue;
799 struct isp1362_ep *ep;
800
801 if (atomic_read(&epq->finishing)) {
802 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
803 return;
804 }
805
806 list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
807 struct urb *urb = get_urb(ep);
808 int ret;
809
810 if (!list_empty(&ep->active)) {
811 DBG(1, "%s: Skipping active %s ep %p\n", __func__,
812 epq->name, ep);
813 continue;
814 }
815
816 DBG(1, "%s: Processing %s ep %p req %d\n", __func__,
817 epq->name, ep, ep->num_req);
818 ret = submit_req(isp1362_hcd, urb, ep, epq);
819 if (ret == -ENOMEM)
820 break;
821 else if (ret == -EOVERFLOW)
822 continue;
823 ptd_count++;
824 }
825
826 if (ptd_count) {
827 static int last_count;
828
829 if (ptd_count != last_count) {
830 DBG(0, "%s: ptd_count: %d\n", __func__, ptd_count);
831 last_count = ptd_count;
832 }
833 enable_intl_transfers(isp1362_hcd);
834 }
835
836 epq->ptd_count += ptd_count;
837 if (epq->ptd_count > epq->stat_maxptds)
838 epq->stat_maxptds = epq->ptd_count;
839}
840
841static inline int next_ptd(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
842{
843 u16 ptd_offset = ep->ptd_offset;
844 int num_ptds = (ep->length + PTD_HEADER_SIZE + (epq->blk_size - 1)) / epq->blk_size;
845
846 DBG(2, "%s: PTD offset $%04x + %04x => %d * %04x -> $%04x\n", __func__, ptd_offset,
847 ep->length, num_ptds, epq->blk_size, ptd_offset + num_ptds * epq->blk_size);
848
849 ptd_offset += num_ptds * epq->blk_size;
850 if (ptd_offset < epq->buf_start + epq->buf_size)
851 return ptd_offset;
852 else
853 return -ENOMEM;
854}
855
856static void start_iso_transfers(struct isp1362_hcd *isp1362_hcd)
857{
858 int ptd_count = 0;
859 int flip = isp1362_hcd->istl_flip;
860 struct isp1362_ep_queue *epq;
861 int ptd_offset;
862 struct isp1362_ep *ep;
863 struct isp1362_ep *tmp;
864 u16 fno = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
865
866 fill2:
867 epq = &isp1362_hcd->istl_queue[flip];
868 if (atomic_read(&epq->finishing)) {
869 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
870 return;
871 }
872
873 if (!list_empty(&epq->active))
874 return;
875
876 ptd_offset = epq->buf_start;
877 list_for_each_entry_safe(ep, tmp, &isp1362_hcd->isoc, schedule) {
878 struct urb *urb = get_urb(ep);
879 s16 diff = fno - (u16)urb->start_frame;
880
881 DBG(1, "%s: Processing %s ep %p\n", __func__, epq->name, ep);
882
883 if (diff > urb->number_of_packets) {
884 /* time frame for this URB has elapsed */
885 finish_request(isp1362_hcd, ep, urb, -EOVERFLOW);
886 continue;
887 } else if (diff < -1) {
888 /* URB is not due in this frame or the next one.
889 * Comparing with '-1' instead of '0' accounts for double
890 * buffering in the ISP1362 which enables us to queue the PTD
891 * one frame ahead of time
892 */
893 } else if (diff == -1) {
894 /* submit PTD's that are due in the next frame */
895 prepare_ptd(isp1362_hcd, urb, ep, epq, fno);
896 if (ptd_offset + PTD_HEADER_SIZE + ep->length >
897 epq->buf_start + epq->buf_size) {
898 pr_err("%s: Not enough ISO buffer space for %d byte PTD\n",
899 __func__, ep->length);
900 continue;
901 }
902 ep->ptd_offset = ptd_offset;
903 list_add_tail(&ep->active, &epq->active);
904
905 ptd_offset = next_ptd(epq, ep);
906 if (ptd_offset < 0) {
907 pr_warn("%s: req %d No more %s PTD buffers available\n",
908 __func__, ep->num_req, epq->name);
909 break;
910 }
911 }
912 }
913 list_for_each_entry(ep, &epq->active, active) {
914 if (epq->active.next == &ep->active)
915 ep->ptd.mps |= PTD_LAST_MSK;
916 isp1362_write_ptd(isp1362_hcd, ep, epq);
917 ptd_count++;
918 }
919
920 if (ptd_count)
921 enable_istl_transfers(isp1362_hcd, flip);
922
923 epq->ptd_count += ptd_count;
924 if (epq->ptd_count > epq->stat_maxptds)
925 epq->stat_maxptds = epq->ptd_count;
926
927 /* check, whether the second ISTL buffer may also be filled */
928 if (!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
929 (flip ? HCBUFSTAT_ISTL0_FULL : HCBUFSTAT_ISTL1_FULL))) {
930 fno++;
931 ptd_count = 0;
932 flip = 1 - flip;
933 goto fill2;
934 }
935}
936
937static void finish_transfers(struct isp1362_hcd *isp1362_hcd, unsigned long done_map,
938 struct isp1362_ep_queue *epq)
939{
940 struct isp1362_ep *ep;
941 struct isp1362_ep *tmp;
942
943 if (list_empty(&epq->active)) {
944 DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
945 return;
946 }
947
948 DBG(1, "%s: Finishing %s transfers %08lx\n", __func__, epq->name, done_map);
949
950 atomic_inc(&epq->finishing);
951 list_for_each_entry_safe(ep, tmp, &epq->active, active) {
952 int index = ep->ptd_index;
953
954 DBG(1, "%s: Checking %s PTD[%02x] $%04x\n", __func__, epq->name,
955 index, ep->ptd_offset);
956
957 BUG_ON(index < 0);
958 if (__test_and_clear_bit(index, &done_map)) {
959 isp1362_read_ptd(isp1362_hcd, ep, epq);
960 epq->free_ptd = index;
961 BUG_ON(ep->num_ptds == 0);
962 release_ptd_buffers(epq, ep);
963
964 DBG(1, "%s: ep %p req %d removed from active list\n", __func__,
965 ep, ep->num_req);
966 if (!list_empty(&ep->remove_list)) {
967 list_del_init(&ep->remove_list);
968 DBG(1, "%s: ep %p removed from remove list\n", __func__, ep);
969 }
970 DBG(1, "%s: Postprocessing %s ep %p req %d\n", __func__, epq->name,
971 ep, ep->num_req);
972 postproc_ep(isp1362_hcd, ep);
973 }
974 if (!done_map)
975 break;
976 }
977 if (done_map)
978 pr_warn("%s: done_map not clear: %08lx:%08lx\n",
979 __func__, done_map, epq->skip_map);
980 atomic_dec(&epq->finishing);
981}
982
983static void finish_iso_transfers(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep_queue *epq)
984{
985 struct isp1362_ep *ep;
986 struct isp1362_ep *tmp;
987
988 if (list_empty(&epq->active)) {
989 DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
990 return;
991 }
992
993 DBG(1, "%s: Finishing %s transfers\n", __func__, epq->name);
994
995 atomic_inc(&epq->finishing);
996 list_for_each_entry_safe(ep, tmp, &epq->active, active) {
997 DBG(1, "%s: Checking PTD $%04x\n", __func__, ep->ptd_offset);
998
999 isp1362_read_ptd(isp1362_hcd, ep, epq);
1000 DBG(1, "%s: Postprocessing %s ep %p\n", __func__, epq->name, ep);
1001 postproc_ep(isp1362_hcd, ep);
1002 }
1003 WARN_ON(epq->blk_size != 0);
1004 atomic_dec(&epq->finishing);
1005}
1006
1007static irqreturn_t isp1362_irq(struct usb_hcd *hcd)
1008{
1009 int handled = 0;
1010 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1011 u16 irqstat;
1012 u16 svc_mask;
1013
1014 spin_lock(&isp1362_hcd->lock);
1015
1016 BUG_ON(isp1362_hcd->irq_active++);
1017
1018 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1019
1020 irqstat = isp1362_read_reg16(isp1362_hcd, HCuPINT);
1021 DBG(3, "%s: got IRQ %04x:%04x\n", __func__, irqstat, isp1362_hcd->irqenb);
1022
1023 /* only handle interrupts that are currently enabled */
1024 irqstat &= isp1362_hcd->irqenb;
1025 isp1362_write_reg16(isp1362_hcd, HCuPINT, irqstat);
1026 svc_mask = irqstat;
1027
1028 if (irqstat & HCuPINT_SOF) {
1029 isp1362_hcd->irqenb &= ~HCuPINT_SOF;
1030 isp1362_hcd->irq_stat[ISP1362_INT_SOF]++;
1031 handled = 1;
1032 svc_mask &= ~HCuPINT_SOF;
1033 DBG(3, "%s: SOF\n", __func__);
1034 isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1035 if (!list_empty(&isp1362_hcd->remove_list))
1036 finish_unlinks(isp1362_hcd);
1037 if (!list_empty(&isp1362_hcd->async) && !(irqstat & HCuPINT_ATL)) {
1038 if (list_empty(&isp1362_hcd->atl_queue.active)) {
1039 start_atl_transfers(isp1362_hcd);
1040 } else {
1041 isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
1042 isp1362_write_reg32(isp1362_hcd, HCATLSKIP,
1043 isp1362_hcd->atl_queue.skip_map);
1044 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1045 }
1046 }
1047 }
1048
1049 if (irqstat & HCuPINT_ISTL0) {
1050 isp1362_hcd->irq_stat[ISP1362_INT_ISTL0]++;
1051 handled = 1;
1052 svc_mask &= ~HCuPINT_ISTL0;
1053 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL0_FULL);
1054 DBG(1, "%s: ISTL0\n", __func__);
1055 WARN_ON((int)!!isp1362_hcd->istl_flip);
1056 WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1057 HCBUFSTAT_ISTL0_ACTIVE);
1058 WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1059 HCBUFSTAT_ISTL0_DONE));
1060 isp1362_hcd->irqenb &= ~HCuPINT_ISTL0;
1061 }
1062
1063 if (irqstat & HCuPINT_ISTL1) {
1064 isp1362_hcd->irq_stat[ISP1362_INT_ISTL1]++;
1065 handled = 1;
1066 svc_mask &= ~HCuPINT_ISTL1;
1067 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL1_FULL);
1068 DBG(1, "%s: ISTL1\n", __func__);
1069 WARN_ON(!(int)isp1362_hcd->istl_flip);
1070 WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1071 HCBUFSTAT_ISTL1_ACTIVE);
1072 WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1073 HCBUFSTAT_ISTL1_DONE));
1074 isp1362_hcd->irqenb &= ~HCuPINT_ISTL1;
1075 }
1076
1077 if (irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) {
1078 WARN_ON((irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) ==
1079 (HCuPINT_ISTL0 | HCuPINT_ISTL1));
1080 finish_iso_transfers(isp1362_hcd,
1081 &isp1362_hcd->istl_queue[isp1362_hcd->istl_flip]);
1082 start_iso_transfers(isp1362_hcd);
1083 isp1362_hcd->istl_flip = 1 - isp1362_hcd->istl_flip;
1084 }
1085
1086 if (irqstat & HCuPINT_INTL) {
1087 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1088 u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCINTLSKIP);
1089 isp1362_hcd->irq_stat[ISP1362_INT_INTL]++;
1090
1091 DBG(2, "%s: INTL\n", __func__);
1092
1093 svc_mask &= ~HCuPINT_INTL;
1094
1095 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, skip_map | done_map);
1096 if (~(done_map | skip_map) == 0)
1097 /* All PTDs are finished, disable INTL processing entirely */
1098 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
1099
1100 handled = 1;
1101 WARN_ON(!done_map);
1102 if (done_map) {
1103 DBG(3, "%s: INTL done_map %08x\n", __func__, done_map);
1104 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1105 start_intl_transfers(isp1362_hcd);
1106 }
1107 }
1108
1109 if (irqstat & HCuPINT_ATL) {
1110 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1111 u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCATLSKIP);
1112 isp1362_hcd->irq_stat[ISP1362_INT_ATL]++;
1113
1114 DBG(2, "%s: ATL\n", __func__);
1115
1116 svc_mask &= ~HCuPINT_ATL;
1117
1118 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, skip_map | done_map);
1119 if (~(done_map | skip_map) == 0)
1120 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1121 if (done_map) {
1122 DBG(3, "%s: ATL done_map %08x\n", __func__, done_map);
1123 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1124 start_atl_transfers(isp1362_hcd);
1125 }
1126 handled = 1;
1127 }
1128
1129 if (irqstat & HCuPINT_OPR) {
1130 u32 intstat = isp1362_read_reg32(isp1362_hcd, HCINTSTAT);
1131 isp1362_hcd->irq_stat[ISP1362_INT_OPR]++;
1132
1133 svc_mask &= ~HCuPINT_OPR;
1134 DBG(2, "%s: OPR %08x:%08x\n", __func__, intstat, isp1362_hcd->intenb);
1135 intstat &= isp1362_hcd->intenb;
1136 if (intstat & OHCI_INTR_UE) {
1137 pr_err("Unrecoverable error\n");
1138 /* FIXME: do here reset or cleanup or whatever */
1139 }
1140 if (intstat & OHCI_INTR_RHSC) {
1141 isp1362_hcd->rhstatus = isp1362_read_reg32(isp1362_hcd, HCRHSTATUS);
1142 isp1362_hcd->rhport[0] = isp1362_read_reg32(isp1362_hcd, HCRHPORT1);
1143 isp1362_hcd->rhport[1] = isp1362_read_reg32(isp1362_hcd, HCRHPORT2);
1144 }
1145 if (intstat & OHCI_INTR_RD) {
1146 pr_info("%s: RESUME DETECTED\n", __func__);
1147 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1148 usb_hcd_resume_root_hub(hcd);
1149 }
1150 isp1362_write_reg32(isp1362_hcd, HCINTSTAT, intstat);
1151 irqstat &= ~HCuPINT_OPR;
1152 handled = 1;
1153 }
1154
1155 if (irqstat & HCuPINT_SUSP) {
1156 isp1362_hcd->irq_stat[ISP1362_INT_SUSP]++;
1157 handled = 1;
1158 svc_mask &= ~HCuPINT_SUSP;
1159
1160 pr_info("%s: SUSPEND IRQ\n", __func__);
1161 }
1162
1163 if (irqstat & HCuPINT_CLKRDY) {
1164 isp1362_hcd->irq_stat[ISP1362_INT_CLKRDY]++;
1165 handled = 1;
1166 isp1362_hcd->irqenb &= ~HCuPINT_CLKRDY;
1167 svc_mask &= ~HCuPINT_CLKRDY;
1168 pr_info("%s: CLKRDY IRQ\n", __func__);
1169 }
1170
1171 if (svc_mask)
1172 pr_err("%s: Unserviced interrupt(s) %04x\n", __func__, svc_mask);
1173
1174 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
1175 isp1362_hcd->irq_active--;
1176 spin_unlock(&isp1362_hcd->lock);
1177
1178 return IRQ_RETVAL(handled);
1179}
1180
1181/*-------------------------------------------------------------------------*/
1182
1183#define MAX_PERIODIC_LOAD 900 /* out of 1000 usec */
1184static int balance(struct isp1362_hcd *isp1362_hcd, u16 interval, u16 load)
1185{
1186 int i, branch = -ENOSPC;
1187
1188 /* search for the least loaded schedule branch of that interval
1189 * which has enough bandwidth left unreserved.
1190 */
1191 for (i = 0; i < interval; i++) {
1192 if (branch < 0 || isp1362_hcd->load[branch] > isp1362_hcd->load[i]) {
1193 int j;
1194
1195 for (j = i; j < PERIODIC_SIZE; j += interval) {
1196 if ((isp1362_hcd->load[j] + load) > MAX_PERIODIC_LOAD) {
1197 pr_err("%s: new load %d load[%02x] %d max %d\n", __func__,
1198 load, j, isp1362_hcd->load[j], MAX_PERIODIC_LOAD);
1199 break;
1200 }
1201 }
1202 if (j < PERIODIC_SIZE)
1203 continue;
1204 branch = i;
1205 }
1206 }
1207 return branch;
1208}
1209
1210/* NB! ALL the code above this point runs with isp1362_hcd->lock
1211 held, irqs off
1212*/
1213
1214/*-------------------------------------------------------------------------*/
1215
1216static int isp1362_urb_enqueue(struct usb_hcd *hcd,
1217 struct urb *urb,
1218 gfp_t mem_flags)
1219{
1220 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1221 struct usb_device *udev = urb->dev;
1222 unsigned int pipe = urb->pipe;
1223 int is_out = !usb_pipein(pipe);
1224 int type = usb_pipetype(pipe);
1225 int epnum = usb_pipeendpoint(pipe);
1226 struct usb_host_endpoint *hep = urb->ep;
1227 struct isp1362_ep *ep = NULL;
1228 unsigned long flags;
1229 int retval = 0;
1230
1231 DBG(3, "%s: urb %p\n", __func__, urb);
1232
1233 if (type == PIPE_ISOCHRONOUS) {
1234 pr_err("Isochronous transfers not supported\n");
1235 return -ENOSPC;
1236 }
1237
1238 URB_DBG("%s: FA %d ep%d%s %s: len %d %s%s\n", __func__,
1239 usb_pipedevice(pipe), epnum,
1240 is_out ? "out" : "in",
1241 usb_pipecontrol(pipe) ? "ctrl" :
1242 usb_pipeint(pipe) ? "int" :
1243 usb_pipebulk(pipe) ? "bulk" :
1244 "iso",
1245 urb->transfer_buffer_length,
1246 (urb->transfer_flags & URB_ZERO_PACKET) ? "ZERO_PACKET " : "",
1247 !(urb->transfer_flags & URB_SHORT_NOT_OK) ?
1248 "short_ok" : "");
1249
1250 /* avoid all allocations within spinlocks: request or endpoint */
1251 if (!hep->hcpriv) {
1252 ep = kzalloc(sizeof *ep, mem_flags);
1253 if (!ep)
1254 return -ENOMEM;
1255 }
1256 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1257
1258 /* don't submit to a dead or disabled port */
1259 if (!((isp1362_hcd->rhport[0] | isp1362_hcd->rhport[1]) &
1260 USB_PORT_STAT_ENABLE) ||
1261 !HC_IS_RUNNING(hcd->state)) {
1262 kfree(ep);
1263 retval = -ENODEV;
1264 goto fail_not_linked;
1265 }
1266
1267 retval = usb_hcd_link_urb_to_ep(hcd, urb);
1268 if (retval) {
1269 kfree(ep);
1270 goto fail_not_linked;
1271 }
1272
1273 if (hep->hcpriv) {
1274 ep = hep->hcpriv;
1275 } else {
1276 INIT_LIST_HEAD(&ep->schedule);
1277 INIT_LIST_HEAD(&ep->active);
1278 INIT_LIST_HEAD(&ep->remove_list);
1279 ep->udev = usb_get_dev(udev);
1280 ep->hep = hep;
1281 ep->epnum = epnum;
1282 ep->maxpacket = usb_maxpacket(udev, urb->pipe);
1283 ep->ptd_offset = -EINVAL;
1284 ep->ptd_index = -EINVAL;
1285 usb_settoggle(udev, epnum, is_out, 0);
1286
1287 if (type == PIPE_CONTROL)
1288 ep->nextpid = USB_PID_SETUP;
1289 else if (is_out)
1290 ep->nextpid = USB_PID_OUT;
1291 else
1292 ep->nextpid = USB_PID_IN;
1293
1294 switch (type) {
1295 case PIPE_ISOCHRONOUS:
1296 case PIPE_INTERRUPT:
1297 if (urb->interval > PERIODIC_SIZE)
1298 urb->interval = PERIODIC_SIZE;
1299 ep->interval = urb->interval;
1300 ep->branch = PERIODIC_SIZE;
1301 ep->load = usb_calc_bus_time(udev->speed, !is_out,
1302 type == PIPE_ISOCHRONOUS,
1303 usb_maxpacket(udev, pipe)) / 1000;
1304 break;
1305 }
1306 hep->hcpriv = ep;
1307 }
1308 ep->num_req = isp1362_hcd->req_serial++;
1309
1310 /* maybe put endpoint into schedule */
1311 switch (type) {
1312 case PIPE_CONTROL:
1313 case PIPE_BULK:
1314 if (list_empty(&ep->schedule)) {
1315 DBG(1, "%s: Adding ep %p req %d to async schedule\n",
1316 __func__, ep, ep->num_req);
1317 list_add_tail(&ep->schedule, &isp1362_hcd->async);
1318 }
1319 break;
1320 case PIPE_ISOCHRONOUS:
1321 case PIPE_INTERRUPT:
1322 urb->interval = ep->interval;
1323
1324 /* urb submitted for already existing EP */
1325 if (ep->branch < PERIODIC_SIZE)
1326 break;
1327
1328 retval = balance(isp1362_hcd, ep->interval, ep->load);
1329 if (retval < 0) {
1330 pr_err("%s: balance returned %d\n", __func__, retval);
1331 goto fail;
1332 }
1333 ep->branch = retval;
1334 retval = 0;
1335 isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1336 DBG(1, "%s: Current frame %04x branch %02x start_frame %04x(%04x)\n",
1337 __func__, isp1362_hcd->fmindex, ep->branch,
1338 ((isp1362_hcd->fmindex + PERIODIC_SIZE - 1) &
1339 ~(PERIODIC_SIZE - 1)) + ep->branch,
1340 (isp1362_hcd->fmindex & (PERIODIC_SIZE - 1)) + ep->branch);
1341
1342 if (list_empty(&ep->schedule)) {
1343 if (type == PIPE_ISOCHRONOUS) {
1344 u16 frame = isp1362_hcd->fmindex;
1345
1346 frame += max_t(u16, 8, ep->interval);
1347 frame &= ~(ep->interval - 1);
1348 frame |= ep->branch;
1349 if (frame_before(frame, isp1362_hcd->fmindex))
1350 frame += ep->interval;
1351 urb->start_frame = frame;
1352
1353 DBG(1, "%s: Adding ep %p to isoc schedule\n", __func__, ep);
1354 list_add_tail(&ep->schedule, &isp1362_hcd->isoc);
1355 } else {
1356 DBG(1, "%s: Adding ep %p to periodic schedule\n", __func__, ep);
1357 list_add_tail(&ep->schedule, &isp1362_hcd->periodic);
1358 }
1359 } else
1360 DBG(1, "%s: ep %p already scheduled\n", __func__, ep);
1361
1362 DBG(2, "%s: load %d bandwidth %d -> %d\n", __func__,
1363 ep->load / ep->interval, isp1362_hcd->load[ep->branch],
1364 isp1362_hcd->load[ep->branch] + ep->load);
1365 isp1362_hcd->load[ep->branch] += ep->load;
1366 }
1367
1368 urb->hcpriv = hep;
1369 ALIGNSTAT(isp1362_hcd, urb->transfer_buffer);
1370
1371 switch (type) {
1372 case PIPE_CONTROL:
1373 case PIPE_BULK:
1374 start_atl_transfers(isp1362_hcd);
1375 break;
1376 case PIPE_INTERRUPT:
1377 start_intl_transfers(isp1362_hcd);
1378 break;
1379 case PIPE_ISOCHRONOUS:
1380 start_iso_transfers(isp1362_hcd);
1381 break;
1382 default:
1383 BUG();
1384 }
1385 fail:
1386 if (retval)
1387 usb_hcd_unlink_urb_from_ep(hcd, urb);
1388
1389
1390 fail_not_linked:
1391 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1392 if (retval)
1393 DBG(0, "%s: urb %p failed with %d\n", __func__, urb, retval);
1394 return retval;
1395}
1396
1397static int isp1362_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status)
1398{
1399 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1400 struct usb_host_endpoint *hep;
1401 unsigned long flags;
1402 struct isp1362_ep *ep;
1403 int retval = 0;
1404
1405 DBG(3, "%s: urb %p\n", __func__, urb);
1406
1407 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1408 retval = usb_hcd_check_unlink_urb(hcd, urb, status);
1409 if (retval)
1410 goto done;
1411
1412 hep = urb->hcpriv;
1413
1414 if (!hep) {
1415 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1416 return -EIDRM;
1417 }
1418
1419 ep = hep->hcpriv;
1420 if (ep) {
1421 /* In front of queue? */
1422 if (ep->hep->urb_list.next == &urb->urb_list) {
1423 if (!list_empty(&ep->active)) {
1424 DBG(1, "%s: urb %p ep %p req %d active PTD[%d] $%04x\n", __func__,
1425 urb, ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1426 /* disable processing and queue PTD for removal */
1427 remove_ptd(isp1362_hcd, ep);
1428 urb = NULL;
1429 }
1430 }
1431 if (urb) {
1432 DBG(1, "%s: Finishing ep %p req %d\n", __func__, ep,
1433 ep->num_req);
1434 finish_request(isp1362_hcd, ep, urb, status);
1435 } else
1436 DBG(1, "%s: urb %p active; wait4irq\n", __func__, urb);
1437 } else {
1438 pr_warn("%s: No EP in URB %p\n", __func__, urb);
1439 retval = -EINVAL;
1440 }
1441done:
1442 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1443
1444 DBG(3, "%s: exit\n", __func__);
1445
1446 return retval;
1447}
1448
1449static void isp1362_endpoint_disable(struct usb_hcd *hcd, struct usb_host_endpoint *hep)
1450{
1451 struct isp1362_ep *ep = hep->hcpriv;
1452 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1453 unsigned long flags;
1454
1455 DBG(1, "%s: ep %p\n", __func__, ep);
1456 if (!ep)
1457 return;
1458 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1459 if (!list_empty(&hep->urb_list)) {
1460 if (!list_empty(&ep->active) && list_empty(&ep->remove_list)) {
1461 DBG(1, "%s: Removing ep %p req %d PTD[%d] $%04x\n", __func__,
1462 ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1463 remove_ptd(isp1362_hcd, ep);
1464 pr_info("%s: Waiting for Interrupt to clean up\n", __func__);
1465 }
1466 }
1467 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1468 /* Wait for interrupt to clear out active list */
1469 while (!list_empty(&ep->active))
1470 msleep(1);
1471
1472 DBG(1, "%s: Freeing EP %p\n", __func__, ep);
1473
1474 usb_put_dev(ep->udev);
1475 kfree(ep);
1476 hep->hcpriv = NULL;
1477}
1478
1479static int isp1362_get_frame(struct usb_hcd *hcd)
1480{
1481 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1482 u32 fmnum;
1483 unsigned long flags;
1484
1485 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1486 fmnum = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1487 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1488
1489 return (int)fmnum;
1490}
1491
1492/*-------------------------------------------------------------------------*/
1493
1494/* Adapted from ohci-hub.c */
1495static int isp1362_hub_status_data(struct usb_hcd *hcd, char *buf)
1496{
1497 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1498 int ports, i, changed = 0;
1499 unsigned long flags;
1500
1501 if (!HC_IS_RUNNING(hcd->state))
1502 return -ESHUTDOWN;
1503
1504 /* Report no status change now, if we are scheduled to be
1505 called later */
1506 if (timer_pending(&hcd->rh_timer))
1507 return 0;
1508
1509 ports = isp1362_hcd->rhdesca & RH_A_NDP;
1510 BUG_ON(ports > 2);
1511
1512 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1513 /* init status */
1514 if (isp1362_hcd->rhstatus & (RH_HS_LPSC | RH_HS_OCIC))
1515 buf[0] = changed = 1;
1516 else
1517 buf[0] = 0;
1518
1519 for (i = 0; i < ports; i++) {
1520 u32 status = isp1362_hcd->rhport[i];
1521
1522 if (status & (RH_PS_CSC | RH_PS_PESC | RH_PS_PSSC |
1523 RH_PS_OCIC | RH_PS_PRSC)) {
1524 changed = 1;
1525 buf[0] |= 1 << (i + 1);
1526 continue;
1527 }
1528
1529 if (!(status & RH_PS_CCS))
1530 continue;
1531 }
1532 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1533 return changed;
1534}
1535
1536static void isp1362_hub_descriptor(struct isp1362_hcd *isp1362_hcd,
1537 struct usb_hub_descriptor *desc)
1538{
1539 u32 reg = isp1362_hcd->rhdesca;
1540
1541 DBG(3, "%s: enter\n", __func__);
1542
1543 desc->bDescriptorType = USB_DT_HUB;
1544 desc->bDescLength = 9;
1545 desc->bHubContrCurrent = 0;
1546 desc->bNbrPorts = reg & 0x3;
1547 /* Power switching, device type, overcurrent. */
1548 desc->wHubCharacteristics = cpu_to_le16((reg >> 8) &
1549 (HUB_CHAR_LPSM |
1550 HUB_CHAR_COMPOUND |
1551 HUB_CHAR_OCPM));
1552 DBG(0, "%s: hubcharacteristics = %02x\n", __func__,
1553 desc->wHubCharacteristics);
1554 desc->bPwrOn2PwrGood = (reg >> 24) & 0xff;
1555 /* ports removable, and legacy PortPwrCtrlMask */
1556 desc->u.hs.DeviceRemovable[0] = desc->bNbrPorts == 1 ? 1 << 1 : 3 << 1;
1557 desc->u.hs.DeviceRemovable[1] = ~0;
1558
1559 DBG(3, "%s: exit\n", __func__);
1560}
1561
1562/* Adapted from ohci-hub.c */
1563static int isp1362_hub_control(struct usb_hcd *hcd, u16 typeReq, u16 wValue,
1564 u16 wIndex, char *buf, u16 wLength)
1565{
1566 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1567 int retval = 0;
1568 unsigned long flags;
1569 unsigned long t1;
1570 int ports = isp1362_hcd->rhdesca & RH_A_NDP;
1571 u32 tmp = 0;
1572
1573 switch (typeReq) {
1574 case ClearHubFeature:
1575 DBG(0, "ClearHubFeature: ");
1576 switch (wValue) {
1577 case C_HUB_OVER_CURRENT:
1578 DBG(0, "C_HUB_OVER_CURRENT\n");
1579 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1580 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_OCIC);
1581 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1582 break;
1583 case C_HUB_LOCAL_POWER:
1584 DBG(0, "C_HUB_LOCAL_POWER\n");
1585 break;
1586 default:
1587 goto error;
1588 }
1589 break;
1590 case SetHubFeature:
1591 DBG(0, "SetHubFeature: ");
1592 switch (wValue) {
1593 case C_HUB_OVER_CURRENT:
1594 case C_HUB_LOCAL_POWER:
1595 DBG(0, "C_HUB_OVER_CURRENT or C_HUB_LOCAL_POWER\n");
1596 break;
1597 default:
1598 goto error;
1599 }
1600 break;
1601 case GetHubDescriptor:
1602 DBG(0, "GetHubDescriptor\n");
1603 isp1362_hub_descriptor(isp1362_hcd, (struct usb_hub_descriptor *)buf);
1604 break;
1605 case GetHubStatus:
1606 DBG(0, "GetHubStatus\n");
1607 put_unaligned(cpu_to_le32(0), (__le32 *) buf);
1608 break;
1609 case GetPortStatus:
1610#ifndef VERBOSE
1611 DBG(0, "GetPortStatus\n");
1612#endif
1613 if (!wIndex || wIndex > ports)
1614 goto error;
1615 tmp = isp1362_hcd->rhport[--wIndex];
1616 put_unaligned(cpu_to_le32(tmp), (__le32 *) buf);
1617 break;
1618 case ClearPortFeature:
1619 DBG(0, "ClearPortFeature: ");
1620 if (!wIndex || wIndex > ports)
1621 goto error;
1622 wIndex--;
1623
1624 switch (wValue) {
1625 case USB_PORT_FEAT_ENABLE:
1626 DBG(0, "USB_PORT_FEAT_ENABLE\n");
1627 tmp = RH_PS_CCS;
1628 break;
1629 case USB_PORT_FEAT_C_ENABLE:
1630 DBG(0, "USB_PORT_FEAT_C_ENABLE\n");
1631 tmp = RH_PS_PESC;
1632 break;
1633 case USB_PORT_FEAT_SUSPEND:
1634 DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1635 tmp = RH_PS_POCI;
1636 break;
1637 case USB_PORT_FEAT_C_SUSPEND:
1638 DBG(0, "USB_PORT_FEAT_C_SUSPEND\n");
1639 tmp = RH_PS_PSSC;
1640 break;
1641 case USB_PORT_FEAT_POWER:
1642 DBG(0, "USB_PORT_FEAT_POWER\n");
1643 tmp = RH_PS_LSDA;
1644
1645 break;
1646 case USB_PORT_FEAT_C_CONNECTION:
1647 DBG(0, "USB_PORT_FEAT_C_CONNECTION\n");
1648 tmp = RH_PS_CSC;
1649 break;
1650 case USB_PORT_FEAT_C_OVER_CURRENT:
1651 DBG(0, "USB_PORT_FEAT_C_OVER_CURRENT\n");
1652 tmp = RH_PS_OCIC;
1653 break;
1654 case USB_PORT_FEAT_C_RESET:
1655 DBG(0, "USB_PORT_FEAT_C_RESET\n");
1656 tmp = RH_PS_PRSC;
1657 break;
1658 default:
1659 goto error;
1660 }
1661
1662 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1663 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, tmp);
1664 isp1362_hcd->rhport[wIndex] =
1665 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1666 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1667 break;
1668 case SetPortFeature:
1669 DBG(0, "SetPortFeature: ");
1670 if (!wIndex || wIndex > ports)
1671 goto error;
1672 wIndex--;
1673 switch (wValue) {
1674 case USB_PORT_FEAT_SUSPEND:
1675 DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1676 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1677 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PSS);
1678 isp1362_hcd->rhport[wIndex] =
1679 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1680 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1681 break;
1682 case USB_PORT_FEAT_POWER:
1683 DBG(0, "USB_PORT_FEAT_POWER\n");
1684 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1685 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PPS);
1686 isp1362_hcd->rhport[wIndex] =
1687 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1688 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1689 break;
1690 case USB_PORT_FEAT_RESET:
1691 DBG(0, "USB_PORT_FEAT_RESET\n");
1692 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1693
1694 t1 = jiffies + msecs_to_jiffies(USB_RESET_WIDTH);
1695 while (time_before(jiffies, t1)) {
1696 /* spin until any current reset finishes */
1697 for (;;) {
1698 tmp = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1699 if (!(tmp & RH_PS_PRS))
1700 break;
1701 udelay(500);
1702 }
1703 if (!(tmp & RH_PS_CCS))
1704 break;
1705 /* Reset lasts 10ms (claims datasheet) */
1706 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, (RH_PS_PRS));
1707
1708 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1709 msleep(10);
1710 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1711 }
1712
1713 isp1362_hcd->rhport[wIndex] = isp1362_read_reg32(isp1362_hcd,
1714 HCRHPORT1 + wIndex);
1715 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1716 break;
1717 default:
1718 goto error;
1719 }
1720 break;
1721
1722 default:
1723 error:
1724 /* "protocol stall" on error */
1725 DBG(0, "PROTOCOL STALL\n");
1726 retval = -EPIPE;
1727 }
1728
1729 return retval;
1730}
1731
1732#ifdef CONFIG_PM
1733static int isp1362_bus_suspend(struct usb_hcd *hcd)
1734{
1735 int status = 0;
1736 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1737 unsigned long flags;
1738
1739 if (time_before(jiffies, isp1362_hcd->next_statechange))
1740 msleep(5);
1741
1742 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1743
1744 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1745 switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1746 case OHCI_USB_RESUME:
1747 DBG(0, "%s: resume/suspend?\n", __func__);
1748 isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1749 isp1362_hcd->hc_control |= OHCI_USB_RESET;
1750 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1751 fallthrough;
1752 case OHCI_USB_RESET:
1753 status = -EBUSY;
1754 pr_warn("%s: needs reinit!\n", __func__);
1755 goto done;
1756 case OHCI_USB_SUSPEND:
1757 pr_warn("%s: already suspended?\n", __func__);
1758 goto done;
1759 }
1760 DBG(0, "%s: suspend root hub\n", __func__);
1761
1762 /* First stop any processing */
1763 hcd->state = HC_STATE_QUIESCING;
1764 if (!list_empty(&isp1362_hcd->atl_queue.active) ||
1765 !list_empty(&isp1362_hcd->intl_queue.active) ||
1766 !list_empty(&isp1362_hcd->istl_queue[0] .active) ||
1767 !list_empty(&isp1362_hcd->istl_queue[1] .active)) {
1768 int limit;
1769
1770 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
1771 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
1772 isp1362_write_reg16(isp1362_hcd, HCBUFSTAT, 0);
1773 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1774 isp1362_write_reg32(isp1362_hcd, HCINTSTAT, OHCI_INTR_SF);
1775
1776 DBG(0, "%s: stopping schedules ...\n", __func__);
1777 limit = 2000;
1778 while (limit > 0) {
1779 udelay(250);
1780 limit -= 250;
1781 if (isp1362_read_reg32(isp1362_hcd, HCINTSTAT) & OHCI_INTR_SF)
1782 break;
1783 }
1784 mdelay(7);
1785 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ATL) {
1786 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1787 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1788 }
1789 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_INTL) {
1790 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1791 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1792 }
1793 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL0)
1794 finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[0]);
1795 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL1)
1796 finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[1]);
1797 }
1798 DBG(0, "%s: HCINTSTAT: %08x\n", __func__,
1799 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1800 isp1362_write_reg32(isp1362_hcd, HCINTSTAT,
1801 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1802
1803 /* Suspend hub */
1804 isp1362_hcd->hc_control = OHCI_USB_SUSPEND;
1805 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1806 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1807 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1808
1809#if 1
1810 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1811 if ((isp1362_hcd->hc_control & OHCI_CTRL_HCFS) != OHCI_USB_SUSPEND) {
1812 pr_err("%s: controller won't suspend %08x\n", __func__,
1813 isp1362_hcd->hc_control);
1814 status = -EBUSY;
1815 } else
1816#endif
1817 {
1818 /* no resumes until devices finish suspending */
1819 isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(5);
1820 }
1821done:
1822 if (status == 0) {
1823 hcd->state = HC_STATE_SUSPENDED;
1824 DBG(0, "%s: HCD suspended: %08x\n", __func__,
1825 isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1826 }
1827 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1828 return status;
1829}
1830
1831static int isp1362_bus_resume(struct usb_hcd *hcd)
1832{
1833 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1834 u32 port;
1835 unsigned long flags;
1836 int status = -EINPROGRESS;
1837
1838 if (time_before(jiffies, isp1362_hcd->next_statechange))
1839 msleep(5);
1840
1841 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1842 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1843 pr_info("%s: HCCONTROL: %08x\n", __func__, isp1362_hcd->hc_control);
1844 if (hcd->state == HC_STATE_RESUMING) {
1845 pr_warn("%s: duplicate resume\n", __func__);
1846 status = 0;
1847 } else
1848 switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1849 case OHCI_USB_SUSPEND:
1850 DBG(0, "%s: resume root hub\n", __func__);
1851 isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1852 isp1362_hcd->hc_control |= OHCI_USB_RESUME;
1853 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1854 break;
1855 case OHCI_USB_RESUME:
1856 /* HCFS changes sometime after INTR_RD */
1857 DBG(0, "%s: remote wakeup\n", __func__);
1858 break;
1859 case OHCI_USB_OPER:
1860 DBG(0, "%s: odd resume\n", __func__);
1861 status = 0;
1862 hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1863 break;
1864 default: /* RESET, we lost power */
1865 DBG(0, "%s: root hub hardware reset\n", __func__);
1866 status = -EBUSY;
1867 }
1868 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1869 if (status == -EBUSY) {
1870 DBG(0, "%s: Restarting HC\n", __func__);
1871 isp1362_hc_stop(hcd);
1872 return isp1362_hc_start(hcd);
1873 }
1874 if (status != -EINPROGRESS)
1875 return status;
1876 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1877 port = isp1362_read_reg32(isp1362_hcd, HCRHDESCA) & RH_A_NDP;
1878 while (port--) {
1879 u32 stat = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + port);
1880
1881 /* force global, not selective, resume */
1882 if (!(stat & RH_PS_PSS)) {
1883 DBG(0, "%s: Not Resuming RH port %d\n", __func__, port);
1884 continue;
1885 }
1886 DBG(0, "%s: Resuming RH port %d\n", __func__, port);
1887 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + port, RH_PS_POCI);
1888 }
1889 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1890
1891 /* Some controllers (lucent) need extra-long delays */
1892 hcd->state = HC_STATE_RESUMING;
1893 mdelay(20 /* usb 11.5.1.10 */ + 15);
1894
1895 isp1362_hcd->hc_control = OHCI_USB_OPER;
1896 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1897 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1898 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1899 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1900 /* TRSMRCY */
1901 msleep(10);
1902
1903 /* keep it alive for ~5x suspend + resume costs */
1904 isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(250);
1905
1906 hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1907 hcd->state = HC_STATE_RUNNING;
1908 return 0;
1909}
1910#else
1911#define isp1362_bus_suspend NULL
1912#define isp1362_bus_resume NULL
1913#endif
1914
1915/*-------------------------------------------------------------------------*/
1916
1917static void dump_irq(struct seq_file *s, char *label, u16 mask)
1918{
1919 seq_printf(s, "%-15s %04x%s%s%s%s%s%s\n", label, mask,
1920 mask & HCuPINT_CLKRDY ? " clkrdy" : "",
1921 mask & HCuPINT_SUSP ? " susp" : "",
1922 mask & HCuPINT_OPR ? " opr" : "",
1923 mask & HCuPINT_EOT ? " eot" : "",
1924 mask & HCuPINT_ATL ? " atl" : "",
1925 mask & HCuPINT_SOF ? " sof" : "");
1926}
1927
1928static void dump_int(struct seq_file *s, char *label, u32 mask)
1929{
1930 seq_printf(s, "%-15s %08x%s%s%s%s%s%s%s\n", label, mask,
1931 mask & OHCI_INTR_MIE ? " MIE" : "",
1932 mask & OHCI_INTR_RHSC ? " rhsc" : "",
1933 mask & OHCI_INTR_FNO ? " fno" : "",
1934 mask & OHCI_INTR_UE ? " ue" : "",
1935 mask & OHCI_INTR_RD ? " rd" : "",
1936 mask & OHCI_INTR_SF ? " sof" : "",
1937 mask & OHCI_INTR_SO ? " so" : "");
1938}
1939
1940static void dump_ctrl(struct seq_file *s, char *label, u32 mask)
1941{
1942 seq_printf(s, "%-15s %08x%s%s%s\n", label, mask,
1943 mask & OHCI_CTRL_RWC ? " rwc" : "",
1944 mask & OHCI_CTRL_RWE ? " rwe" : "",
1945 ({
1946 char *hcfs;
1947 switch (mask & OHCI_CTRL_HCFS) {
1948 case OHCI_USB_OPER:
1949 hcfs = " oper";
1950 break;
1951 case OHCI_USB_RESET:
1952 hcfs = " reset";
1953 break;
1954 case OHCI_USB_RESUME:
1955 hcfs = " resume";
1956 break;
1957 case OHCI_USB_SUSPEND:
1958 hcfs = " suspend";
1959 break;
1960 default:
1961 hcfs = " ?";
1962 }
1963 hcfs;
1964 }));
1965}
1966
1967static void dump_regs(struct seq_file *s, struct isp1362_hcd *isp1362_hcd)
1968{
1969 seq_printf(s, "HCREVISION [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCREVISION),
1970 isp1362_read_reg32(isp1362_hcd, HCREVISION));
1971 seq_printf(s, "HCCONTROL [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCONTROL),
1972 isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1973 seq_printf(s, "HCCMDSTAT [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCMDSTAT),
1974 isp1362_read_reg32(isp1362_hcd, HCCMDSTAT));
1975 seq_printf(s, "HCINTSTAT [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTSTAT),
1976 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1977 seq_printf(s, "HCINTENB [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTENB),
1978 isp1362_read_reg32(isp1362_hcd, HCINTENB));
1979 seq_printf(s, "HCFMINTVL [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMINTVL),
1980 isp1362_read_reg32(isp1362_hcd, HCFMINTVL));
1981 seq_printf(s, "HCFMREM [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMREM),
1982 isp1362_read_reg32(isp1362_hcd, HCFMREM));
1983 seq_printf(s, "HCFMNUM [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMNUM),
1984 isp1362_read_reg32(isp1362_hcd, HCFMNUM));
1985 seq_printf(s, "HCLSTHRESH [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCLSTHRESH),
1986 isp1362_read_reg32(isp1362_hcd, HCLSTHRESH));
1987 seq_printf(s, "HCRHDESCA [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCA),
1988 isp1362_read_reg32(isp1362_hcd, HCRHDESCA));
1989 seq_printf(s, "HCRHDESCB [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCB),
1990 isp1362_read_reg32(isp1362_hcd, HCRHDESCB));
1991 seq_printf(s, "HCRHSTATUS [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHSTATUS),
1992 isp1362_read_reg32(isp1362_hcd, HCRHSTATUS));
1993 seq_printf(s, "HCRHPORT1 [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT1),
1994 isp1362_read_reg32(isp1362_hcd, HCRHPORT1));
1995 seq_printf(s, "HCRHPORT2 [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT2),
1996 isp1362_read_reg32(isp1362_hcd, HCRHPORT2));
1997 seq_printf(s, "\n");
1998 seq_printf(s, "HCHWCFG [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCHWCFG),
1999 isp1362_read_reg16(isp1362_hcd, HCHWCFG));
2000 seq_printf(s, "HCDMACFG [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCDMACFG),
2001 isp1362_read_reg16(isp1362_hcd, HCDMACFG));
2002 seq_printf(s, "HCXFERCTR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCXFERCTR),
2003 isp1362_read_reg16(isp1362_hcd, HCXFERCTR));
2004 seq_printf(s, "HCuPINT [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINT),
2005 isp1362_read_reg16(isp1362_hcd, HCuPINT));
2006 seq_printf(s, "HCuPINTENB [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINTENB),
2007 isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2008 seq_printf(s, "HCCHIPID [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCCHIPID),
2009 isp1362_read_reg16(isp1362_hcd, HCCHIPID));
2010 seq_printf(s, "HCSCRATCH [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCSCRATCH),
2011 isp1362_read_reg16(isp1362_hcd, HCSCRATCH));
2012 seq_printf(s, "HCBUFSTAT [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCBUFSTAT),
2013 isp1362_read_reg16(isp1362_hcd, HCBUFSTAT));
2014 seq_printf(s, "HCDIRADDR [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCDIRADDR),
2015 isp1362_read_reg32(isp1362_hcd, HCDIRADDR));
2016#if 0
2017 seq_printf(s, "HCDIRDATA [%02x] %04x\n", ISP1362_REG_NO(HCDIRDATA),
2018 isp1362_read_reg16(isp1362_hcd, HCDIRDATA));
2019#endif
2020 seq_printf(s, "HCISTLBUFSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLBUFSZ),
2021 isp1362_read_reg16(isp1362_hcd, HCISTLBUFSZ));
2022 seq_printf(s, "HCISTLRATE [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLRATE),
2023 isp1362_read_reg16(isp1362_hcd, HCISTLRATE));
2024 seq_printf(s, "\n");
2025 seq_printf(s, "HCINTLBUFSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBUFSZ),
2026 isp1362_read_reg16(isp1362_hcd, HCINTLBUFSZ));
2027 seq_printf(s, "HCINTLBLKSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBLKSZ),
2028 isp1362_read_reg16(isp1362_hcd, HCINTLBLKSZ));
2029 seq_printf(s, "HCINTLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLDONE),
2030 isp1362_read_reg32(isp1362_hcd, HCINTLDONE));
2031 seq_printf(s, "HCINTLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLSKIP),
2032 isp1362_read_reg32(isp1362_hcd, HCINTLSKIP));
2033 seq_printf(s, "HCINTLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLLAST),
2034 isp1362_read_reg32(isp1362_hcd, HCINTLLAST));
2035 seq_printf(s, "HCINTLCURR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLCURR),
2036 isp1362_read_reg16(isp1362_hcd, HCINTLCURR));
2037 seq_printf(s, "\n");
2038 seq_printf(s, "HCATLBUFSZ [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBUFSZ),
2039 isp1362_read_reg16(isp1362_hcd, HCATLBUFSZ));
2040 seq_printf(s, "HCATLBLKSZ [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBLKSZ),
2041 isp1362_read_reg16(isp1362_hcd, HCATLBLKSZ));
2042#if 0
2043 seq_printf(s, "HCATLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDONE),
2044 isp1362_read_reg32(isp1362_hcd, HCATLDONE));
2045#endif
2046 seq_printf(s, "HCATLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLSKIP),
2047 isp1362_read_reg32(isp1362_hcd, HCATLSKIP));
2048 seq_printf(s, "HCATLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLLAST),
2049 isp1362_read_reg32(isp1362_hcd, HCATLLAST));
2050 seq_printf(s, "HCATLCURR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLCURR),
2051 isp1362_read_reg16(isp1362_hcd, HCATLCURR));
2052 seq_printf(s, "\n");
2053 seq_printf(s, "HCATLDTC [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTC),
2054 isp1362_read_reg16(isp1362_hcd, HCATLDTC));
2055 seq_printf(s, "HCATLDTCTO [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTCTO),
2056 isp1362_read_reg16(isp1362_hcd, HCATLDTCTO));
2057}
2058
2059static int isp1362_show(struct seq_file *s, void *unused)
2060{
2061 struct isp1362_hcd *isp1362_hcd = s->private;
2062 struct isp1362_ep *ep;
2063 int i;
2064
2065 seq_printf(s, "%s\n%s version %s\n",
2066 isp1362_hcd_to_hcd(isp1362_hcd)->product_desc, hcd_name, DRIVER_VERSION);
2067
2068 /* collect statistics to help estimate potential win for
2069 * DMA engines that care about alignment (PXA)
2070 */
2071 seq_printf(s, "alignment: 16b/%ld 8b/%ld 4b/%ld 2b/%ld 1b/%ld\n",
2072 isp1362_hcd->stat16, isp1362_hcd->stat8, isp1362_hcd->stat4,
2073 isp1362_hcd->stat2, isp1362_hcd->stat1);
2074 seq_printf(s, "max # ptds in ATL fifo: %d\n", isp1362_hcd->atl_queue.stat_maxptds);
2075 seq_printf(s, "max # ptds in INTL fifo: %d\n", isp1362_hcd->intl_queue.stat_maxptds);
2076 seq_printf(s, "max # ptds in ISTL fifo: %d\n",
2077 max(isp1362_hcd->istl_queue[0] .stat_maxptds,
2078 isp1362_hcd->istl_queue[1] .stat_maxptds));
2079
2080 /* FIXME: don't show the following in suspended state */
2081 spin_lock_irq(&isp1362_hcd->lock);
2082
2083 dump_irq(s, "hc_irq_enable", isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2084 dump_irq(s, "hc_irq_status", isp1362_read_reg16(isp1362_hcd, HCuPINT));
2085 dump_int(s, "ohci_int_enable", isp1362_read_reg32(isp1362_hcd, HCINTENB));
2086 dump_int(s, "ohci_int_status", isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
2087 dump_ctrl(s, "ohci_control", isp1362_read_reg32(isp1362_hcd, HCCONTROL));
2088
2089 for (i = 0; i < NUM_ISP1362_IRQS; i++)
2090 if (isp1362_hcd->irq_stat[i])
2091 seq_printf(s, "%-15s: %d\n",
2092 ISP1362_INT_NAME(i), isp1362_hcd->irq_stat[i]);
2093
2094 dump_regs(s, isp1362_hcd);
2095 list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
2096 struct urb *urb;
2097
2098 seq_printf(s, "%p, ep%d%s, maxpacket %d:\n", ep, ep->epnum,
2099 ({
2100 char *s;
2101 switch (ep->nextpid) {
2102 case USB_PID_IN:
2103 s = "in";
2104 break;
2105 case USB_PID_OUT:
2106 s = "out";
2107 break;
2108 case USB_PID_SETUP:
2109 s = "setup";
2110 break;
2111 case USB_PID_ACK:
2112 s = "status";
2113 break;
2114 default:
2115 s = "?";
2116 break;
2117 }
2118 s;}), ep->maxpacket) ;
2119 list_for_each_entry(urb, &ep->hep->urb_list, urb_list) {
2120 seq_printf(s, " urb%p, %d/%d\n", urb,
2121 urb->actual_length,
2122 urb->transfer_buffer_length);
2123 }
2124 }
2125 if (!list_empty(&isp1362_hcd->async))
2126 seq_printf(s, "\n");
2127 dump_ptd_queue(&isp1362_hcd->atl_queue);
2128
2129 seq_printf(s, "periodic size= %d\n", PERIODIC_SIZE);
2130
2131 list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
2132 seq_printf(s, "branch:%2d load:%3d PTD[%d] $%04x:\n", ep->branch,
2133 isp1362_hcd->load[ep->branch], ep->ptd_index, ep->ptd_offset);
2134
2135 seq_printf(s, " %d/%p (%sdev%d ep%d%s max %d)\n",
2136 ep->interval, ep,
2137 (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2138 ep->udev->devnum, ep->epnum,
2139 (ep->epnum == 0) ? "" :
2140 ((ep->nextpid == USB_PID_IN) ?
2141 "in" : "out"), ep->maxpacket);
2142 }
2143 dump_ptd_queue(&isp1362_hcd->intl_queue);
2144
2145 seq_printf(s, "ISO:\n");
2146
2147 list_for_each_entry(ep, &isp1362_hcd->isoc, schedule) {
2148 seq_printf(s, " %d/%p (%sdev%d ep%d%s max %d)\n",
2149 ep->interval, ep,
2150 (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2151 ep->udev->devnum, ep->epnum,
2152 (ep->epnum == 0) ? "" :
2153 ((ep->nextpid == USB_PID_IN) ?
2154 "in" : "out"), ep->maxpacket);
2155 }
2156
2157 spin_unlock_irq(&isp1362_hcd->lock);
2158 seq_printf(s, "\n");
2159
2160 return 0;
2161}
2162DEFINE_SHOW_ATTRIBUTE(isp1362);
2163
2164/* expect just one isp1362_hcd per system */
2165static void create_debug_file(struct isp1362_hcd *isp1362_hcd)
2166{
2167 debugfs_create_file("isp1362", S_IRUGO, usb_debug_root, isp1362_hcd,
2168 &isp1362_fops);
2169}
2170
2171static void remove_debug_file(struct isp1362_hcd *isp1362_hcd)
2172{
2173 debugfs_lookup_and_remove("isp1362", usb_debug_root);
2174}
2175
2176/*-------------------------------------------------------------------------*/
2177
2178static void __isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2179{
2180 int tmp = 20;
2181
2182 isp1362_write_reg16(isp1362_hcd, HCSWRES, HCSWRES_MAGIC);
2183 isp1362_write_reg32(isp1362_hcd, HCCMDSTAT, OHCI_HCR);
2184 while (--tmp) {
2185 mdelay(1);
2186 if (!(isp1362_read_reg32(isp1362_hcd, HCCMDSTAT) & OHCI_HCR))
2187 break;
2188 }
2189 if (!tmp)
2190 pr_err("Software reset timeout\n");
2191}
2192
2193static void isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2194{
2195 unsigned long flags;
2196
2197 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2198 __isp1362_sw_reset(isp1362_hcd);
2199 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2200}
2201
2202static int isp1362_mem_config(struct usb_hcd *hcd)
2203{
2204 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2205 unsigned long flags;
2206 u32 total;
2207 u16 istl_size = ISP1362_ISTL_BUFSIZE;
2208 u16 intl_blksize = ISP1362_INTL_BLKSIZE + PTD_HEADER_SIZE;
2209 u16 intl_size = ISP1362_INTL_BUFFERS * intl_blksize;
2210 u16 atl_blksize = ISP1362_ATL_BLKSIZE + PTD_HEADER_SIZE;
2211 u16 atl_buffers = (ISP1362_BUF_SIZE - (istl_size + intl_size)) / atl_blksize;
2212 u16 atl_size;
2213 int i;
2214
2215 WARN_ON(istl_size & 3);
2216 WARN_ON(atl_blksize & 3);
2217 WARN_ON(intl_blksize & 3);
2218 WARN_ON(atl_blksize < PTD_HEADER_SIZE);
2219 WARN_ON(intl_blksize < PTD_HEADER_SIZE);
2220
2221 BUG_ON((unsigned)ISP1362_INTL_BUFFERS > 32);
2222 if (atl_buffers > 32)
2223 atl_buffers = 32;
2224 atl_size = atl_buffers * atl_blksize;
2225 total = atl_size + intl_size + istl_size;
2226 dev_info(hcd->self.controller, "ISP1362 Memory usage:\n");
2227 dev_info(hcd->self.controller, " ISTL: 2 * %4d: %4d @ $%04x:$%04x\n",
2228 istl_size / 2, istl_size, 0, istl_size / 2);
2229 dev_info(hcd->self.controller, " INTL: %4d * (%3zu+8): %4d @ $%04x\n",
2230 ISP1362_INTL_BUFFERS, intl_blksize - PTD_HEADER_SIZE,
2231 intl_size, istl_size);
2232 dev_info(hcd->self.controller, " ATL : %4d * (%3zu+8): %4d @ $%04x\n",
2233 atl_buffers, atl_blksize - PTD_HEADER_SIZE,
2234 atl_size, istl_size + intl_size);
2235 dev_info(hcd->self.controller, " USED/FREE: %4d %4d\n", total,
2236 ISP1362_BUF_SIZE - total);
2237
2238 if (total > ISP1362_BUF_SIZE) {
2239 dev_err(hcd->self.controller, "%s: Memory requested: %d, available %d\n",
2240 __func__, total, ISP1362_BUF_SIZE);
2241 return -ENOMEM;
2242 }
2243
2244 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2245
2246 for (i = 0; i < 2; i++) {
2247 isp1362_hcd->istl_queue[i].buf_start = i * istl_size / 2,
2248 isp1362_hcd->istl_queue[i].buf_size = istl_size / 2;
2249 isp1362_hcd->istl_queue[i].blk_size = 4;
2250 INIT_LIST_HEAD(&isp1362_hcd->istl_queue[i].active);
2251 snprintf(isp1362_hcd->istl_queue[i].name,
2252 sizeof(isp1362_hcd->istl_queue[i].name), "ISTL%d", i);
2253 DBG(3, "%s: %5s buf $%04x %d\n", __func__,
2254 isp1362_hcd->istl_queue[i].name,
2255 isp1362_hcd->istl_queue[i].buf_start,
2256 isp1362_hcd->istl_queue[i].buf_size);
2257 }
2258 isp1362_write_reg16(isp1362_hcd, HCISTLBUFSZ, istl_size / 2);
2259
2260 isp1362_hcd->intl_queue.buf_start = istl_size;
2261 isp1362_hcd->intl_queue.buf_size = intl_size;
2262 isp1362_hcd->intl_queue.buf_count = ISP1362_INTL_BUFFERS;
2263 isp1362_hcd->intl_queue.blk_size = intl_blksize;
2264 isp1362_hcd->intl_queue.buf_avail = isp1362_hcd->intl_queue.buf_count;
2265 isp1362_hcd->intl_queue.skip_map = ~0;
2266 INIT_LIST_HEAD(&isp1362_hcd->intl_queue.active);
2267
2268 isp1362_write_reg16(isp1362_hcd, HCINTLBUFSZ,
2269 isp1362_hcd->intl_queue.buf_size);
2270 isp1362_write_reg16(isp1362_hcd, HCINTLBLKSZ,
2271 isp1362_hcd->intl_queue.blk_size - PTD_HEADER_SIZE);
2272 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
2273 isp1362_write_reg32(isp1362_hcd, HCINTLLAST,
2274 1 << (ISP1362_INTL_BUFFERS - 1));
2275
2276 isp1362_hcd->atl_queue.buf_start = istl_size + intl_size;
2277 isp1362_hcd->atl_queue.buf_size = atl_size;
2278 isp1362_hcd->atl_queue.buf_count = atl_buffers;
2279 isp1362_hcd->atl_queue.blk_size = atl_blksize;
2280 isp1362_hcd->atl_queue.buf_avail = isp1362_hcd->atl_queue.buf_count;
2281 isp1362_hcd->atl_queue.skip_map = ~0;
2282 INIT_LIST_HEAD(&isp1362_hcd->atl_queue.active);
2283
2284 isp1362_write_reg16(isp1362_hcd, HCATLBUFSZ,
2285 isp1362_hcd->atl_queue.buf_size);
2286 isp1362_write_reg16(isp1362_hcd, HCATLBLKSZ,
2287 isp1362_hcd->atl_queue.blk_size - PTD_HEADER_SIZE);
2288 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
2289 isp1362_write_reg32(isp1362_hcd, HCATLLAST,
2290 1 << (atl_buffers - 1));
2291
2292 snprintf(isp1362_hcd->atl_queue.name,
2293 sizeof(isp1362_hcd->atl_queue.name), "ATL");
2294 snprintf(isp1362_hcd->intl_queue.name,
2295 sizeof(isp1362_hcd->intl_queue.name), "INTL");
2296 DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2297 isp1362_hcd->intl_queue.name,
2298 isp1362_hcd->intl_queue.buf_start,
2299 ISP1362_INTL_BUFFERS, isp1362_hcd->intl_queue.blk_size,
2300 isp1362_hcd->intl_queue.buf_size);
2301 DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2302 isp1362_hcd->atl_queue.name,
2303 isp1362_hcd->atl_queue.buf_start,
2304 atl_buffers, isp1362_hcd->atl_queue.blk_size,
2305 isp1362_hcd->atl_queue.buf_size);
2306
2307 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2308
2309 return 0;
2310}
2311
2312static int isp1362_hc_reset(struct usb_hcd *hcd)
2313{
2314 int ret = 0;
2315 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2316 unsigned long t;
2317 unsigned long timeout = 100;
2318 unsigned long flags;
2319 int clkrdy = 0;
2320
2321 pr_debug("%s:\n", __func__);
2322
2323 if (isp1362_hcd->board && isp1362_hcd->board->reset) {
2324 isp1362_hcd->board->reset(hcd->self.controller, 1);
2325 msleep(20);
2326 if (isp1362_hcd->board->clock)
2327 isp1362_hcd->board->clock(hcd->self.controller, 1);
2328 isp1362_hcd->board->reset(hcd->self.controller, 0);
2329 } else
2330 isp1362_sw_reset(isp1362_hcd);
2331
2332 /* chip has been reset. First we need to see a clock */
2333 t = jiffies + msecs_to_jiffies(timeout);
2334 while (!clkrdy && time_before_eq(jiffies, t)) {
2335 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2336 clkrdy = isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_CLKRDY;
2337 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2338 if (!clkrdy)
2339 msleep(4);
2340 }
2341
2342 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2343 isp1362_write_reg16(isp1362_hcd, HCuPINT, HCuPINT_CLKRDY);
2344 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2345 if (!clkrdy) {
2346 pr_err("Clock not ready after %lums\n", timeout);
2347 ret = -ENODEV;
2348 }
2349 return ret;
2350}
2351
2352static void isp1362_hc_stop(struct usb_hcd *hcd)
2353{
2354 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2355 unsigned long flags;
2356 u32 tmp;
2357
2358 pr_debug("%s:\n", __func__);
2359
2360 del_timer_sync(&hcd->rh_timer);
2361
2362 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2363
2364 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2365
2366 /* Switch off power for all ports */
2367 tmp = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2368 tmp &= ~(RH_A_NPS | RH_A_PSM);
2369 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, tmp);
2370 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2371
2372 /* Reset the chip */
2373 if (isp1362_hcd->board && isp1362_hcd->board->reset)
2374 isp1362_hcd->board->reset(hcd->self.controller, 1);
2375 else
2376 __isp1362_sw_reset(isp1362_hcd);
2377
2378 if (isp1362_hcd->board && isp1362_hcd->board->clock)
2379 isp1362_hcd->board->clock(hcd->self.controller, 0);
2380
2381 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2382}
2383
2384#ifdef CHIP_BUFFER_TEST
2385static int isp1362_chip_test(struct isp1362_hcd *isp1362_hcd)
2386{
2387 int ret = 0;
2388 u16 *ref;
2389 unsigned long flags;
2390
2391 ref = kmalloc(2 * ISP1362_BUF_SIZE, GFP_KERNEL);
2392 if (ref) {
2393 int offset;
2394 u16 *tst = &ref[ISP1362_BUF_SIZE / 2];
2395
2396 for (offset = 0; offset < ISP1362_BUF_SIZE / 2; offset++) {
2397 ref[offset] = ~offset;
2398 tst[offset] = offset;
2399 }
2400
2401 for (offset = 0; offset < 4; offset++) {
2402 int j;
2403
2404 for (j = 0; j < 8; j++) {
2405 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2406 isp1362_write_buffer(isp1362_hcd, (u8 *)ref + offset, 0, j);
2407 isp1362_read_buffer(isp1362_hcd, (u8 *)tst + offset, 0, j);
2408 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2409
2410 if (memcmp(ref, tst, j)) {
2411 ret = -ENODEV;
2412 pr_err("%s: memory check with %d byte offset %d failed\n",
2413 __func__, j, offset);
2414 dump_data((u8 *)ref + offset, j);
2415 dump_data((u8 *)tst + offset, j);
2416 }
2417 }
2418 }
2419
2420 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2421 isp1362_write_buffer(isp1362_hcd, ref, 0, ISP1362_BUF_SIZE);
2422 isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2423 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2424
2425 if (memcmp(ref, tst, ISP1362_BUF_SIZE)) {
2426 ret = -ENODEV;
2427 pr_err("%s: memory check failed\n", __func__);
2428 dump_data((u8 *)tst, ISP1362_BUF_SIZE / 2);
2429 }
2430
2431 for (offset = 0; offset < 256; offset++) {
2432 int test_size = 0;
2433
2434 yield();
2435
2436 memset(tst, 0, ISP1362_BUF_SIZE);
2437 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2438 isp1362_write_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2439 isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2440 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2441 if (memcmp(tst, tst + (ISP1362_BUF_SIZE / (2 * sizeof(*tst))),
2442 ISP1362_BUF_SIZE / 2)) {
2443 pr_err("%s: Failed to clear buffer\n", __func__);
2444 dump_data((u8 *)tst, ISP1362_BUF_SIZE);
2445 break;
2446 }
2447 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2448 isp1362_write_buffer(isp1362_hcd, ref, offset * 2, PTD_HEADER_SIZE);
2449 isp1362_write_buffer(isp1362_hcd, ref + PTD_HEADER_SIZE / sizeof(*ref),
2450 offset * 2 + PTD_HEADER_SIZE, test_size);
2451 isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2452 PTD_HEADER_SIZE + test_size);
2453 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2454 if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2455 dump_data(((u8 *)ref) + offset, PTD_HEADER_SIZE + test_size);
2456 dump_data((u8 *)tst, PTD_HEADER_SIZE + test_size);
2457 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2458 isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2459 PTD_HEADER_SIZE + test_size);
2460 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2461 if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2462 ret = -ENODEV;
2463 pr_err("%s: memory check with offset %02x failed\n",
2464 __func__, offset);
2465 break;
2466 }
2467 pr_warn("%s: memory check with offset %02x ok after second read\n",
2468 __func__, offset);
2469 }
2470 }
2471 kfree(ref);
2472 }
2473 return ret;
2474}
2475#endif
2476
2477static int isp1362_hc_start(struct usb_hcd *hcd)
2478{
2479 int ret;
2480 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2481 struct isp1362_platform_data *board = isp1362_hcd->board;
2482 u16 hwcfg;
2483 u16 chipid;
2484 unsigned long flags;
2485
2486 pr_debug("%s:\n", __func__);
2487
2488 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2489 chipid = isp1362_read_reg16(isp1362_hcd, HCCHIPID);
2490 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2491
2492 if ((chipid & HCCHIPID_MASK) != HCCHIPID_MAGIC) {
2493 pr_err("%s: Invalid chip ID %04x\n", __func__, chipid);
2494 return -ENODEV;
2495 }
2496
2497#ifdef CHIP_BUFFER_TEST
2498 ret = isp1362_chip_test(isp1362_hcd);
2499 if (ret)
2500 return -ENODEV;
2501#endif
2502 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2503 /* clear interrupt status and disable all interrupt sources */
2504 isp1362_write_reg16(isp1362_hcd, HCuPINT, 0xff);
2505 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2506
2507 /* HW conf */
2508 hwcfg = HCHWCFG_INT_ENABLE | HCHWCFG_DBWIDTH(1);
2509 if (board->sel15Kres)
2510 hwcfg |= HCHWCFG_PULLDOWN_DS2 |
2511 ((MAX_ROOT_PORTS > 1) ? HCHWCFG_PULLDOWN_DS1 : 0);
2512 if (board->clknotstop)
2513 hwcfg |= HCHWCFG_CLKNOTSTOP;
2514 if (board->oc_enable)
2515 hwcfg |= HCHWCFG_ANALOG_OC;
2516 if (board->int_act_high)
2517 hwcfg |= HCHWCFG_INT_POL;
2518 if (board->int_edge_triggered)
2519 hwcfg |= HCHWCFG_INT_TRIGGER;
2520 if (board->dreq_act_high)
2521 hwcfg |= HCHWCFG_DREQ_POL;
2522 if (board->dack_act_high)
2523 hwcfg |= HCHWCFG_DACK_POL;
2524 isp1362_write_reg16(isp1362_hcd, HCHWCFG, hwcfg);
2525 isp1362_show_reg(isp1362_hcd, HCHWCFG);
2526 isp1362_write_reg16(isp1362_hcd, HCDMACFG, 0);
2527 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2528
2529 ret = isp1362_mem_config(hcd);
2530 if (ret)
2531 return ret;
2532
2533 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2534
2535 /* Root hub conf */
2536 isp1362_hcd->rhdesca = 0;
2537 if (board->no_power_switching)
2538 isp1362_hcd->rhdesca |= RH_A_NPS;
2539 if (board->power_switching_mode)
2540 isp1362_hcd->rhdesca |= RH_A_PSM;
2541 if (board->potpg)
2542 isp1362_hcd->rhdesca |= (board->potpg << 24) & RH_A_POTPGT;
2543 else
2544 isp1362_hcd->rhdesca |= (25 << 24) & RH_A_POTPGT;
2545
2546 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca & ~RH_A_OCPM);
2547 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca | RH_A_OCPM);
2548 isp1362_hcd->rhdesca = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2549
2550 isp1362_hcd->rhdescb = RH_B_PPCM;
2551 isp1362_write_reg32(isp1362_hcd, HCRHDESCB, isp1362_hcd->rhdescb);
2552 isp1362_hcd->rhdescb = isp1362_read_reg32(isp1362_hcd, HCRHDESCB);
2553
2554 isp1362_read_reg32(isp1362_hcd, HCFMINTVL);
2555 isp1362_write_reg32(isp1362_hcd, HCFMINTVL, (FSMP(FI) << 16) | FI);
2556 isp1362_write_reg32(isp1362_hcd, HCLSTHRESH, LSTHRESH);
2557
2558 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2559
2560 isp1362_hcd->hc_control = OHCI_USB_OPER;
2561 hcd->state = HC_STATE_RUNNING;
2562
2563 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2564 /* Set up interrupts */
2565 isp1362_hcd->intenb = OHCI_INTR_MIE | OHCI_INTR_RHSC | OHCI_INTR_UE;
2566 isp1362_hcd->intenb |= OHCI_INTR_RD;
2567 isp1362_hcd->irqenb = HCuPINT_OPR | HCuPINT_SUSP;
2568 isp1362_write_reg32(isp1362_hcd, HCINTENB, isp1362_hcd->intenb);
2569 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
2570
2571 /* Go operational */
2572 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
2573 /* enable global power */
2574 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC | RH_HS_DRWE);
2575
2576 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2577
2578 return 0;
2579}
2580
2581/*-------------------------------------------------------------------------*/
2582
2583static const struct hc_driver isp1362_hc_driver = {
2584 .description = hcd_name,
2585 .product_desc = "ISP1362 Host Controller",
2586 .hcd_priv_size = sizeof(struct isp1362_hcd),
2587
2588 .irq = isp1362_irq,
2589 .flags = HCD_USB11 | HCD_MEMORY,
2590
2591 .reset = isp1362_hc_reset,
2592 .start = isp1362_hc_start,
2593 .stop = isp1362_hc_stop,
2594
2595 .urb_enqueue = isp1362_urb_enqueue,
2596 .urb_dequeue = isp1362_urb_dequeue,
2597 .endpoint_disable = isp1362_endpoint_disable,
2598
2599 .get_frame_number = isp1362_get_frame,
2600
2601 .hub_status_data = isp1362_hub_status_data,
2602 .hub_control = isp1362_hub_control,
2603 .bus_suspend = isp1362_bus_suspend,
2604 .bus_resume = isp1362_bus_resume,
2605};
2606
2607/*-------------------------------------------------------------------------*/
2608
2609static void isp1362_remove(struct platform_device *pdev)
2610{
2611 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2612 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2613
2614 remove_debug_file(isp1362_hcd);
2615 DBG(0, "%s: Removing HCD\n", __func__);
2616 usb_remove_hcd(hcd);
2617 DBG(0, "%s: put_hcd\n", __func__);
2618 usb_put_hcd(hcd);
2619 DBG(0, "%s: Done\n", __func__);
2620}
2621
2622static int isp1362_probe(struct platform_device *pdev)
2623{
2624 struct usb_hcd *hcd;
2625 struct isp1362_hcd *isp1362_hcd;
2626 struct resource *data, *irq_res;
2627 void __iomem *addr_reg;
2628 void __iomem *data_reg;
2629 int irq;
2630 int retval = 0;
2631 unsigned int irq_flags = 0;
2632
2633 if (usb_disabled())
2634 return -ENODEV;
2635
2636 /* basic sanity checks first. board-specific init logic should
2637 * have initialized this the three resources and probably board
2638 * specific platform_data. we don't probe for IRQs, and do only
2639 * minimal sanity checking.
2640 */
2641 if (pdev->num_resources < 3)
2642 return -ENODEV;
2643
2644 irq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
2645 if (!irq_res)
2646 return -ENODEV;
2647
2648 irq = irq_res->start;
2649
2650 addr_reg = devm_platform_ioremap_resource(pdev, 1);
2651 if (IS_ERR(addr_reg))
2652 return PTR_ERR(addr_reg);
2653
2654 data_reg = devm_platform_get_and_ioremap_resource(pdev, 0, &data);
2655 if (IS_ERR(data_reg))
2656 return PTR_ERR(data_reg);
2657
2658 /* allocate and initialize hcd */
2659 hcd = usb_create_hcd(&isp1362_hc_driver, &pdev->dev, dev_name(&pdev->dev));
2660 if (!hcd)
2661 return -ENOMEM;
2662
2663 hcd->rsrc_start = data->start;
2664 isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2665 isp1362_hcd->data_reg = data_reg;
2666 isp1362_hcd->addr_reg = addr_reg;
2667
2668 isp1362_hcd->next_statechange = jiffies;
2669 spin_lock_init(&isp1362_hcd->lock);
2670 INIT_LIST_HEAD(&isp1362_hcd->async);
2671 INIT_LIST_HEAD(&isp1362_hcd->periodic);
2672 INIT_LIST_HEAD(&isp1362_hcd->isoc);
2673 INIT_LIST_HEAD(&isp1362_hcd->remove_list);
2674 isp1362_hcd->board = dev_get_platdata(&pdev->dev);
2675#if USE_PLATFORM_DELAY
2676 if (!isp1362_hcd->board->delay) {
2677 dev_err(hcd->self.controller, "No platform delay function given\n");
2678 retval = -ENODEV;
2679 goto err;
2680 }
2681#endif
2682
2683 if (irq_res->flags & IORESOURCE_IRQ_HIGHEDGE)
2684 irq_flags |= IRQF_TRIGGER_RISING;
2685 if (irq_res->flags & IORESOURCE_IRQ_LOWEDGE)
2686 irq_flags |= IRQF_TRIGGER_FALLING;
2687 if (irq_res->flags & IORESOURCE_IRQ_HIGHLEVEL)
2688 irq_flags |= IRQF_TRIGGER_HIGH;
2689 if (irq_res->flags & IORESOURCE_IRQ_LOWLEVEL)
2690 irq_flags |= IRQF_TRIGGER_LOW;
2691
2692 retval = usb_add_hcd(hcd, irq, irq_flags | IRQF_SHARED);
2693 if (retval != 0)
2694 goto err;
2695 device_wakeup_enable(hcd->self.controller);
2696
2697 dev_info(&pdev->dev, "%s, irq %d\n", hcd->product_desc, irq);
2698
2699 create_debug_file(isp1362_hcd);
2700
2701 return 0;
2702
2703 err:
2704 usb_put_hcd(hcd);
2705
2706 return retval;
2707}
2708
2709#ifdef CONFIG_PM
2710static int isp1362_suspend(struct platform_device *pdev, pm_message_t state)
2711{
2712 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2713 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2714 unsigned long flags;
2715 int retval = 0;
2716
2717 DBG(0, "%s: Suspending device\n", __func__);
2718
2719 if (state.event == PM_EVENT_FREEZE) {
2720 DBG(0, "%s: Suspending root hub\n", __func__);
2721 retval = isp1362_bus_suspend(hcd);
2722 } else {
2723 DBG(0, "%s: Suspending RH ports\n", __func__);
2724 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2725 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2726 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2727 }
2728 if (retval == 0)
2729 pdev->dev.power.power_state = state;
2730 return retval;
2731}
2732
2733static int isp1362_resume(struct platform_device *pdev)
2734{
2735 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2736 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2737 unsigned long flags;
2738
2739 DBG(0, "%s: Resuming\n", __func__);
2740
2741 if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
2742 DBG(0, "%s: Resume RH ports\n", __func__);
2743 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2744 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC);
2745 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2746 return 0;
2747 }
2748
2749 pdev->dev.power.power_state = PMSG_ON;
2750
2751 return isp1362_bus_resume(isp1362_hcd_to_hcd(isp1362_hcd));
2752}
2753#else
2754#define isp1362_suspend NULL
2755#define isp1362_resume NULL
2756#endif
2757
2758static struct platform_driver isp1362_driver = {
2759 .probe = isp1362_probe,
2760 .remove_new = isp1362_remove,
2761
2762 .suspend = isp1362_suspend,
2763 .resume = isp1362_resume,
2764 .driver = {
2765 .name = hcd_name,
2766 },
2767};
2768
2769module_platform_driver(isp1362_driver);
1/*
2 * ISP1362 HCD (Host Controller Driver) for USB.
3 *
4 * Copyright (C) 2005 Lothar Wassmann <LW@KARO-electronics.de>
5 *
6 * Derived from the SL811 HCD, rewritten for ISP116x.
7 * Copyright (C) 2005 Olav Kongas <ok@artecdesign.ee>
8 *
9 * Portions:
10 * Copyright (C) 2004 Psion Teklogix (for NetBook PRO)
11 * Copyright (C) 2004 David Brownell
12 */
13
14/*
15 * The ISP1362 chip requires a large delay (300ns and 462ns) between
16 * accesses to the address and data register.
17 * The following timing options exist:
18 *
19 * 1. Configure your memory controller to add such delays if it can (the best)
20 * 2. Implement platform-specific delay function possibly
21 * combined with configuring the memory controller; see
22 * include/linux/usb_isp1362.h for more info.
23 * 3. Use ndelay (easiest, poorest).
24 *
25 * Use the corresponding macros USE_PLATFORM_DELAY and USE_NDELAY in the
26 * platform specific section of isp1362.h to select the appropriate variant.
27 *
28 * Also note that according to the Philips "ISP1362 Errata" document
29 * Rev 1.00 from 27 May data corruption may occur when the #WR signal
30 * is reasserted (even with #CS deasserted) within 132ns after a
31 * write cycle to any controller register. If the hardware doesn't
32 * implement the recommended fix (gating the #WR with #CS) software
33 * must ensure that no further write cycle (not necessarily to the chip!)
34 * is issued by the CPU within this interval.
35
36 * For PXA25x this can be ensured by using VLIO with the maximum
37 * recovery time (MSCx = 0x7f8c) with a memory clock of 99.53 MHz.
38 */
39
40#undef ISP1362_DEBUG
41
42/*
43 * The PXA255 UDC apparently doesn't handle GET_STATUS, GET_CONFIG and
44 * GET_INTERFACE requests correctly when the SETUP and DATA stages of the
45 * requests are carried out in separate frames. This will delay any SETUP
46 * packets until the start of the next frame so that this situation is
47 * unlikely to occur (and makes usbtest happy running with a PXA255 target
48 * device).
49 */
50#undef BUGGY_PXA2XX_UDC_USBTEST
51
52#undef PTD_TRACE
53#undef URB_TRACE
54#undef VERBOSE
55#undef REGISTERS
56
57/* This enables a memory test on the ISP1362 chip memory to make sure the
58 * chip access timing is correct.
59 */
60#undef CHIP_BUFFER_TEST
61
62#include <linux/module.h>
63#include <linux/moduleparam.h>
64#include <linux/kernel.h>
65#include <linux/delay.h>
66#include <linux/ioport.h>
67#include <linux/sched.h>
68#include <linux/slab.h>
69#include <linux/errno.h>
70#include <linux/list.h>
71#include <linux/interrupt.h>
72#include <linux/usb.h>
73#include <linux/usb/isp1362.h>
74#include <linux/usb/hcd.h>
75#include <linux/platform_device.h>
76#include <linux/pm.h>
77#include <linux/io.h>
78#include <linux/bitmap.h>
79#include <linux/prefetch.h>
80#include <linux/debugfs.h>
81#include <linux/seq_file.h>
82
83#include <asm/irq.h>
84#include <asm/byteorder.h>
85#include <asm/unaligned.h>
86
87static int dbg_level;
88#ifdef ISP1362_DEBUG
89module_param(dbg_level, int, 0644);
90#else
91module_param(dbg_level, int, 0);
92#endif
93
94#include "../core/usb.h"
95#include "isp1362.h"
96
97
98#define DRIVER_VERSION "2005-04-04"
99#define DRIVER_DESC "ISP1362 USB Host Controller Driver"
100
101MODULE_DESCRIPTION(DRIVER_DESC);
102MODULE_LICENSE("GPL");
103
104static const char hcd_name[] = "isp1362-hcd";
105
106static void isp1362_hc_stop(struct usb_hcd *hcd);
107static int isp1362_hc_start(struct usb_hcd *hcd);
108
109/*-------------------------------------------------------------------------*/
110
111/*
112 * When called from the interrupthandler only isp1362_hcd->irqenb is modified,
113 * since the interrupt handler will write isp1362_hcd->irqenb to HCuPINT upon
114 * completion.
115 * We don't need a 'disable' counterpart, since interrupts will be disabled
116 * only by the interrupt handler.
117 */
118static inline void isp1362_enable_int(struct isp1362_hcd *isp1362_hcd, u16 mask)
119{
120 if ((isp1362_hcd->irqenb | mask) == isp1362_hcd->irqenb)
121 return;
122 if (mask & ~isp1362_hcd->irqenb)
123 isp1362_write_reg16(isp1362_hcd, HCuPINT, mask & ~isp1362_hcd->irqenb);
124 isp1362_hcd->irqenb |= mask;
125 if (isp1362_hcd->irq_active)
126 return;
127 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
128}
129
130/*-------------------------------------------------------------------------*/
131
132static inline struct isp1362_ep_queue *get_ptd_queue(struct isp1362_hcd *isp1362_hcd,
133 u16 offset)
134{
135 struct isp1362_ep_queue *epq = NULL;
136
137 if (offset < isp1362_hcd->istl_queue[1].buf_start)
138 epq = &isp1362_hcd->istl_queue[0];
139 else if (offset < isp1362_hcd->intl_queue.buf_start)
140 epq = &isp1362_hcd->istl_queue[1];
141 else if (offset < isp1362_hcd->atl_queue.buf_start)
142 epq = &isp1362_hcd->intl_queue;
143 else if (offset < isp1362_hcd->atl_queue.buf_start +
144 isp1362_hcd->atl_queue.buf_size)
145 epq = &isp1362_hcd->atl_queue;
146
147 if (epq)
148 DBG(1, "%s: PTD $%04x is on %s queue\n", __func__, offset, epq->name);
149 else
150 pr_warning("%s: invalid PTD $%04x\n", __func__, offset);
151
152 return epq;
153}
154
155static inline int get_ptd_offset(struct isp1362_ep_queue *epq, u8 index)
156{
157 int offset;
158
159 if (index * epq->blk_size > epq->buf_size) {
160 pr_warning("%s: Bad %s index %d(%d)\n", __func__, epq->name, index,
161 epq->buf_size / epq->blk_size);
162 return -EINVAL;
163 }
164 offset = epq->buf_start + index * epq->blk_size;
165 DBG(3, "%s: %s PTD[%02x] # %04x\n", __func__, epq->name, index, offset);
166
167 return offset;
168}
169
170/*-------------------------------------------------------------------------*/
171
172static inline u16 max_transfer_size(struct isp1362_ep_queue *epq, size_t size,
173 int mps)
174{
175 u16 xfer_size = min_t(size_t, MAX_XFER_SIZE, size);
176
177 xfer_size = min_t(size_t, xfer_size, epq->buf_avail * epq->blk_size - PTD_HEADER_SIZE);
178 if (xfer_size < size && xfer_size % mps)
179 xfer_size -= xfer_size % mps;
180
181 return xfer_size;
182}
183
184static int claim_ptd_buffers(struct isp1362_ep_queue *epq,
185 struct isp1362_ep *ep, u16 len)
186{
187 int ptd_offset = -EINVAL;
188 int num_ptds = ((len + PTD_HEADER_SIZE - 1) / epq->blk_size) + 1;
189 int found;
190
191 BUG_ON(len > epq->buf_size);
192
193 if (!epq->buf_avail)
194 return -ENOMEM;
195
196 if (ep->num_ptds)
197 pr_err("%s: %s len %d/%d num_ptds %d buf_map %08lx skip_map %08lx\n", __func__,
198 epq->name, len, epq->blk_size, num_ptds, epq->buf_map, epq->skip_map);
199 BUG_ON(ep->num_ptds != 0);
200
201 found = bitmap_find_next_zero_area(&epq->buf_map, epq->buf_count, 0,
202 num_ptds, 0);
203 if (found >= epq->buf_count)
204 return -EOVERFLOW;
205
206 DBG(1, "%s: Found %d PTDs[%d] for %d/%d byte\n", __func__,
207 num_ptds, found, len, (int)(epq->blk_size - PTD_HEADER_SIZE));
208 ptd_offset = get_ptd_offset(epq, found);
209 WARN_ON(ptd_offset < 0);
210 ep->ptd_offset = ptd_offset;
211 ep->num_ptds += num_ptds;
212 epq->buf_avail -= num_ptds;
213 BUG_ON(epq->buf_avail > epq->buf_count);
214 ep->ptd_index = found;
215 bitmap_set(&epq->buf_map, found, num_ptds);
216 DBG(1, "%s: Done %s PTD[%d] $%04x, avail %d count %d claimed %d %08lx:%08lx\n",
217 __func__, epq->name, ep->ptd_index, ep->ptd_offset,
218 epq->buf_avail, epq->buf_count, num_ptds, epq->buf_map, epq->skip_map);
219
220 return found;
221}
222
223static inline void release_ptd_buffers(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
224{
225 int last = ep->ptd_index + ep->num_ptds;
226
227 if (last > epq->buf_count)
228 pr_err("%s: ep %p req %d len %d %s PTD[%d] $%04x num_ptds %d buf_count %d buf_avail %d buf_map %08lx skip_map %08lx\n",
229 __func__, ep, ep->num_req, ep->length, epq->name, ep->ptd_index,
230 ep->ptd_offset, ep->num_ptds, epq->buf_count, epq->buf_avail,
231 epq->buf_map, epq->skip_map);
232 BUG_ON(last > epq->buf_count);
233
234 bitmap_clear(&epq->buf_map, ep->ptd_index, ep->num_ptds);
235 bitmap_set(&epq->skip_map, ep->ptd_index, ep->num_ptds);
236 epq->buf_avail += ep->num_ptds;
237 epq->ptd_count--;
238
239 BUG_ON(epq->buf_avail > epq->buf_count);
240 BUG_ON(epq->ptd_count > epq->buf_count);
241
242 DBG(1, "%s: Done %s PTDs $%04x released %d avail %d count %d\n",
243 __func__, epq->name,
244 ep->ptd_offset, ep->num_ptds, epq->buf_avail, epq->buf_count);
245 DBG(1, "%s: buf_map %08lx skip_map %08lx\n", __func__,
246 epq->buf_map, epq->skip_map);
247
248 ep->num_ptds = 0;
249 ep->ptd_offset = -EINVAL;
250 ep->ptd_index = -EINVAL;
251}
252
253/*-------------------------------------------------------------------------*/
254
255/*
256 Set up PTD's.
257*/
258static void prepare_ptd(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
259 struct isp1362_ep *ep, struct isp1362_ep_queue *epq,
260 u16 fno)
261{
262 struct ptd *ptd;
263 int toggle;
264 int dir;
265 u16 len;
266 size_t buf_len = urb->transfer_buffer_length - urb->actual_length;
267
268 DBG(3, "%s: %s ep %p\n", __func__, epq->name, ep);
269
270 ptd = &ep->ptd;
271
272 ep->data = (unsigned char *)urb->transfer_buffer + urb->actual_length;
273
274 switch (ep->nextpid) {
275 case USB_PID_IN:
276 toggle = usb_gettoggle(urb->dev, ep->epnum, 0);
277 dir = PTD_DIR_IN;
278 if (usb_pipecontrol(urb->pipe)) {
279 len = min_t(size_t, ep->maxpacket, buf_len);
280 } else if (usb_pipeisoc(urb->pipe)) {
281 len = min_t(size_t, urb->iso_frame_desc[fno].length, MAX_XFER_SIZE);
282 ep->data = urb->transfer_buffer + urb->iso_frame_desc[fno].offset;
283 } else
284 len = max_transfer_size(epq, buf_len, ep->maxpacket);
285 DBG(1, "%s: IN len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
286 (int)buf_len);
287 break;
288 case USB_PID_OUT:
289 toggle = usb_gettoggle(urb->dev, ep->epnum, 1);
290 dir = PTD_DIR_OUT;
291 if (usb_pipecontrol(urb->pipe))
292 len = min_t(size_t, ep->maxpacket, buf_len);
293 else if (usb_pipeisoc(urb->pipe))
294 len = min_t(size_t, urb->iso_frame_desc[0].length, MAX_XFER_SIZE);
295 else
296 len = max_transfer_size(epq, buf_len, ep->maxpacket);
297 if (len == 0)
298 pr_info("%s: Sending ZERO packet: %d\n", __func__,
299 urb->transfer_flags & URB_ZERO_PACKET);
300 DBG(1, "%s: OUT len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
301 (int)buf_len);
302 break;
303 case USB_PID_SETUP:
304 toggle = 0;
305 dir = PTD_DIR_SETUP;
306 len = sizeof(struct usb_ctrlrequest);
307 DBG(1, "%s: SETUP len %d\n", __func__, len);
308 ep->data = urb->setup_packet;
309 break;
310 case USB_PID_ACK:
311 toggle = 1;
312 len = 0;
313 dir = (urb->transfer_buffer_length && usb_pipein(urb->pipe)) ?
314 PTD_DIR_OUT : PTD_DIR_IN;
315 DBG(1, "%s: ACK len %d\n", __func__, len);
316 break;
317 default:
318 toggle = dir = len = 0;
319 pr_err("%s@%d: ep->nextpid %02x\n", __func__, __LINE__, ep->nextpid);
320 BUG_ON(1);
321 }
322
323 ep->length = len;
324 if (!len)
325 ep->data = NULL;
326
327 ptd->count = PTD_CC_MSK | PTD_ACTIVE_MSK | PTD_TOGGLE(toggle);
328 ptd->mps = PTD_MPS(ep->maxpacket) | PTD_SPD(urb->dev->speed == USB_SPEED_LOW) |
329 PTD_EP(ep->epnum);
330 ptd->len = PTD_LEN(len) | PTD_DIR(dir);
331 ptd->faddr = PTD_FA(usb_pipedevice(urb->pipe));
332
333 if (usb_pipeint(urb->pipe)) {
334 ptd->faddr |= PTD_SF_INT(ep->branch);
335 ptd->faddr |= PTD_PR(ep->interval ? __ffs(ep->interval) : 0);
336 }
337 if (usb_pipeisoc(urb->pipe))
338 ptd->faddr |= PTD_SF_ISO(fno);
339
340 DBG(1, "%s: Finished\n", __func__);
341}
342
343static void isp1362_write_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
344 struct isp1362_ep_queue *epq)
345{
346 struct ptd *ptd = &ep->ptd;
347 int len = PTD_GET_DIR(ptd) == PTD_DIR_IN ? 0 : ep->length;
348
349 prefetch(ptd);
350 isp1362_write_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
351 if (len)
352 isp1362_write_buffer(isp1362_hcd, ep->data,
353 ep->ptd_offset + PTD_HEADER_SIZE, len);
354
355 dump_ptd(ptd);
356 dump_ptd_out_data(ptd, ep->data);
357}
358
359static void isp1362_read_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
360 struct isp1362_ep_queue *epq)
361{
362 struct ptd *ptd = &ep->ptd;
363 int act_len;
364
365 WARN_ON(list_empty(&ep->active));
366 BUG_ON(ep->ptd_offset < 0);
367
368 list_del_init(&ep->active);
369 DBG(1, "%s: ep %p removed from active list %p\n", __func__, ep, &epq->active);
370
371 prefetchw(ptd);
372 isp1362_read_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
373 dump_ptd(ptd);
374 act_len = PTD_GET_COUNT(ptd);
375 if (PTD_GET_DIR(ptd) != PTD_DIR_IN || act_len == 0)
376 return;
377 if (act_len > ep->length)
378 pr_err("%s: ep %p PTD $%04x act_len %d ep->length %d\n", __func__, ep,
379 ep->ptd_offset, act_len, ep->length);
380 BUG_ON(act_len > ep->length);
381 /* Only transfer the amount of data that has actually been overwritten
382 * in the chip buffer. We don't want any data that doesn't belong to the
383 * transfer to leak out of the chip to the callers transfer buffer!
384 */
385 prefetchw(ep->data);
386 isp1362_read_buffer(isp1362_hcd, ep->data,
387 ep->ptd_offset + PTD_HEADER_SIZE, act_len);
388 dump_ptd_in_data(ptd, ep->data);
389}
390
391/*
392 * INT PTDs will stay in the chip until data is available.
393 * This function will remove a PTD from the chip when the URB is dequeued.
394 * Must be called with the spinlock held and IRQs disabled
395 */
396static void remove_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
397
398{
399 int index;
400 struct isp1362_ep_queue *epq;
401
402 DBG(1, "%s: ep %p PTD[%d] $%04x\n", __func__, ep, ep->ptd_index, ep->ptd_offset);
403 BUG_ON(ep->ptd_offset < 0);
404
405 epq = get_ptd_queue(isp1362_hcd, ep->ptd_offset);
406 BUG_ON(!epq);
407
408 /* put ep in remove_list for cleanup */
409 WARN_ON(!list_empty(&ep->remove_list));
410 list_add_tail(&ep->remove_list, &isp1362_hcd->remove_list);
411 /* let SOF interrupt handle the cleanup */
412 isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
413
414 index = ep->ptd_index;
415 if (index < 0)
416 /* ISO queues don't have SKIP registers */
417 return;
418
419 DBG(1, "%s: Disabling PTD[%02x] $%04x %08lx|%08x\n", __func__,
420 index, ep->ptd_offset, epq->skip_map, 1 << index);
421
422 /* prevent further processing of PTD (will be effective after next SOF) */
423 epq->skip_map |= 1 << index;
424 if (epq == &isp1362_hcd->atl_queue) {
425 DBG(2, "%s: ATLSKIP = %08x -> %08lx\n", __func__,
426 isp1362_read_reg32(isp1362_hcd, HCATLSKIP), epq->skip_map);
427 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, epq->skip_map);
428 if (~epq->skip_map == 0)
429 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
430 } else if (epq == &isp1362_hcd->intl_queue) {
431 DBG(2, "%s: INTLSKIP = %08x -> %08lx\n", __func__,
432 isp1362_read_reg32(isp1362_hcd, HCINTLSKIP), epq->skip_map);
433 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, epq->skip_map);
434 if (~epq->skip_map == 0)
435 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
436 }
437}
438
439/*
440 Take done or failed requests out of schedule. Give back
441 processed urbs.
442*/
443static void finish_request(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
444 struct urb *urb, int status)
445 __releases(isp1362_hcd->lock)
446 __acquires(isp1362_hcd->lock)
447{
448 urb->hcpriv = NULL;
449 ep->error_count = 0;
450
451 if (usb_pipecontrol(urb->pipe))
452 ep->nextpid = USB_PID_SETUP;
453
454 URB_DBG("%s: req %d FA %d ep%d%s %s: len %d/%d %s stat %d\n", __func__,
455 ep->num_req, usb_pipedevice(urb->pipe),
456 usb_pipeendpoint(urb->pipe),
457 !usb_pipein(urb->pipe) ? "out" : "in",
458 usb_pipecontrol(urb->pipe) ? "ctrl" :
459 usb_pipeint(urb->pipe) ? "int" :
460 usb_pipebulk(urb->pipe) ? "bulk" :
461 "iso",
462 urb->actual_length, urb->transfer_buffer_length,
463 !(urb->transfer_flags & URB_SHORT_NOT_OK) ?
464 "short_ok" : "", urb->status);
465
466
467 usb_hcd_unlink_urb_from_ep(isp1362_hcd_to_hcd(isp1362_hcd), urb);
468 spin_unlock(&isp1362_hcd->lock);
469 usb_hcd_giveback_urb(isp1362_hcd_to_hcd(isp1362_hcd), urb, status);
470 spin_lock(&isp1362_hcd->lock);
471
472 /* take idle endpoints out of the schedule right away */
473 if (!list_empty(&ep->hep->urb_list))
474 return;
475
476 /* async deschedule */
477 if (!list_empty(&ep->schedule)) {
478 list_del_init(&ep->schedule);
479 return;
480 }
481
482
483 if (ep->interval) {
484 /* periodic deschedule */
485 DBG(1, "deschedule qh%d/%p branch %d load %d bandwidth %d -> %d\n", ep->interval,
486 ep, ep->branch, ep->load,
487 isp1362_hcd->load[ep->branch],
488 isp1362_hcd->load[ep->branch] - ep->load);
489 isp1362_hcd->load[ep->branch] -= ep->load;
490 ep->branch = PERIODIC_SIZE;
491 }
492}
493
494/*
495 * Analyze transfer results, handle partial transfers and errors
496*/
497static void postproc_ep(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
498{
499 struct urb *urb = get_urb(ep);
500 struct usb_device *udev;
501 struct ptd *ptd;
502 int short_ok;
503 u16 len;
504 int urbstat = -EINPROGRESS;
505 u8 cc;
506
507 DBG(2, "%s: ep %p req %d\n", __func__, ep, ep->num_req);
508
509 udev = urb->dev;
510 ptd = &ep->ptd;
511 cc = PTD_GET_CC(ptd);
512 if (cc == PTD_NOTACCESSED) {
513 pr_err("%s: req %d PTD %p Untouched by ISP1362\n", __func__,
514 ep->num_req, ptd);
515 cc = PTD_DEVNOTRESP;
516 }
517
518 short_ok = !(urb->transfer_flags & URB_SHORT_NOT_OK);
519 len = urb->transfer_buffer_length - urb->actual_length;
520
521 /* Data underrun is special. For allowed underrun
522 we clear the error and continue as normal. For
523 forbidden underrun we finish the DATA stage
524 immediately while for control transfer,
525 we do a STATUS stage.
526 */
527 if (cc == PTD_DATAUNDERRUN) {
528 if (short_ok) {
529 DBG(1, "%s: req %d Allowed data underrun short_%sok %d/%d/%d byte\n",
530 __func__, ep->num_req, short_ok ? "" : "not_",
531 PTD_GET_COUNT(ptd), ep->maxpacket, len);
532 cc = PTD_CC_NOERROR;
533 urbstat = 0;
534 } else {
535 DBG(1, "%s: req %d Data Underrun %s nextpid %02x short_%sok %d/%d/%d byte\n",
536 __func__, ep->num_req,
537 usb_pipein(urb->pipe) ? "IN" : "OUT", ep->nextpid,
538 short_ok ? "" : "not_",
539 PTD_GET_COUNT(ptd), ep->maxpacket, len);
540 /* save the data underrun error code for later and
541 * proceed with the status stage
542 */
543 urb->actual_length += PTD_GET_COUNT(ptd);
544 if (usb_pipecontrol(urb->pipe)) {
545 ep->nextpid = USB_PID_ACK;
546 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
547
548 if (urb->status == -EINPROGRESS)
549 urb->status = cc_to_error[PTD_DATAUNDERRUN];
550 } else {
551 usb_settoggle(udev, ep->epnum, ep->nextpid == USB_PID_OUT,
552 PTD_GET_TOGGLE(ptd));
553 urbstat = cc_to_error[PTD_DATAUNDERRUN];
554 }
555 goto out;
556 }
557 }
558
559 if (cc != PTD_CC_NOERROR) {
560 if (++ep->error_count >= 3 || cc == PTD_CC_STALL || cc == PTD_DATAOVERRUN) {
561 urbstat = cc_to_error[cc];
562 DBG(1, "%s: req %d nextpid %02x, status %d, error %d, error_count %d\n",
563 __func__, ep->num_req, ep->nextpid, urbstat, cc,
564 ep->error_count);
565 }
566 goto out;
567 }
568
569 switch (ep->nextpid) {
570 case USB_PID_OUT:
571 if (PTD_GET_COUNT(ptd) != ep->length)
572 pr_err("%s: count=%d len=%d\n", __func__,
573 PTD_GET_COUNT(ptd), ep->length);
574 BUG_ON(PTD_GET_COUNT(ptd) != ep->length);
575 urb->actual_length += ep->length;
576 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
577 usb_settoggle(udev, ep->epnum, 1, PTD_GET_TOGGLE(ptd));
578 if (urb->actual_length == urb->transfer_buffer_length) {
579 DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
580 ep->num_req, len, ep->maxpacket, urbstat);
581 if (usb_pipecontrol(urb->pipe)) {
582 DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
583 ep->num_req,
584 usb_pipein(urb->pipe) ? "IN" : "OUT");
585 ep->nextpid = USB_PID_ACK;
586 } else {
587 if (len % ep->maxpacket ||
588 !(urb->transfer_flags & URB_ZERO_PACKET)) {
589 urbstat = 0;
590 DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
591 __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
592 urbstat, len, ep->maxpacket, urb->actual_length);
593 }
594 }
595 }
596 break;
597 case USB_PID_IN:
598 len = PTD_GET_COUNT(ptd);
599 BUG_ON(len > ep->length);
600 urb->actual_length += len;
601 BUG_ON(urb->actual_length > urb->transfer_buffer_length);
602 usb_settoggle(udev, ep->epnum, 0, PTD_GET_TOGGLE(ptd));
603 /* if transfer completed or (allowed) data underrun */
604 if ((urb->transfer_buffer_length == urb->actual_length) ||
605 len % ep->maxpacket) {
606 DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
607 ep->num_req, len, ep->maxpacket, urbstat);
608 if (usb_pipecontrol(urb->pipe)) {
609 DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
610 ep->num_req,
611 usb_pipein(urb->pipe) ? "IN" : "OUT");
612 ep->nextpid = USB_PID_ACK;
613 } else {
614 urbstat = 0;
615 DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
616 __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
617 urbstat, len, ep->maxpacket, urb->actual_length);
618 }
619 }
620 break;
621 case USB_PID_SETUP:
622 if (urb->transfer_buffer_length == urb->actual_length) {
623 ep->nextpid = USB_PID_ACK;
624 } else if (usb_pipeout(urb->pipe)) {
625 usb_settoggle(udev, 0, 1, 1);
626 ep->nextpid = USB_PID_OUT;
627 } else {
628 usb_settoggle(udev, 0, 0, 1);
629 ep->nextpid = USB_PID_IN;
630 }
631 break;
632 case USB_PID_ACK:
633 DBG(3, "%s: req %d got ACK %d -> 0\n", __func__, ep->num_req,
634 urbstat);
635 WARN_ON(urbstat != -EINPROGRESS);
636 urbstat = 0;
637 ep->nextpid = 0;
638 break;
639 default:
640 BUG_ON(1);
641 }
642
643 out:
644 if (urbstat != -EINPROGRESS) {
645 DBG(2, "%s: Finishing ep %p req %d urb %p status %d\n", __func__,
646 ep, ep->num_req, urb, urbstat);
647 finish_request(isp1362_hcd, ep, urb, urbstat);
648 }
649}
650
651static void finish_unlinks(struct isp1362_hcd *isp1362_hcd)
652{
653 struct isp1362_ep *ep;
654 struct isp1362_ep *tmp;
655
656 list_for_each_entry_safe(ep, tmp, &isp1362_hcd->remove_list, remove_list) {
657 struct isp1362_ep_queue *epq =
658 get_ptd_queue(isp1362_hcd, ep->ptd_offset);
659 int index = ep->ptd_index;
660
661 BUG_ON(epq == NULL);
662 if (index >= 0) {
663 DBG(1, "%s: remove PTD[%d] $%04x\n", __func__, index, ep->ptd_offset);
664 BUG_ON(ep->num_ptds == 0);
665 release_ptd_buffers(epq, ep);
666 }
667 if (!list_empty(&ep->hep->urb_list)) {
668 struct urb *urb = get_urb(ep);
669
670 DBG(1, "%s: Finishing req %d ep %p from remove_list\n", __func__,
671 ep->num_req, ep);
672 finish_request(isp1362_hcd, ep, urb, -ESHUTDOWN);
673 }
674 WARN_ON(list_empty(&ep->active));
675 if (!list_empty(&ep->active)) {
676 list_del_init(&ep->active);
677 DBG(1, "%s: ep %p removed from active list\n", __func__, ep);
678 }
679 list_del_init(&ep->remove_list);
680 DBG(1, "%s: ep %p removed from remove_list\n", __func__, ep);
681 }
682 DBG(1, "%s: Done\n", __func__);
683}
684
685static inline void enable_atl_transfers(struct isp1362_hcd *isp1362_hcd, int count)
686{
687 if (count > 0) {
688 if (count < isp1362_hcd->atl_queue.ptd_count)
689 isp1362_write_reg16(isp1362_hcd, HCATLDTC, count);
690 isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
691 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, isp1362_hcd->atl_queue.skip_map);
692 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
693 } else
694 isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
695}
696
697static inline void enable_intl_transfers(struct isp1362_hcd *isp1362_hcd)
698{
699 isp1362_enable_int(isp1362_hcd, HCuPINT_INTL);
700 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
701 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, isp1362_hcd->intl_queue.skip_map);
702}
703
704static inline void enable_istl_transfers(struct isp1362_hcd *isp1362_hcd, int flip)
705{
706 isp1362_enable_int(isp1362_hcd, flip ? HCuPINT_ISTL1 : HCuPINT_ISTL0);
707 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, flip ?
708 HCBUFSTAT_ISTL1_FULL : HCBUFSTAT_ISTL0_FULL);
709}
710
711static int submit_req(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
712 struct isp1362_ep *ep, struct isp1362_ep_queue *epq)
713{
714 int index = epq->free_ptd;
715
716 prepare_ptd(isp1362_hcd, urb, ep, epq, 0);
717 index = claim_ptd_buffers(epq, ep, ep->length);
718 if (index == -ENOMEM) {
719 DBG(1, "%s: req %d No free %s PTD available: %d, %08lx:%08lx\n", __func__,
720 ep->num_req, epq->name, ep->num_ptds, epq->buf_map, epq->skip_map);
721 return index;
722 } else if (index == -EOVERFLOW) {
723 DBG(1, "%s: req %d Not enough space for %d byte %s PTD %d %08lx:%08lx\n",
724 __func__, ep->num_req, ep->length, epq->name, ep->num_ptds,
725 epq->buf_map, epq->skip_map);
726 return index;
727 } else
728 BUG_ON(index < 0);
729 list_add_tail(&ep->active, &epq->active);
730 DBG(1, "%s: ep %p req %d len %d added to active list %p\n", __func__,
731 ep, ep->num_req, ep->length, &epq->active);
732 DBG(1, "%s: Submitting %s PTD $%04x for ep %p req %d\n", __func__, epq->name,
733 ep->ptd_offset, ep, ep->num_req);
734 isp1362_write_ptd(isp1362_hcd, ep, epq);
735 __clear_bit(ep->ptd_index, &epq->skip_map);
736
737 return 0;
738}
739
740static void start_atl_transfers(struct isp1362_hcd *isp1362_hcd)
741{
742 int ptd_count = 0;
743 struct isp1362_ep_queue *epq = &isp1362_hcd->atl_queue;
744 struct isp1362_ep *ep;
745 int defer = 0;
746
747 if (atomic_read(&epq->finishing)) {
748 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
749 return;
750 }
751
752 list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
753 struct urb *urb = get_urb(ep);
754 int ret;
755
756 if (!list_empty(&ep->active)) {
757 DBG(2, "%s: Skipping active %s ep %p\n", __func__, epq->name, ep);
758 continue;
759 }
760
761 DBG(1, "%s: Processing %s ep %p req %d\n", __func__, epq->name,
762 ep, ep->num_req);
763
764 ret = submit_req(isp1362_hcd, urb, ep, epq);
765 if (ret == -ENOMEM) {
766 defer = 1;
767 break;
768 } else if (ret == -EOVERFLOW) {
769 defer = 1;
770 continue;
771 }
772#ifdef BUGGY_PXA2XX_UDC_USBTEST
773 defer = ep->nextpid == USB_PID_SETUP;
774#endif
775 ptd_count++;
776 }
777
778 /* Avoid starving of endpoints */
779 if (isp1362_hcd->async.next != isp1362_hcd->async.prev) {
780 DBG(2, "%s: Cycling ASYNC schedule %d\n", __func__, ptd_count);
781 list_move(&isp1362_hcd->async, isp1362_hcd->async.next);
782 }
783 if (ptd_count || defer)
784 enable_atl_transfers(isp1362_hcd, defer ? 0 : ptd_count);
785
786 epq->ptd_count += ptd_count;
787 if (epq->ptd_count > epq->stat_maxptds) {
788 epq->stat_maxptds = epq->ptd_count;
789 DBG(0, "%s: max_ptds: %d\n", __func__, epq->stat_maxptds);
790 }
791}
792
793static void start_intl_transfers(struct isp1362_hcd *isp1362_hcd)
794{
795 int ptd_count = 0;
796 struct isp1362_ep_queue *epq = &isp1362_hcd->intl_queue;
797 struct isp1362_ep *ep;
798
799 if (atomic_read(&epq->finishing)) {
800 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
801 return;
802 }
803
804 list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
805 struct urb *urb = get_urb(ep);
806 int ret;
807
808 if (!list_empty(&ep->active)) {
809 DBG(1, "%s: Skipping active %s ep %p\n", __func__,
810 epq->name, ep);
811 continue;
812 }
813
814 DBG(1, "%s: Processing %s ep %p req %d\n", __func__,
815 epq->name, ep, ep->num_req);
816 ret = submit_req(isp1362_hcd, urb, ep, epq);
817 if (ret == -ENOMEM)
818 break;
819 else if (ret == -EOVERFLOW)
820 continue;
821 ptd_count++;
822 }
823
824 if (ptd_count) {
825 static int last_count;
826
827 if (ptd_count != last_count) {
828 DBG(0, "%s: ptd_count: %d\n", __func__, ptd_count);
829 last_count = ptd_count;
830 }
831 enable_intl_transfers(isp1362_hcd);
832 }
833
834 epq->ptd_count += ptd_count;
835 if (epq->ptd_count > epq->stat_maxptds)
836 epq->stat_maxptds = epq->ptd_count;
837}
838
839static inline int next_ptd(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
840{
841 u16 ptd_offset = ep->ptd_offset;
842 int num_ptds = (ep->length + PTD_HEADER_SIZE + (epq->blk_size - 1)) / epq->blk_size;
843
844 DBG(2, "%s: PTD offset $%04x + %04x => %d * %04x -> $%04x\n", __func__, ptd_offset,
845 ep->length, num_ptds, epq->blk_size, ptd_offset + num_ptds * epq->blk_size);
846
847 ptd_offset += num_ptds * epq->blk_size;
848 if (ptd_offset < epq->buf_start + epq->buf_size)
849 return ptd_offset;
850 else
851 return -ENOMEM;
852}
853
854static void start_iso_transfers(struct isp1362_hcd *isp1362_hcd)
855{
856 int ptd_count = 0;
857 int flip = isp1362_hcd->istl_flip;
858 struct isp1362_ep_queue *epq;
859 int ptd_offset;
860 struct isp1362_ep *ep;
861 struct isp1362_ep *tmp;
862 u16 fno = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
863
864 fill2:
865 epq = &isp1362_hcd->istl_queue[flip];
866 if (atomic_read(&epq->finishing)) {
867 DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
868 return;
869 }
870
871 if (!list_empty(&epq->active))
872 return;
873
874 ptd_offset = epq->buf_start;
875 list_for_each_entry_safe(ep, tmp, &isp1362_hcd->isoc, schedule) {
876 struct urb *urb = get_urb(ep);
877 s16 diff = fno - (u16)urb->start_frame;
878
879 DBG(1, "%s: Processing %s ep %p\n", __func__, epq->name, ep);
880
881 if (diff > urb->number_of_packets) {
882 /* time frame for this URB has elapsed */
883 finish_request(isp1362_hcd, ep, urb, -EOVERFLOW);
884 continue;
885 } else if (diff < -1) {
886 /* URB is not due in this frame or the next one.
887 * Comparing with '-1' instead of '0' accounts for double
888 * buffering in the ISP1362 which enables us to queue the PTD
889 * one frame ahead of time
890 */
891 } else if (diff == -1) {
892 /* submit PTD's that are due in the next frame */
893 prepare_ptd(isp1362_hcd, urb, ep, epq, fno);
894 if (ptd_offset + PTD_HEADER_SIZE + ep->length >
895 epq->buf_start + epq->buf_size) {
896 pr_err("%s: Not enough ISO buffer space for %d byte PTD\n",
897 __func__, ep->length);
898 continue;
899 }
900 ep->ptd_offset = ptd_offset;
901 list_add_tail(&ep->active, &epq->active);
902
903 ptd_offset = next_ptd(epq, ep);
904 if (ptd_offset < 0) {
905 pr_warning("%s: req %d No more %s PTD buffers available\n", __func__,
906 ep->num_req, epq->name);
907 break;
908 }
909 }
910 }
911 list_for_each_entry(ep, &epq->active, active) {
912 if (epq->active.next == &ep->active)
913 ep->ptd.mps |= PTD_LAST_MSK;
914 isp1362_write_ptd(isp1362_hcd, ep, epq);
915 ptd_count++;
916 }
917
918 if (ptd_count)
919 enable_istl_transfers(isp1362_hcd, flip);
920
921 epq->ptd_count += ptd_count;
922 if (epq->ptd_count > epq->stat_maxptds)
923 epq->stat_maxptds = epq->ptd_count;
924
925 /* check, whether the second ISTL buffer may also be filled */
926 if (!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
927 (flip ? HCBUFSTAT_ISTL0_FULL : HCBUFSTAT_ISTL1_FULL))) {
928 fno++;
929 ptd_count = 0;
930 flip = 1 - flip;
931 goto fill2;
932 }
933}
934
935static void finish_transfers(struct isp1362_hcd *isp1362_hcd, unsigned long done_map,
936 struct isp1362_ep_queue *epq)
937{
938 struct isp1362_ep *ep;
939 struct isp1362_ep *tmp;
940
941 if (list_empty(&epq->active)) {
942 DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
943 return;
944 }
945
946 DBG(1, "%s: Finishing %s transfers %08lx\n", __func__, epq->name, done_map);
947
948 atomic_inc(&epq->finishing);
949 list_for_each_entry_safe(ep, tmp, &epq->active, active) {
950 int index = ep->ptd_index;
951
952 DBG(1, "%s: Checking %s PTD[%02x] $%04x\n", __func__, epq->name,
953 index, ep->ptd_offset);
954
955 BUG_ON(index < 0);
956 if (__test_and_clear_bit(index, &done_map)) {
957 isp1362_read_ptd(isp1362_hcd, ep, epq);
958 epq->free_ptd = index;
959 BUG_ON(ep->num_ptds == 0);
960 release_ptd_buffers(epq, ep);
961
962 DBG(1, "%s: ep %p req %d removed from active list\n", __func__,
963 ep, ep->num_req);
964 if (!list_empty(&ep->remove_list)) {
965 list_del_init(&ep->remove_list);
966 DBG(1, "%s: ep %p removed from remove list\n", __func__, ep);
967 }
968 DBG(1, "%s: Postprocessing %s ep %p req %d\n", __func__, epq->name,
969 ep, ep->num_req);
970 postproc_ep(isp1362_hcd, ep);
971 }
972 if (!done_map)
973 break;
974 }
975 if (done_map)
976 pr_warning("%s: done_map not clear: %08lx:%08lx\n", __func__, done_map,
977 epq->skip_map);
978 atomic_dec(&epq->finishing);
979}
980
981static void finish_iso_transfers(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep_queue *epq)
982{
983 struct isp1362_ep *ep;
984 struct isp1362_ep *tmp;
985
986 if (list_empty(&epq->active)) {
987 DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
988 return;
989 }
990
991 DBG(1, "%s: Finishing %s transfers\n", __func__, epq->name);
992
993 atomic_inc(&epq->finishing);
994 list_for_each_entry_safe(ep, tmp, &epq->active, active) {
995 DBG(1, "%s: Checking PTD $%04x\n", __func__, ep->ptd_offset);
996
997 isp1362_read_ptd(isp1362_hcd, ep, epq);
998 DBG(1, "%s: Postprocessing %s ep %p\n", __func__, epq->name, ep);
999 postproc_ep(isp1362_hcd, ep);
1000 }
1001 WARN_ON(epq->blk_size != 0);
1002 atomic_dec(&epq->finishing);
1003}
1004
1005static irqreturn_t isp1362_irq(struct usb_hcd *hcd)
1006{
1007 int handled = 0;
1008 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1009 u16 irqstat;
1010 u16 svc_mask;
1011
1012 spin_lock(&isp1362_hcd->lock);
1013
1014 BUG_ON(isp1362_hcd->irq_active++);
1015
1016 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1017
1018 irqstat = isp1362_read_reg16(isp1362_hcd, HCuPINT);
1019 DBG(3, "%s: got IRQ %04x:%04x\n", __func__, irqstat, isp1362_hcd->irqenb);
1020
1021 /* only handle interrupts that are currently enabled */
1022 irqstat &= isp1362_hcd->irqenb;
1023 isp1362_write_reg16(isp1362_hcd, HCuPINT, irqstat);
1024 svc_mask = irqstat;
1025
1026 if (irqstat & HCuPINT_SOF) {
1027 isp1362_hcd->irqenb &= ~HCuPINT_SOF;
1028 isp1362_hcd->irq_stat[ISP1362_INT_SOF]++;
1029 handled = 1;
1030 svc_mask &= ~HCuPINT_SOF;
1031 DBG(3, "%s: SOF\n", __func__);
1032 isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1033 if (!list_empty(&isp1362_hcd->remove_list))
1034 finish_unlinks(isp1362_hcd);
1035 if (!list_empty(&isp1362_hcd->async) && !(irqstat & HCuPINT_ATL)) {
1036 if (list_empty(&isp1362_hcd->atl_queue.active)) {
1037 start_atl_transfers(isp1362_hcd);
1038 } else {
1039 isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
1040 isp1362_write_reg32(isp1362_hcd, HCATLSKIP,
1041 isp1362_hcd->atl_queue.skip_map);
1042 isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1043 }
1044 }
1045 }
1046
1047 if (irqstat & HCuPINT_ISTL0) {
1048 isp1362_hcd->irq_stat[ISP1362_INT_ISTL0]++;
1049 handled = 1;
1050 svc_mask &= ~HCuPINT_ISTL0;
1051 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL0_FULL);
1052 DBG(1, "%s: ISTL0\n", __func__);
1053 WARN_ON((int)!!isp1362_hcd->istl_flip);
1054 WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1055 HCBUFSTAT_ISTL0_ACTIVE);
1056 WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1057 HCBUFSTAT_ISTL0_DONE));
1058 isp1362_hcd->irqenb &= ~HCuPINT_ISTL0;
1059 }
1060
1061 if (irqstat & HCuPINT_ISTL1) {
1062 isp1362_hcd->irq_stat[ISP1362_INT_ISTL1]++;
1063 handled = 1;
1064 svc_mask &= ~HCuPINT_ISTL1;
1065 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL1_FULL);
1066 DBG(1, "%s: ISTL1\n", __func__);
1067 WARN_ON(!(int)isp1362_hcd->istl_flip);
1068 WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1069 HCBUFSTAT_ISTL1_ACTIVE);
1070 WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1071 HCBUFSTAT_ISTL1_DONE));
1072 isp1362_hcd->irqenb &= ~HCuPINT_ISTL1;
1073 }
1074
1075 if (irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) {
1076 WARN_ON((irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) ==
1077 (HCuPINT_ISTL0 | HCuPINT_ISTL1));
1078 finish_iso_transfers(isp1362_hcd,
1079 &isp1362_hcd->istl_queue[isp1362_hcd->istl_flip]);
1080 start_iso_transfers(isp1362_hcd);
1081 isp1362_hcd->istl_flip = 1 - isp1362_hcd->istl_flip;
1082 }
1083
1084 if (irqstat & HCuPINT_INTL) {
1085 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1086 u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCINTLSKIP);
1087 isp1362_hcd->irq_stat[ISP1362_INT_INTL]++;
1088
1089 DBG(2, "%s: INTL\n", __func__);
1090
1091 svc_mask &= ~HCuPINT_INTL;
1092
1093 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, skip_map | done_map);
1094 if (~(done_map | skip_map) == 0)
1095 /* All PTDs are finished, disable INTL processing entirely */
1096 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
1097
1098 handled = 1;
1099 WARN_ON(!done_map);
1100 if (done_map) {
1101 DBG(3, "%s: INTL done_map %08x\n", __func__, done_map);
1102 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1103 start_intl_transfers(isp1362_hcd);
1104 }
1105 }
1106
1107 if (irqstat & HCuPINT_ATL) {
1108 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1109 u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCATLSKIP);
1110 isp1362_hcd->irq_stat[ISP1362_INT_ATL]++;
1111
1112 DBG(2, "%s: ATL\n", __func__);
1113
1114 svc_mask &= ~HCuPINT_ATL;
1115
1116 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, skip_map | done_map);
1117 if (~(done_map | skip_map) == 0)
1118 isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1119 if (done_map) {
1120 DBG(3, "%s: ATL done_map %08x\n", __func__, done_map);
1121 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1122 start_atl_transfers(isp1362_hcd);
1123 }
1124 handled = 1;
1125 }
1126
1127 if (irqstat & HCuPINT_OPR) {
1128 u32 intstat = isp1362_read_reg32(isp1362_hcd, HCINTSTAT);
1129 isp1362_hcd->irq_stat[ISP1362_INT_OPR]++;
1130
1131 svc_mask &= ~HCuPINT_OPR;
1132 DBG(2, "%s: OPR %08x:%08x\n", __func__, intstat, isp1362_hcd->intenb);
1133 intstat &= isp1362_hcd->intenb;
1134 if (intstat & OHCI_INTR_UE) {
1135 pr_err("Unrecoverable error\n");
1136 /* FIXME: do here reset or cleanup or whatever */
1137 }
1138 if (intstat & OHCI_INTR_RHSC) {
1139 isp1362_hcd->rhstatus = isp1362_read_reg32(isp1362_hcd, HCRHSTATUS);
1140 isp1362_hcd->rhport[0] = isp1362_read_reg32(isp1362_hcd, HCRHPORT1);
1141 isp1362_hcd->rhport[1] = isp1362_read_reg32(isp1362_hcd, HCRHPORT2);
1142 }
1143 if (intstat & OHCI_INTR_RD) {
1144 pr_info("%s: RESUME DETECTED\n", __func__);
1145 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1146 usb_hcd_resume_root_hub(hcd);
1147 }
1148 isp1362_write_reg32(isp1362_hcd, HCINTSTAT, intstat);
1149 irqstat &= ~HCuPINT_OPR;
1150 handled = 1;
1151 }
1152
1153 if (irqstat & HCuPINT_SUSP) {
1154 isp1362_hcd->irq_stat[ISP1362_INT_SUSP]++;
1155 handled = 1;
1156 svc_mask &= ~HCuPINT_SUSP;
1157
1158 pr_info("%s: SUSPEND IRQ\n", __func__);
1159 }
1160
1161 if (irqstat & HCuPINT_CLKRDY) {
1162 isp1362_hcd->irq_stat[ISP1362_INT_CLKRDY]++;
1163 handled = 1;
1164 isp1362_hcd->irqenb &= ~HCuPINT_CLKRDY;
1165 svc_mask &= ~HCuPINT_CLKRDY;
1166 pr_info("%s: CLKRDY IRQ\n", __func__);
1167 }
1168
1169 if (svc_mask)
1170 pr_err("%s: Unserviced interrupt(s) %04x\n", __func__, svc_mask);
1171
1172 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
1173 isp1362_hcd->irq_active--;
1174 spin_unlock(&isp1362_hcd->lock);
1175
1176 return IRQ_RETVAL(handled);
1177}
1178
1179/*-------------------------------------------------------------------------*/
1180
1181#define MAX_PERIODIC_LOAD 900 /* out of 1000 usec */
1182static int balance(struct isp1362_hcd *isp1362_hcd, u16 interval, u16 load)
1183{
1184 int i, branch = -ENOSPC;
1185
1186 /* search for the least loaded schedule branch of that interval
1187 * which has enough bandwidth left unreserved.
1188 */
1189 for (i = 0; i < interval; i++) {
1190 if (branch < 0 || isp1362_hcd->load[branch] > isp1362_hcd->load[i]) {
1191 int j;
1192
1193 for (j = i; j < PERIODIC_SIZE; j += interval) {
1194 if ((isp1362_hcd->load[j] + load) > MAX_PERIODIC_LOAD) {
1195 pr_err("%s: new load %d load[%02x] %d max %d\n", __func__,
1196 load, j, isp1362_hcd->load[j], MAX_PERIODIC_LOAD);
1197 break;
1198 }
1199 }
1200 if (j < PERIODIC_SIZE)
1201 continue;
1202 branch = i;
1203 }
1204 }
1205 return branch;
1206}
1207
1208/* NB! ALL the code above this point runs with isp1362_hcd->lock
1209 held, irqs off
1210*/
1211
1212/*-------------------------------------------------------------------------*/
1213
1214static int isp1362_urb_enqueue(struct usb_hcd *hcd,
1215 struct urb *urb,
1216 gfp_t mem_flags)
1217{
1218 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1219 struct usb_device *udev = urb->dev;
1220 unsigned int pipe = urb->pipe;
1221 int is_out = !usb_pipein(pipe);
1222 int type = usb_pipetype(pipe);
1223 int epnum = usb_pipeendpoint(pipe);
1224 struct usb_host_endpoint *hep = urb->ep;
1225 struct isp1362_ep *ep = NULL;
1226 unsigned long flags;
1227 int retval = 0;
1228
1229 DBG(3, "%s: urb %p\n", __func__, urb);
1230
1231 if (type == PIPE_ISOCHRONOUS) {
1232 pr_err("Isochronous transfers not supported\n");
1233 return -ENOSPC;
1234 }
1235
1236 URB_DBG("%s: FA %d ep%d%s %s: len %d %s%s\n", __func__,
1237 usb_pipedevice(pipe), epnum,
1238 is_out ? "out" : "in",
1239 usb_pipecontrol(pipe) ? "ctrl" :
1240 usb_pipeint(pipe) ? "int" :
1241 usb_pipebulk(pipe) ? "bulk" :
1242 "iso",
1243 urb->transfer_buffer_length,
1244 (urb->transfer_flags & URB_ZERO_PACKET) ? "ZERO_PACKET " : "",
1245 !(urb->transfer_flags & URB_SHORT_NOT_OK) ?
1246 "short_ok" : "");
1247
1248 /* avoid all allocations within spinlocks: request or endpoint */
1249 if (!hep->hcpriv) {
1250 ep = kzalloc(sizeof *ep, mem_flags);
1251 if (!ep)
1252 return -ENOMEM;
1253 }
1254 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1255
1256 /* don't submit to a dead or disabled port */
1257 if (!((isp1362_hcd->rhport[0] | isp1362_hcd->rhport[1]) &
1258 USB_PORT_STAT_ENABLE) ||
1259 !HC_IS_RUNNING(hcd->state)) {
1260 kfree(ep);
1261 retval = -ENODEV;
1262 goto fail_not_linked;
1263 }
1264
1265 retval = usb_hcd_link_urb_to_ep(hcd, urb);
1266 if (retval) {
1267 kfree(ep);
1268 goto fail_not_linked;
1269 }
1270
1271 if (hep->hcpriv) {
1272 ep = hep->hcpriv;
1273 } else {
1274 INIT_LIST_HEAD(&ep->schedule);
1275 INIT_LIST_HEAD(&ep->active);
1276 INIT_LIST_HEAD(&ep->remove_list);
1277 ep->udev = usb_get_dev(udev);
1278 ep->hep = hep;
1279 ep->epnum = epnum;
1280 ep->maxpacket = usb_maxpacket(udev, urb->pipe, is_out);
1281 ep->ptd_offset = -EINVAL;
1282 ep->ptd_index = -EINVAL;
1283 usb_settoggle(udev, epnum, is_out, 0);
1284
1285 if (type == PIPE_CONTROL)
1286 ep->nextpid = USB_PID_SETUP;
1287 else if (is_out)
1288 ep->nextpid = USB_PID_OUT;
1289 else
1290 ep->nextpid = USB_PID_IN;
1291
1292 switch (type) {
1293 case PIPE_ISOCHRONOUS:
1294 case PIPE_INTERRUPT:
1295 if (urb->interval > PERIODIC_SIZE)
1296 urb->interval = PERIODIC_SIZE;
1297 ep->interval = urb->interval;
1298 ep->branch = PERIODIC_SIZE;
1299 ep->load = usb_calc_bus_time(udev->speed, !is_out,
1300 (type == PIPE_ISOCHRONOUS),
1301 usb_maxpacket(udev, pipe, is_out)) / 1000;
1302 break;
1303 }
1304 hep->hcpriv = ep;
1305 }
1306 ep->num_req = isp1362_hcd->req_serial++;
1307
1308 /* maybe put endpoint into schedule */
1309 switch (type) {
1310 case PIPE_CONTROL:
1311 case PIPE_BULK:
1312 if (list_empty(&ep->schedule)) {
1313 DBG(1, "%s: Adding ep %p req %d to async schedule\n",
1314 __func__, ep, ep->num_req);
1315 list_add_tail(&ep->schedule, &isp1362_hcd->async);
1316 }
1317 break;
1318 case PIPE_ISOCHRONOUS:
1319 case PIPE_INTERRUPT:
1320 urb->interval = ep->interval;
1321
1322 /* urb submitted for already existing EP */
1323 if (ep->branch < PERIODIC_SIZE)
1324 break;
1325
1326 retval = balance(isp1362_hcd, ep->interval, ep->load);
1327 if (retval < 0) {
1328 pr_err("%s: balance returned %d\n", __func__, retval);
1329 goto fail;
1330 }
1331 ep->branch = retval;
1332 retval = 0;
1333 isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1334 DBG(1, "%s: Current frame %04x branch %02x start_frame %04x(%04x)\n",
1335 __func__, isp1362_hcd->fmindex, ep->branch,
1336 ((isp1362_hcd->fmindex + PERIODIC_SIZE - 1) &
1337 ~(PERIODIC_SIZE - 1)) + ep->branch,
1338 (isp1362_hcd->fmindex & (PERIODIC_SIZE - 1)) + ep->branch);
1339
1340 if (list_empty(&ep->schedule)) {
1341 if (type == PIPE_ISOCHRONOUS) {
1342 u16 frame = isp1362_hcd->fmindex;
1343
1344 frame += max_t(u16, 8, ep->interval);
1345 frame &= ~(ep->interval - 1);
1346 frame |= ep->branch;
1347 if (frame_before(frame, isp1362_hcd->fmindex))
1348 frame += ep->interval;
1349 urb->start_frame = frame;
1350
1351 DBG(1, "%s: Adding ep %p to isoc schedule\n", __func__, ep);
1352 list_add_tail(&ep->schedule, &isp1362_hcd->isoc);
1353 } else {
1354 DBG(1, "%s: Adding ep %p to periodic schedule\n", __func__, ep);
1355 list_add_tail(&ep->schedule, &isp1362_hcd->periodic);
1356 }
1357 } else
1358 DBG(1, "%s: ep %p already scheduled\n", __func__, ep);
1359
1360 DBG(2, "%s: load %d bandwidth %d -> %d\n", __func__,
1361 ep->load / ep->interval, isp1362_hcd->load[ep->branch],
1362 isp1362_hcd->load[ep->branch] + ep->load);
1363 isp1362_hcd->load[ep->branch] += ep->load;
1364 }
1365
1366 urb->hcpriv = hep;
1367 ALIGNSTAT(isp1362_hcd, urb->transfer_buffer);
1368
1369 switch (type) {
1370 case PIPE_CONTROL:
1371 case PIPE_BULK:
1372 start_atl_transfers(isp1362_hcd);
1373 break;
1374 case PIPE_INTERRUPT:
1375 start_intl_transfers(isp1362_hcd);
1376 break;
1377 case PIPE_ISOCHRONOUS:
1378 start_iso_transfers(isp1362_hcd);
1379 break;
1380 default:
1381 BUG();
1382 }
1383 fail:
1384 if (retval)
1385 usb_hcd_unlink_urb_from_ep(hcd, urb);
1386
1387
1388 fail_not_linked:
1389 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1390 if (retval)
1391 DBG(0, "%s: urb %p failed with %d\n", __func__, urb, retval);
1392 return retval;
1393}
1394
1395static int isp1362_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status)
1396{
1397 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1398 struct usb_host_endpoint *hep;
1399 unsigned long flags;
1400 struct isp1362_ep *ep;
1401 int retval = 0;
1402
1403 DBG(3, "%s: urb %p\n", __func__, urb);
1404
1405 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1406 retval = usb_hcd_check_unlink_urb(hcd, urb, status);
1407 if (retval)
1408 goto done;
1409
1410 hep = urb->hcpriv;
1411
1412 if (!hep) {
1413 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1414 return -EIDRM;
1415 }
1416
1417 ep = hep->hcpriv;
1418 if (ep) {
1419 /* In front of queue? */
1420 if (ep->hep->urb_list.next == &urb->urb_list) {
1421 if (!list_empty(&ep->active)) {
1422 DBG(1, "%s: urb %p ep %p req %d active PTD[%d] $%04x\n", __func__,
1423 urb, ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1424 /* disable processing and queue PTD for removal */
1425 remove_ptd(isp1362_hcd, ep);
1426 urb = NULL;
1427 }
1428 }
1429 if (urb) {
1430 DBG(1, "%s: Finishing ep %p req %d\n", __func__, ep,
1431 ep->num_req);
1432 finish_request(isp1362_hcd, ep, urb, status);
1433 } else
1434 DBG(1, "%s: urb %p active; wait4irq\n", __func__, urb);
1435 } else {
1436 pr_warning("%s: No EP in URB %p\n", __func__, urb);
1437 retval = -EINVAL;
1438 }
1439done:
1440 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1441
1442 DBG(3, "%s: exit\n", __func__);
1443
1444 return retval;
1445}
1446
1447static void isp1362_endpoint_disable(struct usb_hcd *hcd, struct usb_host_endpoint *hep)
1448{
1449 struct isp1362_ep *ep = hep->hcpriv;
1450 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1451 unsigned long flags;
1452
1453 DBG(1, "%s: ep %p\n", __func__, ep);
1454 if (!ep)
1455 return;
1456 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1457 if (!list_empty(&hep->urb_list)) {
1458 if (!list_empty(&ep->active) && list_empty(&ep->remove_list)) {
1459 DBG(1, "%s: Removing ep %p req %d PTD[%d] $%04x\n", __func__,
1460 ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1461 remove_ptd(isp1362_hcd, ep);
1462 pr_info("%s: Waiting for Interrupt to clean up\n", __func__);
1463 }
1464 }
1465 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1466 /* Wait for interrupt to clear out active list */
1467 while (!list_empty(&ep->active))
1468 msleep(1);
1469
1470 DBG(1, "%s: Freeing EP %p\n", __func__, ep);
1471
1472 usb_put_dev(ep->udev);
1473 kfree(ep);
1474 hep->hcpriv = NULL;
1475}
1476
1477static int isp1362_get_frame(struct usb_hcd *hcd)
1478{
1479 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1480 u32 fmnum;
1481 unsigned long flags;
1482
1483 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1484 fmnum = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1485 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1486
1487 return (int)fmnum;
1488}
1489
1490/*-------------------------------------------------------------------------*/
1491
1492/* Adapted from ohci-hub.c */
1493static int isp1362_hub_status_data(struct usb_hcd *hcd, char *buf)
1494{
1495 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1496 int ports, i, changed = 0;
1497 unsigned long flags;
1498
1499 if (!HC_IS_RUNNING(hcd->state))
1500 return -ESHUTDOWN;
1501
1502 /* Report no status change now, if we are scheduled to be
1503 called later */
1504 if (timer_pending(&hcd->rh_timer))
1505 return 0;
1506
1507 ports = isp1362_hcd->rhdesca & RH_A_NDP;
1508 BUG_ON(ports > 2);
1509
1510 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1511 /* init status */
1512 if (isp1362_hcd->rhstatus & (RH_HS_LPSC | RH_HS_OCIC))
1513 buf[0] = changed = 1;
1514 else
1515 buf[0] = 0;
1516
1517 for (i = 0; i < ports; i++) {
1518 u32 status = isp1362_hcd->rhport[i];
1519
1520 if (status & (RH_PS_CSC | RH_PS_PESC | RH_PS_PSSC |
1521 RH_PS_OCIC | RH_PS_PRSC)) {
1522 changed = 1;
1523 buf[0] |= 1 << (i + 1);
1524 continue;
1525 }
1526
1527 if (!(status & RH_PS_CCS))
1528 continue;
1529 }
1530 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1531 return changed;
1532}
1533
1534static void isp1362_hub_descriptor(struct isp1362_hcd *isp1362_hcd,
1535 struct usb_hub_descriptor *desc)
1536{
1537 u32 reg = isp1362_hcd->rhdesca;
1538
1539 DBG(3, "%s: enter\n", __func__);
1540
1541 desc->bDescriptorType = 0x29;
1542 desc->bDescLength = 9;
1543 desc->bHubContrCurrent = 0;
1544 desc->bNbrPorts = reg & 0x3;
1545 /* Power switching, device type, overcurrent. */
1546 desc->wHubCharacteristics = cpu_to_le16((reg >> 8) & 0x1f);
1547 DBG(0, "%s: hubcharacteristics = %02x\n", __func__, cpu_to_le16((reg >> 8) & 0x1f));
1548 desc->bPwrOn2PwrGood = (reg >> 24) & 0xff;
1549 /* ports removable, and legacy PortPwrCtrlMask */
1550 desc->u.hs.DeviceRemovable[0] = desc->bNbrPorts == 1 ? 1 << 1 : 3 << 1;
1551 desc->u.hs.DeviceRemovable[1] = ~0;
1552
1553 DBG(3, "%s: exit\n", __func__);
1554}
1555
1556/* Adapted from ohci-hub.c */
1557static int isp1362_hub_control(struct usb_hcd *hcd, u16 typeReq, u16 wValue,
1558 u16 wIndex, char *buf, u16 wLength)
1559{
1560 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1561 int retval = 0;
1562 unsigned long flags;
1563 unsigned long t1;
1564 int ports = isp1362_hcd->rhdesca & RH_A_NDP;
1565 u32 tmp = 0;
1566
1567 switch (typeReq) {
1568 case ClearHubFeature:
1569 DBG(0, "ClearHubFeature: ");
1570 switch (wValue) {
1571 case C_HUB_OVER_CURRENT:
1572 DBG(0, "C_HUB_OVER_CURRENT\n");
1573 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1574 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_OCIC);
1575 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1576 case C_HUB_LOCAL_POWER:
1577 DBG(0, "C_HUB_LOCAL_POWER\n");
1578 break;
1579 default:
1580 goto error;
1581 }
1582 break;
1583 case SetHubFeature:
1584 DBG(0, "SetHubFeature: ");
1585 switch (wValue) {
1586 case C_HUB_OVER_CURRENT:
1587 case C_HUB_LOCAL_POWER:
1588 DBG(0, "C_HUB_OVER_CURRENT or C_HUB_LOCAL_POWER\n");
1589 break;
1590 default:
1591 goto error;
1592 }
1593 break;
1594 case GetHubDescriptor:
1595 DBG(0, "GetHubDescriptor\n");
1596 isp1362_hub_descriptor(isp1362_hcd, (struct usb_hub_descriptor *)buf);
1597 break;
1598 case GetHubStatus:
1599 DBG(0, "GetHubStatus\n");
1600 put_unaligned(cpu_to_le32(0), (__le32 *) buf);
1601 break;
1602 case GetPortStatus:
1603#ifndef VERBOSE
1604 DBG(0, "GetPortStatus\n");
1605#endif
1606 if (!wIndex || wIndex > ports)
1607 goto error;
1608 tmp = isp1362_hcd->rhport[--wIndex];
1609 put_unaligned(cpu_to_le32(tmp), (__le32 *) buf);
1610 break;
1611 case ClearPortFeature:
1612 DBG(0, "ClearPortFeature: ");
1613 if (!wIndex || wIndex > ports)
1614 goto error;
1615 wIndex--;
1616
1617 switch (wValue) {
1618 case USB_PORT_FEAT_ENABLE:
1619 DBG(0, "USB_PORT_FEAT_ENABLE\n");
1620 tmp = RH_PS_CCS;
1621 break;
1622 case USB_PORT_FEAT_C_ENABLE:
1623 DBG(0, "USB_PORT_FEAT_C_ENABLE\n");
1624 tmp = RH_PS_PESC;
1625 break;
1626 case USB_PORT_FEAT_SUSPEND:
1627 DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1628 tmp = RH_PS_POCI;
1629 break;
1630 case USB_PORT_FEAT_C_SUSPEND:
1631 DBG(0, "USB_PORT_FEAT_C_SUSPEND\n");
1632 tmp = RH_PS_PSSC;
1633 break;
1634 case USB_PORT_FEAT_POWER:
1635 DBG(0, "USB_PORT_FEAT_POWER\n");
1636 tmp = RH_PS_LSDA;
1637
1638 break;
1639 case USB_PORT_FEAT_C_CONNECTION:
1640 DBG(0, "USB_PORT_FEAT_C_CONNECTION\n");
1641 tmp = RH_PS_CSC;
1642 break;
1643 case USB_PORT_FEAT_C_OVER_CURRENT:
1644 DBG(0, "USB_PORT_FEAT_C_OVER_CURRENT\n");
1645 tmp = RH_PS_OCIC;
1646 break;
1647 case USB_PORT_FEAT_C_RESET:
1648 DBG(0, "USB_PORT_FEAT_C_RESET\n");
1649 tmp = RH_PS_PRSC;
1650 break;
1651 default:
1652 goto error;
1653 }
1654
1655 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1656 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, tmp);
1657 isp1362_hcd->rhport[wIndex] =
1658 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1659 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1660 break;
1661 case SetPortFeature:
1662 DBG(0, "SetPortFeature: ");
1663 if (!wIndex || wIndex > ports)
1664 goto error;
1665 wIndex--;
1666 switch (wValue) {
1667 case USB_PORT_FEAT_SUSPEND:
1668 DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1669 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1670 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PSS);
1671 isp1362_hcd->rhport[wIndex] =
1672 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1673 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1674 break;
1675 case USB_PORT_FEAT_POWER:
1676 DBG(0, "USB_PORT_FEAT_POWER\n");
1677 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1678 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PPS);
1679 isp1362_hcd->rhport[wIndex] =
1680 isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1681 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1682 break;
1683 case USB_PORT_FEAT_RESET:
1684 DBG(0, "USB_PORT_FEAT_RESET\n");
1685 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1686
1687 t1 = jiffies + msecs_to_jiffies(USB_RESET_WIDTH);
1688 while (time_before(jiffies, t1)) {
1689 /* spin until any current reset finishes */
1690 for (;;) {
1691 tmp = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1692 if (!(tmp & RH_PS_PRS))
1693 break;
1694 udelay(500);
1695 }
1696 if (!(tmp & RH_PS_CCS))
1697 break;
1698 /* Reset lasts 10ms (claims datasheet) */
1699 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, (RH_PS_PRS));
1700
1701 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1702 msleep(10);
1703 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1704 }
1705
1706 isp1362_hcd->rhport[wIndex] = isp1362_read_reg32(isp1362_hcd,
1707 HCRHPORT1 + wIndex);
1708 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1709 break;
1710 default:
1711 goto error;
1712 }
1713 break;
1714
1715 default:
1716 error:
1717 /* "protocol stall" on error */
1718 DBG(0, "PROTOCOL STALL\n");
1719 retval = -EPIPE;
1720 }
1721
1722 return retval;
1723}
1724
1725#ifdef CONFIG_PM
1726static int isp1362_bus_suspend(struct usb_hcd *hcd)
1727{
1728 int status = 0;
1729 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1730 unsigned long flags;
1731
1732 if (time_before(jiffies, isp1362_hcd->next_statechange))
1733 msleep(5);
1734
1735 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1736
1737 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1738 switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1739 case OHCI_USB_RESUME:
1740 DBG(0, "%s: resume/suspend?\n", __func__);
1741 isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1742 isp1362_hcd->hc_control |= OHCI_USB_RESET;
1743 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1744 /* FALL THROUGH */
1745 case OHCI_USB_RESET:
1746 status = -EBUSY;
1747 pr_warning("%s: needs reinit!\n", __func__);
1748 goto done;
1749 case OHCI_USB_SUSPEND:
1750 pr_warning("%s: already suspended?\n", __func__);
1751 goto done;
1752 }
1753 DBG(0, "%s: suspend root hub\n", __func__);
1754
1755 /* First stop any processing */
1756 hcd->state = HC_STATE_QUIESCING;
1757 if (!list_empty(&isp1362_hcd->atl_queue.active) ||
1758 !list_empty(&isp1362_hcd->intl_queue.active) ||
1759 !list_empty(&isp1362_hcd->istl_queue[0] .active) ||
1760 !list_empty(&isp1362_hcd->istl_queue[1] .active)) {
1761 int limit;
1762
1763 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
1764 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
1765 isp1362_write_reg16(isp1362_hcd, HCBUFSTAT, 0);
1766 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1767 isp1362_write_reg32(isp1362_hcd, HCINTSTAT, OHCI_INTR_SF);
1768
1769 DBG(0, "%s: stopping schedules ...\n", __func__);
1770 limit = 2000;
1771 while (limit > 0) {
1772 udelay(250);
1773 limit -= 250;
1774 if (isp1362_read_reg32(isp1362_hcd, HCINTSTAT) & OHCI_INTR_SF)
1775 break;
1776 }
1777 mdelay(7);
1778 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ATL) {
1779 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1780 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1781 }
1782 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_INTL) {
1783 u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1784 finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1785 }
1786 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL0)
1787 finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[0]);
1788 if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL1)
1789 finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[1]);
1790 }
1791 DBG(0, "%s: HCINTSTAT: %08x\n", __func__,
1792 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1793 isp1362_write_reg32(isp1362_hcd, HCINTSTAT,
1794 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1795
1796 /* Suspend hub */
1797 isp1362_hcd->hc_control = OHCI_USB_SUSPEND;
1798 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1799 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1800 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1801
1802#if 1
1803 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1804 if ((isp1362_hcd->hc_control & OHCI_CTRL_HCFS) != OHCI_USB_SUSPEND) {
1805 pr_err("%s: controller won't suspend %08x\n", __func__,
1806 isp1362_hcd->hc_control);
1807 status = -EBUSY;
1808 } else
1809#endif
1810 {
1811 /* no resumes until devices finish suspending */
1812 isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(5);
1813 }
1814done:
1815 if (status == 0) {
1816 hcd->state = HC_STATE_SUSPENDED;
1817 DBG(0, "%s: HCD suspended: %08x\n", __func__,
1818 isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1819 }
1820 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1821 return status;
1822}
1823
1824static int isp1362_bus_resume(struct usb_hcd *hcd)
1825{
1826 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1827 u32 port;
1828 unsigned long flags;
1829 int status = -EINPROGRESS;
1830
1831 if (time_before(jiffies, isp1362_hcd->next_statechange))
1832 msleep(5);
1833
1834 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1835 isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1836 pr_info("%s: HCCONTROL: %08x\n", __func__, isp1362_hcd->hc_control);
1837 if (hcd->state == HC_STATE_RESUMING) {
1838 pr_warning("%s: duplicate resume\n", __func__);
1839 status = 0;
1840 } else
1841 switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1842 case OHCI_USB_SUSPEND:
1843 DBG(0, "%s: resume root hub\n", __func__);
1844 isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1845 isp1362_hcd->hc_control |= OHCI_USB_RESUME;
1846 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1847 break;
1848 case OHCI_USB_RESUME:
1849 /* HCFS changes sometime after INTR_RD */
1850 DBG(0, "%s: remote wakeup\n", __func__);
1851 break;
1852 case OHCI_USB_OPER:
1853 DBG(0, "%s: odd resume\n", __func__);
1854 status = 0;
1855 hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1856 break;
1857 default: /* RESET, we lost power */
1858 DBG(0, "%s: root hub hardware reset\n", __func__);
1859 status = -EBUSY;
1860 }
1861 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1862 if (status == -EBUSY) {
1863 DBG(0, "%s: Restarting HC\n", __func__);
1864 isp1362_hc_stop(hcd);
1865 return isp1362_hc_start(hcd);
1866 }
1867 if (status != -EINPROGRESS)
1868 return status;
1869 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1870 port = isp1362_read_reg32(isp1362_hcd, HCRHDESCA) & RH_A_NDP;
1871 while (port--) {
1872 u32 stat = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + port);
1873
1874 /* force global, not selective, resume */
1875 if (!(stat & RH_PS_PSS)) {
1876 DBG(0, "%s: Not Resuming RH port %d\n", __func__, port);
1877 continue;
1878 }
1879 DBG(0, "%s: Resuming RH port %d\n", __func__, port);
1880 isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + port, RH_PS_POCI);
1881 }
1882 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1883
1884 /* Some controllers (lucent) need extra-long delays */
1885 hcd->state = HC_STATE_RESUMING;
1886 mdelay(20 /* usb 11.5.1.10 */ + 15);
1887
1888 isp1362_hcd->hc_control = OHCI_USB_OPER;
1889 spin_lock_irqsave(&isp1362_hcd->lock, flags);
1890 isp1362_show_reg(isp1362_hcd, HCCONTROL);
1891 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1892 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1893 /* TRSMRCY */
1894 msleep(10);
1895
1896 /* keep it alive for ~5x suspend + resume costs */
1897 isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(250);
1898
1899 hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1900 hcd->state = HC_STATE_RUNNING;
1901 return 0;
1902}
1903#else
1904#define isp1362_bus_suspend NULL
1905#define isp1362_bus_resume NULL
1906#endif
1907
1908/*-------------------------------------------------------------------------*/
1909
1910static void dump_irq(struct seq_file *s, char *label, u16 mask)
1911{
1912 seq_printf(s, "%-15s %04x%s%s%s%s%s%s\n", label, mask,
1913 mask & HCuPINT_CLKRDY ? " clkrdy" : "",
1914 mask & HCuPINT_SUSP ? " susp" : "",
1915 mask & HCuPINT_OPR ? " opr" : "",
1916 mask & HCuPINT_EOT ? " eot" : "",
1917 mask & HCuPINT_ATL ? " atl" : "",
1918 mask & HCuPINT_SOF ? " sof" : "");
1919}
1920
1921static void dump_int(struct seq_file *s, char *label, u32 mask)
1922{
1923 seq_printf(s, "%-15s %08x%s%s%s%s%s%s%s\n", label, mask,
1924 mask & OHCI_INTR_MIE ? " MIE" : "",
1925 mask & OHCI_INTR_RHSC ? " rhsc" : "",
1926 mask & OHCI_INTR_FNO ? " fno" : "",
1927 mask & OHCI_INTR_UE ? " ue" : "",
1928 mask & OHCI_INTR_RD ? " rd" : "",
1929 mask & OHCI_INTR_SF ? " sof" : "",
1930 mask & OHCI_INTR_SO ? " so" : "");
1931}
1932
1933static void dump_ctrl(struct seq_file *s, char *label, u32 mask)
1934{
1935 seq_printf(s, "%-15s %08x%s%s%s\n", label, mask,
1936 mask & OHCI_CTRL_RWC ? " rwc" : "",
1937 mask & OHCI_CTRL_RWE ? " rwe" : "",
1938 ({
1939 char *hcfs;
1940 switch (mask & OHCI_CTRL_HCFS) {
1941 case OHCI_USB_OPER:
1942 hcfs = " oper";
1943 break;
1944 case OHCI_USB_RESET:
1945 hcfs = " reset";
1946 break;
1947 case OHCI_USB_RESUME:
1948 hcfs = " resume";
1949 break;
1950 case OHCI_USB_SUSPEND:
1951 hcfs = " suspend";
1952 break;
1953 default:
1954 hcfs = " ?";
1955 }
1956 hcfs;
1957 }));
1958}
1959
1960static void dump_regs(struct seq_file *s, struct isp1362_hcd *isp1362_hcd)
1961{
1962 seq_printf(s, "HCREVISION [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCREVISION),
1963 isp1362_read_reg32(isp1362_hcd, HCREVISION));
1964 seq_printf(s, "HCCONTROL [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCONTROL),
1965 isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1966 seq_printf(s, "HCCMDSTAT [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCMDSTAT),
1967 isp1362_read_reg32(isp1362_hcd, HCCMDSTAT));
1968 seq_printf(s, "HCINTSTAT [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTSTAT),
1969 isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1970 seq_printf(s, "HCINTENB [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTENB),
1971 isp1362_read_reg32(isp1362_hcd, HCINTENB));
1972 seq_printf(s, "HCFMINTVL [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMINTVL),
1973 isp1362_read_reg32(isp1362_hcd, HCFMINTVL));
1974 seq_printf(s, "HCFMREM [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMREM),
1975 isp1362_read_reg32(isp1362_hcd, HCFMREM));
1976 seq_printf(s, "HCFMNUM [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMNUM),
1977 isp1362_read_reg32(isp1362_hcd, HCFMNUM));
1978 seq_printf(s, "HCLSTHRESH [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCLSTHRESH),
1979 isp1362_read_reg32(isp1362_hcd, HCLSTHRESH));
1980 seq_printf(s, "HCRHDESCA [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCA),
1981 isp1362_read_reg32(isp1362_hcd, HCRHDESCA));
1982 seq_printf(s, "HCRHDESCB [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCB),
1983 isp1362_read_reg32(isp1362_hcd, HCRHDESCB));
1984 seq_printf(s, "HCRHSTATUS [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHSTATUS),
1985 isp1362_read_reg32(isp1362_hcd, HCRHSTATUS));
1986 seq_printf(s, "HCRHPORT1 [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT1),
1987 isp1362_read_reg32(isp1362_hcd, HCRHPORT1));
1988 seq_printf(s, "HCRHPORT2 [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT2),
1989 isp1362_read_reg32(isp1362_hcd, HCRHPORT2));
1990 seq_printf(s, "\n");
1991 seq_printf(s, "HCHWCFG [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCHWCFG),
1992 isp1362_read_reg16(isp1362_hcd, HCHWCFG));
1993 seq_printf(s, "HCDMACFG [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCDMACFG),
1994 isp1362_read_reg16(isp1362_hcd, HCDMACFG));
1995 seq_printf(s, "HCXFERCTR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCXFERCTR),
1996 isp1362_read_reg16(isp1362_hcd, HCXFERCTR));
1997 seq_printf(s, "HCuPINT [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINT),
1998 isp1362_read_reg16(isp1362_hcd, HCuPINT));
1999 seq_printf(s, "HCuPINTENB [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINTENB),
2000 isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2001 seq_printf(s, "HCCHIPID [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCCHIPID),
2002 isp1362_read_reg16(isp1362_hcd, HCCHIPID));
2003 seq_printf(s, "HCSCRATCH [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCSCRATCH),
2004 isp1362_read_reg16(isp1362_hcd, HCSCRATCH));
2005 seq_printf(s, "HCBUFSTAT [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCBUFSTAT),
2006 isp1362_read_reg16(isp1362_hcd, HCBUFSTAT));
2007 seq_printf(s, "HCDIRADDR [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCDIRADDR),
2008 isp1362_read_reg32(isp1362_hcd, HCDIRADDR));
2009#if 0
2010 seq_printf(s, "HCDIRDATA [%02x] %04x\n", ISP1362_REG_NO(HCDIRDATA),
2011 isp1362_read_reg16(isp1362_hcd, HCDIRDATA));
2012#endif
2013 seq_printf(s, "HCISTLBUFSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLBUFSZ),
2014 isp1362_read_reg16(isp1362_hcd, HCISTLBUFSZ));
2015 seq_printf(s, "HCISTLRATE [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLRATE),
2016 isp1362_read_reg16(isp1362_hcd, HCISTLRATE));
2017 seq_printf(s, "\n");
2018 seq_printf(s, "HCINTLBUFSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBUFSZ),
2019 isp1362_read_reg16(isp1362_hcd, HCINTLBUFSZ));
2020 seq_printf(s, "HCINTLBLKSZ[%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBLKSZ),
2021 isp1362_read_reg16(isp1362_hcd, HCINTLBLKSZ));
2022 seq_printf(s, "HCINTLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLDONE),
2023 isp1362_read_reg32(isp1362_hcd, HCINTLDONE));
2024 seq_printf(s, "HCINTLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLSKIP),
2025 isp1362_read_reg32(isp1362_hcd, HCINTLSKIP));
2026 seq_printf(s, "HCINTLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLLAST),
2027 isp1362_read_reg32(isp1362_hcd, HCINTLLAST));
2028 seq_printf(s, "HCINTLCURR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLCURR),
2029 isp1362_read_reg16(isp1362_hcd, HCINTLCURR));
2030 seq_printf(s, "\n");
2031 seq_printf(s, "HCATLBUFSZ [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBUFSZ),
2032 isp1362_read_reg16(isp1362_hcd, HCATLBUFSZ));
2033 seq_printf(s, "HCATLBLKSZ [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBLKSZ),
2034 isp1362_read_reg16(isp1362_hcd, HCATLBLKSZ));
2035#if 0
2036 seq_printf(s, "HCATLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDONE),
2037 isp1362_read_reg32(isp1362_hcd, HCATLDONE));
2038#endif
2039 seq_printf(s, "HCATLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLSKIP),
2040 isp1362_read_reg32(isp1362_hcd, HCATLSKIP));
2041 seq_printf(s, "HCATLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLLAST),
2042 isp1362_read_reg32(isp1362_hcd, HCATLLAST));
2043 seq_printf(s, "HCATLCURR [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLCURR),
2044 isp1362_read_reg16(isp1362_hcd, HCATLCURR));
2045 seq_printf(s, "\n");
2046 seq_printf(s, "HCATLDTC [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTC),
2047 isp1362_read_reg16(isp1362_hcd, HCATLDTC));
2048 seq_printf(s, "HCATLDTCTO [%02x] %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTCTO),
2049 isp1362_read_reg16(isp1362_hcd, HCATLDTCTO));
2050}
2051
2052static int isp1362_show(struct seq_file *s, void *unused)
2053{
2054 struct isp1362_hcd *isp1362_hcd = s->private;
2055 struct isp1362_ep *ep;
2056 int i;
2057
2058 seq_printf(s, "%s\n%s version %s\n",
2059 isp1362_hcd_to_hcd(isp1362_hcd)->product_desc, hcd_name, DRIVER_VERSION);
2060
2061 /* collect statistics to help estimate potential win for
2062 * DMA engines that care about alignment (PXA)
2063 */
2064 seq_printf(s, "alignment: 16b/%ld 8b/%ld 4b/%ld 2b/%ld 1b/%ld\n",
2065 isp1362_hcd->stat16, isp1362_hcd->stat8, isp1362_hcd->stat4,
2066 isp1362_hcd->stat2, isp1362_hcd->stat1);
2067 seq_printf(s, "max # ptds in ATL fifo: %d\n", isp1362_hcd->atl_queue.stat_maxptds);
2068 seq_printf(s, "max # ptds in INTL fifo: %d\n", isp1362_hcd->intl_queue.stat_maxptds);
2069 seq_printf(s, "max # ptds in ISTL fifo: %d\n",
2070 max(isp1362_hcd->istl_queue[0] .stat_maxptds,
2071 isp1362_hcd->istl_queue[1] .stat_maxptds));
2072
2073 /* FIXME: don't show the following in suspended state */
2074 spin_lock_irq(&isp1362_hcd->lock);
2075
2076 dump_irq(s, "hc_irq_enable", isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2077 dump_irq(s, "hc_irq_status", isp1362_read_reg16(isp1362_hcd, HCuPINT));
2078 dump_int(s, "ohci_int_enable", isp1362_read_reg32(isp1362_hcd, HCINTENB));
2079 dump_int(s, "ohci_int_status", isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
2080 dump_ctrl(s, "ohci_control", isp1362_read_reg32(isp1362_hcd, HCCONTROL));
2081
2082 for (i = 0; i < NUM_ISP1362_IRQS; i++)
2083 if (isp1362_hcd->irq_stat[i])
2084 seq_printf(s, "%-15s: %d\n",
2085 ISP1362_INT_NAME(i), isp1362_hcd->irq_stat[i]);
2086
2087 dump_regs(s, isp1362_hcd);
2088 list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
2089 struct urb *urb;
2090
2091 seq_printf(s, "%p, ep%d%s, maxpacket %d:\n", ep, ep->epnum,
2092 ({
2093 char *s;
2094 switch (ep->nextpid) {
2095 case USB_PID_IN:
2096 s = "in";
2097 break;
2098 case USB_PID_OUT:
2099 s = "out";
2100 break;
2101 case USB_PID_SETUP:
2102 s = "setup";
2103 break;
2104 case USB_PID_ACK:
2105 s = "status";
2106 break;
2107 default:
2108 s = "?";
2109 break;
2110 }
2111 s;}), ep->maxpacket) ;
2112 list_for_each_entry(urb, &ep->hep->urb_list, urb_list) {
2113 seq_printf(s, " urb%p, %d/%d\n", urb,
2114 urb->actual_length,
2115 urb->transfer_buffer_length);
2116 }
2117 }
2118 if (!list_empty(&isp1362_hcd->async))
2119 seq_printf(s, "\n");
2120 dump_ptd_queue(&isp1362_hcd->atl_queue);
2121
2122 seq_printf(s, "periodic size= %d\n", PERIODIC_SIZE);
2123
2124 list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
2125 seq_printf(s, "branch:%2d load:%3d PTD[%d] $%04x:\n", ep->branch,
2126 isp1362_hcd->load[ep->branch], ep->ptd_index, ep->ptd_offset);
2127
2128 seq_printf(s, " %d/%p (%sdev%d ep%d%s max %d)\n",
2129 ep->interval, ep,
2130 (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2131 ep->udev->devnum, ep->epnum,
2132 (ep->epnum == 0) ? "" :
2133 ((ep->nextpid == USB_PID_IN) ?
2134 "in" : "out"), ep->maxpacket);
2135 }
2136 dump_ptd_queue(&isp1362_hcd->intl_queue);
2137
2138 seq_printf(s, "ISO:\n");
2139
2140 list_for_each_entry(ep, &isp1362_hcd->isoc, schedule) {
2141 seq_printf(s, " %d/%p (%sdev%d ep%d%s max %d)\n",
2142 ep->interval, ep,
2143 (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2144 ep->udev->devnum, ep->epnum,
2145 (ep->epnum == 0) ? "" :
2146 ((ep->nextpid == USB_PID_IN) ?
2147 "in" : "out"), ep->maxpacket);
2148 }
2149
2150 spin_unlock_irq(&isp1362_hcd->lock);
2151 seq_printf(s, "\n");
2152
2153 return 0;
2154}
2155
2156static int isp1362_open(struct inode *inode, struct file *file)
2157{
2158 return single_open(file, isp1362_show, inode);
2159}
2160
2161static const struct file_operations debug_ops = {
2162 .open = isp1362_open,
2163 .read = seq_read,
2164 .llseek = seq_lseek,
2165 .release = single_release,
2166};
2167
2168/* expect just one isp1362_hcd per system */
2169static void create_debug_file(struct isp1362_hcd *isp1362_hcd)
2170{
2171 isp1362_hcd->debug_file = debugfs_create_file("isp1362", S_IRUGO,
2172 usb_debug_root,
2173 isp1362_hcd, &debug_ops);
2174}
2175
2176static void remove_debug_file(struct isp1362_hcd *isp1362_hcd)
2177{
2178 debugfs_remove(isp1362_hcd->debug_file);
2179}
2180
2181/*-------------------------------------------------------------------------*/
2182
2183static void __isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2184{
2185 int tmp = 20;
2186
2187 isp1362_write_reg16(isp1362_hcd, HCSWRES, HCSWRES_MAGIC);
2188 isp1362_write_reg32(isp1362_hcd, HCCMDSTAT, OHCI_HCR);
2189 while (--tmp) {
2190 mdelay(1);
2191 if (!(isp1362_read_reg32(isp1362_hcd, HCCMDSTAT) & OHCI_HCR))
2192 break;
2193 }
2194 if (!tmp)
2195 pr_err("Software reset timeout\n");
2196}
2197
2198static void isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2199{
2200 unsigned long flags;
2201
2202 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2203 __isp1362_sw_reset(isp1362_hcd);
2204 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2205}
2206
2207static int isp1362_mem_config(struct usb_hcd *hcd)
2208{
2209 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2210 unsigned long flags;
2211 u32 total;
2212 u16 istl_size = ISP1362_ISTL_BUFSIZE;
2213 u16 intl_blksize = ISP1362_INTL_BLKSIZE + PTD_HEADER_SIZE;
2214 u16 intl_size = ISP1362_INTL_BUFFERS * intl_blksize;
2215 u16 atl_blksize = ISP1362_ATL_BLKSIZE + PTD_HEADER_SIZE;
2216 u16 atl_buffers = (ISP1362_BUF_SIZE - (istl_size + intl_size)) / atl_blksize;
2217 u16 atl_size;
2218 int i;
2219
2220 WARN_ON(istl_size & 3);
2221 WARN_ON(atl_blksize & 3);
2222 WARN_ON(intl_blksize & 3);
2223 WARN_ON(atl_blksize < PTD_HEADER_SIZE);
2224 WARN_ON(intl_blksize < PTD_HEADER_SIZE);
2225
2226 BUG_ON((unsigned)ISP1362_INTL_BUFFERS > 32);
2227 if (atl_buffers > 32)
2228 atl_buffers = 32;
2229 atl_size = atl_buffers * atl_blksize;
2230 total = atl_size + intl_size + istl_size;
2231 dev_info(hcd->self.controller, "ISP1362 Memory usage:\n");
2232 dev_info(hcd->self.controller, " ISTL: 2 * %4d: %4d @ $%04x:$%04x\n",
2233 istl_size / 2, istl_size, 0, istl_size / 2);
2234 dev_info(hcd->self.controller, " INTL: %4d * (%3zu+8): %4d @ $%04x\n",
2235 ISP1362_INTL_BUFFERS, intl_blksize - PTD_HEADER_SIZE,
2236 intl_size, istl_size);
2237 dev_info(hcd->self.controller, " ATL : %4d * (%3zu+8): %4d @ $%04x\n",
2238 atl_buffers, atl_blksize - PTD_HEADER_SIZE,
2239 atl_size, istl_size + intl_size);
2240 dev_info(hcd->self.controller, " USED/FREE: %4d %4d\n", total,
2241 ISP1362_BUF_SIZE - total);
2242
2243 if (total > ISP1362_BUF_SIZE) {
2244 dev_err(hcd->self.controller, "%s: Memory requested: %d, available %d\n",
2245 __func__, total, ISP1362_BUF_SIZE);
2246 return -ENOMEM;
2247 }
2248
2249 total = istl_size + intl_size + atl_size;
2250 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2251
2252 for (i = 0; i < 2; i++) {
2253 isp1362_hcd->istl_queue[i].buf_start = i * istl_size / 2,
2254 isp1362_hcd->istl_queue[i].buf_size = istl_size / 2;
2255 isp1362_hcd->istl_queue[i].blk_size = 4;
2256 INIT_LIST_HEAD(&isp1362_hcd->istl_queue[i].active);
2257 snprintf(isp1362_hcd->istl_queue[i].name,
2258 sizeof(isp1362_hcd->istl_queue[i].name), "ISTL%d", i);
2259 DBG(3, "%s: %5s buf $%04x %d\n", __func__,
2260 isp1362_hcd->istl_queue[i].name,
2261 isp1362_hcd->istl_queue[i].buf_start,
2262 isp1362_hcd->istl_queue[i].buf_size);
2263 }
2264 isp1362_write_reg16(isp1362_hcd, HCISTLBUFSZ, istl_size / 2);
2265
2266 isp1362_hcd->intl_queue.buf_start = istl_size;
2267 isp1362_hcd->intl_queue.buf_size = intl_size;
2268 isp1362_hcd->intl_queue.buf_count = ISP1362_INTL_BUFFERS;
2269 isp1362_hcd->intl_queue.blk_size = intl_blksize;
2270 isp1362_hcd->intl_queue.buf_avail = isp1362_hcd->intl_queue.buf_count;
2271 isp1362_hcd->intl_queue.skip_map = ~0;
2272 INIT_LIST_HEAD(&isp1362_hcd->intl_queue.active);
2273
2274 isp1362_write_reg16(isp1362_hcd, HCINTLBUFSZ,
2275 isp1362_hcd->intl_queue.buf_size);
2276 isp1362_write_reg16(isp1362_hcd, HCINTLBLKSZ,
2277 isp1362_hcd->intl_queue.blk_size - PTD_HEADER_SIZE);
2278 isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
2279 isp1362_write_reg32(isp1362_hcd, HCINTLLAST,
2280 1 << (ISP1362_INTL_BUFFERS - 1));
2281
2282 isp1362_hcd->atl_queue.buf_start = istl_size + intl_size;
2283 isp1362_hcd->atl_queue.buf_size = atl_size;
2284 isp1362_hcd->atl_queue.buf_count = atl_buffers;
2285 isp1362_hcd->atl_queue.blk_size = atl_blksize;
2286 isp1362_hcd->atl_queue.buf_avail = isp1362_hcd->atl_queue.buf_count;
2287 isp1362_hcd->atl_queue.skip_map = ~0;
2288 INIT_LIST_HEAD(&isp1362_hcd->atl_queue.active);
2289
2290 isp1362_write_reg16(isp1362_hcd, HCATLBUFSZ,
2291 isp1362_hcd->atl_queue.buf_size);
2292 isp1362_write_reg16(isp1362_hcd, HCATLBLKSZ,
2293 isp1362_hcd->atl_queue.blk_size - PTD_HEADER_SIZE);
2294 isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
2295 isp1362_write_reg32(isp1362_hcd, HCATLLAST,
2296 1 << (atl_buffers - 1));
2297
2298 snprintf(isp1362_hcd->atl_queue.name,
2299 sizeof(isp1362_hcd->atl_queue.name), "ATL");
2300 snprintf(isp1362_hcd->intl_queue.name,
2301 sizeof(isp1362_hcd->intl_queue.name), "INTL");
2302 DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2303 isp1362_hcd->intl_queue.name,
2304 isp1362_hcd->intl_queue.buf_start,
2305 ISP1362_INTL_BUFFERS, isp1362_hcd->intl_queue.blk_size,
2306 isp1362_hcd->intl_queue.buf_size);
2307 DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2308 isp1362_hcd->atl_queue.name,
2309 isp1362_hcd->atl_queue.buf_start,
2310 atl_buffers, isp1362_hcd->atl_queue.blk_size,
2311 isp1362_hcd->atl_queue.buf_size);
2312
2313 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2314
2315 return 0;
2316}
2317
2318static int isp1362_hc_reset(struct usb_hcd *hcd)
2319{
2320 int ret = 0;
2321 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2322 unsigned long t;
2323 unsigned long timeout = 100;
2324 unsigned long flags;
2325 int clkrdy = 0;
2326
2327 pr_debug("%s:\n", __func__);
2328
2329 if (isp1362_hcd->board && isp1362_hcd->board->reset) {
2330 isp1362_hcd->board->reset(hcd->self.controller, 1);
2331 msleep(20);
2332 if (isp1362_hcd->board->clock)
2333 isp1362_hcd->board->clock(hcd->self.controller, 1);
2334 isp1362_hcd->board->reset(hcd->self.controller, 0);
2335 } else
2336 isp1362_sw_reset(isp1362_hcd);
2337
2338 /* chip has been reset. First we need to see a clock */
2339 t = jiffies + msecs_to_jiffies(timeout);
2340 while (!clkrdy && time_before_eq(jiffies, t)) {
2341 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2342 clkrdy = isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_CLKRDY;
2343 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2344 if (!clkrdy)
2345 msleep(4);
2346 }
2347
2348 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2349 isp1362_write_reg16(isp1362_hcd, HCuPINT, HCuPINT_CLKRDY);
2350 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2351 if (!clkrdy) {
2352 pr_err("Clock not ready after %lums\n", timeout);
2353 ret = -ENODEV;
2354 }
2355 return ret;
2356}
2357
2358static void isp1362_hc_stop(struct usb_hcd *hcd)
2359{
2360 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2361 unsigned long flags;
2362 u32 tmp;
2363
2364 pr_debug("%s:\n", __func__);
2365
2366 del_timer_sync(&hcd->rh_timer);
2367
2368 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2369
2370 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2371
2372 /* Switch off power for all ports */
2373 tmp = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2374 tmp &= ~(RH_A_NPS | RH_A_PSM);
2375 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, tmp);
2376 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2377
2378 /* Reset the chip */
2379 if (isp1362_hcd->board && isp1362_hcd->board->reset)
2380 isp1362_hcd->board->reset(hcd->self.controller, 1);
2381 else
2382 __isp1362_sw_reset(isp1362_hcd);
2383
2384 if (isp1362_hcd->board && isp1362_hcd->board->clock)
2385 isp1362_hcd->board->clock(hcd->self.controller, 0);
2386
2387 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2388}
2389
2390#ifdef CHIP_BUFFER_TEST
2391static int isp1362_chip_test(struct isp1362_hcd *isp1362_hcd)
2392{
2393 int ret = 0;
2394 u16 *ref;
2395 unsigned long flags;
2396
2397 ref = kmalloc(2 * ISP1362_BUF_SIZE, GFP_KERNEL);
2398 if (ref) {
2399 int offset;
2400 u16 *tst = &ref[ISP1362_BUF_SIZE / 2];
2401
2402 for (offset = 0; offset < ISP1362_BUF_SIZE / 2; offset++) {
2403 ref[offset] = ~offset;
2404 tst[offset] = offset;
2405 }
2406
2407 for (offset = 0; offset < 4; offset++) {
2408 int j;
2409
2410 for (j = 0; j < 8; j++) {
2411 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2412 isp1362_write_buffer(isp1362_hcd, (u8 *)ref + offset, 0, j);
2413 isp1362_read_buffer(isp1362_hcd, (u8 *)tst + offset, 0, j);
2414 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2415
2416 if (memcmp(ref, tst, j)) {
2417 ret = -ENODEV;
2418 pr_err("%s: memory check with %d byte offset %d failed\n",
2419 __func__, j, offset);
2420 dump_data((u8 *)ref + offset, j);
2421 dump_data((u8 *)tst + offset, j);
2422 }
2423 }
2424 }
2425
2426 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2427 isp1362_write_buffer(isp1362_hcd, ref, 0, ISP1362_BUF_SIZE);
2428 isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2429 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2430
2431 if (memcmp(ref, tst, ISP1362_BUF_SIZE)) {
2432 ret = -ENODEV;
2433 pr_err("%s: memory check failed\n", __func__);
2434 dump_data((u8 *)tst, ISP1362_BUF_SIZE / 2);
2435 }
2436
2437 for (offset = 0; offset < 256; offset++) {
2438 int test_size = 0;
2439
2440 yield();
2441
2442 memset(tst, 0, ISP1362_BUF_SIZE);
2443 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2444 isp1362_write_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2445 isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2446 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2447 if (memcmp(tst, tst + (ISP1362_BUF_SIZE / (2 * sizeof(*tst))),
2448 ISP1362_BUF_SIZE / 2)) {
2449 pr_err("%s: Failed to clear buffer\n", __func__);
2450 dump_data((u8 *)tst, ISP1362_BUF_SIZE);
2451 break;
2452 }
2453 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2454 isp1362_write_buffer(isp1362_hcd, ref, offset * 2, PTD_HEADER_SIZE);
2455 isp1362_write_buffer(isp1362_hcd, ref + PTD_HEADER_SIZE / sizeof(*ref),
2456 offset * 2 + PTD_HEADER_SIZE, test_size);
2457 isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2458 PTD_HEADER_SIZE + test_size);
2459 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2460 if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2461 dump_data(((u8 *)ref) + offset, PTD_HEADER_SIZE + test_size);
2462 dump_data((u8 *)tst, PTD_HEADER_SIZE + test_size);
2463 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2464 isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2465 PTD_HEADER_SIZE + test_size);
2466 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2467 if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2468 ret = -ENODEV;
2469 pr_err("%s: memory check with offset %02x failed\n",
2470 __func__, offset);
2471 break;
2472 }
2473 pr_warning("%s: memory check with offset %02x ok after second read\n",
2474 __func__, offset);
2475 }
2476 }
2477 kfree(ref);
2478 }
2479 return ret;
2480}
2481#endif
2482
2483static int isp1362_hc_start(struct usb_hcd *hcd)
2484{
2485 int ret;
2486 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2487 struct isp1362_platform_data *board = isp1362_hcd->board;
2488 u16 hwcfg;
2489 u16 chipid;
2490 unsigned long flags;
2491
2492 pr_debug("%s:\n", __func__);
2493
2494 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2495 chipid = isp1362_read_reg16(isp1362_hcd, HCCHIPID);
2496 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2497
2498 if ((chipid & HCCHIPID_MASK) != HCCHIPID_MAGIC) {
2499 pr_err("%s: Invalid chip ID %04x\n", __func__, chipid);
2500 return -ENODEV;
2501 }
2502
2503#ifdef CHIP_BUFFER_TEST
2504 ret = isp1362_chip_test(isp1362_hcd);
2505 if (ret)
2506 return -ENODEV;
2507#endif
2508 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2509 /* clear interrupt status and disable all interrupt sources */
2510 isp1362_write_reg16(isp1362_hcd, HCuPINT, 0xff);
2511 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2512
2513 /* HW conf */
2514 hwcfg = HCHWCFG_INT_ENABLE | HCHWCFG_DBWIDTH(1);
2515 if (board->sel15Kres)
2516 hwcfg |= HCHWCFG_PULLDOWN_DS2 |
2517 ((MAX_ROOT_PORTS > 1) ? HCHWCFG_PULLDOWN_DS1 : 0);
2518 if (board->clknotstop)
2519 hwcfg |= HCHWCFG_CLKNOTSTOP;
2520 if (board->oc_enable)
2521 hwcfg |= HCHWCFG_ANALOG_OC;
2522 if (board->int_act_high)
2523 hwcfg |= HCHWCFG_INT_POL;
2524 if (board->int_edge_triggered)
2525 hwcfg |= HCHWCFG_INT_TRIGGER;
2526 if (board->dreq_act_high)
2527 hwcfg |= HCHWCFG_DREQ_POL;
2528 if (board->dack_act_high)
2529 hwcfg |= HCHWCFG_DACK_POL;
2530 isp1362_write_reg16(isp1362_hcd, HCHWCFG, hwcfg);
2531 isp1362_show_reg(isp1362_hcd, HCHWCFG);
2532 isp1362_write_reg16(isp1362_hcd, HCDMACFG, 0);
2533 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2534
2535 ret = isp1362_mem_config(hcd);
2536 if (ret)
2537 return ret;
2538
2539 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2540
2541 /* Root hub conf */
2542 isp1362_hcd->rhdesca = 0;
2543 if (board->no_power_switching)
2544 isp1362_hcd->rhdesca |= RH_A_NPS;
2545 if (board->power_switching_mode)
2546 isp1362_hcd->rhdesca |= RH_A_PSM;
2547 if (board->potpg)
2548 isp1362_hcd->rhdesca |= (board->potpg << 24) & RH_A_POTPGT;
2549 else
2550 isp1362_hcd->rhdesca |= (25 << 24) & RH_A_POTPGT;
2551
2552 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca & ~RH_A_OCPM);
2553 isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca | RH_A_OCPM);
2554 isp1362_hcd->rhdesca = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2555
2556 isp1362_hcd->rhdescb = RH_B_PPCM;
2557 isp1362_write_reg32(isp1362_hcd, HCRHDESCB, isp1362_hcd->rhdescb);
2558 isp1362_hcd->rhdescb = isp1362_read_reg32(isp1362_hcd, HCRHDESCB);
2559
2560 isp1362_read_reg32(isp1362_hcd, HCFMINTVL);
2561 isp1362_write_reg32(isp1362_hcd, HCFMINTVL, (FSMP(FI) << 16) | FI);
2562 isp1362_write_reg32(isp1362_hcd, HCLSTHRESH, LSTHRESH);
2563
2564 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2565
2566 isp1362_hcd->hc_control = OHCI_USB_OPER;
2567 hcd->state = HC_STATE_RUNNING;
2568
2569 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2570 /* Set up interrupts */
2571 isp1362_hcd->intenb = OHCI_INTR_MIE | OHCI_INTR_RHSC | OHCI_INTR_UE;
2572 isp1362_hcd->intenb |= OHCI_INTR_RD;
2573 isp1362_hcd->irqenb = HCuPINT_OPR | HCuPINT_SUSP;
2574 isp1362_write_reg32(isp1362_hcd, HCINTENB, isp1362_hcd->intenb);
2575 isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
2576
2577 /* Go operational */
2578 isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
2579 /* enable global power */
2580 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC | RH_HS_DRWE);
2581
2582 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2583
2584 return 0;
2585}
2586
2587/*-------------------------------------------------------------------------*/
2588
2589static struct hc_driver isp1362_hc_driver = {
2590 .description = hcd_name,
2591 .product_desc = "ISP1362 Host Controller",
2592 .hcd_priv_size = sizeof(struct isp1362_hcd),
2593
2594 .irq = isp1362_irq,
2595 .flags = HCD_USB11 | HCD_MEMORY,
2596
2597 .reset = isp1362_hc_reset,
2598 .start = isp1362_hc_start,
2599 .stop = isp1362_hc_stop,
2600
2601 .urb_enqueue = isp1362_urb_enqueue,
2602 .urb_dequeue = isp1362_urb_dequeue,
2603 .endpoint_disable = isp1362_endpoint_disable,
2604
2605 .get_frame_number = isp1362_get_frame,
2606
2607 .hub_status_data = isp1362_hub_status_data,
2608 .hub_control = isp1362_hub_control,
2609 .bus_suspend = isp1362_bus_suspend,
2610 .bus_resume = isp1362_bus_resume,
2611};
2612
2613/*-------------------------------------------------------------------------*/
2614
2615static int isp1362_remove(struct platform_device *pdev)
2616{
2617 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2618 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2619 struct resource *res;
2620
2621 remove_debug_file(isp1362_hcd);
2622 DBG(0, "%s: Removing HCD\n", __func__);
2623 usb_remove_hcd(hcd);
2624
2625 DBG(0, "%s: Unmapping data_reg @ %p\n", __func__,
2626 isp1362_hcd->data_reg);
2627 iounmap(isp1362_hcd->data_reg);
2628
2629 DBG(0, "%s: Unmapping addr_reg @ %p\n", __func__,
2630 isp1362_hcd->addr_reg);
2631 iounmap(isp1362_hcd->addr_reg);
2632
2633 res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
2634 DBG(0, "%s: release mem_region: %08lx\n", __func__, (long unsigned int)res->start);
2635 if (res)
2636 release_mem_region(res->start, resource_size(res));
2637
2638 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
2639 DBG(0, "%s: release mem_region: %08lx\n", __func__, (long unsigned int)res->start);
2640 if (res)
2641 release_mem_region(res->start, resource_size(res));
2642
2643 DBG(0, "%s: put_hcd\n", __func__);
2644 usb_put_hcd(hcd);
2645 DBG(0, "%s: Done\n", __func__);
2646
2647 return 0;
2648}
2649
2650static int isp1362_probe(struct platform_device *pdev)
2651{
2652 struct usb_hcd *hcd;
2653 struct isp1362_hcd *isp1362_hcd;
2654 struct resource *addr, *data;
2655 void __iomem *addr_reg;
2656 void __iomem *data_reg;
2657 int irq;
2658 int retval = 0;
2659 struct resource *irq_res;
2660 unsigned int irq_flags = 0;
2661
2662 if (usb_disabled())
2663 return -ENODEV;
2664
2665 /* basic sanity checks first. board-specific init logic should
2666 * have initialized this the three resources and probably board
2667 * specific platform_data. we don't probe for IRQs, and do only
2668 * minimal sanity checking.
2669 */
2670 if (pdev->num_resources < 3) {
2671 retval = -ENODEV;
2672 goto err1;
2673 }
2674
2675 data = platform_get_resource(pdev, IORESOURCE_MEM, 0);
2676 addr = platform_get_resource(pdev, IORESOURCE_MEM, 1);
2677 irq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
2678 if (!addr || !data || !irq_res) {
2679 retval = -ENODEV;
2680 goto err1;
2681 }
2682 irq = irq_res->start;
2683
2684 if (pdev->dev.dma_mask) {
2685 DBG(1, "won't do DMA");
2686 retval = -ENODEV;
2687 goto err1;
2688 }
2689
2690 if (!request_mem_region(addr->start, resource_size(addr), hcd_name)) {
2691 retval = -EBUSY;
2692 goto err1;
2693 }
2694 addr_reg = ioremap(addr->start, resource_size(addr));
2695 if (addr_reg == NULL) {
2696 retval = -ENOMEM;
2697 goto err2;
2698 }
2699
2700 if (!request_mem_region(data->start, resource_size(data), hcd_name)) {
2701 retval = -EBUSY;
2702 goto err3;
2703 }
2704 data_reg = ioremap(data->start, resource_size(data));
2705 if (data_reg == NULL) {
2706 retval = -ENOMEM;
2707 goto err4;
2708 }
2709
2710 /* allocate and initialize hcd */
2711 hcd = usb_create_hcd(&isp1362_hc_driver, &pdev->dev, dev_name(&pdev->dev));
2712 if (!hcd) {
2713 retval = -ENOMEM;
2714 goto err5;
2715 }
2716 hcd->rsrc_start = data->start;
2717 isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2718 isp1362_hcd->data_reg = data_reg;
2719 isp1362_hcd->addr_reg = addr_reg;
2720
2721 isp1362_hcd->next_statechange = jiffies;
2722 spin_lock_init(&isp1362_hcd->lock);
2723 INIT_LIST_HEAD(&isp1362_hcd->async);
2724 INIT_LIST_HEAD(&isp1362_hcd->periodic);
2725 INIT_LIST_HEAD(&isp1362_hcd->isoc);
2726 INIT_LIST_HEAD(&isp1362_hcd->remove_list);
2727 isp1362_hcd->board = dev_get_platdata(&pdev->dev);
2728#if USE_PLATFORM_DELAY
2729 if (!isp1362_hcd->board->delay) {
2730 dev_err(hcd->self.controller, "No platform delay function given\n");
2731 retval = -ENODEV;
2732 goto err6;
2733 }
2734#endif
2735
2736 if (irq_res->flags & IORESOURCE_IRQ_HIGHEDGE)
2737 irq_flags |= IRQF_TRIGGER_RISING;
2738 if (irq_res->flags & IORESOURCE_IRQ_LOWEDGE)
2739 irq_flags |= IRQF_TRIGGER_FALLING;
2740 if (irq_res->flags & IORESOURCE_IRQ_HIGHLEVEL)
2741 irq_flags |= IRQF_TRIGGER_HIGH;
2742 if (irq_res->flags & IORESOURCE_IRQ_LOWLEVEL)
2743 irq_flags |= IRQF_TRIGGER_LOW;
2744
2745 retval = usb_add_hcd(hcd, irq, irq_flags | IRQF_SHARED);
2746 if (retval != 0)
2747 goto err6;
2748 device_wakeup_enable(hcd->self.controller);
2749
2750 pr_info("%s, irq %d\n", hcd->product_desc, irq);
2751
2752 create_debug_file(isp1362_hcd);
2753
2754 return 0;
2755
2756 err6:
2757 DBG(0, "%s: Freeing dev %p\n", __func__, isp1362_hcd);
2758 usb_put_hcd(hcd);
2759 err5:
2760 DBG(0, "%s: Unmapping data_reg @ %p\n", __func__, data_reg);
2761 iounmap(data_reg);
2762 err4:
2763 DBG(0, "%s: Releasing mem region %08lx\n", __func__, (long unsigned int)data->start);
2764 release_mem_region(data->start, resource_size(data));
2765 err3:
2766 DBG(0, "%s: Unmapping addr_reg @ %p\n", __func__, addr_reg);
2767 iounmap(addr_reg);
2768 err2:
2769 DBG(0, "%s: Releasing mem region %08lx\n", __func__, (long unsigned int)addr->start);
2770 release_mem_region(addr->start, resource_size(addr));
2771 err1:
2772 pr_err("%s: init error, %d\n", __func__, retval);
2773
2774 return retval;
2775}
2776
2777#ifdef CONFIG_PM
2778static int isp1362_suspend(struct platform_device *pdev, pm_message_t state)
2779{
2780 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2781 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2782 unsigned long flags;
2783 int retval = 0;
2784
2785 DBG(0, "%s: Suspending device\n", __func__);
2786
2787 if (state.event == PM_EVENT_FREEZE) {
2788 DBG(0, "%s: Suspending root hub\n", __func__);
2789 retval = isp1362_bus_suspend(hcd);
2790 } else {
2791 DBG(0, "%s: Suspending RH ports\n", __func__);
2792 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2793 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2794 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2795 }
2796 if (retval == 0)
2797 pdev->dev.power.power_state = state;
2798 return retval;
2799}
2800
2801static int isp1362_resume(struct platform_device *pdev)
2802{
2803 struct usb_hcd *hcd = platform_get_drvdata(pdev);
2804 struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2805 unsigned long flags;
2806
2807 DBG(0, "%s: Resuming\n", __func__);
2808
2809 if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
2810 DBG(0, "%s: Resume RH ports\n", __func__);
2811 spin_lock_irqsave(&isp1362_hcd->lock, flags);
2812 isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC);
2813 spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2814 return 0;
2815 }
2816
2817 pdev->dev.power.power_state = PMSG_ON;
2818
2819 return isp1362_bus_resume(isp1362_hcd_to_hcd(isp1362_hcd));
2820}
2821#else
2822#define isp1362_suspend NULL
2823#define isp1362_resume NULL
2824#endif
2825
2826static struct platform_driver isp1362_driver = {
2827 .probe = isp1362_probe,
2828 .remove = isp1362_remove,
2829
2830 .suspend = isp1362_suspend,
2831 .resume = isp1362_resume,
2832 .driver = {
2833 .name = hcd_name,
2834 .owner = THIS_MODULE,
2835 },
2836};
2837
2838module_platform_driver(isp1362_driver);