Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * mtu3_gadget.c - MediaTek usb3 DRD peripheral support
4 *
5 * Copyright (C) 2016 MediaTek Inc.
6 *
7 * Author: Chunfeng Yun <chunfeng.yun@mediatek.com>
8 */
9
10#include "mtu3.h"
11#include "mtu3_trace.h"
12
13void mtu3_req_complete(struct mtu3_ep *mep,
14 struct usb_request *req, int status)
15__releases(mep->mtu->lock)
16__acquires(mep->mtu->lock)
17{
18 struct mtu3_request *mreq = to_mtu3_request(req);
19 struct mtu3 *mtu = mreq->mtu;
20
21 list_del(&mreq->list);
22 if (req->status == -EINPROGRESS)
23 req->status = status;
24
25 trace_mtu3_req_complete(mreq);
26
27 /* ep0 makes use of PIO, needn't unmap it */
28 if (mep->epnum)
29 usb_gadget_unmap_request(&mtu->g, req, mep->is_in);
30
31 dev_dbg(mtu->dev, "%s complete req: %p, sts %d, %d/%d\n",
32 mep->name, req, req->status, req->actual, req->length);
33
34 spin_unlock(&mtu->lock);
35 usb_gadget_giveback_request(&mep->ep, req);
36 spin_lock(&mtu->lock);
37}
38
39static void nuke(struct mtu3_ep *mep, const int status)
40{
41 struct mtu3_request *mreq = NULL;
42
43 if (list_empty(&mep->req_list))
44 return;
45
46 dev_dbg(mep->mtu->dev, "abort %s's req: sts %d\n", mep->name, status);
47
48 /* exclude EP0 */
49 if (mep->epnum)
50 mtu3_qmu_flush(mep);
51
52 while (!list_empty(&mep->req_list)) {
53 mreq = list_first_entry(&mep->req_list,
54 struct mtu3_request, list);
55 mtu3_req_complete(mep, &mreq->request, status);
56 }
57}
58
59static int mtu3_ep_enable(struct mtu3_ep *mep)
60{
61 const struct usb_endpoint_descriptor *desc;
62 const struct usb_ss_ep_comp_descriptor *comp_desc;
63 struct mtu3 *mtu = mep->mtu;
64 u32 interval = 0;
65 u32 mult = 0;
66 u32 burst = 0;
67 int ret;
68
69 desc = mep->desc;
70 comp_desc = mep->comp_desc;
71 mep->type = usb_endpoint_type(desc);
72 mep->maxp = usb_endpoint_maxp(desc);
73
74 switch (mtu->g.speed) {
75 case USB_SPEED_SUPER:
76 case USB_SPEED_SUPER_PLUS:
77 if (usb_endpoint_xfer_int(desc) ||
78 usb_endpoint_xfer_isoc(desc)) {
79 interval = desc->bInterval;
80 interval = clamp_val(interval, 1, 16);
81 if (usb_endpoint_xfer_isoc(desc) && comp_desc)
82 mult = comp_desc->bmAttributes;
83 }
84 if (comp_desc)
85 burst = comp_desc->bMaxBurst;
86
87 break;
88 case USB_SPEED_HIGH:
89 if (usb_endpoint_xfer_isoc(desc) ||
90 usb_endpoint_xfer_int(desc)) {
91 interval = desc->bInterval;
92 interval = clamp_val(interval, 1, 16);
93 mult = usb_endpoint_maxp_mult(desc) - 1;
94 }
95 break;
96 case USB_SPEED_FULL:
97 if (usb_endpoint_xfer_isoc(desc))
98 interval = clamp_val(desc->bInterval, 1, 16);
99 else if (usb_endpoint_xfer_int(desc))
100 interval = clamp_val(desc->bInterval, 1, 255);
101
102 break;
103 default:
104 break; /*others are ignored */
105 }
106
107 dev_dbg(mtu->dev, "%s maxp:%d, interval:%d, burst:%d, mult:%d\n",
108 __func__, mep->maxp, interval, burst, mult);
109
110 mep->ep.maxpacket = mep->maxp;
111 mep->ep.desc = desc;
112 mep->ep.comp_desc = comp_desc;
113
114 /* slot mainly affects bulk/isoc transfer, so ignore int */
115 mep->slot = usb_endpoint_xfer_int(desc) ? 0 : mtu->slot;
116
117 ret = mtu3_config_ep(mtu, mep, interval, burst, mult);
118 if (ret < 0)
119 return ret;
120
121 ret = mtu3_gpd_ring_alloc(mep);
122 if (ret < 0) {
123 mtu3_deconfig_ep(mtu, mep);
124 return ret;
125 }
126
127 mtu3_qmu_start(mep);
128
129 return 0;
130}
131
132static int mtu3_ep_disable(struct mtu3_ep *mep)
133{
134 struct mtu3 *mtu = mep->mtu;
135
136 /* abort all pending requests */
137 nuke(mep, -ESHUTDOWN);
138 mtu3_qmu_stop(mep);
139 mtu3_deconfig_ep(mtu, mep);
140 mtu3_gpd_ring_free(mep);
141
142 mep->desc = NULL;
143 mep->ep.desc = NULL;
144 mep->comp_desc = NULL;
145 mep->type = 0;
146 mep->flags = 0;
147
148 return 0;
149}
150
151static int mtu3_gadget_ep_enable(struct usb_ep *ep,
152 const struct usb_endpoint_descriptor *desc)
153{
154 struct mtu3_ep *mep;
155 struct mtu3 *mtu;
156 unsigned long flags;
157 int ret = -EINVAL;
158
159 if (!ep || !desc || desc->bDescriptorType != USB_DT_ENDPOINT) {
160 pr_debug("%s invalid parameters\n", __func__);
161 return -EINVAL;
162 }
163
164 if (!desc->wMaxPacketSize) {
165 pr_debug("%s missing wMaxPacketSize\n", __func__);
166 return -EINVAL;
167 }
168 mep = to_mtu3_ep(ep);
169 mtu = mep->mtu;
170
171 /* check ep number and direction against endpoint */
172 if (usb_endpoint_num(desc) != mep->epnum)
173 return -EINVAL;
174
175 if (!!usb_endpoint_dir_in(desc) ^ !!mep->is_in)
176 return -EINVAL;
177
178 dev_dbg(mtu->dev, "%s %s\n", __func__, ep->name);
179
180 if (mep->flags & MTU3_EP_ENABLED) {
181 dev_WARN_ONCE(mtu->dev, true, "%s is already enabled\n",
182 mep->name);
183 return 0;
184 }
185
186 spin_lock_irqsave(&mtu->lock, flags);
187 mep->desc = desc;
188 mep->comp_desc = ep->comp_desc;
189
190 ret = mtu3_ep_enable(mep);
191 if (ret)
192 goto error;
193
194 mep->flags = MTU3_EP_ENABLED;
195 mtu->active_ep++;
196
197error:
198 spin_unlock_irqrestore(&mtu->lock, flags);
199
200 dev_dbg(mtu->dev, "%s active_ep=%d\n", __func__, mtu->active_ep);
201 trace_mtu3_gadget_ep_enable(mep);
202
203 return ret;
204}
205
206static int mtu3_gadget_ep_disable(struct usb_ep *ep)
207{
208 struct mtu3_ep *mep = to_mtu3_ep(ep);
209 struct mtu3 *mtu = mep->mtu;
210 unsigned long flags;
211
212 dev_dbg(mtu->dev, "%s %s\n", __func__, mep->name);
213 trace_mtu3_gadget_ep_disable(mep);
214
215 if (!(mep->flags & MTU3_EP_ENABLED)) {
216 dev_warn(mtu->dev, "%s is already disabled\n", mep->name);
217 return 0;
218 }
219
220 spin_lock_irqsave(&mtu->lock, flags);
221 mtu3_ep_disable(mep);
222 mep->flags = 0;
223 mtu->active_ep--;
224 spin_unlock_irqrestore(&(mtu->lock), flags);
225
226 dev_dbg(mtu->dev, "%s active_ep=%d, mtu3 is_active=%d\n",
227 __func__, mtu->active_ep, mtu->is_active);
228
229 return 0;
230}
231
232struct usb_request *mtu3_alloc_request(struct usb_ep *ep, gfp_t gfp_flags)
233{
234 struct mtu3_ep *mep = to_mtu3_ep(ep);
235 struct mtu3_request *mreq;
236
237 mreq = kzalloc(sizeof(*mreq), gfp_flags);
238 if (!mreq)
239 return NULL;
240
241 mreq->request.dma = DMA_ADDR_INVALID;
242 mreq->epnum = mep->epnum;
243 mreq->mep = mep;
244 INIT_LIST_HEAD(&mreq->list);
245 trace_mtu3_alloc_request(mreq);
246
247 return &mreq->request;
248}
249
250void mtu3_free_request(struct usb_ep *ep, struct usb_request *req)
251{
252 struct mtu3_request *mreq = to_mtu3_request(req);
253
254 trace_mtu3_free_request(mreq);
255 kfree(mreq);
256}
257
258static int mtu3_gadget_queue(struct usb_ep *ep,
259 struct usb_request *req, gfp_t gfp_flags)
260{
261 struct mtu3_ep *mep = to_mtu3_ep(ep);
262 struct mtu3_request *mreq = to_mtu3_request(req);
263 struct mtu3 *mtu = mep->mtu;
264 unsigned long flags;
265 int ret = 0;
266
267 if (!req->buf)
268 return -ENODATA;
269
270 if (mreq->mep != mep)
271 return -EINVAL;
272
273 dev_dbg(mtu->dev, "%s %s EP%d(%s), req=%p, maxp=%d, len#%d\n",
274 __func__, mep->is_in ? "TX" : "RX", mreq->epnum, ep->name,
275 mreq, ep->maxpacket, mreq->request.length);
276
277 if (req->length > GPD_BUF_SIZE ||
278 (mtu->gen2cp && req->length > GPD_BUF_SIZE_EL)) {
279 dev_warn(mtu->dev,
280 "req length > supported MAX:%d requested:%d\n",
281 mtu->gen2cp ? GPD_BUF_SIZE_EL : GPD_BUF_SIZE,
282 req->length);
283 return -EOPNOTSUPP;
284 }
285
286 /* don't queue if the ep is down */
287 if (!mep->desc) {
288 dev_dbg(mtu->dev, "req=%p queued to %s while it's disabled\n",
289 req, ep->name);
290 return -ESHUTDOWN;
291 }
292
293 mreq->mtu = mtu;
294 mreq->request.actual = 0;
295 mreq->request.status = -EINPROGRESS;
296
297 ret = usb_gadget_map_request(&mtu->g, req, mep->is_in);
298 if (ret) {
299 dev_err(mtu->dev, "dma mapping failed\n");
300 return ret;
301 }
302
303 spin_lock_irqsave(&mtu->lock, flags);
304
305 if (mtu3_prepare_transfer(mep)) {
306 ret = -EAGAIN;
307 goto error;
308 }
309
310 list_add_tail(&mreq->list, &mep->req_list);
311 mtu3_insert_gpd(mep, mreq);
312 mtu3_qmu_resume(mep);
313
314error:
315 spin_unlock_irqrestore(&mtu->lock, flags);
316 trace_mtu3_gadget_queue(mreq);
317
318 return ret;
319}
320
321static int mtu3_gadget_dequeue(struct usb_ep *ep, struct usb_request *req)
322{
323 struct mtu3_ep *mep = to_mtu3_ep(ep);
324 struct mtu3_request *mreq = to_mtu3_request(req);
325 struct mtu3_request *r;
326 struct mtu3 *mtu = mep->mtu;
327 unsigned long flags;
328 int ret = 0;
329
330 if (mreq->mep != mep)
331 return -EINVAL;
332
333 dev_dbg(mtu->dev, "%s : req=%p\n", __func__, req);
334 trace_mtu3_gadget_dequeue(mreq);
335
336 spin_lock_irqsave(&mtu->lock, flags);
337
338 list_for_each_entry(r, &mep->req_list, list) {
339 if (r == mreq)
340 break;
341 }
342 if (r != mreq) {
343 dev_dbg(mtu->dev, "req=%p not queued to %s\n", req, ep->name);
344 ret = -EINVAL;
345 goto done;
346 }
347
348 mtu3_qmu_flush(mep); /* REVISIT: set BPS ?? */
349 mtu3_req_complete(mep, req, -ECONNRESET);
350 mtu3_qmu_start(mep);
351
352done:
353 spin_unlock_irqrestore(&mtu->lock, flags);
354
355 return ret;
356}
357
358/*
359 * Set or clear the halt bit of an EP.
360 * A halted EP won't TX/RX any data but will queue requests.
361 */
362static int mtu3_gadget_ep_set_halt(struct usb_ep *ep, int value)
363{
364 struct mtu3_ep *mep = to_mtu3_ep(ep);
365 struct mtu3 *mtu = mep->mtu;
366 struct mtu3_request *mreq;
367 unsigned long flags;
368 int ret = 0;
369
370 dev_dbg(mtu->dev, "%s : %s...", __func__, ep->name);
371
372 spin_lock_irqsave(&mtu->lock, flags);
373
374 if (mep->type == USB_ENDPOINT_XFER_ISOC) {
375 ret = -EINVAL;
376 goto done;
377 }
378
379 mreq = next_request(mep);
380 if (value) {
381 /*
382 * If there is not request for TX-EP, QMU will not transfer
383 * data to TX-FIFO, so no need check whether TX-FIFO
384 * holds bytes or not here
385 */
386 if (mreq) {
387 dev_dbg(mtu->dev, "req in progress, cannot halt %s\n",
388 ep->name);
389 ret = -EAGAIN;
390 goto done;
391 }
392 } else {
393 mep->flags &= ~MTU3_EP_WEDGE;
394 }
395
396 dev_dbg(mtu->dev, "%s %s stall\n", ep->name, value ? "set" : "clear");
397
398 mtu3_ep_stall_set(mep, value);
399
400done:
401 spin_unlock_irqrestore(&mtu->lock, flags);
402 trace_mtu3_gadget_ep_set_halt(mep);
403
404 return ret;
405}
406
407/* Sets the halt feature with the clear requests ignored */
408static int mtu3_gadget_ep_set_wedge(struct usb_ep *ep)
409{
410 struct mtu3_ep *mep = to_mtu3_ep(ep);
411
412 mep->flags |= MTU3_EP_WEDGE;
413
414 return usb_ep_set_halt(ep);
415}
416
417static const struct usb_ep_ops mtu3_ep_ops = {
418 .enable = mtu3_gadget_ep_enable,
419 .disable = mtu3_gadget_ep_disable,
420 .alloc_request = mtu3_alloc_request,
421 .free_request = mtu3_free_request,
422 .queue = mtu3_gadget_queue,
423 .dequeue = mtu3_gadget_dequeue,
424 .set_halt = mtu3_gadget_ep_set_halt,
425 .set_wedge = mtu3_gadget_ep_set_wedge,
426};
427
428static int mtu3_gadget_get_frame(struct usb_gadget *gadget)
429{
430 struct mtu3 *mtu = gadget_to_mtu3(gadget);
431
432 return (int)mtu3_readl(mtu->mac_base, U3D_USB20_FRAME_NUM);
433}
434
435static void function_wake_notif(struct mtu3 *mtu, u8 intf)
436{
437 mtu3_writel(mtu->mac_base, U3D_DEV_NOTIF_0,
438 TYPE_FUNCTION_WAKE | DEV_NOTIF_VAL_FW(intf));
439 mtu3_setbits(mtu->mac_base, U3D_DEV_NOTIF_0, SEND_DEV_NOTIF);
440}
441
442static int mtu3_gadget_wakeup(struct usb_gadget *gadget)
443{
444 struct mtu3 *mtu = gadget_to_mtu3(gadget);
445 unsigned long flags;
446
447 dev_dbg(mtu->dev, "%s\n", __func__);
448
449 /* remote wakeup feature is not enabled by host */
450 if (!mtu->may_wakeup)
451 return -EOPNOTSUPP;
452
453 spin_lock_irqsave(&mtu->lock, flags);
454 if (mtu->g.speed >= USB_SPEED_SUPER) {
455 /*
456 * class driver may do function wakeup even UFP is in U0,
457 * and UX_EXIT only takes effect in U1/U2/U3;
458 */
459 mtu3_setbits(mtu->mac_base, U3D_LINK_POWER_CONTROL, UX_EXIT);
460 /*
461 * Assume there's only one function on the composite device
462 * and enable remote wake for the first interface.
463 * FIXME if the IAD (interface association descriptor) shows
464 * there is more than one function.
465 */
466 function_wake_notif(mtu, 0);
467 } else {
468 mtu3_setbits(mtu->mac_base, U3D_POWER_MANAGEMENT, RESUME);
469 spin_unlock_irqrestore(&mtu->lock, flags);
470 usleep_range(10000, 11000);
471 spin_lock_irqsave(&mtu->lock, flags);
472 mtu3_clrbits(mtu->mac_base, U3D_POWER_MANAGEMENT, RESUME);
473 }
474 spin_unlock_irqrestore(&mtu->lock, flags);
475 return 0;
476}
477
478static int mtu3_gadget_set_self_powered(struct usb_gadget *gadget,
479 int is_selfpowered)
480{
481 struct mtu3 *mtu = gadget_to_mtu3(gadget);
482
483 mtu->is_self_powered = !!is_selfpowered;
484 return 0;
485}
486
487static int mtu3_gadget_pullup(struct usb_gadget *gadget, int is_on)
488{
489 struct mtu3 *mtu = gadget_to_mtu3(gadget);
490 unsigned long flags;
491
492 dev_dbg(mtu->dev, "%s (%s) for %sactive device\n", __func__,
493 is_on ? "on" : "off", mtu->is_active ? "" : "in");
494
495 pm_runtime_get_sync(mtu->dev);
496
497 /* we'd rather not pullup unless the device is active. */
498 spin_lock_irqsave(&mtu->lock, flags);
499
500 is_on = !!is_on;
501 if (!mtu->is_active) {
502 /* save it for mtu3_start() to process the request */
503 mtu->softconnect = is_on;
504 } else if (is_on != mtu->softconnect) {
505 mtu->softconnect = is_on;
506 mtu3_dev_on_off(mtu, is_on);
507 }
508
509 spin_unlock_irqrestore(&mtu->lock, flags);
510 pm_runtime_put(mtu->dev);
511
512 return 0;
513}
514
515static int mtu3_gadget_start(struct usb_gadget *gadget,
516 struct usb_gadget_driver *driver)
517{
518 struct mtu3 *mtu = gadget_to_mtu3(gadget);
519 unsigned long flags;
520
521 if (mtu->gadget_driver) {
522 dev_err(mtu->dev, "%s is already bound to %s\n",
523 mtu->g.name, mtu->gadget_driver->driver.name);
524 return -EBUSY;
525 }
526
527 dev_dbg(mtu->dev, "bind driver %s\n", driver->function);
528 pm_runtime_get_sync(mtu->dev);
529
530 spin_lock_irqsave(&mtu->lock, flags);
531
532 mtu->softconnect = 0;
533 mtu->gadget_driver = driver;
534
535 if (mtu->ssusb->dr_mode == USB_DR_MODE_PERIPHERAL)
536 mtu3_start(mtu);
537
538 spin_unlock_irqrestore(&mtu->lock, flags);
539 pm_runtime_put(mtu->dev);
540
541 return 0;
542}
543
544static void stop_activity(struct mtu3 *mtu)
545{
546 struct usb_gadget_driver *driver = mtu->gadget_driver;
547 int i;
548
549 /* don't disconnect if it's not connected */
550 if (mtu->g.speed == USB_SPEED_UNKNOWN)
551 driver = NULL;
552 else
553 mtu->g.speed = USB_SPEED_UNKNOWN;
554
555 /* deactivate the hardware */
556 if (mtu->softconnect) {
557 mtu->softconnect = 0;
558 mtu3_dev_on_off(mtu, 0);
559 }
560
561 /*
562 * killing any outstanding requests will quiesce the driver;
563 * then report disconnect
564 */
565 nuke(mtu->ep0, -ESHUTDOWN);
566 for (i = 1; i < mtu->num_eps; i++) {
567 nuke(mtu->in_eps + i, -ESHUTDOWN);
568 nuke(mtu->out_eps + i, -ESHUTDOWN);
569 }
570
571 if (driver) {
572 spin_unlock(&mtu->lock);
573 driver->disconnect(&mtu->g);
574 spin_lock(&mtu->lock);
575 }
576}
577
578static int mtu3_gadget_stop(struct usb_gadget *g)
579{
580 struct mtu3 *mtu = gadget_to_mtu3(g);
581 unsigned long flags;
582
583 dev_dbg(mtu->dev, "%s\n", __func__);
584
585 spin_lock_irqsave(&mtu->lock, flags);
586
587 stop_activity(mtu);
588 mtu->gadget_driver = NULL;
589
590 if (mtu->ssusb->dr_mode == USB_DR_MODE_PERIPHERAL)
591 mtu3_stop(mtu);
592
593 spin_unlock_irqrestore(&mtu->lock, flags);
594
595 synchronize_irq(mtu->irq);
596 return 0;
597}
598
599static void
600mtu3_gadget_set_speed(struct usb_gadget *g, enum usb_device_speed speed)
601{
602 struct mtu3 *mtu = gadget_to_mtu3(g);
603 unsigned long flags;
604
605 dev_dbg(mtu->dev, "%s %s\n", __func__, usb_speed_string(speed));
606
607 spin_lock_irqsave(&mtu->lock, flags);
608 mtu->speed = speed;
609 spin_unlock_irqrestore(&mtu->lock, flags);
610}
611
612static void mtu3_gadget_async_callbacks(struct usb_gadget *g, bool enable)
613{
614 struct mtu3 *mtu = gadget_to_mtu3(g);
615 unsigned long flags;
616
617 dev_dbg(mtu->dev, "%s %s\n", __func__, enable ? "en" : "dis");
618
619 spin_lock_irqsave(&mtu->lock, flags);
620 mtu->async_callbacks = enable;
621 spin_unlock_irqrestore(&mtu->lock, flags);
622}
623
624static const struct usb_gadget_ops mtu3_gadget_ops = {
625 .get_frame = mtu3_gadget_get_frame,
626 .wakeup = mtu3_gadget_wakeup,
627 .set_selfpowered = mtu3_gadget_set_self_powered,
628 .pullup = mtu3_gadget_pullup,
629 .udc_start = mtu3_gadget_start,
630 .udc_stop = mtu3_gadget_stop,
631 .udc_set_speed = mtu3_gadget_set_speed,
632 .udc_async_callbacks = mtu3_gadget_async_callbacks,
633};
634
635static void mtu3_state_reset(struct mtu3 *mtu)
636{
637 mtu->address = 0;
638 mtu->ep0_state = MU3D_EP0_STATE_SETUP;
639 mtu->may_wakeup = 0;
640 mtu->u1_enable = 0;
641 mtu->u2_enable = 0;
642 mtu->delayed_status = false;
643 mtu->test_mode = false;
644}
645
646static void init_hw_ep(struct mtu3 *mtu, struct mtu3_ep *mep,
647 u32 epnum, u32 is_in)
648{
649 mep->epnum = epnum;
650 mep->mtu = mtu;
651 mep->is_in = is_in;
652
653 INIT_LIST_HEAD(&mep->req_list);
654
655 sprintf(mep->name, "ep%d%s", epnum,
656 !epnum ? "" : (is_in ? "in" : "out"));
657
658 mep->ep.name = mep->name;
659 INIT_LIST_HEAD(&mep->ep.ep_list);
660
661 /* initialize maxpacket as SS */
662 if (!epnum) {
663 usb_ep_set_maxpacket_limit(&mep->ep, 512);
664 mep->ep.caps.type_control = true;
665 mep->ep.ops = &mtu3_ep0_ops;
666 mtu->g.ep0 = &mep->ep;
667 } else {
668 usb_ep_set_maxpacket_limit(&mep->ep, 1024);
669 mep->ep.caps.type_iso = true;
670 mep->ep.caps.type_bulk = true;
671 mep->ep.caps.type_int = true;
672 mep->ep.ops = &mtu3_ep_ops;
673 list_add_tail(&mep->ep.ep_list, &mtu->g.ep_list);
674 }
675
676 dev_dbg(mtu->dev, "%s, name=%s, maxp=%d\n", __func__, mep->ep.name,
677 mep->ep.maxpacket);
678
679 if (!epnum) {
680 mep->ep.caps.dir_in = true;
681 mep->ep.caps.dir_out = true;
682 } else if (is_in) {
683 mep->ep.caps.dir_in = true;
684 } else {
685 mep->ep.caps.dir_out = true;
686 }
687}
688
689static void mtu3_gadget_init_eps(struct mtu3 *mtu)
690{
691 u8 epnum;
692
693 /* initialize endpoint list just once */
694 INIT_LIST_HEAD(&(mtu->g.ep_list));
695
696 dev_dbg(mtu->dev, "%s num_eps(1 for a pair of tx&rx ep)=%d\n",
697 __func__, mtu->num_eps);
698
699 init_hw_ep(mtu, mtu->ep0, 0, 0);
700 for (epnum = 1; epnum < mtu->num_eps; epnum++) {
701 init_hw_ep(mtu, mtu->in_eps + epnum, epnum, 1);
702 init_hw_ep(mtu, mtu->out_eps + epnum, epnum, 0);
703 }
704}
705
706int mtu3_gadget_setup(struct mtu3 *mtu)
707{
708 mtu->g.ops = &mtu3_gadget_ops;
709 mtu->g.max_speed = mtu->max_speed;
710 mtu->g.speed = USB_SPEED_UNKNOWN;
711 mtu->g.sg_supported = 0;
712 mtu->g.name = MTU3_DRIVER_NAME;
713 mtu->g.irq = mtu->irq;
714 mtu->is_active = 0;
715 mtu->delayed_status = false;
716
717 mtu3_gadget_init_eps(mtu);
718
719 return usb_add_gadget_udc(mtu->dev, &mtu->g);
720}
721
722void mtu3_gadget_cleanup(struct mtu3 *mtu)
723{
724 usb_del_gadget_udc(&mtu->g);
725}
726
727void mtu3_gadget_resume(struct mtu3 *mtu)
728{
729 dev_dbg(mtu->dev, "gadget RESUME\n");
730 if (mtu->async_callbacks && mtu->gadget_driver && mtu->gadget_driver->resume) {
731 spin_unlock(&mtu->lock);
732 mtu->gadget_driver->resume(&mtu->g);
733 spin_lock(&mtu->lock);
734 }
735}
736
737/* called when SOF packets stop for 3+ msec or enters U3 */
738void mtu3_gadget_suspend(struct mtu3 *mtu)
739{
740 dev_dbg(mtu->dev, "gadget SUSPEND\n");
741 if (mtu->async_callbacks && mtu->gadget_driver && mtu->gadget_driver->suspend) {
742 spin_unlock(&mtu->lock);
743 mtu->gadget_driver->suspend(&mtu->g);
744 spin_lock(&mtu->lock);
745 }
746}
747
748/* called when VBUS drops below session threshold, and in other cases */
749void mtu3_gadget_disconnect(struct mtu3 *mtu)
750{
751 dev_dbg(mtu->dev, "gadget DISCONNECT\n");
752 if (mtu->async_callbacks && mtu->gadget_driver && mtu->gadget_driver->disconnect) {
753 spin_unlock(&mtu->lock);
754 mtu->gadget_driver->disconnect(&mtu->g);
755 spin_lock(&mtu->lock);
756 }
757
758 mtu3_state_reset(mtu);
759 usb_gadget_set_state(&mtu->g, USB_STATE_NOTATTACHED);
760}
761
762void mtu3_gadget_reset(struct mtu3 *mtu)
763{
764 dev_dbg(mtu->dev, "gadget RESET\n");
765
766 /* report disconnect, if we didn't flush EP state */
767 if (mtu->g.speed != USB_SPEED_UNKNOWN)
768 mtu3_gadget_disconnect(mtu);
769 else
770 mtu3_state_reset(mtu);
771}
1/*
2 * mtu3_gadget.c - MediaTek usb3 DRD peripheral support
3 *
4 * Copyright (C) 2016 MediaTek Inc.
5 *
6 * Author: Chunfeng Yun <chunfeng.yun@mediatek.com>
7 *
8 * This software is licensed under the terms of the GNU General Public
9 * License version 2, as published by the Free Software Foundation, and
10 * may be copied, distributed, and modified under those terms.
11 *
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 * GNU General Public License for more details.
16 *
17 */
18
19#include "mtu3.h"
20
21void mtu3_req_complete(struct mtu3_ep *mep,
22 struct usb_request *req, int status)
23__releases(mep->mtu->lock)
24__acquires(mep->mtu->lock)
25{
26 struct mtu3_request *mreq;
27 struct mtu3 *mtu;
28 int busy = mep->busy;
29
30 mreq = to_mtu3_request(req);
31 list_del(&mreq->list);
32 if (mreq->request.status == -EINPROGRESS)
33 mreq->request.status = status;
34
35 mtu = mreq->mtu;
36 mep->busy = 1;
37 spin_unlock(&mtu->lock);
38
39 /* ep0 makes use of PIO, needn't unmap it */
40 if (mep->epnum)
41 usb_gadget_unmap_request(&mtu->g, req, mep->is_in);
42
43 dev_dbg(mtu->dev, "%s complete req: %p, sts %d, %d/%d\n", mep->name,
44 req, req->status, mreq->request.actual, mreq->request.length);
45
46 usb_gadget_giveback_request(&mep->ep, &mreq->request);
47
48 spin_lock(&mtu->lock);
49 mep->busy = busy;
50}
51
52static void nuke(struct mtu3_ep *mep, const int status)
53{
54 struct mtu3_request *mreq = NULL;
55
56 mep->busy = 1;
57 if (list_empty(&mep->req_list))
58 return;
59
60 dev_dbg(mep->mtu->dev, "abort %s's req: sts %d\n", mep->name, status);
61
62 /* exclude EP0 */
63 if (mep->epnum)
64 mtu3_qmu_flush(mep);
65
66 while (!list_empty(&mep->req_list)) {
67 mreq = list_first_entry(&mep->req_list,
68 struct mtu3_request, list);
69 mtu3_req_complete(mep, &mreq->request, status);
70 }
71}
72
73static int mtu3_ep_enable(struct mtu3_ep *mep)
74{
75 const struct usb_endpoint_descriptor *desc;
76 const struct usb_ss_ep_comp_descriptor *comp_desc;
77 struct mtu3 *mtu = mep->mtu;
78 u32 interval = 0;
79 u32 mult = 0;
80 u32 burst = 0;
81 int max_packet;
82 int ret;
83
84 desc = mep->desc;
85 comp_desc = mep->comp_desc;
86 mep->type = usb_endpoint_type(desc);
87 max_packet = usb_endpoint_maxp(desc);
88 mep->maxp = max_packet & GENMASK(10, 0);
89
90 switch (mtu->g.speed) {
91 case USB_SPEED_SUPER:
92 if (usb_endpoint_xfer_int(desc) ||
93 usb_endpoint_xfer_isoc(desc)) {
94 interval = desc->bInterval;
95 interval = clamp_val(interval, 1, 16) - 1;
96 if (usb_endpoint_xfer_isoc(desc) && comp_desc)
97 mult = comp_desc->bmAttributes;
98 }
99 if (comp_desc)
100 burst = comp_desc->bMaxBurst;
101
102 break;
103 case USB_SPEED_HIGH:
104 if (usb_endpoint_xfer_isoc(desc) ||
105 usb_endpoint_xfer_int(desc)) {
106 interval = desc->bInterval;
107 interval = clamp_val(interval, 1, 16) - 1;
108 burst = (max_packet & GENMASK(12, 11)) >> 11;
109 }
110 break;
111 default:
112 break; /*others are ignored */
113 }
114
115 dev_dbg(mtu->dev, "%s maxp:%d, interval:%d, burst:%d, mult:%d\n",
116 __func__, mep->maxp, interval, burst, mult);
117
118 mep->ep.maxpacket = mep->maxp;
119 mep->ep.desc = desc;
120 mep->ep.comp_desc = comp_desc;
121
122 /* slot mainly affects bulk/isoc transfer, so ignore int */
123 mep->slot = usb_endpoint_xfer_int(desc) ? 0 : mtu->slot;
124
125 ret = mtu3_config_ep(mtu, mep, interval, burst, mult);
126 if (ret < 0)
127 return ret;
128
129 ret = mtu3_gpd_ring_alloc(mep);
130 if (ret < 0) {
131 mtu3_deconfig_ep(mtu, mep);
132 return ret;
133 }
134
135 mtu3_qmu_start(mep);
136
137 return 0;
138}
139
140static int mtu3_ep_disable(struct mtu3_ep *mep)
141{
142 struct mtu3 *mtu = mep->mtu;
143
144 mtu3_qmu_stop(mep);
145
146 /* abort all pending requests */
147 nuke(mep, -ESHUTDOWN);
148 mtu3_deconfig_ep(mtu, mep);
149 mtu3_gpd_ring_free(mep);
150
151 mep->desc = NULL;
152 mep->ep.desc = NULL;
153 mep->comp_desc = NULL;
154 mep->type = 0;
155 mep->flags = 0;
156
157 return 0;
158}
159
160static int mtu3_gadget_ep_enable(struct usb_ep *ep,
161 const struct usb_endpoint_descriptor *desc)
162{
163 struct mtu3_ep *mep;
164 struct mtu3 *mtu;
165 unsigned long flags;
166 int ret = -EINVAL;
167
168 if (!ep || !desc || desc->bDescriptorType != USB_DT_ENDPOINT) {
169 pr_debug("%s invalid parameters\n", __func__);
170 return -EINVAL;
171 }
172
173 if (!desc->wMaxPacketSize) {
174 pr_debug("%s missing wMaxPacketSize\n", __func__);
175 return -EINVAL;
176 }
177 mep = to_mtu3_ep(ep);
178 mtu = mep->mtu;
179
180 /* check ep number and direction against endpoint */
181 if (usb_endpoint_num(desc) != mep->epnum)
182 return -EINVAL;
183
184 if (!!usb_endpoint_dir_in(desc) ^ !!mep->is_in)
185 return -EINVAL;
186
187 dev_dbg(mtu->dev, "%s %s\n", __func__, ep->name);
188
189 if (mep->flags & MTU3_EP_ENABLED) {
190 dev_WARN_ONCE(mtu->dev, true, "%s is already enabled\n",
191 mep->name);
192 return 0;
193 }
194
195 spin_lock_irqsave(&mtu->lock, flags);
196 mep->desc = desc;
197 mep->comp_desc = ep->comp_desc;
198
199 ret = mtu3_ep_enable(mep);
200 if (ret)
201 goto error;
202
203 mep->busy = 0;
204 mep->wedged = 0;
205 mep->flags |= MTU3_EP_ENABLED;
206 mtu->active_ep++;
207
208error:
209 spin_unlock_irqrestore(&mtu->lock, flags);
210
211 dev_dbg(mtu->dev, "%s active_ep=%d\n", __func__, mtu->active_ep);
212
213 return ret;
214}
215
216static int mtu3_gadget_ep_disable(struct usb_ep *ep)
217{
218 struct mtu3_ep *mep = to_mtu3_ep(ep);
219 struct mtu3 *mtu = mep->mtu;
220 unsigned long flags;
221
222 dev_dbg(mtu->dev, "%s %s\n", __func__, mep->name);
223
224 if (!(mep->flags & MTU3_EP_ENABLED)) {
225 dev_warn(mtu->dev, "%s is already disabled\n", mep->name);
226 return 0;
227 }
228
229 spin_lock_irqsave(&mtu->lock, flags);
230 mtu3_ep_disable(mep);
231 mep->flags &= ~MTU3_EP_ENABLED;
232 mtu->active_ep--;
233 spin_unlock_irqrestore(&(mtu->lock), flags);
234
235 dev_dbg(mtu->dev, "%s active_ep=%d, mtu3 is_active=%d\n",
236 __func__, mtu->active_ep, mtu->is_active);
237
238 return 0;
239}
240
241struct usb_request *mtu3_alloc_request(struct usb_ep *ep, gfp_t gfp_flags)
242{
243 struct mtu3_ep *mep = to_mtu3_ep(ep);
244 struct mtu3_request *mreq;
245
246 mreq = kzalloc(sizeof(*mreq), gfp_flags);
247 if (!mreq)
248 return NULL;
249
250 mreq->request.dma = DMA_ADDR_INVALID;
251 mreq->epnum = mep->epnum;
252 mreq->mep = mep;
253
254 return &mreq->request;
255}
256
257void mtu3_free_request(struct usb_ep *ep, struct usb_request *req)
258{
259 kfree(to_mtu3_request(req));
260}
261
262static int mtu3_gadget_queue(struct usb_ep *ep,
263 struct usb_request *req, gfp_t gfp_flags)
264{
265 struct mtu3_ep *mep;
266 struct mtu3_request *mreq;
267 struct mtu3 *mtu;
268 unsigned long flags;
269 int ret = 0;
270
271 if (!ep || !req)
272 return -EINVAL;
273
274 if (!req->buf)
275 return -ENODATA;
276
277 mep = to_mtu3_ep(ep);
278 mtu = mep->mtu;
279 mreq = to_mtu3_request(req);
280 mreq->mtu = mtu;
281
282 if (mreq->mep != mep)
283 return -EINVAL;
284
285 dev_dbg(mtu->dev, "%s %s EP%d(%s), req=%p, maxp=%d, len#%d\n",
286 __func__, mep->is_in ? "TX" : "RX", mreq->epnum, ep->name,
287 mreq, ep->maxpacket, mreq->request.length);
288
289 if (req->length > GPD_BUF_SIZE) {
290 dev_warn(mtu->dev,
291 "req length > supported MAX:%d requested:%d\n",
292 GPD_BUF_SIZE, req->length);
293 return -EOPNOTSUPP;
294 }
295
296 /* don't queue if the ep is down */
297 if (!mep->desc) {
298 dev_dbg(mtu->dev, "req=%p queued to %s while it's disabled\n",
299 req, ep->name);
300 return -ESHUTDOWN;
301 }
302
303 mreq->request.actual = 0;
304 mreq->request.status = -EINPROGRESS;
305
306 ret = usb_gadget_map_request(&mtu->g, req, mep->is_in);
307 if (ret) {
308 dev_err(mtu->dev, "dma mapping failed\n");
309 return ret;
310 }
311
312 spin_lock_irqsave(&mtu->lock, flags);
313
314 if (mtu3_prepare_transfer(mep)) {
315 ret = -EAGAIN;
316 goto error;
317 }
318
319 list_add_tail(&mreq->list, &mep->req_list);
320 mtu3_insert_gpd(mep, mreq);
321 mtu3_qmu_resume(mep);
322
323error:
324 spin_unlock_irqrestore(&mtu->lock, flags);
325
326 return ret;
327}
328
329static int mtu3_gadget_dequeue(struct usb_ep *ep, struct usb_request *req)
330{
331 struct mtu3_ep *mep = to_mtu3_ep(ep);
332 struct mtu3_request *mreq = to_mtu3_request(req);
333 struct mtu3_request *r;
334 unsigned long flags;
335 int ret = 0;
336 struct mtu3 *mtu = mep->mtu;
337
338 if (!ep || !req || mreq->mep != mep)
339 return -EINVAL;
340
341 dev_dbg(mtu->dev, "%s : req=%p\n", __func__, req);
342
343 spin_lock_irqsave(&mtu->lock, flags);
344
345 list_for_each_entry(r, &mep->req_list, list) {
346 if (r == mreq)
347 break;
348 }
349 if (r != mreq) {
350 dev_dbg(mtu->dev, "req=%p not queued to %s\n", req, ep->name);
351 ret = -EINVAL;
352 goto done;
353 }
354
355 mtu3_qmu_flush(mep); /* REVISIT: set BPS ?? */
356 mtu3_req_complete(mep, req, -ECONNRESET);
357 mtu3_qmu_start(mep);
358
359done:
360 spin_unlock_irqrestore(&mtu->lock, flags);
361
362 return ret;
363}
364
365/*
366 * Set or clear the halt bit of an EP.
367 * A halted EP won't TX/RX any data but will queue requests.
368 */
369static int mtu3_gadget_ep_set_halt(struct usb_ep *ep, int value)
370{
371 struct mtu3_ep *mep = to_mtu3_ep(ep);
372 struct mtu3 *mtu = mep->mtu;
373 struct mtu3_request *mreq;
374 unsigned long flags;
375 int ret = 0;
376
377 if (!ep)
378 return -EINVAL;
379
380 dev_dbg(mtu->dev, "%s : %s...", __func__, ep->name);
381
382 spin_lock_irqsave(&mtu->lock, flags);
383
384 if (mep->type == USB_ENDPOINT_XFER_ISOC) {
385 ret = -EINVAL;
386 goto done;
387 }
388
389 mreq = next_request(mep);
390 if (value) {
391 /*
392 * If there is not request for TX-EP, QMU will not transfer
393 * data to TX-FIFO, so no need check whether TX-FIFO
394 * holds bytes or not here
395 */
396 if (mreq) {
397 dev_dbg(mtu->dev, "req in progress, cannot halt %s\n",
398 ep->name);
399 ret = -EAGAIN;
400 goto done;
401 }
402 } else {
403 mep->wedged = 0;
404 }
405
406 dev_dbg(mtu->dev, "%s %s stall\n", ep->name, value ? "set" : "clear");
407
408 mtu3_ep_stall_set(mep, value);
409
410done:
411 spin_unlock_irqrestore(&mtu->lock, flags);
412
413 return ret;
414}
415
416/* Sets the halt feature with the clear requests ignored */
417static int mtu3_gadget_ep_set_wedge(struct usb_ep *ep)
418{
419 struct mtu3_ep *mep = to_mtu3_ep(ep);
420
421 if (!ep)
422 return -EINVAL;
423
424 mep->wedged = 1;
425
426 return usb_ep_set_halt(ep);
427}
428
429static const struct usb_ep_ops mtu3_ep_ops = {
430 .enable = mtu3_gadget_ep_enable,
431 .disable = mtu3_gadget_ep_disable,
432 .alloc_request = mtu3_alloc_request,
433 .free_request = mtu3_free_request,
434 .queue = mtu3_gadget_queue,
435 .dequeue = mtu3_gadget_dequeue,
436 .set_halt = mtu3_gadget_ep_set_halt,
437 .set_wedge = mtu3_gadget_ep_set_wedge,
438};
439
440static int mtu3_gadget_get_frame(struct usb_gadget *gadget)
441{
442 struct mtu3 *mtu = gadget_to_mtu3(gadget);
443
444 return (int)mtu3_readl(mtu->mac_base, U3D_USB20_FRAME_NUM);
445}
446
447static int mtu3_gadget_wakeup(struct usb_gadget *gadget)
448{
449 struct mtu3 *mtu = gadget_to_mtu3(gadget);
450 unsigned long flags;
451
452 dev_dbg(mtu->dev, "%s\n", __func__);
453
454 /* remote wakeup feature is not enabled by host */
455 if (!mtu->may_wakeup)
456 return -EOPNOTSUPP;
457
458 spin_lock_irqsave(&mtu->lock, flags);
459 if (mtu->g.speed == USB_SPEED_SUPER) {
460 mtu3_setbits(mtu->mac_base, U3D_LINK_POWER_CONTROL, UX_EXIT);
461 } else {
462 mtu3_setbits(mtu->mac_base, U3D_POWER_MANAGEMENT, RESUME);
463 spin_unlock_irqrestore(&mtu->lock, flags);
464 usleep_range(10000, 11000);
465 spin_lock_irqsave(&mtu->lock, flags);
466 mtu3_clrbits(mtu->mac_base, U3D_POWER_MANAGEMENT, RESUME);
467 }
468 spin_unlock_irqrestore(&mtu->lock, flags);
469 return 0;
470}
471
472static int mtu3_gadget_set_self_powered(struct usb_gadget *gadget,
473 int is_selfpowered)
474{
475 struct mtu3 *mtu = gadget_to_mtu3(gadget);
476
477 mtu->is_self_powered = !!is_selfpowered;
478 return 0;
479}
480
481static int mtu3_gadget_pullup(struct usb_gadget *gadget, int is_on)
482{
483 struct mtu3 *mtu = gadget_to_mtu3(gadget);
484 unsigned long flags;
485
486 dev_dbg(mtu->dev, "%s (%s) for %sactive device\n", __func__,
487 is_on ? "on" : "off", mtu->is_active ? "" : "in");
488
489 /* we'd rather not pullup unless the device is active. */
490 spin_lock_irqsave(&mtu->lock, flags);
491
492 is_on = !!is_on;
493 if (!mtu->is_active) {
494 /* save it for mtu3_start() to process the request */
495 mtu->softconnect = is_on;
496 } else if (is_on != mtu->softconnect) {
497 mtu->softconnect = is_on;
498 mtu3_dev_on_off(mtu, is_on);
499 }
500
501 spin_unlock_irqrestore(&mtu->lock, flags);
502
503 return 0;
504}
505
506static int mtu3_gadget_start(struct usb_gadget *gadget,
507 struct usb_gadget_driver *driver)
508{
509 struct mtu3 *mtu = gadget_to_mtu3(gadget);
510 unsigned long flags;
511
512 if (mtu->gadget_driver) {
513 dev_err(mtu->dev, "%s is already bound to %s\n",
514 mtu->g.name, mtu->gadget_driver->driver.name);
515 return -EBUSY;
516 }
517
518 dev_dbg(mtu->dev, "bind driver %s\n", driver->function);
519
520 spin_lock_irqsave(&mtu->lock, flags);
521
522 mtu->softconnect = 0;
523 mtu->gadget_driver = driver;
524
525 if (mtu->ssusb->dr_mode == USB_DR_MODE_PERIPHERAL)
526 mtu3_start(mtu);
527
528 spin_unlock_irqrestore(&mtu->lock, flags);
529
530 return 0;
531}
532
533static void stop_activity(struct mtu3 *mtu)
534{
535 struct usb_gadget_driver *driver = mtu->gadget_driver;
536 int i;
537
538 /* don't disconnect if it's not connected */
539 if (mtu->g.speed == USB_SPEED_UNKNOWN)
540 driver = NULL;
541 else
542 mtu->g.speed = USB_SPEED_UNKNOWN;
543
544 /* deactivate the hardware */
545 if (mtu->softconnect) {
546 mtu->softconnect = 0;
547 mtu3_dev_on_off(mtu, 0);
548 }
549
550 /*
551 * killing any outstanding requests will quiesce the driver;
552 * then report disconnect
553 */
554 nuke(mtu->ep0, -ESHUTDOWN);
555 for (i = 1; i < mtu->num_eps; i++) {
556 nuke(mtu->in_eps + i, -ESHUTDOWN);
557 nuke(mtu->out_eps + i, -ESHUTDOWN);
558 }
559
560 if (driver) {
561 spin_unlock(&mtu->lock);
562 driver->disconnect(&mtu->g);
563 spin_lock(&mtu->lock);
564 }
565}
566
567static int mtu3_gadget_stop(struct usb_gadget *g)
568{
569 struct mtu3 *mtu = gadget_to_mtu3(g);
570 unsigned long flags;
571
572 dev_dbg(mtu->dev, "%s\n", __func__);
573
574 spin_lock_irqsave(&mtu->lock, flags);
575
576 stop_activity(mtu);
577 mtu->gadget_driver = NULL;
578
579 if (mtu->ssusb->dr_mode == USB_DR_MODE_PERIPHERAL)
580 mtu3_stop(mtu);
581
582 spin_unlock_irqrestore(&mtu->lock, flags);
583
584 return 0;
585}
586
587static const struct usb_gadget_ops mtu3_gadget_ops = {
588 .get_frame = mtu3_gadget_get_frame,
589 .wakeup = mtu3_gadget_wakeup,
590 .set_selfpowered = mtu3_gadget_set_self_powered,
591 .pullup = mtu3_gadget_pullup,
592 .udc_start = mtu3_gadget_start,
593 .udc_stop = mtu3_gadget_stop,
594};
595
596static void init_hw_ep(struct mtu3 *mtu, struct mtu3_ep *mep,
597 u32 epnum, u32 is_in)
598{
599 mep->epnum = epnum;
600 mep->mtu = mtu;
601 mep->is_in = is_in;
602
603 INIT_LIST_HEAD(&mep->req_list);
604
605 sprintf(mep->name, "ep%d%s", epnum,
606 !epnum ? "" : (is_in ? "in" : "out"));
607
608 mep->ep.name = mep->name;
609 INIT_LIST_HEAD(&mep->ep.ep_list);
610
611 /* initialize maxpacket as SS */
612 if (!epnum) {
613 usb_ep_set_maxpacket_limit(&mep->ep, 512);
614 mep->ep.caps.type_control = true;
615 mep->ep.ops = &mtu3_ep0_ops;
616 mtu->g.ep0 = &mep->ep;
617 } else {
618 usb_ep_set_maxpacket_limit(&mep->ep, 1024);
619 mep->ep.caps.type_iso = true;
620 mep->ep.caps.type_bulk = true;
621 mep->ep.caps.type_int = true;
622 mep->ep.ops = &mtu3_ep_ops;
623 list_add_tail(&mep->ep.ep_list, &mtu->g.ep_list);
624 }
625
626 dev_dbg(mtu->dev, "%s, name=%s, maxp=%d\n", __func__, mep->ep.name,
627 mep->ep.maxpacket);
628
629 if (!epnum) {
630 mep->ep.caps.dir_in = true;
631 mep->ep.caps.dir_out = true;
632 } else if (is_in) {
633 mep->ep.caps.dir_in = true;
634 } else {
635 mep->ep.caps.dir_out = true;
636 }
637}
638
639static void mtu3_gadget_init_eps(struct mtu3 *mtu)
640{
641 u8 epnum;
642
643 /* initialize endpoint list just once */
644 INIT_LIST_HEAD(&(mtu->g.ep_list));
645
646 dev_dbg(mtu->dev, "%s num_eps(1 for a pair of tx&rx ep)=%d\n",
647 __func__, mtu->num_eps);
648
649 init_hw_ep(mtu, mtu->ep0, 0, 0);
650 for (epnum = 1; epnum < mtu->num_eps; epnum++) {
651 init_hw_ep(mtu, mtu->in_eps + epnum, epnum, 1);
652 init_hw_ep(mtu, mtu->out_eps + epnum, epnum, 0);
653 }
654}
655
656int mtu3_gadget_setup(struct mtu3 *mtu)
657{
658 int ret;
659
660 mtu->g.ops = &mtu3_gadget_ops;
661 mtu->g.max_speed = mtu->max_speed;
662 mtu->g.speed = USB_SPEED_UNKNOWN;
663 mtu->g.sg_supported = 0;
664 mtu->g.name = MTU3_DRIVER_NAME;
665 mtu->is_active = 0;
666
667 mtu3_gadget_init_eps(mtu);
668
669 ret = usb_add_gadget_udc(mtu->dev, &mtu->g);
670 if (ret) {
671 dev_err(mtu->dev, "failed to register udc\n");
672 return ret;
673 }
674
675 usb_gadget_set_state(&mtu->g, USB_STATE_NOTATTACHED);
676
677 return 0;
678}
679
680void mtu3_gadget_cleanup(struct mtu3 *mtu)
681{
682 usb_del_gadget_udc(&mtu->g);
683}
684
685void mtu3_gadget_resume(struct mtu3 *mtu)
686{
687 dev_dbg(mtu->dev, "gadget RESUME\n");
688 if (mtu->gadget_driver && mtu->gadget_driver->resume) {
689 spin_unlock(&mtu->lock);
690 mtu->gadget_driver->resume(&mtu->g);
691 spin_lock(&mtu->lock);
692 }
693}
694
695/* called when SOF packets stop for 3+ msec or enters U3 */
696void mtu3_gadget_suspend(struct mtu3 *mtu)
697{
698 dev_dbg(mtu->dev, "gadget SUSPEND\n");
699 if (mtu->gadget_driver && mtu->gadget_driver->suspend) {
700 spin_unlock(&mtu->lock);
701 mtu->gadget_driver->suspend(&mtu->g);
702 spin_lock(&mtu->lock);
703 }
704}
705
706/* called when VBUS drops below session threshold, and in other cases */
707void mtu3_gadget_disconnect(struct mtu3 *mtu)
708{
709 dev_dbg(mtu->dev, "gadget DISCONNECT\n");
710 if (mtu->gadget_driver && mtu->gadget_driver->disconnect) {
711 spin_unlock(&mtu->lock);
712 mtu->gadget_driver->disconnect(&mtu->g);
713 spin_lock(&mtu->lock);
714 }
715
716 usb_gadget_set_state(&mtu->g, USB_STATE_NOTATTACHED);
717}
718
719void mtu3_gadget_reset(struct mtu3 *mtu)
720{
721 dev_dbg(mtu->dev, "gadget RESET\n");
722
723 /* report disconnect, if we didn't flush EP state */
724 if (mtu->g.speed != USB_SPEED_UNKNOWN)
725 mtu3_gadget_disconnect(mtu);
726
727 mtu->address = 0;
728 mtu->ep0_state = MU3D_EP0_STATE_SETUP;
729 mtu->may_wakeup = 0;
730}