blob: fc92e4f9f74aa0cd8174ed18a1bee1a8c33fc426 [file] [log] [blame]
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001/**
2 * gadget.c - DesignWare USB3 DRD Controller Gadget Framework Link
3 *
4 * Copyright (C) 2010-2011 Texas Instruments Incorporated - http://www.ti.com
Felipe Balbi4dc64e52011-08-19 18:10:58 +03005 *
6 * Authors: Felipe Balbi <balbi@ti.com>,
7 * Sebastian Andrzej Siewior <bigeasy@linutronix.de>
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions, and the following disclaimer,
14 * without modification.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. The names of the above-listed copyright holders may not be used
19 * to endorse or promote products derived from this software without
20 * specific prior written permission.
21 *
22 * ALTERNATIVELY, this software may be distributed under the terms of the
23 * GNU General Public License ("GPL") version 2, as published by the Free
24 * Software Foundation.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
27 * IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
28 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
30 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
31 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
32 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
33 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
34 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
35 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
36 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
37 */
38
39#include <linux/kernel.h>
40#include <linux/delay.h>
41#include <linux/slab.h>
42#include <linux/spinlock.h>
43#include <linux/platform_device.h>
44#include <linux/pm_runtime.h>
45#include <linux/interrupt.h>
46#include <linux/io.h>
47#include <linux/list.h>
48#include <linux/dma-mapping.h>
49
50#include <linux/usb/ch9.h>
51#include <linux/usb/gadget.h>
52
53#include "core.h"
54#include "gadget.h"
55#include "io.h"
56
57#define DMA_ADDR_INVALID (~(dma_addr_t)0)
58
59void dwc3_map_buffer_to_dma(struct dwc3_request *req)
60{
61 struct dwc3 *dwc = req->dep->dwc;
62
Sebastian Andrzej Siewior4ae8e1c2011-08-31 17:12:02 +020063 if (req->request.length == 0) {
64 /* req->request.dma = dwc->setup_buf_addr; */
65 return;
66 }
67
Felipe Balbi4dc64e52011-08-19 18:10:58 +030068 if (req->request.dma == DMA_ADDR_INVALID) {
69 req->request.dma = dma_map_single(dwc->dev, req->request.buf,
70 req->request.length, req->direction
71 ? DMA_TO_DEVICE : DMA_FROM_DEVICE);
72 req->mapped = true;
Felipe Balbi4dc64e52011-08-19 18:10:58 +030073 }
74}
75
76void dwc3_unmap_buffer_from_dma(struct dwc3_request *req)
77{
78 struct dwc3 *dwc = req->dep->dwc;
79
Sebastian Andrzej Siewior4ae8e1c2011-08-31 17:12:02 +020080 if (req->request.length == 0) {
81 req->request.dma = DMA_ADDR_INVALID;
82 return;
83 }
84
Felipe Balbi4dc64e52011-08-19 18:10:58 +030085 if (req->mapped) {
86 dma_unmap_single(dwc->dev, req->request.dma,
87 req->request.length, req->direction
88 ? DMA_TO_DEVICE : DMA_FROM_DEVICE);
89 req->mapped = 0;
Felipe Balbi162e1282011-08-27 15:10:09 +030090 req->request.dma = DMA_ADDR_INVALID;
Felipe Balbi4dc64e52011-08-19 18:10:58 +030091 }
92}
93
94void dwc3_gadget_giveback(struct dwc3_ep *dep, struct dwc3_request *req,
95 int status)
96{
97 struct dwc3 *dwc = dep->dwc;
98
99 if (req->queued) {
100 dep->busy_slot++;
101 /*
102 * Skip LINK TRB. We can't use req->trb and check for
103 * DWC3_TRBCTL_LINK_TRB because it points the TRB we just
104 * completed (not the LINK TRB).
105 */
106 if (((dep->busy_slot & DWC3_TRB_MASK) == DWC3_TRB_NUM - 1) &&
107 usb_endpoint_xfer_isoc(dep->desc))
108 dep->busy_slot++;
109 }
110 list_del(&req->list);
111
112 if (req->request.status == -EINPROGRESS)
113 req->request.status = status;
114
115 dwc3_unmap_buffer_from_dma(req);
116
117 dev_dbg(dwc->dev, "request %p from %s completed %d/%d ===> %d\n",
118 req, dep->name, req->request.actual,
119 req->request.length, status);
120
121 spin_unlock(&dwc->lock);
122 req->request.complete(&req->dep->endpoint, &req->request);
123 spin_lock(&dwc->lock);
124}
125
126static const char *dwc3_gadget_ep_cmd_string(u8 cmd)
127{
128 switch (cmd) {
129 case DWC3_DEPCMD_DEPSTARTCFG:
130 return "Start New Configuration";
131 case DWC3_DEPCMD_ENDTRANSFER:
132 return "End Transfer";
133 case DWC3_DEPCMD_UPDATETRANSFER:
134 return "Update Transfer";
135 case DWC3_DEPCMD_STARTTRANSFER:
136 return "Start Transfer";
137 case DWC3_DEPCMD_CLEARSTALL:
138 return "Clear Stall";
139 case DWC3_DEPCMD_SETSTALL:
140 return "Set Stall";
141 case DWC3_DEPCMD_GETSEQNUMBER:
142 return "Get Data Sequence Number";
143 case DWC3_DEPCMD_SETTRANSFRESOURCE:
144 return "Set Endpoint Transfer Resource";
145 case DWC3_DEPCMD_SETEPCONFIG:
146 return "Set Endpoint Configuration";
147 default:
148 return "UNKNOWN command";
149 }
150}
151
152int dwc3_send_gadget_ep_cmd(struct dwc3 *dwc, unsigned ep,
153 unsigned cmd, struct dwc3_gadget_ep_cmd_params *params)
154{
155 struct dwc3_ep *dep = dwc->eps[ep];
Sebastian Andrzej Siewior6062cac2011-08-29 16:46:38 +0200156 u32 timeout = 500;
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300157 u32 reg;
158
159 dev_vdbg(dwc->dev, "%s: cmd '%s' params %08x %08x %08x\n",
160 dep->name,
161 dwc3_gadget_ep_cmd_string(cmd), params->param0.raw,
162 params->param1.raw, params->param2.raw);
163
164 dwc3_writel(dwc->regs, DWC3_DEPCMDPAR0(ep), params->param0.raw);
165 dwc3_writel(dwc->regs, DWC3_DEPCMDPAR1(ep), params->param1.raw);
166 dwc3_writel(dwc->regs, DWC3_DEPCMDPAR2(ep), params->param2.raw);
167
168 dwc3_writel(dwc->regs, DWC3_DEPCMD(ep), cmd | DWC3_DEPCMD_CMDACT);
169 do {
170 reg = dwc3_readl(dwc->regs, DWC3_DEPCMD(ep));
171 if (!(reg & DWC3_DEPCMD_CMDACT)) {
Felipe Balbic7dbe4f2011-08-27 20:29:58 +0300172 dev_vdbg(dwc->dev, "Command Complete --> %d\n",
173 DWC3_DEPCMD_STATUS(reg));
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300174 return 0;
175 }
176
177 /*
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300178 * We can't sleep here, because it is also called from
179 * interrupt context.
180 */
181 timeout--;
182 if (!timeout)
183 return -ETIMEDOUT;
184
Sebastian Andrzej Siewior6062cac2011-08-29 16:46:38 +0200185 udelay(1);
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300186 } while (1);
187}
188
189static dma_addr_t dwc3_trb_dma_offset(struct dwc3_ep *dep,
190 struct dwc3_trb_hw *trb)
191{
192 u32 offset = trb - dep->trb_pool;
193
194 return dep->trb_pool_dma + offset;
195}
196
197static int dwc3_alloc_trb_pool(struct dwc3_ep *dep)
198{
199 struct dwc3 *dwc = dep->dwc;
200
201 if (dep->trb_pool)
202 return 0;
203
204 if (dep->number == 0 || dep->number == 1)
205 return 0;
206
207 dep->trb_pool = dma_alloc_coherent(dwc->dev,
208 sizeof(struct dwc3_trb) * DWC3_TRB_NUM,
209 &dep->trb_pool_dma, GFP_KERNEL);
210 if (!dep->trb_pool) {
211 dev_err(dep->dwc->dev, "failed to allocate trb pool for %s\n",
212 dep->name);
213 return -ENOMEM;
214 }
215
216 return 0;
217}
218
219static void dwc3_free_trb_pool(struct dwc3_ep *dep)
220{
221 struct dwc3 *dwc = dep->dwc;
222
223 dma_free_coherent(dwc->dev, sizeof(struct dwc3_trb) * DWC3_TRB_NUM,
224 dep->trb_pool, dep->trb_pool_dma);
225
226 dep->trb_pool = NULL;
227 dep->trb_pool_dma = 0;
228}
229
230static int dwc3_gadget_start_config(struct dwc3 *dwc, struct dwc3_ep *dep)
231{
232 struct dwc3_gadget_ep_cmd_params params;
233 u32 cmd;
234
235 memset(&params, 0x00, sizeof(params));
236
237 if (dep->number != 1) {
238 cmd = DWC3_DEPCMD_DEPSTARTCFG;
239 /* XferRscIdx == 0 for ep0 and 2 for the remaining */
Paul Zimmermana6544d02011-09-30 10:58:42 +0300240 if (dep->number > 1) {
241 if (dwc->start_config_issued)
242 return 0;
243 dwc->start_config_issued = true;
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300244 cmd |= DWC3_DEPCMD_PARAM(2);
Paul Zimmermana6544d02011-09-30 10:58:42 +0300245 }
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300246
247 return dwc3_send_gadget_ep_cmd(dwc, 0, cmd, &params);
248 }
249
250 return 0;
251}
252
253static int dwc3_gadget_set_ep_config(struct dwc3 *dwc, struct dwc3_ep *dep,
254 const struct usb_endpoint_descriptor *desc)
255{
256 struct dwc3_gadget_ep_cmd_params params;
257
258 memset(&params, 0x00, sizeof(params));
259
260 params.param0.depcfg.ep_type = usb_endpoint_type(desc);
261 params.param0.depcfg.max_packet_size =
262 le16_to_cpu(desc->wMaxPacketSize);
263
264 params.param1.depcfg.xfer_complete_enable = true;
265 params.param1.depcfg.xfer_not_ready_enable = true;
266
267 if (usb_endpoint_xfer_isoc(desc))
268 params.param1.depcfg.xfer_in_progress_enable = true;
269
270 /*
271 * We are doing 1:1 mapping for endpoints, meaning
272 * Physical Endpoints 2 maps to Logical Endpoint 2 and
273 * so on. We consider the direction bit as part of the physical
274 * endpoint number. So USB endpoint 0x81 is 0x03.
275 */
276 params.param1.depcfg.ep_number = dep->number;
277
278 /*
279 * We must use the lower 16 TX FIFOs even though
280 * HW might have more
281 */
282 if (dep->direction)
283 params.param0.depcfg.fifo_number = dep->number >> 1;
284
285 if (desc->bInterval) {
286 params.param1.depcfg.binterval_m1 = desc->bInterval - 1;
287 dep->interval = 1 << (desc->bInterval - 1);
288 }
289
290 return dwc3_send_gadget_ep_cmd(dwc, dep->number,
291 DWC3_DEPCMD_SETEPCONFIG, &params);
292}
293
294static int dwc3_gadget_set_xfer_resource(struct dwc3 *dwc, struct dwc3_ep *dep)
295{
296 struct dwc3_gadget_ep_cmd_params params;
297
298 memset(&params, 0x00, sizeof(params));
299
300 params.param0.depxfercfg.number_xfer_resources = 1;
301
302 return dwc3_send_gadget_ep_cmd(dwc, dep->number,
303 DWC3_DEPCMD_SETTRANSFRESOURCE, &params);
304}
305
306/**
307 * __dwc3_gadget_ep_enable - Initializes a HW endpoint
308 * @dep: endpoint to be initialized
309 * @desc: USB Endpoint Descriptor
310 *
311 * Caller should take care of locking
312 */
313static int __dwc3_gadget_ep_enable(struct dwc3_ep *dep,
314 const struct usb_endpoint_descriptor *desc)
315{
316 struct dwc3 *dwc = dep->dwc;
317 u32 reg;
318 int ret = -ENOMEM;
319
320 if (!(dep->flags & DWC3_EP_ENABLED)) {
321 ret = dwc3_gadget_start_config(dwc, dep);
322 if (ret)
323 return ret;
324 }
325
326 ret = dwc3_gadget_set_ep_config(dwc, dep, desc);
327 if (ret)
328 return ret;
329
330 if (!(dep->flags & DWC3_EP_ENABLED)) {
331 struct dwc3_trb_hw *trb_st_hw;
332 struct dwc3_trb_hw *trb_link_hw;
333 struct dwc3_trb trb_link;
334
335 ret = dwc3_gadget_set_xfer_resource(dwc, dep);
336 if (ret)
337 return ret;
338
339 dep->desc = desc;
340 dep->type = usb_endpoint_type(desc);
341 dep->flags |= DWC3_EP_ENABLED;
342
343 reg = dwc3_readl(dwc->regs, DWC3_DALEPENA);
344 reg |= DWC3_DALEPENA_EP(dep->number);
345 dwc3_writel(dwc->regs, DWC3_DALEPENA, reg);
346
347 if (!usb_endpoint_xfer_isoc(desc))
348 return 0;
349
350 memset(&trb_link, 0, sizeof(trb_link));
351
352 /* Link TRB for ISOC. The HWO but is never reset */
353 trb_st_hw = &dep->trb_pool[0];
354
355 trb_link.bplh = dwc3_trb_dma_offset(dep, trb_st_hw);
356 trb_link.trbctl = DWC3_TRBCTL_LINK_TRB;
357 trb_link.hwo = true;
358
359 trb_link_hw = &dep->trb_pool[DWC3_TRB_NUM - 1];
360 dwc3_trb_to_hw(&trb_link, trb_link_hw);
361 }
362
363 return 0;
364}
365
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +0200366static void dwc3_stop_active_transfer(struct dwc3 *dwc, u32 epnum);
367static void dwc3_remove_requests(struct dwc3 *dwc, struct dwc3_ep *dep)
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300368{
369 struct dwc3_request *req;
370
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +0200371 if (!list_empty(&dep->req_queued))
372 dwc3_stop_active_transfer(dwc, dep->number);
373
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300374 while (!list_empty(&dep->request_list)) {
375 req = next_request(&dep->request_list);
376
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +0200377 dwc3_gadget_giveback(dep, req, -ESHUTDOWN);
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300378 }
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300379}
380
381/**
382 * __dwc3_gadget_ep_disable - Disables a HW endpoint
383 * @dep: the endpoint to disable
384 *
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +0200385 * This function also removes requests which are currently processed ny the
386 * hardware and those which are not yet scheduled.
387 * Caller should take care of locking.
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300388 */
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300389static int __dwc3_gadget_ep_disable(struct dwc3_ep *dep)
390{
391 struct dwc3 *dwc = dep->dwc;
392 u32 reg;
393
394 dep->flags &= ~DWC3_EP_ENABLED;
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +0200395 dwc3_remove_requests(dwc, dep);
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300396
397 reg = dwc3_readl(dwc->regs, DWC3_DALEPENA);
398 reg &= ~DWC3_DALEPENA_EP(dep->number);
399 dwc3_writel(dwc->regs, DWC3_DALEPENA, reg);
400
401 dep->desc = NULL;
402 dep->type = 0;
403
404 return 0;
405}
406
407/* -------------------------------------------------------------------------- */
408
409static int dwc3_gadget_ep0_enable(struct usb_ep *ep,
410 const struct usb_endpoint_descriptor *desc)
411{
412 return -EINVAL;
413}
414
415static int dwc3_gadget_ep0_disable(struct usb_ep *ep)
416{
417 return -EINVAL;
418}
419
420/* -------------------------------------------------------------------------- */
421
422static int dwc3_gadget_ep_enable(struct usb_ep *ep,
423 const struct usb_endpoint_descriptor *desc)
424{
425 struct dwc3_ep *dep;
426 struct dwc3 *dwc;
427 unsigned long flags;
428 int ret;
429
430 if (!ep || !desc || desc->bDescriptorType != USB_DT_ENDPOINT) {
431 pr_debug("dwc3: invalid parameters\n");
432 return -EINVAL;
433 }
434
435 if (!desc->wMaxPacketSize) {
436 pr_debug("dwc3: missing wMaxPacketSize\n");
437 return -EINVAL;
438 }
439
440 dep = to_dwc3_ep(ep);
441 dwc = dep->dwc;
442
443 switch (usb_endpoint_type(desc)) {
444 case USB_ENDPOINT_XFER_CONTROL:
445 strncat(dep->name, "-control", sizeof(dep->name));
446 break;
447 case USB_ENDPOINT_XFER_ISOC:
448 strncat(dep->name, "-isoc", sizeof(dep->name));
449 break;
450 case USB_ENDPOINT_XFER_BULK:
451 strncat(dep->name, "-bulk", sizeof(dep->name));
452 break;
453 case USB_ENDPOINT_XFER_INT:
454 strncat(dep->name, "-int", sizeof(dep->name));
455 break;
456 default:
457 dev_err(dwc->dev, "invalid endpoint transfer type\n");
458 }
459
460 if (dep->flags & DWC3_EP_ENABLED) {
461 dev_WARN_ONCE(dwc->dev, true, "%s is already enabled\n",
462 dep->name);
463 return 0;
464 }
465
466 dev_vdbg(dwc->dev, "Enabling %s\n", dep->name);
467
468 spin_lock_irqsave(&dwc->lock, flags);
469 ret = __dwc3_gadget_ep_enable(dep, desc);
470 spin_unlock_irqrestore(&dwc->lock, flags);
471
472 return ret;
473}
474
475static int dwc3_gadget_ep_disable(struct usb_ep *ep)
476{
477 struct dwc3_ep *dep;
478 struct dwc3 *dwc;
479 unsigned long flags;
480 int ret;
481
482 if (!ep) {
483 pr_debug("dwc3: invalid parameters\n");
484 return -EINVAL;
485 }
486
487 dep = to_dwc3_ep(ep);
488 dwc = dep->dwc;
489
490 if (!(dep->flags & DWC3_EP_ENABLED)) {
491 dev_WARN_ONCE(dwc->dev, true, "%s is already disabled\n",
492 dep->name);
493 return 0;
494 }
495
496 snprintf(dep->name, sizeof(dep->name), "ep%d%s",
497 dep->number >> 1,
498 (dep->number & 1) ? "in" : "out");
499
500 spin_lock_irqsave(&dwc->lock, flags);
501 ret = __dwc3_gadget_ep_disable(dep);
502 spin_unlock_irqrestore(&dwc->lock, flags);
503
504 return ret;
505}
506
507static struct usb_request *dwc3_gadget_ep_alloc_request(struct usb_ep *ep,
508 gfp_t gfp_flags)
509{
510 struct dwc3_request *req;
511 struct dwc3_ep *dep = to_dwc3_ep(ep);
512 struct dwc3 *dwc = dep->dwc;
513
514 req = kzalloc(sizeof(*req), gfp_flags);
515 if (!req) {
516 dev_err(dwc->dev, "not enough memory\n");
517 return NULL;
518 }
519
520 req->epnum = dep->number;
521 req->dep = dep;
522 req->request.dma = DMA_ADDR_INVALID;
523
524 return &req->request;
525}
526
527static void dwc3_gadget_ep_free_request(struct usb_ep *ep,
528 struct usb_request *request)
529{
530 struct dwc3_request *req = to_dwc3_request(request);
531
532 kfree(req);
533}
534
535/*
536 * dwc3_prepare_trbs - setup TRBs from requests
537 * @dep: endpoint for which requests are being prepared
538 * @starting: true if the endpoint is idle and no requests are queued.
539 *
540 * The functions goes through the requests list and setups TRBs for the
541 * transfers. The functions returns once there are not more TRBs available or
542 * it run out of requests.
543 */
544static struct dwc3_request *dwc3_prepare_trbs(struct dwc3_ep *dep,
545 bool starting)
546{
547 struct dwc3_request *req, *n, *ret = NULL;
548 struct dwc3_trb_hw *trb_hw;
549 struct dwc3_trb trb;
550 u32 trbs_left;
551
552 BUILD_BUG_ON_NOT_POWER_OF_2(DWC3_TRB_NUM);
553
554 /* the first request must not be queued */
555 trbs_left = (dep->busy_slot - dep->free_slot) & DWC3_TRB_MASK;
556 /*
557 * if busy & slot are equal than it is either full or empty. If we are
558 * starting to proceed requests then we are empty. Otherwise we ar
559 * full and don't do anything
560 */
561 if (!trbs_left) {
562 if (!starting)
563 return NULL;
564 trbs_left = DWC3_TRB_NUM;
565 /*
566 * In case we start from scratch, we queue the ISOC requests
567 * starting from slot 1. This is done because we use ring
568 * buffer and have no LST bit to stop us. Instead, we place
569 * IOC bit TRB_NUM/4. We try to avoid to having an interrupt
570 * after the first request so we start at slot 1 and have
571 * 7 requests proceed before we hit the first IOC.
572 * Other transfer types don't use the ring buffer and are
573 * processed from the first TRB until the last one. Since we
574 * don't wrap around we have to start at the beginning.
575 */
576 if (usb_endpoint_xfer_isoc(dep->desc)) {
577 dep->busy_slot = 1;
578 dep->free_slot = 1;
579 } else {
580 dep->busy_slot = 0;
581 dep->free_slot = 0;
582 }
583 }
584
585 /* The last TRB is a link TRB, not used for xfer */
586 if ((trbs_left <= 1) && usb_endpoint_xfer_isoc(dep->desc))
587 return NULL;
588
589 list_for_each_entry_safe(req, n, &dep->request_list, list) {
590 unsigned int last_one = 0;
591 unsigned int cur_slot;
592
593 trb_hw = &dep->trb_pool[dep->free_slot & DWC3_TRB_MASK];
594 cur_slot = dep->free_slot;
595 dep->free_slot++;
596
597 /* Skip the LINK-TRB on ISOC */
598 if (((cur_slot & DWC3_TRB_MASK) == DWC3_TRB_NUM - 1) &&
599 usb_endpoint_xfer_isoc(dep->desc))
600 continue;
601
602 dwc3_gadget_move_request_queued(req);
603 memset(&trb, 0, sizeof(trb));
604 trbs_left--;
605
606 /* Is our TRB pool empty? */
607 if (!trbs_left)
608 last_one = 1;
609 /* Is this the last request? */
610 if (list_empty(&dep->request_list))
611 last_one = 1;
612
613 /*
614 * FIXME we shouldn't need to set LST bit always but we are
615 * facing some weird problem with the Hardware where it doesn't
616 * complete even though it has been previously started.
617 *
618 * While we're debugging the problem, as a workaround to
619 * multiple TRBs handling, use only one TRB at a time.
620 */
621 last_one = 1;
622
623 req->trb = trb_hw;
624 if (!ret)
625 ret = req;
626
627 trb.bplh = req->request.dma;
628
629 if (usb_endpoint_xfer_isoc(dep->desc)) {
630 trb.isp_imi = true;
631 trb.csp = true;
632 } else {
633 trb.lst = last_one;
634 }
635
636 switch (usb_endpoint_type(dep->desc)) {
637 case USB_ENDPOINT_XFER_CONTROL:
638 trb.trbctl = DWC3_TRBCTL_CONTROL_SETUP;
639 break;
640
641 case USB_ENDPOINT_XFER_ISOC:
Sebastian Andrzej Siewior15623d72011-08-22 17:42:19 +0200642 trb.trbctl = DWC3_TRBCTL_ISOCHRONOUS_FIRST;
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300643
644 /* IOC every DWC3_TRB_NUM / 4 so we can refill */
645 if (!(cur_slot % (DWC3_TRB_NUM / 4)))
646 trb.ioc = last_one;
647 break;
648
649 case USB_ENDPOINT_XFER_BULK:
650 case USB_ENDPOINT_XFER_INT:
651 trb.trbctl = DWC3_TRBCTL_NORMAL;
652 break;
653 default:
654 /*
655 * This is only possible with faulty memory because we
656 * checked it already :)
657 */
658 BUG();
659 }
660
661 trb.length = req->request.length;
662 trb.hwo = true;
663
664 dwc3_trb_to_hw(&trb, trb_hw);
665 req->trb_dma = dwc3_trb_dma_offset(dep, trb_hw);
666
667 if (last_one)
668 break;
669 }
670
671 return ret;
672}
673
674static int __dwc3_gadget_kick_transfer(struct dwc3_ep *dep, u16 cmd_param,
675 int start_new)
676{
677 struct dwc3_gadget_ep_cmd_params params;
678 struct dwc3_request *req;
679 struct dwc3 *dwc = dep->dwc;
680 int ret;
681 u32 cmd;
682
683 if (start_new && (dep->flags & DWC3_EP_BUSY)) {
684 dev_vdbg(dwc->dev, "%s: endpoint busy\n", dep->name);
685 return -EBUSY;
686 }
687 dep->flags &= ~DWC3_EP_PENDING_REQUEST;
688
689 /*
690 * If we are getting here after a short-out-packet we don't enqueue any
691 * new requests as we try to set the IOC bit only on the last request.
692 */
693 if (start_new) {
694 if (list_empty(&dep->req_queued))
695 dwc3_prepare_trbs(dep, start_new);
696
697 /* req points to the first request which will be sent */
698 req = next_request(&dep->req_queued);
699 } else {
700 /*
701 * req points to the first request where HWO changed
702 * from 0 to 1
703 */
704 req = dwc3_prepare_trbs(dep, start_new);
705 }
706 if (!req) {
707 dep->flags |= DWC3_EP_PENDING_REQUEST;
708 return 0;
709 }
710
711 memset(&params, 0, sizeof(params));
712 params.param0.depstrtxfer.transfer_desc_addr_high =
713 upper_32_bits(req->trb_dma);
714 params.param1.depstrtxfer.transfer_desc_addr_low =
715 lower_32_bits(req->trb_dma);
716
717 if (start_new)
718 cmd = DWC3_DEPCMD_STARTTRANSFER;
719 else
720 cmd = DWC3_DEPCMD_UPDATETRANSFER;
721
722 cmd |= DWC3_DEPCMD_PARAM(cmd_param);
723 ret = dwc3_send_gadget_ep_cmd(dwc, dep->number, cmd, &params);
724 if (ret < 0) {
725 dev_dbg(dwc->dev, "failed to send STARTTRANSFER command\n");
726
727 /*
728 * FIXME we need to iterate over the list of requests
729 * here and stop, unmap, free and del each of the linked
730 * requests instead of we do now.
731 */
732 dwc3_unmap_buffer_from_dma(req);
733 list_del(&req->list);
734 return ret;
735 }
736
737 dep->flags |= DWC3_EP_BUSY;
738 dep->res_trans_idx = dwc3_gadget_ep_get_transfer_index(dwc,
739 dep->number);
740 if (!dep->res_trans_idx)
741 printk_once(KERN_ERR "%s() res_trans_idx is invalid\n", __func__);
742 return 0;
743}
744
745static int __dwc3_gadget_ep_queue(struct dwc3_ep *dep, struct dwc3_request *req)
746{
747 req->request.actual = 0;
748 req->request.status = -EINPROGRESS;
749 req->direction = dep->direction;
750 req->epnum = dep->number;
751
752 /*
753 * We only add to our list of requests now and
754 * start consuming the list once we get XferNotReady
755 * IRQ.
756 *
757 * That way, we avoid doing anything that we don't need
758 * to do now and defer it until the point we receive a
759 * particular token from the Host side.
760 *
761 * This will also avoid Host cancelling URBs due to too
762 * many NACKs.
763 */
764 dwc3_map_buffer_to_dma(req);
765 list_add_tail(&req->list, &dep->request_list);
766
767 /*
768 * There is one special case: XferNotReady with
769 * empty list of requests. We need to kick the
770 * transfer here in that situation, otherwise
771 * we will be NAKing forever.
772 *
773 * If we get XferNotReady before gadget driver
774 * has a chance to queue a request, we will ACK
775 * the IRQ but won't be able to receive the data
776 * until the next request is queued. The following
777 * code is handling exactly that.
778 */
779 if (dep->flags & DWC3_EP_PENDING_REQUEST) {
780 int ret;
781 int start_trans;
782
783 start_trans = 1;
784 if (usb_endpoint_xfer_isoc(dep->endpoint.desc) &&
785 dep->flags & DWC3_EP_BUSY)
786 start_trans = 0;
787
788 ret = __dwc3_gadget_kick_transfer(dep, 0, start_trans);
789 if (ret && ret != -EBUSY) {
790 struct dwc3 *dwc = dep->dwc;
791
792 dev_dbg(dwc->dev, "%s: failed to kick transfers\n",
793 dep->name);
794 }
795 };
796
797 return 0;
798}
799
800static int dwc3_gadget_ep_queue(struct usb_ep *ep, struct usb_request *request,
801 gfp_t gfp_flags)
802{
803 struct dwc3_request *req = to_dwc3_request(request);
804 struct dwc3_ep *dep = to_dwc3_ep(ep);
805 struct dwc3 *dwc = dep->dwc;
806
807 unsigned long flags;
808
809 int ret;
810
811 if (!dep->desc) {
812 dev_dbg(dwc->dev, "trying to queue request %p to disabled %s\n",
813 request, ep->name);
814 return -ESHUTDOWN;
815 }
816
817 dev_vdbg(dwc->dev, "queing request %p to %s length %d\n",
818 request, ep->name, request->length);
819
820 spin_lock_irqsave(&dwc->lock, flags);
821 ret = __dwc3_gadget_ep_queue(dep, req);
822 spin_unlock_irqrestore(&dwc->lock, flags);
823
824 return ret;
825}
826
827static int dwc3_gadget_ep_dequeue(struct usb_ep *ep,
828 struct usb_request *request)
829{
830 struct dwc3_request *req = to_dwc3_request(request);
831 struct dwc3_request *r = NULL;
832
833 struct dwc3_ep *dep = to_dwc3_ep(ep);
834 struct dwc3 *dwc = dep->dwc;
835
836 unsigned long flags;
837 int ret = 0;
838
839 spin_lock_irqsave(&dwc->lock, flags);
840
841 list_for_each_entry(r, &dep->request_list, list) {
842 if (r == req)
843 break;
844 }
845
846 if (r != req) {
847 list_for_each_entry(r, &dep->req_queued, list) {
848 if (r == req)
849 break;
850 }
851 if (r == req) {
852 /* wait until it is processed */
853 dwc3_stop_active_transfer(dwc, dep->number);
854 goto out0;
855 }
856 dev_err(dwc->dev, "request %p was not queued to %s\n",
857 request, ep->name);
858 ret = -EINVAL;
859 goto out0;
860 }
861
862 /* giveback the request */
863 dwc3_gadget_giveback(dep, req, -ECONNRESET);
864
865out0:
866 spin_unlock_irqrestore(&dwc->lock, flags);
867
868 return ret;
869}
870
871int __dwc3_gadget_ep_set_halt(struct dwc3_ep *dep, int value)
872{
873 struct dwc3_gadget_ep_cmd_params params;
874 struct dwc3 *dwc = dep->dwc;
875 int ret;
876
877 memset(&params, 0x00, sizeof(params));
878
879 if (value) {
Felipe Balbiaa7b4d02011-08-30 15:48:08 +0300880 if (dep->number == 0 || dep->number == 1) {
881 /*
882 * Whenever EP0 is stalled, we will restart
883 * the state machine, thus moving back to
884 * Setup Phase
885 */
886 dwc->ep0state = EP0_SETUP_PHASE;
887 }
Felipe Balbi4dc64e52011-08-19 18:10:58 +0300888
889 ret = dwc3_send_gadget_ep_cmd(dwc, dep->number,
890 DWC3_DEPCMD_SETSTALL, &params);
891 if (ret)
892 dev_err(dwc->dev, "failed to %s STALL on %s\n",
893 value ? "set" : "clear",
894 dep->name);
895 else
896 dep->flags |= DWC3_EP_STALL;
897 } else {
898 ret = dwc3_send_gadget_ep_cmd(dwc, dep->number,
899 DWC3_DEPCMD_CLEARSTALL, &params);
900 if (ret)
901 dev_err(dwc->dev, "failed to %s STALL on %s\n",
902 value ? "set" : "clear",
903 dep->name);
904 else
905 dep->flags &= ~DWC3_EP_STALL;
906 }
907 return ret;
908}
909
910static int dwc3_gadget_ep_set_halt(struct usb_ep *ep, int value)
911{
912 struct dwc3_ep *dep = to_dwc3_ep(ep);
913 struct dwc3 *dwc = dep->dwc;
914
915 unsigned long flags;
916
917 int ret;
918
919 spin_lock_irqsave(&dwc->lock, flags);
920
921 if (usb_endpoint_xfer_isoc(dep->desc)) {
922 dev_err(dwc->dev, "%s is of Isochronous type\n", dep->name);
923 ret = -EINVAL;
924 goto out;
925 }
926
927 ret = __dwc3_gadget_ep_set_halt(dep, value);
928out:
929 spin_unlock_irqrestore(&dwc->lock, flags);
930
931 return ret;
932}
933
934static int dwc3_gadget_ep_set_wedge(struct usb_ep *ep)
935{
936 struct dwc3_ep *dep = to_dwc3_ep(ep);
937
938 dep->flags |= DWC3_EP_WEDGE;
939
940 return usb_ep_set_halt(ep);
941}
942
943/* -------------------------------------------------------------------------- */
944
945static struct usb_endpoint_descriptor dwc3_gadget_ep0_desc = {
946 .bLength = USB_DT_ENDPOINT_SIZE,
947 .bDescriptorType = USB_DT_ENDPOINT,
948 .bmAttributes = USB_ENDPOINT_XFER_CONTROL,
949};
950
951static const struct usb_ep_ops dwc3_gadget_ep0_ops = {
952 .enable = dwc3_gadget_ep0_enable,
953 .disable = dwc3_gadget_ep0_disable,
954 .alloc_request = dwc3_gadget_ep_alloc_request,
955 .free_request = dwc3_gadget_ep_free_request,
956 .queue = dwc3_gadget_ep0_queue,
957 .dequeue = dwc3_gadget_ep_dequeue,
958 .set_halt = dwc3_gadget_ep_set_halt,
959 .set_wedge = dwc3_gadget_ep_set_wedge,
960};
961
962static const struct usb_ep_ops dwc3_gadget_ep_ops = {
963 .enable = dwc3_gadget_ep_enable,
964 .disable = dwc3_gadget_ep_disable,
965 .alloc_request = dwc3_gadget_ep_alloc_request,
966 .free_request = dwc3_gadget_ep_free_request,
967 .queue = dwc3_gadget_ep_queue,
968 .dequeue = dwc3_gadget_ep_dequeue,
969 .set_halt = dwc3_gadget_ep_set_halt,
970 .set_wedge = dwc3_gadget_ep_set_wedge,
971};
972
973/* -------------------------------------------------------------------------- */
974
975static int dwc3_gadget_get_frame(struct usb_gadget *g)
976{
977 struct dwc3 *dwc = gadget_to_dwc(g);
978 u32 reg;
979
980 reg = dwc3_readl(dwc->regs, DWC3_DSTS);
981 return DWC3_DSTS_SOFFN(reg);
982}
983
984static int dwc3_gadget_wakeup(struct usb_gadget *g)
985{
986 struct dwc3 *dwc = gadget_to_dwc(g);
987
988 unsigned long timeout;
989 unsigned long flags;
990
991 u32 reg;
992
993 int ret = 0;
994
995 u8 link_state;
996 u8 speed;
997
998 spin_lock_irqsave(&dwc->lock, flags);
999
1000 /*
1001 * According to the Databook Remote wakeup request should
1002 * be issued only when the device is in early suspend state.
1003 *
1004 * We can check that via USB Link State bits in DSTS register.
1005 */
1006 reg = dwc3_readl(dwc->regs, DWC3_DSTS);
1007
1008 speed = reg & DWC3_DSTS_CONNECTSPD;
1009 if (speed == DWC3_DSTS_SUPERSPEED) {
1010 dev_dbg(dwc->dev, "no wakeup on SuperSpeed\n");
1011 ret = -EINVAL;
1012 goto out;
1013 }
1014
1015 link_state = DWC3_DSTS_USBLNKST(reg);
1016
1017 switch (link_state) {
1018 case DWC3_LINK_STATE_RX_DET: /* in HS, means Early Suspend */
1019 case DWC3_LINK_STATE_U3: /* in HS, means SUSPEND */
1020 break;
1021 default:
1022 dev_dbg(dwc->dev, "can't wakeup from link state %d\n",
1023 link_state);
1024 ret = -EINVAL;
1025 goto out;
1026 }
1027
1028 reg = dwc3_readl(dwc->regs, DWC3_DCTL);
1029
1030 /*
1031 * Switch link state to Recovery. In HS/FS/LS this means
1032 * RemoteWakeup Request
1033 */
1034 reg |= DWC3_DCTL_ULSTCHNG_RECOVERY;
1035 dwc3_writel(dwc->regs, DWC3_DCTL, reg);
1036
1037 /* wait for at least 2000us */
1038 usleep_range(2000, 2500);
1039
1040 /* write zeroes to Link Change Request */
1041 reg &= ~DWC3_DCTL_ULSTCHNGREQ_MASK;
1042 dwc3_writel(dwc->regs, DWC3_DCTL, reg);
1043
1044 /* pool until Link State change to ON */
1045 timeout = jiffies + msecs_to_jiffies(100);
1046
1047 while (!(time_after(jiffies, timeout))) {
1048 reg = dwc3_readl(dwc->regs, DWC3_DSTS);
1049
1050 /* in HS, means ON */
1051 if (DWC3_DSTS_USBLNKST(reg) == DWC3_LINK_STATE_U0)
1052 break;
1053 }
1054
1055 if (DWC3_DSTS_USBLNKST(reg) != DWC3_LINK_STATE_U0) {
1056 dev_err(dwc->dev, "failed to send remote wakeup\n");
1057 ret = -EINVAL;
1058 }
1059
1060out:
1061 spin_unlock_irqrestore(&dwc->lock, flags);
1062
1063 return ret;
1064}
1065
1066static int dwc3_gadget_set_selfpowered(struct usb_gadget *g,
1067 int is_selfpowered)
1068{
1069 struct dwc3 *dwc = gadget_to_dwc(g);
1070
1071 dwc->is_selfpowered = !!is_selfpowered;
1072
1073 return 0;
1074}
1075
1076static void dwc3_gadget_run_stop(struct dwc3 *dwc, int is_on)
1077{
1078 u32 reg;
Sebastian Andrzej Siewior6062cac2011-08-29 16:46:38 +02001079 u32 timeout = 500;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001080
1081 reg = dwc3_readl(dwc->regs, DWC3_DCTL);
1082 if (is_on)
1083 reg |= DWC3_DCTL_RUN_STOP;
1084 else
1085 reg &= ~DWC3_DCTL_RUN_STOP;
1086
1087 dwc3_writel(dwc->regs, DWC3_DCTL, reg);
1088
1089 do {
1090 reg = dwc3_readl(dwc->regs, DWC3_DSTS);
1091 if (is_on) {
1092 if (!(reg & DWC3_DSTS_DEVCTRLHLT))
1093 break;
1094 } else {
1095 if (reg & DWC3_DSTS_DEVCTRLHLT)
1096 break;
1097 }
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001098 timeout--;
1099 if (!timeout)
1100 break;
Sebastian Andrzej Siewior6062cac2011-08-29 16:46:38 +02001101 udelay(1);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001102 } while (1);
1103
1104 dev_vdbg(dwc->dev, "gadget %s data soft-%s\n",
1105 dwc->gadget_driver
1106 ? dwc->gadget_driver->function : "no-function",
1107 is_on ? "connect" : "disconnect");
1108}
1109
1110static int dwc3_gadget_pullup(struct usb_gadget *g, int is_on)
1111{
1112 struct dwc3 *dwc = gadget_to_dwc(g);
1113 unsigned long flags;
1114
1115 is_on = !!is_on;
1116
1117 spin_lock_irqsave(&dwc->lock, flags);
1118 dwc3_gadget_run_stop(dwc, is_on);
1119 spin_unlock_irqrestore(&dwc->lock, flags);
1120
1121 return 0;
1122}
1123
1124static int dwc3_gadget_start(struct usb_gadget *g,
1125 struct usb_gadget_driver *driver)
1126{
1127 struct dwc3 *dwc = gadget_to_dwc(g);
1128 struct dwc3_ep *dep;
1129 unsigned long flags;
1130 int ret = 0;
1131 u32 reg;
1132
1133 spin_lock_irqsave(&dwc->lock, flags);
1134
1135 if (dwc->gadget_driver) {
1136 dev_err(dwc->dev, "%s is already bound to %s\n",
1137 dwc->gadget.name,
1138 dwc->gadget_driver->driver.name);
1139 ret = -EBUSY;
1140 goto err0;
1141 }
1142
1143 dwc->gadget_driver = driver;
1144 dwc->gadget.dev.driver = &driver->driver;
1145
1146 reg = dwc3_readl(dwc->regs, DWC3_GCTL);
1147
Felipe Balbid195b322011-09-08 17:42:11 +03001148 reg &= ~DWC3_GCTL_SCALEDOWN(3);
1149 reg &= ~DWC3_GCTL_PRTCAPDIR(DWC3_GCTL_PRTCAP_OTG);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001150 reg &= ~DWC3_GCTL_DISSCRAMBLE;
Felipe Balbid195b322011-09-08 17:42:11 +03001151 reg |= DWC3_GCTL_PRTCAPDIR(DWC3_GCTL_PRTCAP_DEVICE);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001152
1153 /*
1154 * WORKAROUND: DWC3 revisions <1.90a have a bug
1155 * when The device fails to connect at SuperSpeed
1156 * and falls back to high-speed mode which causes
1157 * the device to enter in a Connect/Disconnect loop
1158 */
1159 if (dwc->revision < DWC3_REVISION_190A)
1160 reg |= DWC3_GCTL_U2RSTECN;
1161
1162 dwc3_writel(dwc->regs, DWC3_GCTL, reg);
1163
1164 reg = dwc3_readl(dwc->regs, DWC3_DCFG);
1165 reg &= ~(DWC3_DCFG_SPEED_MASK);
1166 reg |= DWC3_DCFG_SUPERSPEED;
1167 dwc3_writel(dwc->regs, DWC3_DCFG, reg);
1168
Paul Zimmermana6544d02011-09-30 10:58:42 +03001169 dwc->start_config_issued = false;
1170
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001171 /* Start with SuperSpeed Default */
1172 dwc3_gadget_ep0_desc.wMaxPacketSize = cpu_to_le16(512);
1173
1174 dep = dwc->eps[0];
1175 ret = __dwc3_gadget_ep_enable(dep, &dwc3_gadget_ep0_desc);
1176 if (ret) {
1177 dev_err(dwc->dev, "failed to enable %s\n", dep->name);
1178 goto err0;
1179 }
1180
1181 dep = dwc->eps[1];
1182 ret = __dwc3_gadget_ep_enable(dep, &dwc3_gadget_ep0_desc);
1183 if (ret) {
1184 dev_err(dwc->dev, "failed to enable %s\n", dep->name);
1185 goto err1;
1186 }
1187
1188 /* begin to receive SETUP packets */
Felipe Balbi32e132e2011-08-27 22:28:36 +03001189 dwc->ep0state = EP0_SETUP_PHASE;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001190 dwc3_ep0_out_start(dwc);
1191
1192 spin_unlock_irqrestore(&dwc->lock, flags);
1193
1194 return 0;
1195
1196err1:
1197 __dwc3_gadget_ep_disable(dwc->eps[0]);
1198
1199err0:
1200 spin_unlock_irqrestore(&dwc->lock, flags);
1201
1202 return ret;
1203}
1204
1205static int dwc3_gadget_stop(struct usb_gadget *g,
1206 struct usb_gadget_driver *driver)
1207{
1208 struct dwc3 *dwc = gadget_to_dwc(g);
1209 unsigned long flags;
1210
1211 spin_lock_irqsave(&dwc->lock, flags);
1212
1213 __dwc3_gadget_ep_disable(dwc->eps[0]);
1214 __dwc3_gadget_ep_disable(dwc->eps[1]);
1215
1216 dwc->gadget_driver = NULL;
1217 dwc->gadget.dev.driver = NULL;
1218
1219 spin_unlock_irqrestore(&dwc->lock, flags);
1220
1221 return 0;
1222}
1223static const struct usb_gadget_ops dwc3_gadget_ops = {
1224 .get_frame = dwc3_gadget_get_frame,
1225 .wakeup = dwc3_gadget_wakeup,
1226 .set_selfpowered = dwc3_gadget_set_selfpowered,
1227 .pullup = dwc3_gadget_pullup,
1228 .udc_start = dwc3_gadget_start,
1229 .udc_stop = dwc3_gadget_stop,
1230};
1231
1232/* -------------------------------------------------------------------------- */
1233
1234static int __devinit dwc3_gadget_init_endpoints(struct dwc3 *dwc)
1235{
1236 struct dwc3_ep *dep;
1237 u8 epnum;
1238
1239 INIT_LIST_HEAD(&dwc->gadget.ep_list);
1240
1241 for (epnum = 0; epnum < DWC3_ENDPOINTS_NUM; epnum++) {
1242 dep = kzalloc(sizeof(*dep), GFP_KERNEL);
1243 if (!dep) {
1244 dev_err(dwc->dev, "can't allocate endpoint %d\n",
1245 epnum);
1246 return -ENOMEM;
1247 }
1248
1249 dep->dwc = dwc;
1250 dep->number = epnum;
1251 dwc->eps[epnum] = dep;
1252
1253 snprintf(dep->name, sizeof(dep->name), "ep%d%s", epnum >> 1,
1254 (epnum & 1) ? "in" : "out");
1255 dep->endpoint.name = dep->name;
1256 dep->direction = (epnum & 1);
1257
1258 if (epnum == 0 || epnum == 1) {
1259 dep->endpoint.maxpacket = 512;
1260 dep->endpoint.ops = &dwc3_gadget_ep0_ops;
1261 if (!epnum)
1262 dwc->gadget.ep0 = &dep->endpoint;
1263 } else {
1264 int ret;
1265
1266 dep->endpoint.maxpacket = 1024;
1267 dep->endpoint.ops = &dwc3_gadget_ep_ops;
1268 list_add_tail(&dep->endpoint.ep_list,
1269 &dwc->gadget.ep_list);
1270
1271 ret = dwc3_alloc_trb_pool(dep);
1272 if (ret) {
1273 dev_err(dwc->dev, "%s: failed to allocate TRB pool\n", dep->name);
1274 return ret;
1275 }
1276 }
1277 INIT_LIST_HEAD(&dep->request_list);
1278 INIT_LIST_HEAD(&dep->req_queued);
1279 }
1280
1281 return 0;
1282}
1283
1284static void dwc3_gadget_free_endpoints(struct dwc3 *dwc)
1285{
1286 struct dwc3_ep *dep;
1287 u8 epnum;
1288
1289 for (epnum = 0; epnum < DWC3_ENDPOINTS_NUM; epnum++) {
1290 dep = dwc->eps[epnum];
1291 dwc3_free_trb_pool(dep);
1292
1293 if (epnum != 0 && epnum != 1)
1294 list_del(&dep->endpoint.ep_list);
1295
1296 kfree(dep);
1297 }
1298}
1299
1300static void dwc3_gadget_release(struct device *dev)
1301{
1302 dev_dbg(dev, "%s\n", __func__);
1303}
1304
1305/* -------------------------------------------------------------------------- */
1306static int dwc3_cleanup_done_reqs(struct dwc3 *dwc, struct dwc3_ep *dep,
1307 const struct dwc3_event_depevt *event, int status)
1308{
1309 struct dwc3_request *req;
1310 struct dwc3_trb trb;
1311 unsigned int count;
1312 unsigned int s_pkt = 0;
1313
1314 do {
1315 req = next_request(&dep->req_queued);
1316 if (!req)
1317 break;
1318
1319 dwc3_trb_to_nat(req->trb, &trb);
1320
Sebastian Andrzej Siewior679dc462011-08-19 19:59:12 +02001321 if (trb.hwo && status != -ESHUTDOWN)
1322 /*
1323 * We continue despite the error. There is not much we
1324 * can do. If we don't clean in up we loop for ever. If
1325 * we skip the TRB than it gets overwritten reused after
1326 * a while since we use them in a ring buffer. a BUG()
1327 * would help. Lets hope that if this occures, someone
1328 * fixes the root cause instead of looking away :)
1329 */
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001330 dev_err(dwc->dev, "%s's TRB (%p) still owned by HW\n",
1331 dep->name, req->trb);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001332 count = trb.length;
1333
1334 if (dep->direction) {
1335 if (count) {
1336 dev_err(dwc->dev, "incomplete IN transfer %s\n",
1337 dep->name);
1338 status = -ECONNRESET;
1339 }
1340 } else {
1341 if (count && (event->status & DEPEVT_STATUS_SHORT))
1342 s_pkt = 1;
1343 }
1344
1345 /*
1346 * We assume here we will always receive the entire data block
1347 * which we should receive. Meaning, if we program RX to
1348 * receive 4K but we receive only 2K, we assume that's all we
1349 * should receive and we simply bounce the request back to the
1350 * gadget driver for further processing.
1351 */
1352 req->request.actual += req->request.length - count;
1353 dwc3_gadget_giveback(dep, req, status);
1354 if (s_pkt)
1355 break;
1356 if ((event->status & DEPEVT_STATUS_LST) && trb.lst)
1357 break;
1358 if ((event->status & DEPEVT_STATUS_IOC) && trb.ioc)
1359 break;
1360 } while (1);
1361
1362 if ((event->status & DEPEVT_STATUS_IOC) && trb.ioc)
1363 return 0;
1364 return 1;
1365}
1366
1367static void dwc3_endpoint_transfer_complete(struct dwc3 *dwc,
1368 struct dwc3_ep *dep, const struct dwc3_event_depevt *event,
1369 int start_new)
1370{
1371 unsigned status = 0;
1372 int clean_busy;
1373
1374 if (event->status & DEPEVT_STATUS_BUSERR)
1375 status = -ECONNRESET;
1376
1377 clean_busy = dwc3_cleanup_done_reqs(dwc, dep, event, status);
Sebastian Andrzej Siewior4df39772011-08-22 17:42:18 +02001378 if (clean_busy) {
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001379 dep->flags &= ~DWC3_EP_BUSY;
Sebastian Andrzej Siewior4df39772011-08-22 17:42:18 +02001380 dep->res_trans_idx = 0;
1381 }
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001382}
1383
1384static void dwc3_gadget_start_isoc(struct dwc3 *dwc,
1385 struct dwc3_ep *dep, const struct dwc3_event_depevt *event)
1386{
1387 u32 uf;
1388
1389 if (list_empty(&dep->request_list)) {
1390 dev_vdbg(dwc->dev, "ISOC ep %s run out for requests.\n",
1391 dep->name);
1392 return;
1393 }
1394
1395 if (event->parameters) {
1396 u32 mask;
1397
1398 mask = ~(dep->interval - 1);
1399 uf = event->parameters & mask;
1400 /* 4 micro frames in the future */
1401 uf += dep->interval * 4;
1402 } else {
1403 uf = 0;
1404 }
1405
1406 __dwc3_gadget_kick_transfer(dep, uf, 1);
1407}
1408
1409static void dwc3_process_ep_cmd_complete(struct dwc3_ep *dep,
1410 const struct dwc3_event_depevt *event)
1411{
1412 struct dwc3 *dwc = dep->dwc;
1413 struct dwc3_event_depevt mod_ev = *event;
1414
1415 /*
1416 * We were asked to remove one requests. It is possible that this
1417 * request and a few other were started together and have the same
1418 * transfer index. Since we stopped the complete endpoint we don't
1419 * know how many requests were already completed (and not yet)
1420 * reported and how could be done (later). We purge them all until
1421 * the end of the list.
1422 */
1423 mod_ev.status = DEPEVT_STATUS_LST;
1424 dwc3_cleanup_done_reqs(dwc, dep, &mod_ev, -ESHUTDOWN);
1425 dep->flags &= ~DWC3_EP_BUSY;
1426 /* pending requets are ignored and are queued on XferNotReady */
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001427}
1428
1429static void dwc3_ep_cmd_compl(struct dwc3_ep *dep,
1430 const struct dwc3_event_depevt *event)
1431{
1432 u32 param = event->parameters;
1433 u32 cmd_type = (param >> 8) & ((1 << 5) - 1);
1434
1435 switch (cmd_type) {
1436 case DWC3_DEPCMD_ENDTRANSFER:
1437 dwc3_process_ep_cmd_complete(dep, event);
1438 break;
1439 case DWC3_DEPCMD_STARTTRANSFER:
1440 dep->res_trans_idx = param & 0x7f;
1441 break;
1442 default:
1443 printk(KERN_ERR "%s() unknown /unexpected type: %d\n",
1444 __func__, cmd_type);
1445 break;
1446 };
1447}
1448
1449static void dwc3_endpoint_interrupt(struct dwc3 *dwc,
1450 const struct dwc3_event_depevt *event)
1451{
1452 struct dwc3_ep *dep;
1453 u8 epnum = event->endpoint_number;
1454
1455 dep = dwc->eps[epnum];
1456
1457 dev_vdbg(dwc->dev, "%s: %s\n", dep->name,
1458 dwc3_ep_event_string(event->endpoint_event));
1459
1460 if (epnum == 0 || epnum == 1) {
1461 dwc3_ep0_interrupt(dwc, event);
1462 return;
1463 }
1464
1465 switch (event->endpoint_event) {
1466 case DWC3_DEPEVT_XFERCOMPLETE:
1467 if (usb_endpoint_xfer_isoc(dep->desc)) {
1468 dev_dbg(dwc->dev, "%s is an Isochronous endpoint\n",
1469 dep->name);
1470 return;
1471 }
1472
1473 dwc3_endpoint_transfer_complete(dwc, dep, event, 1);
1474 break;
1475 case DWC3_DEPEVT_XFERINPROGRESS:
1476 if (!usb_endpoint_xfer_isoc(dep->desc)) {
1477 dev_dbg(dwc->dev, "%s is not an Isochronous endpoint\n",
1478 dep->name);
1479 return;
1480 }
1481
1482 dwc3_endpoint_transfer_complete(dwc, dep, event, 0);
1483 break;
1484 case DWC3_DEPEVT_XFERNOTREADY:
1485 if (usb_endpoint_xfer_isoc(dep->desc)) {
1486 dwc3_gadget_start_isoc(dwc, dep, event);
1487 } else {
1488 int ret;
1489
1490 dev_vdbg(dwc->dev, "%s: reason %s\n",
1491 dep->name, event->status
1492 ? "Transfer Active"
1493 : "Transfer Not Active");
1494
1495 ret = __dwc3_gadget_kick_transfer(dep, 0, 1);
1496 if (!ret || ret == -EBUSY)
1497 return;
1498
1499 dev_dbg(dwc->dev, "%s: failed to kick transfers\n",
1500 dep->name);
1501 }
1502
1503 break;
1504 case DWC3_DEPEVT_RXTXFIFOEVT:
1505 dev_dbg(dwc->dev, "%s FIFO Overrun\n", dep->name);
1506 break;
1507 case DWC3_DEPEVT_STREAMEVT:
1508 dev_dbg(dwc->dev, "%s Stream Event\n", dep->name);
1509 break;
1510 case DWC3_DEPEVT_EPCMDCMPLT:
1511 dwc3_ep_cmd_compl(dep, event);
1512 break;
1513 }
1514}
1515
1516static void dwc3_disconnect_gadget(struct dwc3 *dwc)
1517{
1518 if (dwc->gadget_driver && dwc->gadget_driver->disconnect) {
1519 spin_unlock(&dwc->lock);
1520 dwc->gadget_driver->disconnect(&dwc->gadget);
1521 spin_lock(&dwc->lock);
1522 }
1523}
1524
1525static void dwc3_stop_active_transfer(struct dwc3 *dwc, u32 epnum)
1526{
1527 struct dwc3_ep *dep;
1528 struct dwc3_gadget_ep_cmd_params params;
1529 u32 cmd;
1530 int ret;
1531
1532 dep = dwc->eps[epnum];
1533
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +02001534 WARN_ON(!dep->res_trans_idx);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001535 if (dep->res_trans_idx) {
1536 cmd = DWC3_DEPCMD_ENDTRANSFER;
1537 cmd |= DWC3_DEPCMD_HIPRI_FORCERM | DWC3_DEPCMD_CMDIOC;
1538 cmd |= DWC3_DEPCMD_PARAM(dep->res_trans_idx);
1539 memset(&params, 0, sizeof(params));
1540 ret = dwc3_send_gadget_ep_cmd(dwc, dep->number, cmd, &params);
1541 WARN_ON_ONCE(ret);
Sebastian Andrzej Siewior4df39772011-08-22 17:42:18 +02001542 dep->res_trans_idx = 0;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001543 }
1544}
1545
1546static void dwc3_stop_active_transfers(struct dwc3 *dwc)
1547{
1548 u32 epnum;
1549
1550 for (epnum = 2; epnum < DWC3_ENDPOINTS_NUM; epnum++) {
1551 struct dwc3_ep *dep;
1552
1553 dep = dwc->eps[epnum];
1554 if (!(dep->flags & DWC3_EP_ENABLED))
1555 continue;
1556
Sebastian Andrzej Siewiorb55db3b2011-08-29 13:56:37 +02001557 dwc3_remove_requests(dwc, dep);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001558 }
1559}
1560
1561static void dwc3_clear_stall_all_ep(struct dwc3 *dwc)
1562{
1563 u32 epnum;
1564
1565 for (epnum = 1; epnum < DWC3_ENDPOINTS_NUM; epnum++) {
1566 struct dwc3_ep *dep;
1567 struct dwc3_gadget_ep_cmd_params params;
1568 int ret;
1569
1570 dep = dwc->eps[epnum];
1571
1572 if (!(dep->flags & DWC3_EP_STALL))
1573 continue;
1574
1575 dep->flags &= ~DWC3_EP_STALL;
1576
1577 memset(&params, 0, sizeof(params));
1578 ret = dwc3_send_gadget_ep_cmd(dwc, dep->number,
1579 DWC3_DEPCMD_CLEARSTALL, &params);
1580 WARN_ON_ONCE(ret);
1581 }
1582}
1583
1584static void dwc3_gadget_disconnect_interrupt(struct dwc3 *dwc)
1585{
1586 dev_vdbg(dwc->dev, "%s\n", __func__);
1587#if 0
1588 XXX
1589 U1/U2 is powersave optimization. Skip it for now. Anyway we need to
1590 enable it before we can disable it.
1591
1592 reg = dwc3_readl(dwc->regs, DWC3_DCTL);
1593 reg &= ~DWC3_DCTL_INITU1ENA;
1594 dwc3_writel(dwc->regs, DWC3_DCTL, reg);
1595
1596 reg &= ~DWC3_DCTL_INITU2ENA;
1597 dwc3_writel(dwc->regs, DWC3_DCTL, reg);
1598#endif
1599
1600 dwc3_stop_active_transfers(dwc);
1601 dwc3_disconnect_gadget(dwc);
Paul Zimmermana6544d02011-09-30 10:58:42 +03001602 dwc->start_config_issued = false;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001603
1604 dwc->gadget.speed = USB_SPEED_UNKNOWN;
1605}
1606
1607static void dwc3_gadget_usb3_phy_power(struct dwc3 *dwc, int on)
1608{
1609 u32 reg;
1610
1611 reg = dwc3_readl(dwc->regs, DWC3_GUSB3PIPECTL(0));
1612
1613 if (on)
1614 reg &= ~DWC3_GUSB3PIPECTL_SUSPHY;
1615 else
1616 reg |= DWC3_GUSB3PIPECTL_SUSPHY;
1617
1618 dwc3_writel(dwc->regs, DWC3_GUSB3PIPECTL(0), reg);
1619}
1620
1621static void dwc3_gadget_usb2_phy_power(struct dwc3 *dwc, int on)
1622{
1623 u32 reg;
1624
1625 reg = dwc3_readl(dwc->regs, DWC3_GUSB2PHYCFG(0));
1626
1627 if (on)
1628 reg &= ~DWC3_GUSB2PHYCFG_SUSPHY;
1629 else
1630 reg |= DWC3_GUSB2PHYCFG_SUSPHY;
1631
1632 dwc3_writel(dwc->regs, DWC3_GUSB2PHYCFG(0), reg);
1633}
1634
1635static void dwc3_gadget_reset_interrupt(struct dwc3 *dwc)
1636{
1637 u32 reg;
1638
1639 dev_vdbg(dwc->dev, "%s\n", __func__);
1640
1641 /* Enable PHYs */
1642 dwc3_gadget_usb2_phy_power(dwc, true);
1643 dwc3_gadget_usb3_phy_power(dwc, true);
1644
1645 if (dwc->gadget.speed != USB_SPEED_UNKNOWN)
1646 dwc3_disconnect_gadget(dwc);
1647
1648 reg = dwc3_readl(dwc->regs, DWC3_DCTL);
1649 reg &= ~DWC3_DCTL_TSTCTRL_MASK;
1650 dwc3_writel(dwc->regs, DWC3_DCTL, reg);
1651
1652 dwc3_stop_active_transfers(dwc);
1653 dwc3_clear_stall_all_ep(dwc);
Paul Zimmermana6544d02011-09-30 10:58:42 +03001654 dwc->start_config_issued = false;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001655
1656 /* Reset device address to zero */
1657 reg = dwc3_readl(dwc->regs, DWC3_DCFG);
1658 reg &= ~(DWC3_DCFG_DEVADDR_MASK);
1659 dwc3_writel(dwc->regs, DWC3_DCFG, reg);
1660
1661 /*
1662 * Wait for RxFifo to drain
1663 *
1664 * REVISIT probably shouldn't wait forever.
1665 * In case Hardware ends up in a screwed up
1666 * case, we error out, notify the user and,
1667 * maybe, WARN() or BUG() but leave the rest
1668 * of the kernel working fine.
1669 *
1670 * REVISIT the below is rather CPU intensive,
1671 * maybe we should read and if it doesn't work
1672 * sleep (not busy wait) for a few useconds.
1673 *
1674 * REVISIT why wait until the RXFIFO is empty anyway?
1675 */
1676 while (!(dwc3_readl(dwc->regs, DWC3_DSTS)
1677 & DWC3_DSTS_RXFIFOEMPTY))
1678 cpu_relax();
1679}
1680
1681static void dwc3_update_ram_clk_sel(struct dwc3 *dwc, u32 speed)
1682{
1683 u32 reg;
1684 u32 usb30_clock = DWC3_GCTL_CLK_BUS;
1685
1686 /*
1687 * We change the clock only at SS but I dunno why I would want to do
1688 * this. Maybe it becomes part of the power saving plan.
1689 */
1690
1691 if (speed != DWC3_DSTS_SUPERSPEED)
1692 return;
1693
1694 /*
1695 * RAMClkSel is reset to 0 after USB reset, so it must be reprogrammed
1696 * each time on Connect Done.
1697 */
1698 if (!usb30_clock)
1699 return;
1700
1701 reg = dwc3_readl(dwc->regs, DWC3_GCTL);
1702 reg |= DWC3_GCTL_RAMCLKSEL(usb30_clock);
1703 dwc3_writel(dwc->regs, DWC3_GCTL, reg);
1704}
1705
1706static void dwc3_gadget_disable_phy(struct dwc3 *dwc, u8 speed)
1707{
1708 switch (speed) {
1709 case USB_SPEED_SUPER:
1710 dwc3_gadget_usb2_phy_power(dwc, false);
1711 break;
1712 case USB_SPEED_HIGH:
1713 case USB_SPEED_FULL:
1714 case USB_SPEED_LOW:
1715 dwc3_gadget_usb3_phy_power(dwc, false);
1716 break;
1717 }
1718}
1719
1720static void dwc3_gadget_conndone_interrupt(struct dwc3 *dwc)
1721{
1722 struct dwc3_gadget_ep_cmd_params params;
1723 struct dwc3_ep *dep;
1724 int ret;
1725 u32 reg;
1726 u8 speed;
1727
1728 dev_vdbg(dwc->dev, "%s\n", __func__);
1729
1730 memset(&params, 0x00, sizeof(params));
1731
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001732 reg = dwc3_readl(dwc->regs, DWC3_DSTS);
1733 speed = reg & DWC3_DSTS_CONNECTSPD;
1734 dwc->speed = speed;
1735
1736 dwc3_update_ram_clk_sel(dwc, speed);
1737
1738 switch (speed) {
1739 case DWC3_DCFG_SUPERSPEED:
1740 dwc3_gadget_ep0_desc.wMaxPacketSize = cpu_to_le16(512);
1741 dwc->gadget.ep0->maxpacket = 512;
1742 dwc->gadget.speed = USB_SPEED_SUPER;
1743 break;
1744 case DWC3_DCFG_HIGHSPEED:
1745 dwc3_gadget_ep0_desc.wMaxPacketSize = cpu_to_le16(64);
1746 dwc->gadget.ep0->maxpacket = 64;
1747 dwc->gadget.speed = USB_SPEED_HIGH;
1748 break;
1749 case DWC3_DCFG_FULLSPEED2:
1750 case DWC3_DCFG_FULLSPEED1:
1751 dwc3_gadget_ep0_desc.wMaxPacketSize = cpu_to_le16(64);
1752 dwc->gadget.ep0->maxpacket = 64;
1753 dwc->gadget.speed = USB_SPEED_FULL;
1754 break;
1755 case DWC3_DCFG_LOWSPEED:
1756 dwc3_gadget_ep0_desc.wMaxPacketSize = cpu_to_le16(8);
1757 dwc->gadget.ep0->maxpacket = 8;
1758 dwc->gadget.speed = USB_SPEED_LOW;
1759 break;
1760 }
1761
1762 /* Disable unneded PHY */
1763 dwc3_gadget_disable_phy(dwc, dwc->gadget.speed);
1764
1765 dep = dwc->eps[0];
1766 ret = __dwc3_gadget_ep_enable(dep, &dwc3_gadget_ep0_desc);
1767 if (ret) {
1768 dev_err(dwc->dev, "failed to enable %s\n", dep->name);
1769 return;
1770 }
1771
1772 dep = dwc->eps[1];
1773 ret = __dwc3_gadget_ep_enable(dep, &dwc3_gadget_ep0_desc);
1774 if (ret) {
1775 dev_err(dwc->dev, "failed to enable %s\n", dep->name);
1776 return;
1777 }
1778
1779 /*
1780 * Configure PHY via GUSB3PIPECTLn if required.
1781 *
1782 * Update GTXFIFOSIZn
1783 *
1784 * In both cases reset values should be sufficient.
1785 */
1786}
1787
1788static void dwc3_gadget_wakeup_interrupt(struct dwc3 *dwc)
1789{
1790 dev_vdbg(dwc->dev, "%s\n", __func__);
1791
1792 /*
1793 * TODO take core out of low power mode when that's
1794 * implemented.
1795 */
1796
1797 dwc->gadget_driver->resume(&dwc->gadget);
1798}
1799
1800static void dwc3_gadget_linksts_change_interrupt(struct dwc3 *dwc,
1801 unsigned int evtinfo)
1802{
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001803 /* The fith bit says SuperSpeed yes or no. */
1804 dwc->link_state = evtinfo & DWC3_LINK_STATE_MASK;
Felipe Balbi6edd5bd2011-09-08 21:18:47 +03001805
1806 dev_vdbg(dwc->dev, "%s link %d\n", __func__, dwc->link_state);
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001807}
1808
1809static void dwc3_gadget_interrupt(struct dwc3 *dwc,
1810 const struct dwc3_event_devt *event)
1811{
1812 switch (event->type) {
1813 case DWC3_DEVICE_EVENT_DISCONNECT:
1814 dwc3_gadget_disconnect_interrupt(dwc);
1815 break;
1816 case DWC3_DEVICE_EVENT_RESET:
1817 dwc3_gadget_reset_interrupt(dwc);
1818 break;
1819 case DWC3_DEVICE_EVENT_CONNECT_DONE:
1820 dwc3_gadget_conndone_interrupt(dwc);
1821 break;
1822 case DWC3_DEVICE_EVENT_WAKEUP:
1823 dwc3_gadget_wakeup_interrupt(dwc);
1824 break;
1825 case DWC3_DEVICE_EVENT_LINK_STATUS_CHANGE:
1826 dwc3_gadget_linksts_change_interrupt(dwc, event->event_info);
1827 break;
1828 case DWC3_DEVICE_EVENT_EOPF:
1829 dev_vdbg(dwc->dev, "End of Periodic Frame\n");
1830 break;
1831 case DWC3_DEVICE_EVENT_SOF:
1832 dev_vdbg(dwc->dev, "Start of Periodic Frame\n");
1833 break;
1834 case DWC3_DEVICE_EVENT_ERRATIC_ERROR:
1835 dev_vdbg(dwc->dev, "Erratic Error\n");
1836 break;
1837 case DWC3_DEVICE_EVENT_CMD_CMPL:
1838 dev_vdbg(dwc->dev, "Command Complete\n");
1839 break;
1840 case DWC3_DEVICE_EVENT_OVERFLOW:
1841 dev_vdbg(dwc->dev, "Overflow\n");
1842 break;
1843 default:
1844 dev_dbg(dwc->dev, "UNKNOWN IRQ %d\n", event->type);
1845 }
1846}
1847
1848static void dwc3_process_event_entry(struct dwc3 *dwc,
1849 const union dwc3_event *event)
1850{
1851 /* Endpoint IRQ, handle it and return early */
1852 if (event->type.is_devspec == 0) {
1853 /* depevt */
1854 return dwc3_endpoint_interrupt(dwc, &event->depevt);
1855 }
1856
1857 switch (event->type.type) {
1858 case DWC3_EVENT_TYPE_DEV:
1859 dwc3_gadget_interrupt(dwc, &event->devt);
1860 break;
1861 /* REVISIT what to do with Carkit and I2C events ? */
1862 default:
1863 dev_err(dwc->dev, "UNKNOWN IRQ type %d\n", event->raw);
1864 }
1865}
1866
1867static irqreturn_t dwc3_process_event_buf(struct dwc3 *dwc, u32 buf)
1868{
1869 struct dwc3_event_buffer *evt;
1870 int left;
1871 u32 count;
1872
1873 count = dwc3_readl(dwc->regs, DWC3_GEVNTCOUNT(buf));
1874 count &= DWC3_GEVNTCOUNT_MASK;
1875 if (!count)
1876 return IRQ_NONE;
1877
1878 evt = dwc->ev_buffs[buf];
1879 left = count;
1880
1881 while (left > 0) {
1882 union dwc3_event event;
1883
1884 memcpy(&event.raw, (evt->buf + evt->lpos), sizeof(event.raw));
1885 dwc3_process_event_entry(dwc, &event);
1886 /*
1887 * XXX we wrap around correctly to the next entry as almost all
1888 * entries are 4 bytes in size. There is one entry which has 12
1889 * bytes which is a regular entry followed by 8 bytes data. ATM
1890 * I don't know how things are organized if were get next to the
1891 * a boundary so I worry about that once we try to handle that.
1892 */
1893 evt->lpos = (evt->lpos + 4) % DWC3_EVENT_BUFFERS_SIZE;
1894 left -= 4;
1895
1896 dwc3_writel(dwc->regs, DWC3_GEVNTCOUNT(buf), 4);
1897 }
1898
1899 return IRQ_HANDLED;
1900}
1901
1902static irqreturn_t dwc3_interrupt(int irq, void *_dwc)
1903{
1904 struct dwc3 *dwc = _dwc;
1905 int i;
1906 irqreturn_t ret = IRQ_NONE;
1907
1908 spin_lock(&dwc->lock);
1909
1910 for (i = 0; i < DWC3_EVENT_BUFFERS_NUM; i++) {
1911 irqreturn_t status;
1912
1913 status = dwc3_process_event_buf(dwc, i);
1914 if (status == IRQ_HANDLED)
1915 ret = status;
1916 }
1917
1918 spin_unlock(&dwc->lock);
1919
1920 return ret;
1921}
1922
1923/**
1924 * dwc3_gadget_init - Initializes gadget related registers
1925 * @dwc: Pointer to out controller context structure
1926 *
1927 * Returns 0 on success otherwise negative errno.
1928 */
1929int __devinit dwc3_gadget_init(struct dwc3 *dwc)
1930{
1931 u32 reg;
1932 int ret;
1933 int irq;
1934
1935 dwc->ctrl_req = dma_alloc_coherent(dwc->dev, sizeof(*dwc->ctrl_req),
1936 &dwc->ctrl_req_addr, GFP_KERNEL);
1937 if (!dwc->ctrl_req) {
1938 dev_err(dwc->dev, "failed to allocate ctrl request\n");
1939 ret = -ENOMEM;
1940 goto err0;
1941 }
1942
1943 dwc->ep0_trb = dma_alloc_coherent(dwc->dev, sizeof(*dwc->ep0_trb),
1944 &dwc->ep0_trb_addr, GFP_KERNEL);
1945 if (!dwc->ep0_trb) {
1946 dev_err(dwc->dev, "failed to allocate ep0 trb\n");
1947 ret = -ENOMEM;
1948 goto err1;
1949 }
1950
1951 dwc->setup_buf = dma_alloc_coherent(dwc->dev,
1952 sizeof(*dwc->setup_buf) * 2,
1953 &dwc->setup_buf_addr, GFP_KERNEL);
1954 if (!dwc->setup_buf) {
1955 dev_err(dwc->dev, "failed to allocate setup buffer\n");
1956 ret = -ENOMEM;
1957 goto err2;
1958 }
1959
Felipe Balbi64e96342011-08-27 22:07:53 +03001960 dwc->ep0_bounce = dma_alloc_coherent(dwc->dev,
1961 512, &dwc->ep0_bounce_addr, GFP_KERNEL);
1962 if (!dwc->ep0_bounce) {
1963 dev_err(dwc->dev, "failed to allocate ep0 bounce buffer\n");
1964 ret = -ENOMEM;
1965 goto err3;
1966 }
1967
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001968 dev_set_name(&dwc->gadget.dev, "gadget");
1969
1970 dwc->gadget.ops = &dwc3_gadget_ops;
1971 dwc->gadget.is_dualspeed = true;
1972 dwc->gadget.speed = USB_SPEED_UNKNOWN;
1973 dwc->gadget.dev.parent = dwc->dev;
1974
1975 dma_set_coherent_mask(&dwc->gadget.dev, dwc->dev->coherent_dma_mask);
1976
1977 dwc->gadget.dev.dma_parms = dwc->dev->dma_parms;
1978 dwc->gadget.dev.dma_mask = dwc->dev->dma_mask;
1979 dwc->gadget.dev.release = dwc3_gadget_release;
1980 dwc->gadget.name = "dwc3-gadget";
1981
1982 /*
1983 * REVISIT: Here we should clear all pending IRQs to be
1984 * sure we're starting from a well known location.
1985 */
1986
1987 ret = dwc3_gadget_init_endpoints(dwc);
1988 if (ret)
Felipe Balbi64e96342011-08-27 22:07:53 +03001989 goto err4;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001990
1991 irq = platform_get_irq(to_platform_device(dwc->dev), 0);
1992
1993 ret = request_irq(irq, dwc3_interrupt, IRQF_SHARED,
1994 "dwc3", dwc);
1995 if (ret) {
1996 dev_err(dwc->dev, "failed to request irq #%d --> %d\n",
1997 irq, ret);
Felipe Balbi64e96342011-08-27 22:07:53 +03001998 goto err5;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03001999 }
2000
2001 /* Enable all but Start and End of Frame IRQs */
2002 reg = (DWC3_DEVTEN_VNDRDEVTSTRCVEDEN |
2003 DWC3_DEVTEN_EVNTOVERFLOWEN |
2004 DWC3_DEVTEN_CMDCMPLTEN |
2005 DWC3_DEVTEN_ERRTICERREN |
2006 DWC3_DEVTEN_WKUPEVTEN |
2007 DWC3_DEVTEN_ULSTCNGEN |
2008 DWC3_DEVTEN_CONNECTDONEEN |
2009 DWC3_DEVTEN_USBRSTEN |
2010 DWC3_DEVTEN_DISCONNEVTEN);
2011 dwc3_writel(dwc->regs, DWC3_DEVTEN, reg);
2012
2013 ret = device_register(&dwc->gadget.dev);
2014 if (ret) {
2015 dev_err(dwc->dev, "failed to register gadget device\n");
2016 put_device(&dwc->gadget.dev);
Felipe Balbi64e96342011-08-27 22:07:53 +03002017 goto err6;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002018 }
2019
2020 ret = usb_add_gadget_udc(dwc->dev, &dwc->gadget);
2021 if (ret) {
2022 dev_err(dwc->dev, "failed to register udc\n");
Felipe Balbi64e96342011-08-27 22:07:53 +03002023 goto err7;
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002024 }
2025
2026 return 0;
2027
Felipe Balbi64e96342011-08-27 22:07:53 +03002028err7:
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002029 device_unregister(&dwc->gadget.dev);
2030
Felipe Balbi64e96342011-08-27 22:07:53 +03002031err6:
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002032 dwc3_writel(dwc->regs, DWC3_DEVTEN, 0x00);
2033 free_irq(irq, dwc);
2034
Felipe Balbi64e96342011-08-27 22:07:53 +03002035err5:
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002036 dwc3_gadget_free_endpoints(dwc);
2037
Felipe Balbi64e96342011-08-27 22:07:53 +03002038err4:
2039 dma_free_coherent(dwc->dev, 512, dwc->ep0_bounce,
2040 dwc->ep0_bounce_addr);
2041
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002042err3:
2043 dma_free_coherent(dwc->dev, sizeof(*dwc->setup_buf) * 2,
2044 dwc->setup_buf, dwc->setup_buf_addr);
2045
2046err2:
2047 dma_free_coherent(dwc->dev, sizeof(*dwc->ep0_trb),
2048 dwc->ep0_trb, dwc->ep0_trb_addr);
2049
2050err1:
2051 dma_free_coherent(dwc->dev, sizeof(*dwc->ctrl_req),
2052 dwc->ctrl_req, dwc->ctrl_req_addr);
2053
2054err0:
2055 return ret;
2056}
2057
2058void dwc3_gadget_exit(struct dwc3 *dwc)
2059{
2060 int irq;
2061 int i;
2062
2063 usb_del_gadget_udc(&dwc->gadget);
2064 irq = platform_get_irq(to_platform_device(dwc->dev), 0);
2065
2066 dwc3_writel(dwc->regs, DWC3_DEVTEN, 0x00);
2067 free_irq(irq, dwc);
2068
2069 for (i = 0; i < ARRAY_SIZE(dwc->eps); i++)
2070 __dwc3_gadget_ep_disable(dwc->eps[i]);
2071
2072 dwc3_gadget_free_endpoints(dwc);
2073
Felipe Balbi64e96342011-08-27 22:07:53 +03002074 dma_free_coherent(dwc->dev, 512, dwc->ep0_bounce,
2075 dwc->ep0_bounce_addr);
2076
Felipe Balbi4dc64e52011-08-19 18:10:58 +03002077 dma_free_coherent(dwc->dev, sizeof(*dwc->setup_buf) * 2,
2078 dwc->setup_buf, dwc->setup_buf_addr);
2079
2080 dma_free_coherent(dwc->dev, sizeof(*dwc->ep0_trb),
2081 dwc->ep0_trb, dwc->ep0_trb_addr);
2082
2083 dma_free_coherent(dwc->dev, sizeof(*dwc->ctrl_req),
2084 dwc->ctrl_req, dwc->ctrl_req_addr);
2085
2086 device_unregister(&dwc->gadget.dev);
2087}