|  | /* | 
|  | * amd5536.c -- AMD 5536 UDC high/full speed USB device controller | 
|  | * | 
|  | * Copyright (C) 2005-2007 AMD (http://www.amd.com) | 
|  | * Author: Thomas Dahlmann | 
|  | * | 
|  | * This program is free software; you can redistribute it and/or modify | 
|  | * it under the terms of the GNU General Public License as published by | 
|  | * the Free Software Foundation; either version 2 of the License, or | 
|  | * (at your option) any later version. | 
|  | */ | 
|  |  | 
|  | /* | 
|  | * The AMD5536 UDC is part of the x86 southbridge AMD Geode CS5536. | 
|  | * It is a USB Highspeed DMA capable USB device controller. Beside ep0 it | 
|  | * provides 4 IN and 4 OUT endpoints (bulk or interrupt type). | 
|  | * | 
|  | * Make sure that UDC is assigned to port 4 by BIOS settings (port can also | 
|  | * be used as host port) and UOC bits PAD_EN and APU are set (should be done | 
|  | * by BIOS init). | 
|  | * | 
|  | * UDC DMA requires 32-bit aligned buffers so DMA with gadget ether does not | 
|  | * work without updating NET_IP_ALIGN. Or PIO mode (module param "use_dma=0") | 
|  | * can be used with gadget ether. | 
|  | */ | 
|  |  | 
|  | /* debug control */ | 
|  | /* #define UDC_VERBOSE */ | 
|  |  | 
|  | /* Driver strings */ | 
|  | #define UDC_MOD_DESCRIPTION		"AMD 5536 UDC - USB Device Controller" | 
|  | #define UDC_DRIVER_VERSION_STRING	"01.00.0206" | 
|  |  | 
|  | /* system */ | 
|  | #include <linux/module.h> | 
|  | #include <linux/pci.h> | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/delay.h> | 
|  | #include <linux/ioport.h> | 
|  | #include <linux/sched.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/errno.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/timer.h> | 
|  | #include <linux/list.h> | 
|  | #include <linux/interrupt.h> | 
|  | #include <linux/ioctl.h> | 
|  | #include <linux/fs.h> | 
|  | #include <linux/dmapool.h> | 
|  | #include <linux/moduleparam.h> | 
|  | #include <linux/device.h> | 
|  | #include <linux/io.h> | 
|  | #include <linux/irq.h> | 
|  | #include <linux/prefetch.h> | 
|  |  | 
|  | #include <asm/byteorder.h> | 
|  | #include <asm/unaligned.h> | 
|  |  | 
|  | /* gadget stack */ | 
|  | #include <linux/usb/ch9.h> | 
|  | #include <linux/usb/gadget.h> | 
|  |  | 
|  | /* udc specific */ | 
|  | #include "amd5536udc.h" | 
|  |  | 
|  |  | 
|  | static void udc_tasklet_disconnect(unsigned long); | 
|  | static void empty_req_queue(struct udc_ep *); | 
|  | static int udc_probe(struct udc *dev); | 
|  | static void udc_basic_init(struct udc *dev); | 
|  | static void udc_setup_endpoints(struct udc *dev); | 
|  | static void udc_soft_reset(struct udc *dev); | 
|  | static struct udc_request *udc_alloc_bna_dummy(struct udc_ep *ep); | 
|  | static void udc_free_request(struct usb_ep *usbep, struct usb_request *usbreq); | 
|  | static int udc_free_dma_chain(struct udc *dev, struct udc_request *req); | 
|  | static int udc_create_dma_chain(struct udc_ep *ep, struct udc_request *req, | 
|  | unsigned long buf_len, gfp_t gfp_flags); | 
|  | static int udc_remote_wakeup(struct udc *dev); | 
|  | static int udc_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id); | 
|  | static void udc_pci_remove(struct pci_dev *pdev); | 
|  |  | 
|  | /* description */ | 
|  | static const char mod_desc[] = UDC_MOD_DESCRIPTION; | 
|  | static const char name[] = "amd5536udc"; | 
|  |  | 
|  | /* structure to hold endpoint function pointers */ | 
|  | static const struct usb_ep_ops udc_ep_ops; | 
|  |  | 
|  | /* received setup data */ | 
|  | static union udc_setup_data setup_data; | 
|  |  | 
|  | /* pointer to device object */ | 
|  | static struct udc *udc; | 
|  |  | 
|  | /* irq spin lock for soft reset */ | 
|  | static DEFINE_SPINLOCK(udc_irq_spinlock); | 
|  | /* stall spin lock */ | 
|  | static DEFINE_SPINLOCK(udc_stall_spinlock); | 
|  |  | 
|  | /* | 
|  | * slave mode: pending bytes in rx fifo after nyet, | 
|  | * used if EPIN irq came but no req was available | 
|  | */ | 
|  | static unsigned int udc_rxfifo_pending; | 
|  |  | 
|  | /* count soft resets after suspend to avoid loop */ | 
|  | static int soft_reset_occured; | 
|  | static int soft_reset_after_usbreset_occured; | 
|  |  | 
|  | /* timer */ | 
|  | static struct timer_list udc_timer; | 
|  | static int stop_timer; | 
|  |  | 
|  | /* set_rde -- Is used to control enabling of RX DMA. Problem is | 
|  | * that UDC has only one bit (RDE) to enable/disable RX DMA for | 
|  | * all OUT endpoints. So we have to handle race conditions like | 
|  | * when OUT data reaches the fifo but no request was queued yet. | 
|  | * This cannot be solved by letting the RX DMA disabled until a | 
|  | * request gets queued because there may be other OUT packets | 
|  | * in the FIFO (important for not blocking control traffic). | 
|  | * The value of set_rde controls the correspondig timer. | 
|  | * | 
|  | * set_rde -1 == not used, means it is alloed to be set to 0 or 1 | 
|  | * set_rde  0 == do not touch RDE, do no start the RDE timer | 
|  | * set_rde  1 == timer function will look whether FIFO has data | 
|  | * set_rde  2 == set by timer function to enable RX DMA on next call | 
|  | */ | 
|  | static int set_rde = -1; | 
|  |  | 
|  | static DECLARE_COMPLETION(on_exit); | 
|  | static struct timer_list udc_pollstall_timer; | 
|  | static int stop_pollstall_timer; | 
|  | static DECLARE_COMPLETION(on_pollstall_exit); | 
|  |  | 
|  | /* tasklet for usb disconnect */ | 
|  | static DECLARE_TASKLET(disconnect_tasklet, udc_tasklet_disconnect, | 
|  | (unsigned long) &udc); | 
|  |  | 
|  |  | 
|  | /* endpoint names used for print */ | 
|  | static const char ep0_string[] = "ep0in"; | 
|  | static const char *const ep_string[] = { | 
|  | ep0_string, | 
|  | "ep1in-int", "ep2in-bulk", "ep3in-bulk", "ep4in-bulk", "ep5in-bulk", | 
|  | "ep6in-bulk", "ep7in-bulk", "ep8in-bulk", "ep9in-bulk", "ep10in-bulk", | 
|  | "ep11in-bulk", "ep12in-bulk", "ep13in-bulk", "ep14in-bulk", | 
|  | "ep15in-bulk", "ep0out", "ep1out-bulk", "ep2out-bulk", "ep3out-bulk", | 
|  | "ep4out-bulk", "ep5out-bulk", "ep6out-bulk", "ep7out-bulk", | 
|  | "ep8out-bulk", "ep9out-bulk", "ep10out-bulk", "ep11out-bulk", | 
|  | "ep12out-bulk", "ep13out-bulk", "ep14out-bulk", "ep15out-bulk" | 
|  | }; | 
|  |  | 
|  | /* DMA usage flag */ | 
|  | static bool use_dma = 1; | 
|  | /* packet per buffer dma */ | 
|  | static bool use_dma_ppb = 1; | 
|  | /* with per descr. update */ | 
|  | static bool use_dma_ppb_du; | 
|  | /* buffer fill mode */ | 
|  | static int use_dma_bufferfill_mode; | 
|  | /* full speed only mode */ | 
|  | static bool use_fullspeed; | 
|  | /* tx buffer size for high speed */ | 
|  | static unsigned long hs_tx_buf = UDC_EPIN_BUFF_SIZE; | 
|  |  | 
|  | /* module parameters */ | 
|  | module_param(use_dma, bool, S_IRUGO); | 
|  | MODULE_PARM_DESC(use_dma, "true for DMA"); | 
|  | module_param(use_dma_ppb, bool, S_IRUGO); | 
|  | MODULE_PARM_DESC(use_dma_ppb, "true for DMA in packet per buffer mode"); | 
|  | module_param(use_dma_ppb_du, bool, S_IRUGO); | 
|  | MODULE_PARM_DESC(use_dma_ppb_du, | 
|  | "true for DMA in packet per buffer mode with descriptor update"); | 
|  | module_param(use_fullspeed, bool, S_IRUGO); | 
|  | MODULE_PARM_DESC(use_fullspeed, "true for fullspeed only"); | 
|  |  | 
|  | /*---------------------------------------------------------------------------*/ | 
|  | /* Prints UDC device registers and endpoint irq registers */ | 
|  | static void print_regs(struct udc *dev) | 
|  | { | 
|  | DBG(dev, "------- Device registers -------\n"); | 
|  | DBG(dev, "dev config     = %08x\n", readl(&dev->regs->cfg)); | 
|  | DBG(dev, "dev control    = %08x\n", readl(&dev->regs->ctl)); | 
|  | DBG(dev, "dev status     = %08x\n", readl(&dev->regs->sts)); | 
|  | DBG(dev, "\n"); | 
|  | DBG(dev, "dev int's      = %08x\n", readl(&dev->regs->irqsts)); | 
|  | DBG(dev, "dev intmask    = %08x\n", readl(&dev->regs->irqmsk)); | 
|  | DBG(dev, "\n"); | 
|  | DBG(dev, "dev ep int's   = %08x\n", readl(&dev->regs->ep_irqsts)); | 
|  | DBG(dev, "dev ep intmask = %08x\n", readl(&dev->regs->ep_irqmsk)); | 
|  | DBG(dev, "\n"); | 
|  | DBG(dev, "USE DMA        = %d\n", use_dma); | 
|  | if (use_dma && use_dma_ppb && !use_dma_ppb_du) { | 
|  | DBG(dev, "DMA mode       = PPBNDU (packet per buffer " | 
|  | "WITHOUT desc. update)\n"); | 
|  | dev_info(&dev->pdev->dev, "DMA mode (%s)\n", "PPBNDU"); | 
|  | } else if (use_dma && use_dma_ppb && use_dma_ppb_du) { | 
|  | DBG(dev, "DMA mode       = PPBDU (packet per buffer " | 
|  | "WITH desc. update)\n"); | 
|  | dev_info(&dev->pdev->dev, "DMA mode (%s)\n", "PPBDU"); | 
|  | } | 
|  | if (use_dma && use_dma_bufferfill_mode) { | 
|  | DBG(dev, "DMA mode       = BF (buffer fill mode)\n"); | 
|  | dev_info(&dev->pdev->dev, "DMA mode (%s)\n", "BF"); | 
|  | } | 
|  | if (!use_dma) | 
|  | dev_info(&dev->pdev->dev, "FIFO mode\n"); | 
|  | DBG(dev, "-------------------------------------------------------\n"); | 
|  | } | 
|  |  | 
|  | /* Masks unused interrupts */ | 
|  | static int udc_mask_unused_interrupts(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | /* mask all dev interrupts */ | 
|  | tmp =	AMD_BIT(UDC_DEVINT_SVC) | | 
|  | AMD_BIT(UDC_DEVINT_ENUM) | | 
|  | AMD_BIT(UDC_DEVINT_US) | | 
|  | AMD_BIT(UDC_DEVINT_UR) | | 
|  | AMD_BIT(UDC_DEVINT_ES) | | 
|  | AMD_BIT(UDC_DEVINT_SI) | | 
|  | AMD_BIT(UDC_DEVINT_SOF)| | 
|  | AMD_BIT(UDC_DEVINT_SC); | 
|  | writel(tmp, &dev->regs->irqmsk); | 
|  |  | 
|  | /* mask all ep interrupts */ | 
|  | writel(UDC_EPINT_MSK_DISABLE_ALL, &dev->regs->ep_irqmsk); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Enables endpoint 0 interrupts */ | 
|  | static int udc_enable_ep0_interrupts(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | DBG(dev, "udc_enable_ep0_interrupts()\n"); | 
|  |  | 
|  | /* read irq mask */ | 
|  | tmp = readl(&dev->regs->ep_irqmsk); | 
|  | /* enable ep0 irq's */ | 
|  | tmp &= AMD_UNMASK_BIT(UDC_EPINT_IN_EP0) | 
|  | & AMD_UNMASK_BIT(UDC_EPINT_OUT_EP0); | 
|  | writel(tmp, &dev->regs->ep_irqmsk); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Enables device interrupts for SET_INTF and SET_CONFIG */ | 
|  | static int udc_enable_dev_setup_interrupts(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | DBG(dev, "enable device interrupts for setup data\n"); | 
|  |  | 
|  | /* read irq mask */ | 
|  | tmp = readl(&dev->regs->irqmsk); | 
|  |  | 
|  | /* enable SET_INTERFACE, SET_CONFIG and other needed irq's */ | 
|  | tmp &= AMD_UNMASK_BIT(UDC_DEVINT_SI) | 
|  | & AMD_UNMASK_BIT(UDC_DEVINT_SC) | 
|  | & AMD_UNMASK_BIT(UDC_DEVINT_UR) | 
|  | & AMD_UNMASK_BIT(UDC_DEVINT_SVC) | 
|  | & AMD_UNMASK_BIT(UDC_DEVINT_ENUM); | 
|  | writel(tmp, &dev->regs->irqmsk); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Calculates fifo start of endpoint based on preceding endpoints */ | 
|  | static int udc_set_txfifo_addr(struct udc_ep *ep) | 
|  | { | 
|  | struct udc	*dev; | 
|  | u32 tmp; | 
|  | int i; | 
|  |  | 
|  | if (!ep || !(ep->in)) | 
|  | return -EINVAL; | 
|  |  | 
|  | dev = ep->dev; | 
|  | ep->txfifo = dev->txfifo; | 
|  |  | 
|  | /* traverse ep's */ | 
|  | for (i = 0; i < ep->num; i++) { | 
|  | if (dev->ep[i].regs) { | 
|  | /* read fifo size */ | 
|  | tmp = readl(&dev->ep[i].regs->bufin_framenum); | 
|  | tmp = AMD_GETBITS(tmp, UDC_EPIN_BUFF_SIZE); | 
|  | ep->txfifo += tmp; | 
|  | } | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* CNAK pending field: bit0 = ep0in, bit16 = ep0out */ | 
|  | static u32 cnak_pending; | 
|  |  | 
|  | static void UDC_QUEUE_CNAK(struct udc_ep *ep, unsigned num) | 
|  | { | 
|  | if (readl(&ep->regs->ctl) & AMD_BIT(UDC_EPCTL_NAK)) { | 
|  | DBG(ep->dev, "NAK could not be cleared for ep%d\n", num); | 
|  | cnak_pending |= 1 << (num); | 
|  | ep->naking = 1; | 
|  | } else | 
|  | cnak_pending = cnak_pending & (~(1 << (num))); | 
|  | } | 
|  |  | 
|  |  | 
|  | /* Enables endpoint, is called by gadget driver */ | 
|  | static int | 
|  | udc_ep_enable(struct usb_ep *usbep, const struct usb_endpoint_descriptor *desc) | 
|  | { | 
|  | struct udc_ep		*ep; | 
|  | struct udc		*dev; | 
|  | u32			tmp; | 
|  | unsigned long		iflags; | 
|  | u8 udc_csr_epix; | 
|  | unsigned		maxpacket; | 
|  |  | 
|  | if (!usbep | 
|  | || usbep->name == ep0_string | 
|  | || !desc | 
|  | || desc->bDescriptorType != USB_DT_ENDPOINT) | 
|  | return -EINVAL; | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  | dev = ep->dev; | 
|  |  | 
|  | DBG(dev, "udc_ep_enable() ep %d\n", ep->num); | 
|  |  | 
|  | if (!dev->driver || dev->gadget.speed == USB_SPEED_UNKNOWN) | 
|  | return -ESHUTDOWN; | 
|  |  | 
|  | spin_lock_irqsave(&dev->lock, iflags); | 
|  | ep->ep.desc = desc; | 
|  |  | 
|  | ep->halted = 0; | 
|  |  | 
|  | /* set traffic type */ | 
|  | tmp = readl(&dev->ep[ep->num].regs->ctl); | 
|  | tmp = AMD_ADDBITS(tmp, desc->bmAttributes, UDC_EPCTL_ET); | 
|  | writel(tmp, &dev->ep[ep->num].regs->ctl); | 
|  |  | 
|  | /* set max packet size */ | 
|  | maxpacket = usb_endpoint_maxp(desc); | 
|  | tmp = readl(&dev->ep[ep->num].regs->bufout_maxpkt); | 
|  | tmp = AMD_ADDBITS(tmp, maxpacket, UDC_EP_MAX_PKT_SIZE); | 
|  | ep->ep.maxpacket = maxpacket; | 
|  | writel(tmp, &dev->ep[ep->num].regs->bufout_maxpkt); | 
|  |  | 
|  | /* IN ep */ | 
|  | if (ep->in) { | 
|  |  | 
|  | /* ep ix in UDC CSR register space */ | 
|  | udc_csr_epix = ep->num; | 
|  |  | 
|  | /* set buffer size (tx fifo entries) */ | 
|  | tmp = readl(&dev->ep[ep->num].regs->bufin_framenum); | 
|  | /* double buffering: fifo size = 2 x max packet size */ | 
|  | tmp = AMD_ADDBITS( | 
|  | tmp, | 
|  | maxpacket * UDC_EPIN_BUFF_SIZE_MULT | 
|  | / UDC_DWORD_BYTES, | 
|  | UDC_EPIN_BUFF_SIZE); | 
|  | writel(tmp, &dev->ep[ep->num].regs->bufin_framenum); | 
|  |  | 
|  | /* calc. tx fifo base addr */ | 
|  | udc_set_txfifo_addr(ep); | 
|  |  | 
|  | /* flush fifo */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_F); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  |  | 
|  | /* OUT ep */ | 
|  | } else { | 
|  | /* ep ix in UDC CSR register space */ | 
|  | udc_csr_epix = ep->num - UDC_CSR_EP_OUT_IX_OFS; | 
|  |  | 
|  | /* set max packet size UDC CSR	*/ | 
|  | tmp = readl(&dev->csr->ne[ep->num - UDC_CSR_EP_OUT_IX_OFS]); | 
|  | tmp = AMD_ADDBITS(tmp, maxpacket, | 
|  | UDC_CSR_NE_MAX_PKT); | 
|  | writel(tmp, &dev->csr->ne[ep->num - UDC_CSR_EP_OUT_IX_OFS]); | 
|  |  | 
|  | if (use_dma && !ep->in) { | 
|  | /* alloc and init BNA dummy request */ | 
|  | ep->bna_dummy_req = udc_alloc_bna_dummy(ep); | 
|  | ep->bna_occurred = 0; | 
|  | } | 
|  |  | 
|  | if (ep->num != UDC_EP0OUT_IX) | 
|  | dev->data_ep_enabled = 1; | 
|  | } | 
|  |  | 
|  | /* set ep values */ | 
|  | tmp = readl(&dev->csr->ne[udc_csr_epix]); | 
|  | /* max packet */ | 
|  | tmp = AMD_ADDBITS(tmp, maxpacket, UDC_CSR_NE_MAX_PKT); | 
|  | /* ep number */ | 
|  | tmp = AMD_ADDBITS(tmp, desc->bEndpointAddress, UDC_CSR_NE_NUM); | 
|  | /* ep direction */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->in, UDC_CSR_NE_DIR); | 
|  | /* ep type */ | 
|  | tmp = AMD_ADDBITS(tmp, desc->bmAttributes, UDC_CSR_NE_TYPE); | 
|  | /* ep config */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->dev->cur_config, UDC_CSR_NE_CFG); | 
|  | /* ep interface */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->dev->cur_intf, UDC_CSR_NE_INTF); | 
|  | /* ep alt */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->dev->cur_alt, UDC_CSR_NE_ALT); | 
|  | /* write reg */ | 
|  | writel(tmp, &dev->csr->ne[udc_csr_epix]); | 
|  |  | 
|  | /* enable ep irq */ | 
|  | tmp = readl(&dev->regs->ep_irqmsk); | 
|  | tmp &= AMD_UNMASK_BIT(ep->num); | 
|  | writel(tmp, &dev->regs->ep_irqmsk); | 
|  |  | 
|  | /* | 
|  | * clear NAK by writing CNAK | 
|  | * avoid BNA for OUT DMA, don't clear NAK until DMA desc. written | 
|  | */ | 
|  | if (!use_dma || ep->in) { | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->naking = 0; | 
|  | UDC_QUEUE_CNAK(ep, ep->num); | 
|  | } | 
|  | tmp = desc->bEndpointAddress; | 
|  | DBG(dev, "%s enabled\n", usbep->name); | 
|  |  | 
|  | spin_unlock_irqrestore(&dev->lock, iflags); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Resets endpoint */ | 
|  | static void ep_init(struct udc_regs __iomem *regs, struct udc_ep *ep) | 
|  | { | 
|  | u32		tmp; | 
|  |  | 
|  | VDBG(ep->dev, "ep-%d reset\n", ep->num); | 
|  | ep->ep.desc = NULL; | 
|  | ep->ep.ops = &udc_ep_ops; | 
|  | INIT_LIST_HEAD(&ep->queue); | 
|  |  | 
|  | ep->ep.maxpacket = (u16) ~0; | 
|  | /* set NAK */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_SNAK); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->naking = 1; | 
|  |  | 
|  | /* disable interrupt */ | 
|  | tmp = readl(®s->ep_irqmsk); | 
|  | tmp |= AMD_BIT(ep->num); | 
|  | writel(tmp, ®s->ep_irqmsk); | 
|  |  | 
|  | if (ep->in) { | 
|  | /* unset P and IN bit of potential former DMA */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp &= AMD_UNMASK_BIT(UDC_EPCTL_P); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  |  | 
|  | tmp = readl(&ep->regs->sts); | 
|  | tmp |= AMD_BIT(UDC_EPSTS_IN); | 
|  | writel(tmp, &ep->regs->sts); | 
|  |  | 
|  | /* flush the fifo */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_F); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  |  | 
|  | } | 
|  | /* reset desc pointer */ | 
|  | writel(0, &ep->regs->desptr); | 
|  | } | 
|  |  | 
|  | /* Disables endpoint, is called by gadget driver */ | 
|  | static int udc_ep_disable(struct usb_ep *usbep) | 
|  | { | 
|  | struct udc_ep	*ep = NULL; | 
|  | unsigned long	iflags; | 
|  |  | 
|  | if (!usbep) | 
|  | return -EINVAL; | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  | if (usbep->name == ep0_string || !ep->ep.desc) | 
|  | return -EINVAL; | 
|  |  | 
|  | DBG(ep->dev, "Disable ep-%d\n", ep->num); | 
|  |  | 
|  | spin_lock_irqsave(&ep->dev->lock, iflags); | 
|  | udc_free_request(&ep->ep, &ep->bna_dummy_req->req); | 
|  | empty_req_queue(ep); | 
|  | ep_init(ep->dev->regs, ep); | 
|  | spin_unlock_irqrestore(&ep->dev->lock, iflags); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Allocates request packet, called by gadget driver */ | 
|  | static struct usb_request * | 
|  | udc_alloc_request(struct usb_ep *usbep, gfp_t gfp) | 
|  | { | 
|  | struct udc_request	*req; | 
|  | struct udc_data_dma	*dma_desc; | 
|  | struct udc_ep	*ep; | 
|  |  | 
|  | if (!usbep) | 
|  | return NULL; | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  |  | 
|  | VDBG(ep->dev, "udc_alloc_req(): ep%d\n", ep->num); | 
|  | req = kzalloc(sizeof(struct udc_request), gfp); | 
|  | if (!req) | 
|  | return NULL; | 
|  |  | 
|  | req->req.dma = DMA_DONT_USE; | 
|  | INIT_LIST_HEAD(&req->queue); | 
|  |  | 
|  | if (ep->dma) { | 
|  | /* ep0 in requests are allocated from data pool here */ | 
|  | dma_desc = pci_pool_alloc(ep->dev->data_requests, gfp, | 
|  | &req->td_phys); | 
|  | if (!dma_desc) { | 
|  | kfree(req); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | VDBG(ep->dev, "udc_alloc_req: req = %p dma_desc = %p, " | 
|  | "td_phys = %lx\n", | 
|  | req, dma_desc, | 
|  | (unsigned long)req->td_phys); | 
|  | /* prevent from using desc. - set HOST BUSY */ | 
|  | dma_desc->status = AMD_ADDBITS(dma_desc->status, | 
|  | UDC_DMA_STP_STS_BS_HOST_BUSY, | 
|  | UDC_DMA_STP_STS_BS); | 
|  | dma_desc->bufptr = cpu_to_le32(DMA_DONT_USE); | 
|  | req->td_data = dma_desc; | 
|  | req->td_data_last = NULL; | 
|  | req->chain_len = 1; | 
|  | } | 
|  |  | 
|  | return &req->req; | 
|  | } | 
|  |  | 
|  | /* Frees request packet, called by gadget driver */ | 
|  | static void | 
|  | udc_free_request(struct usb_ep *usbep, struct usb_request *usbreq) | 
|  | { | 
|  | struct udc_ep	*ep; | 
|  | struct udc_request	*req; | 
|  |  | 
|  | if (!usbep || !usbreq) | 
|  | return; | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  | req = container_of(usbreq, struct udc_request, req); | 
|  | VDBG(ep->dev, "free_req req=%p\n", req); | 
|  | BUG_ON(!list_empty(&req->queue)); | 
|  | if (req->td_data) { | 
|  | VDBG(ep->dev, "req->td_data=%p\n", req->td_data); | 
|  |  | 
|  | /* free dma chain if created */ | 
|  | if (req->chain_len > 1) | 
|  | udc_free_dma_chain(ep->dev, req); | 
|  |  | 
|  | pci_pool_free(ep->dev->data_requests, req->td_data, | 
|  | req->td_phys); | 
|  | } | 
|  | kfree(req); | 
|  | } | 
|  |  | 
|  | /* Init BNA dummy descriptor for HOST BUSY and pointing to itself */ | 
|  | static void udc_init_bna_dummy(struct udc_request *req) | 
|  | { | 
|  | if (req) { | 
|  | /* set last bit */ | 
|  | req->td_data->status |= AMD_BIT(UDC_DMA_IN_STS_L); | 
|  | /* set next pointer to itself */ | 
|  | req->td_data->next = req->td_phys; | 
|  | /* set HOST BUSY */ | 
|  | req->td_data->status | 
|  | = AMD_ADDBITS(req->td_data->status, | 
|  | UDC_DMA_STP_STS_BS_DMA_DONE, | 
|  | UDC_DMA_STP_STS_BS); | 
|  | #ifdef UDC_VERBOSE | 
|  | pr_debug("bna desc = %p, sts = %08x\n", | 
|  | req->td_data, req->td_data->status); | 
|  | #endif | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Allocate BNA dummy descriptor */ | 
|  | static struct udc_request *udc_alloc_bna_dummy(struct udc_ep *ep) | 
|  | { | 
|  | struct udc_request *req = NULL; | 
|  | struct usb_request *_req = NULL; | 
|  |  | 
|  | /* alloc the dummy request */ | 
|  | _req = udc_alloc_request(&ep->ep, GFP_ATOMIC); | 
|  | if (_req) { | 
|  | req = container_of(_req, struct udc_request, req); | 
|  | ep->bna_dummy_req = req; | 
|  | udc_init_bna_dummy(req); | 
|  | } | 
|  | return req; | 
|  | } | 
|  |  | 
|  | /* Write data to TX fifo for IN packets */ | 
|  | static void | 
|  | udc_txfifo_write(struct udc_ep *ep, struct usb_request *req) | 
|  | { | 
|  | u8			*req_buf; | 
|  | u32			*buf; | 
|  | int			i, j; | 
|  | unsigned		bytes = 0; | 
|  | unsigned		remaining = 0; | 
|  |  | 
|  | if (!req || !ep) | 
|  | return; | 
|  |  | 
|  | req_buf = req->buf + req->actual; | 
|  | prefetch(req_buf); | 
|  | remaining = req->length - req->actual; | 
|  |  | 
|  | buf = (u32 *) req_buf; | 
|  |  | 
|  | bytes = ep->ep.maxpacket; | 
|  | if (bytes > remaining) | 
|  | bytes = remaining; | 
|  |  | 
|  | /* dwords first */ | 
|  | for (i = 0; i < bytes / UDC_DWORD_BYTES; i++) | 
|  | writel(*(buf + i), ep->txfifo); | 
|  |  | 
|  | /* remaining bytes must be written by byte access */ | 
|  | for (j = 0; j < bytes % UDC_DWORD_BYTES; j++) { | 
|  | writeb((u8)(*(buf + i) >> (j << UDC_BITS_PER_BYTE_SHIFT)), | 
|  | ep->txfifo); | 
|  | } | 
|  |  | 
|  | /* dummy write confirm */ | 
|  | writel(0, &ep->regs->confirm); | 
|  | } | 
|  |  | 
|  | /* Read dwords from RX fifo for OUT transfers */ | 
|  | static int udc_rxfifo_read_dwords(struct udc *dev, u32 *buf, int dwords) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | VDBG(dev, "udc_read_dwords(): %d dwords\n", dwords); | 
|  |  | 
|  | for (i = 0; i < dwords; i++) | 
|  | *(buf + i) = readl(dev->rxfifo); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Read bytes from RX fifo for OUT transfers */ | 
|  | static int udc_rxfifo_read_bytes(struct udc *dev, u8 *buf, int bytes) | 
|  | { | 
|  | int i, j; | 
|  | u32 tmp; | 
|  |  | 
|  | VDBG(dev, "udc_read_bytes(): %d bytes\n", bytes); | 
|  |  | 
|  | /* dwords first */ | 
|  | for (i = 0; i < bytes / UDC_DWORD_BYTES; i++) | 
|  | *((u32 *)(buf + (i<<2))) = readl(dev->rxfifo); | 
|  |  | 
|  | /* remaining bytes must be read by byte access */ | 
|  | if (bytes % UDC_DWORD_BYTES) { | 
|  | tmp = readl(dev->rxfifo); | 
|  | for (j = 0; j < bytes % UDC_DWORD_BYTES; j++) { | 
|  | *(buf + (i<<2) + j) = (u8)(tmp & UDC_BYTE_MASK); | 
|  | tmp = tmp >> UDC_BITS_PER_BYTE; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Read data from RX fifo for OUT transfers */ | 
|  | static int | 
|  | udc_rxfifo_read(struct udc_ep *ep, struct udc_request *req) | 
|  | { | 
|  | u8 *buf; | 
|  | unsigned buf_space; | 
|  | unsigned bytes = 0; | 
|  | unsigned finished = 0; | 
|  |  | 
|  | /* received number bytes */ | 
|  | bytes = readl(&ep->regs->sts); | 
|  | bytes = AMD_GETBITS(bytes, UDC_EPSTS_RX_PKT_SIZE); | 
|  |  | 
|  | buf_space = req->req.length - req->req.actual; | 
|  | buf = req->req.buf + req->req.actual; | 
|  | if (bytes > buf_space) { | 
|  | if ((buf_space % ep->ep.maxpacket) != 0) { | 
|  | DBG(ep->dev, | 
|  | "%s: rx %d bytes, rx-buf space = %d bytesn\n", | 
|  | ep->ep.name, bytes, buf_space); | 
|  | req->req.status = -EOVERFLOW; | 
|  | } | 
|  | bytes = buf_space; | 
|  | } | 
|  | req->req.actual += bytes; | 
|  |  | 
|  | /* last packet ? */ | 
|  | if (((bytes % ep->ep.maxpacket) != 0) || (!bytes) | 
|  | || ((req->req.actual == req->req.length) && !req->req.zero)) | 
|  | finished = 1; | 
|  |  | 
|  | /* read rx fifo bytes */ | 
|  | VDBG(ep->dev, "ep %s: rxfifo read %d bytes\n", ep->ep.name, bytes); | 
|  | udc_rxfifo_read_bytes(ep->dev, buf, bytes); | 
|  |  | 
|  | return finished; | 
|  | } | 
|  |  | 
|  | /* create/re-init a DMA descriptor or a DMA descriptor chain */ | 
|  | static int prep_dma(struct udc_ep *ep, struct udc_request *req, gfp_t gfp) | 
|  | { | 
|  | int	retval = 0; | 
|  | u32	tmp; | 
|  |  | 
|  | VDBG(ep->dev, "prep_dma\n"); | 
|  | VDBG(ep->dev, "prep_dma ep%d req->td_data=%p\n", | 
|  | ep->num, req->td_data); | 
|  |  | 
|  | /* set buffer pointer */ | 
|  | req->td_data->bufptr = req->req.dma; | 
|  |  | 
|  | /* set last bit */ | 
|  | req->td_data->status |= AMD_BIT(UDC_DMA_IN_STS_L); | 
|  |  | 
|  | /* build/re-init dma chain if maxpkt scatter mode, not for EP0 */ | 
|  | if (use_dma_ppb) { | 
|  |  | 
|  | retval = udc_create_dma_chain(ep, req, ep->ep.maxpacket, gfp); | 
|  | if (retval != 0) { | 
|  | if (retval == -ENOMEM) | 
|  | DBG(ep->dev, "Out of DMA memory\n"); | 
|  | return retval; | 
|  | } | 
|  | if (ep->in) { | 
|  | if (req->req.length == ep->ep.maxpacket) { | 
|  | /* write tx bytes */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | ep->ep.maxpacket, | 
|  | UDC_DMA_IN_STS_TXBYTES); | 
|  |  | 
|  | } | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | if (ep->in) { | 
|  | VDBG(ep->dev, "IN: use_dma_ppb=%d req->req.len=%d " | 
|  | "maxpacket=%d ep%d\n", | 
|  | use_dma_ppb, req->req.length, | 
|  | ep->ep.maxpacket, ep->num); | 
|  | /* | 
|  | * if bytes < max packet then tx bytes must | 
|  | * be written in packet per buffer mode | 
|  | */ | 
|  | if (!use_dma_ppb || req->req.length < ep->ep.maxpacket | 
|  | || ep->num == UDC_EP0OUT_IX | 
|  | || ep->num == UDC_EP0IN_IX) { | 
|  | /* write tx bytes */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | req->req.length, | 
|  | UDC_DMA_IN_STS_TXBYTES); | 
|  | /* reset frame num */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | 0, | 
|  | UDC_DMA_IN_STS_FRAMENUM); | 
|  | } | 
|  | /* set HOST BUSY */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | UDC_DMA_STP_STS_BS_HOST_BUSY, | 
|  | UDC_DMA_STP_STS_BS); | 
|  | } else { | 
|  | VDBG(ep->dev, "OUT set host ready\n"); | 
|  | /* set HOST READY */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | UDC_DMA_STP_STS_BS_HOST_READY, | 
|  | UDC_DMA_STP_STS_BS); | 
|  |  | 
|  |  | 
|  | /* clear NAK by writing CNAK */ | 
|  | if (ep->naking) { | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->naking = 0; | 
|  | UDC_QUEUE_CNAK(ep, ep->num); | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* Completes request packet ... caller MUST hold lock */ | 
|  | static void | 
|  | complete_req(struct udc_ep *ep, struct udc_request *req, int sts) | 
|  | __releases(ep->dev->lock) | 
|  | __acquires(ep->dev->lock) | 
|  | { | 
|  | struct udc		*dev; | 
|  | unsigned		halted; | 
|  |  | 
|  | VDBG(ep->dev, "complete_req(): ep%d\n", ep->num); | 
|  |  | 
|  | dev = ep->dev; | 
|  | /* unmap DMA */ | 
|  | if (ep->dma) | 
|  | usb_gadget_unmap_request(&dev->gadget, &req->req, ep->in); | 
|  |  | 
|  | halted = ep->halted; | 
|  | ep->halted = 1; | 
|  |  | 
|  | /* set new status if pending */ | 
|  | if (req->req.status == -EINPROGRESS) | 
|  | req->req.status = sts; | 
|  |  | 
|  | /* remove from ep queue */ | 
|  | list_del_init(&req->queue); | 
|  |  | 
|  | VDBG(ep->dev, "req %p => complete %d bytes at %s with sts %d\n", | 
|  | &req->req, req->req.length, ep->ep.name, sts); | 
|  |  | 
|  | spin_unlock(&dev->lock); | 
|  | req->req.complete(&ep->ep, &req->req); | 
|  | spin_lock(&dev->lock); | 
|  | ep->halted = halted; | 
|  | } | 
|  |  | 
|  | /* frees pci pool descriptors of a DMA chain */ | 
|  | static int udc_free_dma_chain(struct udc *dev, struct udc_request *req) | 
|  | { | 
|  |  | 
|  | int ret_val = 0; | 
|  | struct udc_data_dma	*td; | 
|  | struct udc_data_dma	*td_last = NULL; | 
|  | unsigned int i; | 
|  |  | 
|  | DBG(dev, "free chain req = %p\n", req); | 
|  |  | 
|  | /* do not free first desc., will be done by free for request */ | 
|  | td_last = req->td_data; | 
|  | td = phys_to_virt(td_last->next); | 
|  |  | 
|  | for (i = 1; i < req->chain_len; i++) { | 
|  |  | 
|  | pci_pool_free(dev->data_requests, td, | 
|  | (dma_addr_t) td_last->next); | 
|  | td_last = td; | 
|  | td = phys_to_virt(td_last->next); | 
|  | } | 
|  |  | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /* Iterates to the end of a DMA chain and returns last descriptor */ | 
|  | static struct udc_data_dma *udc_get_last_dma_desc(struct udc_request *req) | 
|  | { | 
|  | struct udc_data_dma	*td; | 
|  |  | 
|  | td = req->td_data; | 
|  | while (td && !(td->status & AMD_BIT(UDC_DMA_IN_STS_L))) | 
|  | td = phys_to_virt(td->next); | 
|  |  | 
|  | return td; | 
|  |  | 
|  | } | 
|  |  | 
|  | /* Iterates to the end of a DMA chain and counts bytes received */ | 
|  | static u32 udc_get_ppbdu_rxbytes(struct udc_request *req) | 
|  | { | 
|  | struct udc_data_dma	*td; | 
|  | u32 count; | 
|  |  | 
|  | td = req->td_data; | 
|  | /* received number bytes */ | 
|  | count = AMD_GETBITS(td->status, UDC_DMA_OUT_STS_RXBYTES); | 
|  |  | 
|  | while (td && !(td->status & AMD_BIT(UDC_DMA_IN_STS_L))) { | 
|  | td = phys_to_virt(td->next); | 
|  | /* received number bytes */ | 
|  | if (td) { | 
|  | count += AMD_GETBITS(td->status, | 
|  | UDC_DMA_OUT_STS_RXBYTES); | 
|  | } | 
|  | } | 
|  |  | 
|  | return count; | 
|  |  | 
|  | } | 
|  |  | 
|  | /* Creates or re-inits a DMA chain */ | 
|  | static int udc_create_dma_chain( | 
|  | struct udc_ep *ep, | 
|  | struct udc_request *req, | 
|  | unsigned long buf_len, gfp_t gfp_flags | 
|  | ) | 
|  | { | 
|  | unsigned long bytes = req->req.length; | 
|  | unsigned int i; | 
|  | dma_addr_t dma_addr; | 
|  | struct udc_data_dma	*td = NULL; | 
|  | struct udc_data_dma	*last = NULL; | 
|  | unsigned long txbytes; | 
|  | unsigned create_new_chain = 0; | 
|  | unsigned len; | 
|  |  | 
|  | VDBG(ep->dev, "udc_create_dma_chain: bytes=%ld buf_len=%ld\n", | 
|  | bytes, buf_len); | 
|  | dma_addr = DMA_DONT_USE; | 
|  |  | 
|  | /* unset L bit in first desc for OUT */ | 
|  | if (!ep->in) | 
|  | req->td_data->status &= AMD_CLEAR_BIT(UDC_DMA_IN_STS_L); | 
|  |  | 
|  | /* alloc only new desc's if not already available */ | 
|  | len = req->req.length / ep->ep.maxpacket; | 
|  | if (req->req.length % ep->ep.maxpacket) | 
|  | len++; | 
|  |  | 
|  | if (len > req->chain_len) { | 
|  | /* shorter chain already allocated before */ | 
|  | if (req->chain_len > 1) | 
|  | udc_free_dma_chain(ep->dev, req); | 
|  | req->chain_len = len; | 
|  | create_new_chain = 1; | 
|  | } | 
|  |  | 
|  | td = req->td_data; | 
|  | /* gen. required number of descriptors and buffers */ | 
|  | for (i = buf_len; i < bytes; i += buf_len) { | 
|  | /* create or determine next desc. */ | 
|  | if (create_new_chain) { | 
|  |  | 
|  | td = pci_pool_alloc(ep->dev->data_requests, | 
|  | gfp_flags, &dma_addr); | 
|  | if (!td) | 
|  | return -ENOMEM; | 
|  |  | 
|  | td->status = 0; | 
|  | } else if (i == buf_len) { | 
|  | /* first td */ | 
|  | td = (struct udc_data_dma *) phys_to_virt( | 
|  | req->td_data->next); | 
|  | td->status = 0; | 
|  | } else { | 
|  | td = (struct udc_data_dma *) phys_to_virt(last->next); | 
|  | td->status = 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | if (td) | 
|  | td->bufptr = req->req.dma + i; /* assign buffer */ | 
|  | else | 
|  | break; | 
|  |  | 
|  | /* short packet ? */ | 
|  | if ((bytes - i) >= buf_len) { | 
|  | txbytes = buf_len; | 
|  | } else { | 
|  | /* short packet */ | 
|  | txbytes = bytes - i; | 
|  | } | 
|  |  | 
|  | /* link td and assign tx bytes */ | 
|  | if (i == buf_len) { | 
|  | if (create_new_chain) | 
|  | req->td_data->next = dma_addr; | 
|  | /* | 
|  | else | 
|  | req->td_data->next = virt_to_phys(td); | 
|  | */ | 
|  | /* write tx bytes */ | 
|  | if (ep->in) { | 
|  | /* first desc */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | ep->ep.maxpacket, | 
|  | UDC_DMA_IN_STS_TXBYTES); | 
|  | /* second desc */ | 
|  | td->status = AMD_ADDBITS(td->status, | 
|  | txbytes, | 
|  | UDC_DMA_IN_STS_TXBYTES); | 
|  | } | 
|  | } else { | 
|  | if (create_new_chain) | 
|  | last->next = dma_addr; | 
|  | /* | 
|  | else | 
|  | last->next = virt_to_phys(td); | 
|  | */ | 
|  | if (ep->in) { | 
|  | /* write tx bytes */ | 
|  | td->status = AMD_ADDBITS(td->status, | 
|  | txbytes, | 
|  | UDC_DMA_IN_STS_TXBYTES); | 
|  | } | 
|  | } | 
|  | last = td; | 
|  | } | 
|  | /* set last bit */ | 
|  | if (td) { | 
|  | td->status |= AMD_BIT(UDC_DMA_IN_STS_L); | 
|  | /* last desc. points to itself */ | 
|  | req->td_data_last = td; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Enabling RX DMA */ | 
|  | static void udc_set_rde(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | VDBG(dev, "udc_set_rde()\n"); | 
|  | /* stop RDE timer */ | 
|  | if (timer_pending(&udc_timer)) { | 
|  | set_rde = 0; | 
|  | mod_timer(&udc_timer, jiffies - 1); | 
|  | } | 
|  | /* set RDE */ | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_RDE); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  | } | 
|  |  | 
|  | /* Queues a request packet, called by gadget driver */ | 
|  | static int | 
|  | udc_queue(struct usb_ep *usbep, struct usb_request *usbreq, gfp_t gfp) | 
|  | { | 
|  | int			retval = 0; | 
|  | u8			open_rxfifo = 0; | 
|  | unsigned long		iflags; | 
|  | struct udc_ep		*ep; | 
|  | struct udc_request	*req; | 
|  | struct udc		*dev; | 
|  | u32			tmp; | 
|  |  | 
|  | /* check the inputs */ | 
|  | req = container_of(usbreq, struct udc_request, req); | 
|  |  | 
|  | if (!usbep || !usbreq || !usbreq->complete || !usbreq->buf | 
|  | || !list_empty(&req->queue)) | 
|  | return -EINVAL; | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  | if (!ep->ep.desc && (ep->num != 0 && ep->num != UDC_EP0OUT_IX)) | 
|  | return -EINVAL; | 
|  |  | 
|  | VDBG(ep->dev, "udc_queue(): ep%d-in=%d\n", ep->num, ep->in); | 
|  | dev = ep->dev; | 
|  |  | 
|  | if (!dev->driver || dev->gadget.speed == USB_SPEED_UNKNOWN) | 
|  | return -ESHUTDOWN; | 
|  |  | 
|  | /* map dma (usually done before) */ | 
|  | if (ep->dma) { | 
|  | VDBG(dev, "DMA map req %p\n", req); | 
|  | retval = usb_gadget_map_request(&udc->gadget, usbreq, ep->in); | 
|  | if (retval) | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | VDBG(dev, "%s queue req %p, len %d req->td_data=%p buf %p\n", | 
|  | usbep->name, usbreq, usbreq->length, | 
|  | req->td_data, usbreq->buf); | 
|  |  | 
|  | spin_lock_irqsave(&dev->lock, iflags); | 
|  | usbreq->actual = 0; | 
|  | usbreq->status = -EINPROGRESS; | 
|  | req->dma_done = 0; | 
|  |  | 
|  | /* on empty queue just do first transfer */ | 
|  | if (list_empty(&ep->queue)) { | 
|  | /* zlp */ | 
|  | if (usbreq->length == 0) { | 
|  | /* IN zlp's are handled by hardware */ | 
|  | complete_req(ep, req, 0); | 
|  | VDBG(dev, "%s: zlp\n", ep->ep.name); | 
|  | /* | 
|  | * if set_config or set_intf is waiting for ack by zlp | 
|  | * then set CSR_DONE | 
|  | */ | 
|  | if (dev->set_cfg_not_acked) { | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_CSR_DONE); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  | dev->set_cfg_not_acked = 0; | 
|  | } | 
|  | /* setup command is ACK'ed now by zlp */ | 
|  | if (dev->waiting_zlp_ack_ep0in) { | 
|  | /* clear NAK by writing CNAK in EP0_IN */ | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0IN_IX].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[UDC_EP0IN_IX], | 
|  | UDC_EP0IN_IX); | 
|  | dev->waiting_zlp_ack_ep0in = 0; | 
|  | } | 
|  | goto finished; | 
|  | } | 
|  | if (ep->dma) { | 
|  | retval = prep_dma(ep, req, gfp); | 
|  | if (retval != 0) | 
|  | goto finished; | 
|  | /* write desc pointer to enable DMA */ | 
|  | if (ep->in) { | 
|  | /* set HOST READY */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS(req->td_data->status, | 
|  | UDC_DMA_IN_STS_BS_HOST_READY, | 
|  | UDC_DMA_IN_STS_BS); | 
|  | } | 
|  |  | 
|  | /* disabled rx dma while descriptor update */ | 
|  | if (!ep->in) { | 
|  | /* stop RDE timer */ | 
|  | if (timer_pending(&udc_timer)) { | 
|  | set_rde = 0; | 
|  | mod_timer(&udc_timer, jiffies - 1); | 
|  | } | 
|  | /* clear RDE */ | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp &= AMD_UNMASK_BIT(UDC_DEVCTL_RDE); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  | open_rxfifo = 1; | 
|  |  | 
|  | /* | 
|  | * if BNA occurred then let BNA dummy desc. | 
|  | * point to current desc. | 
|  | */ | 
|  | if (ep->bna_occurred) { | 
|  | VDBG(dev, "copy to BNA dummy desc.\n"); | 
|  | memcpy(ep->bna_dummy_req->td_data, | 
|  | req->td_data, | 
|  | sizeof(struct udc_data_dma)); | 
|  | } | 
|  | } | 
|  | /* write desc pointer */ | 
|  | writel(req->td_phys, &ep->regs->desptr); | 
|  |  | 
|  | /* clear NAK by writing CNAK */ | 
|  | if (ep->naking) { | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->naking = 0; | 
|  | UDC_QUEUE_CNAK(ep, ep->num); | 
|  | } | 
|  |  | 
|  | if (ep->in) { | 
|  | /* enable ep irq */ | 
|  | tmp = readl(&dev->regs->ep_irqmsk); | 
|  | tmp &= AMD_UNMASK_BIT(ep->num); | 
|  | writel(tmp, &dev->regs->ep_irqmsk); | 
|  | } | 
|  | } else if (ep->in) { | 
|  | /* enable ep irq */ | 
|  | tmp = readl(&dev->regs->ep_irqmsk); | 
|  | tmp &= AMD_UNMASK_BIT(ep->num); | 
|  | writel(tmp, &dev->regs->ep_irqmsk); | 
|  | } | 
|  |  | 
|  | } else if (ep->dma) { | 
|  |  | 
|  | /* | 
|  | * prep_dma not used for OUT ep's, this is not possible | 
|  | * for PPB modes, because of chain creation reasons | 
|  | */ | 
|  | if (ep->in) { | 
|  | retval = prep_dma(ep, req, gfp); | 
|  | if (retval != 0) | 
|  | goto finished; | 
|  | } | 
|  | } | 
|  | VDBG(dev, "list_add\n"); | 
|  | /* add request to ep queue */ | 
|  | if (req) { | 
|  |  | 
|  | list_add_tail(&req->queue, &ep->queue); | 
|  |  | 
|  | /* open rxfifo if out data queued */ | 
|  | if (open_rxfifo) { | 
|  | /* enable DMA */ | 
|  | req->dma_going = 1; | 
|  | udc_set_rde(dev); | 
|  | if (ep->num != UDC_EP0OUT_IX) | 
|  | dev->data_ep_queued = 1; | 
|  | } | 
|  | /* stop OUT naking */ | 
|  | if (!ep->in) { | 
|  | if (!use_dma && udc_rxfifo_pending) { | 
|  | DBG(dev, "udc_queue(): pending bytes in " | 
|  | "rxfifo after nyet\n"); | 
|  | /* | 
|  | * read pending bytes afer nyet: | 
|  | * referring to isr | 
|  | */ | 
|  | if (udc_rxfifo_read(ep, req)) { | 
|  | /* finish */ | 
|  | complete_req(ep, req, 0); | 
|  | } | 
|  | udc_rxfifo_pending = 0; | 
|  |  | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | finished: | 
|  | spin_unlock_irqrestore(&dev->lock, iflags); | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* Empty request queue of an endpoint; caller holds spinlock */ | 
|  | static void empty_req_queue(struct udc_ep *ep) | 
|  | { | 
|  | struct udc_request	*req; | 
|  |  | 
|  | ep->halted = 1; | 
|  | while (!list_empty(&ep->queue)) { | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, | 
|  | queue); | 
|  | complete_req(ep, req, -ESHUTDOWN); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Dequeues a request packet, called by gadget driver */ | 
|  | static int udc_dequeue(struct usb_ep *usbep, struct usb_request *usbreq) | 
|  | { | 
|  | struct udc_ep		*ep; | 
|  | struct udc_request	*req; | 
|  | unsigned		halted; | 
|  | unsigned long		iflags; | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  | if (!usbep || !usbreq || (!ep->ep.desc && (ep->num != 0 | 
|  | && ep->num != UDC_EP0OUT_IX))) | 
|  | return -EINVAL; | 
|  |  | 
|  | req = container_of(usbreq, struct udc_request, req); | 
|  |  | 
|  | spin_lock_irqsave(&ep->dev->lock, iflags); | 
|  | halted = ep->halted; | 
|  | ep->halted = 1; | 
|  | /* request in processing or next one */ | 
|  | if (ep->queue.next == &req->queue) { | 
|  | if (ep->dma && req->dma_going) { | 
|  | if (ep->in) | 
|  | ep->cancel_transfer = 1; | 
|  | else { | 
|  | u32 tmp; | 
|  | u32 dma_sts; | 
|  | /* stop potential receive DMA */ | 
|  | tmp = readl(&udc->regs->ctl); | 
|  | writel(tmp & AMD_UNMASK_BIT(UDC_DEVCTL_RDE), | 
|  | &udc->regs->ctl); | 
|  | /* | 
|  | * Cancel transfer later in ISR | 
|  | * if descriptor was touched. | 
|  | */ | 
|  | dma_sts = AMD_GETBITS(req->td_data->status, | 
|  | UDC_DMA_OUT_STS_BS); | 
|  | if (dma_sts != UDC_DMA_OUT_STS_BS_HOST_READY) | 
|  | ep->cancel_transfer = 1; | 
|  | else { | 
|  | udc_init_bna_dummy(ep->req); | 
|  | writel(ep->bna_dummy_req->td_phys, | 
|  | &ep->regs->desptr); | 
|  | } | 
|  | writel(tmp, &udc->regs->ctl); | 
|  | } | 
|  | } | 
|  | } | 
|  | complete_req(ep, req, -ECONNRESET); | 
|  | ep->halted = halted; | 
|  |  | 
|  | spin_unlock_irqrestore(&ep->dev->lock, iflags); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Halt or clear halt of endpoint */ | 
|  | static int | 
|  | udc_set_halt(struct usb_ep *usbep, int halt) | 
|  | { | 
|  | struct udc_ep	*ep; | 
|  | u32 tmp; | 
|  | unsigned long iflags; | 
|  | int retval = 0; | 
|  |  | 
|  | if (!usbep) | 
|  | return -EINVAL; | 
|  |  | 
|  | pr_debug("set_halt %s: halt=%d\n", usbep->name, halt); | 
|  |  | 
|  | ep = container_of(usbep, struct udc_ep, ep); | 
|  | if (!ep->ep.desc && (ep->num != 0 && ep->num != UDC_EP0OUT_IX)) | 
|  | return -EINVAL; | 
|  | if (!ep->dev->driver || ep->dev->gadget.speed == USB_SPEED_UNKNOWN) | 
|  | return -ESHUTDOWN; | 
|  |  | 
|  | spin_lock_irqsave(&udc_stall_spinlock, iflags); | 
|  | /* halt or clear halt */ | 
|  | if (halt) { | 
|  | if (ep->num == 0) | 
|  | ep->dev->stall_ep0in = 1; | 
|  | else { | 
|  | /* | 
|  | * set STALL | 
|  | * rxfifo empty not taken into acount | 
|  | */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_S); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->halted = 1; | 
|  |  | 
|  | /* setup poll timer */ | 
|  | if (!timer_pending(&udc_pollstall_timer)) { | 
|  | udc_pollstall_timer.expires = jiffies + | 
|  | HZ * UDC_POLLSTALL_TIMER_USECONDS | 
|  | / (1000 * 1000); | 
|  | if (!stop_pollstall_timer) { | 
|  | DBG(ep->dev, "start polltimer\n"); | 
|  | add_timer(&udc_pollstall_timer); | 
|  | } | 
|  | } | 
|  | } | 
|  | } else { | 
|  | /* ep is halted by set_halt() before */ | 
|  | if (ep->halted) { | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | /* clear stall bit */ | 
|  | tmp = tmp & AMD_CLEAR_BIT(UDC_EPCTL_S); | 
|  | /* clear NAK by writing CNAK */ | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->halted = 0; | 
|  | UDC_QUEUE_CNAK(ep, ep->num); | 
|  | } | 
|  | } | 
|  | spin_unlock_irqrestore(&udc_stall_spinlock, iflags); | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* gadget interface */ | 
|  | static const struct usb_ep_ops udc_ep_ops = { | 
|  | .enable		= udc_ep_enable, | 
|  | .disable	= udc_ep_disable, | 
|  |  | 
|  | .alloc_request	= udc_alloc_request, | 
|  | .free_request	= udc_free_request, | 
|  |  | 
|  | .queue		= udc_queue, | 
|  | .dequeue	= udc_dequeue, | 
|  |  | 
|  | .set_halt	= udc_set_halt, | 
|  | /* fifo ops not implemented */ | 
|  | }; | 
|  |  | 
|  | /*-------------------------------------------------------------------------*/ | 
|  |  | 
|  | /* Get frame counter (not implemented) */ | 
|  | static int udc_get_frame(struct usb_gadget *gadget) | 
|  | { | 
|  | return -EOPNOTSUPP; | 
|  | } | 
|  |  | 
|  | /* Remote wakeup gadget interface */ | 
|  | static int udc_wakeup(struct usb_gadget *gadget) | 
|  | { | 
|  | struct udc		*dev; | 
|  |  | 
|  | if (!gadget) | 
|  | return -EINVAL; | 
|  | dev = container_of(gadget, struct udc, gadget); | 
|  | udc_remote_wakeup(dev); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int amd5536_start(struct usb_gadget_driver *driver, | 
|  | int (*bind)(struct usb_gadget *)); | 
|  | static int amd5536_stop(struct usb_gadget_driver *driver); | 
|  | /* gadget operations */ | 
|  | static const struct usb_gadget_ops udc_ops = { | 
|  | .wakeup		= udc_wakeup, | 
|  | .get_frame	= udc_get_frame, | 
|  | .start		= amd5536_start, | 
|  | .stop		= amd5536_stop, | 
|  | }; | 
|  |  | 
|  | /* Setups endpoint parameters, adds endpoints to linked list */ | 
|  | static void make_ep_lists(struct udc *dev) | 
|  | { | 
|  | /* make gadget ep lists */ | 
|  | INIT_LIST_HEAD(&dev->gadget.ep_list); | 
|  | list_add_tail(&dev->ep[UDC_EPIN_STATUS_IX].ep.ep_list, | 
|  | &dev->gadget.ep_list); | 
|  | list_add_tail(&dev->ep[UDC_EPIN_IX].ep.ep_list, | 
|  | &dev->gadget.ep_list); | 
|  | list_add_tail(&dev->ep[UDC_EPOUT_IX].ep.ep_list, | 
|  | &dev->gadget.ep_list); | 
|  |  | 
|  | /* fifo config */ | 
|  | dev->ep[UDC_EPIN_STATUS_IX].fifo_depth = UDC_EPIN_SMALLINT_BUFF_SIZE; | 
|  | if (dev->gadget.speed == USB_SPEED_FULL) | 
|  | dev->ep[UDC_EPIN_IX].fifo_depth = UDC_FS_EPIN_BUFF_SIZE; | 
|  | else if (dev->gadget.speed == USB_SPEED_HIGH) | 
|  | dev->ep[UDC_EPIN_IX].fifo_depth = hs_tx_buf; | 
|  | dev->ep[UDC_EPOUT_IX].fifo_depth = UDC_RXFIFO_SIZE; | 
|  | } | 
|  |  | 
|  | /* init registers at driver load time */ | 
|  | static int startup_registers(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | /* init controller by soft reset */ | 
|  | udc_soft_reset(dev); | 
|  |  | 
|  | /* mask not needed interrupts */ | 
|  | udc_mask_unused_interrupts(dev); | 
|  |  | 
|  | /* put into initial config */ | 
|  | udc_basic_init(dev); | 
|  | /* link up all endpoints */ | 
|  | udc_setup_endpoints(dev); | 
|  |  | 
|  | /* program speed */ | 
|  | tmp = readl(&dev->regs->cfg); | 
|  | if (use_fullspeed) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_DEVCFG_SPD_FS, UDC_DEVCFG_SPD); | 
|  | else | 
|  | tmp = AMD_ADDBITS(tmp, UDC_DEVCFG_SPD_HS, UDC_DEVCFG_SPD); | 
|  | writel(tmp, &dev->regs->cfg); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Inits UDC context */ | 
|  | static void udc_basic_init(struct udc *dev) | 
|  | { | 
|  | u32	tmp; | 
|  |  | 
|  | DBG(dev, "udc_basic_init()\n"); | 
|  |  | 
|  | dev->gadget.speed = USB_SPEED_UNKNOWN; | 
|  |  | 
|  | /* stop RDE timer */ | 
|  | if (timer_pending(&udc_timer)) { | 
|  | set_rde = 0; | 
|  | mod_timer(&udc_timer, jiffies - 1); | 
|  | } | 
|  | /* stop poll stall timer */ | 
|  | if (timer_pending(&udc_pollstall_timer)) | 
|  | mod_timer(&udc_pollstall_timer, jiffies - 1); | 
|  | /* disable DMA */ | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp &= AMD_UNMASK_BIT(UDC_DEVCTL_RDE); | 
|  | tmp &= AMD_UNMASK_BIT(UDC_DEVCTL_TDE); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  |  | 
|  | /* enable dynamic CSR programming */ | 
|  | tmp = readl(&dev->regs->cfg); | 
|  | tmp |= AMD_BIT(UDC_DEVCFG_CSR_PRG); | 
|  | /* set self powered */ | 
|  | tmp |= AMD_BIT(UDC_DEVCFG_SP); | 
|  | /* set remote wakeupable */ | 
|  | tmp |= AMD_BIT(UDC_DEVCFG_RWKP); | 
|  | writel(tmp, &dev->regs->cfg); | 
|  |  | 
|  | make_ep_lists(dev); | 
|  |  | 
|  | dev->data_ep_enabled = 0; | 
|  | dev->data_ep_queued = 0; | 
|  | } | 
|  |  | 
|  | /* Sets initial endpoint parameters */ | 
|  | static void udc_setup_endpoints(struct udc *dev) | 
|  | { | 
|  | struct udc_ep	*ep; | 
|  | u32	tmp; | 
|  | u32	reg; | 
|  |  | 
|  | DBG(dev, "udc_setup_endpoints()\n"); | 
|  |  | 
|  | /* read enum speed */ | 
|  | tmp = readl(&dev->regs->sts); | 
|  | tmp = AMD_GETBITS(tmp, UDC_DEVSTS_ENUM_SPEED); | 
|  | if (tmp == UDC_DEVSTS_ENUM_SPEED_HIGH) | 
|  | dev->gadget.speed = USB_SPEED_HIGH; | 
|  | else if (tmp == UDC_DEVSTS_ENUM_SPEED_FULL) | 
|  | dev->gadget.speed = USB_SPEED_FULL; | 
|  |  | 
|  | /* set basic ep parameters */ | 
|  | for (tmp = 0; tmp < UDC_EP_NUM; tmp++) { | 
|  | ep = &dev->ep[tmp]; | 
|  | ep->dev = dev; | 
|  | ep->ep.name = ep_string[tmp]; | 
|  | ep->num = tmp; | 
|  | /* txfifo size is calculated at enable time */ | 
|  | ep->txfifo = dev->txfifo; | 
|  |  | 
|  | /* fifo size */ | 
|  | if (tmp < UDC_EPIN_NUM) { | 
|  | ep->fifo_depth = UDC_TXFIFO_SIZE; | 
|  | ep->in = 1; | 
|  | } else { | 
|  | ep->fifo_depth = UDC_RXFIFO_SIZE; | 
|  | ep->in = 0; | 
|  |  | 
|  | } | 
|  | ep->regs = &dev->ep_regs[tmp]; | 
|  | /* | 
|  | * ep will be reset only if ep was not enabled before to avoid | 
|  | * disabling ep interrupts when ENUM interrupt occurs but ep is | 
|  | * not enabled by gadget driver | 
|  | */ | 
|  | if (!ep->ep.desc) | 
|  | ep_init(dev->regs, ep); | 
|  |  | 
|  | if (use_dma) { | 
|  | /* | 
|  | * ep->dma is not really used, just to indicate that | 
|  | * DMA is active: remove this | 
|  | * dma regs = dev control regs | 
|  | */ | 
|  | ep->dma = &dev->regs->ctl; | 
|  |  | 
|  | /* nak OUT endpoints until enable - not for ep0 */ | 
|  | if (tmp != UDC_EP0IN_IX && tmp != UDC_EP0OUT_IX | 
|  | && tmp > UDC_EPIN_NUM) { | 
|  | /* set NAK */ | 
|  | reg = readl(&dev->ep[tmp].regs->ctl); | 
|  | reg |= AMD_BIT(UDC_EPCTL_SNAK); | 
|  | writel(reg, &dev->ep[tmp].regs->ctl); | 
|  | dev->ep[tmp].naking = 1; | 
|  |  | 
|  | } | 
|  | } | 
|  | } | 
|  | /* EP0 max packet */ | 
|  | if (dev->gadget.speed == USB_SPEED_FULL) { | 
|  | dev->ep[UDC_EP0IN_IX].ep.maxpacket = UDC_FS_EP0IN_MAX_PKT_SIZE; | 
|  | dev->ep[UDC_EP0OUT_IX].ep.maxpacket = | 
|  | UDC_FS_EP0OUT_MAX_PKT_SIZE; | 
|  | } else if (dev->gadget.speed == USB_SPEED_HIGH) { | 
|  | dev->ep[UDC_EP0IN_IX].ep.maxpacket = UDC_EP0IN_MAX_PKT_SIZE; | 
|  | dev->ep[UDC_EP0OUT_IX].ep.maxpacket = UDC_EP0OUT_MAX_PKT_SIZE; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * with suspend bug workaround, ep0 params for gadget driver | 
|  | * are set at gadget driver bind() call | 
|  | */ | 
|  | dev->gadget.ep0 = &dev->ep[UDC_EP0IN_IX].ep; | 
|  | dev->ep[UDC_EP0IN_IX].halted = 0; | 
|  | INIT_LIST_HEAD(&dev->gadget.ep0->ep_list); | 
|  |  | 
|  | /* init cfg/alt/int */ | 
|  | dev->cur_config = 0; | 
|  | dev->cur_intf = 0; | 
|  | dev->cur_alt = 0; | 
|  | } | 
|  |  | 
|  | /* Bringup after Connect event, initial bringup to be ready for ep0 events */ | 
|  | static void usb_connect(struct udc *dev) | 
|  | { | 
|  |  | 
|  | dev_info(&dev->pdev->dev, "USB Connect\n"); | 
|  |  | 
|  | dev->connected = 1; | 
|  |  | 
|  | /* put into initial config */ | 
|  | udc_basic_init(dev); | 
|  |  | 
|  | /* enable device setup interrupts */ | 
|  | udc_enable_dev_setup_interrupts(dev); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Calls gadget with disconnect event and resets the UDC and makes | 
|  | * initial bringup to be ready for ep0 events | 
|  | */ | 
|  | static void usb_disconnect(struct udc *dev) | 
|  | { | 
|  |  | 
|  | dev_info(&dev->pdev->dev, "USB Disconnect\n"); | 
|  |  | 
|  | dev->connected = 0; | 
|  |  | 
|  | /* mask interrupts */ | 
|  | udc_mask_unused_interrupts(dev); | 
|  |  | 
|  | /* REVISIT there doesn't seem to be a point to having this | 
|  | * talk to a tasklet ... do it directly, we already hold | 
|  | * the spinlock needed to process the disconnect. | 
|  | */ | 
|  |  | 
|  | tasklet_schedule(&disconnect_tasklet); | 
|  | } | 
|  |  | 
|  | /* Tasklet for disconnect to be outside of interrupt context */ | 
|  | static void udc_tasklet_disconnect(unsigned long par) | 
|  | { | 
|  | struct udc *dev = (struct udc *)(*((struct udc **) par)); | 
|  | u32 tmp; | 
|  |  | 
|  | DBG(dev, "Tasklet disconnect\n"); | 
|  | spin_lock_irq(&dev->lock); | 
|  |  | 
|  | if (dev->driver) { | 
|  | spin_unlock(&dev->lock); | 
|  | dev->driver->disconnect(&dev->gadget); | 
|  | spin_lock(&dev->lock); | 
|  |  | 
|  | /* empty queues */ | 
|  | for (tmp = 0; tmp < UDC_EP_NUM; tmp++) | 
|  | empty_req_queue(&dev->ep[tmp]); | 
|  |  | 
|  | } | 
|  |  | 
|  | /* disable ep0 */ | 
|  | ep_init(dev->regs, | 
|  | &dev->ep[UDC_EP0IN_IX]); | 
|  |  | 
|  |  | 
|  | if (!soft_reset_occured) { | 
|  | /* init controller by soft reset */ | 
|  | udc_soft_reset(dev); | 
|  | soft_reset_occured++; | 
|  | } | 
|  |  | 
|  | /* re-enable dev interrupts */ | 
|  | udc_enable_dev_setup_interrupts(dev); | 
|  | /* back to full speed ? */ | 
|  | if (use_fullspeed) { | 
|  | tmp = readl(&dev->regs->cfg); | 
|  | tmp = AMD_ADDBITS(tmp, UDC_DEVCFG_SPD_FS, UDC_DEVCFG_SPD); | 
|  | writel(tmp, &dev->regs->cfg); | 
|  | } | 
|  |  | 
|  | spin_unlock_irq(&dev->lock); | 
|  | } | 
|  |  | 
|  | /* Reset the UDC core */ | 
|  | static void udc_soft_reset(struct udc *dev) | 
|  | { | 
|  | unsigned long	flags; | 
|  |  | 
|  | DBG(dev, "Soft reset\n"); | 
|  | /* | 
|  | * reset possible waiting interrupts, because int. | 
|  | * status is lost after soft reset, | 
|  | * ep int. status reset | 
|  | */ | 
|  | writel(UDC_EPINT_MSK_DISABLE_ALL, &dev->regs->ep_irqsts); | 
|  | /* device int. status reset */ | 
|  | writel(UDC_DEV_MSK_DISABLE, &dev->regs->irqsts); | 
|  |  | 
|  | spin_lock_irqsave(&udc_irq_spinlock, flags); | 
|  | writel(AMD_BIT(UDC_DEVCFG_SOFTRESET), &dev->regs->cfg); | 
|  | readl(&dev->regs->cfg); | 
|  | spin_unlock_irqrestore(&udc_irq_spinlock, flags); | 
|  |  | 
|  | } | 
|  |  | 
|  | /* RDE timer callback to set RDE bit */ | 
|  | static void udc_timer_function(unsigned long v) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | spin_lock_irq(&udc_irq_spinlock); | 
|  |  | 
|  | if (set_rde > 0) { | 
|  | /* | 
|  | * open the fifo if fifo was filled on last timer call | 
|  | * conditionally | 
|  | */ | 
|  | if (set_rde > 1) { | 
|  | /* set RDE to receive setup data */ | 
|  | tmp = readl(&udc->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_RDE); | 
|  | writel(tmp, &udc->regs->ctl); | 
|  | set_rde = -1; | 
|  | } else if (readl(&udc->regs->sts) | 
|  | & AMD_BIT(UDC_DEVSTS_RXFIFO_EMPTY)) { | 
|  | /* | 
|  | * if fifo empty setup polling, do not just | 
|  | * open the fifo | 
|  | */ | 
|  | udc_timer.expires = jiffies + HZ/UDC_RDE_TIMER_DIV; | 
|  | if (!stop_timer) | 
|  | add_timer(&udc_timer); | 
|  | } else { | 
|  | /* | 
|  | * fifo contains data now, setup timer for opening | 
|  | * the fifo when timer expires to be able to receive | 
|  | * setup packets, when data packets gets queued by | 
|  | * gadget layer then timer will forced to expire with | 
|  | * set_rde=0 (RDE is set in udc_queue()) | 
|  | */ | 
|  | set_rde++; | 
|  | /* debug: lhadmot_timer_start = 221070 */ | 
|  | udc_timer.expires = jiffies + HZ*UDC_RDE_TIMER_SECONDS; | 
|  | if (!stop_timer) | 
|  | add_timer(&udc_timer); | 
|  | } | 
|  |  | 
|  | } else | 
|  | set_rde = -1; /* RDE was set by udc_queue() */ | 
|  | spin_unlock_irq(&udc_irq_spinlock); | 
|  | if (stop_timer) | 
|  | complete(&on_exit); | 
|  |  | 
|  | } | 
|  |  | 
|  | /* Handle halt state, used in stall poll timer */ | 
|  | static void udc_handle_halt_state(struct udc_ep *ep) | 
|  | { | 
|  | u32 tmp; | 
|  | /* set stall as long not halted */ | 
|  | if (ep->halted == 1) { | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | /* STALL cleared ? */ | 
|  | if (!(tmp & AMD_BIT(UDC_EPCTL_S))) { | 
|  | /* | 
|  | * FIXME: MSC spec requires that stall remains | 
|  | * even on receivng of CLEAR_FEATURE HALT. So | 
|  | * we would set STALL again here to be compliant. | 
|  | * But with current mass storage drivers this does | 
|  | * not work (would produce endless host retries). | 
|  | * So we clear halt on CLEAR_FEATURE. | 
|  | * | 
|  | DBG(ep->dev, "ep %d: set STALL again\n", ep->num); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_S); | 
|  | writel(tmp, &ep->regs->ctl);*/ | 
|  |  | 
|  | /* clear NAK by writing CNAK */ | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | ep->halted = 0; | 
|  | UDC_QUEUE_CNAK(ep, ep->num); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Stall timer callback to poll S bit and set it again after */ | 
|  | static void udc_pollstall_timer_function(unsigned long v) | 
|  | { | 
|  | struct udc_ep *ep; | 
|  | int halted = 0; | 
|  |  | 
|  | spin_lock_irq(&udc_stall_spinlock); | 
|  | /* | 
|  | * only one IN and OUT endpoints are handled | 
|  | * IN poll stall | 
|  | */ | 
|  | ep = &udc->ep[UDC_EPIN_IX]; | 
|  | udc_handle_halt_state(ep); | 
|  | if (ep->halted) | 
|  | halted = 1; | 
|  | /* OUT poll stall */ | 
|  | ep = &udc->ep[UDC_EPOUT_IX]; | 
|  | udc_handle_halt_state(ep); | 
|  | if (ep->halted) | 
|  | halted = 1; | 
|  |  | 
|  | /* setup timer again when still halted */ | 
|  | if (!stop_pollstall_timer && halted) { | 
|  | udc_pollstall_timer.expires = jiffies + | 
|  | HZ * UDC_POLLSTALL_TIMER_USECONDS | 
|  | / (1000 * 1000); | 
|  | add_timer(&udc_pollstall_timer); | 
|  | } | 
|  | spin_unlock_irq(&udc_stall_spinlock); | 
|  |  | 
|  | if (stop_pollstall_timer) | 
|  | complete(&on_pollstall_exit); | 
|  | } | 
|  |  | 
|  | /* Inits endpoint 0 so that SETUP packets are processed */ | 
|  | static void activate_control_endpoints(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  |  | 
|  | DBG(dev, "activate_control_endpoints\n"); | 
|  |  | 
|  | /* flush fifo */ | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_F); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  |  | 
|  | /* set ep0 directions */ | 
|  | dev->ep[UDC_EP0IN_IX].in = 1; | 
|  | dev->ep[UDC_EP0OUT_IX].in = 0; | 
|  |  | 
|  | /* set buffer size (tx fifo entries) of EP0_IN */ | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->bufin_framenum); | 
|  | if (dev->gadget.speed == USB_SPEED_FULL) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_FS_EPIN0_BUFF_SIZE, | 
|  | UDC_EPIN_BUFF_SIZE); | 
|  | else if (dev->gadget.speed == USB_SPEED_HIGH) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_EPIN0_BUFF_SIZE, | 
|  | UDC_EPIN_BUFF_SIZE); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->bufin_framenum); | 
|  |  | 
|  | /* set max packet size of EP0_IN */ | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->bufout_maxpkt); | 
|  | if (dev->gadget.speed == USB_SPEED_FULL) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_FS_EP0IN_MAX_PKT_SIZE, | 
|  | UDC_EP_MAX_PKT_SIZE); | 
|  | else if (dev->gadget.speed == USB_SPEED_HIGH) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_EP0IN_MAX_PKT_SIZE, | 
|  | UDC_EP_MAX_PKT_SIZE); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->bufout_maxpkt); | 
|  |  | 
|  | /* set max packet size of EP0_OUT */ | 
|  | tmp = readl(&dev->ep[UDC_EP0OUT_IX].regs->bufout_maxpkt); | 
|  | if (dev->gadget.speed == USB_SPEED_FULL) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_FS_EP0OUT_MAX_PKT_SIZE, | 
|  | UDC_EP_MAX_PKT_SIZE); | 
|  | else if (dev->gadget.speed == USB_SPEED_HIGH) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_EP0OUT_MAX_PKT_SIZE, | 
|  | UDC_EP_MAX_PKT_SIZE); | 
|  | writel(tmp, &dev->ep[UDC_EP0OUT_IX].regs->bufout_maxpkt); | 
|  |  | 
|  | /* set max packet size of EP0 in UDC CSR */ | 
|  | tmp = readl(&dev->csr->ne[0]); | 
|  | if (dev->gadget.speed == USB_SPEED_FULL) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_FS_EP0OUT_MAX_PKT_SIZE, | 
|  | UDC_CSR_NE_MAX_PKT); | 
|  | else if (dev->gadget.speed == USB_SPEED_HIGH) | 
|  | tmp = AMD_ADDBITS(tmp, UDC_EP0OUT_MAX_PKT_SIZE, | 
|  | UDC_CSR_NE_MAX_PKT); | 
|  | writel(tmp, &dev->csr->ne[0]); | 
|  |  | 
|  | if (use_dma) { | 
|  | dev->ep[UDC_EP0OUT_IX].td->status |= | 
|  | AMD_BIT(UDC_DMA_OUT_STS_L); | 
|  | /* write dma desc address */ | 
|  | writel(dev->ep[UDC_EP0OUT_IX].td_stp_dma, | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->subptr); | 
|  | writel(dev->ep[UDC_EP0OUT_IX].td_phys, | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->desptr); | 
|  | /* stop RDE timer */ | 
|  | if (timer_pending(&udc_timer)) { | 
|  | set_rde = 0; | 
|  | mod_timer(&udc_timer, jiffies - 1); | 
|  | } | 
|  | /* stop pollstall timer */ | 
|  | if (timer_pending(&udc_pollstall_timer)) | 
|  | mod_timer(&udc_pollstall_timer, jiffies - 1); | 
|  | /* enable DMA */ | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_MODE) | 
|  | | AMD_BIT(UDC_DEVCTL_RDE) | 
|  | | AMD_BIT(UDC_DEVCTL_TDE); | 
|  | if (use_dma_bufferfill_mode) | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_BF); | 
|  | else if (use_dma_ppb_du) | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_DU); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  | } | 
|  |  | 
|  | /* clear NAK by writing CNAK for EP0IN */ | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0IN_IX].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[UDC_EP0IN_IX], UDC_EP0IN_IX); | 
|  |  | 
|  | /* clear NAK by writing CNAK for EP0OUT */ | 
|  | tmp = readl(&dev->ep[UDC_EP0OUT_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &dev->ep[UDC_EP0OUT_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0OUT_IX].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[UDC_EP0OUT_IX], UDC_EP0OUT_IX); | 
|  | } | 
|  |  | 
|  | /* Make endpoint 0 ready for control traffic */ | 
|  | static int setup_ep0(struct udc *dev) | 
|  | { | 
|  | activate_control_endpoints(dev); | 
|  | /* enable ep0 interrupts */ | 
|  | udc_enable_ep0_interrupts(dev); | 
|  | /* enable device setup interrupts */ | 
|  | udc_enable_dev_setup_interrupts(dev); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Called by gadget driver to register itself */ | 
|  | static int amd5536_start(struct usb_gadget_driver *driver, | 
|  | int (*bind)(struct usb_gadget *)) | 
|  | { | 
|  | struct udc		*dev = udc; | 
|  | int			retval; | 
|  | u32 tmp; | 
|  |  | 
|  | if (!driver || !bind || !driver->setup | 
|  | || driver->max_speed < USB_SPEED_HIGH) | 
|  | return -EINVAL; | 
|  | if (!dev) | 
|  | return -ENODEV; | 
|  | if (dev->driver) | 
|  | return -EBUSY; | 
|  |  | 
|  | driver->driver.bus = NULL; | 
|  | dev->driver = driver; | 
|  | dev->gadget.dev.driver = &driver->driver; | 
|  |  | 
|  | retval = bind(&dev->gadget); | 
|  |  | 
|  | /* Some gadget drivers use both ep0 directions. | 
|  | * NOTE: to gadget driver, ep0 is just one endpoint... | 
|  | */ | 
|  | dev->ep[UDC_EP0OUT_IX].ep.driver_data = | 
|  | dev->ep[UDC_EP0IN_IX].ep.driver_data; | 
|  |  | 
|  | if (retval) { | 
|  | DBG(dev, "binding to %s returning %d\n", | 
|  | driver->driver.name, retval); | 
|  | dev->driver = NULL; | 
|  | dev->gadget.dev.driver = NULL; | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* get ready for ep0 traffic */ | 
|  | setup_ep0(dev); | 
|  |  | 
|  | /* clear SD */ | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp = tmp & AMD_CLEAR_BIT(UDC_DEVCTL_SD); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  |  | 
|  | usb_connect(dev); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* shutdown requests and disconnect from gadget */ | 
|  | static void | 
|  | shutdown(struct udc *dev, struct usb_gadget_driver *driver) | 
|  | __releases(dev->lock) | 
|  | __acquires(dev->lock) | 
|  | { | 
|  | int tmp; | 
|  |  | 
|  | if (dev->gadget.speed != USB_SPEED_UNKNOWN) { | 
|  | spin_unlock(&dev->lock); | 
|  | driver->disconnect(&dev->gadget); | 
|  | spin_lock(&dev->lock); | 
|  | } | 
|  |  | 
|  | /* empty queues and init hardware */ | 
|  | udc_basic_init(dev); | 
|  | for (tmp = 0; tmp < UDC_EP_NUM; tmp++) | 
|  | empty_req_queue(&dev->ep[tmp]); | 
|  |  | 
|  | udc_setup_endpoints(dev); | 
|  | } | 
|  |  | 
|  | /* Called by gadget driver to unregister itself */ | 
|  | static int amd5536_stop(struct usb_gadget_driver *driver) | 
|  | { | 
|  | struct udc	*dev = udc; | 
|  | unsigned long	flags; | 
|  | u32 tmp; | 
|  |  | 
|  | if (!dev) | 
|  | return -ENODEV; | 
|  | if (!driver || driver != dev->driver || !driver->unbind) | 
|  | return -EINVAL; | 
|  |  | 
|  | spin_lock_irqsave(&dev->lock, flags); | 
|  | udc_mask_unused_interrupts(dev); | 
|  | shutdown(dev, driver); | 
|  | spin_unlock_irqrestore(&dev->lock, flags); | 
|  |  | 
|  | driver->unbind(&dev->gadget); | 
|  | dev->gadget.dev.driver = NULL; | 
|  | dev->driver = NULL; | 
|  |  | 
|  | /* set SD */ | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_SD); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  |  | 
|  |  | 
|  | DBG(dev, "%s: unregistered\n", driver->driver.name); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Clear pending NAK bits */ | 
|  | static void udc_process_cnak_queue(struct udc *dev) | 
|  | { | 
|  | u32 tmp; | 
|  | u32 reg; | 
|  |  | 
|  | /* check epin's */ | 
|  | DBG(dev, "CNAK pending queue processing\n"); | 
|  | for (tmp = 0; tmp < UDC_EPIN_NUM_USED; tmp++) { | 
|  | if (cnak_pending & (1 << tmp)) { | 
|  | DBG(dev, "CNAK pending for ep%d\n", tmp); | 
|  | /* clear NAK by writing CNAK */ | 
|  | reg = readl(&dev->ep[tmp].regs->ctl); | 
|  | reg |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(reg, &dev->ep[tmp].regs->ctl); | 
|  | dev->ep[tmp].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[tmp], dev->ep[tmp].num); | 
|  | } | 
|  | } | 
|  | /* ...	and ep0out */ | 
|  | if (cnak_pending & (1 << UDC_EP0OUT_IX)) { | 
|  | DBG(dev, "CNAK pending for ep%d\n", UDC_EP0OUT_IX); | 
|  | /* clear NAK by writing CNAK */ | 
|  | reg = readl(&dev->ep[UDC_EP0OUT_IX].regs->ctl); | 
|  | reg |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(reg, &dev->ep[UDC_EP0OUT_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0OUT_IX].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[UDC_EP0OUT_IX], | 
|  | dev->ep[UDC_EP0OUT_IX].num); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Enabling RX DMA after setup packet */ | 
|  | static void udc_ep0_set_rde(struct udc *dev) | 
|  | { | 
|  | if (use_dma) { | 
|  | /* | 
|  | * only enable RXDMA when no data endpoint enabled | 
|  | * or data is queued | 
|  | */ | 
|  | if (!dev->data_ep_enabled || dev->data_ep_queued) { | 
|  | udc_set_rde(dev); | 
|  | } else { | 
|  | /* | 
|  | * setup timer for enabling RDE (to not enable | 
|  | * RXFIFO DMA for data endpoints to early) | 
|  | */ | 
|  | if (set_rde != 0 && !timer_pending(&udc_timer)) { | 
|  | udc_timer.expires = | 
|  | jiffies + HZ/UDC_RDE_TIMER_DIV; | 
|  | set_rde = 1; | 
|  | if (!stop_timer) | 
|  | add_timer(&udc_timer); | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | /* Interrupt handler for data OUT traffic */ | 
|  | static irqreturn_t udc_data_out_isr(struct udc *dev, int ep_ix) | 
|  | { | 
|  | irqreturn_t		ret_val = IRQ_NONE; | 
|  | u32			tmp; | 
|  | struct udc_ep		*ep; | 
|  | struct udc_request	*req; | 
|  | unsigned int		count; | 
|  | struct udc_data_dma	*td = NULL; | 
|  | unsigned		dma_done; | 
|  |  | 
|  | VDBG(dev, "ep%d irq\n", ep_ix); | 
|  | ep = &dev->ep[ep_ix]; | 
|  |  | 
|  | tmp = readl(&ep->regs->sts); | 
|  | if (use_dma) { | 
|  | /* BNA event ? */ | 
|  | if (tmp & AMD_BIT(UDC_EPSTS_BNA)) { | 
|  | DBG(dev, "BNA ep%dout occurred - DESPTR = %x\n", | 
|  | ep->num, readl(&ep->regs->desptr)); | 
|  | /* clear BNA */ | 
|  | writel(tmp | AMD_BIT(UDC_EPSTS_BNA), &ep->regs->sts); | 
|  | if (!ep->cancel_transfer) | 
|  | ep->bna_occurred = 1; | 
|  | else | 
|  | ep->cancel_transfer = 0; | 
|  | ret_val = IRQ_HANDLED; | 
|  | goto finished; | 
|  | } | 
|  | } | 
|  | /* HE event ? */ | 
|  | if (tmp & AMD_BIT(UDC_EPSTS_HE)) { | 
|  | dev_err(&dev->pdev->dev, "HE ep%dout occurred\n", ep->num); | 
|  |  | 
|  | /* clear HE */ | 
|  | writel(tmp | AMD_BIT(UDC_EPSTS_HE), &ep->regs->sts); | 
|  | ret_val = IRQ_HANDLED; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | if (!list_empty(&ep->queue)) { | 
|  |  | 
|  | /* next request */ | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, queue); | 
|  | } else { | 
|  | req = NULL; | 
|  | udc_rxfifo_pending = 1; | 
|  | } | 
|  | VDBG(dev, "req = %p\n", req); | 
|  | /* fifo mode */ | 
|  | if (!use_dma) { | 
|  |  | 
|  | /* read fifo */ | 
|  | if (req && udc_rxfifo_read(ep, req)) { | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | /* finish */ | 
|  | complete_req(ep, req, 0); | 
|  | /* next request */ | 
|  | if (!list_empty(&ep->queue) && !ep->halted) { | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, queue); | 
|  | } else | 
|  | req = NULL; | 
|  | } | 
|  |  | 
|  | /* DMA */ | 
|  | } else if (!ep->cancel_transfer && req != NULL) { | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | /* check for DMA done */ | 
|  | if (!use_dma_ppb) { | 
|  | dma_done = AMD_GETBITS(req->td_data->status, | 
|  | UDC_DMA_OUT_STS_BS); | 
|  | /* packet per buffer mode - rx bytes */ | 
|  | } else { | 
|  | /* | 
|  | * if BNA occurred then recover desc. from | 
|  | * BNA dummy desc. | 
|  | */ | 
|  | if (ep->bna_occurred) { | 
|  | VDBG(dev, "Recover desc. from BNA dummy\n"); | 
|  | memcpy(req->td_data, ep->bna_dummy_req->td_data, | 
|  | sizeof(struct udc_data_dma)); | 
|  | ep->bna_occurred = 0; | 
|  | udc_init_bna_dummy(ep->req); | 
|  | } | 
|  | td = udc_get_last_dma_desc(req); | 
|  | dma_done = AMD_GETBITS(td->status, UDC_DMA_OUT_STS_BS); | 
|  | } | 
|  | if (dma_done == UDC_DMA_OUT_STS_BS_DMA_DONE) { | 
|  | /* buffer fill mode - rx bytes */ | 
|  | if (!use_dma_ppb) { | 
|  | /* received number bytes */ | 
|  | count = AMD_GETBITS(req->td_data->status, | 
|  | UDC_DMA_OUT_STS_RXBYTES); | 
|  | VDBG(dev, "rx bytes=%u\n", count); | 
|  | /* packet per buffer mode - rx bytes */ | 
|  | } else { | 
|  | VDBG(dev, "req->td_data=%p\n", req->td_data); | 
|  | VDBG(dev, "last desc = %p\n", td); | 
|  | /* received number bytes */ | 
|  | if (use_dma_ppb_du) { | 
|  | /* every desc. counts bytes */ | 
|  | count = udc_get_ppbdu_rxbytes(req); | 
|  | } else { | 
|  | /* last desc. counts bytes */ | 
|  | count = AMD_GETBITS(td->status, | 
|  | UDC_DMA_OUT_STS_RXBYTES); | 
|  | if (!count && req->req.length | 
|  | == UDC_DMA_MAXPACKET) { | 
|  | /* | 
|  | * on 64k packets the RXBYTES | 
|  | * field is zero | 
|  | */ | 
|  | count = UDC_DMA_MAXPACKET; | 
|  | } | 
|  | } | 
|  | VDBG(dev, "last desc rx bytes=%u\n", count); | 
|  | } | 
|  |  | 
|  | tmp = req->req.length - req->req.actual; | 
|  | if (count > tmp) { | 
|  | if ((tmp % ep->ep.maxpacket) != 0) { | 
|  | DBG(dev, "%s: rx %db, space=%db\n", | 
|  | ep->ep.name, count, tmp); | 
|  | req->req.status = -EOVERFLOW; | 
|  | } | 
|  | count = tmp; | 
|  | } | 
|  | req->req.actual += count; | 
|  | req->dma_going = 0; | 
|  | /* complete request */ | 
|  | complete_req(ep, req, 0); | 
|  |  | 
|  | /* next request */ | 
|  | if (!list_empty(&ep->queue) && !ep->halted) { | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, | 
|  | queue); | 
|  | /* | 
|  | * DMA may be already started by udc_queue() | 
|  | * called by gadget drivers completion | 
|  | * routine. This happens when queue | 
|  | * holds one request only. | 
|  | */ | 
|  | if (req->dma_going == 0) { | 
|  | /* next dma */ | 
|  | if (prep_dma(ep, req, GFP_ATOMIC) != 0) | 
|  | goto finished; | 
|  | /* write desc pointer */ | 
|  | writel(req->td_phys, | 
|  | &ep->regs->desptr); | 
|  | req->dma_going = 1; | 
|  | /* enable DMA */ | 
|  | udc_set_rde(dev); | 
|  | } | 
|  | } else { | 
|  | /* | 
|  | * implant BNA dummy descriptor to allow | 
|  | * RXFIFO opening by RDE | 
|  | */ | 
|  | if (ep->bna_dummy_req) { | 
|  | /* write desc pointer */ | 
|  | writel(ep->bna_dummy_req->td_phys, | 
|  | &ep->regs->desptr); | 
|  | ep->bna_occurred = 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * schedule timer for setting RDE if queue | 
|  | * remains empty to allow ep0 packets pass | 
|  | * through | 
|  | */ | 
|  | if (set_rde != 0 | 
|  | && !timer_pending(&udc_timer)) { | 
|  | udc_timer.expires = | 
|  | jiffies | 
|  | + HZ*UDC_RDE_TIMER_SECONDS; | 
|  | set_rde = 1; | 
|  | if (!stop_timer) | 
|  | add_timer(&udc_timer); | 
|  | } | 
|  | if (ep->num != UDC_EP0OUT_IX) | 
|  | dev->data_ep_queued = 0; | 
|  | } | 
|  |  | 
|  | } else { | 
|  | /* | 
|  | * RX DMA must be reenabled for each desc in PPBDU mode | 
|  | * and must be enabled for PPBNDU mode in case of BNA | 
|  | */ | 
|  | udc_set_rde(dev); | 
|  | } | 
|  |  | 
|  | } else if (ep->cancel_transfer) { | 
|  | ret_val = IRQ_HANDLED; | 
|  | ep->cancel_transfer = 0; | 
|  | } | 
|  |  | 
|  | /* check pending CNAKS */ | 
|  | if (cnak_pending) { | 
|  | /* CNAk processing when rxfifo empty only */ | 
|  | if (readl(&dev->regs->sts) & AMD_BIT(UDC_DEVSTS_RXFIFO_EMPTY)) | 
|  | udc_process_cnak_queue(dev); | 
|  | } | 
|  |  | 
|  | /* clear OUT bits in ep status */ | 
|  | writel(UDC_EPSTS_OUT_CLEAR, &ep->regs->sts); | 
|  | finished: | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /* Interrupt handler for data IN traffic */ | 
|  | static irqreturn_t udc_data_in_isr(struct udc *dev, int ep_ix) | 
|  | { | 
|  | irqreturn_t ret_val = IRQ_NONE; | 
|  | u32 tmp; | 
|  | u32 epsts; | 
|  | struct udc_ep *ep; | 
|  | struct udc_request *req; | 
|  | struct udc_data_dma *td; | 
|  | unsigned dma_done; | 
|  | unsigned len; | 
|  |  | 
|  | ep = &dev->ep[ep_ix]; | 
|  |  | 
|  | epsts = readl(&ep->regs->sts); | 
|  | if (use_dma) { | 
|  | /* BNA ? */ | 
|  | if (epsts & AMD_BIT(UDC_EPSTS_BNA)) { | 
|  | dev_err(&dev->pdev->dev, | 
|  | "BNA ep%din occurred - DESPTR = %08lx\n", | 
|  | ep->num, | 
|  | (unsigned long) readl(&ep->regs->desptr)); | 
|  |  | 
|  | /* clear BNA */ | 
|  | writel(epsts, &ep->regs->sts); | 
|  | ret_val = IRQ_HANDLED; | 
|  | goto finished; | 
|  | } | 
|  | } | 
|  | /* HE event ? */ | 
|  | if (epsts & AMD_BIT(UDC_EPSTS_HE)) { | 
|  | dev_err(&dev->pdev->dev, | 
|  | "HE ep%dn occurred - DESPTR = %08lx\n", | 
|  | ep->num, (unsigned long) readl(&ep->regs->desptr)); | 
|  |  | 
|  | /* clear HE */ | 
|  | writel(epsts | AMD_BIT(UDC_EPSTS_HE), &ep->regs->sts); | 
|  | ret_val = IRQ_HANDLED; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | /* DMA completion */ | 
|  | if (epsts & AMD_BIT(UDC_EPSTS_TDC)) { | 
|  | VDBG(dev, "TDC set- completion\n"); | 
|  | ret_val = IRQ_HANDLED; | 
|  | if (!ep->cancel_transfer && !list_empty(&ep->queue)) { | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, queue); | 
|  | /* | 
|  | * length bytes transferred | 
|  | * check dma done of last desc. in PPBDU mode | 
|  | */ | 
|  | if (use_dma_ppb_du) { | 
|  | td = udc_get_last_dma_desc(req); | 
|  | if (td) { | 
|  | dma_done = | 
|  | AMD_GETBITS(td->status, | 
|  | UDC_DMA_IN_STS_BS); | 
|  | /* don't care DMA done */ | 
|  | req->req.actual = req->req.length; | 
|  | } | 
|  | } else { | 
|  | /* assume all bytes transferred */ | 
|  | req->req.actual = req->req.length; | 
|  | } | 
|  |  | 
|  | if (req->req.actual == req->req.length) { | 
|  | /* complete req */ | 
|  | complete_req(ep, req, 0); | 
|  | req->dma_going = 0; | 
|  | /* further request available ? */ | 
|  | if (list_empty(&ep->queue)) { | 
|  | /* disable interrupt */ | 
|  | tmp = readl(&dev->regs->ep_irqmsk); | 
|  | tmp |= AMD_BIT(ep->num); | 
|  | writel(tmp, &dev->regs->ep_irqmsk); | 
|  | } | 
|  | } | 
|  | } | 
|  | ep->cancel_transfer = 0; | 
|  |  | 
|  | } | 
|  | /* | 
|  | * status reg has IN bit set and TDC not set (if TDC was handled, | 
|  | * IN must not be handled (UDC defect) ? | 
|  | */ | 
|  | if ((epsts & AMD_BIT(UDC_EPSTS_IN)) | 
|  | && !(epsts & AMD_BIT(UDC_EPSTS_TDC))) { | 
|  | ret_val = IRQ_HANDLED; | 
|  | if (!list_empty(&ep->queue)) { | 
|  | /* next request */ | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, queue); | 
|  | /* FIFO mode */ | 
|  | if (!use_dma) { | 
|  | /* write fifo */ | 
|  | udc_txfifo_write(ep, &req->req); | 
|  | len = req->req.length - req->req.actual; | 
|  | if (len > ep->ep.maxpacket) | 
|  | len = ep->ep.maxpacket; | 
|  | req->req.actual += len; | 
|  | if (req->req.actual == req->req.length | 
|  | || (len != ep->ep.maxpacket)) { | 
|  | /* complete req */ | 
|  | complete_req(ep, req, 0); | 
|  | } | 
|  | /* DMA */ | 
|  | } else if (req && !req->dma_going) { | 
|  | VDBG(dev, "IN DMA : req=%p req->td_data=%p\n", | 
|  | req, req->td_data); | 
|  | if (req->td_data) { | 
|  |  | 
|  | req->dma_going = 1; | 
|  |  | 
|  | /* | 
|  | * unset L bit of first desc. | 
|  | * for chain | 
|  | */ | 
|  | if (use_dma_ppb && req->req.length > | 
|  | ep->ep.maxpacket) { | 
|  | req->td_data->status &= | 
|  | AMD_CLEAR_BIT( | 
|  | UDC_DMA_IN_STS_L); | 
|  | } | 
|  |  | 
|  | /* write desc pointer */ | 
|  | writel(req->td_phys, &ep->regs->desptr); | 
|  |  | 
|  | /* set HOST READY */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS( | 
|  | req->td_data->status, | 
|  | UDC_DMA_IN_STS_BS_HOST_READY, | 
|  | UDC_DMA_IN_STS_BS); | 
|  |  | 
|  | /* set poll demand bit */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_P); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | } | 
|  | } | 
|  |  | 
|  | } else if (!use_dma && ep->in) { | 
|  | /* disable interrupt */ | 
|  | tmp = readl( | 
|  | &dev->regs->ep_irqmsk); | 
|  | tmp |= AMD_BIT(ep->num); | 
|  | writel(tmp, | 
|  | &dev->regs->ep_irqmsk); | 
|  | } | 
|  | } | 
|  | /* clear status bits */ | 
|  | writel(epsts, &ep->regs->sts); | 
|  |  | 
|  | finished: | 
|  | return ret_val; | 
|  |  | 
|  | } | 
|  |  | 
|  | /* Interrupt handler for Control OUT traffic */ | 
|  | static irqreturn_t udc_control_out_isr(struct udc *dev) | 
|  | __releases(dev->lock) | 
|  | __acquires(dev->lock) | 
|  | { | 
|  | irqreturn_t ret_val = IRQ_NONE; | 
|  | u32 tmp; | 
|  | int setup_supported; | 
|  | u32 count; | 
|  | int set = 0; | 
|  | struct udc_ep	*ep; | 
|  | struct udc_ep	*ep_tmp; | 
|  |  | 
|  | ep = &dev->ep[UDC_EP0OUT_IX]; | 
|  |  | 
|  | /* clear irq */ | 
|  | writel(AMD_BIT(UDC_EPINT_OUT_EP0), &dev->regs->ep_irqsts); | 
|  |  | 
|  | tmp = readl(&dev->ep[UDC_EP0OUT_IX].regs->sts); | 
|  | /* check BNA and clear if set */ | 
|  | if (tmp & AMD_BIT(UDC_EPSTS_BNA)) { | 
|  | VDBG(dev, "ep0: BNA set\n"); | 
|  | writel(AMD_BIT(UDC_EPSTS_BNA), | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->sts); | 
|  | ep->bna_occurred = 1; | 
|  | ret_val = IRQ_HANDLED; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | /* type of data: SETUP or DATA 0 bytes */ | 
|  | tmp = AMD_GETBITS(tmp, UDC_EPSTS_OUT); | 
|  | VDBG(dev, "data_typ = %x\n", tmp); | 
|  |  | 
|  | /* setup data */ | 
|  | if (tmp == UDC_EPSTS_OUT_SETUP) { | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | ep->dev->stall_ep0in = 0; | 
|  | dev->waiting_zlp_ack_ep0in = 0; | 
|  |  | 
|  | /* set NAK for EP0_IN */ | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_SNAK); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0IN_IX].naking = 1; | 
|  | /* get setup data */ | 
|  | if (use_dma) { | 
|  |  | 
|  | /* clear OUT bits in ep status */ | 
|  | writel(UDC_EPSTS_OUT_CLEAR, | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->sts); | 
|  |  | 
|  | setup_data.data[0] = | 
|  | dev->ep[UDC_EP0OUT_IX].td_stp->data12; | 
|  | setup_data.data[1] = | 
|  | dev->ep[UDC_EP0OUT_IX].td_stp->data34; | 
|  | /* set HOST READY */ | 
|  | dev->ep[UDC_EP0OUT_IX].td_stp->status = | 
|  | UDC_DMA_STP_STS_BS_HOST_READY; | 
|  | } else { | 
|  | /* read fifo */ | 
|  | udc_rxfifo_read_dwords(dev, setup_data.data, 2); | 
|  | } | 
|  |  | 
|  | /* determine direction of control data */ | 
|  | if ((setup_data.request.bRequestType & USB_DIR_IN) != 0) { | 
|  | dev->gadget.ep0 = &dev->ep[UDC_EP0IN_IX].ep; | 
|  | /* enable RDE */ | 
|  | udc_ep0_set_rde(dev); | 
|  | set = 0; | 
|  | } else { | 
|  | dev->gadget.ep0 = &dev->ep[UDC_EP0OUT_IX].ep; | 
|  | /* | 
|  | * implant BNA dummy descriptor to allow RXFIFO opening | 
|  | * by RDE | 
|  | */ | 
|  | if (ep->bna_dummy_req) { | 
|  | /* write desc pointer */ | 
|  | writel(ep->bna_dummy_req->td_phys, | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->desptr); | 
|  | ep->bna_occurred = 0; | 
|  | } | 
|  |  | 
|  | set = 1; | 
|  | dev->ep[UDC_EP0OUT_IX].naking = 1; | 
|  | /* | 
|  | * setup timer for enabling RDE (to not enable | 
|  | * RXFIFO DMA for data to early) | 
|  | */ | 
|  | set_rde = 1; | 
|  | if (!timer_pending(&udc_timer)) { | 
|  | udc_timer.expires = jiffies + | 
|  | HZ/UDC_RDE_TIMER_DIV; | 
|  | if (!stop_timer) | 
|  | add_timer(&udc_timer); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * mass storage reset must be processed here because | 
|  | * next packet may be a CLEAR_FEATURE HALT which would not | 
|  | * clear the stall bit when no STALL handshake was received | 
|  | * before (autostall can cause this) | 
|  | */ | 
|  | if (setup_data.data[0] == UDC_MSCRES_DWORD0 | 
|  | && setup_data.data[1] == UDC_MSCRES_DWORD1) { | 
|  | DBG(dev, "MSC Reset\n"); | 
|  | /* | 
|  | * clear stall bits | 
|  | * only one IN and OUT endpoints are handled | 
|  | */ | 
|  | ep_tmp = &udc->ep[UDC_EPIN_IX]; | 
|  | udc_set_halt(&ep_tmp->ep, 0); | 
|  | ep_tmp = &udc->ep[UDC_EPOUT_IX]; | 
|  | udc_set_halt(&ep_tmp->ep, 0); | 
|  | } | 
|  |  | 
|  | /* call gadget with setup data received */ | 
|  | spin_unlock(&dev->lock); | 
|  | setup_supported = dev->driver->setup(&dev->gadget, | 
|  | &setup_data.request); | 
|  | spin_lock(&dev->lock); | 
|  |  | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | /* ep0 in returns data (not zlp) on IN phase */ | 
|  | if (setup_supported >= 0 && setup_supported < | 
|  | UDC_EP0IN_MAXPACKET) { | 
|  | /* clear NAK by writing CNAK in EP0_IN */ | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0IN_IX].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[UDC_EP0IN_IX], UDC_EP0IN_IX); | 
|  |  | 
|  | /* if unsupported request then stall */ | 
|  | } else if (setup_supported < 0) { | 
|  | tmp |= AMD_BIT(UDC_EPCTL_S); | 
|  | writel(tmp, &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | } else | 
|  | dev->waiting_zlp_ack_ep0in = 1; | 
|  |  | 
|  |  | 
|  | /* clear NAK by writing CNAK in EP0_OUT */ | 
|  | if (!set) { | 
|  | tmp = readl(&dev->ep[UDC_EP0OUT_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_CNAK); | 
|  | writel(tmp, &dev->ep[UDC_EP0OUT_IX].regs->ctl); | 
|  | dev->ep[UDC_EP0OUT_IX].naking = 0; | 
|  | UDC_QUEUE_CNAK(&dev->ep[UDC_EP0OUT_IX], UDC_EP0OUT_IX); | 
|  | } | 
|  |  | 
|  | if (!use_dma) { | 
|  | /* clear OUT bits in ep status */ | 
|  | writel(UDC_EPSTS_OUT_CLEAR, | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->sts); | 
|  | } | 
|  |  | 
|  | /* data packet 0 bytes */ | 
|  | } else if (tmp == UDC_EPSTS_OUT_DATA) { | 
|  | /* clear OUT bits in ep status */ | 
|  | writel(UDC_EPSTS_OUT_CLEAR, &dev->ep[UDC_EP0OUT_IX].regs->sts); | 
|  |  | 
|  | /* get setup data: only 0 packet */ | 
|  | if (use_dma) { | 
|  | /* no req if 0 packet, just reactivate */ | 
|  | if (list_empty(&dev->ep[UDC_EP0OUT_IX].queue)) { | 
|  | VDBG(dev, "ZLP\n"); | 
|  |  | 
|  | /* set HOST READY */ | 
|  | dev->ep[UDC_EP0OUT_IX].td->status = | 
|  | AMD_ADDBITS( | 
|  | dev->ep[UDC_EP0OUT_IX].td->status, | 
|  | UDC_DMA_OUT_STS_BS_HOST_READY, | 
|  | UDC_DMA_OUT_STS_BS); | 
|  | /* enable RDE */ | 
|  | udc_ep0_set_rde(dev); | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | } else { | 
|  | /* control write */ | 
|  | ret_val |= udc_data_out_isr(dev, UDC_EP0OUT_IX); | 
|  | /* re-program desc. pointer for possible ZLPs */ | 
|  | writel(dev->ep[UDC_EP0OUT_IX].td_phys, | 
|  | &dev->ep[UDC_EP0OUT_IX].regs->desptr); | 
|  | /* enable RDE */ | 
|  | udc_ep0_set_rde(dev); | 
|  | } | 
|  | } else { | 
|  |  | 
|  | /* received number bytes */ | 
|  | count = readl(&dev->ep[UDC_EP0OUT_IX].regs->sts); | 
|  | count = AMD_GETBITS(count, UDC_EPSTS_RX_PKT_SIZE); | 
|  | /* out data for fifo mode not working */ | 
|  | count = 0; | 
|  |  | 
|  | /* 0 packet or real data ? */ | 
|  | if (count != 0) { | 
|  | ret_val |= udc_data_out_isr(dev, UDC_EP0OUT_IX); | 
|  | } else { | 
|  | /* dummy read confirm */ | 
|  | readl(&dev->ep[UDC_EP0OUT_IX].regs->confirm); | 
|  | ret_val = IRQ_HANDLED; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | /* check pending CNAKS */ | 
|  | if (cnak_pending) { | 
|  | /* CNAk processing when rxfifo empty only */ | 
|  | if (readl(&dev->regs->sts) & AMD_BIT(UDC_DEVSTS_RXFIFO_EMPTY)) | 
|  | udc_process_cnak_queue(dev); | 
|  | } | 
|  |  | 
|  | finished: | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /* Interrupt handler for Control IN traffic */ | 
|  | static irqreturn_t udc_control_in_isr(struct udc *dev) | 
|  | { | 
|  | irqreturn_t ret_val = IRQ_NONE; | 
|  | u32 tmp; | 
|  | struct udc_ep *ep; | 
|  | struct udc_request *req; | 
|  | unsigned len; | 
|  |  | 
|  | ep = &dev->ep[UDC_EP0IN_IX]; | 
|  |  | 
|  | /* clear irq */ | 
|  | writel(AMD_BIT(UDC_EPINT_IN_EP0), &dev->regs->ep_irqsts); | 
|  |  | 
|  | tmp = readl(&dev->ep[UDC_EP0IN_IX].regs->sts); | 
|  | /* DMA completion */ | 
|  | if (tmp & AMD_BIT(UDC_EPSTS_TDC)) { | 
|  | VDBG(dev, "isr: TDC clear\n"); | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | /* clear TDC bit */ | 
|  | writel(AMD_BIT(UDC_EPSTS_TDC), | 
|  | &dev->ep[UDC_EP0IN_IX].regs->sts); | 
|  |  | 
|  | /* status reg has IN bit set ? */ | 
|  | } else if (tmp & AMD_BIT(UDC_EPSTS_IN)) { | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | if (ep->dma) { | 
|  | /* clear IN bit */ | 
|  | writel(AMD_BIT(UDC_EPSTS_IN), | 
|  | &dev->ep[UDC_EP0IN_IX].regs->sts); | 
|  | } | 
|  | if (dev->stall_ep0in) { | 
|  | DBG(dev, "stall ep0in\n"); | 
|  | /* halt ep0in */ | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_S); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | } else { | 
|  | if (!list_empty(&ep->queue)) { | 
|  | /* next request */ | 
|  | req = list_entry(ep->queue.next, | 
|  | struct udc_request, queue); | 
|  |  | 
|  | if (ep->dma) { | 
|  | /* write desc pointer */ | 
|  | writel(req->td_phys, &ep->regs->desptr); | 
|  | /* set HOST READY */ | 
|  | req->td_data->status = | 
|  | AMD_ADDBITS( | 
|  | req->td_data->status, | 
|  | UDC_DMA_STP_STS_BS_HOST_READY, | 
|  | UDC_DMA_STP_STS_BS); | 
|  |  | 
|  | /* set poll demand bit */ | 
|  | tmp = | 
|  | readl(&dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_EPCTL_P); | 
|  | writel(tmp, | 
|  | &dev->ep[UDC_EP0IN_IX].regs->ctl); | 
|  |  | 
|  | /* all bytes will be transferred */ | 
|  | req->req.actual = req->req.length; | 
|  |  | 
|  | /* complete req */ | 
|  | complete_req(ep, req, 0); | 
|  |  | 
|  | } else { | 
|  | /* write fifo */ | 
|  | udc_txfifo_write(ep, &req->req); | 
|  |  | 
|  | /* lengh bytes transferred */ | 
|  | len = req->req.length - req->req.actual; | 
|  | if (len > ep->ep.maxpacket) | 
|  | len = ep->ep.maxpacket; | 
|  |  | 
|  | req->req.actual += len; | 
|  | if (req->req.actual == req->req.length | 
|  | || (len != ep->ep.maxpacket)) { | 
|  | /* complete req */ | 
|  | complete_req(ep, req, 0); | 
|  | } | 
|  | } | 
|  |  | 
|  | } | 
|  | } | 
|  | ep->halted = 0; | 
|  | dev->stall_ep0in = 0; | 
|  | if (!ep->dma) { | 
|  | /* clear IN bit */ | 
|  | writel(AMD_BIT(UDC_EPSTS_IN), | 
|  | &dev->ep[UDC_EP0IN_IX].regs->sts); | 
|  | } | 
|  | } | 
|  |  | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  |  | 
|  | /* Interrupt handler for global device events */ | 
|  | static irqreturn_t udc_dev_isr(struct udc *dev, u32 dev_irq) | 
|  | __releases(dev->lock) | 
|  | __acquires(dev->lock) | 
|  | { | 
|  | irqreturn_t ret_val = IRQ_NONE; | 
|  | u32 tmp; | 
|  | u32 cfg; | 
|  | struct udc_ep *ep; | 
|  | u16 i; | 
|  | u8 udc_csr_epix; | 
|  |  | 
|  | /* SET_CONFIG irq ? */ | 
|  | if (dev_irq & AMD_BIT(UDC_DEVINT_SC)) { | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | /* read config value */ | 
|  | tmp = readl(&dev->regs->sts); | 
|  | cfg = AMD_GETBITS(tmp, UDC_DEVSTS_CFG); | 
|  | DBG(dev, "SET_CONFIG interrupt: config=%d\n", cfg); | 
|  | dev->cur_config = cfg; | 
|  | dev->set_cfg_not_acked = 1; | 
|  |  | 
|  | /* make usb request for gadget driver */ | 
|  | memset(&setup_data, 0 , sizeof(union udc_setup_data)); | 
|  | setup_data.request.bRequest = USB_REQ_SET_CONFIGURATION; | 
|  | setup_data.request.wValue = cpu_to_le16(dev->cur_config); | 
|  |  | 
|  | /* programm the NE registers */ | 
|  | for (i = 0; i < UDC_EP_NUM; i++) { | 
|  | ep = &dev->ep[i]; | 
|  | if (ep->in) { | 
|  |  | 
|  | /* ep ix in UDC CSR register space */ | 
|  | udc_csr_epix = ep->num; | 
|  |  | 
|  |  | 
|  | /* OUT ep */ | 
|  | } else { | 
|  | /* ep ix in UDC CSR register space */ | 
|  | udc_csr_epix = ep->num - UDC_CSR_EP_OUT_IX_OFS; | 
|  | } | 
|  |  | 
|  | tmp = readl(&dev->csr->ne[udc_csr_epix]); | 
|  | /* ep cfg */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->dev->cur_config, | 
|  | UDC_CSR_NE_CFG); | 
|  | /* write reg */ | 
|  | writel(tmp, &dev->csr->ne[udc_csr_epix]); | 
|  |  | 
|  | /* clear stall bits */ | 
|  | ep->halted = 0; | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp = tmp & AMD_CLEAR_BIT(UDC_EPCTL_S); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | } | 
|  | /* call gadget zero with setup data received */ | 
|  | spin_unlock(&dev->lock); | 
|  | tmp = dev->driver->setup(&dev->gadget, &setup_data.request); | 
|  | spin_lock(&dev->lock); | 
|  |  | 
|  | } /* SET_INTERFACE ? */ | 
|  | if (dev_irq & AMD_BIT(UDC_DEVINT_SI)) { | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | dev->set_cfg_not_acked = 1; | 
|  | /* read interface and alt setting values */ | 
|  | tmp = readl(&dev->regs->sts); | 
|  | dev->cur_alt = AMD_GETBITS(tmp, UDC_DEVSTS_ALT); | 
|  | dev->cur_intf = AMD_GETBITS(tmp, UDC_DEVSTS_INTF); | 
|  |  | 
|  | /* make usb request for gadget driver */ | 
|  | memset(&setup_data, 0 , sizeof(union udc_setup_data)); | 
|  | setup_data.request.bRequest = USB_REQ_SET_INTERFACE; | 
|  | setup_data.request.bRequestType = USB_RECIP_INTERFACE; | 
|  | setup_data.request.wValue = cpu_to_le16(dev->cur_alt); | 
|  | setup_data.request.wIndex = cpu_to_le16(dev->cur_intf); | 
|  |  | 
|  | DBG(dev, "SET_INTERFACE interrupt: alt=%d intf=%d\n", | 
|  | dev->cur_alt, dev->cur_intf); | 
|  |  | 
|  | /* programm the NE registers */ | 
|  | for (i = 0; i < UDC_EP_NUM; i++) { | 
|  | ep = &dev->ep[i]; | 
|  | if (ep->in) { | 
|  |  | 
|  | /* ep ix in UDC CSR register space */ | 
|  | udc_csr_epix = ep->num; | 
|  |  | 
|  |  | 
|  | /* OUT ep */ | 
|  | } else { | 
|  | /* ep ix in UDC CSR register space */ | 
|  | udc_csr_epix = ep->num - UDC_CSR_EP_OUT_IX_OFS; | 
|  | } | 
|  |  | 
|  | /* UDC CSR reg */ | 
|  | /* set ep values */ | 
|  | tmp = readl(&dev->csr->ne[udc_csr_epix]); | 
|  | /* ep interface */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->dev->cur_intf, | 
|  | UDC_CSR_NE_INTF); | 
|  | /* tmp = AMD_ADDBITS(tmp, 2, UDC_CSR_NE_INTF); */ | 
|  | /* ep alt */ | 
|  | tmp = AMD_ADDBITS(tmp, ep->dev->cur_alt, | 
|  | UDC_CSR_NE_ALT); | 
|  | /* write reg */ | 
|  | writel(tmp, &dev->csr->ne[udc_csr_epix]); | 
|  |  | 
|  | /* clear stall bits */ | 
|  | ep->halted = 0; | 
|  | tmp = readl(&ep->regs->ctl); | 
|  | tmp = tmp & AMD_CLEAR_BIT(UDC_EPCTL_S); | 
|  | writel(tmp, &ep->regs->ctl); | 
|  | } | 
|  |  | 
|  | /* call gadget zero with setup data received */ | 
|  | spin_unlock(&dev->lock); | 
|  | tmp = dev->driver->setup(&dev->gadget, &setup_data.request); | 
|  | spin_lock(&dev->lock); | 
|  |  | 
|  | } /* USB reset */ | 
|  | if (dev_irq & AMD_BIT(UDC_DEVINT_UR)) { | 
|  | DBG(dev, "USB Reset interrupt\n"); | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | /* allow soft reset when suspend occurs */ | 
|  | soft_reset_occured = 0; | 
|  |  | 
|  | dev->waiting_zlp_ack_ep0in = 0; | 
|  | dev->set_cfg_not_acked = 0; | 
|  |  | 
|  | /* mask not needed interrupts */ | 
|  | udc_mask_unused_interrupts(dev); | 
|  |  | 
|  | /* call gadget to resume and reset configs etc. */ | 
|  | spin_unlock(&dev->lock); | 
|  | if (dev->sys_suspended && dev->driver->resume) { | 
|  | dev->driver->resume(&dev->gadget); | 
|  | dev->sys_suspended = 0; | 
|  | } | 
|  | dev->driver->disconnect(&dev->gadget); | 
|  | spin_lock(&dev->lock); | 
|  |  | 
|  | /* disable ep0 to empty req queue */ | 
|  | empty_req_queue(&dev->ep[UDC_EP0IN_IX]); | 
|  | ep_init(dev->regs, &dev->ep[UDC_EP0IN_IX]); | 
|  |  | 
|  | /* soft reset when rxfifo not empty */ | 
|  | tmp = readl(&dev->regs->sts); | 
|  | if (!(tmp & AMD_BIT(UDC_DEVSTS_RXFIFO_EMPTY)) | 
|  | && !soft_reset_after_usbreset_occured) { | 
|  | udc_soft_reset(dev); | 
|  | soft_reset_after_usbreset_occured++; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * DMA reset to kill potential old DMA hw hang, | 
|  | * POLL bit is already reset by ep_init() through | 
|  | * disconnect() | 
|  | */ | 
|  | DBG(dev, "DMA machine reset\n"); | 
|  | tmp = readl(&dev->regs->cfg); | 
|  | writel(tmp | AMD_BIT(UDC_DEVCFG_DMARST), &dev->regs->cfg); | 
|  | writel(tmp, &dev->regs->cfg); | 
|  |  | 
|  | /* put into initial config */ | 
|  | udc_basic_init(dev); | 
|  |  | 
|  | /* enable device setup interrupts */ | 
|  | udc_enable_dev_setup_interrupts(dev); | 
|  |  | 
|  | /* enable suspend interrupt */ | 
|  | tmp = readl(&dev->regs->irqmsk); | 
|  | tmp &= AMD_UNMASK_BIT(UDC_DEVINT_US); | 
|  | writel(tmp, &dev->regs->irqmsk); | 
|  |  | 
|  | } /* USB suspend */ | 
|  | if (dev_irq & AMD_BIT(UDC_DEVINT_US)) { | 
|  | DBG(dev, "USB Suspend interrupt\n"); | 
|  | ret_val = IRQ_HANDLED; | 
|  | if (dev->driver->suspend) { | 
|  | spin_unlock(&dev->lock); | 
|  | dev->sys_suspended = 1; | 
|  | dev->driver->suspend(&dev->gadget); | 
|  | spin_lock(&dev->lock); | 
|  | } | 
|  | } /* new speed ? */ | 
|  | if (dev_irq & AMD_BIT(UDC_DEVINT_ENUM)) { | 
|  | DBG(dev, "ENUM interrupt\n"); | 
|  | ret_val = IRQ_HANDLED; | 
|  | soft_reset_after_usbreset_occured = 0; | 
|  |  | 
|  | /* disable ep0 to empty req queue */ | 
|  | empty_req_queue(&dev->ep[UDC_EP0IN_IX]); | 
|  | ep_init(dev->regs, &dev->ep[UDC_EP0IN_IX]); | 
|  |  | 
|  | /* link up all endpoints */ | 
|  | udc_setup_endpoints(dev); | 
|  | dev_info(&dev->pdev->dev, "Connect: %s\n", | 
|  | usb_speed_string(dev->gadget.speed)); | 
|  |  | 
|  | /* init ep 0 */ | 
|  | activate_control_endpoints(dev); | 
|  |  | 
|  | /* enable ep0 interrupts */ | 
|  | udc_enable_ep0_interrupts(dev); | 
|  | } | 
|  | /* session valid change interrupt */ | 
|  | if (dev_irq & AMD_BIT(UDC_DEVINT_SVC)) { | 
|  | DBG(dev, "USB SVC interrupt\n"); | 
|  | ret_val = IRQ_HANDLED; | 
|  |  | 
|  | /* check that session is not valid to detect disconnect */ | 
|  | tmp = readl(&dev->regs->sts); | 
|  | if (!(tmp & AMD_BIT(UDC_DEVSTS_SESSVLD))) { | 
|  | /* disable suspend interrupt */ | 
|  | tmp = readl(&dev->regs->irqmsk); | 
|  | tmp |= AMD_BIT(UDC_DEVINT_US); | 
|  | writel(tmp, &dev->regs->irqmsk); | 
|  | DBG(dev, "USB Disconnect (session valid low)\n"); | 
|  | /* cleanup on disconnect */ | 
|  | usb_disconnect(udc); | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /* Interrupt Service Routine, see Linux Kernel Doc for parameters */ | 
|  | static irqreturn_t udc_irq(int irq, void *pdev) | 
|  | { | 
|  | struct udc *dev = pdev; | 
|  | u32 reg; | 
|  | u16 i; | 
|  | u32 ep_irq; | 
|  | irqreturn_t ret_val = IRQ_NONE; | 
|  |  | 
|  | spin_lock(&dev->lock); | 
|  |  | 
|  | /* check for ep irq */ | 
|  | reg = readl(&dev->regs->ep_irqsts); | 
|  | if (reg) { | 
|  | if (reg & AMD_BIT(UDC_EPINT_OUT_EP0)) | 
|  | ret_val |= udc_control_out_isr(dev); | 
|  | if (reg & AMD_BIT(UDC_EPINT_IN_EP0)) | 
|  | ret_val |= udc_control_in_isr(dev); | 
|  |  | 
|  | /* | 
|  | * data endpoint | 
|  | * iterate ep's | 
|  | */ | 
|  | for (i = 1; i < UDC_EP_NUM; i++) { | 
|  | ep_irq = 1 << i; | 
|  | if (!(reg & ep_irq) || i == UDC_EPINT_OUT_EP0) | 
|  | continue; | 
|  |  | 
|  | /* clear irq status */ | 
|  | writel(ep_irq, &dev->regs->ep_irqsts); | 
|  |  | 
|  | /* irq for out ep ? */ | 
|  | if (i > UDC_EPIN_NUM) | 
|  | ret_val |= udc_data_out_isr(dev, i); | 
|  | else | 
|  | ret_val |= udc_data_in_isr(dev, i); | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  |  | 
|  | /* check for dev irq */ | 
|  | reg = readl(&dev->regs->irqsts); | 
|  | if (reg) { | 
|  | /* clear irq */ | 
|  | writel(reg, &dev->regs->irqsts); | 
|  | ret_val |= udc_dev_isr(dev, reg); | 
|  | } | 
|  |  | 
|  |  | 
|  | spin_unlock(&dev->lock); | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /* Tears down device */ | 
|  | static void gadget_release(struct device *pdev) | 
|  | { | 
|  | struct amd5536udc *dev = dev_get_drvdata(pdev); | 
|  | kfree(dev); | 
|  | } | 
|  |  | 
|  | /* Cleanup on device remove */ | 
|  | static void udc_remove(struct udc *dev) | 
|  | { | 
|  | /* remove timer */ | 
|  | stop_timer++; | 
|  | if (timer_pending(&udc_timer)) | 
|  | wait_for_completion(&on_exit); | 
|  | if (udc_timer.data) | 
|  | del_timer_sync(&udc_timer); | 
|  | /* remove pollstall timer */ | 
|  | stop_pollstall_timer++; | 
|  | if (timer_pending(&udc_pollstall_timer)) | 
|  | wait_for_completion(&on_pollstall_exit); | 
|  | if (udc_pollstall_timer.data) | 
|  | del_timer_sync(&udc_pollstall_timer); | 
|  | udc = NULL; | 
|  | } | 
|  |  | 
|  | /* Reset all pci context */ | 
|  | static void udc_pci_remove(struct pci_dev *pdev) | 
|  | { | 
|  | struct udc		*dev; | 
|  |  | 
|  | dev = pci_get_drvdata(pdev); | 
|  |  | 
|  | usb_del_gadget_udc(&udc->gadget); | 
|  | /* gadget driver must not be registered */ | 
|  | BUG_ON(dev->driver != NULL); | 
|  |  | 
|  | /* dma pool cleanup */ | 
|  | if (dev->data_requests) | 
|  | pci_pool_destroy(dev->data_requests); | 
|  |  | 
|  | if (dev->stp_requests) { | 
|  | /* cleanup DMA desc's for ep0in */ | 
|  | pci_pool_free(dev->stp_requests, | 
|  | dev->ep[UDC_EP0OUT_IX].td_stp, | 
|  | dev->ep[UDC_EP0OUT_IX].td_stp_dma); | 
|  | pci_pool_free(dev->stp_requests, | 
|  | dev->ep[UDC_EP0OUT_IX].td, | 
|  | dev->ep[UDC_EP0OUT_IX].td_phys); | 
|  |  | 
|  | pci_pool_destroy(dev->stp_requests); | 
|  | } | 
|  |  | 
|  | /* reset controller */ | 
|  | writel(AMD_BIT(UDC_DEVCFG_SOFTRESET), &dev->regs->cfg); | 
|  | if (dev->irq_registered) | 
|  | free_irq(pdev->irq, dev); | 
|  | if (dev->regs) | 
|  | iounmap(dev->regs); | 
|  | if (dev->mem_region) | 
|  | release_mem_region(pci_resource_start(pdev, 0), | 
|  | pci_resource_len(pdev, 0)); | 
|  | if (dev->active) | 
|  | pci_disable_device(pdev); | 
|  |  | 
|  | device_unregister(&dev->gadget.dev); | 
|  | pci_set_drvdata(pdev, NULL); | 
|  |  | 
|  | udc_remove(dev); | 
|  | } | 
|  |  | 
|  | /* create dma pools on init */ | 
|  | static int init_dma_pools(struct udc *dev) | 
|  | { | 
|  | struct udc_stp_dma	*td_stp; | 
|  | struct udc_data_dma	*td_data; | 
|  | int retval; | 
|  |  | 
|  | /* consistent DMA mode setting ? */ | 
|  | if (use_dma_ppb) { | 
|  | use_dma_bufferfill_mode = 0; | 
|  | } else { | 
|  | use_dma_ppb_du = 0; | 
|  | use_dma_bufferfill_mode = 1; | 
|  | } | 
|  |  | 
|  | /* DMA setup */ | 
|  | dev->data_requests = dma_pool_create("data_requests", NULL, | 
|  | sizeof(struct udc_data_dma), 0, 0); | 
|  | if (!dev->data_requests) { | 
|  | DBG(dev, "can't get request data pool\n"); | 
|  | retval = -ENOMEM; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | /* EP0 in dma regs = dev control regs */ | 
|  | dev->ep[UDC_EP0IN_IX].dma = &dev->regs->ctl; | 
|  |  | 
|  | /* dma desc for setup data */ | 
|  | dev->stp_requests = dma_pool_create("setup requests", NULL, | 
|  | sizeof(struct udc_stp_dma), 0, 0); | 
|  | if (!dev->stp_requests) { | 
|  | DBG(dev, "can't get stp request pool\n"); | 
|  | retval = -ENOMEM; | 
|  | goto finished; | 
|  | } | 
|  | /* setup */ | 
|  | td_stp = dma_pool_alloc(dev->stp_requests, GFP_KERNEL, | 
|  | &dev->ep[UDC_EP0OUT_IX].td_stp_dma); | 
|  | if (td_stp == NULL) { | 
|  | retval = -ENOMEM; | 
|  | goto finished; | 
|  | } | 
|  | dev->ep[UDC_EP0OUT_IX].td_stp = td_stp; | 
|  |  | 
|  | /* data: 0 packets !? */ | 
|  | td_data = dma_pool_alloc(dev->stp_requests, GFP_KERNEL, | 
|  | &dev->ep[UDC_EP0OUT_IX].td_phys); | 
|  | if (td_data == NULL) { | 
|  | retval = -ENOMEM; | 
|  | goto finished; | 
|  | } | 
|  | dev->ep[UDC_EP0OUT_IX].td = td_data; | 
|  | return 0; | 
|  |  | 
|  | finished: | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* Called by pci bus driver to init pci context */ | 
|  | static int udc_pci_probe( | 
|  | struct pci_dev *pdev, | 
|  | const struct pci_device_id *id | 
|  | ) | 
|  | { | 
|  | struct udc		*dev; | 
|  | unsigned long		resource; | 
|  | unsigned long		len; | 
|  | int			retval = 0; | 
|  |  | 
|  | /* one udc only */ | 
|  | if (udc) { | 
|  | dev_dbg(&pdev->dev, "already probed\n"); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | /* init */ | 
|  | dev = kzalloc(sizeof(struct udc), GFP_KERNEL); | 
|  | if (!dev) { | 
|  | retval = -ENOMEM; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | /* pci setup */ | 
|  | if (pci_enable_device(pdev) < 0) { | 
|  | kfree(dev); | 
|  | dev = NULL; | 
|  | retval = -ENODEV; | 
|  | goto finished; | 
|  | } | 
|  | dev->active = 1; | 
|  |  | 
|  | /* PCI resource allocation */ | 
|  | resource = pci_resource_start(pdev, 0); | 
|  | len = pci_resource_len(pdev, 0); | 
|  |  | 
|  | if (!request_mem_region(resource, len, name)) { | 
|  | dev_dbg(&pdev->dev, "pci device used already\n"); | 
|  | kfree(dev); | 
|  | dev = NULL; | 
|  | retval = -EBUSY; | 
|  | goto finished; | 
|  | } | 
|  | dev->mem_region = 1; | 
|  |  | 
|  | dev->virt_addr = ioremap_nocache(resource, len); | 
|  | if (dev->virt_addr == NULL) { | 
|  | dev_dbg(&pdev->dev, "start address cannot be mapped\n"); | 
|  | kfree(dev); | 
|  | dev = NULL; | 
|  | retval = -EFAULT; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | if (!pdev->irq) { | 
|  | dev_err(&dev->pdev->dev, "irq not set\n"); | 
|  | kfree(dev); | 
|  | dev = NULL; | 
|  | retval = -ENODEV; | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | spin_lock_init(&dev->lock); | 
|  | /* udc csr registers base */ | 
|  | dev->csr = dev->virt_addr + UDC_CSR_ADDR; | 
|  | /* dev registers base */ | 
|  | dev->regs = dev->virt_addr + UDC_DEVCFG_ADDR; | 
|  | /* ep registers base */ | 
|  | dev->ep_regs = dev->virt_addr + UDC_EPREGS_ADDR; | 
|  | /* fifo's base */ | 
|  | dev->rxfifo = (u32 __iomem *)(dev->virt_addr + UDC_RXFIFO_ADDR); | 
|  | dev->txfifo = (u32 __iomem *)(dev->virt_addr + UDC_TXFIFO_ADDR); | 
|  |  | 
|  | if (request_irq(pdev->irq, udc_irq, IRQF_SHARED, name, dev) != 0) { | 
|  | dev_dbg(&dev->pdev->dev, "request_irq(%d) fail\n", pdev->irq); | 
|  | kfree(dev); | 
|  | dev = NULL; | 
|  | retval = -EBUSY; | 
|  | goto finished; | 
|  | } | 
|  | dev->irq_registered = 1; | 
|  |  | 
|  | pci_set_drvdata(pdev, dev); | 
|  |  | 
|  | /* chip revision for Hs AMD5536 */ | 
|  | dev->chiprev = pdev->revision; | 
|  |  | 
|  | pci_set_master(pdev); | 
|  | pci_try_set_mwi(pdev); | 
|  |  | 
|  | /* init dma pools */ | 
|  | if (use_dma) { | 
|  | retval = init_dma_pools(dev); | 
|  | if (retval != 0) | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | dev->phys_addr = resource; | 
|  | dev->irq = pdev->irq; | 
|  | dev->pdev = pdev; | 
|  | dev->gadget.dev.parent = &pdev->dev; | 
|  | dev->gadget.dev.dma_mask = pdev->dev.dma_mask; | 
|  |  | 
|  | /* general probing */ | 
|  | if (udc_probe(dev) == 0) | 
|  | return 0; | 
|  |  | 
|  | finished: | 
|  | if (dev) | 
|  | udc_pci_remove(pdev); | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* general probe */ | 
|  | static int udc_probe(struct udc *dev) | 
|  | { | 
|  | char		tmp[128]; | 
|  | u32		reg; | 
|  | int		retval; | 
|  |  | 
|  | /* mark timer as not initialized */ | 
|  | udc_timer.data = 0; | 
|  | udc_pollstall_timer.data = 0; | 
|  |  | 
|  | /* device struct setup */ | 
|  | dev->gadget.ops = &udc_ops; | 
|  |  | 
|  | dev_set_name(&dev->gadget.dev, "gadget"); | 
|  | dev->gadget.dev.release = gadget_release; | 
|  | dev->gadget.name = name; | 
|  | dev->gadget.max_speed = USB_SPEED_HIGH; | 
|  |  | 
|  | /* init registers, interrupts, ... */ | 
|  | startup_registers(dev); | 
|  |  | 
|  | dev_info(&dev->pdev->dev, "%s\n", mod_desc); | 
|  |  | 
|  | snprintf(tmp, sizeof tmp, "%d", dev->irq); | 
|  | dev_info(&dev->pdev->dev, | 
|  | "irq %s, pci mem %08lx, chip rev %02x(Geode5536 %s)\n", | 
|  | tmp, dev->phys_addr, dev->chiprev, | 
|  | (dev->chiprev == UDC_HSA0_REV) ? "A0" : "B1"); | 
|  | strcpy(tmp, UDC_DRIVER_VERSION_STRING); | 
|  | if (dev->chiprev == UDC_HSA0_REV) { | 
|  | dev_err(&dev->pdev->dev, "chip revision is A0; too old\n"); | 
|  | retval = -ENODEV; | 
|  | goto finished; | 
|  | } | 
|  | dev_info(&dev->pdev->dev, | 
|  | "driver version: %s(for Geode5536 B1)\n", tmp); | 
|  | udc = dev; | 
|  |  | 
|  | retval = usb_add_gadget_udc(&udc->pdev->dev, &dev->gadget); | 
|  | if (retval) | 
|  | goto finished; | 
|  |  | 
|  | retval = device_register(&dev->gadget.dev); | 
|  | if (retval) { | 
|  | usb_del_gadget_udc(&dev->gadget); | 
|  | put_device(&dev->gadget.dev); | 
|  | goto finished; | 
|  | } | 
|  |  | 
|  | /* timer init */ | 
|  | init_timer(&udc_timer); | 
|  | udc_timer.function = udc_timer_function; | 
|  | udc_timer.data = 1; | 
|  | /* timer pollstall init */ | 
|  | init_timer(&udc_pollstall_timer); | 
|  | udc_pollstall_timer.function = udc_pollstall_timer_function; | 
|  | udc_pollstall_timer.data = 1; | 
|  |  | 
|  | /* set SD */ | 
|  | reg = readl(&dev->regs->ctl); | 
|  | reg |= AMD_BIT(UDC_DEVCTL_SD); | 
|  | writel(reg, &dev->regs->ctl); | 
|  |  | 
|  | /* print dev register info */ | 
|  | print_regs(dev); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | finished: | 
|  | return retval; | 
|  | } | 
|  |  | 
|  | /* Initiates a remote wakeup */ | 
|  | static int udc_remote_wakeup(struct udc *dev) | 
|  | { | 
|  | unsigned long flags; | 
|  | u32 tmp; | 
|  |  | 
|  | DBG(dev, "UDC initiates remote wakeup\n"); | 
|  |  | 
|  | spin_lock_irqsave(&dev->lock, flags); | 
|  |  | 
|  | tmp = readl(&dev->regs->ctl); | 
|  | tmp |= AMD_BIT(UDC_DEVCTL_RES); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  | tmp &= AMD_CLEAR_BIT(UDC_DEVCTL_RES); | 
|  | writel(tmp, &dev->regs->ctl); | 
|  |  | 
|  | spin_unlock_irqrestore(&dev->lock, flags); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* PCI device parameters */ | 
|  | static DEFINE_PCI_DEVICE_TABLE(pci_id) = { | 
|  | { | 
|  | PCI_DEVICE(PCI_VENDOR_ID_AMD, 0x2096), | 
|  | .class =	(PCI_CLASS_SERIAL_USB << 8) | 0xfe, | 
|  | .class_mask =	0xffffffff, | 
|  | }, | 
|  | {}, | 
|  | }; | 
|  | MODULE_DEVICE_TABLE(pci, pci_id); | 
|  |  | 
|  | /* PCI functions */ | 
|  | static struct pci_driver udc_pci_driver = { | 
|  | .name =		(char *) name, | 
|  | .id_table =	pci_id, | 
|  | .probe =	udc_pci_probe, | 
|  | .remove =	udc_pci_remove, | 
|  | }; | 
|  |  | 
|  | module_pci_driver(udc_pci_driver); | 
|  |  | 
|  | MODULE_DESCRIPTION(UDC_MOD_DESCRIPTION); | 
|  | MODULE_AUTHOR("Thomas Dahlmann"); | 
|  | MODULE_LICENSE("GPL"); | 
|  |  |