Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Martin Fuzzey | 9485 | 98.38% | 4 | 12.50% |
Alan Stern | 72 | 0.75% | 2 | 6.25% |
Gustavo A. R. Silva | 15 | 0.16% | 1 | 3.12% |
Peter Chen | 9 | 0.09% | 1 | 3.12% |
Oliver Neukum | 8 | 0.08% | 1 | 3.12% |
John Youn | 8 | 0.08% | 1 | 3.12% |
Jingoo Han | 7 | 0.07% | 2 | 6.25% |
Fabio Estevam | 6 | 0.06% | 2 | 6.25% |
Axel Lin | 5 | 0.05% | 2 | 6.25% |
Kees Cook | 5 | 0.05% | 1 | 3.12% |
Sergei Shtylyov | 4 | 0.04% | 2 | 6.25% |
Tejun Heo | 3 | 0.03% | 1 | 3.12% |
Robert Morell | 2 | 0.02% | 1 | 3.12% |
Greg Kroah-Hartman | 2 | 0.02% | 2 | 6.25% |
Sage Sharp | 2 | 0.02% | 1 | 3.12% |
Yong Zhang | 1 | 0.01% | 1 | 3.12% |
Lucas De Marchi | 1 | 0.01% | 1 | 3.12% |
Uwe Kleine-König | 1 | 0.01% | 1 | 3.12% |
Eric Lescouet | 1 | 0.01% | 1 | 3.12% |
Vasiliy Kulikov | 1 | 0.01% | 1 | 3.12% |
Stefan Weil | 1 | 0.01% | 1 | 3.12% |
Bhumika Goyal | 1 | 0.01% | 1 | 3.12% |
Justin P. Mattock | 1 | 0.01% | 1 | 3.12% |
Total | 9641 | 32 |
// SPDX-License-Identifier: GPL-2.0+ /* * USB Host Controller Driver for IMX21 * * Copyright (C) 2006 Loping Dog Embedded Systems * Copyright (C) 2009 Martin Fuzzey * Originally written by Jay Monkman <jtm@lopingdog.com> * Ported to 2.6.30, debugged and enhanced by Martin Fuzzey */ /* * The i.MX21 USB hardware contains * * 32 transfer descriptors (called ETDs) * * 4Kb of Data memory * * The data memory is shared between the host and function controllers * (but this driver only supports the host controller) * * So setting up a transfer involves: * * Allocating a ETD * * Fill in ETD with appropriate information * * Allocating data memory (and putting the offset in the ETD) * * Activate the ETD * * Get interrupt when done. * * An ETD is assigned to each active endpoint. * * Low resource (ETD and Data memory) situations are handled differently for * isochronous and non insosynchronous transactions : * * Non ISOC transfers are queued if either ETDs or Data memory are unavailable * * ISOC transfers use 2 ETDs per endpoint to achieve double buffering. * They allocate both ETDs and Data memory during URB submission * (and fail if unavailable). */ #include <linux/clk.h> #include <linux/io.h> #include <linux/kernel.h> #include <linux/list.h> #include <linux/platform_device.h> #include <linux/slab.h> #include <linux/usb.h> #include <linux/usb/hcd.h> #include <linux/dma-mapping.h> #include <linux/module.h> #include "imx21-hcd.h" #ifdef CONFIG_DYNAMIC_DEBUG #define DEBUG #endif #ifdef DEBUG #define DEBUG_LOG_FRAME(imx21, etd, event) \ (etd)->event##_frame = readl((imx21)->regs + USBH_FRMNUB) #else #define DEBUG_LOG_FRAME(imx21, etd, event) do { } while (0) #endif static const char hcd_name[] = "imx21-hcd"; static inline struct imx21 *hcd_to_imx21(struct usb_hcd *hcd) { return (struct imx21 *)hcd->hcd_priv; } /* =========================================== */ /* Hardware access helpers */ /* =========================================== */ static inline void set_register_bits(struct imx21 *imx21, u32 offset, u32 mask) { void __iomem *reg = imx21->regs + offset; writel(readl(reg) | mask, reg); } static inline void clear_register_bits(struct imx21 *imx21, u32 offset, u32 mask) { void __iomem *reg = imx21->regs + offset; writel(readl(reg) & ~mask, reg); } static inline void clear_toggle_bit(struct imx21 *imx21, u32 offset, u32 mask) { void __iomem *reg = imx21->regs + offset; if (readl(reg) & mask) writel(mask, reg); } static inline void set_toggle_bit(struct imx21 *imx21, u32 offset, u32 mask) { void __iomem *reg = imx21->regs + offset; if (!(readl(reg) & mask)) writel(mask, reg); } static void etd_writel(struct imx21 *imx21, int etd_num, int dword, u32 value) { writel(value, imx21->regs + USB_ETD_DWORD(etd_num, dword)); } static u32 etd_readl(struct imx21 *imx21, int etd_num, int dword) { return readl(imx21->regs + USB_ETD_DWORD(etd_num, dword)); } static inline int wrap_frame(int counter) { return counter & 0xFFFF; } static inline int frame_after(int frame, int after) { /* handle wrapping like jiffies time_afer */ return (s16)((s16)after - (s16)frame) < 0; } static int imx21_hc_get_frame(struct usb_hcd *hcd) { struct imx21 *imx21 = hcd_to_imx21(hcd); return wrap_frame(readl(imx21->regs + USBH_FRMNUB)); } static inline bool unsuitable_for_dma(dma_addr_t addr) { return (addr & 3) != 0; } #include "imx21-dbg.c" static void nonisoc_urb_completed_for_etd( struct imx21 *imx21, struct etd_priv *etd, int status); static void schedule_nonisoc_etd(struct imx21 *imx21, struct urb *urb); static void free_dmem(struct imx21 *imx21, struct etd_priv *etd); /* =========================================== */ /* ETD management */ /* =========================================== */ static int alloc_etd(struct imx21 *imx21) { int i; struct etd_priv *etd = imx21->etd; for (i = 0; i < USB_NUM_ETD; i++, etd++) { if (etd->alloc == 0) { memset(etd, 0, sizeof(imx21->etd[0])); etd->alloc = 1; debug_etd_allocated(imx21); return i; } } return -1; } static void disactivate_etd(struct imx21 *imx21, int num) { int etd_mask = (1 << num); struct etd_priv *etd = &imx21->etd[num]; writel(etd_mask, imx21->regs + USBH_ETDENCLR); clear_register_bits(imx21, USBH_ETDDONEEN, etd_mask); writel(etd_mask, imx21->regs + USB_ETDDMACHANLCLR); clear_toggle_bit(imx21, USBH_ETDDONESTAT, etd_mask); etd->active_count = 0; DEBUG_LOG_FRAME(imx21, etd, disactivated); } static void reset_etd(struct imx21 *imx21, int num) { struct etd_priv *etd = imx21->etd + num; int i; disactivate_etd(imx21, num); for (i = 0; i < 4; i++) etd_writel(imx21, num, i, 0); etd->urb = NULL; etd->ep = NULL; etd->td = NULL; etd->bounce_buffer = NULL; } static void free_etd(struct imx21 *imx21, int num) { if (num < 0) return; if (num >= USB_NUM_ETD) { dev_err(imx21->dev, "BAD etd=%d!\n", num); return; } if (imx21->etd[num].alloc == 0) { dev_err(imx21->dev, "ETD %d already free!\n", num); return; } debug_etd_freed(imx21); reset_etd(imx21, num); memset(&imx21->etd[num], 0, sizeof(imx21->etd[0])); } static void setup_etd_dword0(struct imx21 *imx21, int etd_num, struct urb *urb, u8 dir, u16 maxpacket) { etd_writel(imx21, etd_num, 0, ((u32) usb_pipedevice(urb->pipe)) << DW0_ADDRESS | ((u32) usb_pipeendpoint(urb->pipe) << DW0_ENDPNT) | ((u32) dir << DW0_DIRECT) | ((u32) ((urb->dev->speed == USB_SPEED_LOW) ? 1 : 0) << DW0_SPEED) | ((u32) fmt_urb_to_etd[usb_pipetype(urb->pipe)] << DW0_FORMAT) | ((u32) maxpacket << DW0_MAXPKTSIZ)); } /** * Copy buffer to data controller data memory. * We cannot use memcpy_toio() because the hardware requires 32bit writes */ static void copy_to_dmem( struct imx21 *imx21, int dmem_offset, void *src, int count) { void __iomem *dmem = imx21->regs + USBOTG_DMEM + dmem_offset; u32 word = 0; u8 *p = src; int byte = 0; int i; for (i = 0; i < count; i++) { byte = i % 4; word += (*p++ << (byte * 8)); if (byte == 3) { writel(word, dmem); dmem += 4; word = 0; } } if (count && byte != 3) writel(word, dmem); } static void activate_etd(struct imx21 *imx21, int etd_num, u8 dir) { u32 etd_mask = 1 << etd_num; struct etd_priv *etd = &imx21->etd[etd_num]; if (etd->dma_handle && unsuitable_for_dma(etd->dma_handle)) { /* For non aligned isoc the condition below is always true */ if (etd->len <= etd->dmem_size) { /* Fits into data memory, use PIO */ if (dir != TD_DIR_IN) { copy_to_dmem(imx21, etd->dmem_offset, etd->cpu_buffer, etd->len); } etd->dma_handle = 0; } else { /* Too big for data memory, use bounce buffer */ enum dma_data_direction dmadir; if (dir == TD_DIR_IN) { dmadir = DMA_FROM_DEVICE; etd->bounce_buffer = kmalloc(etd->len, GFP_ATOMIC); } else { dmadir = DMA_TO_DEVICE; etd->bounce_buffer = kmemdup(etd->cpu_buffer, etd->len, GFP_ATOMIC); } if (!etd->bounce_buffer) { dev_err(imx21->dev, "failed bounce alloc\n"); goto err_bounce_alloc; } etd->dma_handle = dma_map_single(imx21->dev, etd->bounce_buffer, etd->len, dmadir); if (dma_mapping_error(imx21->dev, etd->dma_handle)) { dev_err(imx21->dev, "failed bounce map\n"); goto err_bounce_map; } } } clear_toggle_bit(imx21, USBH_ETDDONESTAT, etd_mask); set_register_bits(imx21, USBH_ETDDONEEN, etd_mask); clear_toggle_bit(imx21, USBH_XFILLSTAT, etd_mask); clear_toggle_bit(imx21, USBH_YFILLSTAT, etd_mask); if (etd->dma_handle) { set_register_bits(imx21, USB_ETDDMACHANLCLR, etd_mask); clear_toggle_bit(imx21, USBH_XBUFSTAT, etd_mask); clear_toggle_bit(imx21, USBH_YBUFSTAT, etd_mask); writel(etd->dma_handle, imx21->regs + USB_ETDSMSA(etd_num)); set_register_bits(imx21, USB_ETDDMAEN, etd_mask); } else { if (dir != TD_DIR_IN) { /* need to set for ZLP and PIO */ set_toggle_bit(imx21, USBH_XFILLSTAT, etd_mask); set_toggle_bit(imx21, USBH_YFILLSTAT, etd_mask); } } DEBUG_LOG_FRAME(imx21, etd, activated); #ifdef DEBUG if (!etd->active_count) { int i; etd->activated_frame = readl(imx21->regs + USBH_FRMNUB); etd->disactivated_frame = -1; etd->last_int_frame = -1; etd->last_req_frame = -1; for (i = 0; i < 4; i++) etd->submitted_dwords[i] = etd_readl(imx21, etd_num, i); } #endif etd->active_count = 1; writel(etd_mask, imx21->regs + USBH_ETDENSET); return; err_bounce_map: kfree(etd->bounce_buffer); err_bounce_alloc: free_dmem(imx21, etd); nonisoc_urb_completed_for_etd(imx21, etd, -ENOMEM); } /* =========================================== */ /* Data memory management */ /* =========================================== */ static int alloc_dmem(struct imx21 *imx21, unsigned int size, struct usb_host_endpoint *ep) { unsigned int offset = 0; struct imx21_dmem_area *area; struct imx21_dmem_area *tmp; size += (~size + 1) & 0x3; /* Round to 4 byte multiple */ if (size > DMEM_SIZE) { dev_err(imx21->dev, "size=%d > DMEM_SIZE(%d)\n", size, DMEM_SIZE); return -EINVAL; } list_for_each_entry(tmp, &imx21->dmem_list, list) { if ((size + offset) < offset) goto fail; if ((size + offset) <= tmp->offset) break; offset = tmp->size + tmp->offset; if ((offset + size) > DMEM_SIZE) goto fail; } area = kmalloc(sizeof(struct imx21_dmem_area), GFP_ATOMIC); if (area == NULL) return -ENOMEM; area->ep = ep; area->offset = offset; area->size = size; list_add_tail(&area->list, &tmp->list); debug_dmem_allocated(imx21, size); return offset; fail: return -ENOMEM; } /* Memory now available for a queued ETD - activate it */ static void activate_queued_etd(struct imx21 *imx21, struct etd_priv *etd, u32 dmem_offset) { struct urb_priv *urb_priv = etd->urb->hcpriv; int etd_num = etd - &imx21->etd[0]; u32 maxpacket = etd_readl(imx21, etd_num, 1) >> DW1_YBUFSRTAD; u8 dir = (etd_readl(imx21, etd_num, 2) >> DW2_DIRPID) & 0x03; dev_dbg(imx21->dev, "activating queued ETD %d now DMEM available\n", etd_num); etd_writel(imx21, etd_num, 1, ((dmem_offset + maxpacket) << DW1_YBUFSRTAD) | dmem_offset); etd->dmem_offset = dmem_offset; urb_priv->active = 1; activate_etd(imx21, etd_num, dir); } static void free_dmem(struct imx21 *imx21, struct etd_priv *etd) { struct imx21_dmem_area *area; struct etd_priv *tmp; int found = 0; int offset; if (!etd->dmem_size) return; etd->dmem_size = 0; offset = etd->dmem_offset; list_for_each_entry(area, &imx21->dmem_list, list) { if (area->offset == offset) { debug_dmem_freed(imx21, area->size); list_del(&area->list); kfree(area); found = 1; break; } } if (!found) { dev_err(imx21->dev, "Trying to free unallocated DMEM %d\n", offset); return; } /* Try again to allocate memory for anything we've queued */ list_for_each_entry_safe(etd, tmp, &imx21->queue_for_dmem, queue) { offset = alloc_dmem(imx21, etd->dmem_size, etd->ep); if (offset >= 0) { list_del(&etd->queue); activate_queued_etd(imx21, etd, (u32)offset); } } } static void free_epdmem(struct imx21 *imx21, struct usb_host_endpoint *ep) { struct imx21_dmem_area *area, *tmp; list_for_each_entry_safe(area, tmp, &imx21->dmem_list, list) { if (area->ep == ep) { dev_err(imx21->dev, "Active DMEM %d for disabled ep=%p\n", area->offset, ep); list_del(&area->list); kfree(area); } } } /* =========================================== */ /* End handling */ /* =========================================== */ /* Endpoint now idle - release its ETD(s) or assign to queued request */ static void ep_idle(struct imx21 *imx21, struct ep_priv *ep_priv) { int i; for (i = 0; i < NUM_ISO_ETDS; i++) { int etd_num = ep_priv->etd[i]; struct etd_priv *etd; if (etd_num < 0) continue; etd = &imx21->etd[etd_num]; ep_priv->etd[i] = -1; free_dmem(imx21, etd); /* for isoc */ if (list_empty(&imx21->queue_for_etd)) { free_etd(imx21, etd_num); continue; } dev_dbg(imx21->dev, "assigning idle etd %d for queued request\n", etd_num); ep_priv = list_first_entry(&imx21->queue_for_etd, struct ep_priv, queue); list_del(&ep_priv->queue); reset_etd(imx21, etd_num); ep_priv->waiting_etd = 0; ep_priv->etd[i] = etd_num; if (list_empty(&ep_priv->ep->urb_list)) { dev_err(imx21->dev, "No urb for queued ep!\n"); continue; } schedule_nonisoc_etd(imx21, list_first_entry( &ep_priv->ep->urb_list, struct urb, urb_list)); } } static void urb_done(struct usb_hcd *hcd, struct urb *urb, int status) __releases(imx21->lock) __acquires(imx21->lock) { struct imx21 *imx21 = hcd_to_imx21(hcd); struct ep_priv *ep_priv = urb->ep->hcpriv; struct urb_priv *urb_priv = urb->hcpriv; debug_urb_completed(imx21, urb, status); dev_vdbg(imx21->dev, "urb %p done %d\n", urb, status); kfree(urb_priv->isoc_td); kfree(urb->hcpriv); urb->hcpriv = NULL; usb_hcd_unlink_urb_from_ep(hcd, urb); spin_unlock(&imx21->lock); usb_hcd_giveback_urb(hcd, urb, status); spin_lock(&imx21->lock); if (list_empty(&ep_priv->ep->urb_list)) ep_idle(imx21, ep_priv); } static void nonisoc_urb_completed_for_etd( struct imx21 *imx21, struct etd_priv *etd, int status) { struct usb_host_endpoint *ep = etd->ep; urb_done(imx21->hcd, etd->urb, status); etd->urb = NULL; if (!list_empty(&ep->urb_list)) { struct urb *urb = list_first_entry( &ep->urb_list, struct urb, urb_list); dev_vdbg(imx21->dev, "next URB %p\n", urb); schedule_nonisoc_etd(imx21, urb); } } /* =========================================== */ /* ISOC Handling ... */ /* =========================================== */ static void schedule_isoc_etds(struct usb_hcd *hcd, struct usb_host_endpoint *ep) { struct imx21 *imx21 = hcd_to_imx21(hcd); struct ep_priv *ep_priv = ep->hcpriv; struct etd_priv *etd; struct urb_priv *urb_priv; struct td *td; int etd_num; int i; int cur_frame; u8 dir; for (i = 0; i < NUM_ISO_ETDS; i++) { too_late: if (list_empty(&ep_priv->td_list)) break; etd_num = ep_priv->etd[i]; if (etd_num < 0) break; etd = &imx21->etd[etd_num]; if (etd->urb) continue; td = list_entry(ep_priv->td_list.next, struct td, list); list_del(&td->list); urb_priv = td->urb->hcpriv; cur_frame = imx21_hc_get_frame(hcd); if (frame_after(cur_frame, td->frame)) { dev_dbg(imx21->dev, "isoc too late frame %d > %d\n", cur_frame, td->frame); urb_priv->isoc_status = -EXDEV; td->urb->iso_frame_desc[ td->isoc_index].actual_length = 0; td->urb->iso_frame_desc[td->isoc_index].status = -EXDEV; if (--urb_priv->isoc_remaining == 0) urb_done(hcd, td->urb, urb_priv->isoc_status); goto too_late; } urb_priv->active = 1; etd->td = td; etd->ep = td->ep; etd->urb = td->urb; etd->len = td->len; etd->dma_handle = td->dma_handle; etd->cpu_buffer = td->cpu_buffer; debug_isoc_submitted(imx21, cur_frame, td); dir = usb_pipeout(td->urb->pipe) ? TD_DIR_OUT : TD_DIR_IN; setup_etd_dword0(imx21, etd_num, td->urb, dir, etd->dmem_size); etd_writel(imx21, etd_num, 1, etd->dmem_offset); etd_writel(imx21, etd_num, 2, (TD_NOTACCESSED << DW2_COMPCODE) | ((td->frame & 0xFFFF) << DW2_STARTFRM)); etd_writel(imx21, etd_num, 3, (TD_NOTACCESSED << DW3_COMPCODE0) | (td->len << DW3_PKTLEN0)); activate_etd(imx21, etd_num, dir); } } static void isoc_etd_done(struct usb_hcd *hcd, int etd_num) { struct imx21 *imx21 = hcd_to_imx21(hcd); int etd_mask = 1 << etd_num; struct etd_priv *etd = imx21->etd + etd_num; struct urb *urb = etd->urb; struct urb_priv *urb_priv = urb->hcpriv; struct td *td = etd->td; struct usb_host_endpoint *ep = etd->ep; int isoc_index = td->isoc_index; unsigned int pipe = urb->pipe; int dir_in = usb_pipein(pipe); int cc; int bytes_xfrd; disactivate_etd(imx21, etd_num); cc = (etd_readl(imx21, etd_num, 3) >> DW3_COMPCODE0) & 0xf; bytes_xfrd = etd_readl(imx21, etd_num, 3) & 0x3ff; /* Input doesn't always fill the buffer, don't generate an error * when this happens. */ if (dir_in && (cc == TD_DATAUNDERRUN)) cc = TD_CC_NOERROR; if (cc == TD_NOTACCESSED) bytes_xfrd = 0; debug_isoc_completed(imx21, imx21_hc_get_frame(hcd), td, cc, bytes_xfrd); if (cc) { urb_priv->isoc_status = -EXDEV; dev_dbg(imx21->dev, "bad iso cc=0x%X frame=%d sched frame=%d " "cnt=%d len=%d urb=%p etd=%d index=%d\n", cc, imx21_hc_get_frame(hcd), td->frame, bytes_xfrd, td->len, urb, etd_num, isoc_index); } if (dir_in) { clear_toggle_bit(imx21, USBH_XFILLSTAT, etd_mask); if (!etd->dma_handle) memcpy_fromio(etd->cpu_buffer, imx21->regs + USBOTG_DMEM + etd->dmem_offset, bytes_xfrd); } urb->actual_length += bytes_xfrd; urb->iso_frame_desc[isoc_index].actual_length = bytes_xfrd; urb->iso_frame_desc[isoc_index].status = cc_to_error[cc]; etd->td = NULL; etd->urb = NULL; etd->ep = NULL; if (--urb_priv->isoc_remaining == 0) urb_done(hcd, urb, urb_priv->isoc_status); schedule_isoc_etds(hcd, ep); } static struct ep_priv *alloc_isoc_ep( struct imx21 *imx21, struct usb_host_endpoint *ep) { struct ep_priv *ep_priv; int i; ep_priv = kzalloc(sizeof(struct ep_priv), GFP_ATOMIC); if (!ep_priv) return NULL; for (i = 0; i < NUM_ISO_ETDS; i++) ep_priv->etd[i] = -1; INIT_LIST_HEAD(&ep_priv->td_list); ep_priv->ep = ep; ep->hcpriv = ep_priv; return ep_priv; } static int alloc_isoc_etds(struct imx21 *imx21, struct ep_priv *ep_priv) { int i, j; int etd_num; /* Allocate the ETDs if required */ for (i = 0; i < NUM_ISO_ETDS; i++) { if (ep_priv->etd[i] < 0) { etd_num = alloc_etd(imx21); if (etd_num < 0) goto alloc_etd_failed; ep_priv->etd[i] = etd_num; imx21->etd[etd_num].ep = ep_priv->ep; } } return 0; alloc_etd_failed: dev_err(imx21->dev, "isoc: Couldn't allocate etd\n"); for (j = 0; j < i; j++) { free_etd(imx21, ep_priv->etd[j]); ep_priv->etd[j] = -1; } return -ENOMEM; } static int imx21_hc_urb_enqueue_isoc(struct usb_hcd *hcd, struct usb_host_endpoint *ep, struct urb *urb, gfp_t mem_flags) { struct imx21 *imx21 = hcd_to_imx21(hcd); struct urb_priv *urb_priv; unsigned long flags; struct ep_priv *ep_priv; struct td *td = NULL; int i; int ret; int cur_frame; u16 maxpacket; urb_priv = kzalloc(sizeof(struct urb_priv), mem_flags); if (urb_priv == NULL) return -ENOMEM; urb_priv->isoc_td = kcalloc(urb->number_of_packets, sizeof(struct td), mem_flags); if (urb_priv->isoc_td == NULL) { ret = -ENOMEM; goto alloc_td_failed; } spin_lock_irqsave(&imx21->lock, flags); if (ep->hcpriv == NULL) { ep_priv = alloc_isoc_ep(imx21, ep); if (ep_priv == NULL) { ret = -ENOMEM; goto alloc_ep_failed; } } else { ep_priv = ep->hcpriv; } ret = alloc_isoc_etds(imx21, ep_priv); if (ret) goto alloc_etd_failed; ret = usb_hcd_link_urb_to_ep(hcd, urb); if (ret) goto link_failed; urb->status = -EINPROGRESS; urb->actual_length = 0; urb->error_count = 0; urb->hcpriv = urb_priv; urb_priv->ep = ep; /* allocate data memory for largest packets if not already done */ maxpacket = usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe)); for (i = 0; i < NUM_ISO_ETDS; i++) { struct etd_priv *etd = &imx21->etd[ep_priv->etd[i]]; if (etd->dmem_size > 0 && etd->dmem_size < maxpacket) { /* not sure if this can really occur.... */ dev_err(imx21->dev, "increasing isoc buffer %d->%d\n", etd->dmem_size, maxpacket); ret = -EMSGSIZE; goto alloc_dmem_failed; } if (etd->dmem_size == 0) { etd->dmem_offset = alloc_dmem(imx21, maxpacket, ep); if (etd->dmem_offset < 0) { dev_dbg(imx21->dev, "failed alloc isoc dmem\n"); ret = -EAGAIN; goto alloc_dmem_failed; } etd->dmem_size = maxpacket; } } /* calculate frame */ cur_frame = imx21_hc_get_frame(hcd); i = 0; if (list_empty(&ep_priv->td_list)) { urb->start_frame = wrap_frame(cur_frame + 5); } else { urb->start_frame = wrap_frame(list_entry(ep_priv->td_list.prev, struct td, list)->frame + urb->interval); if (frame_after(cur_frame, urb->start_frame)) { dev_dbg(imx21->dev, "enqueue: adjusting iso start %d (cur=%d) asap=%d\n", urb->start_frame, cur_frame, (urb->transfer_flags & URB_ISO_ASAP) != 0); i = DIV_ROUND_UP(wrap_frame( cur_frame - urb->start_frame), urb->interval); /* Treat underruns as if URB_ISO_ASAP was set */ if ((urb->transfer_flags & URB_ISO_ASAP) || i >= urb->number_of_packets) { urb->start_frame = wrap_frame(urb->start_frame + i * urb->interval); i = 0; } } } /* set up transfers */ urb_priv->isoc_remaining = urb->number_of_packets - i; td = urb_priv->isoc_td; for (; i < urb->number_of_packets; i++, td++) { unsigned int offset = urb->iso_frame_desc[i].offset; td->ep = ep; td->urb = urb; td->len = urb->iso_frame_desc[i].length; td->isoc_index = i; td->frame = wrap_frame(urb->start_frame + urb->interval * i); td->dma_handle = urb->transfer_dma + offset; td->cpu_buffer = urb->transfer_buffer + offset; list_add_tail(&td->list, &ep_priv->td_list); } dev_vdbg(imx21->dev, "setup %d packets for iso frame %d->%d\n", urb->number_of_packets, urb->start_frame, td->frame); debug_urb_submitted(imx21, urb); schedule_isoc_etds(hcd, ep); spin_unlock_irqrestore(&imx21->lock, flags); return 0; alloc_dmem_failed: usb_hcd_unlink_urb_from_ep(hcd, urb); link_failed: alloc_etd_failed: alloc_ep_failed: spin_unlock_irqrestore(&imx21->lock, flags); kfree(urb_priv->isoc_td); alloc_td_failed: kfree(urb_priv); return ret; } static void dequeue_isoc_urb(struct imx21 *imx21, struct urb *urb, struct ep_priv *ep_priv) { struct urb_priv *urb_priv = urb->hcpriv; struct td *td, *tmp; int i; if (urb_priv->active) { for (i = 0; i < NUM_ISO_ETDS; i++) { int etd_num = ep_priv->etd[i]; if (etd_num != -1 && imx21->etd[etd_num].urb == urb) { struct etd_priv *etd = imx21->etd + etd_num; reset_etd(imx21, etd_num); free_dmem(imx21, etd); } } } list_for_each_entry_safe(td, tmp, &ep_priv->td_list, list) { if (td->urb == urb) { dev_vdbg(imx21->dev, "removing td %p\n", td); list_del(&td->list); } } } /* =========================================== */ /* NON ISOC Handling ... */ /* =========================================== */ static void schedule_nonisoc_etd(struct imx21 *imx21, struct urb *urb) { unsigned int pipe = urb->pipe; struct urb_priv *urb_priv = urb->hcpriv; struct ep_priv *ep_priv = urb_priv->ep->hcpriv; int state = urb_priv->state; int etd_num = ep_priv->etd[0]; struct etd_priv *etd; u32 count; u16 etd_buf_size; u16 maxpacket; u8 dir; u8 bufround; u8 datatoggle; u8 interval = 0; u8 relpolpos = 0; if (etd_num < 0) { dev_err(imx21->dev, "No valid ETD\n"); return; } if (readl(imx21->regs + USBH_ETDENSET) & (1 << etd_num)) dev_err(imx21->dev, "submitting to active ETD %d\n", etd_num); etd = &imx21->etd[etd_num]; maxpacket = usb_maxpacket(urb->dev, pipe, usb_pipeout(pipe)); if (!maxpacket) maxpacket = 8; if (usb_pipecontrol(pipe) && (state != US_CTRL_DATA)) { if (state == US_CTRL_SETUP) { dir = TD_DIR_SETUP; if (unsuitable_for_dma(urb->setup_dma)) usb_hcd_unmap_urb_setup_for_dma(imx21->hcd, urb); etd->dma_handle = urb->setup_dma; etd->cpu_buffer = urb->setup_packet; bufround = 0; count = 8; datatoggle = TD_TOGGLE_DATA0; } else { /* US_CTRL_ACK */ dir = usb_pipeout(pipe) ? TD_DIR_IN : TD_DIR_OUT; bufround = 0; count = 0; datatoggle = TD_TOGGLE_DATA1; } } else { dir = usb_pipeout(pipe) ? TD_DIR_OUT : TD_DIR_IN; bufround = (dir == TD_DIR_IN) ? 1 : 0; if (unsuitable_for_dma(urb->transfer_dma)) usb_hcd_unmap_urb_for_dma(imx21->hcd, urb); etd->dma_handle = urb->transfer_dma; etd->cpu_buffer = urb->transfer_buffer; if (usb_pipebulk(pipe) && (state == US_BULK0)) count = 0; else count = urb->transfer_buffer_length; if (usb_pipecontrol(pipe)) { datatoggle = TD_TOGGLE_DATA1; } else { if (usb_gettoggle( urb->dev, usb_pipeendpoint(urb->pipe), usb_pipeout(urb->pipe))) datatoggle = TD_TOGGLE_DATA1; else datatoggle = TD_TOGGLE_DATA0; } } etd->urb = urb; etd->ep = urb_priv->ep; etd->len = count; if (usb_pipeint(pipe)) { interval = urb->interval; relpolpos = (readl(imx21->regs + USBH_FRMNUB) + 1) & 0xff; } /* Write ETD to device memory */ setup_etd_dword0(imx21, etd_num, urb, dir, maxpacket); etd_writel(imx21, etd_num, 2, (u32) interval << DW2_POLINTERV | ((u32) relpolpos << DW2_RELPOLPOS) | ((u32) dir << DW2_DIRPID) | ((u32) bufround << DW2_BUFROUND) | ((u32) datatoggle << DW2_DATATOG) | ((u32) TD_NOTACCESSED << DW2_COMPCODE)); /* DMA will always transfer buffer size even if TOBYCNT in DWORD3 is smaller. Make sure we don't overrun the buffer! */ if (count && count < maxpacket) etd_buf_size = count; else etd_buf_size = maxpacket; etd_writel(imx21, etd_num, 3, ((u32) (etd_buf_size - 1) << DW3_BUFSIZE) | (u32) count); if (!count) etd->dma_handle = 0; /* allocate x and y buffer space at once */ etd->dmem_size = (count > maxpacket) ? maxpacket * 2 : maxpacket; etd->dmem_offset = alloc_dmem(imx21, etd->dmem_size, urb_priv->ep); if (etd->dmem_offset < 0) { /* Setup everything we can in HW and update when we get DMEM */ etd_writel(imx21, etd_num, 1, (u32)maxpacket << 16); dev_dbg(imx21->dev, "Queuing etd %d for DMEM\n", etd_num); debug_urb_queued_for_dmem(imx21, urb); list_add_tail(&etd->queue, &imx21->queue_for_dmem); return; } etd_writel(imx21, etd_num, 1, (((u32) etd->dmem_offset + (u32) maxpacket) << DW1_YBUFSRTAD) | (u32) etd->dmem_offset); urb_priv->active = 1; /* enable the ETD to kick off transfer */ dev_vdbg(imx21->dev, "Activating etd %d for %d bytes %s\n", etd_num, count, dir != TD_DIR_IN ? "out" : "in"); activate_etd(imx21, etd_num, dir); } static void nonisoc_etd_done(struct usb_hcd *hcd, int etd_num) { struct imx21 *imx21 = hcd_to_imx21(hcd); struct etd_priv *etd = &imx21->etd[etd_num]; struct urb *urb = etd->urb; u32 etd_mask = 1 << etd_num; struct urb_priv *urb_priv = urb->hcpriv; int dir; int cc; u32 bytes_xfrd; int etd_done; disactivate_etd(imx21, etd_num); dir = (etd_readl(imx21, etd_num, 0) >> DW0_DIRECT) & 0x3; cc = (etd_readl(imx21, etd_num, 2) >> DW2_COMPCODE) & 0xf; bytes_xfrd = etd->len - (etd_readl(imx21, etd_num, 3) & 0x1fffff); /* save toggle carry */ usb_settoggle(urb->dev, usb_pipeendpoint(urb->pipe), usb_pipeout(urb->pipe), (etd_readl(imx21, etd_num, 0) >> DW0_TOGCRY) & 0x1); if (dir == TD_DIR_IN) { clear_toggle_bit(imx21, USBH_XFILLSTAT, etd_mask); clear_toggle_bit(imx21, USBH_YFILLSTAT, etd_mask); if (etd->bounce_buffer) { memcpy(etd->cpu_buffer, etd->bounce_buffer, bytes_xfrd); dma_unmap_single(imx21->dev, etd->dma_handle, etd->len, DMA_FROM_DEVICE); } else if (!etd->dma_handle && bytes_xfrd) {/* PIO */ memcpy_fromio(etd->cpu_buffer, imx21->regs + USBOTG_DMEM + etd->dmem_offset, bytes_xfrd); } } kfree(etd->bounce_buffer); etd->bounce_buffer = NULL; free_dmem(imx21, etd); urb->error_count = 0; if (!(urb->transfer_flags & URB_SHORT_NOT_OK) && (cc == TD_DATAUNDERRUN)) cc = TD_CC_NOERROR; if (cc != 0) dev_vdbg(imx21->dev, "cc is 0x%x\n", cc); etd_done = (cc_to_error[cc] != 0); /* stop if error */ switch (usb_pipetype(urb->pipe)) { case PIPE_CONTROL: switch (urb_priv->state) { case US_CTRL_SETUP: if (urb->transfer_buffer_length > 0) urb_priv->state = US_CTRL_DATA; else urb_priv->state = US_CTRL_ACK; break; case US_CTRL_DATA: urb->actual_length += bytes_xfrd; urb_priv->state = US_CTRL_ACK; break; case US_CTRL_ACK: etd_done = 1; break; default: dev_err(imx21->dev, "Invalid pipe state %d\n", urb_priv->state); etd_done = 1; break; } break; case PIPE_BULK: urb->actual_length += bytes_xfrd; if ((urb_priv->state == US_BULK) && (urb->transfer_flags & URB_ZERO_PACKET) && urb->transfer_buffer_length > 0 && ((urb->transfer_buffer_length % usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe))) == 0)) { /* need a 0-packet */ urb_priv->state = US_BULK0; } else { etd_done = 1; } break; case PIPE_INTERRUPT: urb->actual_length += bytes_xfrd; etd_done = 1; break; } if (etd_done) nonisoc_urb_completed_for_etd(imx21, etd, cc_to_error[cc]); else { dev_vdbg(imx21->dev, "next state=%d\n", urb_priv->state); schedule_nonisoc_etd(imx21, urb); } } static struct ep_priv *alloc_ep(void) { int i; struct ep_priv *ep_priv; ep_priv = kzalloc(sizeof(struct ep_priv), GFP_ATOMIC); if (!ep_priv) return NULL; for (i = 0; i < NUM_ISO_ETDS; ++i) ep_priv->etd[i] = -1; return ep_priv; } static int imx21_hc_urb_enqueue(struct usb_hcd *hcd, struct urb *urb, gfp_t mem_flags) { struct imx21 *imx21 = hcd_to_imx21(hcd); struct usb_host_endpoint *ep = urb->ep; struct urb_priv *urb_priv; struct ep_priv *ep_priv; struct etd_priv *etd; int ret; unsigned long flags; dev_vdbg(imx21->dev, "enqueue urb=%p ep=%p len=%d " "buffer=%p dma=%pad setupBuf=%p setupDma=%pad\n", urb, ep, urb->transfer_buffer_length, urb->transfer_buffer, &urb->transfer_dma, urb->setup_packet, &urb->setup_dma); if (usb_pipeisoc(urb->pipe)) return imx21_hc_urb_enqueue_isoc(hcd, ep, urb, mem_flags); urb_priv = kzalloc(sizeof(struct urb_priv), mem_flags); if (!urb_priv) return -ENOMEM; spin_lock_irqsave(&imx21->lock, flags); ep_priv = ep->hcpriv; if (ep_priv == NULL) { ep_priv = alloc_ep(); if (!ep_priv) { ret = -ENOMEM; goto failed_alloc_ep; } ep->hcpriv = ep_priv; ep_priv->ep = ep; } ret = usb_hcd_link_urb_to_ep(hcd, urb); if (ret) goto failed_link; urb->status = -EINPROGRESS; urb->actual_length = 0; urb->error_count = 0; urb->hcpriv = urb_priv; urb_priv->ep = ep; switch (usb_pipetype(urb->pipe)) { case PIPE_CONTROL: urb_priv->state = US_CTRL_SETUP; break; case PIPE_BULK: urb_priv->state = US_BULK; break; } debug_urb_submitted(imx21, urb); if (ep_priv->etd[0] < 0) { if (ep_priv->waiting_etd) { dev_dbg(imx21->dev, "no ETD available already queued %p\n", ep_priv); debug_urb_queued_for_etd(imx21, urb); goto out; } ep_priv->etd[0] = alloc_etd(imx21); if (ep_priv->etd[0] < 0) { dev_dbg(imx21->dev, "no ETD available queueing %p\n", ep_priv); debug_urb_queued_for_etd(imx21, urb); list_add_tail(&ep_priv->queue, &imx21->queue_for_etd); ep_priv->waiting_etd = 1; goto out; } } /* Schedule if no URB already active for this endpoint */ etd = &imx21->etd[ep_priv->etd[0]]; if (etd->urb == NULL) { DEBUG_LOG_FRAME(imx21, etd, last_req); schedule_nonisoc_etd(imx21, urb); } out: spin_unlock_irqrestore(&imx21->lock, flags); return 0; failed_link: failed_alloc_ep: spin_unlock_irqrestore(&imx21->lock, flags); kfree(urb_priv); return ret; } static int imx21_hc_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status) { struct imx21 *imx21 = hcd_to_imx21(hcd); unsigned long flags; struct usb_host_endpoint *ep; struct ep_priv *ep_priv; struct urb_priv *urb_priv = urb->hcpriv; int ret = -EINVAL; dev_vdbg(imx21->dev, "dequeue urb=%p iso=%d status=%d\n", urb, usb_pipeisoc(urb->pipe), status); spin_lock_irqsave(&imx21->lock, flags); ret = usb_hcd_check_unlink_urb(hcd, urb, status); if (ret) goto fail; ep = urb_priv->ep; ep_priv = ep->hcpriv; debug_urb_unlinked(imx21, urb); if (usb_pipeisoc(urb->pipe)) { dequeue_isoc_urb(imx21, urb, ep_priv); schedule_isoc_etds(hcd, ep); } else if (urb_priv->active) { int etd_num = ep_priv->etd[0]; if (etd_num != -1) { struct etd_priv *etd = &imx21->etd[etd_num]; disactivate_etd(imx21, etd_num); free_dmem(imx21, etd); etd->urb = NULL; kfree(etd->bounce_buffer); etd->bounce_buffer = NULL; } } urb_done(hcd, urb, status); spin_unlock_irqrestore(&imx21->lock, flags); return 0; fail: spin_unlock_irqrestore(&imx21->lock, flags); return ret; } /* =========================================== */ /* Interrupt dispatch */ /* =========================================== */ static void process_etds(struct usb_hcd *hcd, struct imx21 *imx21, int sof) { int etd_num; int enable_sof_int = 0; unsigned long flags; spin_lock_irqsave(&imx21->lock, flags); for (etd_num = 0; etd_num < USB_NUM_ETD; etd_num++) { u32 etd_mask = 1 << etd_num; u32 enabled = readl(imx21->regs + USBH_ETDENSET) & etd_mask; u32 done = readl(imx21->regs + USBH_ETDDONESTAT) & etd_mask; struct etd_priv *etd = &imx21->etd[etd_num]; if (done) { DEBUG_LOG_FRAME(imx21, etd, last_int); } else { /* * Kludge warning! * * When multiple transfers are using the bus we sometimes get into a state * where the transfer has completed (the CC field of the ETD is != 0x0F), * the ETD has self disabled but the ETDDONESTAT flag is not set * (and hence no interrupt occurs). * This causes the transfer in question to hang. * The kludge below checks for this condition at each SOF and processes any * blocked ETDs (after an arbitrary 10 frame wait) * * With a single active transfer the usbtest test suite will run for days * without the kludge. * With other bus activity (eg mass storage) even just test1 will hang without * the kludge. */ u32 dword0; int cc; if (etd->active_count && !enabled) /* suspicious... */ enable_sof_int = 1; if (!sof || enabled || !etd->active_count) continue; cc = etd_readl(imx21, etd_num, 2) >> DW2_COMPCODE; if (cc == TD_NOTACCESSED) continue; if (++etd->active_count < 10) continue; dword0 = etd_readl(imx21, etd_num, 0); dev_dbg(imx21->dev, "unblock ETD %d dev=0x%X ep=0x%X cc=0x%02X!\n", etd_num, dword0 & 0x7F, (dword0 >> DW0_ENDPNT) & 0x0F, cc); #ifdef DEBUG dev_dbg(imx21->dev, "frame: act=%d disact=%d" " int=%d req=%d cur=%d\n", etd->activated_frame, etd->disactivated_frame, etd->last_int_frame, etd->last_req_frame, readl(imx21->regs + USBH_FRMNUB)); imx21->debug_unblocks++; #endif etd->active_count = 0; /* End of kludge */ } if (etd->ep == NULL || etd->urb == NULL) { dev_dbg(imx21->dev, "Interrupt for unexpected etd %d" " ep=%p urb=%p\n", etd_num, etd->ep, etd->urb); disactivate_etd(imx21, etd_num); continue; } if (usb_pipeisoc(etd->urb->pipe)) isoc_etd_done(hcd, etd_num); else nonisoc_etd_done(hcd, etd_num); } /* only enable SOF interrupt if it may be needed for the kludge */ if (enable_sof_int) set_register_bits(imx21, USBH_SYSIEN, USBH_SYSIEN_SOFINT); else clear_register_bits(imx21, USBH_SYSIEN, USBH_SYSIEN_SOFINT); spin_unlock_irqrestore(&imx21->lock, flags); } static irqreturn_t imx21_irq(struct usb_hcd *hcd) { struct imx21 *imx21 = hcd_to_imx21(hcd); u32 ints = readl(imx21->regs + USBH_SYSISR); if (ints & USBH_SYSIEN_HERRINT) dev_dbg(imx21->dev, "Scheduling error\n"); if (ints & USBH_SYSIEN_SORINT) dev_dbg(imx21->dev, "Scheduling overrun\n"); if (ints & (USBH_SYSISR_DONEINT | USBH_SYSISR_SOFINT)) process_etds(hcd, imx21, ints & USBH_SYSISR_SOFINT); writel(ints, imx21->regs + USBH_SYSISR); return IRQ_HANDLED; } static void imx21_hc_endpoint_disable(struct usb_hcd *hcd, struct usb_host_endpoint *ep) { struct imx21 *imx21 = hcd_to_imx21(hcd); unsigned long flags; struct ep_priv *ep_priv; int i; if (ep == NULL) return; spin_lock_irqsave(&imx21->lock, flags); ep_priv = ep->hcpriv; dev_vdbg(imx21->dev, "disable ep=%p, ep->hcpriv=%p\n", ep, ep_priv); if (!list_empty(&ep->urb_list)) dev_dbg(imx21->dev, "ep's URB list is not empty\n"); if (ep_priv != NULL) { for (i = 0; i < NUM_ISO_ETDS; i++) { if (ep_priv->etd[i] > -1) dev_dbg(imx21->dev, "free etd %d for disable\n", ep_priv->etd[i]); free_etd(imx21, ep_priv->etd[i]); } kfree(ep_priv); ep->hcpriv = NULL; } for (i = 0; i < USB_NUM_ETD; i++) { if (imx21->etd[i].alloc && imx21->etd[i].ep == ep) { dev_err(imx21->dev, "Active etd %d for disabled ep=%p!\n", i, ep); free_etd(imx21, i); } } free_epdmem(imx21, ep); spin_unlock_irqrestore(&imx21->lock, flags); } /* =========================================== */ /* Hub handling */ /* =========================================== */ static int get_hub_descriptor(struct usb_hcd *hcd, struct usb_hub_descriptor *desc) { struct imx21 *imx21 = hcd_to_imx21(hcd); desc->bDescriptorType = USB_DT_HUB; /* HUB descriptor */ desc->bHubContrCurrent = 0; desc->bNbrPorts = readl(imx21->regs + USBH_ROOTHUBA) & USBH_ROOTHUBA_NDNSTMPRT_MASK; desc->bDescLength = 9; desc->bPwrOn2PwrGood = 0; desc->wHubCharacteristics = (__force __u16) cpu_to_le16( HUB_CHAR_NO_LPSM | /* No power switching */ HUB_CHAR_NO_OCPM); /* No over current protection */ desc->u.hs.DeviceRemovable[0] = 1 << 1; desc->u.hs.DeviceRemovable[1] = ~0; return 0; } static int imx21_hc_hub_status_data(struct usb_hcd *hcd, char *buf) { struct imx21 *imx21 = hcd_to_imx21(hcd); int ports; int changed = 0; int i; unsigned long flags; spin_lock_irqsave(&imx21->lock, flags); ports = readl(imx21->regs + USBH_ROOTHUBA) & USBH_ROOTHUBA_NDNSTMPRT_MASK; if (ports > 7) { ports = 7; dev_err(imx21->dev, "ports %d > 7\n", ports); } for (i = 0; i < ports; i++) { if (readl(imx21->regs + USBH_PORTSTAT(i)) & (USBH_PORTSTAT_CONNECTSC | USBH_PORTSTAT_PRTENBLSC | USBH_PORTSTAT_PRTSTATSC | USBH_PORTSTAT_OVRCURIC | USBH_PORTSTAT_PRTRSTSC)) { changed = 1; buf[0] |= 1 << (i + 1); } } spin_unlock_irqrestore(&imx21->lock, flags); if (changed) dev_info(imx21->dev, "Hub status changed\n"); return changed; } static int imx21_hc_hub_control(struct usb_hcd *hcd, u16 typeReq, u16 wValue, u16 wIndex, char *buf, u16 wLength) { struct imx21 *imx21 = hcd_to_imx21(hcd); int rc = 0; u32 status_write = 0; switch (typeReq) { case ClearHubFeature: dev_dbg(imx21->dev, "ClearHubFeature\n"); switch (wValue) { case C_HUB_OVER_CURRENT: dev_dbg(imx21->dev, " OVER_CURRENT\n"); break; case C_HUB_LOCAL_POWER: dev_dbg(imx21->dev, " LOCAL_POWER\n"); break; default: dev_dbg(imx21->dev, " unknown\n"); rc = -EINVAL; break; } break; case ClearPortFeature: dev_dbg(imx21->dev, "ClearPortFeature\n"); switch (wValue) { case USB_PORT_FEAT_ENABLE: dev_dbg(imx21->dev, " ENABLE\n"); status_write = USBH_PORTSTAT_CURCONST; break; case USB_PORT_FEAT_SUSPEND: dev_dbg(imx21->dev, " SUSPEND\n"); status_write = USBH_PORTSTAT_PRTOVRCURI; break; case USB_PORT_FEAT_POWER: dev_dbg(imx21->dev, " POWER\n"); status_write = USBH_PORTSTAT_LSDEVCON; break; case USB_PORT_FEAT_C_ENABLE: dev_dbg(imx21->dev, " C_ENABLE\n"); status_write = USBH_PORTSTAT_PRTENBLSC; break; case USB_PORT_FEAT_C_SUSPEND: dev_dbg(imx21->dev, " C_SUSPEND\n"); status_write = USBH_PORTSTAT_PRTSTATSC; break; case USB_PORT_FEAT_C_CONNECTION: dev_dbg(imx21->dev, " C_CONNECTION\n"); status_write = USBH_PORTSTAT_CONNECTSC; break; case USB_PORT_FEAT_C_OVER_CURRENT: dev_dbg(imx21->dev, " C_OVER_CURRENT\n"); status_write = USBH_PORTSTAT_OVRCURIC; break; case USB_PORT_FEAT_C_RESET: dev_dbg(imx21->dev, " C_RESET\n"); status_write = USBH_PORTSTAT_PRTRSTSC; break; default: dev_dbg(imx21->dev, " unknown\n"); rc = -EINVAL; break; } break; case GetHubDescriptor: dev_dbg(imx21->dev, "GetHubDescriptor\n"); rc = get_hub_descriptor(hcd, (void *)buf); break; case GetHubStatus: dev_dbg(imx21->dev, " GetHubStatus\n"); *(__le32 *) buf = 0; break; case GetPortStatus: dev_dbg(imx21->dev, "GetPortStatus: port: %d, 0x%x\n", wIndex, USBH_PORTSTAT(wIndex - 1)); *(__le32 *) buf = readl(imx21->regs + USBH_PORTSTAT(wIndex - 1)); break; case SetHubFeature: dev_dbg(imx21->dev, "SetHubFeature\n"); switch (wValue) { case C_HUB_OVER_CURRENT: dev_dbg(imx21->dev, " OVER_CURRENT\n"); break; case C_HUB_LOCAL_POWER: dev_dbg(imx21->dev, " LOCAL_POWER\n"); break; default: dev_dbg(imx21->dev, " unknown\n"); rc = -EINVAL; break; } break; case SetPortFeature: dev_dbg(imx21->dev, "SetPortFeature\n"); switch (wValue) { case USB_PORT_FEAT_SUSPEND: dev_dbg(imx21->dev, " SUSPEND\n"); status_write = USBH_PORTSTAT_PRTSUSPST; break; case USB_PORT_FEAT_POWER: dev_dbg(imx21->dev, " POWER\n"); status_write = USBH_PORTSTAT_PRTPWRST; break; case USB_PORT_FEAT_RESET: dev_dbg(imx21->dev, " RESET\n"); status_write = USBH_PORTSTAT_PRTRSTST; break; default: dev_dbg(imx21->dev, " unknown\n"); rc = -EINVAL; break; } break; default: dev_dbg(imx21->dev, " unknown\n"); rc = -EINVAL; break; } if (status_write) writel(status_write, imx21->regs + USBH_PORTSTAT(wIndex - 1)); return rc; } /* =========================================== */ /* Host controller management */ /* =========================================== */ static int imx21_hc_reset(struct usb_hcd *hcd) { struct imx21 *imx21 = hcd_to_imx21(hcd); unsigned long timeout; unsigned long flags; spin_lock_irqsave(&imx21->lock, flags); /* Reset the Host controller modules */ writel(USBOTG_RST_RSTCTRL | USBOTG_RST_RSTRH | USBOTG_RST_RSTHSIE | USBOTG_RST_RSTHC, imx21->regs + USBOTG_RST_CTRL); /* Wait for reset to finish */ timeout = jiffies + HZ; while (readl(imx21->regs + USBOTG_RST_CTRL) != 0) { if (time_after(jiffies, timeout)) { spin_unlock_irqrestore(&imx21->lock, flags); dev_err(imx21->dev, "timeout waiting for reset\n"); return -ETIMEDOUT; } spin_unlock_irq(&imx21->lock); schedule_timeout_uninterruptible(1); spin_lock_irq(&imx21->lock); } spin_unlock_irqrestore(&imx21->lock, flags); return 0; } static int imx21_hc_start(struct usb_hcd *hcd) { struct imx21 *imx21 = hcd_to_imx21(hcd); unsigned long flags; int i, j; u32 hw_mode = USBOTG_HWMODE_CRECFG_HOST; u32 usb_control = 0; hw_mode |= ((imx21->pdata->host_xcvr << USBOTG_HWMODE_HOSTXCVR_SHIFT) & USBOTG_HWMODE_HOSTXCVR_MASK); hw_mode |= ((imx21->pdata->otg_xcvr << USBOTG_HWMODE_OTGXCVR_SHIFT) & USBOTG_HWMODE_OTGXCVR_MASK); if (imx21->pdata->host1_txenoe) usb_control |= USBCTRL_HOST1_TXEN_OE; if (!imx21->pdata->host1_xcverless) usb_control |= USBCTRL_HOST1_BYP_TLL; if (imx21->pdata->otg_ext_xcvr) usb_control |= USBCTRL_OTC_RCV_RXDP; spin_lock_irqsave(&imx21->lock, flags); writel((USBOTG_CLK_CTRL_HST | USBOTG_CLK_CTRL_MAIN), imx21->regs + USBOTG_CLK_CTRL); writel(hw_mode, imx21->regs + USBOTG_HWMODE); writel(usb_control, imx21->regs + USBCTRL); writel(USB_MISCCONTROL_SKPRTRY | USB_MISCCONTROL_ARBMODE, imx21->regs + USB_MISCCONTROL); /* Clear the ETDs */ for (i = 0; i < USB_NUM_ETD; i++) for (j = 0; j < 4; j++) etd_writel(imx21, i, j, 0); /* Take the HC out of reset */ writel(USBH_HOST_CTRL_HCUSBSTE_OPERATIONAL | USBH_HOST_CTRL_CTLBLKSR_1, imx21->regs + USBH_HOST_CTRL); /* Enable ports */ if (imx21->pdata->enable_otg_host) writel(USBH_PORTSTAT_PRTPWRST | USBH_PORTSTAT_PRTENABST, imx21->regs + USBH_PORTSTAT(0)); if (imx21->pdata->enable_host1) writel(USBH_PORTSTAT_PRTPWRST | USBH_PORTSTAT_PRTENABST, imx21->regs + USBH_PORTSTAT(1)); if (imx21->pdata->enable_host2) writel(USBH_PORTSTAT_PRTPWRST | USBH_PORTSTAT_PRTENABST, imx21->regs + USBH_PORTSTAT(2)); hcd->state = HC_STATE_RUNNING; /* Enable host controller interrupts */ set_register_bits(imx21, USBH_SYSIEN, USBH_SYSIEN_HERRINT | USBH_SYSIEN_DONEINT | USBH_SYSIEN_SORINT); set_register_bits(imx21, USBOTG_CINT_STEN, USBOTG_HCINT); spin_unlock_irqrestore(&imx21->lock, flags); return 0; } static void imx21_hc_stop(struct usb_hcd *hcd) { struct imx21 *imx21 = hcd_to_imx21(hcd); unsigned long flags; spin_lock_irqsave(&imx21->lock, flags); writel(0, imx21->regs + USBH_SYSIEN); clear_register_bits(imx21, USBOTG_CINT_STEN, USBOTG_HCINT); clear_register_bits(imx21, USBOTG_CLK_CTRL_HST | USBOTG_CLK_CTRL_MAIN, USBOTG_CLK_CTRL); spin_unlock_irqrestore(&imx21->lock, flags); } /* =========================================== */ /* Driver glue */ /* =========================================== */ static const struct hc_driver imx21_hc_driver = { .description = hcd_name, .product_desc = "IMX21 USB Host Controller", .hcd_priv_size = sizeof(struct imx21), .flags = HCD_USB11, .irq = imx21_irq, .reset = imx21_hc_reset, .start = imx21_hc_start, .stop = imx21_hc_stop, /* I/O requests */ .urb_enqueue = imx21_hc_urb_enqueue, .urb_dequeue = imx21_hc_urb_dequeue, .endpoint_disable = imx21_hc_endpoint_disable, /* scheduling support */ .get_frame_number = imx21_hc_get_frame, /* Root hub support */ .hub_status_data = imx21_hc_hub_status_data, .hub_control = imx21_hc_hub_control, }; static struct mx21_usbh_platform_data default_pdata = { .host_xcvr = MX21_USBXCVR_TXDIF_RXDIF, .otg_xcvr = MX21_USBXCVR_TXDIF_RXDIF, .enable_host1 = 1, .enable_host2 = 1, .enable_otg_host = 1, }; static int imx21_remove(struct platform_device *pdev) { struct usb_hcd *hcd = platform_get_drvdata(pdev); struct imx21 *imx21 = hcd_to_imx21(hcd); struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, 0); remove_debug_files(imx21); usb_remove_hcd(hcd); if (res != NULL) { clk_disable_unprepare(imx21->clk); clk_put(imx21->clk); iounmap(imx21->regs); release_mem_region(res->start, resource_size(res)); } kfree(hcd); return 0; } static int imx21_probe(struct platform_device *pdev) { struct usb_hcd *hcd; struct imx21 *imx21; struct resource *res; int ret; int irq; printk(KERN_INFO "%s\n", imx21_hc_driver.product_desc); res = platform_get_resource(pdev, IORESOURCE_MEM, 0); if (!res) return -ENODEV; irq = platform_get_irq(pdev, 0); if (irq < 0) { dev_err(&pdev->dev, "Failed to get IRQ: %d\n", irq); return irq; } hcd = usb_create_hcd(&imx21_hc_driver, &pdev->dev, dev_name(&pdev->dev)); if (hcd == NULL) { dev_err(&pdev->dev, "Cannot create hcd (%s)\n", dev_name(&pdev->dev)); return -ENOMEM; } imx21 = hcd_to_imx21(hcd); imx21->hcd = hcd; imx21->dev = &pdev->dev; imx21->pdata = dev_get_platdata(&pdev->dev); if (!imx21->pdata) imx21->pdata = &default_pdata; spin_lock_init(&imx21->lock); INIT_LIST_HEAD(&imx21->dmem_list); INIT_LIST_HEAD(&imx21->queue_for_etd); INIT_LIST_HEAD(&imx21->queue_for_dmem); create_debug_files(imx21); res = request_mem_region(res->start, resource_size(res), hcd_name); if (!res) { ret = -EBUSY; goto failed_request_mem; } imx21->regs = ioremap(res->start, resource_size(res)); if (imx21->regs == NULL) { dev_err(imx21->dev, "Cannot map registers\n"); ret = -ENOMEM; goto failed_ioremap; } /* Enable clocks source */ imx21->clk = clk_get(imx21->dev, NULL); if (IS_ERR(imx21->clk)) { dev_err(imx21->dev, "no clock found\n"); ret = PTR_ERR(imx21->clk); goto failed_clock_get; } ret = clk_set_rate(imx21->clk, clk_round_rate(imx21->clk, 48000000)); if (ret) goto failed_clock_set; ret = clk_prepare_enable(imx21->clk); if (ret) goto failed_clock_enable; dev_info(imx21->dev, "Hardware HC revision: 0x%02X\n", (readl(imx21->regs + USBOTG_HWMODE) >> 16) & 0xFF); ret = usb_add_hcd(hcd, irq, 0); if (ret != 0) { dev_err(imx21->dev, "usb_add_hcd() returned %d\n", ret); goto failed_add_hcd; } device_wakeup_enable(hcd->self.controller); return 0; failed_add_hcd: clk_disable_unprepare(imx21->clk); failed_clock_enable: failed_clock_set: clk_put(imx21->clk); failed_clock_get: iounmap(imx21->regs); failed_ioremap: release_mem_region(res->start, resource_size(res)); failed_request_mem: remove_debug_files(imx21); usb_put_hcd(hcd); return ret; } static struct platform_driver imx21_hcd_driver = { .driver = { .name = hcd_name, }, .probe = imx21_probe, .remove = imx21_remove, .suspend = NULL, .resume = NULL, }; module_platform_driver(imx21_hcd_driver); MODULE_DESCRIPTION("i.MX21 USB Host controller"); MODULE_AUTHOR("Martin Fuzzey"); MODULE_LICENSE("GPL"); MODULE_ALIAS("platform:imx21-hcd");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1