Release 4.17 drivers/usb/gadget/udc/m66592-udc.c
  
  
  
// SPDX-License-Identifier: GPL-2.0
/*
 * M66592 UDC (USB gadget)
 *
 * Copyright (C) 2006-2007 Renesas Solutions Corp.
 *
 * Author : Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
 */
#include <linux/module.h>
#include <linux/interrupt.h>
#include <linux/delay.h>
#include <linux/io.h>
#include <linux/platform_device.h>
#include <linux/slab.h>
#include <linux/err.h>
#include <linux/usb/ch9.h>
#include <linux/usb/gadget.h>
#include "m66592-udc.h"
MODULE_DESCRIPTION("M66592 USB gadget driver");
MODULE_LICENSE("GPL");
MODULE_AUTHOR("Yoshihiro Shimoda");
MODULE_ALIAS("platform:m66592_udc");
#define DRIVER_VERSION	"21 July 2009"
static const char udc_name[] = "m66592_udc";
static const char *m66592_ep_name[] = {
	"ep0", "ep1", "ep2", "ep3", "ep4", "ep5", "ep6", "ep7"
};
static void disable_controller(struct m66592 *m66592);
static void irq_ep0_write(struct m66592_ep *ep, struct m66592_request *req);
static void irq_packet_write(struct m66592_ep *ep, struct m66592_request *req);
static int m66592_queue(struct usb_ep *_ep, struct usb_request *_req,
			gfp_t gfp_flags);
static void transfer_complete(struct m66592_ep *ep,
		struct m66592_request *req, int status);
/*-------------------------------------------------------------------------*/
static inline u16 get_usb_speed(struct m66592 *m66592)
{
	return (m66592_read(m66592, M66592_DVSTCTR) & M66592_RHST);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 24 | 100.00% | 1 | 100.00% | 
| Total | 24 | 100.00% | 1 | 100.00% | 
static void enable_pipe_irq(struct m66592 *m66592, u16 pipenum,
		unsigned long reg)
{
	u16 tmp;
	tmp = m66592_read(m66592, M66592_INTENB0);
	m66592_bclr(m66592, M66592_BEMPE | M66592_NRDYE | M66592_BRDYE,
			M66592_INTENB0);
	m66592_bset(m66592, (1 << pipenum), reg);
	m66592_write(m66592, tmp, M66592_INTENB0);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 65 | 100.00% | 1 | 100.00% | 
| Total | 65 | 100.00% | 1 | 100.00% | 
static void disable_pipe_irq(struct m66592 *m66592, u16 pipenum,
		unsigned long reg)
{
	u16 tmp;
	tmp = m66592_read(m66592, M66592_INTENB0);
	m66592_bclr(m66592, M66592_BEMPE | M66592_NRDYE | M66592_BRDYE,
			M66592_INTENB0);
	m66592_bclr(m66592, (1 << pipenum), reg);
	m66592_write(m66592, tmp, M66592_INTENB0);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 65 | 100.00% | 1 | 100.00% | 
| Total | 65 | 100.00% | 1 | 100.00% | 
static void m66592_usb_connect(struct m66592 *m66592)
{
	m66592_bset(m66592, M66592_CTRE, M66592_INTENB0);
	m66592_bset(m66592, M66592_WDST | M66592_RDST | M66592_CMPL,
			M66592_INTENB0);
	m66592_bset(m66592, M66592_BEMPE | M66592_BRDYE, M66592_INTENB0);
	m66592_bset(m66592, M66592_DPRPU, M66592_SYSCFG);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 53 | 100.00% | 1 | 100.00% | 
| Total | 53 | 100.00% | 1 | 100.00% | 
static void m66592_usb_disconnect(struct m66592 *m66592)
__releases(m66592->lock)
__acquires(m66592->lock)
{
	m66592_bclr(m66592, M66592_CTRE, M66592_INTENB0);
	m66592_bclr(m66592, M66592_WDST | M66592_RDST | M66592_CMPL,
			M66592_INTENB0);
	m66592_bclr(m66592, M66592_BEMPE | M66592_BRDYE, M66592_INTENB0);
	m66592_bclr(m66592, M66592_DPRPU, M66592_SYSCFG);
	m66592->gadget.speed = USB_SPEED_UNKNOWN;
	spin_unlock(&m66592->lock);
	m66592->driver->disconnect(&m66592->gadget);
	spin_lock(&m66592->lock);
	disable_controller(m66592);
	INIT_LIST_HEAD(&m66592->ep[0].queue);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 119 | 100.00% | 2 | 100.00% | 
| Total | 119 | 100.00% | 2 | 100.00% | 
static inline u16 control_reg_get_pid(struct m66592 *m66592, u16 pipenum)
{
	u16 pid = 0;
	unsigned long offset;
	if (pipenum == 0)
		pid = m66592_read(m66592, M66592_DCPCTR) & M66592_PID;
	else if (pipenum < M66592_MAX_NUM_PIPE) {
		offset = get_pipectr_addr(pipenum);
		pid = m66592_read(m66592, offset) & M66592_PID;
	} else
		pr_err("unexpect pipe num (%d)\n", pipenum);
	return pid;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 78 | 98.73% | 1 | 50.00% | 
| David Brownell | 1 | 1.27% | 1 | 50.00% | 
| Total | 79 | 100.00% | 2 | 100.00% | 
static inline void control_reg_set_pid(struct m66592 *m66592, u16 pipenum,
		u16 pid)
{
	unsigned long offset;
	if (pipenum == 0)
		m66592_mdfy(m66592, pid, M66592_PID, M66592_DCPCTR);
	else if (pipenum < M66592_MAX_NUM_PIPE) {
		offset = get_pipectr_addr(pipenum);
		m66592_mdfy(m66592, pid, M66592_PID, offset);
	} else
		pr_err("unexpect pipe num (%d)\n", pipenum);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 73 | 98.65% | 1 | 50.00% | 
| David Brownell | 1 | 1.35% | 1 | 50.00% | 
| Total | 74 | 100.00% | 2 | 100.00% | 
static inline void pipe_start(struct m66592 *m66592, u16 pipenum)
{
	control_reg_set_pid(m66592, pipenum, M66592_PID_BUF);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 24 | 100.00% | 1 | 100.00% | 
| Total | 24 | 100.00% | 1 | 100.00% | 
static inline void pipe_stop(struct m66592 *m66592, u16 pipenum)
{
	control_reg_set_pid(m66592, pipenum, M66592_PID_NAK);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 24 | 100.00% | 1 | 100.00% | 
| Total | 24 | 100.00% | 1 | 100.00% | 
static inline void pipe_stall(struct m66592 *m66592, u16 pipenum)
{
	control_reg_set_pid(m66592, pipenum, M66592_PID_STALL);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 24 | 100.00% | 1 | 100.00% | 
| Total | 24 | 100.00% | 1 | 100.00% | 
static inline u16 control_reg_get(struct m66592 *m66592, u16 pipenum)
{
	u16 ret = 0;
	unsigned long offset;
	if (pipenum == 0)
		ret = m66592_read(m66592, M66592_DCPCTR);
	else if (pipenum < M66592_MAX_NUM_PIPE) {
		offset = get_pipectr_addr(pipenum);
		ret = m66592_read(m66592, offset);
	} else
		pr_err("unexpect pipe num (%d)\n", pipenum);
	return ret;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 74 | 98.67% | 1 | 50.00% | 
| David Brownell | 1 | 1.33% | 1 | 50.00% | 
| Total | 75 | 100.00% | 2 | 100.00% | 
static inline void control_reg_sqclr(struct m66592 *m66592, u16 pipenum)
{
	unsigned long offset;
	pipe_stop(m66592, pipenum);
	if (pipenum == 0)
		m66592_bset(m66592, M66592_SQCLR, M66592_DCPCTR);
	else if (pipenum < M66592_MAX_NUM_PIPE) {
		offset = get_pipectr_addr(pipenum);
		m66592_bset(m66592, M66592_SQCLR, offset);
	} else
		pr_err("unexpect pipe num(%d)\n", pipenum);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 73 | 98.65% | 1 | 50.00% | 
| David Brownell | 1 | 1.35% | 1 | 50.00% | 
| Total | 74 | 100.00% | 2 | 100.00% | 
static inline int get_buffer_size(struct m66592 *m66592, u16 pipenum)
{
	u16 tmp;
	int size;
	if (pipenum == 0) {
		tmp = m66592_read(m66592, M66592_DCPCFG);
		if ((tmp & M66592_CNTMD) != 0)
			size = 256;
		else {
			tmp = m66592_read(m66592, M66592_DCPMAXP);
			size = tmp & M66592_MAXP;
		}
	} else {
		m66592_write(m66592, pipenum, M66592_PIPESEL);
		tmp = m66592_read(m66592, M66592_PIPECFG);
		if ((tmp & M66592_CNTMD) != 0) {
			tmp = m66592_read(m66592, M66592_PIPEBUF);
			size = ((tmp >> 10) + 1) * 64;
		} else {
			tmp = m66592_read(m66592, M66592_PIPEMAXP);
			size = tmp & M66592_MXPS;
		}
	}
	return size;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 147 | 100.00% | 1 | 100.00% | 
| Total | 147 | 100.00% | 1 | 100.00% | 
static inline void pipe_change(struct m66592 *m66592, u16 pipenum)
{
	struct m66592_ep *ep = m66592->pipenum2ep[pipenum];
	unsigned short mbw;
	if (ep->use_dma)
		return;
	m66592_mdfy(m66592, pipenum, M66592_CURPIPE, ep->fifosel);
	ndelay(450);
	if (m66592->pdata->on_chip)
		mbw = M66592_MBW_32;
	else
		mbw = M66592_MBW_16;
	m66592_bset(m66592, mbw, ep->fifosel);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 62 | 73.81% | 1 | 50.00% | 
| Magnus Damm | 22 | 26.19% | 1 | 50.00% | 
| Total | 84 | 100.00% | 2 | 100.00% | 
static int pipe_buffer_setting(struct m66592 *m66592,
		struct m66592_pipe_info *info)
{
	u16 bufnum = 0, buf_bsize = 0;
	u16 pipecfg = 0;
	if (info->pipe == 0)
		return -EINVAL;
	m66592_write(m66592, info->pipe, M66592_PIPESEL);
	if (info->dir_in)
		pipecfg |= M66592_DIR;
	pipecfg |= info->type;
	pipecfg |= info->epnum;
	switch (info->type) {
	case M66592_INT:
		bufnum = 4 + (info->pipe - M66592_BASE_PIPENUM_INT);
		buf_bsize = 0;
		break;
	case M66592_BULK:
		/* isochronous pipes may be used as bulk pipes */
		if (info->pipe >= M66592_BASE_PIPENUM_BULK)
			bufnum = info->pipe - M66592_BASE_PIPENUM_BULK;
		else
			bufnum = info->pipe - M66592_BASE_PIPENUM_ISOC;
		bufnum = M66592_BASE_BUFNUM + (bufnum * 16);
		buf_bsize = 7;
		pipecfg |= M66592_DBLB;
		if (!info->dir_in)
			pipecfg |= M66592_SHTNAK;
		break;
	case M66592_ISO:
		bufnum = M66592_BASE_BUFNUM +
			 (info->pipe - M66592_BASE_PIPENUM_ISOC) * 16;
		buf_bsize = 7;
		break;
	}
	if (buf_bsize && ((bufnum + 16) >= M66592_MAX_BUFNUM)) {
		pr_err("m66592 pipe memory is insufficient\n");
		return -ENOMEM;
	}
	m66592_write(m66592, pipecfg, M66592_PIPECFG);
	m66592_write(m66592, (buf_bsize << 10) | (bufnum), M66592_PIPEBUF);
	m66592_write(m66592, info->maxpacket, M66592_PIPEMAXP);
	if (info->interval)
		info->interval--;
	m66592_write(m66592, info->interval, M66592_PIPEPERI);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 233 | 85.98% | 1 | 25.00% | 
| Magnus Damm | 36 | 13.28% | 1 | 25.00% | 
| Yusuke Goda | 1 | 0.37% | 1 | 25.00% | 
| David Brownell | 1 | 0.37% | 1 | 25.00% | 
| Total | 271 | 100.00% | 4 | 100.00% | 
static void pipe_buffer_release(struct m66592 *m66592,
				struct m66592_pipe_info *info)
{
	if (info->pipe == 0)
		return;
	if (is_bulk_pipe(info->pipe)) {
		m66592->bulk--;
	} else if (is_interrupt_pipe(info->pipe))
		m66592->interrupt--;
	else if (is_isoc_pipe(info->pipe)) {
		m66592->isochronous--;
		if (info->type == M66592_BULK)
			m66592->bulk--;
	} else
		pr_err("ep_release: unexpect pipenum (%d)\n",
				info->pipe);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 95 | 98.96% | 1 | 50.00% | 
| David Brownell | 1 | 1.04% | 1 | 50.00% | 
| Total | 96 | 100.00% | 2 | 100.00% | 
static void pipe_initialize(struct m66592_ep *ep)
{
	struct m66592 *m66592 = ep->m66592;
	unsigned short mbw;
	m66592_mdfy(m66592, 0, M66592_CURPIPE, ep->fifosel);
	m66592_write(m66592, M66592_ACLRM, ep->pipectr);
	m66592_write(m66592, 0, ep->pipectr);
	m66592_write(m66592, M66592_SQCLR, ep->pipectr);
	if (ep->use_dma) {
		m66592_mdfy(m66592, ep->pipenum, M66592_CURPIPE, ep->fifosel);
		ndelay(450);
		if (m66592->pdata->on_chip)
			mbw = M66592_MBW_32;
		else
			mbw = M66592_MBW_16;
		m66592_bset(m66592, mbw, ep->fifosel);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 104 | 82.54% | 1 | 50.00% | 
| Magnus Damm | 22 | 17.46% | 1 | 50.00% | 
| Total | 126 | 100.00% | 2 | 100.00% | 
static void m66592_ep_setting(struct m66592 *m66592, struct m66592_ep *ep,
		const struct usb_endpoint_descriptor *desc,
		u16 pipenum, int dma)
{
	if ((pipenum != 0) && dma) {
		if (m66592->num_dma == 0) {
			m66592->num_dma++;
			ep->use_dma = 1;
			ep->fifoaddr = M66592_D0FIFO;
			ep->fifosel = M66592_D0FIFOSEL;
			ep->fifoctr = M66592_D0FIFOCTR;
			ep->fifotrn = M66592_D0FIFOTRN;
		} else if (!m66592->pdata->on_chip && m66592->num_dma == 1) {
			m66592->num_dma++;
			ep->use_dma = 1;
			ep->fifoaddr = M66592_D1FIFO;
			ep->fifosel = M66592_D1FIFOSEL;
			ep->fifoctr = M66592_D1FIFOCTR;
			ep->fifotrn = M66592_D1FIFOTRN;
		} else {
			ep->use_dma = 0;
			ep->fifoaddr = M66592_CFIFO;
			ep->fifosel = M66592_CFIFOSEL;
			ep->fifoctr = M66592_CFIFOCTR;
			ep->fifotrn = 0;
		}
	} else {
		ep->use_dma = 0;
		ep->fifoaddr = M66592_CFIFO;
		ep->fifosel = M66592_CFIFOSEL;
		ep->fifoctr = M66592_CFIFOCTR;
		ep->fifotrn = 0;
	}
	ep->pipectr = get_pipectr_addr(pipenum);
	ep->pipenum = pipenum;
	ep->ep.maxpacket = usb_endpoint_maxp(desc);
	m66592->pipenum2ep[pipenum] = ep;
	m66592->epaddr2ep[desc->bEndpointAddress&USB_ENDPOINT_NUMBER_MASK] = ep;
	INIT_LIST_HEAD(&ep->queue);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 252 | 96.92% | 2 | 50.00% | 
| Magnus Damm | 7 | 2.69% | 1 | 25.00% | 
| Kuninori Morimoto | 1 | 0.38% | 1 | 25.00% | 
| Total | 260 | 100.00% | 4 | 100.00% | 
static void m66592_ep_release(struct m66592_ep *ep)
{
	struct m66592 *m66592 = ep->m66592;
	u16 pipenum = ep->pipenum;
	if (pipenum == 0)
		return;
	if (ep->use_dma)
		m66592->num_dma--;
	ep->pipenum = 0;
	ep->busy = 0;
	ep->use_dma = 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 63 | 100.00% | 1 | 100.00% | 
| Total | 63 | 100.00% | 1 | 100.00% | 
static int alloc_pipe_config(struct m66592_ep *ep,
		const struct usb_endpoint_descriptor *desc)
{
	struct m66592 *m66592 = ep->m66592;
	struct m66592_pipe_info info;
	int dma = 0;
	int *counter;
	int ret;
	ep->ep.desc = desc;
	BUG_ON(ep->pipenum);
	switch (desc->bmAttributes & USB_ENDPOINT_XFERTYPE_MASK) {
	case USB_ENDPOINT_XFER_BULK:
		if (m66592->bulk >= M66592_MAX_NUM_BULK) {
			if (m66592->isochronous >= M66592_MAX_NUM_ISOC) {
				pr_err("bulk pipe is insufficient\n");
				return -ENODEV;
			} else {
				info.pipe = M66592_BASE_PIPENUM_ISOC
						+ m66592->isochronous;
				counter = &m66592->isochronous;
			}
		} else {
			info.pipe = M66592_BASE_PIPENUM_BULK + m66592->bulk;
			counter = &m66592->bulk;
		}
		info.type = M66592_BULK;
		dma = 1;
		break;
	case USB_ENDPOINT_XFER_INT:
		if (m66592->interrupt >= M66592_MAX_NUM_INT) {
			pr_err("interrupt pipe is insufficient\n");
			return -ENODEV;
		}
		info.pipe = M66592_BASE_PIPENUM_INT + m66592->interrupt;
		info.type = M66592_INT;
		counter = &m66592->interrupt;
		break;
	case USB_ENDPOINT_XFER_ISOC:
		if (m66592->isochronous >= M66592_MAX_NUM_ISOC) {
			pr_err("isochronous pipe is insufficient\n");
			return -ENODEV;
		}
		info.pipe = M66592_BASE_PIPENUM_ISOC + m66592->isochronous;
		info.type = M66592_ISO;
		counter = &m66592->isochronous;
		break;
	default:
		pr_err("unexpect xfer type\n");
		return -EINVAL;
	}
	ep->type = info.type;
	info.epnum = desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK;
	info.maxpacket = usb_endpoint_maxp(desc);
	info.interval = desc->bInterval;
	if (desc->bEndpointAddress & USB_ENDPOINT_DIR_MASK)
		info.dir_in = 1;
	else
		info.dir_in = 0;
	ret = pipe_buffer_setting(m66592, &info);
	if (ret < 0) {
		pr_err("pipe_buffer_setting fail\n");
		return ret;
	}
	(*counter)++;
	if ((counter == &m66592->isochronous) && info.type == M66592_BULK)
		m66592->bulk++;
	m66592_ep_setting(m66592, ep, desc, info.pipe, dma);
	pipe_initialize(ep);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 377 | 97.92% | 2 | 40.00% | 
| David Brownell | 5 | 1.30% | 1 | 20.00% | 
| Ido Shayevitz | 2 | 0.52% | 1 | 20.00% | 
| Kuninori Morimoto | 1 | 0.26% | 1 | 20.00% | 
| Total | 385 | 100.00% | 5 | 100.00% | 
static int free_pipe_config(struct m66592_ep *ep)
{
	struct m66592 *m66592 = ep->m66592;
	struct m66592_pipe_info info;
	info.pipe = ep->pipenum;
	info.type = ep->type;
	pipe_buffer_release(m66592, &info);
	m66592_ep_release(ep);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 56 | 100.00% | 1 | 100.00% | 
| Total | 56 | 100.00% | 1 | 100.00% | 
/*-------------------------------------------------------------------------*/
static void pipe_irq_enable(struct m66592 *m66592, u16 pipenum)
{
	enable_irq_ready(m66592, pipenum);
	enable_irq_nrdy(m66592, pipenum);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 28 | 100.00% | 1 | 100.00% | 
| Total | 28 | 100.00% | 1 | 100.00% | 
static void pipe_irq_disable(struct m66592 *m66592, u16 pipenum)
{
	disable_irq_ready(m66592, pipenum);
	disable_irq_nrdy(m66592, pipenum);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 28 | 100.00% | 1 | 100.00% | 
| Total | 28 | 100.00% | 1 | 100.00% | 
/* if complete is true, gadget driver complete function is not call */
static void control_end(struct m66592 *m66592, unsigned ccpl)
{
	m66592->ep[0].internal_ccpl = ccpl;
	pipe_start(m66592, 0);
	m66592_bset(m66592, M66592_CCPL, M66592_DCPCTR);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 41 | 100.00% | 1 | 100.00% | 
| Total | 41 | 100.00% | 1 | 100.00% | 
static void start_ep0_write(struct m66592_ep *ep, struct m66592_request *req)
{
	struct m66592 *m66592 = ep->m66592;
	pipe_change(m66592, ep->pipenum);
	m66592_mdfy(m66592, M66592_ISEL | M66592_PIPE0,
			(M66592_ISEL | M66592_CURPIPE),
			M66592_CFIFOSEL);
	m66592_write(m66592, M66592_BCLR, ep->fifoctr);
	if (req->req.length == 0) {
		m66592_bset(m66592, M66592_BVAL, ep->fifoctr);
		pipe_start(m66592, 0);
		transfer_complete(ep, req, 0);
	} else {
		m66592_write(m66592, ~M66592_BEMP0, M66592_BEMPSTS);
		irq_ep0_write(ep, req);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 121 | 100.00% | 1 | 100.00% | 
| Total | 121 | 100.00% | 1 | 100.00% | 
static void start_packet_write(struct m66592_ep *ep, struct m66592_request *req)
{
	struct m66592 *m66592 = ep->m66592;
	u16 tmp;
	pipe_change(m66592, ep->pipenum);
	disable_irq_empty(m66592, ep->pipenum);
	pipe_start(m66592, ep->pipenum);
	tmp = m66592_read(m66592, ep->fifoctr);
	if (unlikely((tmp & M66592_FRDY) == 0))
		pipe_irq_enable(m66592, ep->pipenum);
	else
		irq_packet_write(ep, req);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 96 | 100.00% | 1 | 100.00% | 
| Total | 96 | 100.00% | 1 | 100.00% | 
static void start_packet_read(struct m66592_ep *ep, struct m66592_request *req)
{
	struct m66592 *m66592 = ep->m66592;
	u16 pipenum = ep->pipenum;
	if (ep->pipenum == 0) {
		m66592_mdfy(m66592, M66592_PIPE0,
				(M66592_ISEL | M66592_CURPIPE),
				M66592_CFIFOSEL);
		m66592_write(m66592, M66592_BCLR, ep->fifoctr);
		pipe_start(m66592, pipenum);
		pipe_irq_enable(m66592, pipenum);
	} else {
		if (ep->use_dma) {
			m66592_bset(m66592, M66592_TRCLR, ep->fifosel);
			pipe_change(m66592, pipenum);
			m66592_bset(m66592, M66592_TRENB, ep->fifosel);
			m66592_write(m66592,
				(req->req.length + ep->ep.maxpacket - 1)
					/ ep->ep.maxpacket,
				ep->fifotrn);
		}
		pipe_start(m66592, pipenum);	/* trigger once */
		pipe_irq_enable(m66592, pipenum);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 168 | 100.00% | 1 | 100.00% | 
| Total | 168 | 100.00% | 1 | 100.00% | 
static void start_packet(struct m66592_ep *ep, struct m66592_request *req)
{
	if (ep->ep.desc->bEndpointAddress & USB_DIR_IN)
		start_packet_write(ep, req);
	else
		start_packet_read(ep, req);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 41 | 95.35% | 1 | 50.00% | 
| Ido Shayevitz | 2 | 4.65% | 1 | 50.00% | 
| Total | 43 | 100.00% | 2 | 100.00% | 
static void start_ep0(struct m66592_ep *ep, struct m66592_request *req)
{
	u16 ctsq;
	ctsq = m66592_read(ep->m66592, M66592_INTSTS0) & M66592_CTSQ;
	switch (ctsq) {
	case M66592_CS_RDDS:
		start_ep0_write(ep, req);
		break;
	case M66592_CS_WRDS:
		start_packet_read(ep, req);
		break;
	case M66592_CS_WRND:
		control_end(ep->m66592, 0);
		break;
	default:
		pr_err("start_ep0: unexpect ctsq(%x)\n", ctsq);
		break;
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 81 | 98.78% | 1 | 50.00% | 
| David Brownell | 1 | 1.22% | 1 | 50.00% | 
| Total | 82 | 100.00% | 2 | 100.00% | 
static void init_controller(struct m66592 *m66592)
{
	unsigned int endian;
	if (m66592->pdata->on_chip) {
		if (m66592->pdata->endian)
			endian = 0; /* big endian */
		else
			endian = M66592_LITTLE; /* little endian */
		m66592_bset(m66592, M66592_HSE, M66592_SYSCFG);	/* High spd */
		m66592_bclr(m66592, M66592_USBE, M66592_SYSCFG);
		m66592_bclr(m66592, M66592_DPRPU, M66592_SYSCFG);
		m66592_bset(m66592, M66592_USBE, M66592_SYSCFG);
		/* This is a workaound for SH7722 2nd cut */
		m66592_bset(m66592, 0x8000, M66592_DVSTCTR);
		m66592_bset(m66592, 0x1000, M66592_TESTMODE);
		m66592_bclr(m66592, 0x8000, M66592_DVSTCTR);
		m66592_bset(m66592, M66592_INTL, M66592_INTENB1);
		m66592_write(m66592, 0, M66592_CFBCFG);
		m66592_write(m66592, 0, M66592_D0FBCFG);
		m66592_bset(m66592, endian, M66592_CFBCFG);
		m66592_bset(m66592, endian, M66592_D0FBCFG);
	} else {
		unsigned int clock, vif, irq_sense;
		if (m66592->pdata->endian)
			endian = M66592_BIGEND; /* big endian */
		else
			endian = 0; /* little endian */
		if (m66592->pdata->vif)
			vif = M66592_LDRV; /* 3.3v */
		else
			vif = 0; /* 1.5v */
		switch (m66592->pdata->xtal) {
		case M66592_PLATDATA_XTAL_12MHZ:
			clock = M66592_XTAL12;
			break;
		case M66592_PLATDATA_XTAL_24MHZ:
			clock = M66592_XTAL24;
			break;
		case M66592_PLATDATA_XTAL_48MHZ:
			clock = M66592_XTAL48;
			break;
		default:
			pr_warn("m66592-udc: xtal configuration error\n");
			clock = 0;
		}
		switch (m66592->irq_trigger) {
		case IRQF_TRIGGER_LOW:
			irq_sense = M66592_INTL;
			break;
		case IRQF_TRIGGER_FALLING:
			irq_sense = 0;
			break;
		default:
			pr_warn("m66592-udc: irq trigger config error\n");
			irq_sense = 0;
		}
		m66592_bset(m66592,
			    (vif & M66592_LDRV) | (endian & M66592_BIGEND),
			    M66592_PINCFG);
		m66592_bset(m66592, M66592_HSE, M66592_SYSCFG);	/* High spd */
		m66592_mdfy(m66592, clock & M66592_XTAL, M66592_XTAL,
			    M66592_SYSCFG);
		m66592_bclr(m66592, M66592_USBE, M66592_SYSCFG);
		m66592_bclr(m66592, M66592_DPRPU, M66592_SYSCFG);
		m66592_bset(m66592, M66592_USBE, M66592_SYSCFG);
		m66592_bset(m66592, M66592_XCKE, M66592_SYSCFG);
		msleep(3);
		m66592_bset(m66592, M66592_RCKE | M66592_PLLC, M66592_SYSCFG);
		msleep(1);
		m66592_bset(m66592, M66592_SCKE, M66592_SYSCFG);
		m66592_bset(m66592, irq_sense & M66592_INTL, M66592_INTENB1);
		m66592_write(m66592, M66592_BURST | M66592_CPU_ADR_RD_WR,
			     M66592_DMA0CFG);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 255 | 62.04% | 2 | 50.00% | 
| Magnus Damm | 154 | 37.47% | 1 | 25.00% | 
| Joe Perches | 2 | 0.49% | 1 | 25.00% | 
| Total | 411 | 100.00% | 4 | 100.00% | 
static void disable_controller(struct m66592 *m66592)
{
	m66592_bclr(m66592, M66592_UTST, M66592_TESTMODE);
	if (!m66592->pdata->on_chip) {
		m66592_bclr(m66592, M66592_SCKE, M66592_SYSCFG);
		udelay(1);
		m66592_bclr(m66592, M66592_PLLC, M66592_SYSCFG);
		udelay(1);
		m66592_bclr(m66592, M66592_RCKE, M66592_SYSCFG);
		udelay(1);
		m66592_bclr(m66592, M66592_XCKE, M66592_SYSCFG);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 71 | 86.59% | 2 | 50.00% | 
| Magnus Damm | 11 | 13.41% | 2 | 50.00% | 
| Total | 82 | 100.00% | 4 | 100.00% | 
static void m66592_start_xclock(struct m66592 *m66592)
{
	u16 tmp;
	if (!m66592->pdata->on_chip) {
		tmp = m66592_read(m66592, M66592_SYSCFG);
		if (!(tmp & M66592_XCKE))
			m66592_bset(m66592, M66592_XCKE, M66592_SYSCFG);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 41 | 78.85% | 1 | 50.00% | 
| Magnus Damm | 11 | 21.15% | 1 | 50.00% | 
| Total | 52 | 100.00% | 2 | 100.00% | 
/*-------------------------------------------------------------------------*/
static void transfer_complete(struct m66592_ep *ep,
		struct m66592_request *req, int status)
__releases(m66592->lock)
__acquires(m66592->lock)
{
	int restart = 0;
	if (unlikely(ep->pipenum == 0)) {
		if (ep->internal_ccpl) {
			ep->internal_ccpl = 0;
			return;
		}
	}
	list_del_init(&req->queue);
	if (ep->m66592->gadget.speed == USB_SPEED_UNKNOWN)
		req->req.status = -ESHUTDOWN;
	else
		req->req.status = status;
	if (!list_empty(&ep->queue))
		restart = 1;
	spin_unlock(&ep->m66592->lock);
	usb_gadget_giveback_request(&ep->ep, &req->req);
	spin_lock(&ep->m66592->lock);
	if (restart) {
		req = list_entry(ep->queue.next, struct m66592_request, queue);
		if (ep->ep.desc)
			start_packet(ep, req);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 184 | 98.40% | 2 | 50.00% | 
| Ido Shayevitz | 2 | 1.07% | 1 | 25.00% | 
| Michal Sojka | 1 | 0.53% | 1 | 25.00% | 
| Total | 187 | 100.00% | 4 | 100.00% | 
static void irq_ep0_write(struct m66592_ep *ep, struct m66592_request *req)
{
	int i;
	u16 tmp;
	unsigned bufsize;
	size_t size;
	void *buf;
	u16 pipenum = ep->pipenum;
	struct m66592 *m66592 = ep->m66592;
	pipe_change(m66592, pipenum);
	m66592_bset(m66592, M66592_ISEL, ep->fifosel);
	i = 0;
	do {
		tmp = m66592_read(m66592, ep->fifoctr);
		if (i++ > 100000) {
			pr_err("pipe0 is busy. maybe cpu i/o bus "
				"conflict. please power off this controller.");
			return;
		}
		ndelay(1);
	} while ((tmp & M66592_FRDY) == 0);
	/* prepare parameters */
	bufsize = get_buffer_size(m66592, pipenum);
	buf = req->req.buf + req->req.actual;
	size = min(bufsize, req->req.length - req->req.actual);
	/* write fifo */
	if (req->req.buf) {
		if (size > 0)
			m66592_write_fifo(m66592, ep, buf, size);
		if ((size == 0) || ((size % ep->ep.maxpacket) != 0))
			m66592_bset(m66592, M66592_BVAL, ep->fifoctr);
	}
	/* update parameters */
	req->req.actual += size;
	/* check transfer finish */
	if ((!req->req.zero && (req->req.actual == req->req.length))
			|| (size % ep->ep.maxpacket)
			|| (size == 0)) {
		disable_irq_ready(m66592, pipenum);
		disable_irq_empty(m66592, pipenum);
	} else {
		disable_irq_ready(m66592, pipenum);
		enable_irq_empty(m66592, pipenum);
	}
	pipe_start(m66592, pipenum);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 309 | 99.36% | 1 | 33.33% | 
| David Brownell | 1 | 0.32% | 1 | 33.33% | 
| Joe Perches | 1 | 0.32% | 1 | 33.33% | 
| Total | 311 | 100.00% | 3 | 100.00% | 
static void irq_packet_write(struct m66592_ep *ep, struct m66592_request *req)
{
	u16 tmp;
	unsigned bufsize;
	size_t size;
	void *buf;
	u16 pipenum = ep->pipenum;
	struct m66592 *m66592 = ep->m66592;
	pipe_change(m66592, pipenum);
	tmp = m66592_read(m66592, ep->fifoctr);
	if (unlikely((tmp & M66592_FRDY) == 0)) {
		pipe_stop(m66592, pipenum);
		pipe_irq_disable(m66592, pipenum);
		pr_err("write fifo not ready. pipnum=%d\n", pipenum);
		return;
	}
	/* prepare parameters */
	bufsize = get_buffer_size(m66592, pipenum);
	buf = req->req.buf + req->req.actual;
	size = min(bufsize, req->req.length - req->req.actual);
	/* write fifo */
	if (req->req.buf) {
		m66592_write_fifo(m66592, ep, buf, size);
		if ((size == 0)
				|| ((size % ep->ep.maxpacket) != 0)
				|| ((bufsize != ep->ep.maxpacket)
					&& (bufsize > size)))
			m66592_bset(m66592, M66592_BVAL, ep->fifoctr);
	}
	/* update parameters */
	req->req.actual += size;
	/* check transfer finish */
	if ((!req->req.zero && (req->req.actual == req->req.length))
			|| (size % ep->ep.maxpacket)
			|| (size == 0)) {
		disable_irq_ready(m66592, pipenum);
		enable_irq_empty(m66592, pipenum);
	} else {
		disable_irq_empty(m66592, pipenum);
		pipe_irq_enable(m66592, pipenum);
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 299 | 99.67% | 1 | 50.00% | 
| David Brownell | 1 | 0.33% | 1 | 50.00% | 
| Total | 300 | 100.00% | 2 | 100.00% | 
static void irq_packet_read(struct m66592_ep *ep, struct m66592_request *req)
{
	u16 tmp;
	int rcv_len, bufsize, req_len;
	int size;
	void *buf;
	u16 pipenum = ep->pipenum;
	struct m66592 *m66592 = ep->m66592;
	int finish = 0;
	pipe_change(m66592, pipenum);
	tmp = m66592_read(m66592, ep->fifoctr);
	if (unlikely((tmp & M66592_FRDY) == 0)) {
		req->req.status = -EPIPE;
		pipe_stop(m66592, pipenum);
		pipe_irq_disable(m66592, pipenum);
		pr_err("read fifo not ready");
		return;
	}
	/* prepare parameters */
	rcv_len = tmp & M66592_DTLN;
	bufsize = get_buffer_size(m66592, pipenum);
	buf = req->req.buf + req->req.actual;
	req_len = req->req.length - req->req.actual;
	if (rcv_len < bufsize)
		size = min(rcv_len, req_len);
	else
		size = min(bufsize, req_len);
	/* update parameters */
	req->req.actual += size;
	/* check transfer finish */
	if ((!req->req.zero && (req->req.actual == req->req.length))
			|| (size % ep->ep.maxpacket)
			|| (size == 0)) {
		pipe_stop(m66592, pipenum);
		pipe_irq_disable(m66592, pipenum);
		finish = 1;
	}
	/* read fifo */
	if (req->req.buf) {
		if (size == 0)
			m66592_write(m66592, M66592_BCLR, ep->fifoctr);
		else
			m66592_read_fifo(m66592, ep->fifoaddr, buf, size);
	}
	if ((ep->pipenum != 0) && finish)
		transfer_complete(ep, req, 0);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 318 | 99.69% | 1 | 50.00% | 
| David Brownell | 1 | 0.31% | 1 | 50.00% | 
| Total | 319 | 100.00% | 2 | 100.00% | 
static void irq_pipe_ready(struct m66592 *m66592, u16 status, u16 enb)
{
	u16 check;
	u16 pipenum;
	struct m66592_ep *ep;
	struct m66592_request *req;
	if ((status & M66592_BRDY0) && (enb & M66592_BRDY0)) {
		m66592_write(m66592, ~M66592_BRDY0, M66592_BRDYSTS);
		m66592_mdfy(m66592, M66592_PIPE0, M66592_CURPIPE,
				M66592_CFIFOSEL);
		ep = &m66592->ep[0];
		req = list_entry(ep->queue.next, struct m66592_request, queue);
		irq_packet_read(ep, req);
	} else {
		for (pipenum = 1; pipenum < M66592_MAX_NUM_PIPE; pipenum++) {
			check = 1 << pipenum;
			if ((status & check) && (enb & check)) {
				m66592_write(m66592, ~check, M66592_BRDYSTS);
				ep = m66592->pipenum2ep[pipenum];
				req = list_entry(ep->queue.next,
						 struct m66592_request, queue);
				if (ep->ep.desc->bEndpointAddress & USB_DIR_IN)
					irq_packet_write(ep, req);
				else
					irq_packet_read(ep, req);
			}
		}
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 203 | 99.02% | 1 | 50.00% | 
| Ido Shayevitz | 2 | 0.98% | 1 | 50.00% | 
| Total | 205 | 100.00% | 2 | 100.00% | 
static void irq_pipe_empty(struct m66592 *m66592, u16 status, u16 enb)
{
	u16 tmp;
	u16 check;
	u16 pipenum;
	struct m66592_ep *ep;
	struct m66592_request *req;
	if ((status & M66592_BEMP0) && (enb & M66592_BEMP0)) {
		m66592_write(m66592, ~M66592_BEMP0, M66592_BEMPSTS);
		ep = &m66592->ep[0];
		req = list_entry(ep->queue.next, struct m66592_request, queue);
		irq_ep0_write(ep, req);
	} else {
		for (pipenum = 1; pipenum < M66592_MAX_NUM_PIPE; pipenum++) {
			check = 1 << pipenum;
			if ((status & check) && (enb & check)) {
				m66592_write(m66592, ~check, M66592_BEMPSTS);
				tmp = control_reg_get(m66592, pipenum);
				if ((tmp & M66592_INBUFM) == 0) {
					disable_irq_empty(m66592, pipenum);
					pipe_irq_disable(m66592, pipenum);
					pipe_stop(m66592, pipenum);
					ep = m66592->pipenum2ep[pipenum];
					req = list_entry(ep->queue.next,
							 struct m66592_request,
							 queue);
					if (!list_empty(&ep->queue))
						transfer_complete(ep, req, 0);
				}
			}
		}
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 232 | 100.00% | 1 | 100.00% | 
| Total | 232 | 100.00% | 1 | 100.00% | 
static void get_status(struct m66592 *m66592, struct usb_ctrlrequest *ctrl)
__releases(m66592->lock)
__acquires(m66592->lock)
{
	struct m66592_ep *ep;
	u16 pid;
	u16 status = 0;
	u16 w_index = le16_to_cpu(ctrl->wIndex);
	switch (ctrl->bRequestType & USB_RECIP_MASK) {
	case USB_RECIP_DEVICE:
		status = 1 << USB_DEVICE_SELF_POWERED;
		break;
	case USB_RECIP_INTERFACE:
		status = 0;
		break;
	case USB_RECIP_ENDPOINT:
		ep = m66592->epaddr2ep[w_index & USB_ENDPOINT_NUMBER_MASK];
		pid = control_reg_get_pid(m66592, ep->pipenum);
		if (pid == M66592_PID_STALL)
			status = 1 << USB_ENDPOINT_HALT;
		else
			status = 0;
		break;
	default:
		pipe_stall(m66592, 0);
		return;		/* exit */
	}
	m66592->ep0_data = cpu_to_le16(status);
	m66592->ep0_req->buf = &m66592->ep0_data;
	m66592->ep0_req->length = 2;
	/* AV: what happens if we get called again before that gets through? */
	spin_unlock(&m66592->lock);
	m66592_queue(m66592->gadget.ep0, m66592->ep0_req, GFP_KERNEL);
	spin_lock(&m66592->lock);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 191 | 99.48% | 2 | 66.67% | 
| Al Viro | 1 | 0.52% | 1 | 33.33% | 
| Total | 192 | 100.00% | 3 | 100.00% | 
static void clear_feature(struct m66592 *m66592, struct usb_ctrlrequest *ctrl)
{
	switch (ctrl->bRequestType & USB_RECIP_MASK) {
	case USB_RECIP_DEVICE:
		control_end(m66592, 1);
		break;
	case USB_RECIP_INTERFACE:
		control_end(m66592, 1);
		break;
	case USB_RECIP_ENDPOINT: {
		struct m66592_ep *ep;
		struct m66592_request *req;
		u16 w_index = le16_to_cpu(ctrl->wIndex);
		ep = m66592->epaddr2ep[w_index & USB_ENDPOINT_NUMBER_MASK];
		pipe_stop(m66592, ep->pipenum);
		control_reg_sqclr(m66592, ep->pipenum);
		control_end(m66592, 1);
		req = list_entry(ep->queue.next,
		struct m66592_request, queue);
		if (ep->busy) {
			ep->busy = 0;
			if (list_empty(&ep->queue))
				break;
			start_packet(ep, req);
		} else if (!list_empty(&ep->queue))
			pipe_start(m66592, ep->pipenum);
		}
		break;
	default:
		pipe_stall(m66592, 0);
		break;
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 188 | 100.00% | 2 | 100.00% | 
| Total | 188 | 100.00% | 2 | 100.00% | 
static void set_feature(struct m66592 *m66592, struct usb_ctrlrequest *ctrl)
{
	u16 tmp;
	int timeout = 3000;
	switch (ctrl->bRequestType & USB_RECIP_MASK) {
	case USB_RECIP_DEVICE:
		switch (le16_to_cpu(ctrl->wValue)) {
		case USB_DEVICE_TEST_MODE:
			control_end(m66592, 1);
			/* Wait for the completion of status stage */
			do {
				tmp = m66592_read(m66592, M66592_INTSTS0) &
								M66592_CTSQ;
				udelay(1);
			} while (tmp != M66592_CS_IDST && timeout-- > 0);
			if (tmp == M66592_CS_IDST)
				m66592_bset(m66592,
					    le16_to_cpu(ctrl->wIndex >> 8),
					    M66592_TESTMODE);
			break;
		default:
			pipe_stall(m66592, 0);
			break;
		}
		break;
	case USB_RECIP_INTERFACE:
		control_end(m66592, 1);
		break;
	case USB_RECIP_ENDPOINT: {
		struct m66592_ep *ep;
		u16 w_index = le16_to_cpu(ctrl->wIndex);
		ep = m66592->epaddr2ep[w_index & USB_ENDPOINT_NUMBER_MASK];
		pipe_stall(m66592, ep->pipenum);
		control_end(m66592, 1);
		}
		break;
	default:
		pipe_stall(m66592, 0);
		break;
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 190 | 99.48% | 3 | 75.00% | 
| Dan Carpenter | 1 | 0.52% | 1 | 25.00% | 
| Total | 191 | 100.00% | 4 | 100.00% | 
/* if return value is true, call class driver's setup() */
static int setup_packet(struct m66592 *m66592, struct usb_ctrlrequest *ctrl)
{
	u16 *p = (u16 *)ctrl;
	unsigned long offset = M66592_USBREQ;
	int i, ret = 0;
	/* read fifo */
	m66592_write(m66592, ~M66592_VALID, M66592_INTSTS0);
	for (i = 0; i < 4; i++)
		p[i] = m66592_read(m66592, offset + i*2);
	/* check request */
	if ((ctrl->bRequestType & USB_TYPE_MASK) == USB_TYPE_STANDARD) {
		switch (ctrl->bRequest) {
		case USB_REQ_GET_STATUS:
			get_status(m66592, ctrl);
			break;
		case USB_REQ_CLEAR_FEATURE:
			clear_feature(m66592, ctrl);
			break;
		case USB_REQ_SET_FEATURE:
			set_feature(m66592, ctrl);
			break;
		default:
			ret = 1;
			break;
		}
	} else
		ret = 1;
	return ret;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 149 | 100.00% | 1 | 100.00% | 
| Total | 149 | 100.00% | 1 | 100.00% | 
static void m66592_update_usb_speed(struct m66592 *m66592)
{
	u16 speed = get_usb_speed(m66592);
	switch (speed) {
	case M66592_HSMODE:
		m66592->gadget.speed = USB_SPEED_HIGH;
		break;
	case M66592_FSMODE:
		m66592->gadget.speed = USB_SPEED_FULL;
		break;
	default:
		m66592->gadget.speed = USB_SPEED_UNKNOWN;
		pr_err("USB speed unknown\n");
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 62 | 98.41% | 1 | 50.00% | 
| David Brownell | 1 | 1.59% | 1 | 50.00% | 
| Total | 63 | 100.00% | 2 | 100.00% | 
static void irq_device_state(struct m66592 *m66592)
{
	u16 dvsq;
	dvsq = m66592_read(m66592, M66592_INTSTS0) & M66592_DVSQ;
	m66592_write(m66592, ~M66592_DVST, M66592_INTSTS0);
	if (dvsq == M66592_DS_DFLT) {	/* bus reset */
		usb_gadget_udc_reset(&m66592->gadget, m66592->driver);
		m66592_update_usb_speed(m66592);
	}
	if (m66592->old_dvsq == M66592_DS_CNFG && dvsq != M66592_DS_CNFG)
		m66592_update_usb_speed(m66592);
	if ((dvsq == M66592_DS_CNFG || dvsq == M66592_DS_ADDS)
			&& m66592->gadget.speed == USB_SPEED_UNKNOWN)
		m66592_update_usb_speed(m66592);
	m66592->old_dvsq = dvsq;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 104 | 95.41% | 1 | 50.00% | 
| Peter Chen | 5 | 4.59% | 1 | 50.00% | 
| Total | 109 | 100.00% | 2 | 100.00% | 
static void irq_control_stage(struct m66592 *m66592)
__releases(m66592->lock)
__acquires(m66592->lock)
{
	struct usb_ctrlrequest ctrl;
	u16 ctsq;
	ctsq = m66592_read(m66592, M66592_INTSTS0) & M66592_CTSQ;
	m66592_write(m66592, ~M66592_CTRT, M66592_INTSTS0);
	switch (ctsq) {
	case M66592_CS_IDST: {
		struct m66592_ep *ep;
		struct m66592_request *req;
		ep = &m66592->ep[0];
		req = list_entry(ep->queue.next, struct m66592_request, queue);
		transfer_complete(ep, req, 0);
		}
		break;
	case M66592_CS_RDDS:
	case M66592_CS_WRDS:
	case M66592_CS_WRND:
		if (setup_packet(m66592, &ctrl)) {
			spin_unlock(&m66592->lock);
			if (m66592->driver->setup(&m66592->gadget, &ctrl) < 0)
				pipe_stall(m66592, 0);
			spin_lock(&m66592->lock);
		}
		break;
	case M66592_CS_RDSS:
	case M66592_CS_WRSS:
		control_end(m66592, 0);
		break;
	default:
		pr_err("ctrl_stage: unexpect ctsq(%x)\n", ctsq);
		break;
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 194 | 99.49% | 2 | 66.67% | 
| David Brownell | 1 | 0.51% | 1 | 33.33% | 
| Total | 195 | 100.00% | 3 | 100.00% | 
static irqreturn_t m66592_irq(int irq, void *_m66592)
{
	struct m66592 *m66592 = _m66592;
	u16 intsts0;
	u16 intenb0;
	u16 savepipe;
	u16 mask0;
	spin_lock(&m66592->lock);
	intsts0 = m66592_read(m66592, M66592_INTSTS0);
	intenb0 = m66592_read(m66592, M66592_INTENB0);
	if (m66592->pdata->on_chip && !intsts0 && !intenb0) {
		/*
                 * When USB clock stops, it cannot read register. Even if a
                 * clock stops, the interrupt occurs. So this driver turn on
                 * a clock by this timing and do re-reading of register.
                 */
		m66592_start_xclock(m66592);
		intsts0 = m66592_read(m66592, M66592_INTSTS0);
		intenb0 = m66592_read(m66592, M66592_INTENB0);
	}
	savepipe = m66592_read(m66592, M66592_CFIFOSEL);
	mask0 = intsts0 & intenb0;
	if (mask0) {
		u16 brdysts = m66592_read(m66592, M66592_BRDYSTS);
		u16 bempsts = m66592_read(m66592, M66592_BEMPSTS);
		u16 brdyenb = m66592_read(m66592, M66592_BRDYENB);
		u16 bempenb = m66592_read(m66592, M66592_BEMPENB);
		if (mask0 & M66592_VBINT) {
			m66592_write(m66592,  0xffff & ~M66592_VBINT,
					M66592_INTSTS0);
			m66592_start_xclock(m66592);
			/* start vbus sampling */
			m66592->old_vbus = m66592_read(m66592, M66592_INTSTS0)
					& M66592_VBSTS;
			m66592->scount = M66592_MAX_SAMPLING;
			mod_timer(&m66592->timer,
					jiffies + msecs_to_jiffies(50));
		}
		if (intsts0 & M66592_DVSQ)
			irq_device_state(m66592);
		if ((intsts0 & M66592_BRDY) && (intenb0 & M66592_BRDYE)
				&& (brdysts & brdyenb)) {
			irq_pipe_ready(m66592, brdysts, brdyenb);
		}
		if ((intsts0 & M66592_BEMP) && (intenb0 & M66592_BEMPE)
				&& (bempsts & bempenb)) {
			irq_pipe_empty(m66592, bempsts, bempenb);
		}
		if (intsts0 & M66592_CTRT)
			irq_control_stage(m66592);
	}
	m66592_write(m66592, savepipe, M66592_CFIFOSEL);
	spin_unlock(&m66592->lock);
	return IRQ_HANDLED;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 305 | 94.43% | 3 | 60.00% | 
| Michal Nazarewicz | 12 | 3.72% | 1 | 20.00% | 
| Magnus Damm | 6 | 1.86% | 1 | 20.00% | 
| Total | 323 | 100.00% | 5 | 100.00% | 
static void m66592_timer(struct timer_list *t)
{
	struct m66592 *m66592 = from_timer(m66592, t, timer);
	unsigned long flags;
	u16 tmp;
	spin_lock_irqsave(&m66592->lock, flags);
	tmp = m66592_read(m66592, M66592_SYSCFG);
	if (!(tmp & M66592_RCKE)) {
		m66592_bset(m66592, M66592_RCKE | M66592_PLLC, M66592_SYSCFG);
		udelay(10);
		m66592_bset(m66592, M66592_SCKE, M66592_SYSCFG);
	}
	if (m66592->scount > 0) {
		tmp = m66592_read(m66592, M66592_INTSTS0) & M66592_VBSTS;
		if (tmp == m66592->old_vbus) {
			m66592->scount--;
			if (m66592->scount == 0) {
				if (tmp == M66592_VBSTS)
					m66592_usb_connect(m66592);
				else
					m66592_usb_disconnect(m66592);
			} else {
				mod_timer(&m66592->timer,
					jiffies + msecs_to_jiffies(50));
			}
		} else {
			m66592->scount = M66592_MAX_SAMPLING;
			m66592->old_vbus = tmp;
			mod_timer(&m66592->timer,
					jiffies + msecs_to_jiffies(50));
		}
	}
	spin_unlock_irqrestore(&m66592->lock, flags);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 197 | 94.71% | 1 | 50.00% | 
| Kees Cook | 11 | 5.29% | 1 | 50.00% | 
| Total | 208 | 100.00% | 2 | 100.00% | 
/*-------------------------------------------------------------------------*/
static int m66592_enable(struct usb_ep *_ep,
			 const struct usb_endpoint_descriptor *desc)
{
	struct m66592_ep *ep;
	ep = container_of(_ep, struct m66592_ep, ep);
	return alloc_pipe_config(ep, desc);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 42 | 100.00% | 1 | 100.00% | 
| Total | 42 | 100.00% | 1 | 100.00% | 
static int m66592_disable(struct usb_ep *_ep)
{
	struct m66592_ep *ep;
	struct m66592_request *req;
	unsigned long flags;
	ep = container_of(_ep, struct m66592_ep, ep);
	BUG_ON(!ep);
	while (!list_empty(&ep->queue)) {
		req = list_entry(ep->queue.next, struct m66592_request, queue);
		spin_lock_irqsave(&ep->m66592->lock, flags);
		transfer_complete(ep, req, -ECONNRESET);
		spin_unlock_irqrestore(&ep->m66592->lock, flags);
	}
	pipe_irq_disable(ep->m66592, ep->pipenum);
	return free_pipe_config(ep);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 123 | 100.00% | 1 | 100.00% | 
| Total | 123 | 100.00% | 1 | 100.00% | 
static struct usb_request *m66592_alloc_request(struct usb_ep *_ep,
						gfp_t gfp_flags)
{
	struct m66592_request *req;
	req = kzalloc(sizeof(struct m66592_request), gfp_flags);
	if (!req)
		return NULL;
	INIT_LIST_HEAD(&req->queue);
	return &req->req;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 56 | 100.00% | 1 | 100.00% | 
| Total | 56 | 100.00% | 1 | 100.00% | 
static void m66592_free_request(struct usb_ep *_ep, struct usb_request *_req)
{
	struct m66592_request *req;
	req = container_of(_req, struct m66592_request, req);
	kfree(req);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 38 | 100.00% | 1 | 100.00% | 
| Total | 38 | 100.00% | 1 | 100.00% | 
static int m66592_queue(struct usb_ep *_ep, struct usb_request *_req,
			gfp_t gfp_flags)
{
	struct m66592_ep *ep;
	struct m66592_request *req;
	unsigned long flags;
	int request = 0;
	ep = container_of(_ep, struct m66592_ep, ep);
	req = container_of(_req, struct m66592_request, req);
	if (ep->m66592->gadget.speed == USB_SPEED_UNKNOWN)
		return -ESHUTDOWN;
	spin_lock_irqsave(&ep->m66592->lock, flags);
	if (list_empty(&ep->queue))
		request = 1;
	list_add_tail(&req->queue, &ep->queue);
	req->req.actual = 0;
	req->req.status = -EINPROGRESS;
	if (ep->ep.desc == NULL)	/* control */
		start_ep0(ep, req);
	else {
		if (request && !ep->busy)
			start_packet(ep, req);
	}
	spin_unlock_irqrestore(&ep->m66592->lock, flags);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 183 | 98.39% | 1 | 33.33% | 
| Ido Shayevitz | 2 | 1.08% | 1 | 33.33% | 
| David Brownell | 1 | 0.54% | 1 | 33.33% | 
| Total | 186 | 100.00% | 3 | 100.00% | 
static int m66592_dequeue(struct usb_ep *_ep, struct usb_request *_req)
{
	struct m66592_ep *ep;
	struct m66592_request *req;
	unsigned long flags;
	ep = container_of(_ep, struct m66592_ep, ep);
	req = container_of(_req, struct m66592_request, req);
	spin_lock_irqsave(&ep->m66592->lock, flags);
	if (!list_empty(&ep->queue))
		transfer_complete(ep, req, -ECONNRESET);
	spin_unlock_irqrestore(&ep->m66592->lock, flags);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 102 | 100.00% | 1 | 100.00% | 
| Total | 102 | 100.00% | 1 | 100.00% | 
static int m66592_set_halt(struct usb_ep *_ep, int value)
{
	struct m66592_ep *ep = container_of(_ep, struct m66592_ep, ep);
	unsigned long flags;
	int ret = 0;
	spin_lock_irqsave(&ep->m66592->lock, flags);
	if (!list_empty(&ep->queue)) {
		ret = -EAGAIN;
	} else if (value) {
		ep->busy = 1;
		pipe_stall(ep->m66592, ep->pipenum);
	} else {
		ep->busy = 0;
		pipe_stop(ep->m66592, ep->pipenum);
	}
	spin_unlock_irqrestore(&ep->m66592->lock, flags);
	return ret;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 116 | 91.34% | 1 | 50.00% | 
| Michal Nazarewicz | 11 | 8.66% | 1 | 50.00% | 
| Total | 127 | 100.00% | 2 | 100.00% | 
static void m66592_fifo_flush(struct usb_ep *_ep)
{
	struct m66592_ep *ep;
	unsigned long flags;
	ep = container_of(_ep, struct m66592_ep, ep);
	spin_lock_irqsave(&ep->m66592->lock, flags);
	if (list_empty(&ep->queue) && !ep->busy) {
		pipe_stop(ep->m66592, ep->pipenum);
		m66592_bclr(ep->m66592, M66592_BCLR, ep->fifoctr);
	}
	spin_unlock_irqrestore(&ep->m66592->lock, flags);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 97 | 100.00% | 1 | 100.00% | 
| Total | 97 | 100.00% | 1 | 100.00% | 
static const struct usb_ep_ops m66592_ep_ops = {
	.enable		= m66592_enable,
	.disable	= m66592_disable,
	.alloc_request	= m66592_alloc_request,
	.free_request	= m66592_free_request,
	.queue		= m66592_queue,
	.dequeue	= m66592_dequeue,
	.set_halt	= m66592_set_halt,
	.fifo_flush	= m66592_fifo_flush,
};
/*-------------------------------------------------------------------------*/
static int m66592_udc_start(struct usb_gadget *g,
		struct usb_gadget_driver *driver)
{
	struct m66592 *m66592 = to_m66592(g);
	/* hook up the driver */
	driver->driver.bus = NULL;
	m66592->driver = driver;
	m66592_bset(m66592, M66592_VBSE | M66592_URST, M66592_INTENB0);
	if (m66592_read(m66592, M66592_INTSTS0) & M66592_VBSTS) {
		m66592_start_xclock(m66592);
		/* start vbus sampling */
		m66592->old_vbus = m66592_read(m66592,
					 M66592_INTSTS0) & M66592_VBSTS;
		m66592->scount = M66592_MAX_SAMPLING;
		mod_timer(&m66592->timer, jiffies + msecs_to_jiffies(50));
	}
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 92 | 85.19% | 1 | 20.00% | 
| Felipe Balbi | 7 | 6.48% | 1 | 20.00% | 
| Sebastian Andrzej Siewior | 5 | 4.63% | 2 | 40.00% | 
| Uwe Kleine-König | 4 | 3.70% | 1 | 20.00% | 
| Total | 108 | 100.00% | 5 | 100.00% | 
static int m66592_udc_stop(struct usb_gadget *g)
{
	struct m66592 *m66592 = to_m66592(g);
	m66592_bclr(m66592, M66592_VBSE | M66592_URST, M66592_INTENB0);
	init_controller(m66592);
	disable_controller(m66592);
	m66592->driver = NULL;
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 43 | 84.31% | 1 | 33.33% | 
| Felipe Balbi | 7 | 13.73% | 1 | 33.33% | 
| Sebastian Andrzej Siewior | 1 | 1.96% | 1 | 33.33% | 
| Total | 51 | 100.00% | 3 | 100.00% | 
/*-------------------------------------------------------------------------*/
static int m66592_get_frame(struct usb_gadget *_gadget)
{
	struct m66592 *m66592 = gadget_to_m66592(_gadget);
	return m66592_read(m66592, M66592_FRMNUM) & 0x03FF;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 31 | 100.00% | 1 | 100.00% | 
| Total | 31 | 100.00% | 1 | 100.00% | 
static int m66592_pullup(struct usb_gadget *gadget, int is_on)
{
	struct m66592 *m66592 = gadget_to_m66592(gadget);
	unsigned long flags;
	spin_lock_irqsave(&m66592->lock, flags);
	if (is_on)
		m66592_bset(m66592, M66592_DPRPU, M66592_SYSCFG);
	else
		m66592_bclr(m66592, M66592_DPRPU, M66592_SYSCFG);
	spin_unlock_irqrestore(&m66592->lock, flags);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 74 | 100.00% | 1 | 100.00% | 
| Total | 74 | 100.00% | 1 | 100.00% | 
static const struct usb_gadget_ops m66592_gadget_ops = {
	.get_frame		= m66592_get_frame,
	.udc_start		= m66592_udc_start,
	.udc_stop		= m66592_udc_stop,
	.pullup			= m66592_pullup,
};
static int m66592_remove(struct platform_device *pdev)
{
	struct m66592		*m66592 = platform_get_drvdata(pdev);
	usb_del_gadget_udc(&m66592->gadget);
	del_timer_sync(&m66592->timer);
	iounmap(m66592->reg);
	free_irq(platform_get_irq(pdev, 0), m66592);
	m66592_free_request(&m66592->ep[0].ep, m66592->ep0_req);
	if (m66592->pdata->on_chip) {
		clk_disable(m66592->clk);
		clk_put(m66592->clk);
	}
	kfree(m66592);
	return 0;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 73 | 69.52% | 3 | 42.86% | 
| Magnus Damm | 23 | 21.90% | 2 | 28.57% | 
| Sebastian Andrzej Siewior | 8 | 7.62% | 1 | 14.29% | 
| Jingoo Han | 1 | 0.95% | 1 | 14.29% | 
| Total | 105 | 100.00% | 7 | 100.00% | 
static void nop_completion(struct usb_ep *ep, struct usb_request *r)
{
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 15 | 100.00% | 1 | 100.00% | 
| Total | 15 | 100.00% | 1 | 100.00% | 
static int m66592_probe(struct platform_device *pdev)
{
	struct resource *res, *ires;
	void __iomem *reg = NULL;
	struct m66592 *m66592 = NULL;
	char clk_name[8];
	int ret = 0;
	int i;
	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
	if (!res) {
		ret = -ENODEV;
		pr_err("platform_get_resource error.\n");
		goto clean_up;
	}
	ires = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
	if (!ires) {
		ret = -ENODEV;
		dev_err(&pdev->dev,
			"platform_get_resource IORESOURCE_IRQ error.\n");
		goto clean_up;
	}
	reg = ioremap(res->start, resource_size(res));
	if (reg == NULL) {
		ret = -ENOMEM;
		pr_err("ioremap error.\n");
		goto clean_up;
	}
	if (dev_get_platdata(&pdev->dev) == NULL) {
		dev_err(&pdev->dev, "no platform data\n");
		ret = -ENODEV;
		goto clean_up;
	}
	/* initialize ucd */
	m66592 = kzalloc(sizeof(struct m66592), GFP_KERNEL);
	if (m66592 == NULL) {
		ret = -ENOMEM;
		goto clean_up;
	}
	m66592->pdata = dev_get_platdata(&pdev->dev);
	m66592->irq_trigger = ires->flags & IRQF_TRIGGER_MASK;
	spin_lock_init(&m66592->lock);
	platform_set_drvdata(pdev, m66592);
	m66592->gadget.ops = &m66592_gadget_ops;
	m66592->gadget.max_speed = USB_SPEED_HIGH;
	m66592->gadget.name = udc_name;
	timer_setup(&m66592->timer, m66592_timer, 0);
	m66592->reg = reg;
	ret = request_irq(ires->start, m66592_irq, IRQF_SHARED,
			udc_name, m66592);
	if (ret < 0) {
		pr_err("request_irq error (%d)\n", ret);
		goto clean_up;
	}
	if (m66592->pdata->on_chip) {
		snprintf(clk_name, sizeof(clk_name), "usbf%d", pdev->id);
		m66592->clk = clk_get(&pdev->dev, clk_name);
		if (IS_ERR(m66592->clk)) {
			dev_err(&pdev->dev, "cannot get clock \"%s\"\n",
				clk_name);
			ret = PTR_ERR(m66592->clk);
			goto clean_up2;
		}
		clk_enable(m66592->clk);
	}
	INIT_LIST_HEAD(&m66592->gadget.ep_list);
	m66592->gadget.ep0 = &m66592->ep[0].ep;
	INIT_LIST_HEAD(&m66592->gadget.ep0->ep_list);
	for (i = 0; i < M66592_MAX_NUM_PIPE; i++) {
		struct m66592_ep *ep = &m66592->ep[i];
		if (i != 0) {
			INIT_LIST_HEAD(&m66592->ep[i].ep.ep_list);
			list_add_tail(&m66592->ep[i].ep.ep_list,
					&m66592->gadget.ep_list);
		}
		ep->m66592 = m66592;
		INIT_LIST_HEAD(&ep->queue);
		ep->ep.name = m66592_ep_name[i];
		ep->ep.ops = &m66592_ep_ops;
		usb_ep_set_maxpacket_limit(&ep->ep, 512);
		if (i == 0) {
			ep->ep.caps.type_control = true;
		} else {
			ep->ep.caps.type_iso = true;
			ep->ep.caps.type_bulk = true;
			ep->ep.caps.type_int = true;
		}
		ep->ep.caps.dir_in = true;
		ep->ep.caps.dir_out = true;
	}
	usb_ep_set_maxpacket_limit(&m66592->ep[0].ep, 64);
	m66592->ep[0].pipenum = 0;
	m66592->ep[0].fifoaddr = M66592_CFIFO;
	m66592->ep[0].fifosel = M66592_CFIFOSEL;
	m66592->ep[0].fifoctr = M66592_CFIFOCTR;
	m66592->ep[0].fifotrn = 0;
	m66592->ep[0].pipectr = get_pipectr_addr(0);
	m66592->pipenum2ep[0] = &m66592->ep[0];
	m66592->epaddr2ep[0] = &m66592->ep[0];
	m66592->ep0_req = m66592_alloc_request(&m66592->ep[0].ep, GFP_KERNEL);
	if (m66592->ep0_req == NULL) {
		ret = -ENOMEM;
		goto clean_up3;
	}
	m66592->ep0_req->complete = nop_completion;
	init_controller(m66592);
	ret = usb_add_gadget_udc(&pdev->dev, &m66592->gadget);
	if (ret)
		goto err_add_udc;
	dev_info(&pdev->dev, "version %s\n", DRIVER_VERSION);
	return 0;
err_add_udc:
	m66592_free_request(&m66592->ep[0].ep, m66592->ep0_req);
clean_up3:
	if (m66592->pdata->on_chip) {
		clk_disable(m66592->clk);
		clk_put(m66592->clk);
	}
clean_up2:
	free_irq(ires->start, m66592);
clean_up:
	if (m66592) {
		if (m66592->ep0_req)
			m66592_free_request(&m66592->ep[0].ep, m66592->ep0_req);
		kfree(m66592);
	}
	if (reg)
		iounmap(reg);
	return ret;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 599 | 63.79% | 3 | 17.65% | 
| Magnus Damm | 187 | 19.91% | 3 | 17.65% | 
| Robert Baldyga | 81 | 8.63% | 2 | 11.76% | 
| Sebastian Andrzej Siewior | 41 | 4.37% | 1 | 5.88% | 
| Jingoo Han | 9 | 0.96% | 2 | 11.76% | 
| Wei Yongjun | 7 | 0.75% | 1 | 5.88% | 
| Julia Lawall | 5 | 0.53% | 1 | 5.88% | 
| David Brownell | 3 | 0.32% | 1 | 5.88% | 
| Allen | 3 | 0.32% | 1 | 5.88% | 
| Kees Cook | 2 | 0.21% | 1 | 5.88% | 
| Michal Nazarewicz | 2 | 0.21% | 1 | 5.88% | 
| Total | 939 | 100.00% | 17 | 100.00% | 
/*-------------------------------------------------------------------------*/
static struct platform_driver m66592_driver = {
	.remove =	m66592_remove,
	.driver		= {
		.name =	(char *) udc_name,
        },
};
module_platform_driver_probe(m66592_driver, m66592_probe);
Overall Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Yoshihiro Shimoda | 8120 | 91.35% | 6 | 13.64% | 
| Magnus Damm | 480 | 5.40% | 4 | 9.09% | 
| Robert Baldyga | 81 | 0.91% | 2 | 4.55% | 
| Sebastian Andrzej Siewior | 61 | 0.69% | 2 | 4.55% | 
| Michal Nazarewicz | 25 | 0.28% | 2 | 4.55% | 
| David Brownell | 22 | 0.25% | 3 | 6.82% | 
| Felipe Balbi | 19 | 0.21% | 2 | 4.55% | 
| Kees Cook | 13 | 0.15% | 1 | 2.27% | 
| Ido Shayevitz | 10 | 0.11% | 1 | 2.27% | 
| Jingoo Han | 10 | 0.11% | 2 | 4.55% | 
| Wei Yongjun | 7 | 0.08% | 1 | 2.27% | 
| Kay Sievers | 5 | 0.06% | 1 | 2.27% | 
| Peter Chen | 5 | 0.06% | 1 | 2.27% | 
| Julia Lawall | 5 | 0.06% | 1 | 2.27% | 
| Uwe Kleine-König | 4 | 0.04% | 1 | 2.27% | 
| Joe Perches | 3 | 0.03% | 2 | 4.55% | 
| Tejun Heo | 3 | 0.03% | 1 | 2.27% | 
| Paul Mundt | 3 | 0.03% | 1 | 2.27% | 
| Allen | 3 | 0.03% | 1 | 2.27% | 
| Kuninori Morimoto | 2 | 0.02% | 1 | 2.27% | 
| Greg Kroah-Hartman | 2 | 0.02% | 2 | 4.55% | 
| Al Viro | 1 | 0.01% | 1 | 2.27% | 
| Fabio Porcedda | 1 | 0.01% | 1 | 2.27% | 
| Michal Sojka | 1 | 0.01% | 1 | 2.27% | 
| Bhumika Goyal | 1 | 0.01% | 1 | 2.27% | 
| Dan Carpenter | 1 | 0.01% | 1 | 2.27% | 
| Yusuke Goda | 1 | 0.01% | 1 | 2.27% | 
| Total | 8889 | 100.00% | 44 | 100.00% | 
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.