1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3  * driver/usb/gadget/fsl_qe_udc.c
4  *
5  * Copyright (c) 2006-2008 Freescale Semiconductor, Inc. All rights reserved.
6  *
7  * 	Xie Xiaobo <X.Xie@freescale.com>
8  * 	Li Yang <leoli@freescale.com>
9  * 	Based on bareboard code from Shlomi Gridish.
10  *
11  * Description:
12  * Freescle QE/CPM USB Pheripheral Controller Driver
13  * The controller can be found on MPC8360, MPC8272, and etc.
14  * MPC8360 Rev 1.1 may need QE mircocode update
15  */
16 
17 #undef USB_TRACE
18 
19 #include <linux/module.h>
20 #include <linux/kernel.h>
21 #include <linux/ioport.h>
22 #include <linux/types.h>
23 #include <linux/errno.h>
24 #include <linux/err.h>
25 #include <linux/slab.h>
26 #include <linux/list.h>
27 #include <linux/interrupt.h>
28 #include <linux/io.h>
29 #include <linux/moduleparam.h>
30 #include <linux/of_address.h>
31 #include <linux/of_irq.h>
32 #include <linux/of_platform.h>
33 #include <linux/dma-mapping.h>
34 #include <linux/usb/ch9.h>
35 #include <linux/usb/gadget.h>
36 #include <linux/usb/otg.h>
37 #include <soc/fsl/qe/qe.h>
38 #include <asm/cpm.h>
39 #include <asm/dma.h>
40 #include <asm/reg.h>
41 #include "fsl_qe_udc.h"
42 
43 #define DRIVER_DESC     "Freescale QE/CPM USB Device Controller driver"
44 #define DRIVER_AUTHOR   "Xie XiaoBo"
45 #define DRIVER_VERSION  "1.0"
46 
47 #define DMA_ADDR_INVALID        (~(dma_addr_t)0)
48 
49 static const char driver_name[] = "fsl_qe_udc";
50 static const char driver_desc[] = DRIVER_DESC;
51 
52 /*ep name is important in gadget, it should obey the convention of ep_match()*/
53 static const char *const ep_name[] = {
54 	"ep0-control", /* everyone has ep0 */
55 	/* 3 configurable endpoints */
56 	"ep1",
57 	"ep2",
58 	"ep3",
59 };
60 
61 static const struct usb_endpoint_descriptor qe_ep0_desc = {
62 	.bLength =		USB_DT_ENDPOINT_SIZE,
63 	.bDescriptorType =	USB_DT_ENDPOINT,
64 
65 	.bEndpointAddress =	0,
66 	.bmAttributes =		USB_ENDPOINT_XFER_CONTROL,
67 	.wMaxPacketSize =	USB_MAX_CTRL_PAYLOAD,
68 };
69 
70 /********************************************************************
71  *      Internal Used Function Start
72 ********************************************************************/
73 /*-----------------------------------------------------------------
74  * done() - retire a request; caller blocked irqs
75  *--------------------------------------------------------------*/
done(struct qe_ep * ep,struct qe_req * req,int status)76 static void done(struct qe_ep *ep, struct qe_req *req, int status)
77 {
78 	struct qe_udc *udc = ep->udc;
79 	unsigned char stopped = ep->stopped;
80 
81 	/* the req->queue pointer is used by ep_queue() func, in which
82 	 * the request will be added into a udc_ep->queue 'd tail
83 	 * so here the req will be dropped from the ep->queue
84 	 */
85 	list_del_init(&req->queue);
86 
87 	/* req.status should be set as -EINPROGRESS in ep_queue() */
88 	if (req->req.status == -EINPROGRESS)
89 		req->req.status = status;
90 	else
91 		status = req->req.status;
92 
93 	if (req->mapped) {
94 		dma_unmap_single(udc->gadget.dev.parent,
95 			req->req.dma, req->req.length,
96 			ep_is_in(ep)
97 				? DMA_TO_DEVICE
98 				: DMA_FROM_DEVICE);
99 		req->req.dma = DMA_ADDR_INVALID;
100 		req->mapped = 0;
101 	} else
102 		dma_sync_single_for_cpu(udc->gadget.dev.parent,
103 			req->req.dma, req->req.length,
104 			ep_is_in(ep)
105 				? DMA_TO_DEVICE
106 				: DMA_FROM_DEVICE);
107 
108 	if (status && (status != -ESHUTDOWN))
109 		dev_vdbg(udc->dev, "complete %s req %p stat %d len %u/%u\n",
110 			ep->ep.name, &req->req, status,
111 			req->req.actual, req->req.length);
112 
113 	/* don't modify queue heads during completion callback */
114 	ep->stopped = 1;
115 	spin_unlock(&udc->lock);
116 
117 	usb_gadget_giveback_request(&ep->ep, &req->req);
118 
119 	spin_lock(&udc->lock);
120 
121 	ep->stopped = stopped;
122 }
123 
124 /*-----------------------------------------------------------------
125  * nuke(): delete all requests related to this ep
126  *--------------------------------------------------------------*/
nuke(struct qe_ep * ep,int status)127 static void nuke(struct qe_ep *ep, int status)
128 {
129 	/* Whether this eq has request linked */
130 	while (!list_empty(&ep->queue)) {
131 		struct qe_req *req = NULL;
132 		req = list_entry(ep->queue.next, struct qe_req, queue);
133 
134 		done(ep, req, status);
135 	}
136 }
137 
138 /*---------------------------------------------------------------------------*
139  * USB and Endpoint manipulate process, include parameter and register       *
140  *---------------------------------------------------------------------------*/
141 /* @value: 1--set stall 0--clean stall */
qe_eprx_stall_change(struct qe_ep * ep,int value)142 static int qe_eprx_stall_change(struct qe_ep *ep, int value)
143 {
144 	u16 tem_usep;
145 	u8 epnum = ep->epnum;
146 	struct qe_udc *udc = ep->udc;
147 
148 	tem_usep = in_be16(&udc->usb_regs->usb_usep[epnum]);
149 	tem_usep = tem_usep & ~USB_RHS_MASK;
150 	if (value == 1)
151 		tem_usep |= USB_RHS_STALL;
152 	else if (ep->dir == USB_DIR_IN)
153 		tem_usep |= USB_RHS_IGNORE_OUT;
154 
155 	out_be16(&udc->usb_regs->usb_usep[epnum], tem_usep);
156 	return 0;
157 }
158 
qe_eptx_stall_change(struct qe_ep * ep,int value)159 static int qe_eptx_stall_change(struct qe_ep *ep, int value)
160 {
161 	u16 tem_usep;
162 	u8 epnum = ep->epnum;
163 	struct qe_udc *udc = ep->udc;
164 
165 	tem_usep = in_be16(&udc->usb_regs->usb_usep[epnum]);
166 	tem_usep = tem_usep & ~USB_THS_MASK;
167 	if (value == 1)
168 		tem_usep |= USB_THS_STALL;
169 	else if (ep->dir == USB_DIR_OUT)
170 		tem_usep |= USB_THS_IGNORE_IN;
171 
172 	out_be16(&udc->usb_regs->usb_usep[epnum], tem_usep);
173 
174 	return 0;
175 }
176 
qe_ep0_stall(struct qe_udc * udc)177 static int qe_ep0_stall(struct qe_udc *udc)
178 {
179 	qe_eptx_stall_change(&udc->eps[0], 1);
180 	qe_eprx_stall_change(&udc->eps[0], 1);
181 	udc->ep0_state = WAIT_FOR_SETUP;
182 	udc->ep0_dir = 0;
183 	return 0;
184 }
185 
qe_eprx_nack(struct qe_ep * ep)186 static int qe_eprx_nack(struct qe_ep *ep)
187 {
188 	u8 epnum = ep->epnum;
189 	struct qe_udc *udc = ep->udc;
190 
191 	if (ep->state == EP_STATE_IDLE) {
192 		/* Set the ep's nack */
193 		clrsetbits_be16(&udc->usb_regs->usb_usep[epnum],
194 				USB_RHS_MASK, USB_RHS_NACK);
195 
196 		/* Mask Rx and Busy interrupts */
197 		clrbits16(&udc->usb_regs->usb_usbmr,
198 				(USB_E_RXB_MASK | USB_E_BSY_MASK));
199 
200 		ep->state = EP_STATE_NACK;
201 	}
202 	return 0;
203 }
204 
qe_eprx_normal(struct qe_ep * ep)205 static int qe_eprx_normal(struct qe_ep *ep)
206 {
207 	struct qe_udc *udc = ep->udc;
208 
209 	if (ep->state == EP_STATE_NACK) {
210 		clrsetbits_be16(&udc->usb_regs->usb_usep[ep->epnum],
211 				USB_RTHS_MASK, USB_THS_IGNORE_IN);
212 
213 		/* Unmask RX interrupts */
214 		out_be16(&udc->usb_regs->usb_usber,
215 				USB_E_BSY_MASK | USB_E_RXB_MASK);
216 		setbits16(&udc->usb_regs->usb_usbmr,
217 				(USB_E_RXB_MASK | USB_E_BSY_MASK));
218 
219 		ep->state = EP_STATE_IDLE;
220 		ep->has_data = 0;
221 	}
222 
223 	return 0;
224 }
225 
qe_ep_cmd_stoptx(struct qe_ep * ep)226 static int qe_ep_cmd_stoptx(struct qe_ep *ep)
227 {
228 	if (ep->udc->soc_type == PORT_CPM)
229 		cpm_command(CPM_USB_STOP_TX | (ep->epnum << CPM_USB_EP_SHIFT),
230 				CPM_USB_STOP_TX_OPCODE);
231 	else
232 		qe_issue_cmd(QE_USB_STOP_TX, QE_CR_SUBBLOCK_USB,
233 				ep->epnum, 0);
234 
235 	return 0;
236 }
237 
qe_ep_cmd_restarttx(struct qe_ep * ep)238 static int qe_ep_cmd_restarttx(struct qe_ep *ep)
239 {
240 	if (ep->udc->soc_type == PORT_CPM)
241 		cpm_command(CPM_USB_RESTART_TX | (ep->epnum <<
242 				CPM_USB_EP_SHIFT), CPM_USB_RESTART_TX_OPCODE);
243 	else
244 		qe_issue_cmd(QE_USB_RESTART_TX, QE_CR_SUBBLOCK_USB,
245 				ep->epnum, 0);
246 
247 	return 0;
248 }
249 
qe_ep_flushtxfifo(struct qe_ep * ep)250 static int qe_ep_flushtxfifo(struct qe_ep *ep)
251 {
252 	struct qe_udc *udc = ep->udc;
253 	int i;
254 
255 	i = (int)ep->epnum;
256 
257 	qe_ep_cmd_stoptx(ep);
258 	out_8(&udc->usb_regs->usb_uscom,
259 		USB_CMD_FLUSH_FIFO | (USB_CMD_EP_MASK & (ep->epnum)));
260 	out_be16(&udc->ep_param[i]->tbptr, in_be16(&udc->ep_param[i]->tbase));
261 	out_be32(&udc->ep_param[i]->tstate, 0);
262 	out_be16(&udc->ep_param[i]->tbcnt, 0);
263 
264 	ep->c_txbd = ep->txbase;
265 	ep->n_txbd = ep->txbase;
266 	qe_ep_cmd_restarttx(ep);
267 	return 0;
268 }
269 
qe_ep_filltxfifo(struct qe_ep * ep)270 static int qe_ep_filltxfifo(struct qe_ep *ep)
271 {
272 	struct qe_udc *udc = ep->udc;
273 
274 	out_8(&udc->usb_regs->usb_uscom,
275 			USB_CMD_STR_FIFO | (USB_CMD_EP_MASK & (ep->epnum)));
276 	return 0;
277 }
278 
qe_epbds_reset(struct qe_udc * udc,int pipe_num)279 static int qe_epbds_reset(struct qe_udc *udc, int pipe_num)
280 {
281 	struct qe_ep *ep;
282 	u32 bdring_len;
283 	struct qe_bd __iomem *bd;
284 	int i;
285 
286 	ep = &udc->eps[pipe_num];
287 
288 	if (ep->dir == USB_DIR_OUT)
289 		bdring_len = USB_BDRING_LEN_RX;
290 	else
291 		bdring_len = USB_BDRING_LEN;
292 
293 	bd = ep->rxbase;
294 	for (i = 0; i < (bdring_len - 1); i++) {
295 		out_be32((u32 __iomem *)bd, R_E | R_I);
296 		bd++;
297 	}
298 	out_be32((u32 __iomem *)bd, R_E | R_I | R_W);
299 
300 	bd = ep->txbase;
301 	for (i = 0; i < USB_BDRING_LEN_TX - 1; i++) {
302 		out_be32(&bd->buf, 0);
303 		out_be32((u32 __iomem *)bd, 0);
304 		bd++;
305 	}
306 	out_be32((u32 __iomem *)bd, T_W);
307 
308 	return 0;
309 }
310 
qe_ep_reset(struct qe_udc * udc,int pipe_num)311 static int qe_ep_reset(struct qe_udc *udc, int pipe_num)
312 {
313 	struct qe_ep *ep;
314 	u16 tmpusep;
315 
316 	ep = &udc->eps[pipe_num];
317 	tmpusep = in_be16(&udc->usb_regs->usb_usep[pipe_num]);
318 	tmpusep &= ~USB_RTHS_MASK;
319 
320 	switch (ep->dir) {
321 	case USB_DIR_BOTH:
322 		qe_ep_flushtxfifo(ep);
323 		break;
324 	case USB_DIR_OUT:
325 		tmpusep |= USB_THS_IGNORE_IN;
326 		break;
327 	case USB_DIR_IN:
328 		qe_ep_flushtxfifo(ep);
329 		tmpusep |= USB_RHS_IGNORE_OUT;
330 		break;
331 	default:
332 		break;
333 	}
334 	out_be16(&udc->usb_regs->usb_usep[pipe_num], tmpusep);
335 
336 	qe_epbds_reset(udc, pipe_num);
337 
338 	return 0;
339 }
340 
qe_ep_toggledata01(struct qe_ep * ep)341 static int qe_ep_toggledata01(struct qe_ep *ep)
342 {
343 	ep->data01 ^= 0x1;
344 	return 0;
345 }
346 
qe_ep_bd_init(struct qe_udc * udc,unsigned char pipe_num)347 static int qe_ep_bd_init(struct qe_udc *udc, unsigned char pipe_num)
348 {
349 	struct qe_ep *ep = &udc->eps[pipe_num];
350 	unsigned long tmp_addr = 0;
351 	struct usb_ep_para __iomem *epparam;
352 	int i;
353 	struct qe_bd __iomem *bd;
354 	int bdring_len;
355 
356 	if (ep->dir == USB_DIR_OUT)
357 		bdring_len = USB_BDRING_LEN_RX;
358 	else
359 		bdring_len = USB_BDRING_LEN;
360 
361 	epparam = udc->ep_param[pipe_num];
362 	/* alloc multi-ram for BD rings and set the ep parameters */
363 	tmp_addr = cpm_muram_alloc(sizeof(struct qe_bd) * (bdring_len +
364 				USB_BDRING_LEN_TX), QE_ALIGNMENT_OF_BD);
365 	if (IS_ERR_VALUE(tmp_addr))
366 		return -ENOMEM;
367 
368 	out_be16(&epparam->rbase, (u16)tmp_addr);
369 	out_be16(&epparam->tbase, (u16)(tmp_addr +
370 				(sizeof(struct qe_bd) * bdring_len)));
371 
372 	out_be16(&epparam->rbptr, in_be16(&epparam->rbase));
373 	out_be16(&epparam->tbptr, in_be16(&epparam->tbase));
374 
375 	ep->rxbase = cpm_muram_addr(tmp_addr);
376 	ep->txbase = cpm_muram_addr(tmp_addr + (sizeof(struct qe_bd)
377 				* bdring_len));
378 	ep->n_rxbd = ep->rxbase;
379 	ep->e_rxbd = ep->rxbase;
380 	ep->n_txbd = ep->txbase;
381 	ep->c_txbd = ep->txbase;
382 	ep->data01 = 0; /* data0 */
383 
384 	/* Init TX and RX bds */
385 	bd = ep->rxbase;
386 	for (i = 0; i < bdring_len - 1; i++) {
387 		out_be32(&bd->buf, 0);
388 		out_be32((u32 __iomem *)bd, 0);
389 		bd++;
390 	}
391 	out_be32(&bd->buf, 0);
392 	out_be32((u32 __iomem *)bd, R_W);
393 
394 	bd = ep->txbase;
395 	for (i = 0; i < USB_BDRING_LEN_TX - 1; i++) {
396 		out_be32(&bd->buf, 0);
397 		out_be32((u32 __iomem *)bd, 0);
398 		bd++;
399 	}
400 	out_be32(&bd->buf, 0);
401 	out_be32((u32 __iomem *)bd, T_W);
402 
403 	return 0;
404 }
405 
qe_ep_rxbd_update(struct qe_ep * ep)406 static int qe_ep_rxbd_update(struct qe_ep *ep)
407 {
408 	unsigned int size;
409 	int i;
410 	unsigned int tmp;
411 	struct qe_bd __iomem *bd;
412 	unsigned int bdring_len;
413 
414 	if (ep->rxbase == NULL)
415 		return -EINVAL;
416 
417 	bd = ep->rxbase;
418 
419 	ep->rxframe = kmalloc(sizeof(*ep->rxframe), GFP_ATOMIC);
420 	if (!ep->rxframe)
421 		return -ENOMEM;
422 
423 	qe_frame_init(ep->rxframe);
424 
425 	if (ep->dir == USB_DIR_OUT)
426 		bdring_len = USB_BDRING_LEN_RX;
427 	else
428 		bdring_len = USB_BDRING_LEN;
429 
430 	size = (ep->ep.maxpacket + USB_CRC_SIZE + 2) * (bdring_len + 1);
431 	ep->rxbuffer = kzalloc(size, GFP_ATOMIC);
432 	if (!ep->rxbuffer) {
433 		kfree(ep->rxframe);
434 		return -ENOMEM;
435 	}
436 
437 	ep->rxbuf_d = virt_to_phys((void *)ep->rxbuffer);
438 	if (ep->rxbuf_d == DMA_ADDR_INVALID) {
439 		ep->rxbuf_d = dma_map_single(ep->udc->gadget.dev.parent,
440 					ep->rxbuffer,
441 					size,
442 					DMA_FROM_DEVICE);
443 		ep->rxbufmap = 1;
444 	} else {
445 		dma_sync_single_for_device(ep->udc->gadget.dev.parent,
446 					ep->rxbuf_d, size,
447 					DMA_FROM_DEVICE);
448 		ep->rxbufmap = 0;
449 	}
450 
451 	size = ep->ep.maxpacket + USB_CRC_SIZE + 2;
452 	tmp = ep->rxbuf_d;
453 	tmp = (u32)(((tmp >> 2) << 2) + 4);
454 
455 	for (i = 0; i < bdring_len - 1; i++) {
456 		out_be32(&bd->buf, tmp);
457 		out_be32((u32 __iomem *)bd, (R_E | R_I));
458 		tmp = tmp + size;
459 		bd++;
460 	}
461 	out_be32(&bd->buf, tmp);
462 	out_be32((u32 __iomem *)bd, (R_E | R_I | R_W));
463 
464 	return 0;
465 }
466 
qe_ep_register_init(struct qe_udc * udc,unsigned char pipe_num)467 static int qe_ep_register_init(struct qe_udc *udc, unsigned char pipe_num)
468 {
469 	struct qe_ep *ep = &udc->eps[pipe_num];
470 	struct usb_ep_para __iomem *epparam;
471 	u16 usep, logepnum;
472 	u16 tmp;
473 	u8 rtfcr = 0;
474 
475 	epparam = udc->ep_param[pipe_num];
476 
477 	usep = 0;
478 	logepnum = (ep->ep.desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK);
479 	usep |= (logepnum << USB_EPNUM_SHIFT);
480 
481 	switch (ep->ep.desc->bmAttributes & 0x03) {
482 	case USB_ENDPOINT_XFER_BULK:
483 		usep |= USB_TRANS_BULK;
484 		break;
485 	case USB_ENDPOINT_XFER_ISOC:
486 		usep |=  USB_TRANS_ISO;
487 		break;
488 	case USB_ENDPOINT_XFER_INT:
489 		usep |= USB_TRANS_INT;
490 		break;
491 	default:
492 		usep |= USB_TRANS_CTR;
493 		break;
494 	}
495 
496 	switch (ep->dir) {
497 	case USB_DIR_OUT:
498 		usep |= USB_THS_IGNORE_IN;
499 		break;
500 	case USB_DIR_IN:
501 		usep |= USB_RHS_IGNORE_OUT;
502 		break;
503 	default:
504 		break;
505 	}
506 	out_be16(&udc->usb_regs->usb_usep[pipe_num], usep);
507 
508 	rtfcr = 0x30;
509 	out_8(&epparam->rbmr, rtfcr);
510 	out_8(&epparam->tbmr, rtfcr);
511 
512 	tmp = (u16)(ep->ep.maxpacket + USB_CRC_SIZE);
513 	/* MRBLR must be divisble by 4 */
514 	tmp = (u16)(((tmp >> 2) << 2) + 4);
515 	out_be16(&epparam->mrblr, tmp);
516 
517 	return 0;
518 }
519 
qe_ep_init(struct qe_udc * udc,unsigned char pipe_num,const struct usb_endpoint_descriptor * desc)520 static int qe_ep_init(struct qe_udc *udc,
521 		      unsigned char pipe_num,
522 		      const struct usb_endpoint_descriptor *desc)
523 {
524 	struct qe_ep *ep = &udc->eps[pipe_num];
525 	unsigned long flags;
526 	int reval = 0;
527 	u16 max = 0;
528 
529 	max = usb_endpoint_maxp(desc);
530 
531 	/* check the max package size validate for this endpoint */
532 	/* Refer to USB2.0 spec table 9-13,
533 	*/
534 	if (pipe_num != 0) {
535 		switch (desc->bmAttributes & USB_ENDPOINT_XFERTYPE_MASK) {
536 		case USB_ENDPOINT_XFER_BULK:
537 			if (strstr(ep->ep.name, "-iso")
538 					|| strstr(ep->ep.name, "-int"))
539 				goto en_done;
540 			switch (udc->gadget.speed) {
541 			case USB_SPEED_HIGH:
542 			if ((max == 128) || (max == 256) || (max == 512))
543 				break;
544 			default:
545 				switch (max) {
546 				case 4:
547 				case 8:
548 				case 16:
549 				case 32:
550 				case 64:
551 					break;
552 				default:
553 				case USB_SPEED_LOW:
554 					goto en_done;
555 				}
556 			}
557 			break;
558 		case USB_ENDPOINT_XFER_INT:
559 			if (strstr(ep->ep.name, "-iso"))	/* bulk is ok */
560 				goto en_done;
561 			switch (udc->gadget.speed) {
562 			case USB_SPEED_HIGH:
563 				if (max <= 1024)
564 					break;
565 			case USB_SPEED_FULL:
566 				if (max <= 64)
567 					break;
568 			default:
569 				if (max <= 8)
570 					break;
571 				goto en_done;
572 			}
573 			break;
574 		case USB_ENDPOINT_XFER_ISOC:
575 			if (strstr(ep->ep.name, "-bulk")
576 				|| strstr(ep->ep.name, "-int"))
577 				goto en_done;
578 			switch (udc->gadget.speed) {
579 			case USB_SPEED_HIGH:
580 				if (max <= 1024)
581 					break;
582 			case USB_SPEED_FULL:
583 				if (max <= 1023)
584 					break;
585 			default:
586 				goto en_done;
587 			}
588 			break;
589 		case USB_ENDPOINT_XFER_CONTROL:
590 			if (strstr(ep->ep.name, "-iso")
591 				|| strstr(ep->ep.name, "-int"))
592 				goto en_done;
593 			switch (udc->gadget.speed) {
594 			case USB_SPEED_HIGH:
595 			case USB_SPEED_FULL:
596 				switch (max) {
597 				case 1:
598 				case 2:
599 				case 4:
600 				case 8:
601 				case 16:
602 				case 32:
603 				case 64:
604 					break;
605 				default:
606 					goto en_done;
607 				}
608 			case USB_SPEED_LOW:
609 				switch (max) {
610 				case 1:
611 				case 2:
612 				case 4:
613 				case 8:
614 					break;
615 				default:
616 					goto en_done;
617 				}
618 			default:
619 				goto en_done;
620 			}
621 			break;
622 
623 		default:
624 			goto en_done;
625 		}
626 	} /* if ep0*/
627 
628 	spin_lock_irqsave(&udc->lock, flags);
629 
630 	/* initialize ep structure */
631 	ep->ep.maxpacket = max;
632 	ep->tm = (u8)(desc->bmAttributes & USB_ENDPOINT_XFERTYPE_MASK);
633 	ep->ep.desc = desc;
634 	ep->stopped = 0;
635 	ep->init = 1;
636 
637 	if (pipe_num == 0) {
638 		ep->dir = USB_DIR_BOTH;
639 		udc->ep0_dir = USB_DIR_OUT;
640 		udc->ep0_state = WAIT_FOR_SETUP;
641 	} else	{
642 		switch (desc->bEndpointAddress & USB_ENDPOINT_DIR_MASK) {
643 		case USB_DIR_OUT:
644 			ep->dir = USB_DIR_OUT;
645 			break;
646 		case USB_DIR_IN:
647 			ep->dir = USB_DIR_IN;
648 		default:
649 			break;
650 		}
651 	}
652 
653 	/* hardware special operation */
654 	qe_ep_bd_init(udc, pipe_num);
655 	if ((ep->tm == USBP_TM_CTL) || (ep->dir == USB_DIR_OUT)) {
656 		reval = qe_ep_rxbd_update(ep);
657 		if (reval)
658 			goto en_done1;
659 	}
660 
661 	if ((ep->tm == USBP_TM_CTL) || (ep->dir == USB_DIR_IN)) {
662 		ep->txframe = kmalloc(sizeof(*ep->txframe), GFP_ATOMIC);
663 		if (!ep->txframe)
664 			goto en_done2;
665 		qe_frame_init(ep->txframe);
666 	}
667 
668 	qe_ep_register_init(udc, pipe_num);
669 
670 	/* Now HW will be NAKing transfers to that EP,
671 	 * until a buffer is queued to it. */
672 	spin_unlock_irqrestore(&udc->lock, flags);
673 
674 	return 0;
675 en_done2:
676 	kfree(ep->rxbuffer);
677 	kfree(ep->rxframe);
678 en_done1:
679 	spin_unlock_irqrestore(&udc->lock, flags);
680 en_done:
681 	dev_err(udc->dev, "failed to initialize %s\n", ep->ep.name);
682 	return -ENODEV;
683 }
684 
qe_usb_enable(struct qe_udc * udc)685 static inline void qe_usb_enable(struct qe_udc *udc)
686 {
687 	setbits8(&udc->usb_regs->usb_usmod, USB_MODE_EN);
688 }
689 
qe_usb_disable(struct qe_udc * udc)690 static inline void qe_usb_disable(struct qe_udc *udc)
691 {
692 	clrbits8(&udc->usb_regs->usb_usmod, USB_MODE_EN);
693 }
694 
695 /*----------------------------------------------------------------------------*
696  *		USB and EP basic manipulate function end		      *
697  *----------------------------------------------------------------------------*/
698 
699 
700 /******************************************************************************
701 		UDC transmit and receive process
702  ******************************************************************************/
recycle_one_rxbd(struct qe_ep * ep)703 static void recycle_one_rxbd(struct qe_ep *ep)
704 {
705 	u32 bdstatus;
706 
707 	bdstatus = in_be32((u32 __iomem *)ep->e_rxbd);
708 	bdstatus = R_I | R_E | (bdstatus & R_W);
709 	out_be32((u32 __iomem *)ep->e_rxbd, bdstatus);
710 
711 	if (bdstatus & R_W)
712 		ep->e_rxbd = ep->rxbase;
713 	else
714 		ep->e_rxbd++;
715 }
716 
recycle_rxbds(struct qe_ep * ep,unsigned char stopatnext)717 static void recycle_rxbds(struct qe_ep *ep, unsigned char stopatnext)
718 {
719 	u32 bdstatus;
720 	struct qe_bd __iomem *bd, *nextbd;
721 	unsigned char stop = 0;
722 
723 	nextbd = ep->n_rxbd;
724 	bd = ep->e_rxbd;
725 	bdstatus = in_be32((u32 __iomem *)bd);
726 
727 	while (!(bdstatus & R_E) && !(bdstatus & BD_LENGTH_MASK) && !stop) {
728 		bdstatus = R_E | R_I | (bdstatus & R_W);
729 		out_be32((u32 __iomem *)bd, bdstatus);
730 
731 		if (bdstatus & R_W)
732 			bd = ep->rxbase;
733 		else
734 			bd++;
735 
736 		bdstatus = in_be32((u32 __iomem *)bd);
737 		if (stopatnext && (bd == nextbd))
738 			stop = 1;
739 	}
740 
741 	ep->e_rxbd = bd;
742 }
743 
ep_recycle_rxbds(struct qe_ep * ep)744 static void ep_recycle_rxbds(struct qe_ep *ep)
745 {
746 	struct qe_bd __iomem *bd = ep->n_rxbd;
747 	u32 bdstatus;
748 	u8 epnum = ep->epnum;
749 	struct qe_udc *udc = ep->udc;
750 
751 	bdstatus = in_be32((u32 __iomem *)bd);
752 	if (!(bdstatus & R_E) && !(bdstatus & BD_LENGTH_MASK)) {
753 		bd = ep->rxbase +
754 				((in_be16(&udc->ep_param[epnum]->rbptr) -
755 				  in_be16(&udc->ep_param[epnum]->rbase))
756 				 >> 3);
757 		bdstatus = in_be32((u32 __iomem *)bd);
758 
759 		if (bdstatus & R_W)
760 			bd = ep->rxbase;
761 		else
762 			bd++;
763 
764 		ep->e_rxbd = bd;
765 		recycle_rxbds(ep, 0);
766 		ep->e_rxbd = ep->n_rxbd;
767 	} else
768 		recycle_rxbds(ep, 1);
769 
770 	if (in_be16(&udc->usb_regs->usb_usber) & USB_E_BSY_MASK)
771 		out_be16(&udc->usb_regs->usb_usber, USB_E_BSY_MASK);
772 
773 	if (ep->has_data <= 0 && (!list_empty(&ep->queue)))
774 		qe_eprx_normal(ep);
775 
776 	ep->localnack = 0;
777 }
778 
779 static void setup_received_handle(struct qe_udc *udc,
780 					struct usb_ctrlrequest *setup);
781 static int qe_ep_rxframe_handle(struct qe_ep *ep);
782 static void ep0_req_complete(struct qe_udc *udc, struct qe_req *req);
783 /* when BD PID is setup, handle the packet */
ep0_setup_handle(struct qe_udc * udc)784 static int ep0_setup_handle(struct qe_udc *udc)
785 {
786 	struct qe_ep *ep = &udc->eps[0];
787 	struct qe_frame *pframe;
788 	unsigned int fsize;
789 	u8 *cp;
790 
791 	pframe = ep->rxframe;
792 	if ((frame_get_info(pframe) & PID_SETUP)
793 			&& (udc->ep0_state == WAIT_FOR_SETUP)) {
794 		fsize = frame_get_length(pframe);
795 		if (unlikely(fsize != 8))
796 			return -EINVAL;
797 		cp = (u8 *)&udc->local_setup_buff;
798 		memcpy(cp, pframe->data, fsize);
799 		ep->data01 = 1;
800 
801 		/* handle the usb command base on the usb_ctrlrequest */
802 		setup_received_handle(udc, &udc->local_setup_buff);
803 		return 0;
804 	}
805 	return -EINVAL;
806 }
807 
qe_ep0_rx(struct qe_udc * udc)808 static int qe_ep0_rx(struct qe_udc *udc)
809 {
810 	struct qe_ep *ep = &udc->eps[0];
811 	struct qe_frame *pframe;
812 	struct qe_bd __iomem *bd;
813 	u32 bdstatus, length;
814 	u32 vaddr;
815 
816 	pframe = ep->rxframe;
817 
818 	if (ep->dir == USB_DIR_IN) {
819 		dev_err(udc->dev, "ep0 not a control endpoint\n");
820 		return -EINVAL;
821 	}
822 
823 	bd = ep->n_rxbd;
824 	bdstatus = in_be32((u32 __iomem *)bd);
825 	length = bdstatus & BD_LENGTH_MASK;
826 
827 	while (!(bdstatus & R_E) && length) {
828 		if ((bdstatus & R_F) && (bdstatus & R_L)
829 			&& !(bdstatus & R_ERROR)) {
830 			if (length == USB_CRC_SIZE) {
831 				udc->ep0_state = WAIT_FOR_SETUP;
832 				dev_vdbg(udc->dev,
833 					"receive a ZLP in status phase\n");
834 			} else {
835 				qe_frame_clean(pframe);
836 				vaddr = (u32)phys_to_virt(in_be32(&bd->buf));
837 				frame_set_data(pframe, (u8 *)vaddr);
838 				frame_set_length(pframe,
839 						(length - USB_CRC_SIZE));
840 				frame_set_status(pframe, FRAME_OK);
841 				switch (bdstatus & R_PID) {
842 				case R_PID_SETUP:
843 					frame_set_info(pframe, PID_SETUP);
844 					break;
845 				case R_PID_DATA1:
846 					frame_set_info(pframe, PID_DATA1);
847 					break;
848 				default:
849 					frame_set_info(pframe, PID_DATA0);
850 					break;
851 				}
852 
853 				if ((bdstatus & R_PID) == R_PID_SETUP)
854 					ep0_setup_handle(udc);
855 				else
856 					qe_ep_rxframe_handle(ep);
857 			}
858 		} else {
859 			dev_err(udc->dev, "The receive frame with error!\n");
860 		}
861 
862 		/* note: don't clear the rxbd's buffer address */
863 		recycle_one_rxbd(ep);
864 
865 		/* Get next BD */
866 		if (bdstatus & R_W)
867 			bd = ep->rxbase;
868 		else
869 			bd++;
870 
871 		bdstatus = in_be32((u32 __iomem *)bd);
872 		length = bdstatus & BD_LENGTH_MASK;
873 
874 	}
875 
876 	ep->n_rxbd = bd;
877 
878 	return 0;
879 }
880 
qe_ep_rxframe_handle(struct qe_ep * ep)881 static int qe_ep_rxframe_handle(struct qe_ep *ep)
882 {
883 	struct qe_frame *pframe;
884 	u8 framepid = 0;
885 	unsigned int fsize;
886 	u8 *cp;
887 	struct qe_req *req;
888 
889 	pframe = ep->rxframe;
890 
891 	if (frame_get_info(pframe) & PID_DATA1)
892 		framepid = 0x1;
893 
894 	if (framepid != ep->data01) {
895 		dev_err(ep->udc->dev, "the data01 error!\n");
896 		return -EIO;
897 	}
898 
899 	fsize = frame_get_length(pframe);
900 	if (list_empty(&ep->queue)) {
901 		dev_err(ep->udc->dev, "the %s have no requeue!\n", ep->name);
902 	} else {
903 		req = list_entry(ep->queue.next, struct qe_req, queue);
904 
905 		cp = (u8 *)(req->req.buf) + req->req.actual;
906 		if (cp) {
907 			memcpy(cp, pframe->data, fsize);
908 			req->req.actual += fsize;
909 			if ((fsize < ep->ep.maxpacket) ||
910 					(req->req.actual >= req->req.length)) {
911 				if (ep->epnum == 0)
912 					ep0_req_complete(ep->udc, req);
913 				else
914 					done(ep, req, 0);
915 				if (list_empty(&ep->queue) && ep->epnum != 0)
916 					qe_eprx_nack(ep);
917 			}
918 		}
919 	}
920 
921 	qe_ep_toggledata01(ep);
922 
923 	return 0;
924 }
925 
ep_rx_tasklet(unsigned long data)926 static void ep_rx_tasklet(unsigned long data)
927 {
928 	struct qe_udc *udc = (struct qe_udc *)data;
929 	struct qe_ep *ep;
930 	struct qe_frame *pframe;
931 	struct qe_bd __iomem *bd;
932 	unsigned long flags;
933 	u32 bdstatus, length;
934 	u32 vaddr, i;
935 
936 	spin_lock_irqsave(&udc->lock, flags);
937 
938 	for (i = 1; i < USB_MAX_ENDPOINTS; i++) {
939 		ep = &udc->eps[i];
940 
941 		if (ep->dir == USB_DIR_IN || ep->enable_tasklet == 0) {
942 			dev_dbg(udc->dev,
943 				"This is a transmit ep or disable tasklet!\n");
944 			continue;
945 		}
946 
947 		pframe = ep->rxframe;
948 		bd = ep->n_rxbd;
949 		bdstatus = in_be32((u32 __iomem *)bd);
950 		length = bdstatus & BD_LENGTH_MASK;
951 
952 		while (!(bdstatus & R_E) && length) {
953 			if (list_empty(&ep->queue)) {
954 				qe_eprx_nack(ep);
955 				dev_dbg(udc->dev,
956 					"The rxep have noreq %d\n",
957 					ep->has_data);
958 				break;
959 			}
960 
961 			if ((bdstatus & R_F) && (bdstatus & R_L)
962 				&& !(bdstatus & R_ERROR)) {
963 				qe_frame_clean(pframe);
964 				vaddr = (u32)phys_to_virt(in_be32(&bd->buf));
965 				frame_set_data(pframe, (u8 *)vaddr);
966 				frame_set_length(pframe,
967 						(length - USB_CRC_SIZE));
968 				frame_set_status(pframe, FRAME_OK);
969 				switch (bdstatus & R_PID) {
970 				case R_PID_DATA1:
971 					frame_set_info(pframe, PID_DATA1);
972 					break;
973 				case R_PID_SETUP:
974 					frame_set_info(pframe, PID_SETUP);
975 					break;
976 				default:
977 					frame_set_info(pframe, PID_DATA0);
978 					break;
979 				}
980 				/* handle the rx frame */
981 				qe_ep_rxframe_handle(ep);
982 			} else {
983 				dev_err(udc->dev,
984 					"error in received frame\n");
985 			}
986 			/* note: don't clear the rxbd's buffer address */
987 			/*clear the length */
988 			out_be32((u32 __iomem *)bd, bdstatus & BD_STATUS_MASK);
989 			ep->has_data--;
990 			if (!(ep->localnack))
991 				recycle_one_rxbd(ep);
992 
993 			/* Get next BD */
994 			if (bdstatus & R_W)
995 				bd = ep->rxbase;
996 			else
997 				bd++;
998 
999 			bdstatus = in_be32((u32 __iomem *)bd);
1000 			length = bdstatus & BD_LENGTH_MASK;
1001 		}
1002 
1003 		ep->n_rxbd = bd;
1004 
1005 		if (ep->localnack)
1006 			ep_recycle_rxbds(ep);
1007 
1008 		ep->enable_tasklet = 0;
1009 	} /* for i=1 */
1010 
1011 	spin_unlock_irqrestore(&udc->lock, flags);
1012 }
1013 
qe_ep_rx(struct qe_ep * ep)1014 static int qe_ep_rx(struct qe_ep *ep)
1015 {
1016 	struct qe_udc *udc;
1017 	struct qe_frame *pframe;
1018 	struct qe_bd __iomem *bd;
1019 	u16 swoffs, ucoffs, emptybds;
1020 
1021 	udc = ep->udc;
1022 	pframe = ep->rxframe;
1023 
1024 	if (ep->dir == USB_DIR_IN) {
1025 		dev_err(udc->dev, "transmit ep in rx function\n");
1026 		return -EINVAL;
1027 	}
1028 
1029 	bd = ep->n_rxbd;
1030 
1031 	swoffs = (u16)(bd - ep->rxbase);
1032 	ucoffs = (u16)((in_be16(&udc->ep_param[ep->epnum]->rbptr) -
1033 			in_be16(&udc->ep_param[ep->epnum]->rbase)) >> 3);
1034 	if (swoffs < ucoffs)
1035 		emptybds = USB_BDRING_LEN_RX - ucoffs + swoffs;
1036 	else
1037 		emptybds = swoffs - ucoffs;
1038 
1039 	if (emptybds < MIN_EMPTY_BDS) {
1040 		qe_eprx_nack(ep);
1041 		ep->localnack = 1;
1042 		dev_vdbg(udc->dev, "%d empty bds, send NACK\n", emptybds);
1043 	}
1044 	ep->has_data = USB_BDRING_LEN_RX - emptybds;
1045 
1046 	if (list_empty(&ep->queue)) {
1047 		qe_eprx_nack(ep);
1048 		dev_vdbg(udc->dev, "The rxep have no req queued with %d BDs\n",
1049 				ep->has_data);
1050 		return 0;
1051 	}
1052 
1053 	tasklet_schedule(&udc->rx_tasklet);
1054 	ep->enable_tasklet = 1;
1055 
1056 	return 0;
1057 }
1058 
1059 /* send data from a frame, no matter what tx_req */
qe_ep_tx(struct qe_ep * ep,struct qe_frame * frame)1060 static int qe_ep_tx(struct qe_ep *ep, struct qe_frame *frame)
1061 {
1062 	struct qe_udc *udc = ep->udc;
1063 	struct qe_bd __iomem *bd;
1064 	u16 saveusbmr;
1065 	u32 bdstatus, pidmask;
1066 	u32 paddr;
1067 
1068 	if (ep->dir == USB_DIR_OUT) {
1069 		dev_err(udc->dev, "receive ep passed to tx function\n");
1070 		return -EINVAL;
1071 	}
1072 
1073 	/* Disable the Tx interrupt */
1074 	saveusbmr = in_be16(&udc->usb_regs->usb_usbmr);
1075 	out_be16(&udc->usb_regs->usb_usbmr,
1076 			saveusbmr & ~(USB_E_TXB_MASK | USB_E_TXE_MASK));
1077 
1078 	bd = ep->n_txbd;
1079 	bdstatus = in_be32((u32 __iomem *)bd);
1080 
1081 	if (!(bdstatus & (T_R | BD_LENGTH_MASK))) {
1082 		if (frame_get_length(frame) == 0) {
1083 			frame_set_data(frame, udc->nullbuf);
1084 			frame_set_length(frame, 2);
1085 			frame->info |= (ZLP | NO_CRC);
1086 			dev_vdbg(udc->dev, "the frame size = 0\n");
1087 		}
1088 		paddr = virt_to_phys((void *)frame->data);
1089 		out_be32(&bd->buf, paddr);
1090 		bdstatus = (bdstatus&T_W);
1091 		if (!(frame_get_info(frame) & NO_CRC))
1092 			bdstatus |= T_R | T_I | T_L | T_TC
1093 					| frame_get_length(frame);
1094 		else
1095 			bdstatus |= T_R | T_I | T_L | frame_get_length(frame);
1096 
1097 		/* if the packet is a ZLP in status phase */
1098 		if ((ep->epnum == 0) && (udc->ep0_state == DATA_STATE_NEED_ZLP))
1099 			ep->data01 = 0x1;
1100 
1101 		if (ep->data01) {
1102 			pidmask = T_PID_DATA1;
1103 			frame->info |= PID_DATA1;
1104 		} else {
1105 			pidmask = T_PID_DATA0;
1106 			frame->info |= PID_DATA0;
1107 		}
1108 		bdstatus |= T_CNF;
1109 		bdstatus |= pidmask;
1110 		out_be32((u32 __iomem *)bd, bdstatus);
1111 		qe_ep_filltxfifo(ep);
1112 
1113 		/* enable the TX interrupt */
1114 		out_be16(&udc->usb_regs->usb_usbmr, saveusbmr);
1115 
1116 		qe_ep_toggledata01(ep);
1117 		if (bdstatus & T_W)
1118 			ep->n_txbd = ep->txbase;
1119 		else
1120 			ep->n_txbd++;
1121 
1122 		return 0;
1123 	} else {
1124 		out_be16(&udc->usb_regs->usb_usbmr, saveusbmr);
1125 		dev_vdbg(udc->dev, "The tx bd is not ready!\n");
1126 		return -EBUSY;
1127 	}
1128 }
1129 
1130 /* when a bd was transmitted, the function can
1131  * handle the tx_req, not include ep0           */
txcomplete(struct qe_ep * ep,unsigned char restart)1132 static int txcomplete(struct qe_ep *ep, unsigned char restart)
1133 {
1134 	if (ep->tx_req != NULL) {
1135 		struct qe_req *req = ep->tx_req;
1136 		unsigned zlp = 0, last_len = 0;
1137 
1138 		last_len = min_t(unsigned, req->req.length - ep->sent,
1139 				ep->ep.maxpacket);
1140 
1141 		if (!restart) {
1142 			int asent = ep->last;
1143 			ep->sent += asent;
1144 			ep->last -= asent;
1145 		} else {
1146 			ep->last = 0;
1147 		}
1148 
1149 		/* zlp needed when req->re.zero is set */
1150 		if (req->req.zero) {
1151 			if (last_len == 0 ||
1152 				(req->req.length % ep->ep.maxpacket) != 0)
1153 				zlp = 0;
1154 			else
1155 				zlp = 1;
1156 		} else
1157 			zlp = 0;
1158 
1159 		/* a request already were transmitted completely */
1160 		if (((ep->tx_req->req.length - ep->sent) <= 0) && !zlp) {
1161 			done(ep, ep->tx_req, 0);
1162 			ep->tx_req = NULL;
1163 			ep->last = 0;
1164 			ep->sent = 0;
1165 		}
1166 	}
1167 
1168 	/* we should gain a new tx_req fot this endpoint */
1169 	if (ep->tx_req == NULL) {
1170 		if (!list_empty(&ep->queue)) {
1171 			ep->tx_req = list_entry(ep->queue.next,	struct qe_req,
1172 							queue);
1173 			ep->last = 0;
1174 			ep->sent = 0;
1175 		}
1176 	}
1177 
1178 	return 0;
1179 }
1180 
1181 /* give a frame and a tx_req, send some data */
qe_usb_senddata(struct qe_ep * ep,struct qe_frame * frame)1182 static int qe_usb_senddata(struct qe_ep *ep, struct qe_frame *frame)
1183 {
1184 	unsigned int size;
1185 	u8 *buf;
1186 
1187 	qe_frame_clean(frame);
1188 	size = min_t(u32, (ep->tx_req->req.length - ep->sent),
1189 				ep->ep.maxpacket);
1190 	buf = (u8 *)ep->tx_req->req.buf + ep->sent;
1191 	if (buf && size) {
1192 		ep->last = size;
1193 		ep->tx_req->req.actual += size;
1194 		frame_set_data(frame, buf);
1195 		frame_set_length(frame, size);
1196 		frame_set_status(frame, FRAME_OK);
1197 		frame_set_info(frame, 0);
1198 		return qe_ep_tx(ep, frame);
1199 	}
1200 	return -EIO;
1201 }
1202 
1203 /* give a frame struct,send a ZLP */
sendnulldata(struct qe_ep * ep,struct qe_frame * frame,uint infor)1204 static int sendnulldata(struct qe_ep *ep, struct qe_frame *frame, uint infor)
1205 {
1206 	struct qe_udc *udc = ep->udc;
1207 
1208 	if (frame == NULL)
1209 		return -ENODEV;
1210 
1211 	qe_frame_clean(frame);
1212 	frame_set_data(frame, (u8 *)udc->nullbuf);
1213 	frame_set_length(frame, 2);
1214 	frame_set_status(frame, FRAME_OK);
1215 	frame_set_info(frame, (ZLP | NO_CRC | infor));
1216 
1217 	return qe_ep_tx(ep, frame);
1218 }
1219 
frame_create_tx(struct qe_ep * ep,struct qe_frame * frame)1220 static int frame_create_tx(struct qe_ep *ep, struct qe_frame *frame)
1221 {
1222 	struct qe_req *req = ep->tx_req;
1223 	int reval;
1224 
1225 	if (req == NULL)
1226 		return -ENODEV;
1227 
1228 	if ((req->req.length - ep->sent) > 0)
1229 		reval = qe_usb_senddata(ep, frame);
1230 	else
1231 		reval = sendnulldata(ep, frame, 0);
1232 
1233 	return reval;
1234 }
1235 
1236 /* if direction is DIR_IN, the status is Device->Host
1237  * if direction is DIR_OUT, the status transaction is Device<-Host
1238  * in status phase, udc create a request and gain status */
ep0_prime_status(struct qe_udc * udc,int direction)1239 static int ep0_prime_status(struct qe_udc *udc, int direction)
1240 {
1241 
1242 	struct qe_ep *ep = &udc->eps[0];
1243 
1244 	if (direction == USB_DIR_IN) {
1245 		udc->ep0_state = DATA_STATE_NEED_ZLP;
1246 		udc->ep0_dir = USB_DIR_IN;
1247 		sendnulldata(ep, ep->txframe, SETUP_STATUS | NO_REQ);
1248 	} else {
1249 		udc->ep0_dir = USB_DIR_OUT;
1250 		udc->ep0_state = WAIT_FOR_OUT_STATUS;
1251 	}
1252 
1253 	return 0;
1254 }
1255 
1256 /* a request complete in ep0, whether gadget request or udc request */
ep0_req_complete(struct qe_udc * udc,struct qe_req * req)1257 static void ep0_req_complete(struct qe_udc *udc, struct qe_req *req)
1258 {
1259 	struct qe_ep *ep = &udc->eps[0];
1260 	/* because usb and ep's status already been set in ch9setaddress() */
1261 
1262 	switch (udc->ep0_state) {
1263 	case DATA_STATE_XMIT:
1264 		done(ep, req, 0);
1265 		/* receive status phase */
1266 		if (ep0_prime_status(udc, USB_DIR_OUT))
1267 			qe_ep0_stall(udc);
1268 		break;
1269 
1270 	case DATA_STATE_NEED_ZLP:
1271 		done(ep, req, 0);
1272 		udc->ep0_state = WAIT_FOR_SETUP;
1273 		break;
1274 
1275 	case DATA_STATE_RECV:
1276 		done(ep, req, 0);
1277 		/* send status phase */
1278 		if (ep0_prime_status(udc, USB_DIR_IN))
1279 			qe_ep0_stall(udc);
1280 		break;
1281 
1282 	case WAIT_FOR_OUT_STATUS:
1283 		done(ep, req, 0);
1284 		udc->ep0_state = WAIT_FOR_SETUP;
1285 		break;
1286 
1287 	case WAIT_FOR_SETUP:
1288 		dev_vdbg(udc->dev, "Unexpected interrupt\n");
1289 		break;
1290 
1291 	default:
1292 		qe_ep0_stall(udc);
1293 		break;
1294 	}
1295 }
1296 
ep0_txcomplete(struct qe_ep * ep,unsigned char restart)1297 static int ep0_txcomplete(struct qe_ep *ep, unsigned char restart)
1298 {
1299 	struct qe_req *tx_req = NULL;
1300 	struct qe_frame *frame = ep->txframe;
1301 
1302 	if ((frame_get_info(frame) & (ZLP | NO_REQ)) == (ZLP | NO_REQ)) {
1303 		if (!restart)
1304 			ep->udc->ep0_state = WAIT_FOR_SETUP;
1305 		else
1306 			sendnulldata(ep, ep->txframe, SETUP_STATUS | NO_REQ);
1307 		return 0;
1308 	}
1309 
1310 	tx_req = ep->tx_req;
1311 	if (tx_req != NULL) {
1312 		if (!restart) {
1313 			int asent = ep->last;
1314 			ep->sent += asent;
1315 			ep->last -= asent;
1316 		} else {
1317 			ep->last = 0;
1318 		}
1319 
1320 		/* a request already were transmitted completely */
1321 		if ((ep->tx_req->req.length - ep->sent) <= 0) {
1322 			ep->tx_req->req.actual = (unsigned int)ep->sent;
1323 			ep0_req_complete(ep->udc, ep->tx_req);
1324 			ep->tx_req = NULL;
1325 			ep->last = 0;
1326 			ep->sent = 0;
1327 		}
1328 	} else {
1329 		dev_vdbg(ep->udc->dev, "the ep0_controller have no req\n");
1330 	}
1331 
1332 	return 0;
1333 }
1334 
ep0_txframe_handle(struct qe_ep * ep)1335 static int ep0_txframe_handle(struct qe_ep *ep)
1336 {
1337 	/* if have error, transmit again */
1338 	if (frame_get_status(ep->txframe) & FRAME_ERROR) {
1339 		qe_ep_flushtxfifo(ep);
1340 		dev_vdbg(ep->udc->dev, "The EP0 transmit data have error!\n");
1341 		if (frame_get_info(ep->txframe) & PID_DATA0)
1342 			ep->data01 = 0;
1343 		else
1344 			ep->data01 = 1;
1345 
1346 		ep0_txcomplete(ep, 1);
1347 	} else
1348 		ep0_txcomplete(ep, 0);
1349 
1350 	frame_create_tx(ep, ep->txframe);
1351 	return 0;
1352 }
1353 
qe_ep0_txconf(struct qe_ep * ep)1354 static int qe_ep0_txconf(struct qe_ep *ep)
1355 {
1356 	struct qe_bd __iomem *bd;
1357 	struct qe_frame *pframe;
1358 	u32 bdstatus;
1359 
1360 	bd = ep->c_txbd;
1361 	bdstatus = in_be32((u32 __iomem *)bd);
1362 	while (!(bdstatus & T_R) && (bdstatus & ~T_W)) {
1363 		pframe = ep->txframe;
1364 
1365 		/* clear and recycle the BD */
1366 		out_be32((u32 __iomem *)bd, bdstatus & T_W);
1367 		out_be32(&bd->buf, 0);
1368 		if (bdstatus & T_W)
1369 			ep->c_txbd = ep->txbase;
1370 		else
1371 			ep->c_txbd++;
1372 
1373 		if (ep->c_txbd == ep->n_txbd) {
1374 			if (bdstatus & DEVICE_T_ERROR) {
1375 				frame_set_status(pframe, FRAME_ERROR);
1376 				if (bdstatus & T_TO)
1377 					pframe->status |= TX_ER_TIMEOUT;
1378 				if (bdstatus & T_UN)
1379 					pframe->status |= TX_ER_UNDERUN;
1380 			}
1381 			ep0_txframe_handle(ep);
1382 		}
1383 
1384 		bd = ep->c_txbd;
1385 		bdstatus = in_be32((u32 __iomem *)bd);
1386 	}
1387 
1388 	return 0;
1389 }
1390 
ep_txframe_handle(struct qe_ep * ep)1391 static int ep_txframe_handle(struct qe_ep *ep)
1392 {
1393 	if (frame_get_status(ep->txframe) & FRAME_ERROR) {
1394 		qe_ep_flushtxfifo(ep);
1395 		dev_vdbg(ep->udc->dev, "The EP0 transmit data have error!\n");
1396 		if (frame_get_info(ep->txframe) & PID_DATA0)
1397 			ep->data01 = 0;
1398 		else
1399 			ep->data01 = 1;
1400 
1401 		txcomplete(ep, 1);
1402 	} else
1403 		txcomplete(ep, 0);
1404 
1405 	frame_create_tx(ep, ep->txframe); /* send the data */
1406 	return 0;
1407 }
1408 
1409 /* confirm the already trainsmited bd */
qe_ep_txconf(struct qe_ep * ep)1410 static int qe_ep_txconf(struct qe_ep *ep)
1411 {
1412 	struct qe_bd __iomem *bd;
1413 	struct qe_frame *pframe = NULL;
1414 	u32 bdstatus;
1415 	unsigned char breakonrxinterrupt = 0;
1416 
1417 	bd = ep->c_txbd;
1418 	bdstatus = in_be32((u32 __iomem *)bd);
1419 	while (!(bdstatus & T_R) && (bdstatus & ~T_W)) {
1420 		pframe = ep->txframe;
1421 		if (bdstatus & DEVICE_T_ERROR) {
1422 			frame_set_status(pframe, FRAME_ERROR);
1423 			if (bdstatus & T_TO)
1424 				pframe->status |= TX_ER_TIMEOUT;
1425 			if (bdstatus & T_UN)
1426 				pframe->status |= TX_ER_UNDERUN;
1427 		}
1428 
1429 		/* clear and recycle the BD */
1430 		out_be32((u32 __iomem *)bd, bdstatus & T_W);
1431 		out_be32(&bd->buf, 0);
1432 		if (bdstatus & T_W)
1433 			ep->c_txbd = ep->txbase;
1434 		else
1435 			ep->c_txbd++;
1436 
1437 		/* handle the tx frame */
1438 		ep_txframe_handle(ep);
1439 		bd = ep->c_txbd;
1440 		bdstatus = in_be32((u32 __iomem *)bd);
1441 	}
1442 	if (breakonrxinterrupt)
1443 		return -EIO;
1444 	else
1445 		return 0;
1446 }
1447 
1448 /* Add a request in queue, and try to transmit a packet */
ep_req_send(struct qe_ep * ep,struct qe_req * req)1449 static int ep_req_send(struct qe_ep *ep, struct qe_req *req)
1450 {
1451 	int reval = 0;
1452 
1453 	if (ep->tx_req == NULL) {
1454 		ep->sent = 0;
1455 		ep->last = 0;
1456 		txcomplete(ep, 0); /* can gain a new tx_req */
1457 		reval = frame_create_tx(ep, ep->txframe);
1458 	}
1459 	return reval;
1460 }
1461 
1462 /* Maybe this is a good ideal */
ep_req_rx(struct qe_ep * ep,struct qe_req * req)1463 static int ep_req_rx(struct qe_ep *ep, struct qe_req *req)
1464 {
1465 	struct qe_udc *udc = ep->udc;
1466 	struct qe_frame *pframe = NULL;
1467 	struct qe_bd __iomem *bd;
1468 	u32 bdstatus, length;
1469 	u32 vaddr, fsize;
1470 	u8 *cp;
1471 	u8 finish_req = 0;
1472 	u8 framepid;
1473 
1474 	if (list_empty(&ep->queue)) {
1475 		dev_vdbg(udc->dev, "the req already finish!\n");
1476 		return 0;
1477 	}
1478 	pframe = ep->rxframe;
1479 
1480 	bd = ep->n_rxbd;
1481 	bdstatus = in_be32((u32 __iomem *)bd);
1482 	length = bdstatus & BD_LENGTH_MASK;
1483 
1484 	while (!(bdstatus & R_E) && length) {
1485 		if (finish_req)
1486 			break;
1487 		if ((bdstatus & R_F) && (bdstatus & R_L)
1488 					&& !(bdstatus & R_ERROR)) {
1489 			qe_frame_clean(pframe);
1490 			vaddr = (u32)phys_to_virt(in_be32(&bd->buf));
1491 			frame_set_data(pframe, (u8 *)vaddr);
1492 			frame_set_length(pframe, (length - USB_CRC_SIZE));
1493 			frame_set_status(pframe, FRAME_OK);
1494 			switch (bdstatus & R_PID) {
1495 			case R_PID_DATA1:
1496 				frame_set_info(pframe, PID_DATA1); break;
1497 			default:
1498 				frame_set_info(pframe, PID_DATA0); break;
1499 			}
1500 			/* handle the rx frame */
1501 
1502 			if (frame_get_info(pframe) & PID_DATA1)
1503 				framepid = 0x1;
1504 			else
1505 				framepid = 0;
1506 
1507 			if (framepid != ep->data01) {
1508 				dev_vdbg(udc->dev, "the data01 error!\n");
1509 			} else {
1510 				fsize = frame_get_length(pframe);
1511 
1512 				cp = (u8 *)(req->req.buf) + req->req.actual;
1513 				if (cp) {
1514 					memcpy(cp, pframe->data, fsize);
1515 					req->req.actual += fsize;
1516 					if ((fsize < ep->ep.maxpacket)
1517 						|| (req->req.actual >=
1518 							req->req.length)) {
1519 						finish_req = 1;
1520 						done(ep, req, 0);
1521 						if (list_empty(&ep->queue))
1522 							qe_eprx_nack(ep);
1523 					}
1524 				}
1525 				qe_ep_toggledata01(ep);
1526 			}
1527 		} else {
1528 			dev_err(udc->dev, "The receive frame with error!\n");
1529 		}
1530 
1531 		/* note: don't clear the rxbd's buffer address *
1532 		 * only Clear the length */
1533 		out_be32((u32 __iomem *)bd, (bdstatus & BD_STATUS_MASK));
1534 		ep->has_data--;
1535 
1536 		/* Get next BD */
1537 		if (bdstatus & R_W)
1538 			bd = ep->rxbase;
1539 		else
1540 			bd++;
1541 
1542 		bdstatus = in_be32((u32 __iomem *)bd);
1543 		length = bdstatus & BD_LENGTH_MASK;
1544 	}
1545 
1546 	ep->n_rxbd = bd;
1547 	ep_recycle_rxbds(ep);
1548 
1549 	return 0;
1550 }
1551 
1552 /* only add the request in queue */
ep_req_receive(struct qe_ep * ep,struct qe_req * req)1553 static int ep_req_receive(struct qe_ep *ep, struct qe_req *req)
1554 {
1555 	if (ep->state == EP_STATE_NACK) {
1556 		if (ep->has_data <= 0) {
1557 			/* Enable rx and unmask rx interrupt */
1558 			qe_eprx_normal(ep);
1559 		} else {
1560 			/* Copy the exist BD data */
1561 			ep_req_rx(ep, req);
1562 		}
1563 	}
1564 
1565 	return 0;
1566 }
1567 
1568 /********************************************************************
1569 	Internal Used Function End
1570 ********************************************************************/
1571 
1572 /*-----------------------------------------------------------------------
1573 	Endpoint Management Functions For Gadget
1574  -----------------------------------------------------------------------*/
qe_ep_enable(struct usb_ep * _ep,const struct usb_endpoint_descriptor * desc)1575 static int qe_ep_enable(struct usb_ep *_ep,
1576 			 const struct usb_endpoint_descriptor *desc)
1577 {
1578 	struct qe_udc *udc;
1579 	struct qe_ep *ep;
1580 	int retval = 0;
1581 	unsigned char epnum;
1582 
1583 	ep = container_of(_ep, struct qe_ep, ep);
1584 
1585 	/* catch various bogus parameters */
1586 	if (!_ep || !desc || _ep->name == ep_name[0] ||
1587 			(desc->bDescriptorType != USB_DT_ENDPOINT))
1588 		return -EINVAL;
1589 
1590 	udc = ep->udc;
1591 	if (!udc->driver || (udc->gadget.speed == USB_SPEED_UNKNOWN))
1592 		return -ESHUTDOWN;
1593 
1594 	epnum = (u8)desc->bEndpointAddress & 0xF;
1595 
1596 	retval = qe_ep_init(udc, epnum, desc);
1597 	if (retval != 0) {
1598 		cpm_muram_free(cpm_muram_offset(ep->rxbase));
1599 		dev_dbg(udc->dev, "enable ep%d failed\n", ep->epnum);
1600 		return -EINVAL;
1601 	}
1602 	dev_dbg(udc->dev, "enable ep%d successful\n", ep->epnum);
1603 	return 0;
1604 }
1605 
qe_ep_disable(struct usb_ep * _ep)1606 static int qe_ep_disable(struct usb_ep *_ep)
1607 {
1608 	struct qe_udc *udc;
1609 	struct qe_ep *ep;
1610 	unsigned long flags;
1611 	unsigned int size;
1612 
1613 	ep = container_of(_ep, struct qe_ep, ep);
1614 	udc = ep->udc;
1615 
1616 	if (!_ep || !ep->ep.desc) {
1617 		dev_dbg(udc->dev, "%s not enabled\n", _ep ? ep->ep.name : NULL);
1618 		return -EINVAL;
1619 	}
1620 
1621 	spin_lock_irqsave(&udc->lock, flags);
1622 	/* Nuke all pending requests (does flush) */
1623 	nuke(ep, -ESHUTDOWN);
1624 	ep->ep.desc = NULL;
1625 	ep->stopped = 1;
1626 	ep->tx_req = NULL;
1627 	qe_ep_reset(udc, ep->epnum);
1628 	spin_unlock_irqrestore(&udc->lock, flags);
1629 
1630 	cpm_muram_free(cpm_muram_offset(ep->rxbase));
1631 
1632 	if (ep->dir == USB_DIR_OUT)
1633 		size = (ep->ep.maxpacket + USB_CRC_SIZE + 2) *
1634 				(USB_BDRING_LEN_RX + 1);
1635 	else
1636 		size = (ep->ep.maxpacket + USB_CRC_SIZE + 2) *
1637 				(USB_BDRING_LEN + 1);
1638 
1639 	if (ep->dir != USB_DIR_IN) {
1640 		kfree(ep->rxframe);
1641 		if (ep->rxbufmap) {
1642 			dma_unmap_single(udc->gadget.dev.parent,
1643 					ep->rxbuf_d, size,
1644 					DMA_FROM_DEVICE);
1645 			ep->rxbuf_d = DMA_ADDR_INVALID;
1646 		} else {
1647 			dma_sync_single_for_cpu(
1648 					udc->gadget.dev.parent,
1649 					ep->rxbuf_d, size,
1650 					DMA_FROM_DEVICE);
1651 		}
1652 		kfree(ep->rxbuffer);
1653 	}
1654 
1655 	if (ep->dir != USB_DIR_OUT)
1656 		kfree(ep->txframe);
1657 
1658 	dev_dbg(udc->dev, "disabled %s OK\n", _ep->name);
1659 	return 0;
1660 }
1661 
qe_alloc_request(struct usb_ep * _ep,gfp_t gfp_flags)1662 static struct usb_request *qe_alloc_request(struct usb_ep *_ep,	gfp_t gfp_flags)
1663 {
1664 	struct qe_req *req;
1665 
1666 	req = kzalloc(sizeof(*req), gfp_flags);
1667 	if (!req)
1668 		return NULL;
1669 
1670 	req->req.dma = DMA_ADDR_INVALID;
1671 
1672 	INIT_LIST_HEAD(&req->queue);
1673 
1674 	return &req->req;
1675 }
1676 
qe_free_request(struct usb_ep * _ep,struct usb_request * _req)1677 static void qe_free_request(struct usb_ep *_ep, struct usb_request *_req)
1678 {
1679 	struct qe_req *req;
1680 
1681 	req = container_of(_req, struct qe_req, req);
1682 
1683 	if (_req)
1684 		kfree(req);
1685 }
1686 
__qe_ep_queue(struct usb_ep * _ep,struct usb_request * _req)1687 static int __qe_ep_queue(struct usb_ep *_ep, struct usb_request *_req)
1688 {
1689 	struct qe_ep *ep = container_of(_ep, struct qe_ep, ep);
1690 	struct qe_req *req = container_of(_req, struct qe_req, req);
1691 	struct qe_udc *udc;
1692 	int reval;
1693 
1694 	udc = ep->udc;
1695 	/* catch various bogus parameters */
1696 	if (!_req || !req->req.complete || !req->req.buf
1697 			|| !list_empty(&req->queue)) {
1698 		dev_dbg(udc->dev, "bad params\n");
1699 		return -EINVAL;
1700 	}
1701 	if (!_ep || (!ep->ep.desc && ep_index(ep))) {
1702 		dev_dbg(udc->dev, "bad ep\n");
1703 		return -EINVAL;
1704 	}
1705 
1706 	if (!udc->driver || udc->gadget.speed == USB_SPEED_UNKNOWN)
1707 		return -ESHUTDOWN;
1708 
1709 	req->ep = ep;
1710 
1711 	/* map virtual address to hardware */
1712 	if (req->req.dma == DMA_ADDR_INVALID) {
1713 		req->req.dma = dma_map_single(ep->udc->gadget.dev.parent,
1714 					req->req.buf,
1715 					req->req.length,
1716 					ep_is_in(ep)
1717 					? DMA_TO_DEVICE :
1718 					DMA_FROM_DEVICE);
1719 		req->mapped = 1;
1720 	} else {
1721 		dma_sync_single_for_device(ep->udc->gadget.dev.parent,
1722 					req->req.dma, req->req.length,
1723 					ep_is_in(ep)
1724 					? DMA_TO_DEVICE :
1725 					DMA_FROM_DEVICE);
1726 		req->mapped = 0;
1727 	}
1728 
1729 	req->req.status = -EINPROGRESS;
1730 	req->req.actual = 0;
1731 
1732 	list_add_tail(&req->queue, &ep->queue);
1733 	dev_vdbg(udc->dev, "gadget have request in %s! %d\n",
1734 			ep->name, req->req.length);
1735 
1736 	/* push the request to device */
1737 	if (ep_is_in(ep))
1738 		reval = ep_req_send(ep, req);
1739 
1740 	/* EP0 */
1741 	if (ep_index(ep) == 0 && req->req.length > 0) {
1742 		if (ep_is_in(ep))
1743 			udc->ep0_state = DATA_STATE_XMIT;
1744 		else
1745 			udc->ep0_state = DATA_STATE_RECV;
1746 	}
1747 
1748 	if (ep->dir == USB_DIR_OUT)
1749 		reval = ep_req_receive(ep, req);
1750 
1751 	return 0;
1752 }
1753 
1754 /* queues (submits) an I/O request to an endpoint */
qe_ep_queue(struct usb_ep * _ep,struct usb_request * _req,gfp_t gfp_flags)1755 static int qe_ep_queue(struct usb_ep *_ep, struct usb_request *_req,
1756 		       gfp_t gfp_flags)
1757 {
1758 	struct qe_ep *ep = container_of(_ep, struct qe_ep, ep);
1759 	struct qe_udc *udc = ep->udc;
1760 	unsigned long flags;
1761 	int ret;
1762 
1763 	spin_lock_irqsave(&udc->lock, flags);
1764 	ret = __qe_ep_queue(_ep, _req);
1765 	spin_unlock_irqrestore(&udc->lock, flags);
1766 	return ret;
1767 }
1768 
1769 /* dequeues (cancels, unlinks) an I/O request from an endpoint */
qe_ep_dequeue(struct usb_ep * _ep,struct usb_request * _req)1770 static int qe_ep_dequeue(struct usb_ep *_ep, struct usb_request *_req)
1771 {
1772 	struct qe_ep *ep = container_of(_ep, struct qe_ep, ep);
1773 	struct qe_req *req;
1774 	unsigned long flags;
1775 
1776 	if (!_ep || !_req)
1777 		return -EINVAL;
1778 
1779 	spin_lock_irqsave(&ep->udc->lock, flags);
1780 
1781 	/* make sure it's actually queued on this endpoint */
1782 	list_for_each_entry(req, &ep->queue, queue) {
1783 		if (&req->req == _req)
1784 			break;
1785 	}
1786 
1787 	if (&req->req != _req) {
1788 		spin_unlock_irqrestore(&ep->udc->lock, flags);
1789 		return -EINVAL;
1790 	}
1791 
1792 	done(ep, req, -ECONNRESET);
1793 
1794 	spin_unlock_irqrestore(&ep->udc->lock, flags);
1795 	return 0;
1796 }
1797 
1798 /*-----------------------------------------------------------------
1799  * modify the endpoint halt feature
1800  * @ep: the non-isochronous endpoint being stalled
1801  * @value: 1--set halt  0--clear halt
1802  * Returns zero, or a negative error code.
1803 *----------------------------------------------------------------*/
qe_ep_set_halt(struct usb_ep * _ep,int value)1804 static int qe_ep_set_halt(struct usb_ep *_ep, int value)
1805 {
1806 	struct qe_ep *ep;
1807 	unsigned long flags;
1808 	int status = -EOPNOTSUPP;
1809 	struct qe_udc *udc;
1810 
1811 	ep = container_of(_ep, struct qe_ep, ep);
1812 	if (!_ep || !ep->ep.desc) {
1813 		status = -EINVAL;
1814 		goto out;
1815 	}
1816 
1817 	udc = ep->udc;
1818 	/* Attempt to halt IN ep will fail if any transfer requests
1819 	 * are still queue */
1820 	if (value && ep_is_in(ep) && !list_empty(&ep->queue)) {
1821 		status = -EAGAIN;
1822 		goto out;
1823 	}
1824 
1825 	status = 0;
1826 	spin_lock_irqsave(&ep->udc->lock, flags);
1827 	qe_eptx_stall_change(ep, value);
1828 	qe_eprx_stall_change(ep, value);
1829 	spin_unlock_irqrestore(&ep->udc->lock, flags);
1830 
1831 	if (ep->epnum == 0) {
1832 		udc->ep0_state = WAIT_FOR_SETUP;
1833 		udc->ep0_dir = 0;
1834 	}
1835 
1836 	/* set data toggle to DATA0 on clear halt */
1837 	if (value == 0)
1838 		ep->data01 = 0;
1839 out:
1840 	dev_vdbg(udc->dev, "%s %s halt stat %d\n", ep->ep.name,
1841 			value ?  "set" : "clear", status);
1842 
1843 	return status;
1844 }
1845 
1846 static const struct usb_ep_ops qe_ep_ops = {
1847 	.enable = qe_ep_enable,
1848 	.disable = qe_ep_disable,
1849 
1850 	.alloc_request = qe_alloc_request,
1851 	.free_request = qe_free_request,
1852 
1853 	.queue = qe_ep_queue,
1854 	.dequeue = qe_ep_dequeue,
1855 
1856 	.set_halt = qe_ep_set_halt,
1857 };
1858 
1859 /*------------------------------------------------------------------------
1860 	Gadget Driver Layer Operations
1861  ------------------------------------------------------------------------*/
1862 
1863 /* Get the current frame number */
qe_get_frame(struct usb_gadget * gadget)1864 static int qe_get_frame(struct usb_gadget *gadget)
1865 {
1866 	struct qe_udc *udc = container_of(gadget, struct qe_udc, gadget);
1867 	u16 tmp;
1868 
1869 	tmp = in_be16(&udc->usb_param->frame_n);
1870 	if (tmp & 0x8000)
1871 		return tmp & 0x07ff;
1872 	return -EINVAL;
1873 }
1874 
1875 static int fsl_qe_start(struct usb_gadget *gadget,
1876 		struct usb_gadget_driver *driver);
1877 static int fsl_qe_stop(struct usb_gadget *gadget);
1878 
1879 /* defined in usb_gadget.h */
1880 static const struct usb_gadget_ops qe_gadget_ops = {
1881 	.get_frame = qe_get_frame,
1882 	.udc_start = fsl_qe_start,
1883 	.udc_stop = fsl_qe_stop,
1884 };
1885 
1886 /*-------------------------------------------------------------------------
1887 	USB ep0 Setup process in BUS Enumeration
1888  -------------------------------------------------------------------------*/
udc_reset_ep_queue(struct qe_udc * udc,u8 pipe)1889 static int udc_reset_ep_queue(struct qe_udc *udc, u8 pipe)
1890 {
1891 	struct qe_ep *ep = &udc->eps[pipe];
1892 
1893 	nuke(ep, -ECONNRESET);
1894 	ep->tx_req = NULL;
1895 	return 0;
1896 }
1897 
reset_queues(struct qe_udc * udc)1898 static int reset_queues(struct qe_udc *udc)
1899 {
1900 	u8 pipe;
1901 
1902 	for (pipe = 0; pipe < USB_MAX_ENDPOINTS; pipe++)
1903 		udc_reset_ep_queue(udc, pipe);
1904 
1905 	/* report disconnect; the driver is already quiesced */
1906 	spin_unlock(&udc->lock);
1907 	usb_gadget_udc_reset(&udc->gadget, udc->driver);
1908 	spin_lock(&udc->lock);
1909 
1910 	return 0;
1911 }
1912 
ch9setaddress(struct qe_udc * udc,u16 value,u16 index,u16 length)1913 static void ch9setaddress(struct qe_udc *udc, u16 value, u16 index,
1914 			u16 length)
1915 {
1916 	/* Save the new address to device struct */
1917 	udc->device_address = (u8) value;
1918 	/* Update usb state */
1919 	udc->usb_state = USB_STATE_ADDRESS;
1920 
1921 	/* Status phase , send a ZLP */
1922 	if (ep0_prime_status(udc, USB_DIR_IN))
1923 		qe_ep0_stall(udc);
1924 }
1925 
ownercomplete(struct usb_ep * _ep,struct usb_request * _req)1926 static void ownercomplete(struct usb_ep *_ep, struct usb_request *_req)
1927 {
1928 	struct qe_req *req = container_of(_req, struct qe_req, req);
1929 
1930 	req->req.buf = NULL;
1931 	kfree(req);
1932 }
1933 
ch9getstatus(struct qe_udc * udc,u8 request_type,u16 value,u16 index,u16 length)1934 static void ch9getstatus(struct qe_udc *udc, u8 request_type, u16 value,
1935 			u16 index, u16 length)
1936 {
1937 	u16 usb_status = 0;
1938 	struct qe_req *req;
1939 	struct qe_ep *ep;
1940 	int status = 0;
1941 
1942 	ep = &udc->eps[0];
1943 	if ((request_type & USB_RECIP_MASK) == USB_RECIP_DEVICE) {
1944 		/* Get device status */
1945 		usb_status = 1 << USB_DEVICE_SELF_POWERED;
1946 	} else if ((request_type & USB_RECIP_MASK) == USB_RECIP_INTERFACE) {
1947 		/* Get interface status */
1948 		/* We don't have interface information in udc driver */
1949 		usb_status = 0;
1950 	} else if ((request_type & USB_RECIP_MASK) == USB_RECIP_ENDPOINT) {
1951 		/* Get endpoint status */
1952 		int pipe = index & USB_ENDPOINT_NUMBER_MASK;
1953 		struct qe_ep *target_ep = &udc->eps[pipe];
1954 		u16 usep;
1955 
1956 		/* stall if endpoint doesn't exist */
1957 		if (!target_ep->ep.desc)
1958 			goto stall;
1959 
1960 		usep = in_be16(&udc->usb_regs->usb_usep[pipe]);
1961 		if (index & USB_DIR_IN) {
1962 			if (target_ep->dir != USB_DIR_IN)
1963 				goto stall;
1964 			if ((usep & USB_THS_MASK) == USB_THS_STALL)
1965 				usb_status = 1 << USB_ENDPOINT_HALT;
1966 		} else {
1967 			if (target_ep->dir != USB_DIR_OUT)
1968 				goto stall;
1969 			if ((usep & USB_RHS_MASK) == USB_RHS_STALL)
1970 				usb_status = 1 << USB_ENDPOINT_HALT;
1971 		}
1972 	}
1973 
1974 	req = container_of(qe_alloc_request(&ep->ep, GFP_KERNEL),
1975 					struct qe_req, req);
1976 	req->req.length = 2;
1977 	req->req.buf = udc->statusbuf;
1978 	*(u16 *)req->req.buf = cpu_to_le16(usb_status);
1979 	req->req.status = -EINPROGRESS;
1980 	req->req.actual = 0;
1981 	req->req.complete = ownercomplete;
1982 
1983 	udc->ep0_dir = USB_DIR_IN;
1984 
1985 	/* data phase */
1986 	status = __qe_ep_queue(&ep->ep, &req->req);
1987 
1988 	if (status == 0)
1989 		return;
1990 stall:
1991 	dev_err(udc->dev, "Can't respond to getstatus request \n");
1992 	qe_ep0_stall(udc);
1993 }
1994 
1995 /* only handle the setup request, suppose the device in normal status */
setup_received_handle(struct qe_udc * udc,struct usb_ctrlrequest * setup)1996 static void setup_received_handle(struct qe_udc *udc,
1997 				struct usb_ctrlrequest *setup)
1998 {
1999 	/* Fix Endian (udc->local_setup_buff is cpu Endian now)*/
2000 	u16 wValue = le16_to_cpu(setup->wValue);
2001 	u16 wIndex = le16_to_cpu(setup->wIndex);
2002 	u16 wLength = le16_to_cpu(setup->wLength);
2003 
2004 	/* clear the previous request in the ep0 */
2005 	udc_reset_ep_queue(udc, 0);
2006 
2007 	if (setup->bRequestType & USB_DIR_IN)
2008 		udc->ep0_dir = USB_DIR_IN;
2009 	else
2010 		udc->ep0_dir = USB_DIR_OUT;
2011 
2012 	switch (setup->bRequest) {
2013 	case USB_REQ_GET_STATUS:
2014 		/* Data+Status phase form udc */
2015 		if ((setup->bRequestType & (USB_DIR_IN | USB_TYPE_MASK))
2016 					!= (USB_DIR_IN | USB_TYPE_STANDARD))
2017 			break;
2018 		ch9getstatus(udc, setup->bRequestType, wValue, wIndex,
2019 					wLength);
2020 		return;
2021 
2022 	case USB_REQ_SET_ADDRESS:
2023 		/* Status phase from udc */
2024 		if (setup->bRequestType != (USB_DIR_OUT | USB_TYPE_STANDARD |
2025 						USB_RECIP_DEVICE))
2026 			break;
2027 		ch9setaddress(udc, wValue, wIndex, wLength);
2028 		return;
2029 
2030 	case USB_REQ_CLEAR_FEATURE:
2031 	case USB_REQ_SET_FEATURE:
2032 		/* Requests with no data phase, status phase from udc */
2033 		if ((setup->bRequestType & USB_TYPE_MASK)
2034 					!= USB_TYPE_STANDARD)
2035 			break;
2036 
2037 		if ((setup->bRequestType & USB_RECIP_MASK)
2038 				== USB_RECIP_ENDPOINT) {
2039 			int pipe = wIndex & USB_ENDPOINT_NUMBER_MASK;
2040 			struct qe_ep *ep;
2041 
2042 			if (wValue != 0 || wLength != 0
2043 				|| pipe >= USB_MAX_ENDPOINTS)
2044 				break;
2045 			ep = &udc->eps[pipe];
2046 
2047 			spin_unlock(&udc->lock);
2048 			qe_ep_set_halt(&ep->ep,
2049 					(setup->bRequest == USB_REQ_SET_FEATURE)
2050 						? 1 : 0);
2051 			spin_lock(&udc->lock);
2052 		}
2053 
2054 		ep0_prime_status(udc, USB_DIR_IN);
2055 
2056 		return;
2057 
2058 	default:
2059 		break;
2060 	}
2061 
2062 	if (wLength) {
2063 		/* Data phase from gadget, status phase from udc */
2064 		if (setup->bRequestType & USB_DIR_IN) {
2065 			udc->ep0_state = DATA_STATE_XMIT;
2066 			udc->ep0_dir = USB_DIR_IN;
2067 		} else {
2068 			udc->ep0_state = DATA_STATE_RECV;
2069 			udc->ep0_dir = USB_DIR_OUT;
2070 		}
2071 		spin_unlock(&udc->lock);
2072 		if (udc->driver->setup(&udc->gadget,
2073 					&udc->local_setup_buff) < 0)
2074 			qe_ep0_stall(udc);
2075 		spin_lock(&udc->lock);
2076 	} else {
2077 		/* No data phase, IN status from gadget */
2078 		udc->ep0_dir = USB_DIR_IN;
2079 		spin_unlock(&udc->lock);
2080 		if (udc->driver->setup(&udc->gadget,
2081 					&udc->local_setup_buff) < 0)
2082 			qe_ep0_stall(udc);
2083 		spin_lock(&udc->lock);
2084 		udc->ep0_state = DATA_STATE_NEED_ZLP;
2085 	}
2086 }
2087 
2088 /*-------------------------------------------------------------------------
2089 	USB Interrupt handlers
2090  -------------------------------------------------------------------------*/
suspend_irq(struct qe_udc * udc)2091 static void suspend_irq(struct qe_udc *udc)
2092 {
2093 	udc->resume_state = udc->usb_state;
2094 	udc->usb_state = USB_STATE_SUSPENDED;
2095 
2096 	/* report suspend to the driver ,serial.c not support this*/
2097 	if (udc->driver->suspend)
2098 		udc->driver->suspend(&udc->gadget);
2099 }
2100 
resume_irq(struct qe_udc * udc)2101 static void resume_irq(struct qe_udc *udc)
2102 {
2103 	udc->usb_state = udc->resume_state;
2104 	udc->resume_state = 0;
2105 
2106 	/* report resume to the driver , serial.c not support this*/
2107 	if (udc->driver->resume)
2108 		udc->driver->resume(&udc->gadget);
2109 }
2110 
idle_irq(struct qe_udc * udc)2111 static void idle_irq(struct qe_udc *udc)
2112 {
2113 	u8 usbs;
2114 
2115 	usbs = in_8(&udc->usb_regs->usb_usbs);
2116 	if (usbs & USB_IDLE_STATUS_MASK) {
2117 		if ((udc->usb_state) != USB_STATE_SUSPENDED)
2118 			suspend_irq(udc);
2119 	} else {
2120 		if (udc->usb_state == USB_STATE_SUSPENDED)
2121 			resume_irq(udc);
2122 	}
2123 }
2124 
reset_irq(struct qe_udc * udc)2125 static int reset_irq(struct qe_udc *udc)
2126 {
2127 	unsigned char i;
2128 
2129 	if (udc->usb_state == USB_STATE_DEFAULT)
2130 		return 0;
2131 
2132 	qe_usb_disable(udc);
2133 	out_8(&udc->usb_regs->usb_usadr, 0);
2134 
2135 	for (i = 0; i < USB_MAX_ENDPOINTS; i++) {
2136 		if (udc->eps[i].init)
2137 			qe_ep_reset(udc, i);
2138 	}
2139 
2140 	reset_queues(udc);
2141 	udc->usb_state = USB_STATE_DEFAULT;
2142 	udc->ep0_state = WAIT_FOR_SETUP;
2143 	udc->ep0_dir = USB_DIR_OUT;
2144 	qe_usb_enable(udc);
2145 	return 0;
2146 }
2147 
bsy_irq(struct qe_udc * udc)2148 static int bsy_irq(struct qe_udc *udc)
2149 {
2150 	return 0;
2151 }
2152 
txe_irq(struct qe_udc * udc)2153 static int txe_irq(struct qe_udc *udc)
2154 {
2155 	return 0;
2156 }
2157 
2158 /* ep0 tx interrupt also in here */
tx_irq(struct qe_udc * udc)2159 static int tx_irq(struct qe_udc *udc)
2160 {
2161 	struct qe_ep *ep;
2162 	struct qe_bd __iomem *bd;
2163 	int i, res = 0;
2164 
2165 	if ((udc->usb_state == USB_STATE_ADDRESS)
2166 		&& (in_8(&udc->usb_regs->usb_usadr) == 0))
2167 		out_8(&udc->usb_regs->usb_usadr, udc->device_address);
2168 
2169 	for (i = (USB_MAX_ENDPOINTS-1); ((i >= 0) && (res == 0)); i--) {
2170 		ep = &udc->eps[i];
2171 		if (ep && ep->init && (ep->dir != USB_DIR_OUT)) {
2172 			bd = ep->c_txbd;
2173 			if (!(in_be32((u32 __iomem *)bd) & T_R)
2174 						&& (in_be32(&bd->buf))) {
2175 				/* confirm the transmitted bd */
2176 				if (ep->epnum == 0)
2177 					res = qe_ep0_txconf(ep);
2178 				else
2179 					res = qe_ep_txconf(ep);
2180 			}
2181 		}
2182 	}
2183 	return res;
2184 }
2185 
2186 
2187 /* setup packect's rx is handle in the function too */
rx_irq(struct qe_udc * udc)2188 static void rx_irq(struct qe_udc *udc)
2189 {
2190 	struct qe_ep *ep;
2191 	struct qe_bd __iomem *bd;
2192 	int i;
2193 
2194 	for (i = 0; i < USB_MAX_ENDPOINTS; i++) {
2195 		ep = &udc->eps[i];
2196 		if (ep && ep->init && (ep->dir != USB_DIR_IN)) {
2197 			bd = ep->n_rxbd;
2198 			if (!(in_be32((u32 __iomem *)bd) & R_E)
2199 						&& (in_be32(&bd->buf))) {
2200 				if (ep->epnum == 0) {
2201 					qe_ep0_rx(udc);
2202 				} else {
2203 					/*non-setup package receive*/
2204 					qe_ep_rx(ep);
2205 				}
2206 			}
2207 		}
2208 	}
2209 }
2210 
qe_udc_irq(int irq,void * _udc)2211 static irqreturn_t qe_udc_irq(int irq, void *_udc)
2212 {
2213 	struct qe_udc *udc = (struct qe_udc *)_udc;
2214 	u16 irq_src;
2215 	irqreturn_t status = IRQ_NONE;
2216 	unsigned long flags;
2217 
2218 	spin_lock_irqsave(&udc->lock, flags);
2219 
2220 	irq_src = in_be16(&udc->usb_regs->usb_usber) &
2221 		in_be16(&udc->usb_regs->usb_usbmr);
2222 	/* Clear notification bits */
2223 	out_be16(&udc->usb_regs->usb_usber, irq_src);
2224 	/* USB Interrupt */
2225 	if (irq_src & USB_E_IDLE_MASK) {
2226 		idle_irq(udc);
2227 		irq_src &= ~USB_E_IDLE_MASK;
2228 		status = IRQ_HANDLED;
2229 	}
2230 
2231 	if (irq_src & USB_E_TXB_MASK) {
2232 		tx_irq(udc);
2233 		irq_src &= ~USB_E_TXB_MASK;
2234 		status = IRQ_HANDLED;
2235 	}
2236 
2237 	if (irq_src & USB_E_RXB_MASK) {
2238 		rx_irq(udc);
2239 		irq_src &= ~USB_E_RXB_MASK;
2240 		status = IRQ_HANDLED;
2241 	}
2242 
2243 	if (irq_src & USB_E_RESET_MASK) {
2244 		reset_irq(udc);
2245 		irq_src &= ~USB_E_RESET_MASK;
2246 		status = IRQ_HANDLED;
2247 	}
2248 
2249 	if (irq_src & USB_E_BSY_MASK) {
2250 		bsy_irq(udc);
2251 		irq_src &= ~USB_E_BSY_MASK;
2252 		status = IRQ_HANDLED;
2253 	}
2254 
2255 	if (irq_src & USB_E_TXE_MASK) {
2256 		txe_irq(udc);
2257 		irq_src &= ~USB_E_TXE_MASK;
2258 		status = IRQ_HANDLED;
2259 	}
2260 
2261 	spin_unlock_irqrestore(&udc->lock, flags);
2262 
2263 	return status;
2264 }
2265 
2266 /*-------------------------------------------------------------------------
2267 	Gadget driver probe and unregister.
2268  --------------------------------------------------------------------------*/
fsl_qe_start(struct usb_gadget * gadget,struct usb_gadget_driver * driver)2269 static int fsl_qe_start(struct usb_gadget *gadget,
2270 		struct usb_gadget_driver *driver)
2271 {
2272 	struct qe_udc *udc;
2273 	unsigned long flags;
2274 
2275 	udc = container_of(gadget, struct qe_udc, gadget);
2276 	/* lock is needed but whether should use this lock or another */
2277 	spin_lock_irqsave(&udc->lock, flags);
2278 
2279 	driver->driver.bus = NULL;
2280 	/* hook up the driver */
2281 	udc->driver = driver;
2282 	udc->gadget.speed = driver->max_speed;
2283 
2284 	/* Enable IRQ reg and Set usbcmd reg EN bit */
2285 	qe_usb_enable(udc);
2286 
2287 	out_be16(&udc->usb_regs->usb_usber, 0xffff);
2288 	out_be16(&udc->usb_regs->usb_usbmr, USB_E_DEFAULT_DEVICE);
2289 	udc->usb_state = USB_STATE_ATTACHED;
2290 	udc->ep0_state = WAIT_FOR_SETUP;
2291 	udc->ep0_dir = USB_DIR_OUT;
2292 	spin_unlock_irqrestore(&udc->lock, flags);
2293 
2294 	return 0;
2295 }
2296 
fsl_qe_stop(struct usb_gadget * gadget)2297 static int fsl_qe_stop(struct usb_gadget *gadget)
2298 {
2299 	struct qe_udc *udc;
2300 	struct qe_ep *loop_ep;
2301 	unsigned long flags;
2302 
2303 	udc = container_of(gadget, struct qe_udc, gadget);
2304 	/* stop usb controller, disable intr */
2305 	qe_usb_disable(udc);
2306 
2307 	/* in fact, no needed */
2308 	udc->usb_state = USB_STATE_ATTACHED;
2309 	udc->ep0_state = WAIT_FOR_SETUP;
2310 	udc->ep0_dir = 0;
2311 
2312 	/* stand operation */
2313 	spin_lock_irqsave(&udc->lock, flags);
2314 	udc->gadget.speed = USB_SPEED_UNKNOWN;
2315 	nuke(&udc->eps[0], -ESHUTDOWN);
2316 	list_for_each_entry(loop_ep, &udc->gadget.ep_list, ep.ep_list)
2317 		nuke(loop_ep, -ESHUTDOWN);
2318 	spin_unlock_irqrestore(&udc->lock, flags);
2319 
2320 	udc->driver = NULL;
2321 
2322 	return 0;
2323 }
2324 
2325 /* udc structure's alloc and setup, include ep-param alloc */
qe_udc_config(struct platform_device * ofdev)2326 static struct qe_udc *qe_udc_config(struct platform_device *ofdev)
2327 {
2328 	struct qe_udc *udc;
2329 	struct device_node *np = ofdev->dev.of_node;
2330 	unsigned long tmp_addr = 0;
2331 	struct usb_device_para __iomem *usbpram;
2332 	unsigned int i;
2333 	u64 size;
2334 	u32 offset;
2335 
2336 	udc = kzalloc(sizeof(*udc), GFP_KERNEL);
2337 	if (!udc)
2338 		goto cleanup;
2339 
2340 	udc->dev = &ofdev->dev;
2341 
2342 	/* get default address of usb parameter in MURAM from device tree */
2343 	offset = *of_get_address(np, 1, &size, NULL);
2344 	udc->usb_param = cpm_muram_addr(offset);
2345 	memset_io(udc->usb_param, 0, size);
2346 
2347 	usbpram = udc->usb_param;
2348 	out_be16(&usbpram->frame_n, 0);
2349 	out_be32(&usbpram->rstate, 0);
2350 
2351 	tmp_addr = cpm_muram_alloc((USB_MAX_ENDPOINTS *
2352 					sizeof(struct usb_ep_para)),
2353 					   USB_EP_PARA_ALIGNMENT);
2354 	if (IS_ERR_VALUE(tmp_addr))
2355 		goto cleanup;
2356 
2357 	for (i = 0; i < USB_MAX_ENDPOINTS; i++) {
2358 		out_be16(&usbpram->epptr[i], (u16)tmp_addr);
2359 		udc->ep_param[i] = cpm_muram_addr(tmp_addr);
2360 		tmp_addr += 32;
2361 	}
2362 
2363 	memset_io(udc->ep_param[0], 0,
2364 			USB_MAX_ENDPOINTS * sizeof(struct usb_ep_para));
2365 
2366 	udc->resume_state = USB_STATE_NOTATTACHED;
2367 	udc->usb_state = USB_STATE_POWERED;
2368 	udc->ep0_dir = 0;
2369 
2370 	spin_lock_init(&udc->lock);
2371 	return udc;
2372 
2373 cleanup:
2374 	kfree(udc);
2375 	return NULL;
2376 }
2377 
2378 /* USB Controller register init */
qe_udc_reg_init(struct qe_udc * udc)2379 static int qe_udc_reg_init(struct qe_udc *udc)
2380 {
2381 	struct usb_ctlr __iomem *qe_usbregs;
2382 	qe_usbregs = udc->usb_regs;
2383 
2384 	/* Spec says that we must enable the USB controller to change mode. */
2385 	out_8(&qe_usbregs->usb_usmod, 0x01);
2386 	/* Mode changed, now disable it, since muram isn't initialized yet. */
2387 	out_8(&qe_usbregs->usb_usmod, 0x00);
2388 
2389 	/* Initialize the rest. */
2390 	out_be16(&qe_usbregs->usb_usbmr, 0);
2391 	out_8(&qe_usbregs->usb_uscom, 0);
2392 	out_be16(&qe_usbregs->usb_usber, USBER_ALL_CLEAR);
2393 
2394 	return 0;
2395 }
2396 
qe_ep_config(struct qe_udc * udc,unsigned char pipe_num)2397 static int qe_ep_config(struct qe_udc *udc, unsigned char pipe_num)
2398 {
2399 	struct qe_ep *ep = &udc->eps[pipe_num];
2400 
2401 	ep->udc = udc;
2402 	strcpy(ep->name, ep_name[pipe_num]);
2403 	ep->ep.name = ep_name[pipe_num];
2404 
2405 	if (pipe_num == 0) {
2406 		ep->ep.caps.type_control = true;
2407 	} else {
2408 		ep->ep.caps.type_iso = true;
2409 		ep->ep.caps.type_bulk = true;
2410 		ep->ep.caps.type_int = true;
2411 	}
2412 
2413 	ep->ep.caps.dir_in = true;
2414 	ep->ep.caps.dir_out = true;
2415 
2416 	ep->ep.ops = &qe_ep_ops;
2417 	ep->stopped = 1;
2418 	usb_ep_set_maxpacket_limit(&ep->ep, (unsigned short) ~0);
2419 	ep->ep.desc = NULL;
2420 	ep->dir = 0xff;
2421 	ep->epnum = (u8)pipe_num;
2422 	ep->sent = 0;
2423 	ep->last = 0;
2424 	ep->init = 0;
2425 	ep->rxframe = NULL;
2426 	ep->txframe = NULL;
2427 	ep->tx_req = NULL;
2428 	ep->state = EP_STATE_IDLE;
2429 	ep->has_data = 0;
2430 
2431 	/* the queue lists any req for this ep */
2432 	INIT_LIST_HEAD(&ep->queue);
2433 
2434 	/* gagdet.ep_list used for ep_autoconfig so no ep0*/
2435 	if (pipe_num != 0)
2436 		list_add_tail(&ep->ep.ep_list, &udc->gadget.ep_list);
2437 
2438 	ep->gadget = &udc->gadget;
2439 
2440 	return 0;
2441 }
2442 
2443 /*-----------------------------------------------------------------------
2444  *	UDC device Driver operation functions				*
2445  *----------------------------------------------------------------------*/
qe_udc_release(struct device * dev)2446 static void qe_udc_release(struct device *dev)
2447 {
2448 	struct qe_udc *udc = container_of(dev, struct qe_udc, gadget.dev);
2449 	int i;
2450 
2451 	complete(udc->done);
2452 	cpm_muram_free(cpm_muram_offset(udc->ep_param[0]));
2453 	for (i = 0; i < USB_MAX_ENDPOINTS; i++)
2454 		udc->ep_param[i] = NULL;
2455 
2456 	kfree(udc);
2457 }
2458 
2459 /* Driver probe functions */
2460 static const struct of_device_id qe_udc_match[];
qe_udc_probe(struct platform_device * ofdev)2461 static int qe_udc_probe(struct platform_device *ofdev)
2462 {
2463 	struct qe_udc *udc;
2464 	const struct of_device_id *match;
2465 	struct device_node *np = ofdev->dev.of_node;
2466 	struct qe_ep *ep;
2467 	unsigned int ret = 0;
2468 	unsigned int i;
2469 	const void *prop;
2470 
2471 	match = of_match_device(qe_udc_match, &ofdev->dev);
2472 	if (!match)
2473 		return -EINVAL;
2474 
2475 	prop = of_get_property(np, "mode", NULL);
2476 	if (!prop || strcmp(prop, "peripheral"))
2477 		return -ENODEV;
2478 
2479 	/* Initialize the udc structure including QH member and other member */
2480 	udc = qe_udc_config(ofdev);
2481 	if (!udc) {
2482 		dev_err(&ofdev->dev, "failed to initialize\n");
2483 		return -ENOMEM;
2484 	}
2485 
2486 	udc->soc_type = (unsigned long)match->data;
2487 	udc->usb_regs = of_iomap(np, 0);
2488 	if (!udc->usb_regs) {
2489 		ret = -ENOMEM;
2490 		goto err1;
2491 	}
2492 
2493 	/* initialize usb hw reg except for regs for EP,
2494 	 * leave usbintr reg untouched*/
2495 	qe_udc_reg_init(udc);
2496 
2497 	/* here comes the stand operations for probe
2498 	 * set the qe_udc->gadget.xxx */
2499 	udc->gadget.ops = &qe_gadget_ops;
2500 
2501 	/* gadget.ep0 is a pointer */
2502 	udc->gadget.ep0 = &udc->eps[0].ep;
2503 
2504 	INIT_LIST_HEAD(&udc->gadget.ep_list);
2505 
2506 	/* modify in register gadget process */
2507 	udc->gadget.speed = USB_SPEED_UNKNOWN;
2508 
2509 	/* name: Identifies the controller hardware type. */
2510 	udc->gadget.name = driver_name;
2511 	udc->gadget.dev.parent = &ofdev->dev;
2512 
2513 	/* initialize qe_ep struct */
2514 	for (i = 0; i < USB_MAX_ENDPOINTS ; i++) {
2515 		/* because the ep type isn't decide here so
2516 		 * qe_ep_init() should be called in ep_enable() */
2517 
2518 		/* setup the qe_ep struct and link ep.ep.list
2519 		 * into gadget.ep_list */
2520 		qe_ep_config(udc, (unsigned char)i);
2521 	}
2522 
2523 	/* ep0 initialization in here */
2524 	ret = qe_ep_init(udc, 0, &qe_ep0_desc);
2525 	if (ret)
2526 		goto err2;
2527 
2528 	/* create a buf for ZLP send, need to remain zeroed */
2529 	udc->nullbuf = devm_kzalloc(&ofdev->dev, 256, GFP_KERNEL);
2530 	if (udc->nullbuf == NULL) {
2531 		ret = -ENOMEM;
2532 		goto err3;
2533 	}
2534 
2535 	/* buffer for data of get_status request */
2536 	udc->statusbuf = devm_kzalloc(&ofdev->dev, 2, GFP_KERNEL);
2537 	if (udc->statusbuf == NULL) {
2538 		ret = -ENOMEM;
2539 		goto err3;
2540 	}
2541 
2542 	udc->nullp = virt_to_phys((void *)udc->nullbuf);
2543 	if (udc->nullp == DMA_ADDR_INVALID) {
2544 		udc->nullp = dma_map_single(
2545 					udc->gadget.dev.parent,
2546 					udc->nullbuf,
2547 					256,
2548 					DMA_TO_DEVICE);
2549 		udc->nullmap = 1;
2550 	} else {
2551 		dma_sync_single_for_device(udc->gadget.dev.parent,
2552 					udc->nullp, 256,
2553 					DMA_TO_DEVICE);
2554 	}
2555 
2556 	tasklet_init(&udc->rx_tasklet, ep_rx_tasklet,
2557 			(unsigned long)udc);
2558 	/* request irq and disable DR  */
2559 	udc->usb_irq = irq_of_parse_and_map(np, 0);
2560 	if (!udc->usb_irq) {
2561 		ret = -EINVAL;
2562 		goto err_noirq;
2563 	}
2564 
2565 	ret = request_irq(udc->usb_irq, qe_udc_irq, 0,
2566 				driver_name, udc);
2567 	if (ret) {
2568 		dev_err(udc->dev, "cannot request irq %d err %d\n",
2569 				udc->usb_irq, ret);
2570 		goto err4;
2571 	}
2572 
2573 	ret = usb_add_gadget_udc_release(&ofdev->dev, &udc->gadget,
2574 			qe_udc_release);
2575 	if (ret)
2576 		goto err5;
2577 
2578 	platform_set_drvdata(ofdev, udc);
2579 	dev_info(udc->dev,
2580 			"%s USB controller initialized as device\n",
2581 			(udc->soc_type == PORT_QE) ? "QE" : "CPM");
2582 	return 0;
2583 
2584 err5:
2585 	free_irq(udc->usb_irq, udc);
2586 err4:
2587 	irq_dispose_mapping(udc->usb_irq);
2588 err_noirq:
2589 	if (udc->nullmap) {
2590 		dma_unmap_single(udc->gadget.dev.parent,
2591 			udc->nullp, 256,
2592 				DMA_TO_DEVICE);
2593 			udc->nullp = DMA_ADDR_INVALID;
2594 	} else {
2595 		dma_sync_single_for_cpu(udc->gadget.dev.parent,
2596 			udc->nullp, 256,
2597 				DMA_TO_DEVICE);
2598 	}
2599 err3:
2600 	ep = &udc->eps[0];
2601 	cpm_muram_free(cpm_muram_offset(ep->rxbase));
2602 	kfree(ep->rxframe);
2603 	kfree(ep->rxbuffer);
2604 	kfree(ep->txframe);
2605 err2:
2606 	iounmap(udc->usb_regs);
2607 err1:
2608 	kfree(udc);
2609 	return ret;
2610 }
2611 
2612 #ifdef CONFIG_PM
qe_udc_suspend(struct platform_device * dev,pm_message_t state)2613 static int qe_udc_suspend(struct platform_device *dev, pm_message_t state)
2614 {
2615 	return -ENOTSUPP;
2616 }
2617 
qe_udc_resume(struct platform_device * dev)2618 static int qe_udc_resume(struct platform_device *dev)
2619 {
2620 	return -ENOTSUPP;
2621 }
2622 #endif
2623 
qe_udc_remove(struct platform_device * ofdev)2624 static int qe_udc_remove(struct platform_device *ofdev)
2625 {
2626 	struct qe_udc *udc = platform_get_drvdata(ofdev);
2627 	struct qe_ep *ep;
2628 	unsigned int size;
2629 	DECLARE_COMPLETION_ONSTACK(done);
2630 
2631 	usb_del_gadget_udc(&udc->gadget);
2632 
2633 	udc->done = &done;
2634 	tasklet_disable(&udc->rx_tasklet);
2635 
2636 	if (udc->nullmap) {
2637 		dma_unmap_single(udc->gadget.dev.parent,
2638 			udc->nullp, 256,
2639 				DMA_TO_DEVICE);
2640 			udc->nullp = DMA_ADDR_INVALID;
2641 	} else {
2642 		dma_sync_single_for_cpu(udc->gadget.dev.parent,
2643 			udc->nullp, 256,
2644 				DMA_TO_DEVICE);
2645 	}
2646 
2647 	ep = &udc->eps[0];
2648 	cpm_muram_free(cpm_muram_offset(ep->rxbase));
2649 	size = (ep->ep.maxpacket + USB_CRC_SIZE + 2) * (USB_BDRING_LEN + 1);
2650 
2651 	kfree(ep->rxframe);
2652 	if (ep->rxbufmap) {
2653 		dma_unmap_single(udc->gadget.dev.parent,
2654 				ep->rxbuf_d, size,
2655 				DMA_FROM_DEVICE);
2656 		ep->rxbuf_d = DMA_ADDR_INVALID;
2657 	} else {
2658 		dma_sync_single_for_cpu(udc->gadget.dev.parent,
2659 				ep->rxbuf_d, size,
2660 				DMA_FROM_DEVICE);
2661 	}
2662 
2663 	kfree(ep->rxbuffer);
2664 	kfree(ep->txframe);
2665 
2666 	free_irq(udc->usb_irq, udc);
2667 	irq_dispose_mapping(udc->usb_irq);
2668 
2669 	tasklet_kill(&udc->rx_tasklet);
2670 
2671 	iounmap(udc->usb_regs);
2672 
2673 	/* wait for release() of gadget.dev to free udc */
2674 	wait_for_completion(&done);
2675 
2676 	return 0;
2677 }
2678 
2679 /*-------------------------------------------------------------------------*/
2680 static const struct of_device_id qe_udc_match[] = {
2681 	{
2682 		.compatible = "fsl,mpc8323-qe-usb",
2683 		.data = (void *)PORT_QE,
2684 	},
2685 	{
2686 		.compatible = "fsl,mpc8360-qe-usb",
2687 		.data = (void *)PORT_QE,
2688 	},
2689 	{
2690 		.compatible = "fsl,mpc8272-cpm-usb",
2691 		.data = (void *)PORT_CPM,
2692 	},
2693 	{},
2694 };
2695 
2696 MODULE_DEVICE_TABLE(of, qe_udc_match);
2697 
2698 static struct platform_driver udc_driver = {
2699 	.driver = {
2700 		.name = driver_name,
2701 		.of_match_table = qe_udc_match,
2702 	},
2703 	.probe          = qe_udc_probe,
2704 	.remove         = qe_udc_remove,
2705 #ifdef CONFIG_PM
2706 	.suspend        = qe_udc_suspend,
2707 	.resume         = qe_udc_resume,
2708 #endif
2709 };
2710 
2711 module_platform_driver(udc_driver);
2712 
2713 MODULE_DESCRIPTION(DRIVER_DESC);
2714 MODULE_AUTHOR(DRIVER_AUTHOR);
2715 MODULE_LICENSE("GPL");
2716