1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Cadence USBSS DRD Driver - gadget side. 4 * 5 * Copyright (C) 2018-2019 Cadence Design Systems. 6 * Copyright (C) 2017-2018 NXP 7 * 8 * Authors: Pawel Jez <pjez@cadence.com>, 9 * Pawel Laszczak <pawell@cadence.com> 10 * Peter Chen <peter.chen@nxp.com> 11 */ 12 13 /* 14 * Work around 1: 15 * At some situations, the controller may get stale data address in TRB 16 * at below sequences: 17 * 1. Controller read TRB includes data address 18 * 2. Software updates TRBs includes data address and Cycle bit 19 * 3. Controller read TRB which includes Cycle bit 20 * 4. DMA run with stale data address 21 * 22 * To fix this problem, driver needs to make the first TRB in TD as invalid. 23 * After preparing all TRBs driver needs to check the position of DMA and 24 * if the DMA point to the first just added TRB and doorbell is 1, 25 * then driver must defer making this TRB as valid. This TRB will be make 26 * as valid during adding next TRB only if DMA is stopped or at TRBERR 27 * interrupt. 28 * 29 * Issue has been fixed in DEV_VER_V3 version of controller. 30 * 31 * Work around 2: 32 * Controller for OUT endpoints has shared on-chip buffers for all incoming 33 * packets, including ep0out. It's FIFO buffer, so packets must be handle by DMA 34 * in correct order. If the first packet in the buffer will not be handled, 35 * then the following packets directed for other endpoints and functions 36 * will be blocked. 37 * Additionally the packets directed to one endpoint can block entire on-chip 38 * buffers. In this case transfer to other endpoints also will blocked. 39 * 40 * To resolve this issue after raising the descriptor missing interrupt 41 * driver prepares internal usb_request object and use it to arm DMA transfer. 42 * 43 * The problematic situation was observed in case when endpoint has been enabled 44 * but no usb_request were queued. Driver try detects such endpoints and will 45 * use this workaround only for these endpoint. 46 * 47 * Driver use limited number of buffer. This number can be set by macro 48 * CDNS3_WA2_NUM_BUFFERS. 49 * 50 * Such blocking situation was observed on ACM gadget. For this function 51 * host send OUT data packet but ACM function is not prepared for this packet. 52 * It's cause that buffer placed in on chip memory block transfer to other 53 * endpoints. 54 * 55 * Issue has been fixed in DEV_VER_V2 version of controller. 56 * 57 */ 58 59 #include <linux/dma-mapping.h> 60 #include <linux/usb/gadget.h> 61 #include <linux/module.h> 62 #include <linux/dmapool.h> 63 #include <linux/iopoll.h> 64 #include <linux/property.h> 65 66 #include "core.h" 67 #include "gadget-export.h" 68 #include "cdns3-gadget.h" 69 #include "cdns3-trace.h" 70 #include "drd.h" 71 72 static int __cdns3_gadget_ep_queue(struct usb_ep *ep, 73 struct usb_request *request, 74 gfp_t gfp_flags); 75 76 static int cdns3_ep_run_transfer(struct cdns3_endpoint *priv_ep, 77 struct usb_request *request); 78 79 static int cdns3_ep_run_stream_transfer(struct cdns3_endpoint *priv_ep, 80 struct usb_request *request); 81 82 /** 83 * cdns3_clear_register_bit - clear bit in given register. 84 * @ptr: address of device controller register to be read and changed 85 * @mask: bits requested to clar 86 */ 87 static void cdns3_clear_register_bit(void __iomem *ptr, u32 mask) 88 { 89 mask = readl(ptr) & ~mask; 90 writel(mask, ptr); 91 } 92 93 /** 94 * cdns3_set_register_bit - set bit in given register. 95 * @ptr: address of device controller register to be read and changed 96 * @mask: bits requested to set 97 */ 98 void cdns3_set_register_bit(void __iomem *ptr, u32 mask) 99 { 100 mask = readl(ptr) | mask; 101 writel(mask, ptr); 102 } 103 104 /** 105 * cdns3_ep_addr_to_index - Macro converts endpoint address to 106 * index of endpoint object in cdns3_device.eps[] container 107 * @ep_addr: endpoint address for which endpoint object is required 108 * 109 */ 110 u8 cdns3_ep_addr_to_index(u8 ep_addr) 111 { 112 return (((ep_addr & 0x7F)) + ((ep_addr & USB_DIR_IN) ? 16 : 0)); 113 } 114 115 static int cdns3_get_dma_pos(struct cdns3_device *priv_dev, 116 struct cdns3_endpoint *priv_ep) 117 { 118 int dma_index; 119 120 dma_index = readl(&priv_dev->regs->ep_traddr) - priv_ep->trb_pool_dma; 121 122 return dma_index / TRB_SIZE; 123 } 124 125 /** 126 * cdns3_next_request - returns next request from list 127 * @list: list containing requests 128 * 129 * Returns request or NULL if no requests in list 130 */ 131 struct usb_request *cdns3_next_request(struct list_head *list) 132 { 133 return list_first_entry_or_null(list, struct usb_request, list); 134 } 135 136 /** 137 * cdns3_next_align_buf - returns next buffer from list 138 * @list: list containing buffers 139 * 140 * Returns buffer or NULL if no buffers in list 141 */ 142 static struct cdns3_aligned_buf *cdns3_next_align_buf(struct list_head *list) 143 { 144 return list_first_entry_or_null(list, struct cdns3_aligned_buf, list); 145 } 146 147 /** 148 * cdns3_next_priv_request - returns next request from list 149 * @list: list containing requests 150 * 151 * Returns request or NULL if no requests in list 152 */ 153 static struct cdns3_request *cdns3_next_priv_request(struct list_head *list) 154 { 155 return list_first_entry_or_null(list, struct cdns3_request, list); 156 } 157 158 /** 159 * cdns3_select_ep - selects endpoint 160 * @priv_dev: extended gadget object 161 * @ep: endpoint address 162 */ 163 void cdns3_select_ep(struct cdns3_device *priv_dev, u32 ep) 164 { 165 if (priv_dev->selected_ep == ep) 166 return; 167 168 priv_dev->selected_ep = ep; 169 writel(ep, &priv_dev->regs->ep_sel); 170 } 171 172 /** 173 * cdns3_get_tdl - gets current tdl for selected endpoint. 174 * @priv_dev: extended gadget object 175 * 176 * Before calling this function the appropriate endpoint must 177 * be selected by means of cdns3_select_ep function. 178 */ 179 static int cdns3_get_tdl(struct cdns3_device *priv_dev) 180 { 181 if (priv_dev->dev_ver < DEV_VER_V3) 182 return EP_CMD_TDL_GET(readl(&priv_dev->regs->ep_cmd)); 183 else 184 return readl(&priv_dev->regs->ep_tdl); 185 } 186 187 dma_addr_t cdns3_trb_virt_to_dma(struct cdns3_endpoint *priv_ep, 188 struct cdns3_trb *trb) 189 { 190 u32 offset = (char *)trb - (char *)priv_ep->trb_pool; 191 192 return priv_ep->trb_pool_dma + offset; 193 } 194 195 static void cdns3_free_trb_pool(struct cdns3_endpoint *priv_ep) 196 { 197 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 198 199 if (priv_ep->trb_pool) { 200 dma_pool_free(priv_dev->eps_dma_pool, 201 priv_ep->trb_pool, priv_ep->trb_pool_dma); 202 priv_ep->trb_pool = NULL; 203 } 204 } 205 206 /** 207 * cdns3_allocate_trb_pool - Allocates TRB's pool for selected endpoint 208 * @priv_ep: endpoint object 209 * 210 * Function will return 0 on success or -ENOMEM on allocation error 211 */ 212 int cdns3_allocate_trb_pool(struct cdns3_endpoint *priv_ep) 213 { 214 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 215 int ring_size = TRB_RING_SIZE; 216 int num_trbs = ring_size / TRB_SIZE; 217 struct cdns3_trb *link_trb; 218 219 if (priv_ep->trb_pool && priv_ep->alloc_ring_size < ring_size) 220 cdns3_free_trb_pool(priv_ep); 221 222 if (!priv_ep->trb_pool) { 223 priv_ep->trb_pool = dma_pool_alloc(priv_dev->eps_dma_pool, 224 GFP_ATOMIC, 225 &priv_ep->trb_pool_dma); 226 227 if (!priv_ep->trb_pool) 228 return -ENOMEM; 229 230 priv_ep->alloc_ring_size = ring_size; 231 } 232 233 memset(priv_ep->trb_pool, 0, ring_size); 234 235 priv_ep->num_trbs = num_trbs; 236 237 if (!priv_ep->num) 238 return 0; 239 240 /* Initialize the last TRB as Link TRB */ 241 link_trb = (priv_ep->trb_pool + (priv_ep->num_trbs - 1)); 242 243 if (priv_ep->use_streams) { 244 /* 245 * For stream capable endpoints driver use single correct TRB. 246 * The last trb has zeroed cycle bit 247 */ 248 link_trb->control = 0; 249 } else { 250 link_trb->buffer = cpu_to_le32(TRB_BUFFER(priv_ep->trb_pool_dma)); 251 link_trb->control = cpu_to_le32(TRB_CYCLE | TRB_TYPE(TRB_LINK) | TRB_TOGGLE); 252 } 253 return 0; 254 } 255 256 /** 257 * cdns3_ep_stall_flush - Stalls and flushes selected endpoint 258 * @priv_ep: endpoint object 259 * 260 * Endpoint must be selected before call to this function 261 */ 262 static void cdns3_ep_stall_flush(struct cdns3_endpoint *priv_ep) 263 { 264 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 265 int val; 266 267 trace_cdns3_halt(priv_ep, 1, 1); 268 269 writel(EP_CMD_DFLUSH | EP_CMD_ERDY | EP_CMD_SSTALL, 270 &priv_dev->regs->ep_cmd); 271 272 /* wait for DFLUSH cleared */ 273 readl_poll_timeout_atomic(&priv_dev->regs->ep_cmd, val, 274 !(val & EP_CMD_DFLUSH), 1, 1000); 275 priv_ep->flags |= EP_STALLED; 276 priv_ep->flags &= ~EP_STALL_PENDING; 277 } 278 279 /** 280 * cdns3_hw_reset_eps_config - reset endpoints configuration kept by controller. 281 * @priv_dev: extended gadget object 282 */ 283 void cdns3_hw_reset_eps_config(struct cdns3_device *priv_dev) 284 { 285 int i; 286 287 writel(USB_CONF_CFGRST, &priv_dev->regs->usb_conf); 288 289 cdns3_allow_enable_l1(priv_dev, 0); 290 priv_dev->hw_configured_flag = 0; 291 priv_dev->onchip_used_size = 0; 292 priv_dev->out_mem_is_allocated = 0; 293 priv_dev->wait_for_setup = 0; 294 priv_dev->using_streams = 0; 295 296 for (i = 0; i < CDNS3_ENDPOINTS_MAX_COUNT; i++) 297 if (priv_dev->eps[i]) 298 priv_dev->eps[i]->flags &= ~EP_CONFIGURED; 299 } 300 301 /** 302 * cdns3_ep_inc_trb - increment a trb index. 303 * @index: Pointer to the TRB index to increment. 304 * @cs: Cycle state 305 * @trb_in_seg: number of TRBs in segment 306 * 307 * The index should never point to the link TRB. After incrementing, 308 * if it is point to the link TRB, wrap around to the beginning and revert 309 * cycle state bit The 310 * link TRB is always at the last TRB entry. 311 */ 312 static void cdns3_ep_inc_trb(int *index, u8 *cs, int trb_in_seg) 313 { 314 (*index)++; 315 if (*index == (trb_in_seg - 1)) { 316 *index = 0; 317 *cs ^= 1; 318 } 319 } 320 321 /** 322 * cdns3_ep_inc_enq - increment endpoint's enqueue pointer 323 * @priv_ep: The endpoint whose enqueue pointer we're incrementing 324 */ 325 static void cdns3_ep_inc_enq(struct cdns3_endpoint *priv_ep) 326 { 327 priv_ep->free_trbs--; 328 cdns3_ep_inc_trb(&priv_ep->enqueue, &priv_ep->pcs, priv_ep->num_trbs); 329 } 330 331 /** 332 * cdns3_ep_inc_deq - increment endpoint's dequeue pointer 333 * @priv_ep: The endpoint whose dequeue pointer we're incrementing 334 */ 335 static void cdns3_ep_inc_deq(struct cdns3_endpoint *priv_ep) 336 { 337 priv_ep->free_trbs++; 338 cdns3_ep_inc_trb(&priv_ep->dequeue, &priv_ep->ccs, priv_ep->num_trbs); 339 } 340 341 /** 342 * cdns3_allow_enable_l1 - enable/disable permits to transition to L1. 343 * @priv_dev: Extended gadget object 344 * @enable: Enable/disable permit to transition to L1. 345 * 346 * If bit USB_CONF_L1EN is set and device receive Extended Token packet, 347 * then controller answer with ACK handshake. 348 * If bit USB_CONF_L1DS is set and device receive Extended Token packet, 349 * then controller answer with NYET handshake. 350 */ 351 void cdns3_allow_enable_l1(struct cdns3_device *priv_dev, int enable) 352 { 353 if (enable) 354 writel(USB_CONF_L1EN, &priv_dev->regs->usb_conf); 355 else 356 writel(USB_CONF_L1DS, &priv_dev->regs->usb_conf); 357 } 358 359 enum usb_device_speed cdns3_get_speed(struct cdns3_device *priv_dev) 360 { 361 u32 reg; 362 363 reg = readl(&priv_dev->regs->usb_sts); 364 365 if (DEV_SUPERSPEED(reg)) 366 return USB_SPEED_SUPER; 367 else if (DEV_HIGHSPEED(reg)) 368 return USB_SPEED_HIGH; 369 else if (DEV_FULLSPEED(reg)) 370 return USB_SPEED_FULL; 371 else if (DEV_LOWSPEED(reg)) 372 return USB_SPEED_LOW; 373 return USB_SPEED_UNKNOWN; 374 } 375 376 /** 377 * cdns3_start_all_request - add to ring all request not started 378 * @priv_dev: Extended gadget object 379 * @priv_ep: The endpoint for whom request will be started. 380 * 381 * Returns return ENOMEM if transfer ring i not enough TRBs to start 382 * all requests. 383 */ 384 static int cdns3_start_all_request(struct cdns3_device *priv_dev, 385 struct cdns3_endpoint *priv_ep) 386 { 387 struct usb_request *request; 388 int ret = 0; 389 u8 pending_empty = list_empty(&priv_ep->pending_req_list); 390 391 /* 392 * If the last pending transfer is INTERNAL 393 * OR streams are enabled for this endpoint 394 * do NOT start new transfer till the last one is pending 395 */ 396 if (!pending_empty) { 397 struct cdns3_request *priv_req; 398 399 request = cdns3_next_request(&priv_ep->pending_req_list); 400 priv_req = to_cdns3_request(request); 401 if ((priv_req->flags & REQUEST_INTERNAL) || 402 (priv_ep->flags & EP_TDLCHK_EN) || 403 priv_ep->use_streams) { 404 dev_dbg(priv_dev->dev, "Blocking external request\n"); 405 return ret; 406 } 407 } 408 409 while (!list_empty(&priv_ep->deferred_req_list)) { 410 request = cdns3_next_request(&priv_ep->deferred_req_list); 411 412 if (!priv_ep->use_streams) { 413 ret = cdns3_ep_run_transfer(priv_ep, request); 414 } else { 415 priv_ep->stream_sg_idx = 0; 416 ret = cdns3_ep_run_stream_transfer(priv_ep, request); 417 } 418 if (ret) 419 return ret; 420 421 list_move_tail(&request->list, &priv_ep->pending_req_list); 422 if (request->stream_id != 0 || (priv_ep->flags & EP_TDLCHK_EN)) 423 break; 424 } 425 426 priv_ep->flags &= ~EP_RING_FULL; 427 return ret; 428 } 429 430 /* 431 * WA2: Set flag for all not ISOC OUT endpoints. If this flag is set 432 * driver try to detect whether endpoint need additional internal 433 * buffer for unblocking on-chip FIFO buffer. This flag will be cleared 434 * if before first DESCMISS interrupt the DMA will be armed. 435 */ 436 #define cdns3_wa2_enable_detection(priv_dev, priv_ep, reg) do { \ 437 if (!priv_ep->dir && priv_ep->type != USB_ENDPOINT_XFER_ISOC) { \ 438 priv_ep->flags |= EP_QUIRK_EXTRA_BUF_DET; \ 439 (reg) |= EP_STS_EN_DESCMISEN; \ 440 } } while (0) 441 442 static void __cdns3_descmiss_copy_data(struct usb_request *request, 443 struct usb_request *descmiss_req) 444 { 445 int length = request->actual + descmiss_req->actual; 446 struct scatterlist *s = request->sg; 447 448 if (!s) { 449 if (length <= request->length) { 450 memcpy(&((u8 *)request->buf)[request->actual], 451 descmiss_req->buf, 452 descmiss_req->actual); 453 request->actual = length; 454 } else { 455 /* It should never occures */ 456 request->status = -ENOMEM; 457 } 458 } else { 459 if (length <= sg_dma_len(s)) { 460 void *p = phys_to_virt(sg_dma_address(s)); 461 462 memcpy(&((u8 *)p)[request->actual], 463 descmiss_req->buf, 464 descmiss_req->actual); 465 request->actual = length; 466 } else { 467 request->status = -ENOMEM; 468 } 469 } 470 } 471 472 /** 473 * cdns3_wa2_descmiss_copy_data - copy data from internal requests to 474 * request queued by class driver. 475 * @priv_ep: extended endpoint object 476 * @request: request object 477 */ 478 static void cdns3_wa2_descmiss_copy_data(struct cdns3_endpoint *priv_ep, 479 struct usb_request *request) 480 { 481 struct usb_request *descmiss_req; 482 struct cdns3_request *descmiss_priv_req; 483 484 while (!list_empty(&priv_ep->wa2_descmiss_req_list)) { 485 int chunk_end; 486 487 descmiss_priv_req = 488 cdns3_next_priv_request(&priv_ep->wa2_descmiss_req_list); 489 descmiss_req = &descmiss_priv_req->request; 490 491 /* driver can't touch pending request */ 492 if (descmiss_priv_req->flags & REQUEST_PENDING) 493 break; 494 495 chunk_end = descmiss_priv_req->flags & REQUEST_INTERNAL_CH; 496 request->status = descmiss_req->status; 497 __cdns3_descmiss_copy_data(request, descmiss_req); 498 list_del_init(&descmiss_priv_req->list); 499 kfree(descmiss_req->buf); 500 cdns3_gadget_ep_free_request(&priv_ep->endpoint, descmiss_req); 501 --priv_ep->wa2_counter; 502 503 if (!chunk_end) 504 break; 505 } 506 } 507 508 static struct usb_request *cdns3_wa2_gadget_giveback(struct cdns3_device *priv_dev, 509 struct cdns3_endpoint *priv_ep, 510 struct cdns3_request *priv_req) 511 { 512 if (priv_ep->flags & EP_QUIRK_EXTRA_BUF_EN && 513 priv_req->flags & REQUEST_INTERNAL) { 514 struct usb_request *req; 515 516 req = cdns3_next_request(&priv_ep->deferred_req_list); 517 518 priv_ep->descmis_req = NULL; 519 520 if (!req) 521 return NULL; 522 523 /* unmap the gadget request before copying data */ 524 usb_gadget_unmap_request_by_dev(priv_dev->sysdev, req, 525 priv_ep->dir); 526 527 cdns3_wa2_descmiss_copy_data(priv_ep, req); 528 if (!(priv_ep->flags & EP_QUIRK_END_TRANSFER) && 529 req->length != req->actual) { 530 /* wait for next part of transfer */ 531 /* re-map the gadget request buffer*/ 532 usb_gadget_map_request_by_dev(priv_dev->sysdev, req, 533 usb_endpoint_dir_in(priv_ep->endpoint.desc)); 534 return NULL; 535 } 536 537 if (req->status == -EINPROGRESS) 538 req->status = 0; 539 540 list_del_init(&req->list); 541 cdns3_start_all_request(priv_dev, priv_ep); 542 return req; 543 } 544 545 return &priv_req->request; 546 } 547 548 static int cdns3_wa2_gadget_ep_queue(struct cdns3_device *priv_dev, 549 struct cdns3_endpoint *priv_ep, 550 struct cdns3_request *priv_req) 551 { 552 int deferred = 0; 553 554 /* 555 * If transfer was queued before DESCMISS appear than we 556 * can disable handling of DESCMISS interrupt. Driver assumes that it 557 * can disable special treatment for this endpoint. 558 */ 559 if (priv_ep->flags & EP_QUIRK_EXTRA_BUF_DET) { 560 u32 reg; 561 562 cdns3_select_ep(priv_dev, priv_ep->num | priv_ep->dir); 563 priv_ep->flags &= ~EP_QUIRK_EXTRA_BUF_DET; 564 reg = readl(&priv_dev->regs->ep_sts_en); 565 reg &= ~EP_STS_EN_DESCMISEN; 566 trace_cdns3_wa2(priv_ep, "workaround disabled\n"); 567 writel(reg, &priv_dev->regs->ep_sts_en); 568 } 569 570 if (priv_ep->flags & EP_QUIRK_EXTRA_BUF_EN) { 571 u8 pending_empty = list_empty(&priv_ep->pending_req_list); 572 u8 descmiss_empty = list_empty(&priv_ep->wa2_descmiss_req_list); 573 574 /* 575 * DESCMISS transfer has been finished, so data will be 576 * directly copied from internal allocated usb_request 577 * objects. 578 */ 579 if (pending_empty && !descmiss_empty && 580 !(priv_req->flags & REQUEST_INTERNAL)) { 581 cdns3_wa2_descmiss_copy_data(priv_ep, 582 &priv_req->request); 583 584 trace_cdns3_wa2(priv_ep, "get internal stored data"); 585 586 list_add_tail(&priv_req->request.list, 587 &priv_ep->pending_req_list); 588 cdns3_gadget_giveback(priv_ep, priv_req, 589 priv_req->request.status); 590 591 /* 592 * Intentionally driver returns positive value as 593 * correct value. It informs that transfer has 594 * been finished. 595 */ 596 return EINPROGRESS; 597 } 598 599 /* 600 * Driver will wait for completion DESCMISS transfer, 601 * before starts new, not DESCMISS transfer. 602 */ 603 if (!pending_empty && !descmiss_empty) { 604 trace_cdns3_wa2(priv_ep, "wait for pending transfer\n"); 605 deferred = 1; 606 } 607 608 if (priv_req->flags & REQUEST_INTERNAL) 609 list_add_tail(&priv_req->list, 610 &priv_ep->wa2_descmiss_req_list); 611 } 612 613 return deferred; 614 } 615 616 static void cdns3_wa2_remove_old_request(struct cdns3_endpoint *priv_ep) 617 { 618 struct cdns3_request *priv_req; 619 620 while (!list_empty(&priv_ep->wa2_descmiss_req_list)) { 621 u8 chain; 622 623 priv_req = cdns3_next_priv_request(&priv_ep->wa2_descmiss_req_list); 624 chain = !!(priv_req->flags & REQUEST_INTERNAL_CH); 625 626 trace_cdns3_wa2(priv_ep, "removes eldest request"); 627 628 kfree(priv_req->request.buf); 629 list_del_init(&priv_req->list); 630 cdns3_gadget_ep_free_request(&priv_ep->endpoint, 631 &priv_req->request); 632 --priv_ep->wa2_counter; 633 634 if (!chain) 635 break; 636 } 637 } 638 639 /** 640 * cdns3_wa2_descmissing_packet - handles descriptor missing event. 641 * @priv_ep: extended gadget object 642 * 643 * This function is used only for WA2. For more information see Work around 2 644 * description. 645 */ 646 static void cdns3_wa2_descmissing_packet(struct cdns3_endpoint *priv_ep) 647 { 648 struct cdns3_request *priv_req; 649 struct usb_request *request; 650 u8 pending_empty = list_empty(&priv_ep->pending_req_list); 651 652 /* check for pending transfer */ 653 if (!pending_empty) { 654 trace_cdns3_wa2(priv_ep, "Ignoring Descriptor missing IRQ\n"); 655 return; 656 } 657 658 if (priv_ep->flags & EP_QUIRK_EXTRA_BUF_DET) { 659 priv_ep->flags &= ~EP_QUIRK_EXTRA_BUF_DET; 660 priv_ep->flags |= EP_QUIRK_EXTRA_BUF_EN; 661 } 662 663 trace_cdns3_wa2(priv_ep, "Description Missing detected\n"); 664 665 if (priv_ep->wa2_counter >= CDNS3_WA2_NUM_BUFFERS) { 666 trace_cdns3_wa2(priv_ep, "WA2 overflow\n"); 667 cdns3_wa2_remove_old_request(priv_ep); 668 } 669 670 request = cdns3_gadget_ep_alloc_request(&priv_ep->endpoint, 671 GFP_ATOMIC); 672 if (!request) 673 goto err; 674 675 priv_req = to_cdns3_request(request); 676 priv_req->flags |= REQUEST_INTERNAL; 677 678 /* if this field is still assigned it indicate that transfer related 679 * with this request has not been finished yet. Driver in this 680 * case simply allocate next request and assign flag REQUEST_INTERNAL_CH 681 * flag to previous one. It will indicate that current request is 682 * part of the previous one. 683 */ 684 if (priv_ep->descmis_req) 685 priv_ep->descmis_req->flags |= REQUEST_INTERNAL_CH; 686 687 priv_req->request.buf = kzalloc(CDNS3_DESCMIS_BUF_SIZE, 688 GFP_ATOMIC); 689 priv_ep->wa2_counter++; 690 691 if (!priv_req->request.buf) { 692 cdns3_gadget_ep_free_request(&priv_ep->endpoint, request); 693 goto err; 694 } 695 696 priv_req->request.length = CDNS3_DESCMIS_BUF_SIZE; 697 priv_ep->descmis_req = priv_req; 698 699 __cdns3_gadget_ep_queue(&priv_ep->endpoint, 700 &priv_ep->descmis_req->request, 701 GFP_ATOMIC); 702 703 return; 704 705 err: 706 dev_err(priv_ep->cdns3_dev->dev, 707 "Failed: No sufficient memory for DESCMIS\n"); 708 } 709 710 static void cdns3_wa2_reset_tdl(struct cdns3_device *priv_dev) 711 { 712 u16 tdl = EP_CMD_TDL_GET(readl(&priv_dev->regs->ep_cmd)); 713 714 if (tdl) { 715 u16 reset_val = EP_CMD_TDL_MAX + 1 - tdl; 716 717 writel(EP_CMD_TDL_SET(reset_val) | EP_CMD_STDL, 718 &priv_dev->regs->ep_cmd); 719 } 720 } 721 722 static void cdns3_wa2_check_outq_status(struct cdns3_device *priv_dev) 723 { 724 u32 ep_sts_reg; 725 726 /* select EP0-out */ 727 cdns3_select_ep(priv_dev, 0); 728 729 ep_sts_reg = readl(&priv_dev->regs->ep_sts); 730 731 if (EP_STS_OUTQ_VAL(ep_sts_reg)) { 732 u32 outq_ep_num = EP_STS_OUTQ_NO(ep_sts_reg); 733 struct cdns3_endpoint *outq_ep = priv_dev->eps[outq_ep_num]; 734 735 if ((outq_ep->flags & EP_ENABLED) && !(outq_ep->use_streams) && 736 outq_ep->type != USB_ENDPOINT_XFER_ISOC && outq_ep_num) { 737 u8 pending_empty = list_empty(&outq_ep->pending_req_list); 738 739 if ((outq_ep->flags & EP_QUIRK_EXTRA_BUF_DET) || 740 (outq_ep->flags & EP_QUIRK_EXTRA_BUF_EN) || 741 !pending_empty) { 742 } else { 743 u32 ep_sts_en_reg; 744 u32 ep_cmd_reg; 745 746 cdns3_select_ep(priv_dev, outq_ep->num | 747 outq_ep->dir); 748 ep_sts_en_reg = readl(&priv_dev->regs->ep_sts_en); 749 ep_cmd_reg = readl(&priv_dev->regs->ep_cmd); 750 751 outq_ep->flags |= EP_TDLCHK_EN; 752 cdns3_set_register_bit(&priv_dev->regs->ep_cfg, 753 EP_CFG_TDL_CHK); 754 755 cdns3_wa2_enable_detection(priv_dev, outq_ep, 756 ep_sts_en_reg); 757 writel(ep_sts_en_reg, 758 &priv_dev->regs->ep_sts_en); 759 /* reset tdl value to zero */ 760 cdns3_wa2_reset_tdl(priv_dev); 761 /* 762 * Memory barrier - Reset tdl before ringing the 763 * doorbell. 764 */ 765 wmb(); 766 if (EP_CMD_DRDY & ep_cmd_reg) { 767 trace_cdns3_wa2(outq_ep, "Enabling WA2 skipping doorbell\n"); 768 769 } else { 770 trace_cdns3_wa2(outq_ep, "Enabling WA2 ringing doorbell\n"); 771 /* 772 * ring doorbell to generate DESCMIS irq 773 */ 774 writel(EP_CMD_DRDY, 775 &priv_dev->regs->ep_cmd); 776 } 777 } 778 } 779 } 780 } 781 782 /** 783 * cdns3_gadget_giveback - call struct usb_request's ->complete callback 784 * @priv_ep: The endpoint to whom the request belongs to 785 * @priv_req: The request we're giving back 786 * @status: completion code for the request 787 * 788 * Must be called with controller's lock held and interrupts disabled. This 789 * function will unmap @req and call its ->complete() callback to notify upper 790 * layers that it has completed. 791 */ 792 void cdns3_gadget_giveback(struct cdns3_endpoint *priv_ep, 793 struct cdns3_request *priv_req, 794 int status) 795 { 796 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 797 struct usb_request *request = &priv_req->request; 798 799 list_del_init(&request->list); 800 801 if (request->status == -EINPROGRESS) 802 request->status = status; 803 804 if (likely(!(priv_req->flags & REQUEST_UNALIGNED))) 805 usb_gadget_unmap_request_by_dev(priv_dev->sysdev, request, 806 priv_ep->dir); 807 808 if ((priv_req->flags & REQUEST_UNALIGNED) && 809 priv_ep->dir == USB_DIR_OUT && !request->status) { 810 /* Make DMA buffer CPU accessible */ 811 dma_sync_single_for_cpu(priv_dev->sysdev, 812 priv_req->aligned_buf->dma, 813 request->actual, 814 priv_req->aligned_buf->dir); 815 memcpy(request->buf, priv_req->aligned_buf->buf, 816 request->actual); 817 } 818 819 priv_req->flags &= ~(REQUEST_PENDING | REQUEST_UNALIGNED); 820 /* All TRBs have finished, clear the counter */ 821 priv_req->finished_trb = 0; 822 trace_cdns3_gadget_giveback(priv_req); 823 824 if (priv_dev->dev_ver < DEV_VER_V2) { 825 request = cdns3_wa2_gadget_giveback(priv_dev, priv_ep, 826 priv_req); 827 if (!request) 828 return; 829 } 830 831 /* 832 * zlp request is appended by driver, needn't call usb_gadget_giveback_request() to notify 833 * gadget composite driver. 834 */ 835 if (request->complete && request->buf != priv_dev->zlp_buf) { 836 spin_unlock(&priv_dev->lock); 837 usb_gadget_giveback_request(&priv_ep->endpoint, 838 request); 839 spin_lock(&priv_dev->lock); 840 } 841 842 if (request->buf == priv_dev->zlp_buf) 843 cdns3_gadget_ep_free_request(&priv_ep->endpoint, request); 844 } 845 846 static void cdns3_wa1_restore_cycle_bit(struct cdns3_endpoint *priv_ep) 847 { 848 /* Work around for stale data address in TRB*/ 849 if (priv_ep->wa1_set) { 850 trace_cdns3_wa1(priv_ep, "restore cycle bit"); 851 852 priv_ep->wa1_set = 0; 853 priv_ep->wa1_trb_index = 0xFFFF; 854 if (priv_ep->wa1_cycle_bit) { 855 priv_ep->wa1_trb->control = 856 priv_ep->wa1_trb->control | cpu_to_le32(0x1); 857 } else { 858 priv_ep->wa1_trb->control = 859 priv_ep->wa1_trb->control & cpu_to_le32(~0x1); 860 } 861 } 862 } 863 864 static void cdns3_free_aligned_request_buf(struct work_struct *work) 865 { 866 struct cdns3_device *priv_dev = container_of(work, struct cdns3_device, 867 aligned_buf_wq); 868 struct cdns3_aligned_buf *buf, *tmp; 869 unsigned long flags; 870 871 spin_lock_irqsave(&priv_dev->lock, flags); 872 873 list_for_each_entry_safe(buf, tmp, &priv_dev->aligned_buf_list, list) { 874 if (!buf->in_use) { 875 list_del(&buf->list); 876 877 /* 878 * Re-enable interrupts to free DMA capable memory. 879 * Driver can't free this memory with disabled 880 * interrupts. 881 */ 882 spin_unlock_irqrestore(&priv_dev->lock, flags); 883 dma_free_noncoherent(priv_dev->sysdev, buf->size, 884 buf->buf, buf->dma, buf->dir); 885 kfree(buf); 886 spin_lock_irqsave(&priv_dev->lock, flags); 887 } 888 } 889 890 spin_unlock_irqrestore(&priv_dev->lock, flags); 891 } 892 893 static int cdns3_prepare_aligned_request_buf(struct cdns3_request *priv_req) 894 { 895 struct cdns3_endpoint *priv_ep = priv_req->priv_ep; 896 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 897 struct cdns3_aligned_buf *buf; 898 899 /* check if buffer is aligned to 8. */ 900 if (!((uintptr_t)priv_req->request.buf & 0x7)) 901 return 0; 902 903 buf = priv_req->aligned_buf; 904 905 if (!buf || priv_req->request.length > buf->size) { 906 buf = kzalloc(sizeof(*buf), GFP_ATOMIC); 907 if (!buf) 908 return -ENOMEM; 909 910 buf->size = priv_req->request.length; 911 buf->dir = usb_endpoint_dir_in(priv_ep->endpoint.desc) ? 912 DMA_TO_DEVICE : DMA_FROM_DEVICE; 913 914 buf->buf = dma_alloc_noncoherent(priv_dev->sysdev, 915 buf->size, 916 &buf->dma, 917 buf->dir, 918 GFP_ATOMIC); 919 if (!buf->buf) { 920 kfree(buf); 921 return -ENOMEM; 922 } 923 924 if (priv_req->aligned_buf) { 925 trace_cdns3_free_aligned_request(priv_req); 926 priv_req->aligned_buf->in_use = 0; 927 queue_work(system_freezable_wq, 928 &priv_dev->aligned_buf_wq); 929 } 930 931 buf->in_use = 1; 932 priv_req->aligned_buf = buf; 933 934 list_add_tail(&buf->list, 935 &priv_dev->aligned_buf_list); 936 } 937 938 if (priv_ep->dir == USB_DIR_IN) { 939 /* Make DMA buffer CPU accessible */ 940 dma_sync_single_for_cpu(priv_dev->sysdev, 941 buf->dma, buf->size, buf->dir); 942 memcpy(buf->buf, priv_req->request.buf, 943 priv_req->request.length); 944 } 945 946 /* Transfer DMA buffer ownership back to device */ 947 dma_sync_single_for_device(priv_dev->sysdev, 948 buf->dma, buf->size, buf->dir); 949 950 priv_req->flags |= REQUEST_UNALIGNED; 951 trace_cdns3_prepare_aligned_request(priv_req); 952 953 return 0; 954 } 955 956 static int cdns3_wa1_update_guard(struct cdns3_endpoint *priv_ep, 957 struct cdns3_trb *trb) 958 { 959 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 960 961 if (!priv_ep->wa1_set) { 962 u32 doorbell; 963 964 doorbell = !!(readl(&priv_dev->regs->ep_cmd) & EP_CMD_DRDY); 965 966 if (doorbell) { 967 priv_ep->wa1_cycle_bit = priv_ep->pcs ? TRB_CYCLE : 0; 968 priv_ep->wa1_set = 1; 969 priv_ep->wa1_trb = trb; 970 priv_ep->wa1_trb_index = priv_ep->enqueue; 971 trace_cdns3_wa1(priv_ep, "set guard"); 972 return 0; 973 } 974 } 975 return 1; 976 } 977 978 static void cdns3_wa1_tray_restore_cycle_bit(struct cdns3_device *priv_dev, 979 struct cdns3_endpoint *priv_ep) 980 { 981 int dma_index; 982 u32 doorbell; 983 984 doorbell = !!(readl(&priv_dev->regs->ep_cmd) & EP_CMD_DRDY); 985 dma_index = cdns3_get_dma_pos(priv_dev, priv_ep); 986 987 if (!doorbell || dma_index != priv_ep->wa1_trb_index) 988 cdns3_wa1_restore_cycle_bit(priv_ep); 989 } 990 991 static int cdns3_ep_run_stream_transfer(struct cdns3_endpoint *priv_ep, 992 struct usb_request *request) 993 { 994 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 995 struct cdns3_request *priv_req; 996 struct cdns3_trb *trb; 997 dma_addr_t trb_dma; 998 int address; 999 u32 control; 1000 u32 length; 1001 u32 tdl; 1002 unsigned int sg_idx = priv_ep->stream_sg_idx; 1003 1004 priv_req = to_cdns3_request(request); 1005 address = priv_ep->endpoint.desc->bEndpointAddress; 1006 1007 priv_ep->flags |= EP_PENDING_REQUEST; 1008 1009 /* must allocate buffer aligned to 8 */ 1010 if (priv_req->flags & REQUEST_UNALIGNED) 1011 trb_dma = priv_req->aligned_buf->dma; 1012 else 1013 trb_dma = request->dma; 1014 1015 /* For stream capable endpoints driver use only single TD. */ 1016 trb = priv_ep->trb_pool + priv_ep->enqueue; 1017 priv_req->start_trb = priv_ep->enqueue; 1018 priv_req->end_trb = priv_req->start_trb; 1019 priv_req->trb = trb; 1020 1021 cdns3_select_ep(priv_ep->cdns3_dev, address); 1022 1023 control = TRB_TYPE(TRB_NORMAL) | TRB_CYCLE | 1024 TRB_STREAM_ID(priv_req->request.stream_id) | TRB_ISP; 1025 1026 if (!request->num_sgs) { 1027 trb->buffer = cpu_to_le32(TRB_BUFFER(trb_dma)); 1028 length = request->length; 1029 } else { 1030 trb->buffer = cpu_to_le32(TRB_BUFFER(request->sg[sg_idx].dma_address)); 1031 length = request->sg[sg_idx].length; 1032 } 1033 1034 tdl = DIV_ROUND_UP(length, priv_ep->endpoint.maxpacket); 1035 1036 trb->length = cpu_to_le32(TRB_BURST_LEN(16) | TRB_LEN(length)); 1037 1038 /* 1039 * For DEV_VER_V2 controller version we have enabled 1040 * USB_CONF2_EN_TDL_TRB in DMULT configuration. 1041 * This enables TDL calculation based on TRB, hence setting TDL in TRB. 1042 */ 1043 if (priv_dev->dev_ver >= DEV_VER_V2) { 1044 if (priv_dev->gadget.speed == USB_SPEED_SUPER) 1045 trb->length |= cpu_to_le32(TRB_TDL_SS_SIZE(tdl)); 1046 } 1047 priv_req->flags |= REQUEST_PENDING; 1048 1049 trb->control = cpu_to_le32(control); 1050 1051 trace_cdns3_prepare_trb(priv_ep, priv_req->trb); 1052 1053 /* 1054 * Memory barrier - Cycle Bit must be set before trb->length and 1055 * trb->buffer fields. 1056 */ 1057 wmb(); 1058 1059 /* always first element */ 1060 writel(EP_TRADDR_TRADDR(priv_ep->trb_pool_dma), 1061 &priv_dev->regs->ep_traddr); 1062 1063 if (!(priv_ep->flags & EP_STALLED)) { 1064 trace_cdns3_ring(priv_ep); 1065 /*clearing TRBERR and EP_STS_DESCMIS before seting DRDY*/ 1066 writel(EP_STS_TRBERR | EP_STS_DESCMIS, &priv_dev->regs->ep_sts); 1067 1068 priv_ep->prime_flag = false; 1069 1070 /* 1071 * Controller version DEV_VER_V2 tdl calculation 1072 * is based on TRB 1073 */ 1074 1075 if (priv_dev->dev_ver < DEV_VER_V2) 1076 writel(EP_CMD_TDL_SET(tdl) | EP_CMD_STDL, 1077 &priv_dev->regs->ep_cmd); 1078 else if (priv_dev->dev_ver > DEV_VER_V2) 1079 writel(tdl, &priv_dev->regs->ep_tdl); 1080 1081 priv_ep->last_stream_id = priv_req->request.stream_id; 1082 writel(EP_CMD_DRDY, &priv_dev->regs->ep_cmd); 1083 writel(EP_CMD_ERDY_SID(priv_req->request.stream_id) | 1084 EP_CMD_ERDY, &priv_dev->regs->ep_cmd); 1085 1086 trace_cdns3_doorbell_epx(priv_ep->name, 1087 readl(&priv_dev->regs->ep_traddr)); 1088 } 1089 1090 /* WORKAROUND for transition to L0 */ 1091 __cdns3_gadget_wakeup(priv_dev); 1092 1093 return 0; 1094 } 1095 1096 static void cdns3_rearm_drdy_if_needed(struct cdns3_endpoint *priv_ep) 1097 { 1098 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 1099 1100 if (priv_dev->dev_ver < DEV_VER_V3) 1101 return; 1102 1103 if (readl(&priv_dev->regs->ep_sts) & EP_STS_TRBERR) { 1104 writel(EP_STS_TRBERR, &priv_dev->regs->ep_sts); 1105 writel(EP_CMD_DRDY, &priv_dev->regs->ep_cmd); 1106 } 1107 } 1108 1109 /** 1110 * cdns3_ep_run_transfer - start transfer on no-default endpoint hardware 1111 * @priv_ep: endpoint object 1112 * @request: request object 1113 * 1114 * Returns zero on success or negative value on failure 1115 */ 1116 static int cdns3_ep_run_transfer(struct cdns3_endpoint *priv_ep, 1117 struct usb_request *request) 1118 { 1119 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 1120 struct cdns3_request *priv_req; 1121 struct cdns3_trb *trb; 1122 struct cdns3_trb *link_trb = NULL; 1123 dma_addr_t trb_dma; 1124 u32 togle_pcs = 1; 1125 int sg_iter = 0; 1126 int num_trb_req; 1127 int trb_burst; 1128 int num_trb; 1129 int address; 1130 u32 control; 1131 int pcs; 1132 u16 total_tdl = 0; 1133 struct scatterlist *s = NULL; 1134 bool sg_supported = !!(request->num_mapped_sgs); 1135 u32 ioc = request->no_interrupt ? 0 : TRB_IOC; 1136 1137 num_trb_req = sg_supported ? request->num_mapped_sgs : 1; 1138 1139 /* ISO transfer require each SOF have a TD, each TD include some TRBs */ 1140 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC) 1141 num_trb = priv_ep->interval * num_trb_req; 1142 else 1143 num_trb = num_trb_req; 1144 1145 priv_req = to_cdns3_request(request); 1146 address = priv_ep->endpoint.desc->bEndpointAddress; 1147 1148 priv_ep->flags |= EP_PENDING_REQUEST; 1149 1150 /* must allocate buffer aligned to 8 */ 1151 if (priv_req->flags & REQUEST_UNALIGNED) 1152 trb_dma = priv_req->aligned_buf->dma; 1153 else 1154 trb_dma = request->dma; 1155 1156 trb = priv_ep->trb_pool + priv_ep->enqueue; 1157 priv_req->start_trb = priv_ep->enqueue; 1158 priv_req->trb = trb; 1159 1160 cdns3_select_ep(priv_ep->cdns3_dev, address); 1161 1162 /* prepare ring */ 1163 if ((priv_ep->enqueue + num_trb) >= (priv_ep->num_trbs - 1)) { 1164 int doorbell, dma_index; 1165 u32 ch_bit = 0; 1166 1167 doorbell = !!(readl(&priv_dev->regs->ep_cmd) & EP_CMD_DRDY); 1168 dma_index = cdns3_get_dma_pos(priv_dev, priv_ep); 1169 1170 /* Driver can't update LINK TRB if it is current processed. */ 1171 if (doorbell && dma_index == priv_ep->num_trbs - 1) { 1172 priv_ep->flags |= EP_DEFERRED_DRDY; 1173 return -ENOBUFS; 1174 } 1175 1176 /*updating C bt in Link TRB before starting DMA*/ 1177 link_trb = priv_ep->trb_pool + (priv_ep->num_trbs - 1); 1178 /* 1179 * For TRs size equal 2 enabling TRB_CHAIN for epXin causes 1180 * that DMA stuck at the LINK TRB. 1181 * On the other hand, removing TRB_CHAIN for longer TRs for 1182 * epXout cause that DMA stuck after handling LINK TRB. 1183 * To eliminate this strange behavioral driver set TRB_CHAIN 1184 * bit only for TR size > 2. 1185 */ 1186 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC || 1187 TRBS_PER_SEGMENT > 2) 1188 ch_bit = TRB_CHAIN; 1189 1190 link_trb->control = cpu_to_le32(((priv_ep->pcs) ? TRB_CYCLE : 0) | 1191 TRB_TYPE(TRB_LINK) | TRB_TOGGLE | ch_bit); 1192 1193 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC) { 1194 /* 1195 * ISO require LINK TRB must be first one of TD. 1196 * Fill LINK TRBs for left trb space to simply software process logic. 1197 */ 1198 while (priv_ep->enqueue) { 1199 *trb = *link_trb; 1200 trace_cdns3_prepare_trb(priv_ep, trb); 1201 1202 cdns3_ep_inc_enq(priv_ep); 1203 trb = priv_ep->trb_pool + priv_ep->enqueue; 1204 priv_req->trb = trb; 1205 } 1206 } 1207 } 1208 1209 if (num_trb > priv_ep->free_trbs) { 1210 priv_ep->flags |= EP_RING_FULL; 1211 return -ENOBUFS; 1212 } 1213 1214 if (priv_dev->dev_ver <= DEV_VER_V2) 1215 togle_pcs = cdns3_wa1_update_guard(priv_ep, trb); 1216 1217 /* set incorrect Cycle Bit for first trb*/ 1218 control = priv_ep->pcs ? 0 : TRB_CYCLE; 1219 trb->length = 0; 1220 if (priv_dev->dev_ver >= DEV_VER_V2) { 1221 u16 td_size; 1222 1223 td_size = DIV_ROUND_UP(request->length, 1224 priv_ep->endpoint.maxpacket); 1225 if (priv_dev->gadget.speed == USB_SPEED_SUPER) 1226 trb->length = cpu_to_le32(TRB_TDL_SS_SIZE(td_size)); 1227 else 1228 control |= TRB_TDL_HS_SIZE(td_size); 1229 } 1230 1231 do { 1232 u32 length; 1233 1234 if (!(sg_iter % num_trb_req) && sg_supported) 1235 s = request->sg; 1236 1237 /* fill TRB */ 1238 control |= TRB_TYPE(TRB_NORMAL); 1239 if (sg_supported) { 1240 trb->buffer = cpu_to_le32(TRB_BUFFER(sg_dma_address(s))); 1241 length = sg_dma_len(s); 1242 } else { 1243 trb->buffer = cpu_to_le32(TRB_BUFFER(trb_dma)); 1244 length = request->length; 1245 } 1246 1247 if (priv_ep->flags & EP_TDLCHK_EN) 1248 total_tdl += DIV_ROUND_UP(length, 1249 priv_ep->endpoint.maxpacket); 1250 1251 trb_burst = priv_ep->trb_burst_size; 1252 1253 /* 1254 * Supposed DMA cross 4k bounder problem should be fixed at DEV_VER_V2, but still 1255 * met problem when do ISO transfer if sg enabled. 1256 * 1257 * Data pattern likes below when sg enabled, package size is 1k and mult is 2 1258 * [UVC Header(8B) ] [data(3k - 8)] ... 1259 * 1260 * The received data at offset 0xd000 will get 0xc000 data, len 0x70. Error happen 1261 * as below pattern: 1262 * 0xd000: wrong 1263 * 0xe000: wrong 1264 * 0xf000: correct 1265 * 0x10000: wrong 1266 * 0x11000: wrong 1267 * 0x12000: correct 1268 * ... 1269 * 1270 * But it is still unclear about why error have not happen below 0xd000, it should 1271 * cross 4k bounder. But anyway, the below code can fix this problem. 1272 * 1273 * To avoid DMA cross 4k bounder at ISO transfer, reduce burst len according to 16. 1274 */ 1275 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && priv_dev->dev_ver <= DEV_VER_V2) 1276 if (ALIGN_DOWN(trb->buffer, SZ_4K) != 1277 ALIGN_DOWN(trb->buffer + length, SZ_4K)) 1278 trb_burst = 16; 1279 1280 trb->length |= cpu_to_le32(TRB_BURST_LEN(trb_burst) | 1281 TRB_LEN(length)); 1282 pcs = priv_ep->pcs ? TRB_CYCLE : 0; 1283 1284 /* 1285 * first trb should be prepared as last to avoid processing 1286 * transfer to early 1287 */ 1288 if (sg_iter != 0) 1289 control |= pcs; 1290 1291 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && !priv_ep->dir) { 1292 control |= ioc | TRB_ISP; 1293 } else { 1294 /* for last element in TD or in SG list */ 1295 if (sg_iter == (num_trb - 1) && sg_iter != 0) 1296 control |= pcs | ioc | TRB_ISP; 1297 } 1298 1299 if (sg_iter) 1300 trb->control = cpu_to_le32(control); 1301 else 1302 priv_req->trb->control = cpu_to_le32(control); 1303 1304 if (sg_supported) { 1305 trb->control |= cpu_to_le32(TRB_ISP); 1306 /* Don't set chain bit for last TRB */ 1307 if ((sg_iter % num_trb_req) < num_trb_req - 1) 1308 trb->control |= cpu_to_le32(TRB_CHAIN); 1309 1310 s = sg_next(s); 1311 } 1312 1313 control = 0; 1314 ++sg_iter; 1315 priv_req->end_trb = priv_ep->enqueue; 1316 cdns3_ep_inc_enq(priv_ep); 1317 trb = priv_ep->trb_pool + priv_ep->enqueue; 1318 trb->length = 0; 1319 } while (sg_iter < num_trb); 1320 1321 trb = priv_req->trb; 1322 1323 priv_req->flags |= REQUEST_PENDING; 1324 priv_req->num_of_trb = num_trb; 1325 1326 if (sg_iter == 1) 1327 trb->control |= cpu_to_le32(ioc | TRB_ISP); 1328 1329 if (priv_dev->dev_ver < DEV_VER_V2 && 1330 (priv_ep->flags & EP_TDLCHK_EN)) { 1331 u16 tdl = total_tdl; 1332 u16 old_tdl = EP_CMD_TDL_GET(readl(&priv_dev->regs->ep_cmd)); 1333 1334 if (tdl > EP_CMD_TDL_MAX) { 1335 tdl = EP_CMD_TDL_MAX; 1336 priv_ep->pending_tdl = total_tdl - EP_CMD_TDL_MAX; 1337 } 1338 1339 if (old_tdl < tdl) { 1340 tdl -= old_tdl; 1341 writel(EP_CMD_TDL_SET(tdl) | EP_CMD_STDL, 1342 &priv_dev->regs->ep_cmd); 1343 } 1344 } 1345 1346 /* 1347 * Memory barrier - cycle bit must be set before other filds in trb. 1348 */ 1349 wmb(); 1350 1351 /* give the TD to the consumer*/ 1352 if (togle_pcs) 1353 trb->control = trb->control ^ cpu_to_le32(1); 1354 1355 if (priv_dev->dev_ver <= DEV_VER_V2) 1356 cdns3_wa1_tray_restore_cycle_bit(priv_dev, priv_ep); 1357 1358 if (num_trb > 1) { 1359 int i = 0; 1360 1361 while (i < num_trb) { 1362 trace_cdns3_prepare_trb(priv_ep, trb + i); 1363 if (trb + i == link_trb) { 1364 trb = priv_ep->trb_pool; 1365 num_trb = num_trb - i; 1366 i = 0; 1367 } else { 1368 i++; 1369 } 1370 } 1371 } else { 1372 trace_cdns3_prepare_trb(priv_ep, priv_req->trb); 1373 } 1374 1375 /* 1376 * Memory barrier - Cycle Bit must be set before trb->length and 1377 * trb->buffer fields. 1378 */ 1379 wmb(); 1380 1381 /* 1382 * For DMULT mode we can set address to transfer ring only once after 1383 * enabling endpoint. 1384 */ 1385 if (priv_ep->flags & EP_UPDATE_EP_TRBADDR) { 1386 /* 1387 * Until SW is not ready to handle the OUT transfer the ISO OUT 1388 * Endpoint should be disabled (EP_CFG.ENABLE = 0). 1389 * EP_CFG_ENABLE must be set before updating ep_traddr. 1390 */ 1391 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && !priv_ep->dir && 1392 !(priv_ep->flags & EP_QUIRK_ISO_OUT_EN)) { 1393 priv_ep->flags |= EP_QUIRK_ISO_OUT_EN; 1394 cdns3_set_register_bit(&priv_dev->regs->ep_cfg, 1395 EP_CFG_ENABLE); 1396 } 1397 1398 writel(EP_TRADDR_TRADDR(priv_ep->trb_pool_dma + 1399 priv_req->start_trb * TRB_SIZE), 1400 &priv_dev->regs->ep_traddr); 1401 1402 priv_ep->flags &= ~EP_UPDATE_EP_TRBADDR; 1403 } 1404 1405 if (!priv_ep->wa1_set && !(priv_ep->flags & EP_STALLED)) { 1406 trace_cdns3_ring(priv_ep); 1407 /*clearing TRBERR and EP_STS_DESCMIS before seting DRDY*/ 1408 writel(EP_STS_TRBERR | EP_STS_DESCMIS, &priv_dev->regs->ep_sts); 1409 writel(EP_CMD_DRDY, &priv_dev->regs->ep_cmd); 1410 cdns3_rearm_drdy_if_needed(priv_ep); 1411 trace_cdns3_doorbell_epx(priv_ep->name, 1412 readl(&priv_dev->regs->ep_traddr)); 1413 } 1414 1415 /* WORKAROUND for transition to L0 */ 1416 __cdns3_gadget_wakeup(priv_dev); 1417 1418 return 0; 1419 } 1420 1421 void cdns3_set_hw_configuration(struct cdns3_device *priv_dev) 1422 { 1423 struct cdns3_endpoint *priv_ep; 1424 struct usb_ep *ep; 1425 1426 if (priv_dev->hw_configured_flag) 1427 return; 1428 1429 writel(USB_CONF_CFGSET, &priv_dev->regs->usb_conf); 1430 1431 cdns3_set_register_bit(&priv_dev->regs->usb_conf, 1432 USB_CONF_U1EN | USB_CONF_U2EN); 1433 1434 priv_dev->hw_configured_flag = 1; 1435 1436 list_for_each_entry(ep, &priv_dev->gadget.ep_list, ep_list) { 1437 if (ep->enabled) { 1438 priv_ep = ep_to_cdns3_ep(ep); 1439 cdns3_start_all_request(priv_dev, priv_ep); 1440 } 1441 } 1442 1443 cdns3_allow_enable_l1(priv_dev, 1); 1444 } 1445 1446 /** 1447 * cdns3_trb_handled - check whether trb has been handled by DMA 1448 * 1449 * @priv_ep: extended endpoint object. 1450 * @priv_req: request object for checking 1451 * 1452 * Endpoint must be selected before invoking this function. 1453 * 1454 * Returns false if request has not been handled by DMA, else returns true. 1455 * 1456 * SR - start ring 1457 * ER - end ring 1458 * DQ = priv_ep->dequeue - dequeue position 1459 * EQ = priv_ep->enqueue - enqueue position 1460 * ST = priv_req->start_trb - index of first TRB in transfer ring 1461 * ET = priv_req->end_trb - index of last TRB in transfer ring 1462 * CI = current_index - index of processed TRB by DMA. 1463 * 1464 * As first step, we check if the TRB between the ST and ET. 1465 * Then, we check if cycle bit for index priv_ep->dequeue 1466 * is correct. 1467 * 1468 * some rules: 1469 * 1. priv_ep->dequeue never equals to current_index. 1470 * 2 priv_ep->enqueue never exceed priv_ep->dequeue 1471 * 3. exception: priv_ep->enqueue == priv_ep->dequeue 1472 * and priv_ep->free_trbs is zero. 1473 * This case indicate that TR is full. 1474 * 1475 * At below two cases, the request have been handled. 1476 * Case 1 - priv_ep->dequeue < current_index 1477 * SR ... EQ ... DQ ... CI ... ER 1478 * SR ... DQ ... CI ... EQ ... ER 1479 * 1480 * Case 2 - priv_ep->dequeue > current_index 1481 * This situation takes place when CI go through the LINK TRB at the end of 1482 * transfer ring. 1483 * SR ... CI ... EQ ... DQ ... ER 1484 */ 1485 static bool cdns3_trb_handled(struct cdns3_endpoint *priv_ep, 1486 struct cdns3_request *priv_req) 1487 { 1488 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 1489 struct cdns3_trb *trb; 1490 int current_index = 0; 1491 int handled = 0; 1492 int doorbell; 1493 1494 current_index = cdns3_get_dma_pos(priv_dev, priv_ep); 1495 doorbell = !!(readl(&priv_dev->regs->ep_cmd) & EP_CMD_DRDY); 1496 1497 /* current trb doesn't belong to this request */ 1498 if (priv_req->start_trb < priv_req->end_trb) { 1499 if (priv_ep->dequeue > priv_req->end_trb) 1500 goto finish; 1501 1502 if (priv_ep->dequeue < priv_req->start_trb) 1503 goto finish; 1504 } 1505 1506 if ((priv_req->start_trb > priv_req->end_trb) && 1507 (priv_ep->dequeue > priv_req->end_trb) && 1508 (priv_ep->dequeue < priv_req->start_trb)) 1509 goto finish; 1510 1511 if ((priv_req->start_trb == priv_req->end_trb) && 1512 (priv_ep->dequeue != priv_req->end_trb)) 1513 goto finish; 1514 1515 trb = &priv_ep->trb_pool[priv_ep->dequeue]; 1516 1517 if ((le32_to_cpu(trb->control) & TRB_CYCLE) != priv_ep->ccs) 1518 goto finish; 1519 1520 if (doorbell == 1 && current_index == priv_ep->dequeue) 1521 goto finish; 1522 1523 /* The corner case for TRBS_PER_SEGMENT equal 2). */ 1524 if (TRBS_PER_SEGMENT == 2 && priv_ep->type != USB_ENDPOINT_XFER_ISOC) { 1525 handled = 1; 1526 goto finish; 1527 } 1528 1529 if (priv_ep->enqueue == priv_ep->dequeue && 1530 priv_ep->free_trbs == 0) { 1531 handled = 1; 1532 } else if (priv_ep->dequeue < current_index) { 1533 if ((current_index == (priv_ep->num_trbs - 1)) && 1534 !priv_ep->dequeue) 1535 goto finish; 1536 1537 handled = 1; 1538 } else if (priv_ep->dequeue > current_index) { 1539 handled = 1; 1540 } 1541 1542 finish: 1543 trace_cdns3_request_handled(priv_req, current_index, handled); 1544 1545 return handled; 1546 } 1547 1548 static void cdns3_transfer_completed(struct cdns3_device *priv_dev, 1549 struct cdns3_endpoint *priv_ep) 1550 { 1551 struct cdns3_request *priv_req; 1552 struct usb_request *request; 1553 struct cdns3_trb *trb; 1554 bool request_handled = false; 1555 bool transfer_end = false; 1556 1557 while (!list_empty(&priv_ep->pending_req_list)) { 1558 request = cdns3_next_request(&priv_ep->pending_req_list); 1559 priv_req = to_cdns3_request(request); 1560 1561 trb = priv_ep->trb_pool + priv_ep->dequeue; 1562 1563 /* The TRB was changed as link TRB, and the request was handled at ep_dequeue */ 1564 while (TRB_FIELD_TO_TYPE(le32_to_cpu(trb->control)) == TRB_LINK) { 1565 1566 /* ISO ep_traddr may stop at LINK TRB */ 1567 if (priv_ep->dequeue == cdns3_get_dma_pos(priv_dev, priv_ep) && 1568 priv_ep->type == USB_ENDPOINT_XFER_ISOC) 1569 break; 1570 1571 trace_cdns3_complete_trb(priv_ep, trb); 1572 cdns3_ep_inc_deq(priv_ep); 1573 trb = priv_ep->trb_pool + priv_ep->dequeue; 1574 } 1575 1576 if (!request->stream_id) { 1577 /* Re-select endpoint. It could be changed by other CPU 1578 * during handling usb_gadget_giveback_request. 1579 */ 1580 cdns3_select_ep(priv_dev, priv_ep->endpoint.address); 1581 1582 while (cdns3_trb_handled(priv_ep, priv_req)) { 1583 priv_req->finished_trb++; 1584 if (priv_req->finished_trb >= priv_req->num_of_trb) 1585 request_handled = true; 1586 1587 trb = priv_ep->trb_pool + priv_ep->dequeue; 1588 trace_cdns3_complete_trb(priv_ep, trb); 1589 1590 if (!transfer_end) 1591 request->actual += 1592 TRB_LEN(le32_to_cpu(trb->length)); 1593 1594 if (priv_req->num_of_trb > 1 && 1595 le32_to_cpu(trb->control) & TRB_SMM && 1596 le32_to_cpu(trb->control) & TRB_CHAIN) 1597 transfer_end = true; 1598 1599 cdns3_ep_inc_deq(priv_ep); 1600 } 1601 1602 if (request_handled) { 1603 /* TRBs are duplicated by priv_ep->interval time for ISO IN */ 1604 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && priv_ep->dir) 1605 request->actual /= priv_ep->interval; 1606 1607 cdns3_gadget_giveback(priv_ep, priv_req, 0); 1608 request_handled = false; 1609 transfer_end = false; 1610 } else { 1611 goto prepare_next_td; 1612 } 1613 1614 if (priv_ep->type != USB_ENDPOINT_XFER_ISOC && 1615 TRBS_PER_SEGMENT == 2) 1616 break; 1617 } else { 1618 /* Re-select endpoint. It could be changed by other CPU 1619 * during handling usb_gadget_giveback_request. 1620 */ 1621 cdns3_select_ep(priv_dev, priv_ep->endpoint.address); 1622 1623 trb = priv_ep->trb_pool; 1624 trace_cdns3_complete_trb(priv_ep, trb); 1625 1626 if (trb != priv_req->trb) 1627 dev_warn(priv_dev->dev, 1628 "request_trb=0x%p, queue_trb=0x%p\n", 1629 priv_req->trb, trb); 1630 1631 request->actual += TRB_LEN(le32_to_cpu(trb->length)); 1632 1633 if (!request->num_sgs || 1634 (request->num_sgs == (priv_ep->stream_sg_idx + 1))) { 1635 priv_ep->stream_sg_idx = 0; 1636 cdns3_gadget_giveback(priv_ep, priv_req, 0); 1637 } else { 1638 priv_ep->stream_sg_idx++; 1639 cdns3_ep_run_stream_transfer(priv_ep, request); 1640 } 1641 break; 1642 } 1643 } 1644 priv_ep->flags &= ~EP_PENDING_REQUEST; 1645 1646 prepare_next_td: 1647 if (!(priv_ep->flags & EP_STALLED) && 1648 !(priv_ep->flags & EP_STALL_PENDING)) 1649 cdns3_start_all_request(priv_dev, priv_ep); 1650 } 1651 1652 void cdns3_rearm_transfer(struct cdns3_endpoint *priv_ep, u8 rearm) 1653 { 1654 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 1655 1656 cdns3_wa1_restore_cycle_bit(priv_ep); 1657 1658 if (rearm) { 1659 trace_cdns3_ring(priv_ep); 1660 1661 /* Cycle Bit must be updated before arming DMA. */ 1662 wmb(); 1663 writel(EP_CMD_DRDY, &priv_dev->regs->ep_cmd); 1664 1665 __cdns3_gadget_wakeup(priv_dev); 1666 1667 trace_cdns3_doorbell_epx(priv_ep->name, 1668 readl(&priv_dev->regs->ep_traddr)); 1669 } 1670 } 1671 1672 static void cdns3_reprogram_tdl(struct cdns3_endpoint *priv_ep) 1673 { 1674 u16 tdl = priv_ep->pending_tdl; 1675 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 1676 1677 if (tdl > EP_CMD_TDL_MAX) { 1678 tdl = EP_CMD_TDL_MAX; 1679 priv_ep->pending_tdl -= EP_CMD_TDL_MAX; 1680 } else { 1681 priv_ep->pending_tdl = 0; 1682 } 1683 1684 writel(EP_CMD_TDL_SET(tdl) | EP_CMD_STDL, &priv_dev->regs->ep_cmd); 1685 } 1686 1687 /** 1688 * cdns3_check_ep_interrupt_proceed - Processes interrupt related to endpoint 1689 * @priv_ep: endpoint object 1690 * 1691 * Returns 0 1692 */ 1693 static int cdns3_check_ep_interrupt_proceed(struct cdns3_endpoint *priv_ep) 1694 { 1695 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 1696 u32 ep_sts_reg; 1697 struct usb_request *deferred_request; 1698 struct usb_request *pending_request; 1699 u32 tdl = 0; 1700 1701 cdns3_select_ep(priv_dev, priv_ep->endpoint.address); 1702 1703 trace_cdns3_epx_irq(priv_dev, priv_ep); 1704 1705 ep_sts_reg = readl(&priv_dev->regs->ep_sts); 1706 writel(ep_sts_reg, &priv_dev->regs->ep_sts); 1707 1708 if ((ep_sts_reg & EP_STS_PRIME) && priv_ep->use_streams) { 1709 bool dbusy = !!(ep_sts_reg & EP_STS_DBUSY); 1710 1711 tdl = cdns3_get_tdl(priv_dev); 1712 1713 /* 1714 * Continue the previous transfer: 1715 * There is some racing between ERDY and PRIME. The device send 1716 * ERDY and almost in the same time Host send PRIME. It cause 1717 * that host ignore the ERDY packet and driver has to send it 1718 * again. 1719 */ 1720 if (tdl && (dbusy || !EP_STS_BUFFEMPTY(ep_sts_reg) || 1721 EP_STS_HOSTPP(ep_sts_reg))) { 1722 writel(EP_CMD_ERDY | 1723 EP_CMD_ERDY_SID(priv_ep->last_stream_id), 1724 &priv_dev->regs->ep_cmd); 1725 ep_sts_reg &= ~(EP_STS_MD_EXIT | EP_STS_IOC); 1726 } else { 1727 priv_ep->prime_flag = true; 1728 1729 pending_request = cdns3_next_request(&priv_ep->pending_req_list); 1730 deferred_request = cdns3_next_request(&priv_ep->deferred_req_list); 1731 1732 if (deferred_request && !pending_request) { 1733 cdns3_start_all_request(priv_dev, priv_ep); 1734 } 1735 } 1736 } 1737 1738 if (ep_sts_reg & EP_STS_TRBERR) { 1739 if (priv_ep->flags & EP_STALL_PENDING && 1740 !(ep_sts_reg & EP_STS_DESCMIS && 1741 priv_dev->dev_ver < DEV_VER_V2)) { 1742 cdns3_ep_stall_flush(priv_ep); 1743 } 1744 1745 /* 1746 * For isochronous transfer driver completes request on 1747 * IOC or on TRBERR. IOC appears only when device receive 1748 * OUT data packet. If host disable stream or lost some packet 1749 * then the only way to finish all queued transfer is to do it 1750 * on TRBERR event. 1751 */ 1752 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && 1753 !priv_ep->wa1_set) { 1754 if (!priv_ep->dir) { 1755 u32 ep_cfg = readl(&priv_dev->regs->ep_cfg); 1756 1757 ep_cfg &= ~EP_CFG_ENABLE; 1758 writel(ep_cfg, &priv_dev->regs->ep_cfg); 1759 priv_ep->flags &= ~EP_QUIRK_ISO_OUT_EN; 1760 priv_ep->flags |= EP_UPDATE_EP_TRBADDR; 1761 } 1762 cdns3_transfer_completed(priv_dev, priv_ep); 1763 } else if (!(priv_ep->flags & EP_STALLED) && 1764 !(priv_ep->flags & EP_STALL_PENDING)) { 1765 if (priv_ep->flags & EP_DEFERRED_DRDY) { 1766 priv_ep->flags &= ~EP_DEFERRED_DRDY; 1767 cdns3_start_all_request(priv_dev, priv_ep); 1768 } else { 1769 cdns3_rearm_transfer(priv_ep, 1770 priv_ep->wa1_set); 1771 } 1772 } 1773 } 1774 1775 if ((ep_sts_reg & EP_STS_IOC) || (ep_sts_reg & EP_STS_ISP) || 1776 (ep_sts_reg & EP_STS_IOT)) { 1777 if (priv_ep->flags & EP_QUIRK_EXTRA_BUF_EN) { 1778 if (ep_sts_reg & EP_STS_ISP) 1779 priv_ep->flags |= EP_QUIRK_END_TRANSFER; 1780 else 1781 priv_ep->flags &= ~EP_QUIRK_END_TRANSFER; 1782 } 1783 1784 if (!priv_ep->use_streams) { 1785 if ((ep_sts_reg & EP_STS_IOC) || 1786 (ep_sts_reg & EP_STS_ISP)) { 1787 cdns3_transfer_completed(priv_dev, priv_ep); 1788 } else if ((priv_ep->flags & EP_TDLCHK_EN) & 1789 priv_ep->pending_tdl) { 1790 /* handle IOT with pending tdl */ 1791 cdns3_reprogram_tdl(priv_ep); 1792 } 1793 } else if (priv_ep->dir == USB_DIR_OUT) { 1794 priv_ep->ep_sts_pending |= ep_sts_reg; 1795 } else if (ep_sts_reg & EP_STS_IOT) { 1796 cdns3_transfer_completed(priv_dev, priv_ep); 1797 } 1798 } 1799 1800 /* 1801 * MD_EXIT interrupt sets when stream capable endpoint exits 1802 * from MOVE DATA state of Bulk IN/OUT stream protocol state machine 1803 */ 1804 if (priv_ep->dir == USB_DIR_OUT && (ep_sts_reg & EP_STS_MD_EXIT) && 1805 (priv_ep->ep_sts_pending & EP_STS_IOT) && priv_ep->use_streams) { 1806 priv_ep->ep_sts_pending = 0; 1807 cdns3_transfer_completed(priv_dev, priv_ep); 1808 } 1809 1810 /* 1811 * WA2: this condition should only be meet when 1812 * priv_ep->flags & EP_QUIRK_EXTRA_BUF_DET or 1813 * priv_ep->flags & EP_QUIRK_EXTRA_BUF_EN. 1814 * In other cases this interrupt will be disabled. 1815 */ 1816 if (ep_sts_reg & EP_STS_DESCMIS && priv_dev->dev_ver < DEV_VER_V2 && 1817 !(priv_ep->flags & EP_STALLED)) 1818 cdns3_wa2_descmissing_packet(priv_ep); 1819 1820 return 0; 1821 } 1822 1823 static void cdns3_disconnect_gadget(struct cdns3_device *priv_dev) 1824 { 1825 if (priv_dev->gadget_driver && priv_dev->gadget_driver->disconnect) 1826 priv_dev->gadget_driver->disconnect(&priv_dev->gadget); 1827 } 1828 1829 /** 1830 * cdns3_check_usb_interrupt_proceed - Processes interrupt related to device 1831 * @priv_dev: extended gadget object 1832 * @usb_ists: bitmap representation of device's reported interrupts 1833 * (usb_ists register value) 1834 */ 1835 static void cdns3_check_usb_interrupt_proceed(struct cdns3_device *priv_dev, 1836 u32 usb_ists) 1837 __must_hold(&priv_dev->lock) 1838 { 1839 int speed = 0; 1840 1841 trace_cdns3_usb_irq(priv_dev, usb_ists); 1842 if (usb_ists & USB_ISTS_L1ENTI) { 1843 /* 1844 * WORKAROUND: CDNS3 controller has issue with hardware resuming 1845 * from L1. To fix it, if any DMA transfer is pending driver 1846 * must starts driving resume signal immediately. 1847 */ 1848 if (readl(&priv_dev->regs->drbl)) 1849 __cdns3_gadget_wakeup(priv_dev); 1850 } 1851 1852 /* Connection detected */ 1853 if (usb_ists & (USB_ISTS_CON2I | USB_ISTS_CONI)) { 1854 speed = cdns3_get_speed(priv_dev); 1855 priv_dev->gadget.speed = speed; 1856 usb_gadget_set_state(&priv_dev->gadget, USB_STATE_POWERED); 1857 cdns3_ep0_config(priv_dev); 1858 } 1859 1860 /* Disconnection detected */ 1861 if (usb_ists & (USB_ISTS_DIS2I | USB_ISTS_DISI)) { 1862 spin_unlock(&priv_dev->lock); 1863 cdns3_disconnect_gadget(priv_dev); 1864 spin_lock(&priv_dev->lock); 1865 priv_dev->gadget.speed = USB_SPEED_UNKNOWN; 1866 usb_gadget_set_state(&priv_dev->gadget, USB_STATE_NOTATTACHED); 1867 cdns3_hw_reset_eps_config(priv_dev); 1868 } 1869 1870 if (usb_ists & (USB_ISTS_L2ENTI | USB_ISTS_U3ENTI)) { 1871 if (priv_dev->gadget_driver && 1872 priv_dev->gadget_driver->suspend) { 1873 spin_unlock(&priv_dev->lock); 1874 priv_dev->gadget_driver->suspend(&priv_dev->gadget); 1875 spin_lock(&priv_dev->lock); 1876 } 1877 } 1878 1879 if (usb_ists & (USB_ISTS_L2EXTI | USB_ISTS_U3EXTI)) { 1880 if (priv_dev->gadget_driver && 1881 priv_dev->gadget_driver->resume) { 1882 spin_unlock(&priv_dev->lock); 1883 priv_dev->gadget_driver->resume(&priv_dev->gadget); 1884 spin_lock(&priv_dev->lock); 1885 } 1886 } 1887 1888 /* reset*/ 1889 if (usb_ists & (USB_ISTS_UWRESI | USB_ISTS_UHRESI | USB_ISTS_U2RESI)) { 1890 if (priv_dev->gadget_driver) { 1891 spin_unlock(&priv_dev->lock); 1892 usb_gadget_udc_reset(&priv_dev->gadget, 1893 priv_dev->gadget_driver); 1894 spin_lock(&priv_dev->lock); 1895 1896 /*read again to check the actual speed*/ 1897 speed = cdns3_get_speed(priv_dev); 1898 priv_dev->gadget.speed = speed; 1899 cdns3_hw_reset_eps_config(priv_dev); 1900 cdns3_ep0_config(priv_dev); 1901 } 1902 } 1903 } 1904 1905 /** 1906 * cdns3_device_irq_handler - interrupt handler for device part of controller 1907 * 1908 * @irq: irq number for cdns3 core device 1909 * @data: structure of cdns3 1910 * 1911 * Returns IRQ_HANDLED or IRQ_NONE 1912 */ 1913 static irqreturn_t cdns3_device_irq_handler(int irq, void *data) 1914 { 1915 struct cdns3_device *priv_dev = data; 1916 struct cdns *cdns = dev_get_drvdata(priv_dev->dev); 1917 irqreturn_t ret = IRQ_NONE; 1918 u32 reg; 1919 1920 if (cdns->in_lpm) 1921 return ret; 1922 1923 /* check USB device interrupt */ 1924 reg = readl(&priv_dev->regs->usb_ists); 1925 if (reg) { 1926 /* After masking interrupts the new interrupts won't be 1927 * reported in usb_ists/ep_ists. In order to not lose some 1928 * of them driver disables only detected interrupts. 1929 * They will be enabled ASAP after clearing source of 1930 * interrupt. This an unusual behavior only applies to 1931 * usb_ists register. 1932 */ 1933 reg = ~reg & readl(&priv_dev->regs->usb_ien); 1934 /* mask deferred interrupt. */ 1935 writel(reg, &priv_dev->regs->usb_ien); 1936 ret = IRQ_WAKE_THREAD; 1937 } 1938 1939 /* check endpoint interrupt */ 1940 reg = readl(&priv_dev->regs->ep_ists); 1941 if (reg) { 1942 writel(0, &priv_dev->regs->ep_ien); 1943 ret = IRQ_WAKE_THREAD; 1944 } 1945 1946 return ret; 1947 } 1948 1949 /** 1950 * cdns3_device_thread_irq_handler - interrupt handler for device part 1951 * of controller 1952 * 1953 * @irq: irq number for cdns3 core device 1954 * @data: structure of cdns3 1955 * 1956 * Returns IRQ_HANDLED or IRQ_NONE 1957 */ 1958 static irqreturn_t cdns3_device_thread_irq_handler(int irq, void *data) 1959 { 1960 struct cdns3_device *priv_dev = data; 1961 irqreturn_t ret = IRQ_NONE; 1962 unsigned long flags; 1963 unsigned int bit; 1964 unsigned long reg; 1965 1966 local_bh_disable(); 1967 spin_lock_irqsave(&priv_dev->lock, flags); 1968 1969 reg = readl(&priv_dev->regs->usb_ists); 1970 if (reg) { 1971 writel(reg, &priv_dev->regs->usb_ists); 1972 writel(USB_IEN_INIT, &priv_dev->regs->usb_ien); 1973 cdns3_check_usb_interrupt_proceed(priv_dev, reg); 1974 ret = IRQ_HANDLED; 1975 } 1976 1977 reg = readl(&priv_dev->regs->ep_ists); 1978 1979 /* handle default endpoint OUT */ 1980 if (reg & EP_ISTS_EP_OUT0) { 1981 cdns3_check_ep0_interrupt_proceed(priv_dev, USB_DIR_OUT); 1982 ret = IRQ_HANDLED; 1983 } 1984 1985 /* handle default endpoint IN */ 1986 if (reg & EP_ISTS_EP_IN0) { 1987 cdns3_check_ep0_interrupt_proceed(priv_dev, USB_DIR_IN); 1988 ret = IRQ_HANDLED; 1989 } 1990 1991 /* check if interrupt from non default endpoint, if no exit */ 1992 reg &= ~(EP_ISTS_EP_OUT0 | EP_ISTS_EP_IN0); 1993 if (!reg) 1994 goto irqend; 1995 1996 for_each_set_bit(bit, ®, 1997 sizeof(u32) * BITS_PER_BYTE) { 1998 cdns3_check_ep_interrupt_proceed(priv_dev->eps[bit]); 1999 ret = IRQ_HANDLED; 2000 } 2001 2002 if (priv_dev->dev_ver < DEV_VER_V2 && priv_dev->using_streams) 2003 cdns3_wa2_check_outq_status(priv_dev); 2004 2005 irqend: 2006 writel(~0, &priv_dev->regs->ep_ien); 2007 spin_unlock_irqrestore(&priv_dev->lock, flags); 2008 local_bh_enable(); 2009 2010 return ret; 2011 } 2012 2013 /** 2014 * cdns3_ep_onchip_buffer_reserve - Try to reserve onchip buf for EP 2015 * 2016 * The real reservation will occur during write to EP_CFG register, 2017 * this function is used to check if the 'size' reservation is allowed. 2018 * 2019 * @priv_dev: extended gadget object 2020 * @size: the size (KB) for EP would like to allocate 2021 * @is_in: endpoint direction 2022 * 2023 * Return 0 if the required size can met or negative value on failure 2024 */ 2025 static int cdns3_ep_onchip_buffer_reserve(struct cdns3_device *priv_dev, 2026 int size, int is_in) 2027 { 2028 int remained; 2029 2030 /* 2KB are reserved for EP0*/ 2031 remained = priv_dev->onchip_buffers - priv_dev->onchip_used_size - 2; 2032 2033 if (is_in) { 2034 if (remained < size) 2035 return -EPERM; 2036 2037 priv_dev->onchip_used_size += size; 2038 } else { 2039 int required; 2040 2041 /** 2042 * ALL OUT EPs are shared the same chunk onchip memory, so 2043 * driver checks if it already has assigned enough buffers 2044 */ 2045 if (priv_dev->out_mem_is_allocated >= size) 2046 return 0; 2047 2048 required = size - priv_dev->out_mem_is_allocated; 2049 2050 if (required > remained) 2051 return -EPERM; 2052 2053 priv_dev->out_mem_is_allocated += required; 2054 priv_dev->onchip_used_size += required; 2055 } 2056 2057 return 0; 2058 } 2059 2060 static void cdns3_configure_dmult(struct cdns3_device *priv_dev, 2061 struct cdns3_endpoint *priv_ep) 2062 { 2063 struct cdns3_usb_regs __iomem *regs = priv_dev->regs; 2064 2065 /* For dev_ver > DEV_VER_V2 DMULT is configured per endpoint */ 2066 if (priv_dev->dev_ver <= DEV_VER_V2) 2067 writel(USB_CONF_DMULT, ®s->usb_conf); 2068 2069 if (priv_dev->dev_ver == DEV_VER_V2) 2070 writel(USB_CONF2_EN_TDL_TRB, ®s->usb_conf2); 2071 2072 if (priv_dev->dev_ver >= DEV_VER_V3 && priv_ep) { 2073 u32 mask; 2074 2075 if (priv_ep->dir) 2076 mask = BIT(priv_ep->num + 16); 2077 else 2078 mask = BIT(priv_ep->num); 2079 2080 if (priv_ep->type != USB_ENDPOINT_XFER_ISOC && !priv_ep->dir) { 2081 cdns3_set_register_bit(®s->tdl_from_trb, mask); 2082 cdns3_set_register_bit(®s->tdl_beh, mask); 2083 cdns3_set_register_bit(®s->tdl_beh2, mask); 2084 cdns3_set_register_bit(®s->dma_adv_td, mask); 2085 } 2086 2087 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && !priv_ep->dir) 2088 cdns3_set_register_bit(®s->tdl_from_trb, mask); 2089 2090 cdns3_set_register_bit(®s->dtrans, mask); 2091 } 2092 } 2093 2094 /** 2095 * cdns3_ep_config - Configure hardware endpoint 2096 * @priv_ep: extended endpoint object 2097 * @enable: set EP_CFG_ENABLE bit in ep_cfg register. 2098 */ 2099 int cdns3_ep_config(struct cdns3_endpoint *priv_ep, bool enable) 2100 { 2101 bool is_iso_ep = (priv_ep->type == USB_ENDPOINT_XFER_ISOC); 2102 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 2103 u32 bEndpointAddress = priv_ep->num | priv_ep->dir; 2104 u32 max_packet_size = priv_ep->wMaxPacketSize; 2105 u8 maxburst = priv_ep->bMaxBurst; 2106 u32 ep_cfg = 0; 2107 u8 buffering; 2108 int ret; 2109 2110 buffering = priv_dev->ep_buf_size - 1; 2111 2112 cdns3_configure_dmult(priv_dev, priv_ep); 2113 2114 switch (priv_ep->type) { 2115 case USB_ENDPOINT_XFER_INT: 2116 ep_cfg = EP_CFG_EPTYPE(USB_ENDPOINT_XFER_INT); 2117 2118 if (priv_dev->dev_ver >= DEV_VER_V2 && !priv_ep->dir) 2119 ep_cfg |= EP_CFG_TDL_CHK; 2120 break; 2121 case USB_ENDPOINT_XFER_BULK: 2122 ep_cfg = EP_CFG_EPTYPE(USB_ENDPOINT_XFER_BULK); 2123 2124 if (priv_dev->dev_ver >= DEV_VER_V2 && !priv_ep->dir) 2125 ep_cfg |= EP_CFG_TDL_CHK; 2126 break; 2127 default: 2128 ep_cfg = EP_CFG_EPTYPE(USB_ENDPOINT_XFER_ISOC); 2129 buffering = (priv_ep->bMaxBurst + 1) * (priv_ep->mult + 1) - 1; 2130 } 2131 2132 switch (priv_dev->gadget.speed) { 2133 case USB_SPEED_FULL: 2134 max_packet_size = is_iso_ep ? 1023 : 64; 2135 break; 2136 case USB_SPEED_HIGH: 2137 max_packet_size = is_iso_ep ? 1024 : 512; 2138 break; 2139 case USB_SPEED_SUPER: 2140 if (priv_ep->type != USB_ENDPOINT_XFER_ISOC) { 2141 max_packet_size = 1024; 2142 maxburst = priv_dev->ep_buf_size - 1; 2143 } 2144 break; 2145 default: 2146 /* all other speed are not supported */ 2147 return -EINVAL; 2148 } 2149 2150 if (max_packet_size == 1024) 2151 priv_ep->trb_burst_size = 128; 2152 else if (max_packet_size >= 512) 2153 priv_ep->trb_burst_size = 64; 2154 else 2155 priv_ep->trb_burst_size = 16; 2156 2157 /* 2158 * In versions preceding DEV_VER_V2, for example, iMX8QM, there exit the bugs 2159 * in the DMA. These bugs occur when the trb_burst_size exceeds 16 and the 2160 * address is not aligned to 128 Bytes (which is a product of the 64-bit AXI 2161 * and AXI maximum burst length of 16 or 0xF+1, dma_axi_ctrl0[3:0]). This 2162 * results in data corruption when it crosses the 4K border. The corruption 2163 * specifically occurs from the position (4K - (address & 0x7F)) to 4K. 2164 * 2165 * So force trb_burst_size to 16 at such platform. 2166 */ 2167 if (priv_dev->dev_ver < DEV_VER_V2) 2168 priv_ep->trb_burst_size = 16; 2169 2170 buffering = min_t(u8, buffering, EP_CFG_BUFFERING_MAX); 2171 maxburst = min_t(u8, maxburst, EP_CFG_MAXBURST_MAX); 2172 2173 /* onchip buffer is only allocated before configuration */ 2174 if (!priv_dev->hw_configured_flag) { 2175 ret = cdns3_ep_onchip_buffer_reserve(priv_dev, buffering + 1, 2176 !!priv_ep->dir); 2177 if (ret) { 2178 dev_err(priv_dev->dev, "onchip mem is full, ep is invalid\n"); 2179 return ret; 2180 } 2181 } 2182 2183 if (enable) 2184 ep_cfg |= EP_CFG_ENABLE; 2185 2186 if (priv_ep->use_streams && priv_dev->gadget.speed >= USB_SPEED_SUPER) { 2187 if (priv_dev->dev_ver >= DEV_VER_V3) { 2188 u32 mask = BIT(priv_ep->num + (priv_ep->dir ? 16 : 0)); 2189 2190 /* 2191 * Stream capable endpoints are handled by using ep_tdl 2192 * register. Other endpoints use TDL from TRB feature. 2193 */ 2194 cdns3_clear_register_bit(&priv_dev->regs->tdl_from_trb, 2195 mask); 2196 } 2197 2198 /* Enable Stream Bit TDL chk and SID chk */ 2199 ep_cfg |= EP_CFG_STREAM_EN | EP_CFG_TDL_CHK | EP_CFG_SID_CHK; 2200 } 2201 2202 ep_cfg |= EP_CFG_MAXPKTSIZE(max_packet_size) | 2203 EP_CFG_MULT(priv_ep->mult) | /* must match EP setting */ 2204 EP_CFG_BUFFERING(buffering) | 2205 EP_CFG_MAXBURST(maxburst); 2206 2207 cdns3_select_ep(priv_dev, bEndpointAddress); 2208 writel(ep_cfg, &priv_dev->regs->ep_cfg); 2209 priv_ep->flags |= EP_CONFIGURED; 2210 2211 dev_dbg(priv_dev->dev, "Configure %s: with val %08x\n", 2212 priv_ep->name, ep_cfg); 2213 2214 return 0; 2215 } 2216 2217 /* Find correct direction for HW endpoint according to description */ 2218 static int cdns3_ep_dir_is_correct(struct usb_endpoint_descriptor *desc, 2219 struct cdns3_endpoint *priv_ep) 2220 { 2221 return (priv_ep->endpoint.caps.dir_in && usb_endpoint_dir_in(desc)) || 2222 (priv_ep->endpoint.caps.dir_out && usb_endpoint_dir_out(desc)); 2223 } 2224 2225 static struct 2226 cdns3_endpoint *cdns3_find_available_ep(struct cdns3_device *priv_dev, 2227 struct usb_endpoint_descriptor *desc) 2228 { 2229 struct usb_ep *ep; 2230 struct cdns3_endpoint *priv_ep; 2231 2232 list_for_each_entry(ep, &priv_dev->gadget.ep_list, ep_list) { 2233 unsigned long num; 2234 int ret; 2235 /* ep name pattern likes epXin or epXout */ 2236 char c[2] = {ep->name[2], '\0'}; 2237 2238 ret = kstrtoul(c, 10, &num); 2239 if (ret) 2240 return ERR_PTR(ret); 2241 2242 priv_ep = ep_to_cdns3_ep(ep); 2243 if (cdns3_ep_dir_is_correct(desc, priv_ep)) { 2244 if (!(priv_ep->flags & EP_CLAIMED)) { 2245 priv_ep->num = num; 2246 return priv_ep; 2247 } 2248 } 2249 } 2250 2251 return ERR_PTR(-ENOENT); 2252 } 2253 2254 /* 2255 * Cadence IP has one limitation that all endpoints must be configured 2256 * (Type & MaxPacketSize) before setting configuration through hardware 2257 * register, it means we can't change endpoints configuration after 2258 * set_configuration. 2259 * 2260 * This function set EP_CLAIMED flag which is added when the gadget driver 2261 * uses usb_ep_autoconfig to configure specific endpoint; 2262 * When the udc driver receives set_configurion request, 2263 * it goes through all claimed endpoints, and configure all endpoints 2264 * accordingly. 2265 * 2266 * At usb_ep_ops.enable/disable, we only enable and disable endpoint through 2267 * ep_cfg register which can be changed after set_configuration, and do 2268 * some software operation accordingly. 2269 */ 2270 static struct 2271 usb_ep *cdns3_gadget_match_ep(struct usb_gadget *gadget, 2272 struct usb_endpoint_descriptor *desc, 2273 struct usb_ss_ep_comp_descriptor *comp_desc) 2274 { 2275 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 2276 struct cdns3_endpoint *priv_ep; 2277 unsigned long flags; 2278 2279 priv_ep = cdns3_find_available_ep(priv_dev, desc); 2280 if (IS_ERR(priv_ep)) { 2281 dev_err(priv_dev->dev, "no available ep\n"); 2282 return NULL; 2283 } 2284 2285 dev_dbg(priv_dev->dev, "match endpoint: %s\n", priv_ep->name); 2286 2287 spin_lock_irqsave(&priv_dev->lock, flags); 2288 priv_ep->endpoint.desc = desc; 2289 priv_ep->dir = usb_endpoint_dir_in(desc) ? USB_DIR_IN : USB_DIR_OUT; 2290 priv_ep->type = usb_endpoint_type(desc); 2291 priv_ep->flags |= EP_CLAIMED; 2292 priv_ep->interval = desc->bInterval ? BIT(desc->bInterval - 1) : 0; 2293 priv_ep->wMaxPacketSize = usb_endpoint_maxp(desc); 2294 priv_ep->mult = USB_EP_MAXP_MULT(priv_ep->wMaxPacketSize); 2295 priv_ep->wMaxPacketSize &= USB_ENDPOINT_MAXP_MASK; 2296 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && comp_desc) { 2297 priv_ep->mult = USB_SS_MULT(comp_desc->bmAttributes) - 1; 2298 priv_ep->bMaxBurst = comp_desc->bMaxBurst; 2299 } 2300 2301 spin_unlock_irqrestore(&priv_dev->lock, flags); 2302 return &priv_ep->endpoint; 2303 } 2304 2305 /** 2306 * cdns3_gadget_ep_alloc_request - Allocates request 2307 * @ep: endpoint object associated with request 2308 * @gfp_flags: gfp flags 2309 * 2310 * Returns allocated request address, NULL on allocation error 2311 */ 2312 struct usb_request *cdns3_gadget_ep_alloc_request(struct usb_ep *ep, 2313 gfp_t gfp_flags) 2314 { 2315 struct cdns3_endpoint *priv_ep = ep_to_cdns3_ep(ep); 2316 struct cdns3_request *priv_req; 2317 2318 priv_req = kzalloc(sizeof(*priv_req), gfp_flags); 2319 if (!priv_req) 2320 return NULL; 2321 2322 priv_req->priv_ep = priv_ep; 2323 2324 trace_cdns3_alloc_request(priv_req); 2325 return &priv_req->request; 2326 } 2327 2328 /** 2329 * cdns3_gadget_ep_free_request - Free memory occupied by request 2330 * @ep: endpoint object associated with request 2331 * @request: request to free memory 2332 */ 2333 void cdns3_gadget_ep_free_request(struct usb_ep *ep, 2334 struct usb_request *request) 2335 { 2336 struct cdns3_request *priv_req = to_cdns3_request(request); 2337 2338 if (priv_req->aligned_buf) 2339 priv_req->aligned_buf->in_use = 0; 2340 2341 trace_cdns3_free_request(priv_req); 2342 kfree(priv_req); 2343 } 2344 2345 /** 2346 * cdns3_gadget_ep_enable - Enable endpoint 2347 * @ep: endpoint object 2348 * @desc: endpoint descriptor 2349 * 2350 * Returns 0 on success, error code elsewhere 2351 */ 2352 static int cdns3_gadget_ep_enable(struct usb_ep *ep, 2353 const struct usb_endpoint_descriptor *desc) 2354 { 2355 struct cdns3_endpoint *priv_ep; 2356 struct cdns3_device *priv_dev; 2357 const struct usb_ss_ep_comp_descriptor *comp_desc; 2358 u32 reg = EP_STS_EN_TRBERREN; 2359 u32 bEndpointAddress; 2360 unsigned long flags; 2361 int enable = 1; 2362 int ret = 0; 2363 int val; 2364 2365 if (!ep) { 2366 pr_debug("usbss: ep not configured?\n"); 2367 return -EINVAL; 2368 } 2369 2370 priv_ep = ep_to_cdns3_ep(ep); 2371 priv_dev = priv_ep->cdns3_dev; 2372 comp_desc = priv_ep->endpoint.comp_desc; 2373 2374 if (!desc || desc->bDescriptorType != USB_DT_ENDPOINT) { 2375 dev_dbg(priv_dev->dev, "usbss: invalid parameters\n"); 2376 return -EINVAL; 2377 } 2378 2379 if (!desc->wMaxPacketSize) { 2380 dev_err(priv_dev->dev, "usbss: missing wMaxPacketSize\n"); 2381 return -EINVAL; 2382 } 2383 2384 if (dev_WARN_ONCE(priv_dev->dev, priv_ep->flags & EP_ENABLED, 2385 "%s is already enabled\n", priv_ep->name)) 2386 return 0; 2387 2388 spin_lock_irqsave(&priv_dev->lock, flags); 2389 2390 priv_ep->endpoint.desc = desc; 2391 priv_ep->type = usb_endpoint_type(desc); 2392 priv_ep->interval = desc->bInterval ? BIT(desc->bInterval - 1) : 0; 2393 2394 if (priv_ep->interval > ISO_MAX_INTERVAL && 2395 priv_ep->type == USB_ENDPOINT_XFER_ISOC) { 2396 dev_err(priv_dev->dev, "Driver is limited to %d period\n", 2397 ISO_MAX_INTERVAL); 2398 2399 ret = -EINVAL; 2400 goto exit; 2401 } 2402 2403 bEndpointAddress = priv_ep->num | priv_ep->dir; 2404 cdns3_select_ep(priv_dev, bEndpointAddress); 2405 2406 /* 2407 * For some versions of controller at some point during ISO OUT traffic 2408 * DMA reads Transfer Ring for the EP which has never got doorbell. 2409 * This issue was detected only on simulation, but to avoid this issue 2410 * driver add protection against it. To fix it driver enable ISO OUT 2411 * endpoint before setting DRBL. This special treatment of ISO OUT 2412 * endpoints are recommended by controller specification. 2413 */ 2414 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC && !priv_ep->dir) 2415 enable = 0; 2416 2417 if (usb_ss_max_streams(comp_desc) && usb_endpoint_xfer_bulk(desc)) { 2418 /* 2419 * Enable stream support (SS mode) related interrupts 2420 * in EP_STS_EN Register 2421 */ 2422 if (priv_dev->gadget.speed >= USB_SPEED_SUPER) { 2423 reg |= EP_STS_EN_IOTEN | EP_STS_EN_PRIMEEEN | 2424 EP_STS_EN_SIDERREN | EP_STS_EN_MD_EXITEN | 2425 EP_STS_EN_STREAMREN; 2426 priv_ep->use_streams = true; 2427 ret = cdns3_ep_config(priv_ep, enable); 2428 priv_dev->using_streams |= true; 2429 } 2430 } else { 2431 ret = cdns3_ep_config(priv_ep, enable); 2432 } 2433 2434 if (ret) 2435 goto exit; 2436 2437 ret = cdns3_allocate_trb_pool(priv_ep); 2438 if (ret) 2439 goto exit; 2440 2441 bEndpointAddress = priv_ep->num | priv_ep->dir; 2442 cdns3_select_ep(priv_dev, bEndpointAddress); 2443 2444 trace_cdns3_gadget_ep_enable(priv_ep); 2445 2446 writel(EP_CMD_EPRST, &priv_dev->regs->ep_cmd); 2447 2448 ret = readl_poll_timeout_atomic(&priv_dev->regs->ep_cmd, val, 2449 !(val & (EP_CMD_CSTALL | EP_CMD_EPRST)), 2450 1, 1000); 2451 2452 if (unlikely(ret)) { 2453 cdns3_free_trb_pool(priv_ep); 2454 ret = -EINVAL; 2455 goto exit; 2456 } 2457 2458 /* enable interrupt for selected endpoint */ 2459 cdns3_set_register_bit(&priv_dev->regs->ep_ien, 2460 BIT(cdns3_ep_addr_to_index(bEndpointAddress))); 2461 2462 if (priv_dev->dev_ver < DEV_VER_V2) 2463 cdns3_wa2_enable_detection(priv_dev, priv_ep, reg); 2464 2465 writel(reg, &priv_dev->regs->ep_sts_en); 2466 2467 ep->desc = desc; 2468 priv_ep->flags &= ~(EP_PENDING_REQUEST | EP_STALLED | EP_STALL_PENDING | 2469 EP_QUIRK_ISO_OUT_EN | EP_QUIRK_EXTRA_BUF_EN); 2470 priv_ep->flags |= EP_ENABLED | EP_UPDATE_EP_TRBADDR; 2471 priv_ep->wa1_set = 0; 2472 priv_ep->enqueue = 0; 2473 priv_ep->dequeue = 0; 2474 reg = readl(&priv_dev->regs->ep_sts); 2475 priv_ep->pcs = !!EP_STS_CCS(reg); 2476 priv_ep->ccs = !!EP_STS_CCS(reg); 2477 /* one TRB is reserved for link TRB used in DMULT mode*/ 2478 priv_ep->free_trbs = priv_ep->num_trbs - 1; 2479 exit: 2480 spin_unlock_irqrestore(&priv_dev->lock, flags); 2481 2482 return ret; 2483 } 2484 2485 /** 2486 * cdns3_gadget_ep_disable - Disable endpoint 2487 * @ep: endpoint object 2488 * 2489 * Returns 0 on success, error code elsewhere 2490 */ 2491 static int cdns3_gadget_ep_disable(struct usb_ep *ep) 2492 { 2493 struct cdns3_endpoint *priv_ep; 2494 struct cdns3_request *priv_req; 2495 struct cdns3_device *priv_dev; 2496 struct usb_request *request; 2497 unsigned long flags; 2498 int ret = 0; 2499 u32 ep_cfg; 2500 int val; 2501 2502 if (!ep) { 2503 pr_err("usbss: invalid parameters\n"); 2504 return -EINVAL; 2505 } 2506 2507 priv_ep = ep_to_cdns3_ep(ep); 2508 priv_dev = priv_ep->cdns3_dev; 2509 2510 if (dev_WARN_ONCE(priv_dev->dev, !(priv_ep->flags & EP_ENABLED), 2511 "%s is already disabled\n", priv_ep->name)) 2512 return 0; 2513 2514 spin_lock_irqsave(&priv_dev->lock, flags); 2515 2516 trace_cdns3_gadget_ep_disable(priv_ep); 2517 2518 cdns3_select_ep(priv_dev, ep->desc->bEndpointAddress); 2519 2520 ep_cfg = readl(&priv_dev->regs->ep_cfg); 2521 ep_cfg &= ~EP_CFG_ENABLE; 2522 writel(ep_cfg, &priv_dev->regs->ep_cfg); 2523 2524 /** 2525 * Driver needs some time before resetting endpoint. 2526 * It need waits for clearing DBUSY bit or for timeout expired. 2527 * 10us is enough time for controller to stop transfer. 2528 */ 2529 readl_poll_timeout_atomic(&priv_dev->regs->ep_sts, val, 2530 !(val & EP_STS_DBUSY), 1, 10); 2531 writel(EP_CMD_EPRST, &priv_dev->regs->ep_cmd); 2532 2533 readl_poll_timeout_atomic(&priv_dev->regs->ep_cmd, val, 2534 !(val & (EP_CMD_CSTALL | EP_CMD_EPRST)), 2535 1, 1000); 2536 if (unlikely(ret)) 2537 dev_err(priv_dev->dev, "Timeout: %s resetting failed.\n", 2538 priv_ep->name); 2539 2540 while (!list_empty(&priv_ep->pending_req_list)) { 2541 request = cdns3_next_request(&priv_ep->pending_req_list); 2542 2543 cdns3_gadget_giveback(priv_ep, to_cdns3_request(request), 2544 -ESHUTDOWN); 2545 } 2546 2547 while (!list_empty(&priv_ep->wa2_descmiss_req_list)) { 2548 priv_req = cdns3_next_priv_request(&priv_ep->wa2_descmiss_req_list); 2549 list_del_init(&priv_req->list); 2550 2551 kfree(priv_req->request.buf); 2552 cdns3_gadget_ep_free_request(&priv_ep->endpoint, 2553 &priv_req->request); 2554 --priv_ep->wa2_counter; 2555 } 2556 2557 while (!list_empty(&priv_ep->deferred_req_list)) { 2558 request = cdns3_next_request(&priv_ep->deferred_req_list); 2559 2560 cdns3_gadget_giveback(priv_ep, to_cdns3_request(request), 2561 -ESHUTDOWN); 2562 } 2563 2564 priv_ep->descmis_req = NULL; 2565 2566 ep->desc = NULL; 2567 priv_ep->flags &= ~EP_ENABLED; 2568 priv_ep->use_streams = false; 2569 2570 spin_unlock_irqrestore(&priv_dev->lock, flags); 2571 2572 return ret; 2573 } 2574 2575 /** 2576 * __cdns3_gadget_ep_queue - Transfer data on endpoint 2577 * @ep: endpoint object 2578 * @request: request object 2579 * @gfp_flags: gfp flags 2580 * 2581 * Returns 0 on success, error code elsewhere 2582 */ 2583 static int __cdns3_gadget_ep_queue(struct usb_ep *ep, 2584 struct usb_request *request, 2585 gfp_t gfp_flags) 2586 { 2587 struct cdns3_endpoint *priv_ep = ep_to_cdns3_ep(ep); 2588 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 2589 struct cdns3_request *priv_req; 2590 int ret = 0; 2591 2592 request->actual = 0; 2593 request->status = -EINPROGRESS; 2594 priv_req = to_cdns3_request(request); 2595 trace_cdns3_ep_queue(priv_req); 2596 2597 if (priv_dev->dev_ver < DEV_VER_V2) { 2598 ret = cdns3_wa2_gadget_ep_queue(priv_dev, priv_ep, 2599 priv_req); 2600 2601 if (ret == EINPROGRESS) 2602 return 0; 2603 } 2604 2605 ret = cdns3_prepare_aligned_request_buf(priv_req); 2606 if (ret < 0) 2607 return ret; 2608 2609 if (likely(!(priv_req->flags & REQUEST_UNALIGNED))) { 2610 ret = usb_gadget_map_request_by_dev(priv_dev->sysdev, request, 2611 usb_endpoint_dir_in(ep->desc)); 2612 if (ret) 2613 return ret; 2614 } 2615 2616 list_add_tail(&request->list, &priv_ep->deferred_req_list); 2617 2618 /* 2619 * For stream capable endpoint if prime irq flag is set then only start 2620 * request. 2621 * If hardware endpoint configuration has not been set yet then 2622 * just queue request in deferred list. Transfer will be started in 2623 * cdns3_set_hw_configuration. 2624 */ 2625 if (!request->stream_id) { 2626 if (priv_dev->hw_configured_flag && 2627 !(priv_ep->flags & EP_STALLED) && 2628 !(priv_ep->flags & EP_STALL_PENDING)) 2629 cdns3_start_all_request(priv_dev, priv_ep); 2630 } else { 2631 if (priv_dev->hw_configured_flag && priv_ep->prime_flag) 2632 cdns3_start_all_request(priv_dev, priv_ep); 2633 } 2634 2635 return 0; 2636 } 2637 2638 static int cdns3_gadget_ep_queue(struct usb_ep *ep, struct usb_request *request, 2639 gfp_t gfp_flags) 2640 { 2641 struct usb_request *zlp_request; 2642 struct cdns3_endpoint *priv_ep; 2643 struct cdns3_device *priv_dev; 2644 unsigned long flags; 2645 int ret; 2646 2647 if (!request || !ep) 2648 return -EINVAL; 2649 2650 priv_ep = ep_to_cdns3_ep(ep); 2651 priv_dev = priv_ep->cdns3_dev; 2652 2653 spin_lock_irqsave(&priv_dev->lock, flags); 2654 2655 ret = __cdns3_gadget_ep_queue(ep, request, gfp_flags); 2656 2657 if (ret == 0 && request->zero && request->length && 2658 (request->length % ep->maxpacket == 0)) { 2659 struct cdns3_request *priv_req; 2660 2661 zlp_request = cdns3_gadget_ep_alloc_request(ep, GFP_ATOMIC); 2662 zlp_request->buf = priv_dev->zlp_buf; 2663 zlp_request->length = 0; 2664 2665 priv_req = to_cdns3_request(zlp_request); 2666 priv_req->flags |= REQUEST_ZLP; 2667 2668 dev_dbg(priv_dev->dev, "Queuing ZLP for endpoint: %s\n", 2669 priv_ep->name); 2670 ret = __cdns3_gadget_ep_queue(ep, zlp_request, gfp_flags); 2671 } 2672 2673 spin_unlock_irqrestore(&priv_dev->lock, flags); 2674 return ret; 2675 } 2676 2677 /** 2678 * cdns3_gadget_ep_dequeue - Remove request from transfer queue 2679 * @ep: endpoint object associated with request 2680 * @request: request object 2681 * 2682 * Returns 0 on success, error code elsewhere 2683 */ 2684 int cdns3_gadget_ep_dequeue(struct usb_ep *ep, 2685 struct usb_request *request) 2686 { 2687 struct cdns3_endpoint *priv_ep = ep_to_cdns3_ep(ep); 2688 struct cdns3_device *priv_dev; 2689 struct usb_request *req, *req_temp; 2690 struct cdns3_request *priv_req; 2691 struct cdns3_trb *link_trb; 2692 u8 req_on_hw_ring = 0; 2693 unsigned long flags; 2694 int ret = 0; 2695 int val; 2696 2697 if (!ep || !request || !ep->desc) 2698 return -EINVAL; 2699 2700 priv_dev = priv_ep->cdns3_dev; 2701 2702 spin_lock_irqsave(&priv_dev->lock, flags); 2703 2704 priv_req = to_cdns3_request(request); 2705 2706 trace_cdns3_ep_dequeue(priv_req); 2707 2708 cdns3_select_ep(priv_dev, ep->desc->bEndpointAddress); 2709 2710 list_for_each_entry_safe(req, req_temp, &priv_ep->pending_req_list, 2711 list) { 2712 if (request == req) { 2713 req_on_hw_ring = 1; 2714 goto found; 2715 } 2716 } 2717 2718 list_for_each_entry_safe(req, req_temp, &priv_ep->deferred_req_list, 2719 list) { 2720 if (request == req) 2721 goto found; 2722 } 2723 2724 goto not_found; 2725 2726 found: 2727 link_trb = priv_req->trb; 2728 2729 /* Update ring only if removed request is on pending_req_list list */ 2730 if (req_on_hw_ring && link_trb) { 2731 /* Stop DMA */ 2732 writel(EP_CMD_DFLUSH, &priv_dev->regs->ep_cmd); 2733 2734 /* wait for DFLUSH cleared */ 2735 readl_poll_timeout_atomic(&priv_dev->regs->ep_cmd, val, 2736 !(val & EP_CMD_DFLUSH), 1, 1000); 2737 2738 link_trb->buffer = cpu_to_le32(TRB_BUFFER(priv_ep->trb_pool_dma + 2739 ((priv_req->end_trb + 1) * TRB_SIZE))); 2740 link_trb->control = cpu_to_le32((le32_to_cpu(link_trb->control) & TRB_CYCLE) | 2741 TRB_TYPE(TRB_LINK) | TRB_CHAIN); 2742 2743 if (priv_ep->wa1_trb == priv_req->trb) 2744 cdns3_wa1_restore_cycle_bit(priv_ep); 2745 } 2746 2747 cdns3_gadget_giveback(priv_ep, priv_req, -ECONNRESET); 2748 2749 req = cdns3_next_request(&priv_ep->pending_req_list); 2750 if (req) 2751 cdns3_rearm_transfer(priv_ep, 1); 2752 2753 not_found: 2754 spin_unlock_irqrestore(&priv_dev->lock, flags); 2755 return ret; 2756 } 2757 2758 /** 2759 * __cdns3_gadget_ep_set_halt - Sets stall on selected endpoint 2760 * Should be called after acquiring spin_lock and selecting ep 2761 * @priv_ep: endpoint object to set stall on. 2762 */ 2763 void __cdns3_gadget_ep_set_halt(struct cdns3_endpoint *priv_ep) 2764 { 2765 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 2766 2767 trace_cdns3_halt(priv_ep, 1, 0); 2768 2769 if (!(priv_ep->flags & EP_STALLED)) { 2770 u32 ep_sts_reg = readl(&priv_dev->regs->ep_sts); 2771 2772 if (!(ep_sts_reg & EP_STS_DBUSY)) 2773 cdns3_ep_stall_flush(priv_ep); 2774 else 2775 priv_ep->flags |= EP_STALL_PENDING; 2776 } 2777 } 2778 2779 /** 2780 * __cdns3_gadget_ep_clear_halt - Clears stall on selected endpoint 2781 * Should be called after acquiring spin_lock and selecting ep 2782 * @priv_ep: endpoint object to clear stall on 2783 */ 2784 int __cdns3_gadget_ep_clear_halt(struct cdns3_endpoint *priv_ep) 2785 { 2786 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 2787 struct usb_request *request; 2788 struct cdns3_request *priv_req; 2789 struct cdns3_trb *trb = NULL; 2790 struct cdns3_trb trb_tmp; 2791 int ret; 2792 int val; 2793 2794 trace_cdns3_halt(priv_ep, 0, 0); 2795 2796 request = cdns3_next_request(&priv_ep->pending_req_list); 2797 if (request) { 2798 priv_req = to_cdns3_request(request); 2799 trb = priv_req->trb; 2800 if (trb) { 2801 trb_tmp = *trb; 2802 trb->control = trb->control ^ cpu_to_le32(TRB_CYCLE); 2803 } 2804 } 2805 2806 writel(EP_CMD_CSTALL | EP_CMD_EPRST, &priv_dev->regs->ep_cmd); 2807 2808 /* wait for EPRST cleared */ 2809 ret = readl_poll_timeout_atomic(&priv_dev->regs->ep_cmd, val, 2810 !(val & EP_CMD_EPRST), 1, 100); 2811 if (ret) 2812 return -EINVAL; 2813 2814 priv_ep->flags &= ~(EP_STALLED | EP_STALL_PENDING); 2815 2816 if (request) { 2817 if (trb) 2818 *trb = trb_tmp; 2819 2820 cdns3_rearm_transfer(priv_ep, 1); 2821 } 2822 2823 cdns3_start_all_request(priv_dev, priv_ep); 2824 return ret; 2825 } 2826 2827 /** 2828 * cdns3_gadget_ep_set_halt - Sets/clears stall on selected endpoint 2829 * @ep: endpoint object to set/clear stall on 2830 * @value: 1 for set stall, 0 for clear stall 2831 * 2832 * Returns 0 on success, error code elsewhere 2833 */ 2834 int cdns3_gadget_ep_set_halt(struct usb_ep *ep, int value) 2835 { 2836 struct cdns3_endpoint *priv_ep = ep_to_cdns3_ep(ep); 2837 struct cdns3_device *priv_dev = priv_ep->cdns3_dev; 2838 unsigned long flags; 2839 int ret = 0; 2840 2841 if (!(priv_ep->flags & EP_ENABLED)) 2842 return -EPERM; 2843 2844 spin_lock_irqsave(&priv_dev->lock, flags); 2845 2846 cdns3_select_ep(priv_dev, ep->desc->bEndpointAddress); 2847 2848 if (!value) { 2849 priv_ep->flags &= ~EP_WEDGE; 2850 ret = __cdns3_gadget_ep_clear_halt(priv_ep); 2851 } else { 2852 __cdns3_gadget_ep_set_halt(priv_ep); 2853 } 2854 2855 spin_unlock_irqrestore(&priv_dev->lock, flags); 2856 2857 return ret; 2858 } 2859 2860 extern const struct usb_ep_ops cdns3_gadget_ep0_ops; 2861 2862 static const struct usb_ep_ops cdns3_gadget_ep_ops = { 2863 .enable = cdns3_gadget_ep_enable, 2864 .disable = cdns3_gadget_ep_disable, 2865 .alloc_request = cdns3_gadget_ep_alloc_request, 2866 .free_request = cdns3_gadget_ep_free_request, 2867 .queue = cdns3_gadget_ep_queue, 2868 .dequeue = cdns3_gadget_ep_dequeue, 2869 .set_halt = cdns3_gadget_ep_set_halt, 2870 .set_wedge = cdns3_gadget_ep_set_wedge, 2871 }; 2872 2873 /** 2874 * cdns3_gadget_get_frame - Returns number of actual ITP frame 2875 * @gadget: gadget object 2876 * 2877 * Returns number of actual ITP frame 2878 */ 2879 static int cdns3_gadget_get_frame(struct usb_gadget *gadget) 2880 { 2881 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 2882 2883 return readl(&priv_dev->regs->usb_itpn); 2884 } 2885 2886 int __cdns3_gadget_wakeup(struct cdns3_device *priv_dev) 2887 { 2888 enum usb_device_speed speed; 2889 2890 speed = cdns3_get_speed(priv_dev); 2891 2892 if (speed >= USB_SPEED_SUPER) 2893 return 0; 2894 2895 /* Start driving resume signaling to indicate remote wakeup. */ 2896 writel(USB_CONF_LGO_L0, &priv_dev->regs->usb_conf); 2897 2898 return 0; 2899 } 2900 2901 static int cdns3_gadget_wakeup(struct usb_gadget *gadget) 2902 { 2903 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 2904 unsigned long flags; 2905 int ret = 0; 2906 2907 spin_lock_irqsave(&priv_dev->lock, flags); 2908 ret = __cdns3_gadget_wakeup(priv_dev); 2909 spin_unlock_irqrestore(&priv_dev->lock, flags); 2910 return ret; 2911 } 2912 2913 static int cdns3_gadget_set_selfpowered(struct usb_gadget *gadget, 2914 int is_selfpowered) 2915 { 2916 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 2917 unsigned long flags; 2918 2919 spin_lock_irqsave(&priv_dev->lock, flags); 2920 priv_dev->is_selfpowered = !!is_selfpowered; 2921 spin_unlock_irqrestore(&priv_dev->lock, flags); 2922 return 0; 2923 } 2924 2925 static int cdns3_gadget_pullup(struct usb_gadget *gadget, int is_on) 2926 { 2927 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 2928 2929 if (is_on) { 2930 writel(USB_CONF_DEVEN, &priv_dev->regs->usb_conf); 2931 } else { 2932 writel(~0, &priv_dev->regs->ep_ists); 2933 writel(~0, &priv_dev->regs->usb_ists); 2934 writel(USB_CONF_DEVDS, &priv_dev->regs->usb_conf); 2935 } 2936 2937 return 0; 2938 } 2939 2940 static void cdns3_gadget_config(struct cdns3_device *priv_dev) 2941 { 2942 struct cdns3_usb_regs __iomem *regs = priv_dev->regs; 2943 u32 reg; 2944 2945 cdns3_ep0_config(priv_dev); 2946 2947 /* enable interrupts for endpoint 0 (in and out) */ 2948 writel(EP_IEN_EP_OUT0 | EP_IEN_EP_IN0, ®s->ep_ien); 2949 2950 /* 2951 * Driver needs to modify LFPS minimal U1 Exit time for DEV_VER_TI_V1 2952 * revision of controller. 2953 */ 2954 if (priv_dev->dev_ver == DEV_VER_TI_V1) { 2955 reg = readl(®s->dbg_link1); 2956 2957 reg &= ~DBG_LINK1_LFPS_MIN_GEN_U1_EXIT_MASK; 2958 reg |= DBG_LINK1_LFPS_MIN_GEN_U1_EXIT(0x55) | 2959 DBG_LINK1_LFPS_MIN_GEN_U1_EXIT_SET; 2960 writel(reg, ®s->dbg_link1); 2961 } 2962 2963 /* 2964 * By default some platforms has set protected access to memory. 2965 * This cause problem with cache, so driver restore non-secure 2966 * access to memory. 2967 */ 2968 reg = readl(®s->dma_axi_ctrl); 2969 reg |= DMA_AXI_CTRL_MARPROT(DMA_AXI_CTRL_NON_SECURE) | 2970 DMA_AXI_CTRL_MAWPROT(DMA_AXI_CTRL_NON_SECURE); 2971 writel(reg, ®s->dma_axi_ctrl); 2972 2973 /* enable generic interrupt*/ 2974 writel(USB_IEN_INIT, ®s->usb_ien); 2975 writel(USB_CONF_CLK2OFFDS | USB_CONF_L1DS, ®s->usb_conf); 2976 /* keep Fast Access bit */ 2977 writel(PUSB_PWR_FST_REG_ACCESS, &priv_dev->regs->usb_pwr); 2978 2979 cdns3_configure_dmult(priv_dev, NULL); 2980 } 2981 2982 /** 2983 * cdns3_gadget_udc_start - Gadget start 2984 * @gadget: gadget object 2985 * @driver: driver which operates on this gadget 2986 * 2987 * Returns 0 on success, error code elsewhere 2988 */ 2989 static int cdns3_gadget_udc_start(struct usb_gadget *gadget, 2990 struct usb_gadget_driver *driver) 2991 { 2992 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 2993 unsigned long flags; 2994 enum usb_device_speed max_speed = driver->max_speed; 2995 2996 spin_lock_irqsave(&priv_dev->lock, flags); 2997 priv_dev->gadget_driver = driver; 2998 2999 /* limit speed if necessary */ 3000 max_speed = min(driver->max_speed, gadget->max_speed); 3001 3002 switch (max_speed) { 3003 case USB_SPEED_FULL: 3004 writel(USB_CONF_SFORCE_FS, &priv_dev->regs->usb_conf); 3005 writel(USB_CONF_USB3DIS, &priv_dev->regs->usb_conf); 3006 break; 3007 case USB_SPEED_HIGH: 3008 writel(USB_CONF_USB3DIS, &priv_dev->regs->usb_conf); 3009 break; 3010 case USB_SPEED_SUPER: 3011 break; 3012 default: 3013 dev_err(priv_dev->dev, 3014 "invalid maximum_speed parameter %d\n", 3015 max_speed); 3016 fallthrough; 3017 case USB_SPEED_UNKNOWN: 3018 /* default to superspeed */ 3019 max_speed = USB_SPEED_SUPER; 3020 break; 3021 } 3022 3023 cdns3_gadget_config(priv_dev); 3024 spin_unlock_irqrestore(&priv_dev->lock, flags); 3025 return 0; 3026 } 3027 3028 /** 3029 * cdns3_gadget_udc_stop - Stops gadget 3030 * @gadget: gadget object 3031 * 3032 * Returns 0 3033 */ 3034 static int cdns3_gadget_udc_stop(struct usb_gadget *gadget) 3035 { 3036 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 3037 struct cdns3_endpoint *priv_ep; 3038 u32 bEndpointAddress; 3039 struct usb_ep *ep; 3040 int val; 3041 3042 priv_dev->gadget_driver = NULL; 3043 3044 priv_dev->onchip_used_size = 0; 3045 priv_dev->out_mem_is_allocated = 0; 3046 priv_dev->gadget.speed = USB_SPEED_UNKNOWN; 3047 3048 list_for_each_entry(ep, &priv_dev->gadget.ep_list, ep_list) { 3049 priv_ep = ep_to_cdns3_ep(ep); 3050 bEndpointAddress = priv_ep->num | priv_ep->dir; 3051 cdns3_select_ep(priv_dev, bEndpointAddress); 3052 writel(EP_CMD_EPRST, &priv_dev->regs->ep_cmd); 3053 readl_poll_timeout_atomic(&priv_dev->regs->ep_cmd, val, 3054 !(val & EP_CMD_EPRST), 1, 100); 3055 3056 priv_ep->flags &= ~EP_CLAIMED; 3057 } 3058 3059 /* disable interrupt for device */ 3060 writel(0, &priv_dev->regs->usb_ien); 3061 writel(0, &priv_dev->regs->usb_pwr); 3062 writel(USB_CONF_DEVDS, &priv_dev->regs->usb_conf); 3063 3064 return 0; 3065 } 3066 3067 /** 3068 * cdns3_gadget_check_config - ensure cdns3 can support the USB configuration 3069 * @gadget: pointer to the USB gadget 3070 * 3071 * Used to record the maximum number of endpoints being used in a USB composite 3072 * device. (across all configurations) This is to be used in the calculation 3073 * of the TXFIFO sizes when resizing internal memory for individual endpoints. 3074 * It will help ensured that the resizing logic reserves enough space for at 3075 * least one max packet. 3076 */ 3077 static int cdns3_gadget_check_config(struct usb_gadget *gadget) 3078 { 3079 struct cdns3_device *priv_dev = gadget_to_cdns3_device(gadget); 3080 struct cdns3_endpoint *priv_ep; 3081 struct usb_ep *ep; 3082 int n_in = 0; 3083 int iso = 0; 3084 int out = 1; 3085 int total; 3086 int n; 3087 3088 list_for_each_entry(ep, &gadget->ep_list, ep_list) { 3089 priv_ep = ep_to_cdns3_ep(ep); 3090 if (!(priv_ep->flags & EP_CLAIMED)) 3091 continue; 3092 3093 n = (priv_ep->mult + 1) * (priv_ep->bMaxBurst + 1); 3094 if (ep->address & USB_DIR_IN) { 3095 /* 3096 * ISO transfer: DMA start move data when get ISO, only transfer 3097 * data as min(TD size, iso). No benefit for allocate bigger 3098 * internal memory than 'iso'. 3099 */ 3100 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC) 3101 iso += n; 3102 else 3103 n_in++; 3104 } else { 3105 if (priv_ep->type == USB_ENDPOINT_XFER_ISOC) 3106 out = max_t(int, out, n); 3107 } 3108 } 3109 3110 /* 2KB are reserved for EP0, 1KB for out*/ 3111 total = 2 + n_in + out + iso; 3112 3113 if (total > priv_dev->onchip_buffers) 3114 return -ENOMEM; 3115 3116 priv_dev->ep_buf_size = (priv_dev->onchip_buffers - 2 - iso) / (n_in + out); 3117 3118 return 0; 3119 } 3120 3121 static const struct usb_gadget_ops cdns3_gadget_ops = { 3122 .get_frame = cdns3_gadget_get_frame, 3123 .wakeup = cdns3_gadget_wakeup, 3124 .set_selfpowered = cdns3_gadget_set_selfpowered, 3125 .pullup = cdns3_gadget_pullup, 3126 .udc_start = cdns3_gadget_udc_start, 3127 .udc_stop = cdns3_gadget_udc_stop, 3128 .match_ep = cdns3_gadget_match_ep, 3129 .check_config = cdns3_gadget_check_config, 3130 }; 3131 3132 static void cdns3_free_all_eps(struct cdns3_device *priv_dev) 3133 { 3134 int i; 3135 3136 /* ep0 OUT point to ep0 IN. */ 3137 priv_dev->eps[16] = NULL; 3138 3139 for (i = 0; i < CDNS3_ENDPOINTS_MAX_COUNT; i++) 3140 if (priv_dev->eps[i]) { 3141 cdns3_free_trb_pool(priv_dev->eps[i]); 3142 devm_kfree(priv_dev->dev, priv_dev->eps[i]); 3143 } 3144 } 3145 3146 /** 3147 * cdns3_init_eps - Initializes software endpoints of gadget 3148 * @priv_dev: extended gadget object 3149 * 3150 * Returns 0 on success, error code elsewhere 3151 */ 3152 static int cdns3_init_eps(struct cdns3_device *priv_dev) 3153 { 3154 u32 ep_enabled_reg, iso_ep_reg; 3155 struct cdns3_endpoint *priv_ep; 3156 int ep_dir, ep_number; 3157 u32 ep_mask; 3158 int ret = 0; 3159 int i; 3160 3161 /* Read it from USB_CAP3 to USB_CAP5 */ 3162 ep_enabled_reg = readl(&priv_dev->regs->usb_cap3); 3163 iso_ep_reg = readl(&priv_dev->regs->usb_cap4); 3164 3165 dev_dbg(priv_dev->dev, "Initializing non-zero endpoints\n"); 3166 3167 for (i = 0; i < CDNS3_ENDPOINTS_MAX_COUNT; i++) { 3168 ep_dir = i >> 4; /* i div 16 */ 3169 ep_number = i & 0xF; /* i % 16 */ 3170 ep_mask = BIT(i); 3171 3172 if (!(ep_enabled_reg & ep_mask)) 3173 continue; 3174 3175 if (ep_dir && !ep_number) { 3176 priv_dev->eps[i] = priv_dev->eps[0]; 3177 continue; 3178 } 3179 3180 priv_ep = devm_kzalloc(priv_dev->dev, sizeof(*priv_ep), 3181 GFP_KERNEL); 3182 if (!priv_ep) 3183 goto err; 3184 3185 /* set parent of endpoint object */ 3186 priv_ep->cdns3_dev = priv_dev; 3187 priv_dev->eps[i] = priv_ep; 3188 priv_ep->num = ep_number; 3189 priv_ep->dir = ep_dir ? USB_DIR_IN : USB_DIR_OUT; 3190 3191 if (!ep_number) { 3192 ret = cdns3_init_ep0(priv_dev, priv_ep); 3193 if (ret) { 3194 dev_err(priv_dev->dev, "Failed to init ep0\n"); 3195 goto err; 3196 } 3197 } else { 3198 snprintf(priv_ep->name, sizeof(priv_ep->name), "ep%d%s", 3199 ep_number, !!ep_dir ? "in" : "out"); 3200 priv_ep->endpoint.name = priv_ep->name; 3201 3202 usb_ep_set_maxpacket_limit(&priv_ep->endpoint, 3203 CDNS3_EP_MAX_PACKET_LIMIT); 3204 priv_ep->endpoint.max_streams = CDNS3_EP_MAX_STREAMS; 3205 priv_ep->endpoint.ops = &cdns3_gadget_ep_ops; 3206 if (ep_dir) 3207 priv_ep->endpoint.caps.dir_in = 1; 3208 else 3209 priv_ep->endpoint.caps.dir_out = 1; 3210 3211 if (iso_ep_reg & ep_mask) 3212 priv_ep->endpoint.caps.type_iso = 1; 3213 3214 priv_ep->endpoint.caps.type_bulk = 1; 3215 priv_ep->endpoint.caps.type_int = 1; 3216 3217 list_add_tail(&priv_ep->endpoint.ep_list, 3218 &priv_dev->gadget.ep_list); 3219 } 3220 3221 priv_ep->flags = 0; 3222 3223 dev_dbg(priv_dev->dev, "Initialized %s support: %s %s\n", 3224 priv_ep->name, 3225 priv_ep->endpoint.caps.type_bulk ? "BULK, INT" : "", 3226 priv_ep->endpoint.caps.type_iso ? "ISO" : ""); 3227 3228 INIT_LIST_HEAD(&priv_ep->pending_req_list); 3229 INIT_LIST_HEAD(&priv_ep->deferred_req_list); 3230 INIT_LIST_HEAD(&priv_ep->wa2_descmiss_req_list); 3231 } 3232 3233 return 0; 3234 err: 3235 cdns3_free_all_eps(priv_dev); 3236 return -ENOMEM; 3237 } 3238 3239 static void cdns3_gadget_release(struct device *dev) 3240 { 3241 struct cdns3_device *priv_dev = container_of(dev, 3242 struct cdns3_device, gadget.dev); 3243 3244 kfree(priv_dev); 3245 } 3246 3247 static void cdns3_gadget_exit(struct cdns *cdns) 3248 { 3249 struct cdns3_device *priv_dev; 3250 3251 priv_dev = cdns->gadget_dev; 3252 3253 3254 pm_runtime_mark_last_busy(cdns->dev); 3255 pm_runtime_put_autosuspend(cdns->dev); 3256 3257 usb_del_gadget(&priv_dev->gadget); 3258 devm_free_irq(cdns->dev, cdns->dev_irq, priv_dev); 3259 3260 cdns3_free_all_eps(priv_dev); 3261 3262 while (!list_empty(&priv_dev->aligned_buf_list)) { 3263 struct cdns3_aligned_buf *buf; 3264 3265 buf = cdns3_next_align_buf(&priv_dev->aligned_buf_list); 3266 dma_free_noncoherent(priv_dev->sysdev, buf->size, 3267 buf->buf, 3268 buf->dma, 3269 buf->dir); 3270 3271 list_del(&buf->list); 3272 kfree(buf); 3273 } 3274 3275 dma_free_coherent(priv_dev->sysdev, 8, priv_dev->setup_buf, 3276 priv_dev->setup_dma); 3277 dma_pool_destroy(priv_dev->eps_dma_pool); 3278 3279 kfree(priv_dev->zlp_buf); 3280 usb_put_gadget(&priv_dev->gadget); 3281 cdns->gadget_dev = NULL; 3282 cdns_drd_gadget_off(cdns); 3283 } 3284 3285 static int cdns3_gadget_start(struct cdns *cdns) 3286 { 3287 struct cdns3_device *priv_dev; 3288 u32 max_speed; 3289 int ret; 3290 3291 priv_dev = kzalloc(sizeof(*priv_dev), GFP_KERNEL); 3292 if (!priv_dev) 3293 return -ENOMEM; 3294 3295 usb_initialize_gadget(cdns->dev, &priv_dev->gadget, 3296 cdns3_gadget_release); 3297 cdns->gadget_dev = priv_dev; 3298 priv_dev->sysdev = cdns->dev; 3299 priv_dev->dev = cdns->dev; 3300 priv_dev->regs = cdns->dev_regs; 3301 3302 device_property_read_u16(priv_dev->dev, "cdns,on-chip-buff-size", 3303 &priv_dev->onchip_buffers); 3304 3305 if (priv_dev->onchip_buffers <= 0) { 3306 u32 reg = readl(&priv_dev->regs->usb_cap2); 3307 3308 priv_dev->onchip_buffers = USB_CAP2_ACTUAL_MEM_SIZE(reg); 3309 } 3310 3311 if (!priv_dev->onchip_buffers) 3312 priv_dev->onchip_buffers = 256; 3313 3314 max_speed = usb_get_maximum_speed(cdns->dev); 3315 3316 /* Check the maximum_speed parameter */ 3317 switch (max_speed) { 3318 case USB_SPEED_FULL: 3319 case USB_SPEED_HIGH: 3320 case USB_SPEED_SUPER: 3321 break; 3322 default: 3323 dev_err(cdns->dev, "invalid maximum_speed parameter %d\n", 3324 max_speed); 3325 fallthrough; 3326 case USB_SPEED_UNKNOWN: 3327 /* default to superspeed */ 3328 max_speed = USB_SPEED_SUPER; 3329 break; 3330 } 3331 3332 /* fill gadget fields */ 3333 priv_dev->gadget.max_speed = max_speed; 3334 priv_dev->gadget.speed = USB_SPEED_UNKNOWN; 3335 priv_dev->gadget.ops = &cdns3_gadget_ops; 3336 priv_dev->gadget.name = "usb-ss-gadget"; 3337 priv_dev->gadget.quirk_avoids_skb_reserve = 1; 3338 priv_dev->gadget.irq = cdns->dev_irq; 3339 3340 spin_lock_init(&priv_dev->lock); 3341 INIT_WORK(&priv_dev->pending_status_wq, 3342 cdns3_pending_setup_status_handler); 3343 3344 INIT_WORK(&priv_dev->aligned_buf_wq, 3345 cdns3_free_aligned_request_buf); 3346 3347 /* initialize endpoint container */ 3348 INIT_LIST_HEAD(&priv_dev->gadget.ep_list); 3349 INIT_LIST_HEAD(&priv_dev->aligned_buf_list); 3350 priv_dev->eps_dma_pool = dma_pool_create("cdns3_eps_dma_pool", 3351 priv_dev->sysdev, 3352 TRB_RING_SIZE, 8, 0); 3353 if (!priv_dev->eps_dma_pool) { 3354 dev_err(priv_dev->dev, "Failed to create TRB dma pool\n"); 3355 ret = -ENOMEM; 3356 goto err1; 3357 } 3358 3359 ret = cdns3_init_eps(priv_dev); 3360 if (ret) { 3361 dev_err(priv_dev->dev, "Failed to create endpoints\n"); 3362 goto err1; 3363 } 3364 3365 /* allocate memory for setup packet buffer */ 3366 priv_dev->setup_buf = dma_alloc_coherent(priv_dev->sysdev, 8, 3367 &priv_dev->setup_dma, GFP_DMA); 3368 if (!priv_dev->setup_buf) { 3369 ret = -ENOMEM; 3370 goto err2; 3371 } 3372 3373 priv_dev->dev_ver = readl(&priv_dev->regs->usb_cap6); 3374 3375 dev_dbg(priv_dev->dev, "Device Controller version: %08x\n", 3376 readl(&priv_dev->regs->usb_cap6)); 3377 dev_dbg(priv_dev->dev, "USB Capabilities:: %08x\n", 3378 readl(&priv_dev->regs->usb_cap1)); 3379 dev_dbg(priv_dev->dev, "On-Chip memory configuration: %08x\n", 3380 readl(&priv_dev->regs->usb_cap2)); 3381 3382 priv_dev->dev_ver = GET_DEV_BASE_VERSION(priv_dev->dev_ver); 3383 if (priv_dev->dev_ver >= DEV_VER_V2) 3384 priv_dev->gadget.sg_supported = 1; 3385 3386 priv_dev->zlp_buf = kzalloc(CDNS3_EP_ZLP_BUF_SIZE, GFP_KERNEL); 3387 if (!priv_dev->zlp_buf) { 3388 ret = -ENOMEM; 3389 goto err3; 3390 } 3391 3392 /* add USB gadget device */ 3393 ret = usb_add_gadget(&priv_dev->gadget); 3394 if (ret < 0) { 3395 dev_err(priv_dev->dev, "Failed to add gadget\n"); 3396 goto err4; 3397 } 3398 3399 return 0; 3400 err4: 3401 kfree(priv_dev->zlp_buf); 3402 err3: 3403 dma_free_coherent(priv_dev->sysdev, 8, priv_dev->setup_buf, 3404 priv_dev->setup_dma); 3405 err2: 3406 cdns3_free_all_eps(priv_dev); 3407 err1: 3408 dma_pool_destroy(priv_dev->eps_dma_pool); 3409 3410 usb_put_gadget(&priv_dev->gadget); 3411 cdns->gadget_dev = NULL; 3412 return ret; 3413 } 3414 3415 static int __cdns3_gadget_init(struct cdns *cdns) 3416 { 3417 int ret = 0; 3418 3419 /* Ensure 32-bit DMA Mask in case we switched back from Host mode */ 3420 ret = dma_set_mask_and_coherent(cdns->dev, DMA_BIT_MASK(32)); 3421 if (ret) { 3422 dev_err(cdns->dev, "Failed to set dma mask: %d\n", ret); 3423 return ret; 3424 } 3425 3426 cdns_drd_gadget_on(cdns); 3427 pm_runtime_get_sync(cdns->dev); 3428 3429 ret = cdns3_gadget_start(cdns); 3430 if (ret) { 3431 pm_runtime_put_sync(cdns->dev); 3432 return ret; 3433 } 3434 3435 /* 3436 * Because interrupt line can be shared with other components in 3437 * driver it can't use IRQF_ONESHOT flag here. 3438 */ 3439 ret = devm_request_threaded_irq(cdns->dev, cdns->dev_irq, 3440 cdns3_device_irq_handler, 3441 cdns3_device_thread_irq_handler, 3442 IRQF_SHARED, dev_name(cdns->dev), 3443 cdns->gadget_dev); 3444 3445 if (ret) 3446 goto err0; 3447 3448 return 0; 3449 err0: 3450 cdns3_gadget_exit(cdns); 3451 return ret; 3452 } 3453 3454 static int cdns3_gadget_suspend(struct cdns *cdns, bool do_wakeup) 3455 __must_hold(&cdns->lock) 3456 { 3457 struct cdns3_device *priv_dev = cdns->gadget_dev; 3458 3459 spin_unlock(&cdns->lock); 3460 cdns3_disconnect_gadget(priv_dev); 3461 spin_lock(&cdns->lock); 3462 3463 priv_dev->gadget.speed = USB_SPEED_UNKNOWN; 3464 usb_gadget_set_state(&priv_dev->gadget, USB_STATE_NOTATTACHED); 3465 cdns3_hw_reset_eps_config(priv_dev); 3466 3467 /* disable interrupt for device */ 3468 writel(0, &priv_dev->regs->usb_ien); 3469 3470 return 0; 3471 } 3472 3473 static int cdns3_gadget_resume(struct cdns *cdns, bool lost_power) 3474 { 3475 struct cdns3_device *priv_dev = cdns->gadget_dev; 3476 3477 if (!priv_dev->gadget_driver) 3478 return 0; 3479 3480 cdns3_gadget_config(priv_dev); 3481 if (lost_power) 3482 writel(USB_CONF_DEVEN, &priv_dev->regs->usb_conf); 3483 3484 return 0; 3485 } 3486 3487 /** 3488 * cdns3_gadget_init - initialize device structure 3489 * 3490 * @cdns: cdns instance 3491 * 3492 * This function initializes the gadget. 3493 */ 3494 int cdns3_gadget_init(struct cdns *cdns) 3495 { 3496 struct cdns_role_driver *rdrv; 3497 3498 rdrv = devm_kzalloc(cdns->dev, sizeof(*rdrv), GFP_KERNEL); 3499 if (!rdrv) 3500 return -ENOMEM; 3501 3502 rdrv->start = __cdns3_gadget_init; 3503 rdrv->stop = cdns3_gadget_exit; 3504 rdrv->suspend = cdns3_gadget_suspend; 3505 rdrv->resume = cdns3_gadget_resume; 3506 rdrv->state = CDNS_ROLE_STATE_INACTIVE; 3507 rdrv->name = "gadget"; 3508 cdns->roles[USB_ROLE_DEVICE] = rdrv; 3509 3510 return 0; 3511 } 3512