|
@@ -50,6 +50,8 @@
|
|
/* For each endpoint, we need 2 QTDs, one for each of IN and OUT */
|
|
/* For each endpoint, we need 2 QTDs, one for each of IN and OUT */
|
|
#define ILIST_SZ (NUM_ENDPOINTS * 2 * ILIST_ENT_SZ)
|
|
#define ILIST_SZ (NUM_ENDPOINTS * 2 * ILIST_ENT_SZ)
|
|
|
|
|
|
|
|
+#define EP_MAX_LENGTH_TRANSFER 0x4000
|
|
|
|
+
|
|
#ifndef DEBUG
|
|
#ifndef DEBUG
|
|
#define DBG(x...) do {} while (0)
|
|
#define DBG(x...) do {} while (0)
|
|
#else
|
|
#else
|
|
@@ -102,13 +104,28 @@ static struct usb_ep_ops ci_ep_ops = {
|
|
};
|
|
};
|
|
|
|
|
|
/* Init values for USB endpoints. */
|
|
/* Init values for USB endpoints. */
|
|
-static const struct usb_ep ci_ep_init[2] = {
|
|
|
|
|
|
+static const struct usb_ep ci_ep_init[5] = {
|
|
[0] = { /* EP 0 */
|
|
[0] = { /* EP 0 */
|
|
.maxpacket = 64,
|
|
.maxpacket = 64,
|
|
.name = "ep0",
|
|
.name = "ep0",
|
|
.ops = &ci_ep_ops,
|
|
.ops = &ci_ep_ops,
|
|
},
|
|
},
|
|
- [1] = { /* EP 1..n */
|
|
|
|
|
|
+ [1] = {
|
|
|
|
+ .maxpacket = 512,
|
|
|
|
+ .name = "ep1in-bulk",
|
|
|
|
+ .ops = &ci_ep_ops,
|
|
|
|
+ },
|
|
|
|
+ [2] = {
|
|
|
|
+ .maxpacket = 512,
|
|
|
|
+ .name = "ep2out-bulk",
|
|
|
|
+ .ops = &ci_ep_ops,
|
|
|
|
+ },
|
|
|
|
+ [3] = {
|
|
|
|
+ .maxpacket = 512,
|
|
|
|
+ .name = "ep3in-int",
|
|
|
|
+ .ops = &ci_ep_ops,
|
|
|
|
+ },
|
|
|
|
+ [4] = {
|
|
.maxpacket = 512,
|
|
.maxpacket = 512,
|
|
.name = "ep-",
|
|
.name = "ep-",
|
|
.ops = &ci_ep_ops,
|
|
.ops = &ci_ep_ops,
|
|
@@ -196,6 +213,19 @@ static void ci_flush_qtd(int ep_num)
|
|
flush_dcache_range(start, end);
|
|
flush_dcache_range(start, end);
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+/**
|
|
|
|
+ * ci_flush_td - flush cache over queue item
|
|
|
|
+ * @td: td pointer
|
|
|
|
+ *
|
|
|
|
+ * This function flushes cache for particular transfer descriptor.
|
|
|
|
+ */
|
|
|
|
+static void ci_flush_td(struct ept_queue_item *td)
|
|
|
|
+{
|
|
|
|
+ const uint32_t start = (uint32_t)td;
|
|
|
|
+ const uint32_t end = (uint32_t) td + ILIST_ENT_SZ;
|
|
|
|
+ flush_dcache_range(start, end);
|
|
|
|
+}
|
|
|
|
+
|
|
/**
|
|
/**
|
|
* ci_invalidate_qtd - invalidate cache over queue item
|
|
* ci_invalidate_qtd - invalidate cache over queue item
|
|
* @ep_num: Endpoint number
|
|
* @ep_num: Endpoint number
|
|
@@ -211,6 +241,19 @@ static void ci_invalidate_qtd(int ep_num)
|
|
invalidate_dcache_range(start, end);
|
|
invalidate_dcache_range(start, end);
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+/**
|
|
|
|
+ * ci_invalidate_td - invalidate cache over queue item
|
|
|
|
+ * @td: td pointer
|
|
|
|
+ *
|
|
|
|
+ * This function invalidates cache for particular transfer descriptor.
|
|
|
|
+ */
|
|
|
|
+static void ci_invalidate_td(struct ept_queue_item *td)
|
|
|
|
+{
|
|
|
|
+ const uint32_t start = (uint32_t)td;
|
|
|
|
+ const uint32_t end = start + ILIST_ENT_SZ;
|
|
|
|
+ invalidate_dcache_range(start, end);
|
|
|
|
+}
|
|
|
|
+
|
|
static struct usb_request *
|
|
static struct usb_request *
|
|
ci_ep_alloc_request(struct usb_ep *ep, unsigned int gfp_flags)
|
|
ci_ep_alloc_request(struct usb_ep *ep, unsigned int gfp_flags)
|
|
{
|
|
{
|
|
@@ -376,6 +419,9 @@ static void ci_ep_submit_next_request(struct ci_ep *ci_ep)
|
|
struct ept_queue_head *head;
|
|
struct ept_queue_head *head;
|
|
int bit, num, len, in;
|
|
int bit, num, len, in;
|
|
struct ci_req *ci_req;
|
|
struct ci_req *ci_req;
|
|
|
|
+ u8 *buf;
|
|
|
|
+ uint32_t length, actlen;
|
|
|
|
+ struct ept_queue_item *dtd, *qtd;
|
|
|
|
|
|
ci_ep->req_primed = true;
|
|
ci_ep->req_primed = true;
|
|
|
|
|
|
@@ -387,16 +433,41 @@ static void ci_ep_submit_next_request(struct ci_ep *ci_ep)
|
|
ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue);
|
|
ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue);
|
|
len = ci_req->req.length;
|
|
len = ci_req->req.length;
|
|
|
|
|
|
- item->info = INFO_BYTES(len) | INFO_ACTIVE;
|
|
|
|
- item->page0 = (unsigned long)ci_req->hw_buf;
|
|
|
|
- item->page1 = ((unsigned long)ci_req->hw_buf & 0xfffff000) + 0x1000;
|
|
|
|
- item->page2 = ((unsigned long)ci_req->hw_buf & 0xfffff000) + 0x2000;
|
|
|
|
- item->page3 = ((unsigned long)ci_req->hw_buf & 0xfffff000) + 0x3000;
|
|
|
|
- item->page4 = ((unsigned long)ci_req->hw_buf & 0xfffff000) + 0x4000;
|
|
|
|
-
|
|
|
|
head->next = (unsigned long)item;
|
|
head->next = (unsigned long)item;
|
|
head->info = 0;
|
|
head->info = 0;
|
|
|
|
|
|
|
|
+ ci_req->dtd_count = 0;
|
|
|
|
+ buf = ci_req->hw_buf;
|
|
|
|
+ actlen = 0;
|
|
|
|
+ dtd = item;
|
|
|
|
+
|
|
|
|
+ do {
|
|
|
|
+ length = min(ci_req->req.length - actlen,
|
|
|
|
+ (unsigned)EP_MAX_LENGTH_TRANSFER);
|
|
|
|
+
|
|
|
|
+ dtd->info = INFO_BYTES(length) | INFO_ACTIVE;
|
|
|
|
+ dtd->page0 = (unsigned long)buf;
|
|
|
|
+ dtd->page1 = ((unsigned long)buf & 0xfffff000) + 0x1000;
|
|
|
|
+ dtd->page2 = ((unsigned long)buf & 0xfffff000) + 0x2000;
|
|
|
|
+ dtd->page3 = ((unsigned long)buf & 0xfffff000) + 0x3000;
|
|
|
|
+ dtd->page4 = ((unsigned long)buf & 0xfffff000) + 0x4000;
|
|
|
|
+
|
|
|
|
+ len -= length;
|
|
|
|
+ actlen += length;
|
|
|
|
+ buf += length;
|
|
|
|
+
|
|
|
|
+ if (len) {
|
|
|
|
+ qtd = (struct ept_queue_item *)
|
|
|
|
+ memalign(ILIST_ALIGN, ILIST_ENT_SZ);
|
|
|
|
+ dtd->next = (uint32_t)qtd;
|
|
|
|
+ dtd = qtd;
|
|
|
|
+ memset(dtd, 0, ILIST_ENT_SZ);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ ci_req->dtd_count++;
|
|
|
|
+ } while (len);
|
|
|
|
+
|
|
|
|
+ item = dtd;
|
|
/*
|
|
/*
|
|
* When sending the data for an IN transaction, the attached host
|
|
* When sending the data for an IN transaction, the attached host
|
|
* knows that all data for the IN is sent when one of the following
|
|
* knows that all data for the IN is sent when one of the following
|
|
@@ -432,6 +503,12 @@ static void ci_ep_submit_next_request(struct ci_ep *ci_ep)
|
|
|
|
|
|
ci_flush_qtd(num);
|
|
ci_flush_qtd(num);
|
|
|
|
|
|
|
|
+ item = (struct ept_queue_item *)head->next;
|
|
|
|
+ while (item->next != TERMINATE) {
|
|
|
|
+ ci_flush_td((struct ept_queue_item *)item->next);
|
|
|
|
+ item = (struct ept_queue_item *)item->next;
|
|
|
|
+ }
|
|
|
|
+
|
|
DBG("ept%d %s queue len %x, req %p, buffer %p\n",
|
|
DBG("ept%d %s queue len %x, req %p, buffer %p\n",
|
|
num, in ? "in" : "out", len, ci_req, ci_req->hw_buf);
|
|
num, in ? "in" : "out", len, ci_req, ci_req->hw_buf);
|
|
ci_flush_qh(num);
|
|
ci_flush_qh(num);
|
|
@@ -497,21 +574,31 @@ static void flip_ep0_direction(void)
|
|
|
|
|
|
static void handle_ep_complete(struct ci_ep *ci_ep)
|
|
static void handle_ep_complete(struct ci_ep *ci_ep)
|
|
{
|
|
{
|
|
- struct ept_queue_item *item;
|
|
|
|
- int num, in, len;
|
|
|
|
|
|
+ struct ept_queue_item *item, *next_td;
|
|
|
|
+ int num, in, len, j;
|
|
struct ci_req *ci_req;
|
|
struct ci_req *ci_req;
|
|
|
|
|
|
num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK;
|
|
num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK;
|
|
in = (ci_ep->desc->bEndpointAddress & USB_DIR_IN) != 0;
|
|
in = (ci_ep->desc->bEndpointAddress & USB_DIR_IN) != 0;
|
|
item = ci_get_qtd(num, in);
|
|
item = ci_get_qtd(num, in);
|
|
ci_invalidate_qtd(num);
|
|
ci_invalidate_qtd(num);
|
|
|
|
+ ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue);
|
|
|
|
|
|
- len = (item->info >> 16) & 0x7fff;
|
|
|
|
- if (item->info & 0xff)
|
|
|
|
- printf("EP%d/%s FAIL info=%x pg0=%x\n",
|
|
|
|
- num, in ? "in" : "out", item->info, item->page0);
|
|
|
|
|
|
+ next_td = item;
|
|
|
|
+ len = 0;
|
|
|
|
+ for (j = 0; j < ci_req->dtd_count; j++) {
|
|
|
|
+ ci_invalidate_td(next_td);
|
|
|
|
+ item = next_td;
|
|
|
|
+ len += (item->info >> 16) & 0x7fff;
|
|
|
|
+ if (item->info & 0xff)
|
|
|
|
+ printf("EP%d/%s FAIL info=%x pg0=%x\n",
|
|
|
|
+ num, in ? "in" : "out", item->info, item->page0);
|
|
|
|
+ if (j != ci_req->dtd_count - 1)
|
|
|
|
+ next_td = (struct ept_queue_item *)item->next;
|
|
|
|
+ if (j != 0)
|
|
|
|
+ free(item);
|
|
|
|
+ }
|
|
|
|
|
|
- ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue);
|
|
|
|
list_del_init(&ci_req->queue);
|
|
list_del_init(&ci_req->queue);
|
|
ci_ep->req_primed = false;
|
|
ci_ep->req_primed = false;
|
|
|
|
|
|
@@ -852,9 +939,19 @@ static int ci_udc_probe(void)
|
|
controller.gadget.ep0 = &controller.ep[0].ep;
|
|
controller.gadget.ep0 = &controller.ep[0].ep;
|
|
INIT_LIST_HEAD(&controller.gadget.ep0->ep_list);
|
|
INIT_LIST_HEAD(&controller.gadget.ep0->ep_list);
|
|
|
|
|
|
- /* Init EP 1..n */
|
|
|
|
- for (i = 1; i < NUM_ENDPOINTS; i++) {
|
|
|
|
- memcpy(&controller.ep[i].ep, &ci_ep_init[1],
|
|
|
|
|
|
+ /* Init EP 1..3 */
|
|
|
|
+ for (i = 1; i < 4; i++) {
|
|
|
|
+ memcpy(&controller.ep[i].ep, &ci_ep_init[i],
|
|
|
|
+ sizeof(*ci_ep_init));
|
|
|
|
+ INIT_LIST_HEAD(&controller.ep[i].queue);
|
|
|
|
+ controller.ep[i].req_primed = false;
|
|
|
|
+ list_add_tail(&controller.ep[i].ep.ep_list,
|
|
|
|
+ &controller.gadget.ep_list);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ /* Init EP 4..n */
|
|
|
|
+ for (i = 4; i < NUM_ENDPOINTS; i++) {
|
|
|
|
+ memcpy(&controller.ep[i].ep, &ci_ep_init[4],
|
|
sizeof(*ci_ep_init));
|
|
sizeof(*ci_ep_init));
|
|
INIT_LIST_HEAD(&controller.ep[i].queue);
|
|
INIT_LIST_HEAD(&controller.ep[i].queue);
|
|
controller.ep[i].req_primed = false;
|
|
controller.ep[i].req_primed = false;
|