
Hi Dan,
On Tue, Sep 3, 2013 at 7:28 PM, Dan Murphy dmurphy@ti.com wrote:
On 08/21/2013 05:12 AM, Vivek Gautam wrote:
This adds stack layer for eXtensible Host Controller Interface which facilitates use of USB 3.0 in host mode.
Adapting xHCI host controller driver in linux-kernel by Sarah Sharp to needs in u-boot.
This adds the basic xHCI host controller driver with bare minimum features:
- Control/Bulk transfer support has been added with required infrastructure for necessary xHC data structures.
- Stream protocol hasn't been supported yet.
- No support for quirky devices has been added.
Signed-off-by: Vikas C Sajjan vikas.sajjan@samsung.com Signed-off-by: Julius Werner jwerner@chromium.org Signed-off-by: Vivek Gautam gautam.vivek@samsung.com Cc: Simon Glass sjg@chromium.org Cc: Minkyu Kang mk7.kang@samsung.com Cc: Dan Murphy dmurphy@ti.com Cc: Marek Vasut marex@denx.de
common/usb.c | 27 +- drivers/usb/host/Makefile | 3 + drivers/usb/host/xhci-mem.c | 731 ++++++++++++++++++++++++ drivers/usb/host/xhci-ring.c | 950 +++++++++++++++++++++++++++++++ drivers/usb/host/xhci.c | 1040 ++++++++++++++++++++++++++++++++++ drivers/usb/host/xhci.h | 1280 ++++++++++++++++++++++++++++++++++++++++++ include/usb.h | 9 +- 7 files changed, 4037 insertions(+), 3 deletions(-) create mode 100644 drivers/usb/host/xhci-mem.c create mode 100644 drivers/usb/host/xhci-ring.c create mode 100644 drivers/usb/host/xhci.c create mode 100644 drivers/usb/host/xhci.h
diff --git a/common/usb.c b/common/usb.c index c97f522..e7800fa 100644 --- a/common/usb.c +++ b/common/usb.c @@ -855,6 +855,16 @@ void usb_free_device(void) }
/*
- XHCI issues Enable Slot command and thereafter
- allocates device contexts. Provide a weak alias
- function for the purpose, so that XHCI overrides it
- and EHCI/OHCI just work out of the box.
- */
+__weak int usb_alloc_device(struct usb_device *udev) +{
return 0;
+} +/*
- By the time we get here, the device has gotten a new device ID
- and is in the default state. We need to identify the thing and
- get the ball rolling..
@@ -867,6 +877,17 @@ int usb_new_device(struct usb_device *dev) int tmp; ALLOC_CACHE_ALIGN_BUFFER(unsigned char, tmpbuf, USB_BUFSIZ);
/*
* Allocate usb 3.0 device context.
* USB 3.0 (xHCI) protocol tries to allocate device slot
* and related data structures first. This call does that.
* Refer to sec 4.3.2 in xHCI spec rev1.0
*/
if (usb_alloc_device(dev)) {
printf("Cannot allocate device context to get SLOT_ID\n");
return -1;
}
/* We still haven't set the Address yet */ addr = dev->devnum; dev->devnum = 0;
@@ -897,7 +918,7 @@ int usb_new_device(struct usb_device *dev) * http://sourceforge.net/mailarchive/forum.php? * thread_id=5729457&forum_id=5398 */
struct usb_device_descriptor *desc;
__maybe_unused struct usb_device_descriptor *desc; int port = -1; struct usb_device *parent = dev->parent; unsigned short portstatus;
@@ -914,6 +935,7 @@ int usb_new_device(struct usb_device *dev) dev->epmaxpacketin[0] = 64; dev->epmaxpacketout[0] = 64;
+#ifndef CONFIG_USB_XHCI
Add a comment why we cannot get the descriptor on xHCI.
Ok, will add a comment here.
err = usb_get_descriptor(dev, USB_DT_DEVICE, 0, desc, 64); if (err < 0) { debug("usb_new_device: usb_get_descriptor() failed\n");
@@ -926,11 +948,12 @@ int usb_new_device(struct usb_device *dev) * to differentiate between HUB and DEVICE. */ dev->descriptor.bDeviceClass = desc->bDeviceClass; +#endif
/* find the port number we're at */ if (parent) { int j;
/* find the port number we're at */ for (j = 0; j < parent->maxchild; j++) { if (parent->children[j] == dev) { port = j;
diff --git a/drivers/usb/host/Makefile b/drivers/usb/host/Makefile index ff6c80e..6bd6c86 100644 --- a/drivers/usb/host/Makefile +++ b/drivers/usb/host/Makefile @@ -42,6 +42,9 @@ COBJS-$(CONFIG_USB_EHCI_SPEAR) += ehci-spear.o COBJS-$(CONFIG_USB_EHCI_TEGRA) += ehci-tegra.o COBJS-$(CONFIG_USB_EHCI_VCT) += ehci-vct.o
+# xhci +COBJS-$(CONFIG_USB_XHCI) += xhci.o xhci-mem.o xhci-ring.o
COBJS := $(COBJS-y) SRCS := $(COBJS:.o=.c) OBJS := $(addprefix $(obj),$(COBJS)) diff --git a/drivers/usb/host/xhci-mem.c b/drivers/usb/host/xhci-mem.c new file mode 100644 index 0000000..709ef7e --- /dev/null +++ b/drivers/usb/host/xhci-mem.c @@ -0,0 +1,731 @@ +/*
- USB HOST XHCI Controller stack
- Copyright (C) 2013 Samsung Electronics Co.Ltd
- Vivek Gautam gautam.vivek@samsung.com
- Vikas Sajjan vikas.sajjan@samsung.com
- Based on xHCI host controller driver in linux-kernel
- by Sarah Sharp.
- This program is free software; you can redistribute it and/or
- modify it under the terms of the GNU General Public License as
- published by the Free Software Foundation; either version 2 of
- the License, or (at your option) any later version.
- This program is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- GNU General Public License for more details.
- You should have received a copy of the GNU General Public License
- along with this program; if not, write to the Free Software
- Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
- MA 02110-1301 USA
- */
Needs new SPDX license.
- SPDX-License-Identifier: GPL-2.0+
Ok, will put SPDX license.
+#include <common.h> +#include <asm/byteorder.h> +#include <usb.h> +#include <asm/io.h>
Move this to the xhci.h as it is needed there for the write/read interfaces
Sure, so that we will not require the patch posted by you.
+#include <malloc.h> +#include <asm/cache.h> +#include <asm-generic/errno.h>
+#include "xhci.h"
+#define CACHELINE_SIZE CONFIG_SYS_CACHELINE_SIZE +/**
- flushes the address passed till the length
- @param addr pointer to memory region to be flushed
- @param len the length of the cache line to be flushed
- @return none
- */
+void xhci_flush_cache(uint32_t addr, u32 len) +{
BUG_ON((void *)addr == NULL || len == 0);
flush_dcache_range(addr & ~(CACHELINE_SIZE - 1),
ALIGN(addr + len, CACHELINE_SIZE));
+}
+/**
- invalidates the address passed till the length
- @param addr pointer to memory region to be invalidates
- @param len the length of the cache line to be invalidated
- @return none
- */
+void xhci_inval_cache(uint32_t addr, u32 len) +{
BUG_ON((void *)addr == NULL || len == 0);
invalidate_dcache_range(addr & ~(CACHELINE_SIZE - 1),
ALIGN(addr + len, CACHELINE_SIZE));
+}
+/**
- frees the "segment" pointer passed
- @param ptr pointer to "segement" to be freed
- @return none
- */
+static void xhci_segment_free(struct xhci_segment *seg) +{
free(seg->trbs);
seg->trbs = NULL;
free(seg);
Do you want to make seg = NULL as well?
Why not ? No one will reference this "seg" after this, right ?
+}
+/**
- frees the "ring" pointer passed
- @param ptr pointer to "ring" to be freed
- @return none
- */
+static void xhci_ring_free(struct xhci_ring *ring) +{
struct xhci_segment *seg;
struct xhci_segment *first_seg;
BUG_ON(!ring);
first_seg = ring->first_seg;
seg = first_seg->next;
while (seg != first_seg) {
struct xhci_segment *next = seg->next;
xhci_segment_free(seg);
seg = next;
}
xhci_segment_free(first_seg);
free(ring);
Do you want to make ring = NULL?
Ditto, this ring pointer also not required further by anyone, isn't it ?
+}
+/**
- frees the "xhci_container_ctx" pointer passed
- @param ptr pointer to "xhci_container_ctx" to be freed
- @return none
- */
+static void xhci_free_container_ctx(struct xhci_container_ctx *ctx) +{
free(ctx->bytes);
free(ctx);
+}
+/**
- frees the virtual devices for "xhci_ctrl" pointer passed
- @param ptr pointer to "xhci_ctrl" whose virtual devices are to be freed
- @return none
- */
+static void xhci_free_virt_devices(struct xhci_ctrl *ctrl) +{
int i;
int slot_id;
struct xhci_virt_device *virt_dev;
/*
* refactored here to loop through all virt_dev
* Slot ID 0 is reserved
*/
for (slot_id = 0; slot_id < MAX_HC_SLOTS; slot_id++) {
virt_dev = ctrl->devs[slot_id];
if (!virt_dev)
continue;
ctrl->dcbaa->dev_context_ptrs[slot_id] = 0;
for (i = 0; i < 31; ++i)
if (virt_dev->eps[i].ring)
xhci_ring_free(virt_dev->eps[i].ring);
if (virt_dev->in_ctx)
xhci_free_container_ctx(virt_dev->in_ctx);
if (virt_dev->out_ctx)
xhci_free_container_ctx(virt_dev->out_ctx);
free(virt_dev);
/* make sure we are pointing to NULL */
ctrl->devs[slot_id] = NULL;
}
+}
+/**
- frees all the memory allocated
- @param ptr pointer to "xhci_ctrl" to be cleaned up
- @return none
- */
+void xhci_cleanup(struct xhci_ctrl *ctrl) +{
xhci_ring_free(ctrl->event_ring);
xhci_ring_free(ctrl->cmd_ring);
xhci_free_virt_devices(ctrl);
free(ctrl->erst.entries);
free(ctrl->dcbaa);
memset(ctrl, '\0', sizeof(struct xhci_ctrl));
+}
+/**
- Malloc the aligned memory
- @param size size of memory to be allocated
- @return allocates the memory and returns the aligned pointer
- */
+static void *xhci_malloc(unsigned int size) +{
void *ptr;
size_t cacheline_size = max(XHCI_ALIGNMENT, CACHELINE_SIZE);
ptr = memalign(cacheline_size, ALIGN(size, cacheline_size));
BUG_ON(!ptr);
memset(ptr, '\0', size);
xhci_flush_cache((uint32_t)ptr, size);
return ptr;
+}
+/**
- Make the prev segment point to the next segment.
- Change the last TRB in the prev segment to be a Link TRB which points to the
- address of the next segment. The caller needs to set any Link TRB
- related flags, such as End TRB, Toggle Cycle, and no snoop.
- @param prev pointer to the previous segment
- @param next pointer to the next segment
- @param link_trbs flag to indicate whether to link the trbs or NOT
- @return none
- */
+static void xhci_link_segments(struct xhci_segment *prev,
struct xhci_segment *next, bool link_trbs)
+{
u32 val;
u64 val_64 = 0;
if (!prev || !next)
return;
prev->next = next;
if (link_trbs) {
val_64 = (uintptr_t)next->trbs;
prev->trbs[TRBS_PER_SEGMENT-1].link.segment_ptr = val_64;
/*
* Set the last TRB in the segment to
* have a TRB type ID of Link TRB
*/
val = le32_to_cpu(prev->trbs[TRBS_PER_SEGMENT-1].link.control);
val &= ~TRB_TYPE_BITMASK;
val |= (TRB_LINK << TRB_TYPE_SHIFT);
prev->trbs[TRBS_PER_SEGMENT-1].link.control = cpu_to_le32(val);
}
+}
+/**
- Initialises the Ring's enqueue,dequeue,enq_seg pointers
- @param ring pointer to the RING to be intialised
- @return none
- */
+static void xhci_initialize_ring_info(struct xhci_ring *ring) +{
/*
* The ring is empty, so the enqueue pointer == dequeue pointer
*/
ring->enqueue = ring->first_seg->trbs;
ring->enq_seg = ring->first_seg;
ring->dequeue = ring->enqueue;
ring->deq_seg = ring->first_seg;
/*
* The ring is initialized to 0. The producer must write 1 to the
* cycle bit to handover ownership of the TRB, so PCS = 1.
* The consumer must compare CCS to the cycle bit to
* check ownership, so CCS = 1.
*/
ring->cycle_state = 1;
+}
+/**
- Allocates a generic ring segment from the ring pool, sets the dma address,
- initializes the segment to zero, and sets the private next pointer to NULL.
- Section 4.11.1.1:
- "All components of all Command and Transfer TRBs shall be initialized to '0'"
- @param none
- @return pointer to the newly allocated SEGMENT
- */
+static struct xhci_segment *xhci_segment_alloc(void) +{
struct xhci_segment *seg;
seg = (struct xhci_segment *)malloc(sizeof(struct xhci_segment));
BUG_ON(!seg);
seg->trbs = (union xhci_trb *)xhci_malloc(SEGMENT_SIZE);
seg->next = NULL;
return seg;
+}
+/**
- Create a new ring with zero or more segments.
- TODO: current code only uses one-time-allocated single-segment rings
- of 1KB anyway, so we might as well get rid of all the segment and
- linking code (and maybe increase the size a bit, e.g. 4KB).
- Link each segment together into a ring.
- Set the end flag and the cycle toggle bit on the last segment.
- See section 4.9.2 and figures 15 and 16 of XHCI spec rev1.0.
- @param num_segs number of segments in the ring
- @param link_trbs flag to indicate whether to link the trbs or NOT
- @return pointer to the newly created RING
- */
+struct xhci_ring *xhci_ring_alloc(unsigned int num_segs, bool link_trbs) +{
struct xhci_ring *ring;
struct xhci_segment *prev;
ring = (struct xhci_ring *)malloc(sizeof(struct xhci_ring));
BUG_ON(!ring);
if (num_segs == 0)
return ring;
ring->first_seg = xhci_segment_alloc();
BUG_ON(!ring->first_seg);
num_segs--;
prev = ring->first_seg;
while (num_segs > 0) {
struct xhci_segment *next;
next = xhci_segment_alloc();
BUG_ON(!next);
xhci_link_segments(prev, next, link_trbs);
prev = next;
num_segs--;
}
xhci_link_segments(prev, ring->first_seg, link_trbs);
if (link_trbs) {
/* See section 4.9.2.1 and 6.4.4.1 */
prev->trbs[TRBS_PER_SEGMENT-1].link.control |=
cpu_to_le32(LINK_TOGGLE);
}
xhci_initialize_ring_info(ring);
return ring;
+}
+/**
- Allocates the Container context
- @param ctrl Host controller data structure
- @param type type of XHCI Container Context
- @return NULL if failed else pointer to the context on success
- */
+static struct xhci_container_ctx
*xhci_alloc_container_ctx(struct xhci_ctrl *ctrl, int type)
+{
struct xhci_container_ctx *ctx;
ctx = (struct xhci_container_ctx *)
malloc(sizeof(struct xhci_container_ctx));
BUG_ON(!ctx);
BUG_ON((type != XHCI_CTX_TYPE_DEVICE) && (type != XHCI_CTX_TYPE_INPUT));
ctx->type = type;
ctx->size = (MAX_EP_CTX_NUM + 1) *
CTX_SIZE(readl(&ctrl->hccr->cr_hccparams));
if (type == XHCI_CTX_TYPE_INPUT)
ctx->size += CTX_SIZE(readl(&ctrl->hccr->cr_hccparams));
ctx->bytes = (u8 *)xhci_malloc(ctx->size);
return ctx;
+}
+/**
- Allocating virtual device
- @param udev pointer to USB deivce structure
- @return 0 on success else -1 on failure
- */
+int xhci_alloc_virt_device(struct usb_device *udev) +{
u64 byte_64 = 0;
unsigned int slot_id = udev->slot_id;
struct xhci_virt_device *virt_dev;
struct xhci_ctrl *ctrl = udev->controller;
/* Slot ID 0 is reserved */
if (ctrl->devs[slot_id]) {
printf("Virt dev for slot[%d] already allocated\n", slot_id);
return -1;
Global comment if you are including errno can't we make the errors returned mean something other then -1? Or I fail to see why errno was included.
Sure, will fix this globally.
}
ctrl->devs[slot_id] = (struct xhci_virt_device *)
malloc(sizeof(struct xhci_virt_device));
if (!ctrl->devs[slot_id]) {
printf("Failed to allocate virtual device\n");
Should use puts when no arguments.
Ok.
return -1;
}
memset(ctrl->devs[slot_id], 0, sizeof(struct xhci_virt_device));
virt_dev = ctrl->devs[slot_id];
/* Allocate the (output) device context that will be used in the HC. */
virt_dev->out_ctx = xhci_alloc_container_ctx(ctrl,
XHCI_CTX_TYPE_DEVICE);
if (!virt_dev->out_ctx) {
printf("Failed to allocate out context for virt dev\n");
Should use puts when no arguments.
ditto
return -1;
}
/* Allocate the (input) device context for address device command */
virt_dev->in_ctx = xhci_alloc_container_ctx(ctrl,
XHCI_CTX_TYPE_INPUT);
if (!virt_dev->in_ctx) {
printf("Failed to allocate in context for virt dev\n");
Should use puts when no arguments.
ditto
return -1;
}
/* Allocate endpoint 0 ring */
virt_dev->eps[0].ring = xhci_ring_alloc(1, true);
byte_64 = (uintptr_t)(virt_dev->out_ctx->bytes);
/* Point to output device context in dcbaa. */
ctrl->dcbaa->dev_context_ptrs[slot_id] = byte_64;
xhci_flush_cache((uint32_t)&ctrl->dcbaa->dev_context_ptrs[slot_id],
sizeof(__le64));
return 0;
+}
+/**
- Allocates the necessary data structures
- for XHCI host controller
- @param ctrl Host controller data structure
- @param hccr pointer to HOST Controller Control Registers
- @param hcor pointer to HOST Controller Operational Registers
- @return 0 if successful else -1 on failure
- */
+int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
struct xhci_hcor *hcor)
+{
uint64_t val_64;
uint64_t trb_64;
uint32_t val;
unsigned long deq;
int i;
struct xhci_segment *seg;
/* DCBAA initialization */
ctrl->dcbaa = (struct xhci_device_context_array *)
xhci_malloc(sizeof(struct xhci_device_context_array));
if (ctrl->dcbaa == NULL) {
printf("unable to allocate DCBA\n");
return -1;
}
val_64 = (uintptr_t)ctrl->dcbaa;
/* Set the pointer in DCBAA register */
xhci_writeq(&hcor->or_dcbaap, val_64);
/* Command ring control pointer register initialization */
ctrl->cmd_ring = xhci_ring_alloc(1, true);
/* Set the address in the Command Ring Control register */
trb_64 = (uintptr_t)ctrl->cmd_ring->first_seg->trbs;
val_64 = xhci_readq(&hcor->or_crcr);
val_64 = (val_64 & (u64) CMD_RING_RSVD_BITS) |
(trb_64 & (u64) ~CMD_RING_RSVD_BITS) |
ctrl->cmd_ring->cycle_state;
xhci_writeq(&hcor->or_crcr, val_64);
/* write the address of db register */
val = xhci_readl(&hccr->cr_dboff);
val &= DBOFF_MASK;
ctrl->dba = (struct xhci_doorbell_array *)((char *)hccr + val);
/* write the address of runtime register */
val = xhci_readl(&hccr->cr_rtsoff);
val &= RTSOFF_MASK;
ctrl->run_regs = (struct xhci_run_regs *)((char *)hccr + val);
/* writting the address of ir_set structure */
ctrl->ir_set = &ctrl->run_regs->ir_set[0];
/* Event ring does not maintain link TRB */
ctrl->event_ring = xhci_ring_alloc(ERST_NUM_SEGS, false);
ctrl->erst.entries = (struct xhci_erst_entry *)
xhci_malloc(sizeof(struct xhci_erst_entry) * ERST_NUM_SEGS);
ctrl->erst.num_entries = ERST_NUM_SEGS;
for (val = 0, seg = ctrl->event_ring->first_seg;
val < ERST_NUM_SEGS;
val++) {
trb_64 = 0;
trb_64 = (uintptr_t)seg->trbs;
struct xhci_erst_entry *entry = &ctrl->erst.entries[val];
xhci_writeq(&entry->seg_addr, trb_64);
entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT);
entry->rsvd = 0;
seg = seg->next;
}
xhci_flush_cache((uint32_t)ctrl->erst.entries,
ERST_NUM_SEGS * sizeof(struct xhci_erst_entry));
deq = (unsigned long)ctrl->event_ring->dequeue;
/* Update HC event ring dequeue pointer */
xhci_writeq(&ctrl->ir_set->erst_dequeue,
(u64)deq & (u64)~ERST_PTR_MASK);
/* set ERST count with the number of entries in the segment table */
val = xhci_readl(&ctrl->ir_set->erst_size);
val &= ERST_SIZE_MASK;
val |= ERST_NUM_SEGS;
xhci_writel(&ctrl->ir_set->erst_size, val);
/* this is the event ring segment table pointer */
val_64 = xhci_readq(&ctrl->ir_set->erst_base);
val_64 &= ERST_PTR_MASK;
val_64 |= ((u32)(ctrl->erst.entries) & ~ERST_PTR_MASK);
xhci_writeq(&ctrl->ir_set->erst_base, val_64);
/* initializing the virtual devices to NULL */
for (i = 0; i < MAX_HC_SLOTS; ++i)
ctrl->devs[i] = NULL;
/*
* Just Zero'ing this register completely,
* or some spurious Device Notification Events
* might screw things here.
*/
xhci_writel(&hcor->or_dnctrl, 0x0);
return 0;
+}
+/**
- Give the input control context for the passed container context
- @param ctx pointer to the context
- @return pointer to the Input control context data
- */
+struct xhci_input_control_ctx
*xhci_get_input_control_ctx(struct xhci_container_ctx *ctx)
+{
BUG_ON(ctx->type != XHCI_CTX_TYPE_INPUT);
return (struct xhci_input_control_ctx *)ctx->bytes;
+}
+/**
- Give the slot context for the passed container context
- @param ctrl Host controller data structure
- @param ctx pointer to the context
- @return pointer to the slot control context data
- */
+struct xhci_slot_ctx *xhci_get_slot_ctx(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *ctx)
+{
if (ctx->type == XHCI_CTX_TYPE_DEVICE)
return (struct xhci_slot_ctx *)ctx->bytes;
return (struct xhci_slot_ctx *)
(ctx->bytes + CTX_SIZE(readl(&ctrl->hccr->cr_hccparams)));
+}
+/**
- Gets the EP context from based on the ep_index
- @param ctrl Host controller data structure
- @param ctx context container
- @param ep_index index of the endpoint
- @return pointer to the End point context
- */
+struct xhci_ep_ctx *xhci_get_ep_ctx(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *ctx,
unsigned int ep_index)
+{
/* increment ep index by offset of start of ep ctx array */
ep_index++;
if (ctx->type == XHCI_CTX_TYPE_INPUT)
ep_index++;
return (struct xhci_ep_ctx *)
(ctx->bytes +
(ep_index * CTX_SIZE(readl(&ctrl->hccr->cr_hccparams))));
+}
+/**
- Copy output xhci_ep_ctx to the input xhci_ep_ctx copy.
- Useful when you want to change one particular aspect of the endpoint
- and then issue a configure endpoint command.
- @param ctrl Host controller data structure
- @param in_ctx contains the input context
- @param out_ctx contains the input context
- @param ep_index index of the end point
- @return none
- */
+void xhci_endpoint_copy(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *in_ctx,
struct xhci_container_ctx *out_ctx,
unsigned int ep_index)
+{
struct xhci_ep_ctx *out_ep_ctx;
struct xhci_ep_ctx *in_ep_ctx;
out_ep_ctx = xhci_get_ep_ctx(ctrl, out_ctx, ep_index);
in_ep_ctx = xhci_get_ep_ctx(ctrl, in_ctx, ep_index);
in_ep_ctx->ep_info = out_ep_ctx->ep_info;
in_ep_ctx->ep_info2 = out_ep_ctx->ep_info2;
in_ep_ctx->deq = out_ep_ctx->deq;
in_ep_ctx->tx_info = out_ep_ctx->tx_info;
+}
+/**
- Copy output xhci_slot_ctx to the input xhci_slot_ctx.
- Useful when you want to change one particular aspect of the endpoint
- and then issue a configure endpoint command.
- Only the context entries field matters, but
- we'll copy the whole thing anyway.
- @param ctrl Host controller data structure
- @param in_ctx contains the inpout context
- @param out_ctx contains the inpout context
- @return none
- */
+void xhci_slot_copy(struct xhci_ctrl *ctrl, struct xhci_container_ctx *in_ctx,
struct xhci_container_ctx *out_ctx)
+{
struct xhci_slot_ctx *in_slot_ctx;
struct xhci_slot_ctx *out_slot_ctx;
in_slot_ctx = xhci_get_slot_ctx(ctrl, in_ctx);
out_slot_ctx = xhci_get_slot_ctx(ctrl, out_ctx);
in_slot_ctx->dev_info = out_slot_ctx->dev_info;
in_slot_ctx->dev_info2 = out_slot_ctx->dev_info2;
in_slot_ctx->tt_info = out_slot_ctx->tt_info;
in_slot_ctx->dev_state = out_slot_ctx->dev_state;
+}
+/**
- Setup an xHCI virtual device for a Set Address command
- @param udev pointer to the Device Data Structure
- @return returns negative value on failure else 0 on success
- */
+void xhci_setup_addressable_virt_dev(struct usb_device *udev) +{
struct usb_device *hop = udev;
struct xhci_virt_device *virt_dev;
struct xhci_ep_ctx *ep0_ctx;
struct xhci_slot_ctx *slot_ctx;
u32 port_num = 0;
u64 trb_64 = 0;
struct xhci_ctrl *ctrl = udev->controller;
virt_dev = ctrl->devs[udev->slot_id];
BUG_ON(!virt_dev);
/* Extract the EP0 and Slot Ctrl */
ep0_ctx = xhci_get_ep_ctx(ctrl, virt_dev->in_ctx, 0);
slot_ctx = xhci_get_slot_ctx(ctrl, virt_dev->in_ctx);
/* Only the control endpoint is valid - one endpoint context */
slot_ctx->dev_info |= cpu_to_le32(LAST_CTX(1) | 0);
switch (udev->speed) {
case USB_SPEED_SUPER:
slot_ctx->dev_info |= cpu_to_le32(SLOT_SPEED_SS);
break;
case USB_SPEED_HIGH:
slot_ctx->dev_info |= cpu_to_le32(SLOT_SPEED_HS);
break;
case USB_SPEED_FULL:
slot_ctx->dev_info |= cpu_to_le32(SLOT_SPEED_FS);
break;
case USB_SPEED_LOW:
slot_ctx->dev_info |= cpu_to_le32(SLOT_SPEED_LS);
break;
default:
/* Speed was set earlier, this shouldn't happen. */
BUG();
}
/* Extract the root hub port number */
if (hop->parent)
while (hop->parent->parent)
hop = hop->parent;
port_num = hop->portnr;
debug("port_num = %d\n", port_num);
slot_ctx->dev_info2 |=
cpu_to_le32(((port_num & ROOT_HUB_PORT_MASK) <<
ROOT_HUB_PORT_SHIFT));
/* Step 4 - ring already allocated */
/* Step 5 */
ep0_ctx->ep_info2 = cpu_to_le32(CTRL_EP << EP_TYPE_SHIFT);
debug("SPEED = %d\n", udev->speed);
switch (udev->speed) {
case USB_SPEED_SUPER:
ep0_ctx->ep_info2 |= cpu_to_le32(((512 & MAX_PACKET_MASK) <<
MAX_PACKET_SHIFT));
debug("Setting Packet size = 512bytes\n");
break;
case USB_SPEED_HIGH:
/* USB core guesses at a 64-byte max packet first for FS devices */
case USB_SPEED_FULL:
ep0_ctx->ep_info2 |= cpu_to_le32(((64 & MAX_PACKET_MASK) <<
MAX_PACKET_SHIFT));
debug("Setting Packet size = 64bytes\n");
break;
case USB_SPEED_LOW:
ep0_ctx->ep_info2 |= cpu_to_le32(((8 & MAX_PACKET_MASK) <<
MAX_PACKET_SHIFT));
debug("Setting Packet size = 8bytes\n");
break;
default:
/* New speed? */
BUG();
}
/* EP 0 can handle "burst" sizes of 1, so Max Burst Size field is 0 */
ep0_ctx->ep_info2 |=
cpu_to_le32(((0 & MAX_BURST_MASK) << MAX_BURST_SHIFT) |
((3 & ERROR_COUNT_MASK) << ERROR_COUNT_SHIFT));
trb_64 = (uintptr_t)virt_dev->eps[0].ring->first_seg->trbs;
ep0_ctx->deq = cpu_to_le64(trb_64 | virt_dev->eps[0].ring->cycle_state);
/* Steps 7 and 8 were done in xhci_alloc_virt_device() */
xhci_flush_cache((uint32_t)ep0_ctx, sizeof(struct xhci_ep_ctx));
xhci_flush_cache((uint32_t)slot_ctx, sizeof(struct xhci_slot_ctx));
+} diff --git a/drivers/usb/host/xhci-ring.c b/drivers/usb/host/xhci-ring.c new file mode 100644 index 0000000..8340850 --- /dev/null +++ b/drivers/usb/host/xhci-ring.c @@ -0,0 +1,950 @@ +/*
- USB HOST XHCI Controller stack
- Copyright (C) 2013 Samsung Electronics Co.Ltd
- Vivek Gautam gautam.vivek@samsung.com
- Vikas Sajjan vikas.sajjan@samsung.com
- Based on xHCI host controller driver in linux-kernel
- by Sarah Sharp.
- This program is free software; you can redistribute it and/or
- modify it under the terms of the GNU General Public License as
- published by the Free Software Foundation; either version 2 of
- the License, or (at your option) any later version.
- This program is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- GNU General Public License for more details.
- You should have received a copy of the GNU General Public License
- along with this program; if not, write to the Free Software
- Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
- MA 02110-1301 USA
- */
+#include <common.h> +#include <asm/byteorder.h> +#include <usb.h> +#include <asm/io.h> +#include <asm/unaligned.h> +#include <asm-generic/errno.h>
+#include "xhci.h"
+/**
- Is this TRB a link TRB or was the last TRB the last TRB in this event ring
- segment? I.e. would the updated event TRB pointer step off the end of the
- event seg ?
- @param ctrl Host controller data structure
- @param ring pointer to the ring
- @param seg poniter to the segment to which TRB belongs
- @param trb poniter to the ring trb
- @return 1 if this TRB a link TRB else 0
- */
+static int last_trb(struct xhci_ctrl *ctrl, struct xhci_ring *ring,
struct xhci_segment *seg, union xhci_trb *trb)
+{
if (ring == ctrl->event_ring)
return trb == &seg->trbs[TRBS_PER_SEGMENT];
else
return TRB_TYPE_LINK_LE32(trb->link.control);
+}
+/**
- Does this link TRB point to the first segment in a ring,
- or was the previous TRB the last TRB on the last segment in the ERST?
- @param ctrl Host controller data structure
- @param ring pointer to the ring
- @param seg poniter to the segment to which TRB belongs
- @param trb poniter to the ring trb
- @return 1 if this TRB is the last TRB on the last segment else 0
- */
+static bool last_trb_on_last_seg(struct xhci_ctrl *ctrl,
struct xhci_ring *ring,
struct xhci_segment *seg,
union xhci_trb *trb)
+{
if (ring == ctrl->event_ring)
return ((trb == &seg->trbs[TRBS_PER_SEGMENT]) &&
(seg->next == ring->first_seg));
else
return le32_to_cpu(trb->link.control) & LINK_TOGGLE;
+}
+/**
- See Cycle bit rules. SW is the consumer for the event ring only.
- Don't make a ring full of link TRBs. That would be dumb and this would loop.
- If we've just enqueued a TRB that is in the middle of a TD (meaning the
- chain bit is set), then set the chain bit in all the following link TRBs.
- If we've enqueued the last TRB in a TD, make sure the following link TRBs
- have their chain bit cleared (so that each Link TRB is a separate TD).
- Section 6.4.4.1 of the 0.95 spec says link TRBs cannot have the chain bit
- set, but other sections talk about dealing with the chain bit set. This was
- fixed in the 0.96 specification errata, but we have to assume that all 0.95
- xHCI hardware can't handle the chain bit being cleared on a link TRB.
- @param ctrl Host controller data structure
- @param ring pointer to the ring
- @param more_trbs_coming flag to indicate whether more trbs
are expected or NOT.
Will you enqueue more TRBs before calling
prepare_ring()?
- @return none
- */
+static void inc_enq(struct xhci_ctrl *ctrl, struct xhci_ring *ring,
bool more_trbs_coming)
+{
u32 chain;
union xhci_trb *next;
chain = le32_to_cpu(ring->enqueue->generic.field[3]) & TRB_CHAIN;
next = ++(ring->enqueue);
/*
* Update the dequeue pointer further if that was a link TRB or we're at
* the end of an event ring segment (which doesn't have link TRBS)
*/
while (last_trb(ctrl, ring, ring->enq_seg, next)) {
if (ring != ctrl->event_ring) {
/*
* If the caller doesn't plan on enqueueing more
* TDs before ringing the doorbell, then we
* don't want to give the link TRB to the
* hardware just yet. We'll give the link TRB
* back in prepare_ring() just before we enqueue
* the TD at the top of the ring.
*/
if (!chain && !more_trbs_coming)
break;
/*
* If we're not dealing with 0.95 hardware or
* isoc rings on AMD 0.96 host,
* carry over the chain bit of the previous TRB
* (which may mean the chain bit is cleared).
*/
next->link.control &= cpu_to_le32(~TRB_CHAIN);
next->link.control |= cpu_to_le32(chain);
next->link.control ^= cpu_to_le32(TRB_CYCLE);
xhci_flush_cache((uint32_t)next,
sizeof(union xhci_trb));
}
/* Toggle the cycle bit after the last ring segment. */
if (last_trb_on_last_seg(ctrl, ring,
ring->enq_seg, next))
ring->cycle_state = (ring->cycle_state ? 0 : 1);
ring->enq_seg = ring->enq_seg->next;
ring->enqueue = ring->enq_seg->trbs;
next = ring->enqueue;
}
+}
+/**
- See Cycle bit rules. SW is the consumer for the event ring only.
- Don't make a ring full of link TRBs. That would be dumb and this would loop.
- @param ctrl Host controller data structure
- @param ring Ring whose Dequeue TRB pointer needs to be incremented.
- return none
- */
+static void inc_deq(struct xhci_ctrl *ctrl, struct xhci_ring *ring) +{
do {
/*
* Update the dequeue pointer further if that was a link TRB or
* we're at the end of an event ring segment (which doesn't have
* link TRBS)
*/
if (last_trb(ctrl, ring, ring->deq_seg, ring->dequeue)) {
if (ring == ctrl->event_ring &&
last_trb_on_last_seg(ctrl, ring,
ring->deq_seg, ring->dequeue)) {
ring->cycle_state = (ring->cycle_state ? 0 : 1);
}
ring->deq_seg = ring->deq_seg->next;
ring->dequeue = ring->deq_seg->trbs;
} else {
ring->dequeue++;
}
} while (last_trb(ctrl, ring, ring->deq_seg, ring->dequeue));
+}
+/**
- Generic function for queueing a TRB on a ring.
- The caller must have checked to make sure there's room on the ring.
- @param more_trbs_coming: Will you enqueue more TRBs before calling
prepare_ring()?
- @param ctrl Host controller data structure
- @param ring pointer to the ring
- @param more_trbs_coming flag to indicate whether more trbs
- @param trb_fields pointer to trb field array containing TRB contents
- @return pointer to the enqueued trb
- */
+static struct xhci_generic_trb *queue_trb(struct xhci_ctrl *ctrl,
struct xhci_ring *ring,
bool more_trbs_coming,
unsigned int *trb_fields)
+{
struct xhci_generic_trb *trb;
int i;
trb = &ring->enqueue->generic;
for (i = 0; i < 4; i++)
trb->field[i] = cpu_to_le32(trb_fields[i]);
xhci_flush_cache((uint32_t)trb, sizeof(struct xhci_generic_trb));
inc_enq(ctrl, ring, more_trbs_coming);
return trb;
+}
+/**
- Does various checks on the endpoint ring, and makes it ready
- to queue num_trbs.
- @param ctrl Host controller data structure
- @param ep_ring pointer to the EP Transfer Ring
- @param ep_state State of the End Point
- @return error code in case of invalid ep_state, 0 on success
- */
+static int prepare_ring(struct xhci_ctrl *ctrl, struct xhci_ring *ep_ring,
u32 ep_state)
+{
union xhci_trb *next = ep_ring->enqueue;
/* Make sure the endpoint has been added to xHC schedule */
switch (ep_state) {
case EP_STATE_DISABLED:
/*
* USB core changed config/interfaces without notifying us,
* or hardware is reporting the wrong state.
*/
printf("WARN urb submitted to disabled ep\n");
return -ENOENT;
case EP_STATE_ERROR:
printf("WARN waiting for error on ep to be cleared\n");
return -EINVAL;
case EP_STATE_HALTED:
printf("WARN halted endpoint, queueing URB anyway.\n");
case EP_STATE_STOPPED:
case EP_STATE_RUNNING:
debug("EP STATE RUNNING.\n");
break;
default:
printf("ERROR unknown endpoint state for ep\n");
return -EINVAL;
}
while (last_trb(ctrl, ep_ring, ep_ring->enq_seg, next)) {
/*
* If we're not dealing with 0.95 hardware or isoc rings
* on AMD 0.96 host, clear the chain bit.
*/
next->link.control &= cpu_to_le32(~TRB_CHAIN);
next->link.control ^= cpu_to_le32(TRB_CYCLE);
xhci_flush_cache((uint32_t)next, sizeof(union xhci_trb));
/* Toggle the cycle bit after the last ring segment. */
if (last_trb_on_last_seg(ctrl, ep_ring,
ep_ring->enq_seg, next))
ep_ring->cycle_state = (ep_ring->cycle_state ? 0 : 1);
ep_ring->enq_seg = ep_ring->enq_seg->next;
ep_ring->enqueue = ep_ring->enq_seg->trbs;
next = ep_ring->enqueue;
}
return 0;
+}
+/**
- Generic function for queueing a command TRB on the command ring.
- Check to make sure there's room on the command ring for one command TRB.
- @param ctrl Host controller data structure
- @param ptr Pointer address to write in the first two fields (opt.)
- @param slot_id Slot ID to encode in the flags field (opt.)
- @param ep_index Endpoint index to encode in the flags field (opt.)
- @param cmd Command type to enqueue
- @return none
- */
+void xhci_queue_command(struct xhci_ctrl *ctrl, u8 *ptr, u32 slot_id,
u32 ep_index, trb_type cmd)
+{
u32 fields[4];
u64 val_64 = (uintptr_t)ptr;
BUG_ON(prepare_ring(ctrl, ctrl->cmd_ring, EP_STATE_RUNNING));
fields[0] = lower_32_bits(val_64);
fields[1] = upper_32_bits(val_64);
fields[2] = 0;
fields[3] = TRB_TYPE(cmd) | EP_ID_FOR_TRB(ep_index) |
SLOT_ID_FOR_TRB(slot_id) | ctrl->cmd_ring->cycle_state;
queue_trb(ctrl, ctrl->cmd_ring, false, fields);
/* Ring the command ring doorbell */
xhci_writel(&ctrl->dba->doorbell[0], DB_VALUE_HOST);
+}
+/**
- The TD size is the number of bytes remaining in the TD (including this TRB),
- right shifted by 10.
- It must fit in bits 21:17, so it can't be bigger than 31.
- @param remainder remaining packets to be sent
- @return remainder if remainder is less than max else max
- */
+static u32 xhci_td_remainder(unsigned int remainder) +{
u32 max = (1 << (21 - 17 + 1)) - 1;
if ((remainder >> 10) >= max)
return max << 17;
else
return (remainder >> 10) << 17;
+}
+/**
- Finds out the remanining packets to be sent
- @param running_total total size sent so far
- @param trb_buff_len length of the TRB Buffer
- @param total_packet_count total packet count
- @param maxpacketsize max packet size of current pipe
- @param num_trbs_left number of TRBs left to be processed
- @return 0 if running_total or trb_buff_len is 0, else remainder
- */
+static u32 xhci_v1_0_td_remainder(int running_total,
int trb_buff_len,
unsigned int total_packet_count,
int maxpacketsize,
unsigned int num_trbs_left)
+{
int packets_transferred;
/* One TRB with a zero-length data packet. */
if (num_trbs_left == 0 || (running_total == 0 && trb_buff_len == 0))
return 0;
/*
* All the TRB queueing functions don't count the current TRB in
* running_total.
*/
packets_transferred = (running_total + trb_buff_len) / maxpacketsize;
if ((total_packet_count - packets_transferred) > 31)
return 31 << 17;
return (total_packet_count - packets_transferred) << 17;
+}
+/**
- Ring the doorbell of the End Point
- @param udev pointer to the USB device structure
- @param ep_index index of the endpoint
- @param start_cycle cycle flag of the first TRB
- @param start_trb pionter to the first TRB
- @return none
- */
+static void giveback_first_trb(struct usb_device *udev, int ep_index,
int start_cycle,
struct xhci_generic_trb *start_trb)
+{
struct xhci_ctrl *ctrl = udev->controller;
/*
* Pass all the TRBs to the hardware at once and make sure this write
* isn't reordered.
*/
if (start_cycle)
start_trb->field[3] |= cpu_to_le32(start_cycle);
else
start_trb->field[3] &= cpu_to_le32(~TRB_CYCLE);
xhci_flush_cache((uint32_t)start_trb, sizeof(struct xhci_generic_trb));
/* Ringing EP doorbell here */
xhci_writel(&ctrl->dba->doorbell[udev->slot_id],
DB_VALUE(ep_index, 0));
return;
+}
+/**** POLLING mechanism for XHCI ****/
+/**
- Finalizes a handled event TRB by advancing our dequeue pointer and giving
- the TRB back to the hardware for recycling. Must call this exactly once at
- the end of each event handler, and not touch the TRB again afterwards.
- @param ctrl Host controller data structure
- @return none
- */
+void xhci_acknowledge_event(struct xhci_ctrl *ctrl) +{
/* Advance our dequeue pointer to the next event */
inc_deq(ctrl, ctrl->event_ring);
/* Inform the hardware */
xhci_writeq(&ctrl->ir_set->erst_dequeue,
(uintptr_t)ctrl->event_ring->dequeue | ERST_EHB);
+}
+/**
- Checks if there is a new event to handle on the event ring.
- @param ctrl Host controller data structure
- @return 0 if failure else 1 on success
- */
+static int event_ready(struct xhci_ctrl *ctrl) +{
union xhci_trb *event;
xhci_inval_cache((uint32_t)ctrl->event_ring->dequeue,
sizeof(union xhci_trb));
event = ctrl->event_ring->dequeue;
/* Does the HC or OS own the TRB? */
if ((le32_to_cpu(event->event_cmd.flags) & TRB_CYCLE) !=
ctrl->event_ring->cycle_state)
return 0;
return 1;
+}
+/**
- Waits for a specific type of event and returns it. Discards unexpected
- events. Caller *must* call xhci_acknowledge_event() after it is finished
- processing the event, and must not access the returned pointer afterwards.
- @param ctrl Host controller data structure
- @param expected TRB type expected from Event TRB
- @return pointer to event trb
- */
+union xhci_trb *xhci_wait_for_event(struct xhci_ctrl *ctrl, trb_type expected) +{
trb_type type;
unsigned long ts = get_timer(0);
do {
union xhci_trb *event = ctrl->event_ring->dequeue;
if (!event_ready(ctrl))
continue;
type = TRB_FIELD_TO_TYPE(le32_to_cpu(event->event_cmd.flags));
if (type == expected)
return event;
if (type == TRB_PORT_STATUS)
/* TODO: remove this once enumeration has been reworked */
/*
* Port status change events always have a
* successful completion code
*/
BUG_ON(GET_COMP_CODE(
le32_to_cpu(event->generic.field[2])) !=
COMP_SUCCESS);
else
printf("Unexpected XHCI event TRB, skipping... "
"(%08x %08x %08x %08x)\n",
le32_to_cpu(event->generic.field[0]),
le32_to_cpu(event->generic.field[1]),
le32_to_cpu(event->generic.field[2]),
le32_to_cpu(event->generic.field[3]));
xhci_acknowledge_event(ctrl);
} while (get_timer(ts) < XHCI_TIMEOUT);
if (expected == TRB_TRANSFER)
return NULL;
printf("XHCI timeout on event type %d... cannot recover.\n", expected);
BUG();
+}
+/*
- Stops transfer processing for an endpoint and throws away all unprocessed
- TRBs by setting the xHC's dequeue pointer to our enqueue pointer. The next
- xhci_bulk_tx/xhci_ctrl_tx on this enpoint will add new transfers there and
- ring the doorbell, causing this endpoint to start working again.
- (Careful: This will BUG() when there was no transfer in progress. Shouldn't
- happen in practice for current uses and is too complicated to fix right now.)
- */
+static void abort_td(struct usb_device *udev, int ep_index) +{
struct xhci_ctrl *ctrl = udev->controller;
struct xhci_ring *ring = ctrl->devs[udev->slot_id]->eps[ep_index].ring;
union xhci_trb *event;
u32 field;
xhci_queue_command(ctrl, NULL, udev->slot_id, ep_index, TRB_STOP_RING);
event = xhci_wait_for_event(ctrl, TRB_TRANSFER);
field = le32_to_cpu(event->trans_event.flags);
BUG_ON(TRB_TO_SLOT_ID(field) != udev->slot_id);
BUG_ON(TRB_TO_EP_INDEX(field) != ep_index);
BUG_ON(GET_COMP_CODE(le32_to_cpu(event->trans_event.transfer_len
!= COMP_STOP)));
xhci_acknowledge_event(ctrl);
event = xhci_wait_for_event(ctrl, TRB_COMPLETION);
BUG_ON(TRB_TO_SLOT_ID(le32_to_cpu(event->event_cmd.flags))
!= udev->slot_id || GET_COMP_CODE(le32_to_cpu(
event->event_cmd.status)) != COMP_SUCCESS);
xhci_acknowledge_event(ctrl);
xhci_queue_command(ctrl, (void *)((uintptr_t)ring->enqueue |
ring->cycle_state), udev->slot_id, ep_index, TRB_SET_DEQ);
event = xhci_wait_for_event(ctrl, TRB_COMPLETION);
BUG_ON(TRB_TO_SLOT_ID(le32_to_cpu(event->event_cmd.flags))
!= udev->slot_id || GET_COMP_CODE(le32_to_cpu(
event->event_cmd.status)) != COMP_SUCCESS);
xhci_acknowledge_event(ctrl);
+}
+static void record_transfer_result(struct usb_device *udev,
union xhci_trb *event, int length)
+{
udev->act_len = min(length, length -
EVENT_TRB_LEN(le32_to_cpu(event->trans_event.transfer_len)));
switch (GET_COMP_CODE(le32_to_cpu(event->trans_event.transfer_len))) {
case COMP_SUCCESS:
BUG_ON(udev->act_len != length);
/* fallthrough */
case COMP_SHORT_TX:
udev->status = 0;
break;
case COMP_STALL:
udev->status = USB_ST_STALLED;
break;
case COMP_DB_ERR:
case COMP_TRB_ERR:
udev->status = USB_ST_BUF_ERR;
break;
case COMP_BABBLE:
udev->status = USB_ST_BABBLE_DET;
break;
default:
udev->status = 0x80; /* USB_ST_TOO_LAZY_TO_MAKE_A_NEW_MACRO */
}
+}
+/**** Bulk and Control transfer methods ****/ +/**
- Queues up the BULK Request
- @param udev pointer to the USB device structure
- @param pipe contains the DIR_IN or OUT , devnum
- @param length length of the buffer
- @param buffer buffer to be read/written based on the request
- @return returns 0 if successful else -1 on failure
- */
+int xhci_bulk_tx(struct usb_device *udev, unsigned long pipe,
int length, void *buffer)
+{
int num_trbs = 0;
struct xhci_generic_trb *start_trb;
bool first_trb = 0;
int start_cycle;
u32 field = 0;
u32 length_field = 0;
struct xhci_ctrl *ctrl = udev->controller;
int slot_id = udev->slot_id;
int ep_index;
struct xhci_virt_device *virt_dev;
struct xhci_ep_ctx *ep_ctx;
struct xhci_ring *ring; /* EP transfer ring */
union xhci_trb *event;
int running_total, trb_buff_len;
unsigned int total_packet_count;
int maxpacketsize;
u64 addr;
int ret;
u32 trb_fields[4];
u64 val_64 = (uintptr_t)buffer;
debug("dev=%p, pipe=%lx, buffer=%p, length=%d\n",
udev, pipe, buffer, length);
ep_index = usb_pipe_ep_index(pipe);
virt_dev = ctrl->devs[slot_id];
xhci_inval_cache((uint32_t)virt_dev->out_ctx->bytes,
virt_dev->out_ctx->size);
ep_ctx = xhci_get_ep_ctx(ctrl, virt_dev->out_ctx, ep_index);
ring = virt_dev->eps[ep_index].ring;
/*
* How much data is (potentially) left before the 64KB boundary?
* XHCI Spec puts restriction( TABLE 49 and 6.4.1 section of XHCI Spec)
* that the buffer should not span 64KB boundary. if so
* we send request in more than 1 TRB by chaining them.
*/
running_total = TRB_MAX_BUFF_SIZE -
(lower_32_bits(val_64) & (TRB_MAX_BUFF_SIZE - 1));
trb_buff_len = running_total;
running_total &= TRB_MAX_BUFF_SIZE - 1;
/*
* If there's some data on this 64KB chunk, or we have to send a
* zero-length transfer, we need at least one TRB
*/
if (running_total != 0 || length == 0)
num_trbs++;
/* How many more 64KB chunks to transfer, how many more TRBs? */
while (running_total < length) {
num_trbs++;
running_total += TRB_MAX_BUFF_SIZE;
}
/*
* XXX: Calling routine prepare_ring() called in place of
* prepare_trasfer() as there in 'Linux' since we are not
* maintaining multiple TDs/transfer at the same time.
*/
ret = prepare_ring(ctrl, ring,
le32_to_cpu(ep_ctx->ep_info) & EP_STATE_MASK);
if (ret < 0)
return ret;
/*
* Don't give the first TRB to the hardware (by toggling the cycle bit)
* until we've finished creating all the other TRBs. The ring's cycle
* state may change as we enqueue the other TRBs, so save it too.
*/
start_trb = &ring->enqueue->generic;
start_cycle = ring->cycle_state;
running_total = 0;
maxpacketsize = usb_maxpacket(udev, pipe);
total_packet_count = DIV_ROUND_UP(length, maxpacketsize);
/* How much data is in the first TRB? */
/*
* How much data is (potentially) left before the 64KB boundary?
* XHCI Spec puts restriction( TABLE 49 and 6.4.1 section of XHCI Spec)
* that the buffer should not span 64KB boundary. if so
* we send request in more than 1 TRB by chaining them.
*/
addr = val_64;
if (trb_buff_len > length)
trb_buff_len = length;
first_trb = true;
/* flush the buffer before use */
xhci_flush_cache((uint32_t)buffer, length);
/* Queue the first TRB, even if it's zero-length */
do {
u32 remainder = 0;
field = 0;
/* Don't change the cycle bit of the first TRB until later */
if (first_trb) {
first_trb = false;
if (start_cycle == 0)
field |= TRB_CYCLE;
} else {
field |= ring->cycle_state;
}
/*
* Chain all the TRBs together; clear the chain bit in the last
* TRB to indicate it's the last TRB in the chain.
*/
if (num_trbs > 1)
field |= TRB_CHAIN;
else
field |= TRB_IOC;
/* Only set interrupt on short packet for IN endpoints */
if (usb_pipein(pipe))
field |= TRB_ISP;
/* Set the TRB length, TD size, and interrupter fields. */
if (HC_VERSION(xhci_readl(&ctrl->hccr->cr_capbase)) < 0x100)
remainder = xhci_td_remainder(length - running_total);
else
remainder = xhci_v1_0_td_remainder(running_total,
trb_buff_len,
total_packet_count,
maxpacketsize,
num_trbs - 1);
length_field = ((trb_buff_len & TRB_LEN_MASK) |
remainder |
((0 & TRB_INTR_TARGET_MASK) <<
TRB_INTR_TARGET_SHIFT));
trb_fields[0] = lower_32_bits(addr);
trb_fields[1] = upper_32_bits(addr);
trb_fields[2] = length_field;
trb_fields[3] = field | (TRB_NORMAL << TRB_TYPE_SHIFT);
queue_trb(ctrl, ring, (num_trbs > 1), trb_fields);
--num_trbs;
running_total += trb_buff_len;
/* Calculate length for next transfer */
addr += trb_buff_len;
trb_buff_len = min((length - running_total), TRB_MAX_BUFF_SIZE);
} while (running_total < length);
giveback_first_trb(udev, ep_index, start_cycle, start_trb);
event = xhci_wait_for_event(ctrl, TRB_TRANSFER);
if (!event) {
debug("XHCI bulk transfer timed out, aborting...\n");
abort_td(udev, ep_index);
udev->status = USB_ST_NAK_REC; /* closest thing to a timeout */
udev->act_len = 0;
return -1;
}
field = le32_to_cpu(event->trans_event.flags);
BUG_ON(TRB_TO_SLOT_ID(field) != slot_id);
BUG_ON(TRB_TO_EP_INDEX(field) != ep_index);
BUG_ON(*(void **)(uintptr_t)le64_to_cpu(event->trans_event.buffer) -
buffer > (size_t)length);
record_transfer_result(udev, event, length);
xhci_acknowledge_event(ctrl);
xhci_inval_cache((uint32_t)buffer, length);
return (udev->status != USB_ST_NOT_PROC) ? 0 : -1;
+}
+/**
- Queues up the Control Transfer Request
- @param udev pointer to the USB device structure
- @param pipe contains the DIR_IN or OUT , devnum
- @param req request type
- @param length length of the buffer
- @param buffer buffer to be read/written based on the request
- @return returns 0 if successful else -1 on failure
- */
+int xhci_ctrl_tx(struct usb_device *udev, unsigned long pipe,
struct devrequest *req, int length,
void *buffer)
+{
int ret;
int start_cycle;
int num_trbs;
u32 field;
u32 length_field;
u64 buf_64 = 0;
struct xhci_generic_trb *start_trb;
struct xhci_ctrl *ctrl = udev->controller;
int slot_id = udev->slot_id;
int ep_index;
u32 trb_fields[4];
struct xhci_virt_device *virt_dev = ctrl->devs[slot_id];
struct xhci_ring *ep_ring;
union xhci_trb *event;
debug("req=%u (%#x), type=%u (%#x), value=%u (%#x), index=%u\n",
req->request, req->request,
req->requesttype, req->requesttype,
le16_to_cpu(req->value), le16_to_cpu(req->value),
le16_to_cpu(req->index));
ep_index = usb_pipe_ep_index(pipe);
ep_ring = virt_dev->eps[ep_index].ring;
/*
* Check to see if the max packet size for the default control
* endpoint changed during FS device enumeration
*/
if (udev->speed == USB_SPEED_FULL) {
ret = xhci_check_maxpacket(udev);
if (ret < 0)
return ret;
}
xhci_inval_cache((uint32_t)virt_dev->out_ctx->bytes,
virt_dev->out_ctx->size);
struct xhci_ep_ctx *ep_ctx = NULL;
ep_ctx = xhci_get_ep_ctx(ctrl, virt_dev->out_ctx, ep_index);
/* 1 TRB for setup, 1 for status */
num_trbs = 2;
/*
* Don't need to check if we need additional event data and normal TRBs,
* since data in control transfers will never get bigger than 16MB
* XXX: can we get a buffer that crosses 64KB boundaries?
*/
if (length > 0)
num_trbs++;
/*
* XXX: Calling routine prepare_ring() called in place of
* prepare_trasfer() as there in 'Linux' since we are not
* maintaining multiple TDs/transfer at the same time.
*/
ret = prepare_ring(ctrl, ep_ring,
le32_to_cpu(ep_ctx->ep_info) & EP_STATE_MASK);
if (ret < 0)
return ret;
/*
* Don't give the first TRB to the hardware (by toggling the cycle bit)
* until we've finished creating all the other TRBs. The ring's cycle
* state may change as we enqueue the other TRBs, so save it too.
*/
start_trb = &ep_ring->enqueue->generic;
start_cycle = ep_ring->cycle_state;
debug("start_trb %p, start_cycle %d\n", start_trb, start_cycle);
/* Queue setup TRB - see section 6.4.1.2.1 */
/* FIXME better way to translate setup_packet into two u32 fields? */
field = 0;
field |= TRB_IDT | (TRB_SETUP << TRB_TYPE_SHIFT);
if (start_cycle == 0)
field |= 0x1;
/* xHCI 1.0 6.4.1.2.1: Transfer Type field */
if (HC_VERSION(xhci_readl(&ctrl->hccr->cr_capbase)) == 0x100) {
if (length > 0) {
if (req->requesttype & USB_DIR_IN)
field |= (TRB_DATA_IN << TRB_TX_TYPE_SHIFT);
else
field |= (TRB_DATA_OUT << TRB_TX_TYPE_SHIFT);
}
}
debug("req->requesttype = %d, req->request = %d,"
"le16_to_cpu(req->value) = %d,"
"le16_to_cpu(req->index) = %d,"
"le16_to_cpu(req->length) = %d\n",
req->requesttype, req->request, le16_to_cpu(req->value),
le16_to_cpu(req->index), le16_to_cpu(req->length));
trb_fields[0] = req->requesttype | req->request << 8 |
le16_to_cpu(req->value) << 16;
trb_fields[1] = le16_to_cpu(req->index) |
le16_to_cpu(req->length) << 16;
/* TRB_LEN | (TRB_INTR_TARGET) */
trb_fields[2] = (8 | ((0 & TRB_INTR_TARGET_MASK) <<
TRB_INTR_TARGET_SHIFT));
/* Immediate data in pointer */
trb_fields[3] = field;
queue_trb(ctrl, ep_ring, true, trb_fields);
/* Re-initializing field to zero */
field = 0;
/* If there's data, queue data TRBs */
/* Only set interrupt on short packet for IN endpoints */
if (usb_pipein(pipe))
field = TRB_ISP | (TRB_DATA << TRB_TYPE_SHIFT);
else
field = (TRB_DATA << TRB_TYPE_SHIFT);
length_field = (length & TRB_LEN_MASK) | xhci_td_remainder(length) |
((0 & TRB_INTR_TARGET_MASK) << TRB_INTR_TARGET_SHIFT);
debug("length_field = %d, length = %d,"
"xhci_td_remainder(length) = %d , TRB_INTR_TARGET(0) = %d\n",
length_field, (length & TRB_LEN_MASK),
xhci_td_remainder(length), 0);
if (length > 0) {
if (req->requesttype & USB_DIR_IN)
field |= TRB_DIR_IN;
buf_64 = (uintptr_t)buffer;
trb_fields[0] = lower_32_bits(buf_64);
trb_fields[1] = upper_32_bits(buf_64);
trb_fields[2] = length_field;
trb_fields[3] = field | ep_ring->cycle_state;
xhci_flush_cache((uint32_t)buffer, length);
queue_trb(ctrl, ep_ring, true, trb_fields);
}
/*
* Queue status TRB -
* see Table 7 and sections 4.11.2.2 and 6.4.1.2.3
*/
/* If the device sent data, the status stage is an OUT transfer */
field = 0;
if (length > 0 && req->requesttype & USB_DIR_IN)
field = 0;
else
field = TRB_DIR_IN;
trb_fields[0] = 0;
trb_fields[1] = 0;
trb_fields[2] = ((0 & TRB_INTR_TARGET_MASK) << TRB_INTR_TARGET_SHIFT);
/* Event on completion */
trb_fields[3] = field | TRB_IOC |
(TRB_STATUS << TRB_TYPE_SHIFT) |
ep_ring->cycle_state;
queue_trb(ctrl, ep_ring, false, trb_fields);
giveback_first_trb(udev, ep_index, start_cycle, start_trb);
event = xhci_wait_for_event(ctrl, TRB_TRANSFER);
if (!event)
goto abort;
field = le32_to_cpu(event->trans_event.flags);
BUG_ON(TRB_TO_SLOT_ID(field) != slot_id);
BUG_ON(TRB_TO_EP_INDEX(field) != ep_index);
record_transfer_result(udev, event, length);
xhci_acknowledge_event(ctrl);
/* Invalidate buffer to make it available to usb-core */
if (length > 0)
xhci_inval_cache((uint32_t)buffer, length);
if (GET_COMP_CODE(le32_to_cpu(event->trans_event.transfer_len))
== COMP_SHORT_TX) {
/* Short data stage, clear up additional status stage event */
event = xhci_wait_for_event(ctrl, TRB_TRANSFER);
if (!event)
goto abort;
BUG_ON(TRB_TO_SLOT_ID(field) != slot_id);
BUG_ON(TRB_TO_EP_INDEX(field) != ep_index);
xhci_acknowledge_event(ctrl);
}
return (udev->status != USB_ST_NOT_PROC) ? 0 : -1;
+abort:
debug("XHCI control transfer timed out, aborting...\n");
abort_td(udev, ep_index);
udev->status = USB_ST_NAK_REC;
udev->act_len = 0;
return -1;
+} diff --git a/drivers/usb/host/xhci.c b/drivers/usb/host/xhci.c new file mode 100644 index 0000000..3e53e3d --- /dev/null +++ b/drivers/usb/host/xhci.c @@ -0,0 +1,1040 @@ +/*
- USB HOST XHCI Controller stack
- Copyright (C) 2013 Samsung Electronics Co.Ltd
- Vivek Gautam gautam.vivek@samsung.com
- Vikas Sajjan vikas.sajjan@samsung.com
- Based on xHCI host controller driver in linux-kernel
- by Sarah Sharp.
- This program is free software; you can redistribute it and/or
- modify it under the terms of the GNU General Public License as
- published by the Free Software Foundation; either version 2 of
- the License, or (at your option) any later version.
- This program is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- GNU General Public License for more details.
- You should have received a copy of the GNU General Public License
- along with this program; if not, write to the Free Software
- Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
- MA 02110-1301 USA
- */
+/**
- This file gives the xhci stack for usb3.0 looking into
- xhci specification Rev1.0 (5/21/10).
- The quirk devices support hasn't been given yet.
- */
Needs new SPDX license.
Sure
- SPDX-License-Identifier: GPL-2.0+
+#include <common.h> +#include <asm/byteorder.h> +#include <usb.h> +#include <asm/io.h>
Move this to xhci.h
Ok will move it.
+#include <malloc.h> +#include <watchdog.h> +#include <asm/cache.h> +#include <asm/unaligned.h> +#include <asm-generic/errno.h> +#include "xhci.h"
+#ifndef CONFIG_USB_MAX_CONTROLLER_COUNT +#define CONFIG_USB_MAX_CONTROLLER_COUNT 1 +#endif
+static struct descriptor {
struct usb_hub_descriptor hub;
struct usb_device_descriptor device;
struct usb_config_descriptor config;
struct usb_interface_descriptor interface;
struct usb_endpoint_descriptor endpoint;
struct usb_ss_ep_comp_descriptor ep_companion;
+} __attribute__ ((packed)) descriptor = {
{
0xc, /* bDescLength */
0x2a, /* bDescriptorType: hub descriptor */
2, /* bNrPorts -- runtime modified */
cpu_to_le16(0x8), /* wHubCharacteristics */
10, /* bPwrOn2PwrGood */
0, /* bHubCntrCurrent */
{}, /* Device removable */
{} /* at most 7 ports! XXX */
},
{
0x12, /* bLength */
1, /* bDescriptorType: UDESC_DEVICE */
cpu_to_le16(0x0300), /* bcdUSB: v3.0 */
9, /* bDeviceClass: UDCLASS_HUB */
0, /* bDeviceSubClass: UDSUBCLASS_HUB */
3, /* bDeviceProtocol: UDPROTO_SSHUBSTT */
9, /* bMaxPacketSize: 512 bytes 2^9 */
0x0000, /* idVendor */
0x0000, /* idProduct */
cpu_to_le16(0x0100), /* bcdDevice */
1, /* iManufacturer */
2, /* iProduct */
0, /* iSerialNumber */
1 /* bNumConfigurations: 1 */
},
{
0x9,
2, /* bDescriptorType: UDESC_CONFIG */
cpu_to_le16(0x1f), /* includes SS endpoint descriptor */
1, /* bNumInterface */
1, /* bConfigurationValue */
0, /* iConfiguration */
0x40, /* bmAttributes: UC_SELF_POWER */
0 /* bMaxPower */
},
{
0x9, /* bLength */
4, /* bDescriptorType: UDESC_INTERFACE */
0, /* bInterfaceNumber */
0, /* bAlternateSetting */
1, /* bNumEndpoints */
9, /* bInterfaceClass: UICLASS_HUB */
0, /* bInterfaceSubClass: UISUBCLASS_HUB */
0, /* bInterfaceProtocol: UIPROTO_HSHUBSTT */
0 /* iInterface */
},
{
0x7, /* bLength */
5, /* bDescriptorType: UDESC_ENDPOINT */
0x81, /* bEndpointAddress: IN endpoint 1 */
3, /* bmAttributes: UE_INTERRUPT */
8, /* wMaxPacketSize */
255 /* bInterval */
},
{
0x06, /* ss_bLength */
0x30, /* ss_bDescriptorType: SS EP Companion */
0x00, /* ss_bMaxBurst: allows 1 TX between ACKs */
/* ss_bmAttributes: 1 packet per service interval */
0x00,
/* ss_wBytesPerInterval: 15 bits for max 15 ports */
cpu_to_le16(0x02),
},
+};
+static struct xhci_ctrl xhcic[CONFIG_USB_MAX_CONTROLLER_COUNT];
+/**
- Waits for as per specified amount of time
- for the "result" to match with "done"
- @param ptr pointer to the register to be read
- @param mask mask for the value read
- @param done value to be campared with result
- @param usec time to wait till
- @return 0 if handshake is success else < 0 on failure
- */
+static int handshake(uint32_t volatile *ptr, uint32_t mask,
uint32_t done, int usec)
+{
uint32_t result;
do {
result = xhci_readl(ptr);
if (result == ~(uint32_t)0)
return -ENODEV;
result &= mask;
if (result == done)
return 0;
usec--;
udelay(1);
} while (usec > 0);
return -ETIMEDOUT;
+}
+/**
- Set the run bit and wait for the host to be running.
- @param hcor pointer to host controller operation registers
- @return status of the Handshake
- */
+static int xhci_start(struct xhci_hcor *hcor) +{
u32 temp;
int ret;
printf("Starting the controller\n");
temp = xhci_readl(&hcor->or_usbcmd);
temp |= (CMD_RUN);
xhci_writel(&hcor->or_usbcmd, temp);
/*
* Wait for the HCHalted Status bit to be 0 to indicate the host is
* running.
*/
ret = handshake(&hcor->or_usbsts, STS_HALT, 0, XHCI_MAX_HALT_USEC);
if (ret)
debug("Host took too long to start, "
"waited %u microseconds.\n",
XHCI_MAX_HALT_USEC);
return ret;
+}
+/**
- Resets the XHCI Controller
- @param hcor pointer to host controller operation registers
- @return -1 if XHCI Controller is halted else status of handshake
- */
+int xhci_reset(struct xhci_hcor *hcor) +{
u32 cmd;
u32 state;
int ret;
/* Halting the Host first */
debug("// Halt the HC\n");
state = xhci_readl(&hcor->or_usbsts) & STS_HALT;
if (!state) {
cmd = xhci_readl(&hcor->or_usbcmd);
cmd &= ~CMD_RUN;
xhci_writel(&hcor->or_usbcmd, cmd);
}
ret = handshake(&hcor->or_usbsts,
STS_HALT, STS_HALT, XHCI_MAX_HALT_USEC);
if (ret) {
printf("Host not halted after %u microseconds.\n",
XHCI_MAX_HALT_USEC);
return -1;
}
debug("// Reset the HC\n");
cmd = xhci_readl(&hcor->or_usbcmd);
cmd |= CMD_RESET;
xhci_writel(&hcor->or_usbcmd, cmd);
ret = handshake(&hcor->or_usbcmd, CMD_RESET, 0, XHCI_MAX_RESET_USEC);
if (ret)
return ret;
/*
* xHCI cannot write to any doorbells or operational registers other
* than status until the "Controller Not Ready" flag is cleared.
*/
return handshake(&hcor->or_usbsts, STS_CNR, 0, XHCI_MAX_RESET_USEC);
+}
+/**
- Used for passing endpoint bitmasks between the core and HCDs.
- Find the index for an endpoint given its descriptor.
- Use the return value to right shift 1 for the bitmask.
- Index = (epnum * 2) + direction - 1,
- where direction = 0 for OUT, 1 for IN.
- For control endpoints, the IN index is used (OUT index is unused), so
- index = (epnum * 2) + direction - 1 = (epnum * 2) + 1 - 1 = (epnum * 2)
- @param desc USB enpdoint Descriptor
- @return index of the Endpoint
- */
+static unsigned int xhci_get_ep_index(struct usb_endpoint_descriptor *desc) +{
unsigned int index;
if (usb_endpoint_xfer_control(desc)) {
index = (unsigned int)(usb_endpoint_num(desc) * 2);
} else {
index = (unsigned int)((usb_endpoint_num(desc) * 2) -
(usb_endpoint_dir_in(desc) ? 0 : 1));
}
No brackets around single line if statements
Oh, right, will fix this.
return index;
+}
+/**
- Issue a configure endpoint command or evaluate context command
- and wait for it to finish.
- @param udev pointer to the Device Data Structure
- @param ctx_change flag to indicate the Context has changed or NOT
- @return 0 on success, -1 on failure
- */
+static int xhci_configure_endpoints(struct usb_device *udev, bool ctx_change) +{
struct xhci_container_ctx *in_ctx;
struct xhci_virt_device *virt_dev;
struct xhci_ctrl *ctrl = udev->controller;
union xhci_trb *event;
virt_dev = ctrl->devs[udev->slot_id];
in_ctx = virt_dev->in_ctx;
xhci_flush_cache((uint32_t)in_ctx->bytes, in_ctx->size);
xhci_queue_command(ctrl, in_ctx->bytes, udev->slot_id, 0,
ctx_change ? TRB_EVAL_CONTEXT : TRB_CONFIG_EP);
event = xhci_wait_for_event(ctrl, TRB_COMPLETION);
BUG_ON(TRB_TO_SLOT_ID(le32_to_cpu(event->event_cmd.flags))
!= udev->slot_id);
switch (GET_COMP_CODE(le32_to_cpu(event->event_cmd.status))) {
case COMP_SUCCESS:
debug("Successful %s command\n",
ctx_change ? "Evaluate Context" : "Configure Endpoint");
break;
default:
printf("ERROR: %s command returned completion code %d.\n",
ctx_change ? "Evaluate Context" : "Configure Endpoint",
GET_COMP_CODE(le32_to_cpu(event->event_cmd.status)));
return -1;
}
xhci_acknowledge_event(ctrl);
return 0;
+}
+/**
- Configure the endpoint, programming the device contexts.
- @param udev pointer to the USB device structure
- @return returns the status of the xhci_configure_endpoints
- */
+static int xhci_set_configuration(struct usb_device *udev) +{
struct xhci_container_ctx *in_ctx;
struct xhci_container_ctx *out_ctx;
struct xhci_input_control_ctx *ctrl_ctx;
struct xhci_slot_ctx *slot_ctx;
struct xhci_ep_ctx *ep_ctx[MAX_EP_CTX_NUM];
int cur_ep;
int max_ep_flag = 0;
int ep_index;
unsigned int dir;
unsigned int ep_type;
struct xhci_ctrl *ctrl = udev->controller;
int num_of_ep;
int ep_flag = 0;
u64 trb_64 = 0;
int slot_id = udev->slot_id;
struct xhci_virt_device *virt_dev = ctrl->devs[slot_id];
struct usb_interface *ifdesc;
out_ctx = virt_dev->out_ctx;
in_ctx = virt_dev->in_ctx;
num_of_ep = udev->config.if_desc[0].no_of_ep;
ifdesc = &udev->config.if_desc[0];
ctrl_ctx = xhci_get_input_control_ctx(in_ctx);
/* Zero the input context control */
ctrl_ctx->add_flags = 0;
ctrl_ctx->drop_flags = 0;
/* EP_FLAG gives values 1 & 4 for EP1OUT and EP2IN */
for (cur_ep = 0; cur_ep < num_of_ep; cur_ep++) {
ep_flag = xhci_get_ep_index(&ifdesc->ep_desc[cur_ep]);
ctrl_ctx->add_flags |= cpu_to_le32(1 << (ep_flag + 1));
if (max_ep_flag < ep_flag)
max_ep_flag = ep_flag;
}
xhci_inval_cache((uint32_t)out_ctx->bytes, out_ctx->size);
/* slot context */
xhci_slot_copy(ctrl, in_ctx, out_ctx);
slot_ctx = xhci_get_slot_ctx(ctrl, in_ctx);
slot_ctx->dev_info &= ~(LAST_CTX_MASK);
slot_ctx->dev_info |= cpu_to_le32(LAST_CTX(max_ep_flag + 1) | 0);
xhci_endpoint_copy(ctrl, in_ctx, out_ctx, 0);
/* filling up ep contexts */
for (cur_ep = 0; cur_ep < num_of_ep; cur_ep++) {
struct usb_endpoint_descriptor *endpt_desc = NULL;
endpt_desc = &ifdesc->ep_desc[cur_ep];
trb_64 = 0;
ep_index = xhci_get_ep_index(endpt_desc);
ep_ctx[ep_index] = xhci_get_ep_ctx(ctrl, in_ctx, ep_index);
/* Allocate the ep rings */
virt_dev->eps[ep_index].ring = xhci_ring_alloc(1, true);
if (!virt_dev->eps[ep_index].ring)
return -1;
/*NOTE: ep_desc[0] actually represents EP1 and so on */
dir = (((endpt_desc->bEndpointAddress) & (0x80)) >> 7);
ep_type = (((endpt_desc->bmAttributes) & (0x3)) | (dir << 2));
ep_ctx[ep_index]->ep_info2 =
cpu_to_le32(ep_type << EP_TYPE_SHIFT);
ep_ctx[ep_index]->ep_info2 |=
cpu_to_le32(MAX_PACKET
(get_unaligned(&endpt_desc->wMaxPacketSize)));
ep_ctx[ep_index]->ep_info2 |=
cpu_to_le32(((0 & MAX_BURST_MASK) << MAX_BURST_SHIFT) |
((3 & ERROR_COUNT_MASK) << ERROR_COUNT_SHIFT));
trb_64 = (uintptr_t)
virt_dev->eps[ep_index].ring->enqueue;
ep_ctx[ep_index]->deq = cpu_to_le64(trb_64 |
virt_dev->eps[ep_index].ring->cycle_state);
}
return xhci_configure_endpoints(udev, false);
+}
+/**
- Issue an Address Device command (which will issue a SetAddress request to
- the device).
- @param udev pointer to the Device Data Structure
- @return 0 if successful else error code on failure
- */
+static int xhci_address_device(struct usb_device *udev) +{
int ret = 0;
struct xhci_ctrl *ctrl = udev->controller;
struct xhci_slot_ctx *slot_ctx;
struct xhci_input_control_ctx *ctrl_ctx;
struct xhci_virt_device *virt_dev;
int slot_id = udev->slot_id;
union xhci_trb *event;
virt_dev = ctrl->devs[slot_id];
/*
* This is the first Set Address since device plug-in
* so setting up the slot context.
*/
debug("Setting up addressable devices\n");
xhci_setup_addressable_virt_dev(udev);
ctrl_ctx = xhci_get_input_control_ctx(virt_dev->in_ctx);
ctrl_ctx->add_flags = cpu_to_le32(SLOT_FLAG | EP0_FLAG);
ctrl_ctx->drop_flags = 0;
xhci_queue_command(ctrl, (void *)ctrl_ctx, slot_id, 0, TRB_ADDR_DEV);
event = xhci_wait_for_event(ctrl, TRB_COMPLETION);
BUG_ON(TRB_TO_SLOT_ID(le32_to_cpu(event->event_cmd.flags)) != slot_id);
switch (GET_COMP_CODE(le32_to_cpu(event->event_cmd.status))) {
case COMP_CTX_STATE:
case COMP_EBADSLT:
printf("Setup ERROR: address device command for slot %d.\n",
slot_id);
ret = -EINVAL;
break;
case COMP_TX_ERR:
printf("Device not responding to set address.\n");
ret = -EPROTO;
break;
case COMP_DEV_ERR:
printf("ERROR: Incompatible device"
"for address device command.\n");
ret = -ENODEV;
break;
case COMP_SUCCESS:
debug("Successful Address Device command\n");
udev->status = 0;
break;
default:
printf("ERROR: unexpected command completion code 0x%x.\n",
GET_COMP_CODE(le32_to_cpu(event->event_cmd.status)));
ret = -EINVAL;
break;
}
xhci_acknowledge_event(ctrl);
if (ret < 0)
/*
* TODO: Unsuccessful Address Device command shall leave the
* slot in default state. So, issue Disable Slot command now.
*/
return ret;
xhci_inval_cache((uint32_t)virt_dev->out_ctx->bytes,
virt_dev->out_ctx->size);
slot_ctx = xhci_get_slot_ctx(ctrl, virt_dev->out_ctx);
debug("xHC internal address is: %d\n",
le32_to_cpu(slot_ctx->dev_state) & DEV_ADDR_MASK);
return 0;
+}
+/**
- Issue Enable slot command to the controller to allocate
- device slot and assign the slot id. It fails if the xHC
- ran out of device slots, the Enable Slot command timed out,
- or allocating memory failed.
- @param udev pointer to the Device Data Structure
- @return Returns 0 on succes else return -1 on failure
- */
+int usb_alloc_device(struct usb_device *udev) +{
union xhci_trb *event;
struct xhci_ctrl *ctrl = udev->controller;
/*
* Root hub will be first device to be initailized.
* If this device is root-hub, don't do any xHC related
* stuff.
*/
if (ctrl->rootdev == 0) {
udev->speed = USB_SPEED_SUPER;
return 0;
}
xhci_queue_command(ctrl, NULL, 0, 0, TRB_ENABLE_SLOT);
event = xhci_wait_for_event(ctrl, TRB_COMPLETION);
BUG_ON(GET_COMP_CODE(le32_to_cpu(event->event_cmd.status))
!= COMP_SUCCESS);
udev->slot_id = TRB_TO_SLOT_ID(le32_to_cpu(event->event_cmd.flags));
xhci_acknowledge_event(ctrl);
if (xhci_alloc_virt_device(udev) < 0) {
/*
* TODO: Unsuccessful Address Device command shall leave
* the slot in default. So, issue Disable Slot command now.
*/
printf("Could not allocate xHCI USB device data structures\n");
return -1;
}
return 0;
+}
+/*
- Full speed devices may have a max packet size greater than 8 bytes, but the
- USB core doesn't know that until it reads the first 8 bytes of the
- descriptor. If the usb_device's max packet size changes after that point,
- we need to issue an evaluate context command and wait on it.
- @param udev pointer to the Device Data Structure
- @return returns the status of the xhci_configure_endpoints
- */
+int xhci_check_maxpacket(struct usb_device *udev) +{
struct xhci_ctrl *ctrl = udev->controller;
unsigned int slot_id = udev->slot_id;
int ep_index = 0; /* control endpoint */
struct xhci_container_ctx *in_ctx;
struct xhci_container_ctx *out_ctx;
struct xhci_input_control_ctx *ctrl_ctx;
struct xhci_ep_ctx *ep_ctx;
int max_packet_size;
int hw_max_packet_size;
int ret = 0;
struct usb_interface *ifdesc;
ifdesc = &udev->config.if_desc[0];
out_ctx = ctrl->devs[slot_id]->out_ctx;
xhci_inval_cache((uint32_t)out_ctx->bytes, out_ctx->size);
ep_ctx = xhci_get_ep_ctx(ctrl, out_ctx, ep_index);
hw_max_packet_size = MAX_PACKET_DECODED(le32_to_cpu(ep_ctx->ep_info2));
max_packet_size = usb_endpoint_maxp(&ifdesc->ep_desc[0]);
if (hw_max_packet_size != max_packet_size) {
debug("Max Packet Size for ep 0 changed.\n");
debug("Max packet size in usb_device = %d\n", max_packet_size);
debug("Max packet size in xHCI HW = %d\n", hw_max_packet_size);
debug("Issuing evaluate context command.\n");
/* Set up the modified control endpoint 0 */
xhci_endpoint_copy(ctrl, ctrl->devs[slot_id]->in_ctx,
ctrl->devs[slot_id]->out_ctx, ep_index);
in_ctx = ctrl->devs[slot_id]->in_ctx;
ep_ctx = xhci_get_ep_ctx(ctrl, in_ctx, ep_index);
ep_ctx->ep_info2 &= cpu_to_le32(~MAX_PACKET_MASK);
ep_ctx->ep_info2 |= cpu_to_le32(MAX_PACKET(max_packet_size));
/*
* Set up the input context flags for the command
* FIXME: This won't work if a non-default control endpoint
* changes max packet sizes.
*/
ctrl_ctx = xhci_get_input_control_ctx(in_ctx);
ctrl_ctx->add_flags = cpu_to_le32(EP0_FLAG);
ctrl_ctx->drop_flags = 0;
ret = xhci_configure_endpoints(udev, true);
}
return ret;
+}
+/**
- Clears the Change bits of the Port Status Register
- @param wValue request value
- @param wIndex request index
- @param addr address of posrt status register
- @param port_status state of port status register
- @return none
- */
+static void xhci_clear_port_change_bit(u16 wValue,
u16 wIndex, volatile uint32_t *addr, u32 port_status)
+{
char *port_change_bit;
u32 status;
switch (wValue) {
case USB_PORT_FEAT_C_RESET:
status = PORT_RC;
port_change_bit = "reset";
break;
case USB_PORT_FEAT_C_CONNECTION:
status = PORT_CSC;
port_change_bit = "connect";
break;
case USB_PORT_FEAT_C_OVER_CURRENT:
status = PORT_OCC;
port_change_bit = "over-current";
break;
case USB_PORT_FEAT_C_ENABLE:
status = PORT_PEC;
port_change_bit = "enable/disable";
break;
case USB_PORT_FEAT_C_SUSPEND:
status = PORT_PLC;
port_change_bit = "suspend/resume";
break;
default:
/* Should never happen */
return;
}
/* Change bits are all write 1 to clear */
xhci_writel(addr, port_status | status);
port_status = xhci_readl(addr);
debug("clear port %s change, actual port %d status = 0x%x\n",
port_change_bit, wIndex, port_status);
+}
+/**
- Save Read Only (RO) bits and save read/write bits where
- writing a 0 clears the bit and writing a 1 sets the bit (RWS).
- For all other types (RW1S, RW1CS, RW, and RZ), writing a '0' has no effect.
- @param state state of the Port Status and Control Regsiter
- @return a value that would result in the port being in the
same state, if the value was written to the port
status control register.
- */
+static u32 xhci_port_state_to_neutral(u32 state) +{
/* Save read-only status and port state */
return (state & XHCI_PORT_RO) | (state & XHCI_PORT_RWS);
+}
+/**
- Submits the Requests to the XHCI Host Controller
- @param udev pointer to the USB device structure
- @param pipe contains the DIR_IN or OUT , devnum
- @param buffer buffer to be read/written based on the request
- @return returns 0 if successful else -1 on failure
- */
+static int xhci_submit_root(struct usb_device *udev, unsigned long pipe,
void *buffer, struct devrequest *req)
+{
uint8_t tmpbuf[4];
u16 typeReq;
void *srcptr = NULL;
int len, srclen;
uint32_t reg;
volatile uint32_t *status_reg;
struct xhci_ctrl *ctrl = udev->controller;
struct xhci_hcor *hcor = ctrl->hcor;
if (((req->requesttype & USB_RT_PORT) &&
le16_to_cpu(req->index)) > CONFIG_SYS_USB_XHCI_MAX_ROOT_PORTS) {
printf("The request port(%d) is not configured\n",
le16_to_cpu(req->index) - 1);
return -1;
}
status_reg = (volatile uint32_t *)
(&hcor->PortRegs[le16_to_cpu(req->index) - 1].or_portsc);
srclen = 0;
typeReq = req->request | req->requesttype << 8;
switch (typeReq) {
case DeviceRequest | USB_REQ_GET_DESCRIPTOR:
switch (le16_to_cpu(req->value) >> 8) {
case USB_DT_DEVICE:
debug("USB_DT_DEVICE request\n");
srcptr = &descriptor.device;
srclen = 0x12;
break;
case USB_DT_CONFIG:
debug("USB_DT_CONFIG config\n");
srcptr = &descriptor.config;
srclen = 0x19;
break;
case USB_DT_STRING:
debug("USB_DT_STRING config\n");
switch (le16_to_cpu(req->value) & 0xff) {
case 0: /* Language */
srcptr = "\4\3\11\4";
srclen = 4;
break;
case 1: /* Vendor String */
srcptr = "\16\3u\0-\0b\0o\0o\0t\0";
srclen = 14;
break;
case 2: /* Product Name */
srcptr = "\52\3X\0H\0C\0I\0 "
"\0H\0o\0s\0t\0 "
"\0C\0o\0n\0t\0r\0o\0l\0l\0e\0r\0";
srclen = 42;
break;
default:
printf("unknown value DT_STRING %x\n",
le16_to_cpu(req->value));
goto unknown;
}
break;
default:
printf("unknown value %x\n", le16_to_cpu(req->value));
goto unknown;
}
break;
case USB_REQ_GET_DESCRIPTOR | ((USB_DIR_IN | USB_RT_HUB) << 8):
switch (le16_to_cpu(req->value) >> 8) {
case USB_DT_HUB:
debug("USB_DT_HUB config\n");
srcptr = &descriptor.hub;
srclen = 0x8;
break;
default:
printf("unknown value %x\n", le16_to_cpu(req->value));
goto unknown;
}
break;
case USB_REQ_SET_ADDRESS | (USB_RECIP_DEVICE << 8):
debug("USB_REQ_SET_ADDRESS\n");
ctrl->rootdev = le16_to_cpu(req->value);
break;
case DeviceOutRequest | USB_REQ_SET_CONFIGURATION:
/* Do nothing */
break;
case USB_REQ_GET_STATUS | ((USB_DIR_IN | USB_RT_HUB) << 8):
tmpbuf[0] = 1; /* USB_STATUS_SELFPOWERED */
tmpbuf[1] = 0;
srcptr = tmpbuf;
srclen = 2;
break;
case USB_REQ_GET_STATUS | ((USB_RT_PORT | USB_DIR_IN) << 8):
memset(tmpbuf, 0, 4);
reg = xhci_readl(status_reg);
if (reg & PORT_CONNECT) {
tmpbuf[0] |= USB_PORT_STAT_CONNECTION;
switch (reg & DEV_SPEED_MASK) {
case XDEV_FS:
debug("SPEED = FULLSPEED\n");
break;
case XDEV_LS:
debug("SPEED = LOWSPEED\n");
tmpbuf[1] |= USB_PORT_STAT_LOW_SPEED >> 8;
break;
case XDEV_HS:
debug("SPEED = HIGHSPEED\n");
tmpbuf[1] |= USB_PORT_STAT_HIGH_SPEED >> 8;
break;
case XDEV_SS:
debug("SPEED = SUPERSPEED\n");
tmpbuf[1] |= USB_PORT_STAT_SUPER_SPEED >> 8;
break;
}
}
if (reg & PORT_PE)
tmpbuf[0] |= USB_PORT_STAT_ENABLE;
if ((reg & PORT_PLS_MASK) == XDEV_U3)
tmpbuf[0] |= USB_PORT_STAT_SUSPEND;
if (reg & PORT_OC)
tmpbuf[0] |= USB_PORT_STAT_OVERCURRENT;
if (reg & PORT_RESET)
tmpbuf[0] |= USB_PORT_STAT_RESET;
if (reg & PORT_POWER)
/*
* XXX: This Port power bit (for USB 3.0 hub)
* we are faking in USB 2.0 hub port status;
* since there's a change in bit positions in
* two:
* USB 2.0 port status PP is at position[8]
* USB 3.0 port status PP is at position[9]
* So, we are still keeping it at position [8]
*/
tmpbuf[1] |= USB_PORT_STAT_POWER >> 8;
if (reg & PORT_CSC)
tmpbuf[2] |= USB_PORT_STAT_C_CONNECTION;
if (reg & PORT_PEC)
tmpbuf[2] |= USB_PORT_STAT_C_ENABLE;
if (reg & PORT_OCC)
tmpbuf[2] |= USB_PORT_STAT_C_OVERCURRENT;
if (reg & PORT_RC)
tmpbuf[2] |= USB_PORT_STAT_C_RESET;
srcptr = tmpbuf;
srclen = 4;
break;
case USB_REQ_SET_FEATURE | ((USB_DIR_OUT | USB_RT_PORT) << 8):
reg = xhci_readl(status_reg);
reg = xhci_port_state_to_neutral(reg);
switch (le16_to_cpu(req->value)) {
case USB_PORT_FEAT_ENABLE:
reg |= PORT_PE;
xhci_writel(status_reg, reg);
break;
case USB_PORT_FEAT_POWER:
reg |= PORT_POWER;
xhci_writel(status_reg, reg);
break;
case USB_PORT_FEAT_RESET:
reg |= PORT_RESET;
xhci_writel(status_reg, reg);
break;
default:
printf("unknown feature %x\n", le16_to_cpu(req->value));
goto unknown;
}
break;
case USB_REQ_CLEAR_FEATURE | ((USB_DIR_OUT | USB_RT_PORT) << 8):
reg = xhci_readl(status_reg);
reg = xhci_port_state_to_neutral(reg);
switch (le16_to_cpu(req->value)) {
case USB_PORT_FEAT_ENABLE:
reg &= ~PORT_PE;
break;
case USB_PORT_FEAT_POWER:
reg &= ~PORT_POWER;
break;
case USB_PORT_FEAT_C_RESET:
case USB_PORT_FEAT_C_CONNECTION:
case USB_PORT_FEAT_C_OVER_CURRENT:
case USB_PORT_FEAT_C_ENABLE:
xhci_clear_port_change_bit((le16_to_cpu(req->value)),
le16_to_cpu(req->index),
status_reg, reg);
break;
default:
printf("unknown feature %x\n", le16_to_cpu(req->value));
goto unknown;
}
xhci_writel(status_reg, reg);
break;
default:
printf("Unknown request\n");
goto unknown;
}
debug("scrlen = %d\n req->length = %d\n",
srclen, le16_to_cpu(req->length));
len = min(srclen, le16_to_cpu(req->length));
if (srcptr != NULL && len > 0)
memcpy(buffer, srcptr, len);
else
debug("Len is 0\n");
udev->act_len = len;
udev->status = 0;
return 0;
+unknown:
udev->act_len = 0;
udev->status = USB_ST_STALLED;
return -1;
+}
+/**
- Submits the INT request to XHCI Host cotroller
- @param udev pointer to the USB device
- @param pipe contains the DIR_IN or OUT , devnum
- @param buffer buffer to be read/written based on the request
- @param length length of the buffer
- @param interval interval of the interrupt
- @return 0
- */
+int +submit_int_msg(struct usb_device *udev, unsigned long pipe, void *buffer,
int length, int interval)
+{
/*
* TODO: Not addressing any interrupt type transfer requests
* Add support for it later.
*/
return -1;
+}
+/**
- submit the BULK type of request to the USB Device
- @param udev pointer to the USB device
- @param pipe contains the DIR_IN or OUT , devnum
- @param buffer buffer to be read/written based on the request
- @param length length of the buffer
- @return returns 0 if successful else -1 on failure
- */
+int +submit_bulk_msg(struct usb_device *udev, unsigned long pipe, void *buffer,
int length)
+{
if (usb_pipetype(pipe) != PIPE_BULK) {
printf("non-bulk pipe (type=%lu)", usb_pipetype(pipe));
return -1;
}
return xhci_bulk_tx(udev, pipe, length, buffer);
+}
+/**
- submit the control type of request to the Root hub/Device based on the devnum
- @param udev pointer to the USB device
- @param pipe contains the DIR_IN or OUT , devnum
- @param buffer buffer to be read/written based on the request
- @param length length of the buffer
- @param setup Request type
- @return returns 0 if successful else -1 on failure
- */
+int +submit_control_msg(struct usb_device *udev, unsigned long pipe, void *buffer,
int length, struct devrequest *setup)
+{
struct xhci_ctrl *ctrl = udev->controller;
int ret = 0;
if (usb_pipetype(pipe) != PIPE_CONTROL) {
printf("non-control pipe (type=%lu)", usb_pipetype(pipe));
return -1;
}
if (usb_pipedevice(pipe) == ctrl->rootdev)
return xhci_submit_root(udev, pipe, buffer, setup);
if (setup->request == USB_REQ_SET_ADDRESS)
return xhci_address_device(udev);
if (setup->request == USB_REQ_SET_CONFIGURATION) {
ret = xhci_set_configuration(udev);
if (ret) {
printf("Failed to configure xHC endpoint\n");
Use puts when there is no args. Also s/xHC/xHCI
Ok.
return ret;
}
}
return xhci_ctrl_tx(udev, pipe, setup, length, buffer);
+}
+/**
- Intialises the XHCI host controller
- and allocates the necessary data structures
- @param index index to the host controller data structure
- @return pointer to the intialised controller
- */
+int usb_lowlevel_init(int index, void **controller) +{
uint32_t val;
uint32_t val2;
uint32_t reg;
struct xhci_hccr *hccr;
struct xhci_hcor *hcor;
struct xhci_ctrl *ctrl;
if (xhci_hcd_init(index, &hccr, (struct xhci_hcor **)&hcor) != 0)
return -ENODEV;
if (xhci_reset(hcor) != 0)
return -ENODEV;
ctrl = &xhcic[index];
ctrl->hccr = hccr;
ctrl->hcor = hcor;
/*
* Program the Number of Device Slots Enabled field in the CONFIG
* register with the max value of slots the HC can handle.
*/
val = (xhci_readl(&hccr->cr_hcsparams1) & HCS_SLOTS_MASK);
val2 = xhci_readl(&hcor->or_config);
val |= (val2 & ~HCS_SLOTS_MASK);
xhci_writel(&hcor->or_config, val);
/* initializing xhci data structures */
if (xhci_mem_init(ctrl, hccr, hcor) < 0)
return -ENOMEM;
reg = xhci_readl(&hccr->cr_hcsparams1);
descriptor.hub.bNbrPorts = ((reg & HCS_MAX_PORTS_MASK) >>
HCS_MAX_PORTS_SHIFT);
printf("Register %x NbrPorts %d\n", reg, descriptor.hub.bNbrPorts);
Nitpick maybe make this debug.
This is what EHCI also prints for information. Will make it debug if you want.
/* Port Indicators */
reg = xhci_readl(&hccr->cr_hccparams);
if (HCS_INDICATOR(reg))
put_unaligned(get_unaligned(&descriptor.hub.wHubCharacteristics)
| 0x80, &descriptor.hub.wHubCharacteristics);
/* Port Power Control */
if (HCC_PPC(reg))
put_unaligned(get_unaligned(&descriptor.hub.wHubCharacteristics)
| 0x01, &descriptor.hub.wHubCharacteristics);
if (xhci_start(hcor)) {
xhci_reset(hcor);
return -ENODEV;
}
/* Zero'ing IRQ control register and IRQ pending register */
xhci_writel(&ctrl->ir_set->irq_control, 0x0);
xhci_writel(&ctrl->ir_set->irq_pending, 0x0);
reg = HC_VERSION(xhci_readl(&hccr->cr_capbase));
printf("USB XHCI %x.%02x\n", reg >> 8, reg & 0xff);
*controller = &xhcic[index];
return 0;
+}
+/**
- Stops the XHCI host controller
- and cleans up all the related data structures
- @param index index to the host controller data structure
- @return none
- */
+int usb_lowlevel_stop(int index) +{
struct xhci_ctrl *ctrl = (xhcic + index);
u32 temp;
xhci_reset(ctrl->hcor);
debug("// Disabling event ring interrupts\n");
temp = xhci_readl(&ctrl->hcor->or_usbsts);
xhci_writel(&ctrl->hcor->or_usbsts, temp & ~STS_EINT);
temp = xhci_readl(&ctrl->ir_set->irq_pending);
xhci_writel(&ctrl->ir_set->irq_pending, ER_IRQ_DISABLE(temp));
xhci_hcd_stop(index);
xhci_cleanup(ctrl);
return 0;
+} diff --git a/drivers/usb/host/xhci.h b/drivers/usb/host/xhci.h new file mode 100644 index 0000000..467afe0 --- /dev/null +++ b/drivers/usb/host/xhci.h @@ -0,0 +1,1280 @@ +/*
- USB HOST XHCI Controller
- Copyright (C) 2013 Samsung Electronics Co.Ltd
- Vivek Gautam gautam.vivek@samsung.com
- Vikas Sajjan vikas.sajjan@samsung.com
- Based on xHCI host controller driver in linux-kernel
- by Sarah Sharp.
- This program is free software; you can redistribute it and/or
- modify it under the terms of the GNU General Public License as
- published by the Free Software Foundation; either version 2 of
- the License, or (at your option) any later version.
- This program is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- GNU General Public License for more details.
- You should have received a copy of the GNU General Public License
- along with this program; if not, write to the Free Software
- Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
- MA 02110-1301 USA
- */
Needs new SPDX license.
Sure
- SPDX-License-Identifier: GPL-2.0+
+#ifndef HOST_XHCI_H_ +#define HOST_XHCI_H_
+#include <asm/cache.h> +#include <linux/list.h>
+/* (shifted) direction/type/recipient from the USB 2.0 spec, table 9.2 */ +#define DeviceRequest \
((USB_DIR_IN | USB_TYPE_STANDARD | USB_RECIP_DEVICE) << 8)
+#define DeviceOutRequest \
((USB_DIR_OUT | USB_TYPE_STANDARD | USB_RECIP_DEVICE) << 8)
+#define InterfaceRequest \
((USB_DIR_IN | USB_TYPE_STANDARD | USB_RECIP_INTERFACE) << 8)
+#define EndpointRequest \
((USB_DIR_IN | USB_TYPE_STANDARD | USB_RECIP_INTERFACE) << 8)
+#define EndpointOutRequest \
((USB_DIR_OUT | USB_TYPE_STANDARD | USB_RECIP_INTERFACE) << 8)
+#define upper_32_bits(n) (u32)(((n) >> 32)) +#define lower_32_bits(n) (u32)(n)
+#define MAX_EP_CTX_NUM 31 +#define XHCI_ALIGNMENT 64 +/* Generic timeout for XHCI events */ +#define XHCI_TIMEOUT 5000 +/* Max number of USB devices for any host controller - limit in section 6.1 */ +#define MAX_HC_SLOTS 256 +/* Section 5.3.3 - MaxPorts */ +#define MAX_HC_PORTS 127
+/* Up to 16 ms to halt an HC */ +#define XHCI_MAX_HALT_USEC (16*1000)
+#define XHCI_MAX_RESET_USEC (250*1000)
+/*
- These bits are Read Only (RO) and should be saved and written to the
- registers: 0, 3, 10:13, 30
- connect status, over-current status, port speed, and device removable.
- connect status and port speed are also sticky - meaning they're in
- the AUX well and they aren't changed by a hot, warm, or cold reset.
- */
+#define XHCI_PORT_RO ((1 << 0) | (1 << 3) | (0xf << 10) | (1 << 30)) +/*
- These bits are RW; writing a 0 clears the bit, writing a 1 sets the bit:
- bits 5:8, 9, 14:15, 25:27
- link state, port power, port indicator state, "wake on" enable state
- */
+#define XHCI_PORT_RWS ((0xf << 5) | (1 << 9) | (0x3 << 14) | (0x7 << 25)) +/*
- These bits are RW; writing a 1 sets the bit, writing a 0 has no effect:
- bit 4 (port reset)
- */
+#define XHCI_PORT_RW1S ((1 << 4)) +/*
- These bits are RW; writing a 1 clears the bit, writing a 0 has no effect:
- bits 1, 17, 18, 19, 20, 21, 22, 23
- port enable/disable, and
- change bits: connect, PED,
- warm port reset changed (reserved zero for USB 2.0 ports),
- over-current, reset, link state, and L1 change
- */
+#define XHCI_PORT_RW1CS ((1 << 1) | (0x7f << 17)) +/*
- Bit 16 is RW, and writing a '1' to it causes the link state control to be
- latched in
- */
+#define XHCI_PORT_RW ((1 << 16)) +/*
- These bits are Reserved Zero (RsvdZ) and zero should be written to them:
- bits 2, 24, 28:31
- */
+#define XHCI_PORT_RZ ((1 << 2) | (1 << 24) | (0xf << 28))
+/*
- XHCI Register Space.
- */
+struct xhci_hccr {
uint32_t cr_capbase;
uint32_t cr_hcsparams1;
uint32_t cr_hcsparams2;
uint32_t cr_hcsparams3;
uint32_t cr_hccparams;
uint32_t cr_dboff;
uint32_t cr_rtsoff;
+/* hc_capbase bitmasks */ +/* bits 7:0 - how long is the Capabilities register */ +#define HC_LENGTH(p) XHCI_HC_LENGTH(p) +/* bits 31:16 */ +#define HC_VERSION(p) (((p) >> 16) & 0xffff)
+/* HCSPARAMS1 - hcs_params1 - bitmasks */ +/* bits 0:7, Max Device Slots */ +#define HCS_MAX_SLOTS(p) (((p) >> 0) & 0xff) +#define HCS_SLOTS_MASK 0xff +/* bits 8:18, Max Interrupters */ +#define HCS_MAX_INTRS(p) (((p) >> 8) & 0x7ff) +/* bits 24:31, Max Ports - max value is 0x7F = 127 ports */ +#define HCS_MAX_PORTS_SHIFT 24 +#define HCS_MAX_PORTS_MASK (0x7f << HCS_MAX_PORTS_SHIFT) +#define HCS_MAX_PORTS(p) (((p) >> 24) & 0x7f)
+/* HCSPARAMS2 - hcs_params2 - bitmasks */ +/* bits 0:3, frames or uframes that SW needs to queue transactions
- ahead of the HW to meet periodic deadlines */
+#define HCS_IST(p) (((p) >> 0) & 0xf) +/* bits 4:7, max number of Event Ring segments */ +#define HCS_ERST_MAX(p) (((p) >> 4) & 0xf) +/* bit 26 Scratchpad restore - for save/restore HW state - not used yet */ +/* bits 27:31 number of Scratchpad buffers SW must allocate for the HW */ +#define HCS_MAX_SCRATCHPAD(p) (((p) >> 27) & 0x1f)
+/* HCSPARAMS3 - hcs_params3 - bitmasks */ +/* bits 0:7, Max U1 to U0 latency for the roothub ports */ +#define HCS_U1_LATENCY(p) (((p) >> 0) & 0xff) +/* bits 16:31, Max U2 to U0 latency for the roothub ports */ +#define HCS_U2_LATENCY(p) (((p) >> 16) & 0xffff)
+/* HCCPARAMS - hcc_params - bitmasks */ +/* true: HC can use 64-bit address pointers */ +#define HCC_64BIT_ADDR(p) ((p) & (1 << 0)) +/* true: HC can do bandwidth negotiation */ +#define HCC_BANDWIDTH_NEG(p) ((p) & (1 << 1)) +/* true: HC uses 64-byte Device Context structures
- FIXME 64-byte context structures aren't supported yet.
- */
+#define HCC_64BYTE_CONTEXT(p) ((p) & (1 << 2)) +/* true: HC has port power switches */ +#define HCC_PPC(p) ((p) & (1 << 3)) +/* true: HC has port indicators */ +#define HCS_INDICATOR(p) ((p) & (1 << 4)) +/* true: HC has Light HC Reset Capability */ +#define HCC_LIGHT_RESET(p) ((p) & (1 << 5)) +/* true: HC supports latency tolerance messaging */ +#define HCC_LTC(p) ((p) & (1 << 6)) +/* true: no secondary Stream ID Support */ +#define HCC_NSS(p) ((p) & (1 << 7)) +/* Max size for Primary Stream Arrays - 2^(n+1), where n is bits 12:15 */ +#define HCC_MAX_PSA(p) (1 << ((((p) >> 12) & 0xf) + 1)) +/* Extended Capabilities pointer from PCI base - section 5.3.6 */ +#define HCC_EXT_CAPS(p) XHCI_HCC_EXT_CAPS(p)
+/* db_off bitmask - bits 0:1 reserved */ +#define DBOFF_MASK (~0x3)
+/* run_regs_off bitmask - bits 0:4 reserved */ +#define RTSOFF_MASK (~0x1f)
+};
+struct xhci_hcor_portRegss {
volatile uint32_t or_portsc;
volatile uint32_t or_portpmsc;
volatile uint32_t or_portli;
volatile uint32_t reserved_3;
+};
+struct xhci_hcor {
volatile uint32_t or_usbcmd;
volatile uint32_t or_usbsts;
volatile uint32_t or_pagesize;
volatile uint32_t reserved_0[2];
volatile uint32_t or_dnctrl;
volatile uint64_t or_crcr;
volatile uint32_t reserved_1[4];
volatile uint64_t or_dcbaap;
volatile uint32_t or_config;
volatile uint32_t reserved_2[241];
struct xhci_hcor_portRegss PortRegs[CONFIG_SYS_USB_XHCI_MAX_ROOT_PORTS];
uint32_t reserved_4[CONFIG_SYS_USB_XHCI_MAX_ROOT_PORTS * 254];
+};
+/* USBCMD - USB command - command bitmasks */ +/* start/stop HC execution - do not write unless HC is halted*/ +#define CMD_RUN XHCI_CMD_RUN +/* Reset HC - resets internal HC state machine and all registers (except
- PCI config regs). HC does NOT drive a USB reset on the downstream ports.
- The xHCI driver must reinitialize the xHC after setting this bit.
- */
+#define CMD_RESET (1 << 1) +/* Event Interrupt Enable - a '1' allows interrupts from the host controller */ +#define CMD_EIE XHCI_CMD_EIE +/* Host System Error Interrupt Enable - get out-of-band signal for HC errors */ +#define CMD_HSEIE XHCI_CMD_HSEIE +/* bits 4:6 are reserved (and should be preserved on writes). */ +/* light reset (port status stays unchanged) - reset completed when this is 0 */ +#define CMD_LRESET (1 << 7) +/* host controller save/restore state. */ +#define CMD_CSS (1 << 8) +#define CMD_CRS (1 << 9) +/* Enable Wrap Event - '1' means xHC generates an event when MFINDEX wraps. */ +#define CMD_EWE XHCI_CMD_EWE +/* MFINDEX power management - '1' means xHC can stop MFINDEX counter if all root
- hubs are in U3 (selective suspend), disconnect, disabled, or powered-off.
- '0' means the xHC can power it off if all ports are in the disconnect,
- disabled, or powered-off state.
- */
+#define CMD_PM_INDEX (1 << 11) +/* bits 12:31 are reserved (and should be preserved on writes). */
+/* USBSTS - USB status - status bitmasks */ +/* HC not running - set to 1 when run/stop bit is cleared. */ +#define STS_HALT XHCI_STS_HALT +/* serious error, e.g. PCI parity error. The HC will clear the run/stop bit. */ +#define STS_FATAL (1 << 2) +/* event interrupt - clear this prior to clearing any IP flags in IR set*/ +#define STS_EINT (1 << 3) +/* port change detect */ +#define STS_PORT (1 << 4) +/* bits 5:7 reserved and zeroed */ +/* save state status - '1' means xHC is saving state */ +#define STS_SAVE (1 << 8) +/* restore state status - '1' means xHC is restoring state */ +#define STS_RESTORE (1 << 9) +/* true: save or restore error */ +#define STS_SRE (1 << 10) +/* true: Controller Not Ready to accept doorbell or op reg writes after reset */ +#define STS_CNR XHCI_STS_CNR +/* true: internal Host Controller Error - SW needs to reset and reinitialize */ +#define STS_HCE (1 << 12) +/* bits 13:31 reserved and should be preserved */
+/*
- DNCTRL - Device Notification Control Register - dev_notification bitmasks
- Generate a device notification event when the HC sees a transaction with a
- notification type that matches a bit set in this bit field.
- */
+#define DEV_NOTE_MASK (0xffff) +#define ENABLE_DEV_NOTE(x) (1 << (x)) +/* Most of the device notification types should only be used for debug.
- SW does need to pay attention to function wake notifications.
- */
+#define DEV_NOTE_FWAKE ENABLE_DEV_NOTE(1)
+/* CRCR - Command Ring Control Register - cmd_ring bitmasks */ +/* bit 0 is the command ring cycle state */ +/* stop ring operation after completion of the currently executing command */ +#define CMD_RING_PAUSE (1 << 1) +/* stop ring immediately - abort the currently executing command */ +#define CMD_RING_ABORT (1 << 2) +/* true: command ring is running */ +#define CMD_RING_RUNNING (1 << 3) +/* bits 4:5 reserved and should be preserved */ +/* Command Ring pointer - bit mask for the lower 32 bits. */ +#define CMD_RING_RSVD_BITS (0x3f)
+/* CONFIG - Configure Register - config_reg bitmasks */ +/* bits 0:7 - maximum number of device slots enabled (NumSlotsEn) */ +#define MAX_DEVS(p) ((p) & 0xff) +/* bits 8:31 - reserved and should be preserved */
+/* PORTSC - Port Status and Control Register - port_status_base bitmasks */ +/* true: device connected */ +#define PORT_CONNECT (1 << 0) +/* true: port enabled */ +#define PORT_PE (1 << 1) +/* bit 2 reserved and zeroed */ +/* true: port has an over-current condition */ +#define PORT_OC (1 << 3) +/* true: port reset signaling asserted */ +#define PORT_RESET (1 << 4) +/* Port Link State - bits 5:8
- A read gives the current link PM state of the port,
- a write with Link State Write Strobe set sets the link state.
- */
+#define PORT_PLS_MASK (0xf << 5) +#define XDEV_U0 (0x0 << 5) +#define XDEV_U2 (0x2 << 5) +#define XDEV_U3 (0x3 << 5) +#define XDEV_RESUME (0xf << 5) +/* true: port has power (see HCC_PPC) */ +#define PORT_POWER (1 << 9) +/* bits 10:13 indicate device speed:
- 0 - undefined speed - port hasn't be initialized by a reset yet
- 1 - full speed
- 2 - low speed
- 3 - high speed
- 4 - super speed
- 5-15 reserved
- */
+#define DEV_SPEED_MASK (0xf << 10) +#define XDEV_FS (0x1 << 10) +#define XDEV_LS (0x2 << 10) +#define XDEV_HS (0x3 << 10) +#define XDEV_SS (0x4 << 10) +#define DEV_UNDEFSPEED(p) (((p) & DEV_SPEED_MASK) == (0x0<<10)) +#define DEV_FULLSPEED(p) (((p) & DEV_SPEED_MASK) == XDEV_FS) +#define DEV_LOWSPEED(p) (((p) & DEV_SPEED_MASK) == XDEV_LS) +#define DEV_HIGHSPEED(p) (((p) & DEV_SPEED_MASK) == XDEV_HS) +#define DEV_SUPERSPEED(p) (((p) & DEV_SPEED_MASK) == XDEV_SS) +/* Bits 20:23 in the Slot Context are the speed for the device */ +#define SLOT_SPEED_FS (XDEV_FS << 10) +#define SLOT_SPEED_LS (XDEV_LS << 10) +#define SLOT_SPEED_HS (XDEV_HS << 10) +#define SLOT_SPEED_SS (XDEV_SS << 10) +/* Port Indicator Control */ +#define PORT_LED_OFF (0 << 14) +#define PORT_LED_AMBER (1 << 14) +#define PORT_LED_GREEN (2 << 14) +#define PORT_LED_MASK (3 << 14) +/* Port Link State Write Strobe - set this when changing link state */ +#define PORT_LINK_STROBE (1 << 16) +/* true: connect status change */ +#define PORT_CSC (1 << 17) +/* true: port enable change */ +#define PORT_PEC (1 << 18) +/* true: warm reset for a USB 3.0 device is done. A "hot" reset puts the port
- into an enabled state, and the device into the default state. A "warm" reset
- also resets the link, forcing the device through the link training sequence.
- SW can also look at the Port Reset register to see when warm reset is done.
- */
+#define PORT_WRC (1 << 19) +/* true: over-current change */ +#define PORT_OCC (1 << 20) +/* true: reset change - 1 to 0 transition of PORT_RESET */ +#define PORT_RC (1 << 21) +/* port link status change - set on some port link state transitions:
- Transition Reason
- U3 to Resume Wakeup signaling from a device
- Resume to Recovery to U0 USB 3.0 device resume
- Resume to U0 USB 2.0 device resume
- U3 to Recovery to U0 Software resume of USB 3.0 device complete
- U3 to U0 Software resume of USB 2.0 device complete
- U2 to U0 L1 resume of USB 2.1 device complete
- U0 to U0 (???) L1 entry rejection by USB 2.1 device
- U0 to disabled L1 entry error with USB 2.1 device
- Any state to inactive Error on USB 3.0 port
- */
+#define PORT_PLC (1 << 22) +/* port configure error change - port failed to configure its link partner */ +#define PORT_CEC (1 << 23) +/* bit 24 reserved */ +/* wake on connect (enable) */ +#define PORT_WKCONN_E (1 << 25) +/* wake on disconnect (enable) */ +#define PORT_WKDISC_E (1 << 26) +/* wake on over-current (enable) */ +#define PORT_WKOC_E (1 << 27) +/* bits 28:29 reserved */ +/* true: device is removable - for USB 3.0 roothub emulation */ +#define PORT_DEV_REMOVE (1 << 30) +/* Initiate a warm port reset - complete when PORT_WRC is '1' */ +#define PORT_WR (1 << 31)
+/* We mark duplicate entries with -1 */ +#define DUPLICATE_ENTRY ((u8)(-1))
+/* Port Power Management Status and Control - port_power_base bitmasks */ +/* Inactivity timer value for transitions into U1, in microseconds.
- Timeout can be up to 127us. 0xFF means an infinite timeout.
- */
+#define PORT_U1_TIMEOUT(p) ((p) & 0xff) +/* Inactivity timer value for transitions into U2 */ +#define PORT_U2_TIMEOUT(p) (((p) & 0xff) << 8) +/* Bits 24:31 for port testing */
+/* USB2 Protocol PORTSPMSC */ +#define PORT_L1S_MASK 7 +#define PORT_L1S_SUCCESS 1 +#define PORT_RWE (1 << 3) +#define PORT_HIRD(p) (((p) & 0xf) << 4) +#define PORT_HIRD_MASK (0xf << 4) +#define PORT_L1DS(p) (((p) & 0xff) << 8) +#define PORT_HLE (1 << 16)
+/** +* struct xhci_intr_reg - Interrupt Register Set +* @irq_pending: IMAN - Interrupt Management Register. Used to enable +* interrupts and check for pending interrupts. +* @irq_control: IMOD - Interrupt Moderation Register. +* Used to throttle interrupts. +* @erst_size: Number of segments in the
Event Ring Segment Table (ERST).
+* @erst_base: ERST base address. +* @erst_dequeue: Event ring dequeue pointer. +* +* Each interrupter (defined by a MSI-X vector) has an event ring and an Event +* Ring Segment Table (ERST) associated with it. +* The event ring is comprised of multiple segments of the same size. +* The HC places events on the ring and "updates the Cycle bit in the TRBs to +* indicate to software the current position of the Enqueue Pointer." +* The HCD (Linux) processes those events and updates the dequeue pointer. +*/ +struct xhci_intr_reg {
volatile __le32 irq_pending;
volatile __le32 irq_control;
volatile __le32 erst_size;
volatile __le32 rsvd;
volatile __le64 erst_base;
volatile __le64 erst_dequeue;
+};
+/* irq_pending bitmasks */ +#define ER_IRQ_PENDING(p) ((p) & 0x1) +/* bits 2:31 need to be preserved */ +/* THIS IS BUGGY - FIXME - IP IS WRITE 1 TO CLEAR */ +#define ER_IRQ_CLEAR(p) ((p) & 0xfffffffe) +#define ER_IRQ_ENABLE(p) ((ER_IRQ_CLEAR(p)) | 0x2) +#define ER_IRQ_DISABLE(p) ((ER_IRQ_CLEAR(p)) & ~(0x2))
+/* irq_control bitmasks */ +/* Minimum interval between interrupts (in 250ns intervals). The interval
- between interrupts will be longer if there are no events on the event ring.
- Default is 4000 (1 ms).
- */
+#define ER_IRQ_INTERVAL_MASK (0xffff) +/* Counter used to count down the time to the next interrupt - HW use only */ +#define ER_IRQ_COUNTER_MASK (0xffff << 16)
+/* erst_size bitmasks */ +/* Preserve bits 16:31 of erst_size */ +#define ERST_SIZE_MASK (0xffff << 16)
+/* erst_dequeue bitmasks */ +/* Dequeue ERST Segment Index (DESI) - Segment number (or alias)
- where the current dequeue pointer lies. This is an optional HW hint.
- */
+#define ERST_DESI_MASK (0x7) +/* Event Handler Busy (EHB) - is the event ring scheduled to be serviced by
- a work queue (or delayed service routine)?
- */
+#define ERST_EHB (1 << 3) +#define ERST_PTR_MASK (0xf)
+/**
- struct xhci_run_regs
- @microframe_index: MFINDEX - current microframe number
- Section 5.5 Host Controller Runtime Registers:
- "Software should read and write these registers using only Dword (32 bit)
- or larger accesses"
- */
+struct xhci_run_regs {
__le32 microframe_index;
__le32 rsvd[7];
struct xhci_intr_reg ir_set[128];
+};
+/**
- struct doorbell_array
- Bits 0 - 7: Endpoint target
- Bits 8 - 15: RsvdZ
- Bits 16 - 31: Stream ID
- Section 5.6
- */
+struct xhci_doorbell_array {
volatile __le32 doorbell[256];
+};
+#define DB_VALUE(ep, stream) ((((ep) + 1) & 0xff) | ((stream) << 16)) +#define DB_VALUE_HOST 0x00000000
+/**
- struct xhci_protocol_caps
- @revision: major revision, minor revision, capability ID,
and next capability pointer.
- @name_string: Four ASCII characters to say which spec this xHC
follows, typically "USB ".
- @port_info: Port offset, count, and protocol-defined information.
- */
+struct xhci_protocol_caps {
u32 revision;
u32 name_string;
u32 port_info;
+};
+#define XHCI_EXT_PORT_MAJOR(x) (((x) >> 24) & 0xff) +#define XHCI_EXT_PORT_OFF(x) ((x) & 0xff) +#define XHCI_EXT_PORT_COUNT(x) (((x) >> 8) & 0xff)
+/**
- struct xhci_container_ctx
- @type: Type of context. Used to calculated offsets to contained contexts.
- @size: Size of the context data
- @bytes: The raw context data given to HW
- @dma: dma address of the bytes
- Represents either a Device or Input context. Holds a pointer to the raw
- memory used for the context (bytes) and dma address of it (dma).
- */
+struct xhci_container_ctx {
unsigned type;
+#define XHCI_CTX_TYPE_DEVICE 0x1 +#define XHCI_CTX_TYPE_INPUT 0x2
int size;
u8 *bytes;
+};
+/**
- struct xhci_slot_ctx
- @dev_info: Route string, device speed, hub info, and last valid endpoint
- @dev_info2: Max exit latency for device number, root hub port number
- @tt_info: tt_info is used to construct split transaction tokens
- @dev_state: slot state and device address
- Slot Context - section 6.2.1.1. This assumes the HC uses 32-byte context
- structures. If the HC uses 64-byte contexts, there is an additional 32 bytes
- reserved at the end of the slot context for HC internal use.
- */
+struct xhci_slot_ctx {
__le32 dev_info;
__le32 dev_info2;
__le32 tt_info;
__le32 dev_state;
/* offset 0x10 to 0x1f reserved for HC internal use */
__le32 reserved[4];
+};
+/* dev_info bitmasks */ +/* Route String - 0:19 */ +#define ROUTE_STRING_MASK (0xfffff) +/* Device speed - values defined by PORTSC Device Speed field - 20:23 */ +#define DEV_SPEED (0xf << 20) +/* bit 24 reserved */ +/* Is this LS/FS device connected through a HS hub? - bit 25 */ +#define DEV_MTT (0x1 << 25) +/* Set if the device is a hub - bit 26 */ +#define DEV_HUB (0x1 << 26) +/* Index of the last valid endpoint context in this device context - 27:31 */ +#define LAST_CTX_MASK (0x1f << 27) +#define LAST_CTX(p) ((p) << 27) +#define LAST_CTX_TO_EP_NUM(p) (((p) >> 27) - 1) +#define SLOT_FLAG (1 << 0) +#define EP0_FLAG (1 << 1)
+/* dev_info2 bitmasks */ +/* Max Exit Latency (ms) - worst case time to wake up all links in dev path */ +#define MAX_EXIT (0xffff) +/* Root hub port number that is needed to access the USB device */ +#define ROOT_HUB_PORT(p) (((p) & 0xff) << 16) +#define ROOT_HUB_PORT_MASK (0xff) +#define ROOT_HUB_PORT_SHIFT (16) +#define DEVINFO_TO_ROOT_HUB_PORT(p) (((p) >> 16) & 0xff) +/* Maximum number of ports under a hub device */ +#define XHCI_MAX_PORTS(p) (((p) & 0xff) << 24)
+/* tt_info bitmasks */ +/*
- TT Hub Slot ID - for low or full speed devices attached to a high-speed hub
- The Slot ID of the hub that isolates the high speed signaling from
- this low or full-speed device. '0' if attached to root hub port.
- */
+#define TT_SLOT (0xff) +/*
- The number of the downstream facing port of the high-speed hub
- '0' if the device is not low or full speed.
- */
+#define TT_PORT (0xff << 8) +#define TT_THINK_TIME(p) (((p) & 0x3) << 16)
+/* dev_state bitmasks */ +/* USB device address - assigned by the HC */ +#define DEV_ADDR_MASK (0xff) +/* bits 8:26 reserved */ +/* Slot state */ +#define SLOT_STATE (0x1f << 27) +#define GET_SLOT_STATE(p) (((p) & (0x1f << 27)) >> 27)
+#define SLOT_STATE_DISABLED 0 +#define SLOT_STATE_ENABLED SLOT_STATE_DISABLED +#define SLOT_STATE_DEFAULT 1 +#define SLOT_STATE_ADDRESSED 2 +#define SLOT_STATE_CONFIGURED 3
+/**
- struct xhci_ep_ctx
- @ep_info: endpoint state, streams, mult, and interval information.
- @ep_info2: information on endpoint type, max packet size, max burst size,
error count, and whether the HC will force an event for all
transactions.
- @deq: 64-bit ring dequeue pointer address. If the endpoint only
defines one stream, this points to the endpoint transfer ring.
Otherwise, it points to a stream context array, which has a
ring pointer for each flow.
- @tx_info:
Average TRB lengths for the endpoint ring and
max payload within an Endpoint Service Interval Time (ESIT).
- Endpoint Context - section 6.2.1.2.This assumes the HC uses 32-byte context
- structures.If the HC uses 64-byte contexts, there is an additional 32 bytes
- reserved at the end of the endpoint context for HC internal use.
- */
+struct xhci_ep_ctx {
__le32 ep_info;
__le32 ep_info2;
__le64 deq;
__le32 tx_info;
/* offset 0x14 - 0x1f reserved for HC internal use */
__le32 reserved[3];
+};
+/* ep_info bitmasks */ +/*
- Endpoint State - bits 0:2
- 0 - disabled
- 1 - running
- 2 - halted due to halt condition - ok to manipulate endpoint ring
- 3 - stopped
- 4 - TRB error
- 5-7 - reserved
- */
+#define EP_STATE_MASK (0xf) +#define EP_STATE_DISABLED 0 +#define EP_STATE_RUNNING 1 +#define EP_STATE_HALTED 2 +#define EP_STATE_STOPPED 3 +#define EP_STATE_ERROR 4 +/* Mult - Max number of burtst within an interval, in EP companion desc. */ +#define EP_MULT(p) (((p) & 0x3) << 8) +#define CTX_TO_EP_MULT(p) (((p) >> 8) & 0x3) +/* bits 10:14 are Max Primary Streams */ +/* bit 15 is Linear Stream Array */ +/* Interval - period between requests to an endpoint - 125u increments. */ +#define EP_INTERVAL(p) (((p) & 0xff) << 16) +#define EP_INTERVAL_TO_UFRAMES(p) (1 << (((p) >> 16) & 0xff)) +#define CTX_TO_EP_INTERVAL(p) (((p) >> 16) & 0xff) +#define EP_MAXPSTREAMS_MASK (0x1f << 10) +#define EP_MAXPSTREAMS(p) (((p) << 10) & EP_MAXPSTREAMS_MASK) +/* Endpoint is set up with a Linear Stream Array (vs. Secondary Stream Array) */ +#define EP_HAS_LSA (1 << 15)
+/* ep_info2 bitmasks */ +/*
- Force Event - generate transfer events for all TRBs for this endpoint
- This will tell the HC to ignore the IOC and ISP flags (for debugging only).
- */
+#define FORCE_EVENT (0x1) +#define ERROR_COUNT(p) (((p) & 0x3) << 1) +#define ERROR_COUNT_SHIFT (1) +#define ERROR_COUNT_MASK (0x3) +#define CTX_TO_EP_TYPE(p) (((p) >> 3) & 0x7) +#define EP_TYPE(p) ((p) << 3) +#define EP_TYPE_SHIFT (3) +#define ISOC_OUT_EP 1 +#define BULK_OUT_EP 2 +#define INT_OUT_EP 3 +#define CTRL_EP 4 +#define ISOC_IN_EP 5 +#define BULK_IN_EP 6 +#define INT_IN_EP 7 +/* bit 6 reserved */ +/* bit 7 is Host Initiate Disable - for disabling stream selection */ +#define MAX_BURST(p) (((p)&0xff) << 8) +#define MAX_BURST_MASK (0xff) +#define MAX_BURST_SHIFT (8) +#define CTX_TO_MAX_BURST(p) (((p) >> 8) & 0xff) +#define MAX_PACKET(p) (((p)&0xffff) << 16) +#define MAX_PACKET_MASK (0xffff) +#define MAX_PACKET_DECODED(p) (((p) >> 16) & 0xffff) +#define MAX_PACKET_SHIFT (16)
+/* Get max packet size from ep desc. Bit 10..0 specify the max packet size.
- USB2.0 spec 9.6.6.
- */
+#define GET_MAX_PACKET(p) ((p) & 0x7ff)
+/* tx_info bitmasks */ +#define AVG_TRB_LENGTH_FOR_EP(p) ((p) & 0xffff) +#define MAX_ESIT_PAYLOAD_FOR_EP(p) (((p) & 0xffff) << 16) +#define CTX_TO_MAX_ESIT_PAYLOAD(p) (((p) >> 16) & 0xffff)
+/* deq bitmasks */ +#define EP_CTX_CYCLE_MASK (1 << 0)
+/**
- struct xhci_input_control_context
- Input control context; see section 6.2.5.
- @drop_context: set the bit of the endpoint context you want to disable
- @add_context: set the bit of the endpoint context you want to enable
- */
+struct xhci_input_control_ctx {
volatile __le32 drop_flags;
volatile __le32 add_flags;
__le32 rsvd2[6];
+};
+/**
- struct xhci_device_context_array
- @dev_context_ptr array of 64-bit DMA addresses for device contexts
- */
+struct xhci_device_context_array {
/* 64-bit device addresses; we only write 32-bit addresses */
__le64 dev_context_ptrs[MAX_HC_SLOTS];
+}; +/* TODO: write function to set the 64-bit device DMA address */ +/*
- TODO: change this to be dynamically sized at HC mem init time since the HC
- might not be able to handle the maximum number of devices possible.
- */
+struct xhci_transfer_event {
/* 64-bit buffer address, or immediate data */
__le64 buffer;
__le32 transfer_len;
/* This field is interpreted differently based on the type of TRB */
volatile __le32 flags;
+};
+/* Transfer event TRB length bit mask */ +/* bits 0:23 */ +#define EVENT_TRB_LEN(p) ((p) & 0xffffff)
+/** Transfer Event bit fields **/ +#define TRB_TO_EP_ID(p) (((p) >> 16) & 0x1f)
+/* Completion Code - only applicable for some types of TRBs */ +#define COMP_CODE_MASK (0xff << 24) +#define COMP_CODE_SHIFT (24) +#define GET_COMP_CODE(p) (((p) & COMP_CODE_MASK) >> 24)
+typedef enum {
COMP_SUCCESS = 1,
/* Data Buffer Error */
COMP_DB_ERR, /* 2 */
/* Babble Detected Error */
COMP_BABBLE, /* 3 */
/* USB Transaction Error */
COMP_TX_ERR, /* 4 */
/* TRB Error - some TRB field is invalid */
COMP_TRB_ERR, /* 5 */
/* Stall Error - USB device is stalled */
COMP_STALL, /* 6 */
/* Resource Error - HC doesn't have memory for that device configuration */
COMP_ENOMEM, /* 7 */
/* Bandwidth Error - not enough room in schedule for this dev config */
COMP_BW_ERR, /* 8 */
/* No Slots Available Error - HC ran out of device slots */
COMP_ENOSLOTS, /* 9 */
/* Invalid Stream Type Error */
COMP_STREAM_ERR, /* 10 */
/* Slot Not Enabled Error - doorbell rung for disabled device slot */
COMP_EBADSLT, /* 11 */
/* Endpoint Not Enabled Error */
COMP_EBADEP,/* 12 */
/* Short Packet */
COMP_SHORT_TX, /* 13 */
/* Ring Underrun - doorbell rung for an empty isoc OUT ep ring */
COMP_UNDERRUN, /* 14 */
/* Ring Overrun - isoc IN ep ring is empty when ep is scheduled to RX */
COMP_OVERRUN, /* 15 */
/* Virtual Function Event Ring Full Error */
COMP_VF_FULL, /* 16 */
/* Parameter Error - Context parameter is invalid */
COMP_EINVAL, /* 17 */
/* Bandwidth Overrun Error - isoc ep exceeded its allocated bandwidth */
COMP_BW_OVER,/* 18 */
/* Context State Error - illegal context state transition requested */
COMP_CTX_STATE,/* 19 */
/* No Ping Response Error - HC didn't get PING_RESPONSE in time to TX */
COMP_PING_ERR,/* 20 */
/* Event Ring is full */
COMP_ER_FULL,/* 21 */
/* Incompatible Device Error */
COMP_DEV_ERR,/* 22 */
/* Missed Service Error - HC couldn't service an isoc ep within interval */
COMP_MISSED_INT,/* 23 */
/* Successfully stopped command ring */
COMP_CMD_STOP, /* 24 */
/* Successfully aborted current command and stopped command ring */
COMP_CMD_ABORT, /* 25 */
/* Stopped - transfer was terminated by a stop endpoint command */
COMP_STOP,/* 26 */
/* Same as COMP_EP_STOPPED, but the transferred length in the event
* is invalid */
COMP_STOP_INVAL, /* 27*/
/* Control Abort Error - Debug Capability - control pipe aborted */
COMP_DBG_ABORT, /* 28 */
/* Max Exit Latency Too Large Error */
COMP_MEL_ERR,/* 29 */
/* TRB type 30 reserved */
/* Isoc Buffer Overrun - an isoc IN ep sent more data than could fit in TD */
COMP_BUFF_OVER = 31,
/* Event Lost Error - xHC has an "internal event overrun condition" */
COMP_ISSUES, /* 32 */
/* Undefined Error - reported when other error codes don't apply */
COMP_UNKNOWN, /* 33 */
/* Invalid Stream ID Error */
COMP_STRID_ERR, /* 34 */
/* Secondary Bandwidth Error - may be returned by a Configure Endpoint cmd */
COMP_2ND_BW_ERR, /* 35 */
/* Split Transaction Error */
COMP_SPLIT_ERR /* 36 */
+} xhci_comp_code;
+struct xhci_link_trb {
/* 64-bit segment pointer*/
volatile __le64 segment_ptr;
volatile __le32 intr_target;
volatile __le32 control;
+};
+/* control bitfields */ +#define LINK_TOGGLE (0x1 << 1)
+/* Command completion event TRB */ +struct xhci_event_cmd {
/* Pointer to command TRB, or the value passed by the event data trb */
volatile __le64 cmd_trb;
volatile __le32 status;
volatile __le32 flags;
+};
+/* flags bitmasks */ +/* bits 16:23 are the virtual function ID */ +/* bits 24:31 are the slot ID */ +#define TRB_TO_SLOT_ID(p) (((p) & (0xff << 24)) >> 24) +#define TRB_TO_SLOT_ID_SHIFT (24) +#define TRB_TO_SLOT_ID_MASK (0xff << TRB_TO_SLOT_ID_SHIFT) +#define SLOT_ID_FOR_TRB(p) (((p) & 0xff) << 24) +#define SLOT_ID_FOR_TRB_MASK (0xff) +#define SLOT_ID_FOR_TRB_SHIFT (24)
+/* Stop Endpoint TRB - ep_index to endpoint ID for this TRB */ +#define TRB_TO_EP_INDEX(p) ((((p) & (0x1f << 16)) >> 16) - 1) +#define EP_ID_FOR_TRB(p) ((((p) + 1) & 0x1f) << 16)
+#define SUSPEND_PORT_FOR_TRB(p) (((p) & 1) << 23) +#define TRB_TO_SUSPEND_PORT(p) (((p) & (1 << 23)) >> 23) +#define LAST_EP_INDEX 30
+/* Set TR Dequeue Pointer command TRB fields */ +#define TRB_TO_STREAM_ID(p) ((((p) & (0xffff << 16)) >> 16)) +#define STREAM_ID_FOR_TRB(p) ((((p)) & 0xffff) << 16)
+/* Port Status Change Event TRB fields */ +/* Port ID - bits 31:24 */ +#define GET_PORT_ID(p) (((p) & (0xff << 24)) >> 24) +#define PORT_ID_SHIFT (24) +#define PORT_ID_MASK (0xff << PORT_ID_SHIFT)
+/* Normal TRB fields */ +/* transfer_len bitmasks - bits 0:16 */ +#define TRB_LEN(p) ((p) & 0x1ffff) +#define TRB_LEN_MASK (0x1ffff) +/* Interrupter Target - which MSI-X vector to target the completion event at */ +#define TRB_INTR_TARGET_SHIFT (22) +#define TRB_INTR_TARGET_MASK (0x3ff) +#define TRB_INTR_TARGET(p) (((p) & 0x3ff) << 22) +#define GET_INTR_TARGET(p) (((p) >> 22) & 0x3ff) +#define TRB_TBC(p) (((p) & 0x3) << 7) +#define TRB_TLBPC(p) (((p) & 0xf) << 16)
+/* Cycle bit - indicates TRB ownership by HC or HCD */ +#define TRB_CYCLE (1<<0) +/*
- Force next event data TRB to be evaluated before task switch.
- Used to pass OS data back after a TD completes.
- */
+#define TRB_ENT (1<<1) +/* Interrupt on short packet */ +#define TRB_ISP (1<<2) +/* Set PCIe no snoop attribute */ +#define TRB_NO_SNOOP (1<<3) +/* Chain multiple TRBs into a TD */ +#define TRB_CHAIN (1<<4) +/* Interrupt on completion */ +#define TRB_IOC (1<<5) +/* The buffer pointer contains immediate data */ +#define TRB_IDT (1<<6)
+/* Block Event Interrupt */ +#define TRB_BEI (1<<9)
+/* Control transfer TRB specific fields */ +#define TRB_DIR_IN (1<<16) +#define TRB_TX_TYPE(p) ((p) << 16) +#define TRB_TX_TYPE_SHIFT (16) +#define TRB_DATA_OUT 2 +#define TRB_DATA_IN 3
+/* Isochronous TRB specific fields */ +#define TRB_SIA (1 << 31)
+struct xhci_generic_trb {
volatile __le32 field[4];
+};
+union xhci_trb {
struct xhci_link_trb link;
struct xhci_transfer_event trans_event;
struct xhci_event_cmd event_cmd;
struct xhci_generic_trb generic;
+};
+/* TRB bit mask */ +#define TRB_TYPE_BITMASK (0xfc00) +#define TRB_TYPE(p) ((p) << 10) +#define TRB_TYPE_SHIFT (10) +#define TRB_FIELD_TO_TYPE(p) (((p) & TRB_TYPE_BITMASK) >> 10)
+/* TRB type IDs */ +typedef enum {
/* bulk, interrupt, isoc scatter/gather, and control data stage */
TRB_NORMAL = 1,
/* setup stage for control transfers */
TRB_SETUP, /* 2 */
/* data stage for control transfers */
TRB_DATA, /* 3 */
/* status stage for control transfers */
TRB_STATUS, /* 4 */
/* isoc transfers */
TRB_ISOC, /* 5 */
/* TRB for linking ring segments */
TRB_LINK, /* 6 */
/* TRB for EVENT DATA */
TRB_EVENT_DATA, /* 7 */
/* Transfer Ring No-op (not for the command ring) */
TRB_TR_NOOP, /* 8 */
/* Command TRBs */
/* Enable Slot Command */
TRB_ENABLE_SLOT, /* 9 */
/* Disable Slot Command */
TRB_DISABLE_SLOT, /* 10 */
/* Address Device Command */
TRB_ADDR_DEV, /* 11 */
/* Configure Endpoint Command */
TRB_CONFIG_EP, /* 12 */
/* Evaluate Context Command */
TRB_EVAL_CONTEXT, /* 13 */
/* Reset Endpoint Command */
TRB_RESET_EP, /* 14 */
/* Stop Transfer Ring Command */
TRB_STOP_RING, /* 15 */
/* Set Transfer Ring Dequeue Pointer Command */
TRB_SET_DEQ, /* 16 */
/* Reset Device Command */
TRB_RESET_DEV, /* 17 */
/* Force Event Command (opt) */
TRB_FORCE_EVENT, /* 18 */
/* Negotiate Bandwidth Command (opt) */
TRB_NEG_BANDWIDTH, /* 19 */
/* Set Latency Tolerance Value Command (opt) */
TRB_SET_LT, /* 20 */
/* Get port bandwidth Command */
TRB_GET_BW, /* 21 */
/* Force Header Command - generate a transaction or link management packet */
TRB_FORCE_HEADER, /* 22 */
/* No-op Command - not for transfer rings */
TRB_CMD_NOOP, /* 23 */
/* TRB IDs 24-31 reserved */
/* Event TRBS */
/* Transfer Event */
TRB_TRANSFER = 32,
/* Command Completion Event */
TRB_COMPLETION, /* 33 */
/* Port Status Change Event */
TRB_PORT_STATUS, /* 34 */
/* Bandwidth Request Event (opt) */
TRB_BANDWIDTH_EVENT, /* 35 */
/* Doorbell Event (opt) */
TRB_DOORBELL, /* 36 */
/* Host Controller Event */
TRB_HC_EVENT, /* 37 */
/* Device Notification Event - device sent function wake notification */
TRB_DEV_NOTE, /* 38 */
/* MFINDEX Wrap Event - microframe counter wrapped */
TRB_MFINDEX_WRAP, /* 39 */
/* TRB IDs 40-47 reserved, 48-63 is vendor-defined */
/* Nec vendor-specific command completion event. */
TRB_NEC_CMD_COMP = 48, /* 48 */
/* Get NEC firmware revision. */
TRB_NEC_GET_FW, /* 49 */
+} trb_type;
+#define TRB_TYPE_LINK(x) (((x) & TRB_TYPE_BITMASK) == TRB_TYPE(TRB_LINK)) +/* Above, but for __le32 types -- can avoid work by swapping constants: */ +#define TRB_TYPE_LINK_LE32(x) (((x) & cpu_to_le32(TRB_TYPE_BITMASK)) == \
cpu_to_le32(TRB_TYPE(TRB_LINK)))
+#define TRB_TYPE_NOOP_LE32(x) (((x) & cpu_to_le32(TRB_TYPE_BITMASK)) == \
cpu_to_le32(TRB_TYPE(TRB_TR_NOOP)))
+/*
- TRBS_PER_SEGMENT must be a multiple of 4,
- since the command ring is 64-byte aligned.
- It must also be greater than 16.
- */
+#define TRBS_PER_SEGMENT 64 +/* Allow two commands + a link TRB, along with any reserved command TRBs */ +#define MAX_RSVD_CMD_TRBS (TRBS_PER_SEGMENT - 3) +#define SEGMENT_SIZE (TRBS_PER_SEGMENT*16) +/* SEGMENT_SHIFT should be log2(SEGMENT_SIZE).
- Change this if you change TRBS_PER_SEGMENT!
- */
+#define SEGMENT_SHIFT 10 +/* TRB buffer pointers can't cross 64KB boundaries */ +#define TRB_MAX_BUFF_SHIFT 16 +#define TRB_MAX_BUFF_SIZE (1 << TRB_MAX_BUFF_SHIFT)
+struct xhci_segment {
union xhci_trb *trbs;
/* private to HCD */
struct xhci_segment *next;
+};
+struct xhci_ring {
struct xhci_segment *first_seg;
union xhci_trb *enqueue;
struct xhci_segment *enq_seg;
union xhci_trb *dequeue;
struct xhci_segment *deq_seg;
/*
* Write the cycle state into the TRB cycle field to give ownership of
* the TRB to the host controller (if we are the producer), or to check
* if we own the TRB (if we are the consumer). See section 4.9.1.
*/
volatile u32 cycle_state;
unsigned int num_segs;
+};
+struct xhci_erst_entry {
/* 64-bit event ring segment address */
__le64 seg_addr;
__le32 seg_size;
/* Set to zero */
__le32 rsvd;
+};
+struct xhci_erst {
struct xhci_erst_entry *entries;
unsigned int num_entries;
/* Num entries the ERST can contain */
unsigned int erst_size;
+};
+/*
- Each segment table entry is 4*32bits long. 1K seems like an ok size:
- (1K bytes * 8bytes/bit) / (4*32 bits) = 64 segment entries in the table,
- meaning 64 ring segments.
- Initial allocated size of the ERST, in number of entries */
+#define ERST_NUM_SEGS 3 +/* Initial number of event segment rings allocated */ +#define ERST_ENTRIES 3 +/* Initial allocated size of the ERST, in number of entries */ +#define ERST_SIZE 64 +/* Poll every 60 seconds */ +#define POLL_TIMEOUT 60 +/* Stop endpoint command timeout (secs) for URB cancellation watchdog timer */ +#define XHCI_STOP_EP_CMD_TIMEOUT 5 +/* XXX: Make these module parameters */
+struct xhci_virt_ep {
struct xhci_ring *ring;
unsigned int ep_state;
+#define SET_DEQ_PENDING (1 << 0) +#define EP_HALTED (1 << 1) /* For stall handling */ +#define EP_HALT_PENDING (1 << 2) /* For URB cancellation */ +/* Transitioning the endpoint to using streams, don't enqueue URBs */ +#define EP_GETTING_STREAMS (1 << 3) +#define EP_HAS_STREAMS (1 << 4) +/* Transitioning the endpoint to not using streams, don't enqueue URBs */ +#define EP_GETTING_NO_STREAMS (1 << 5) +};
+#define CTX_SIZE(_hcc) (HCC_64BYTE_CONTEXT(_hcc) ? 64 : 32)
+struct xhci_virt_device {
struct usb_device *udev;
/*
* Commands to the hardware are passed an "input context" that
* tells the hardware what to change in its data structures.
* The hardware will return changes in an "output context" that
* software must allocate for the hardware. We need to keep
* track of input and output contexts separately because
* these commands might fail and we don't trust the hardware.
*/
struct xhci_container_ctx *out_ctx;
/* Used for addressing devices and configuration changes */
struct xhci_container_ctx *in_ctx;
/* Rings saved to ensure old alt settings can be re-instated */
+#define XHCI_MAX_RINGS_CACHED 31
struct xhci_virt_ep eps[31];
+};
+/* TODO: copied from ehci.h - can be refactored? */ +/* xHCI spec says all registers are little endian */ +static inline unsigned int xhci_readl(uint32_t volatile *regs) +{
return readl(regs);
+}
+static inline void xhci_writel(uint32_t volatile *regs, const unsigned int val) +{
writel(val, regs);
+}
+/*
- Registers should always be accessed with double word or quad word accesses.
- Some xHCI implementations may support 64-bit address pointers. Registers
- with 64-bit address pointers should be written to with dword accesses by
- writing the low dword first (ptr[0]), then the high dword (ptr[1]) second.
- xHCI implementations that do not support 64-bit address pointers will ignore
- the high dword, and write order is irrelevant.
- */
+static inline u64 xhci_readq(__le64 volatile *regs) +{
__u32 *ptr = (__u32 *)regs;
u64 val_lo = readl(ptr);
u64 val_hi = readl(ptr + 1);
return val_lo + (val_hi << 32);
+}
+static inline void xhci_writeq(__le64 volatile *regs, const u64 val) +{
__u32 *ptr = (__u32 *)regs;
u32 val_lo = lower_32_bits(val);
/* FIXME */
u32 val_hi = 0;
writel(val_lo, ptr);
writel(val_hi, ptr + 1);
+}
+int xhci_hcd_init(int index, struct xhci_hccr **ret_hccr,
struct xhci_hcor **ret_hcor);
+void xhci_hcd_stop(int index);
+/*************************************************************
EXTENDED CAPABILITY DEFINITIONS
+*************************************************************/ +/* Up to 16 ms to halt an HC */ +#define XHCI_MAX_HALT_USEC (16*1000) +/* HC not running - set to 1 when run/stop bit is cleared. */ +#define XHCI_STS_HALT (1 << 0)
+/* HCCPARAMS offset from PCI base address */ +#define XHCI_HCC_PARAMS_OFFSET 0x10 +/* HCCPARAMS contains the first extended capability pointer */ +#define XHCI_HCC_EXT_CAPS(p) (((p)>>16)&0xffff)
+/* Command and Status registers offset from the Operational Registers address */ +#define XHCI_CMD_OFFSET 0x00 +#define XHCI_STS_OFFSET 0x04
+#define XHCI_MAX_EXT_CAPS 50
+/* Capability Register */ +/* bits 7:0 - how long is the Capabilities register */ +#define XHCI_HC_LENGTH(p) (((p) >> 00) & 0x00ff)
+/* Extended capability register fields */ +#define XHCI_EXT_CAPS_ID(p) (((p) >> 0) & 0xff) +#define XHCI_EXT_CAPS_NEXT(p) (((p) >> 8) & 0xff) +#define XHCI_EXT_CAPS_VAL(p) ((p) >> 16) +/* Extended capability IDs - ID 0 reserved */ +#define XHCI_EXT_CAPS_LEGACY 1 +#define XHCI_EXT_CAPS_PROTOCOL 2 +#define XHCI_EXT_CAPS_PM 3 +#define XHCI_EXT_CAPS_VIRT 4 +#define XHCI_EXT_CAPS_ROUTE 5 +/* IDs 6-9 reserved */ +#define XHCI_EXT_CAPS_DEBUG 10 +/* USB Legacy Support Capability - section 7.1.1 */ +#define XHCI_HC_BIOS_OWNED (1 << 16) +#define XHCI_HC_OS_OWNED (1 << 24)
+/* USB Legacy Support Capability - section 7.1.1 */ +/* Add this offset, plus the value of xECP in HCCPARAMS to the base address */ +#define XHCI_LEGACY_SUPPORT_OFFSET (0x00)
+/* USB Legacy Support Control and Status Register - section 7.1.2 */ +/* Add this offset, plus the value of xECP in HCCPARAMS to the base address */ +#define XHCI_LEGACY_CONTROL_OFFSET (0x04) +/* bits 1:2, 5:12, and 17:19 need to be preserved; bits 21:28 should be zero */ +#define XHCI_LEGACY_DISABLE_SMI ((0x3 << 1) + (0xff << 5) + (0x7 << 17))
+/* USB 2.0 xHCI 0.96 L1C capability - section 7.2.2.1.3.2 */ +#define XHCI_L1C (1 << 16)
+/* USB 2.0 xHCI 1.0 hardware LMP capability - section 7.2.2.1.3.2 */ +#define XHCI_HLC (1 << 19)
+/* command register values to disable interrupts and halt the HC */ +/* start/stop HC execution - do not write unless HC is halted*/ +#define XHCI_CMD_RUN (1 << 0) +/* Event Interrupt Enable - get irq when EINT bit is set in USBSTS register */ +#define XHCI_CMD_EIE (1 << 2) +/* Host System Error Interrupt Enable - get irq when HSEIE bit set in USBSTS */ +#define XHCI_CMD_HSEIE (1 << 3) +/* Enable Wrap Event - '1' means xHC generates an event when MFINDEX wraps. */ +#define XHCI_CMD_EWE (1 << 10)
+#define XHCI_IRQS (XHCI_CMD_EIE | XHCI_CMD_HSEIE | XHCI_CMD_EWE)
+/* true: Controller Not Ready to accept doorbell or op reg writes after reset */ +#define XHCI_STS_CNR (1 << 11)
+struct xhci_ctrl {
struct xhci_hccr *hccr; /* R/O registers, not need for volatile */
struct xhci_hcor *hcor;
struct xhci_doorbell_array *dba;
struct xhci_run_regs *run_regs;
struct xhci_device_context_array *dcbaa \
__attribute__ ((aligned(ARCH_DMA_MINALIGN)));
struct xhci_ring *event_ring;
struct xhci_ring *cmd_ring;
struct xhci_ring *transfer_ring;
struct xhci_segment *seg;
struct xhci_intr_reg *ir_set;
struct xhci_erst erst;
struct xhci_erst_entry entry[ERST_NUM_SEGS];
struct xhci_virt_device *devs[MAX_HC_SLOTS];
int rootdev;
+};
+unsigned long trb_addr(struct xhci_segment *seg, union xhci_trb *trb); +struct xhci_input_control_ctx
*xhci_get_input_control_ctx(struct xhci_container_ctx *ctx);
+struct xhci_slot_ctx *xhci_get_slot_ctx(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *ctx);
+struct xhci_ep_ctx *xhci_get_ep_ctx(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *ctx,
unsigned int ep_index);
+void xhci_endpoint_copy(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *in_ctx,
struct xhci_container_ctx *out_ctx,
unsigned int ep_index);
+void xhci_slot_copy(struct xhci_ctrl *ctrl,
struct xhci_container_ctx *in_ctx,
struct xhci_container_ctx *out_ctx);
+void xhci_setup_addressable_virt_dev(struct usb_device *udev); +void xhci_queue_command(struct xhci_ctrl *ctrl, u8 *ptr,
u32 slot_id, u32 ep_index, trb_type cmd);
+void xhci_acknowledge_event(struct xhci_ctrl *ctrl); +union xhci_trb *xhci_wait_for_event(struct xhci_ctrl *ctrl, trb_type expected); +int xhci_bulk_tx(struct usb_device *udev, unsigned long pipe,
int length, void *buffer);
+int xhci_ctrl_tx(struct usb_device *udev, unsigned long pipe,
struct devrequest *req, int length, void *buffer);
+int xhci_check_maxpacket(struct usb_device *udev); +void xhci_flush_cache(uint32_t addr, u32 type_len); +void xhci_inval_cache(uint32_t addr, u32 type_len); +void xhci_cleanup(struct xhci_ctrl *ctrl); +struct xhci_ring *xhci_ring_alloc(unsigned int num_segs, bool link_trbs); +int xhci_alloc_virt_device(struct usb_device *udev); +int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
struct xhci_hcor *hcor);
+#endif /* HOST_XHCI_H_ */ diff --git a/include/usb.h b/include/usb.h index 60db897..a96ec23 100644 --- a/include/usb.h +++ b/include/usb.h @@ -125,6 +125,8 @@ struct usb_device { struct usb_device *children[USB_MAXCHILDREN];
void *controller; /* hardware controller private data */
/* slot_id - for xHCI enabled devices */
unsigned int slot_id;
};
/********************************************************************** @@ -138,7 +140,7 @@ struct usb_device { defined(CONFIG_USB_OMAP3) || defined(CONFIG_USB_DA8XX) || \ defined(CONFIG_USB_BLACKFIN) || defined(CONFIG_USB_AM35X) || \ defined(CONFIG_USB_MUSB_DSPS) || defined(CONFIG_USB_MUSB_AM35X) || \
defined(CONFIG_USB_MUSB_OMAP2PLUS)
defined(CONFIG_USB_MUSB_OMAP2PLUS) || defined(CONFIG_USB_XHCI)
int usb_lowlevel_init(int index, void **controller); int usb_lowlevel_stop(int index); @@ -338,6 +340,10 @@ int usb_set_interface(struct usb_device *dev, int interface, int alternate); #define usb_pipecontrol(pipe) (usb_pipetype((pipe)) == PIPE_CONTROL) #define usb_pipebulk(pipe) (usb_pipetype((pipe)) == PIPE_BULK)
+#define usb_pipe_ep_index(pipe) \
usb_pipecontrol(pipe) ? (usb_pipeendpoint(pipe) * 2) : \
((usb_pipeendpoint(pipe) * 2) - \
(usb_pipein(pipe) ? 0 : 1))
/*************************************************************************
- Hub Stuff
@@ -382,5 +388,6 @@ struct usb_device *usb_alloc_new_device(void *controller);
int usb_new_device(struct usb_device *dev); void usb_free_device(void); +int usb_alloc_device(struct usb_device *dev);
#endif /*_USB_H_ */
Various Checkpatch errors, warnings and checks
total: 1 errors, 56 warnings, 48 checks, 4103 lines checked
--
Dan Murphy
U-Boot mailing list U-Boot@lists.denx.de http://lists.denx.de/mailman/listinfo/u-boot