mirror of
https://github.com/torvalds/linux.git
synced 2026-05-04 14:32:27 -04:00
This was done entirely with mindless brute force, using
git grep -l '\<k[vmz]*alloc_objs*(.*, GFP_KERNEL)' |
xargs sed -i 's/\(alloc_objs*(.*\), GFP_KERNEL)/\1)/'
to convert the new alloc_obj() users that had a simple GFP_KERNEL
argument to just drop that argument.
Note that due to the extreme simplicity of the scripting, any slightly
more complex cases spread over multiple lines would not be triggered:
they definitely exist, but this covers the vast bulk of the cases, and
the resulting diff is also then easier to check automatically.
For the same reason the 'flex' versions will be done as a separate
conversion.
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
1561 lines
35 KiB
C
1561 lines
35 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* xhci-dbgcap.c - xHCI debug capability support
|
|
*
|
|
* Copyright (C) 2017 Intel Corporation
|
|
*
|
|
* Author: Lu Baolu <baolu.lu@linux.intel.com>
|
|
*/
|
|
#include <linux/bug.h>
|
|
#include <linux/device.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/kstrtox.h>
|
|
#include <linux/list.h>
|
|
#include <linux/nls.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/string.h>
|
|
#include <linux/sysfs.h>
|
|
#include <linux/types.h>
|
|
#include <linux/workqueue.h>
|
|
|
|
#include <linux/io-64-nonatomic-lo-hi.h>
|
|
|
|
#include <asm/byteorder.h>
|
|
|
|
#include "xhci.h"
|
|
#include "xhci-trace.h"
|
|
#include "xhci-dbgcap.h"
|
|
|
|
static const struct dbc_str dbc_str_default = {
|
|
.manufacturer = "Linux Foundation",
|
|
.product = "Linux USB Debug Target",
|
|
.serial = "0001",
|
|
};
|
|
|
|
static void dbc_free_ctx(struct device *dev, struct xhci_container_ctx *ctx)
|
|
{
|
|
if (!ctx)
|
|
return;
|
|
dma_free_coherent(dev, ctx->size, ctx->bytes, ctx->dma);
|
|
kfree(ctx);
|
|
}
|
|
|
|
/* we use only one segment for DbC rings */
|
|
static void dbc_ring_free(struct device *dev, struct xhci_ring *ring)
|
|
{
|
|
if (!ring)
|
|
return;
|
|
|
|
if (ring->first_seg) {
|
|
dma_free_coherent(dev, TRB_SEGMENT_SIZE,
|
|
ring->first_seg->trbs,
|
|
ring->first_seg->dma);
|
|
kfree(ring->first_seg);
|
|
}
|
|
kfree(ring);
|
|
}
|
|
|
|
static void xhci_dbc_init_ep_contexts(struct xhci_dbc *dbc)
|
|
{
|
|
struct xhci_ep_ctx *ep_ctx;
|
|
unsigned int max_burst;
|
|
dma_addr_t deq;
|
|
|
|
max_burst = DBC_CTRL_MAXBURST(readl(&dbc->regs->control));
|
|
|
|
/* Populate bulk out endpoint context: */
|
|
ep_ctx = dbc_bulkout_ctx(dbc);
|
|
deq = dbc_bulkout_enq(dbc);
|
|
ep_ctx->ep_info = 0;
|
|
ep_ctx->ep_info2 = dbc_epctx_info2(BULK_OUT_EP, 1024, max_burst);
|
|
ep_ctx->deq = cpu_to_le64(deq | dbc->ring_out->cycle_state);
|
|
|
|
/* Populate bulk in endpoint context: */
|
|
ep_ctx = dbc_bulkin_ctx(dbc);
|
|
deq = dbc_bulkin_enq(dbc);
|
|
ep_ctx->ep_info = 0;
|
|
ep_ctx->ep_info2 = dbc_epctx_info2(BULK_IN_EP, 1024, max_burst);
|
|
ep_ctx->deq = cpu_to_le64(deq | dbc->ring_in->cycle_state);
|
|
}
|
|
|
|
static u8 get_str_desc_len(const char *desc)
|
|
{
|
|
return ((struct usb_string_descriptor *)desc)->bLength;
|
|
}
|
|
|
|
static u32 dbc_prepare_info_context_str_len(struct dbc_str_descs *descs)
|
|
{
|
|
u32 len;
|
|
|
|
len = get_str_desc_len(descs->serial);
|
|
len <<= 8;
|
|
len += get_str_desc_len(descs->product);
|
|
len <<= 8;
|
|
len += get_str_desc_len(descs->manufacturer);
|
|
len <<= 8;
|
|
len += get_str_desc_len(descs->string0);
|
|
|
|
return len;
|
|
}
|
|
|
|
static int xhci_dbc_populate_str_desc(char *desc, const char *src)
|
|
{
|
|
struct usb_string_descriptor *s_desc;
|
|
int len;
|
|
|
|
s_desc = (struct usb_string_descriptor *)desc;
|
|
|
|
/* len holds number of 2 byte UTF-16 characters */
|
|
len = utf8s_to_utf16s(src, strlen(src), UTF16_LITTLE_ENDIAN,
|
|
(wchar_t *)s_desc->wData, USB_MAX_STRING_LEN * 2);
|
|
if (len < 0)
|
|
return len;
|
|
|
|
s_desc->bLength = len * 2 + 2;
|
|
s_desc->bDescriptorType = USB_DT_STRING;
|
|
|
|
return s_desc->bLength;
|
|
}
|
|
|
|
static void xhci_dbc_populate_str_descs(struct dbc_str_descs *str_descs,
|
|
struct dbc_str *str)
|
|
{
|
|
/* Serial string: */
|
|
xhci_dbc_populate_str_desc(str_descs->serial, str->serial);
|
|
|
|
/* Product string: */
|
|
xhci_dbc_populate_str_desc(str_descs->product, str->product);
|
|
|
|
/* Manufacturer string: */
|
|
xhci_dbc_populate_str_desc(str_descs->manufacturer, str->manufacturer);
|
|
|
|
/* String0: */
|
|
str_descs->string0[0] = 4;
|
|
str_descs->string0[1] = USB_DT_STRING;
|
|
str_descs->string0[2] = 0x09;
|
|
str_descs->string0[3] = 0x04;
|
|
}
|
|
|
|
static void xhci_dbc_init_contexts(struct xhci_dbc *dbc)
|
|
{
|
|
struct dbc_info_context *info;
|
|
u32 dev_info;
|
|
dma_addr_t dma;
|
|
|
|
if (!dbc)
|
|
return;
|
|
|
|
/* Populate info Context: */
|
|
info = (struct dbc_info_context *)dbc->ctx->bytes;
|
|
dma = dbc->str_descs_dma;
|
|
info->string0 = cpu_to_le64(dma);
|
|
info->manufacturer = cpu_to_le64(dma + USB_MAX_STRING_DESC_LEN);
|
|
info->product = cpu_to_le64(dma + USB_MAX_STRING_DESC_LEN * 2);
|
|
info->serial = cpu_to_le64(dma + USB_MAX_STRING_DESC_LEN * 3);
|
|
info->length = cpu_to_le32(dbc_prepare_info_context_str_len(dbc->str_descs));
|
|
|
|
/* Populate bulk in and out endpoint contexts: */
|
|
xhci_dbc_init_ep_contexts(dbc);
|
|
|
|
/* Set DbC context and info registers: */
|
|
lo_hi_writeq(dbc->ctx->dma, &dbc->regs->dccp);
|
|
|
|
dev_info = (dbc->idVendor << 16) | dbc->bInterfaceProtocol;
|
|
writel(dev_info, &dbc->regs->devinfo1);
|
|
|
|
dev_info = (dbc->bcdDevice << 16) | dbc->idProduct;
|
|
writel(dev_info, &dbc->regs->devinfo2);
|
|
}
|
|
|
|
static void xhci_dbc_giveback(struct dbc_request *req, int status)
|
|
__releases(&dbc->lock)
|
|
__acquires(&dbc->lock)
|
|
{
|
|
struct xhci_dbc *dbc = req->dbc;
|
|
struct device *dev = dbc->dev;
|
|
|
|
list_del_init(&req->list_pending);
|
|
req->trb_dma = 0;
|
|
req->trb = NULL;
|
|
|
|
if (req->status == -EINPROGRESS)
|
|
req->status = status;
|
|
|
|
trace_xhci_dbc_giveback_request(req);
|
|
|
|
dma_unmap_single(dev,
|
|
req->dma,
|
|
req->length,
|
|
dbc_ep_dma_direction(req));
|
|
|
|
/* Give back the transfer request: */
|
|
spin_unlock(&dbc->lock);
|
|
req->complete(dbc, req);
|
|
spin_lock(&dbc->lock);
|
|
}
|
|
|
|
static void trb_to_noop(union xhci_trb *trb)
|
|
{
|
|
trb->generic.field[0] = 0;
|
|
trb->generic.field[1] = 0;
|
|
trb->generic.field[2] = 0;
|
|
trb->generic.field[3] &= cpu_to_le32(TRB_CYCLE);
|
|
trb->generic.field[3] |= cpu_to_le32(TRB_TYPE(TRB_TR_NOOP));
|
|
}
|
|
|
|
static void xhci_dbc_flush_single_request(struct dbc_request *req)
|
|
{
|
|
trb_to_noop(req->trb);
|
|
xhci_dbc_giveback(req, -ESHUTDOWN);
|
|
}
|
|
|
|
static void xhci_dbc_flush_endpoint_requests(struct dbc_ep *dep)
|
|
{
|
|
struct dbc_request *req, *tmp;
|
|
|
|
list_for_each_entry_safe(req, tmp, &dep->list_pending, list_pending)
|
|
xhci_dbc_flush_single_request(req);
|
|
}
|
|
|
|
static void xhci_dbc_flush_requests(struct xhci_dbc *dbc)
|
|
{
|
|
xhci_dbc_flush_endpoint_requests(&dbc->eps[BULK_OUT]);
|
|
xhci_dbc_flush_endpoint_requests(&dbc->eps[BULK_IN]);
|
|
}
|
|
|
|
struct dbc_request *
|
|
dbc_alloc_request(struct xhci_dbc *dbc, unsigned int direction, gfp_t flags)
|
|
{
|
|
struct dbc_request *req;
|
|
|
|
if (direction != BULK_IN &&
|
|
direction != BULK_OUT)
|
|
return NULL;
|
|
|
|
if (!dbc)
|
|
return NULL;
|
|
|
|
req = kzalloc_obj(*req, flags);
|
|
if (!req)
|
|
return NULL;
|
|
|
|
req->dbc = dbc;
|
|
INIT_LIST_HEAD(&req->list_pending);
|
|
INIT_LIST_HEAD(&req->list_pool);
|
|
req->direction = direction;
|
|
|
|
trace_xhci_dbc_alloc_request(req);
|
|
|
|
return req;
|
|
}
|
|
|
|
void
|
|
dbc_free_request(struct dbc_request *req)
|
|
{
|
|
trace_xhci_dbc_free_request(req);
|
|
|
|
kfree(req);
|
|
}
|
|
|
|
static void
|
|
xhci_dbc_queue_trb(struct xhci_ring *ring, u32 field1,
|
|
u32 field2, u32 field3, u32 field4)
|
|
{
|
|
union xhci_trb *trb, *next;
|
|
|
|
trb = ring->enqueue;
|
|
trb->generic.field[0] = cpu_to_le32(field1);
|
|
trb->generic.field[1] = cpu_to_le32(field2);
|
|
trb->generic.field[2] = cpu_to_le32(field3);
|
|
trb->generic.field[3] = cpu_to_le32(field4);
|
|
|
|
trace_xhci_dbc_gadget_ep_queue(ring, &trb->generic,
|
|
xhci_trb_virt_to_dma(ring->enq_seg,
|
|
ring->enqueue));
|
|
ring->num_trbs_free--;
|
|
next = ++(ring->enqueue);
|
|
if (TRB_TYPE_LINK_LE32(next->link.control)) {
|
|
next->link.control ^= cpu_to_le32(TRB_CYCLE);
|
|
ring->enqueue = ring->enq_seg->trbs;
|
|
ring->cycle_state ^= 1;
|
|
}
|
|
}
|
|
|
|
static int xhci_dbc_queue_bulk_tx(struct dbc_ep *dep,
|
|
struct dbc_request *req)
|
|
{
|
|
u64 addr;
|
|
union xhci_trb *trb;
|
|
unsigned int num_trbs;
|
|
struct xhci_dbc *dbc = req->dbc;
|
|
struct xhci_ring *ring = dep->ring;
|
|
u32 length, control, cycle;
|
|
|
|
num_trbs = count_trbs(req->dma, req->length);
|
|
WARN_ON(num_trbs != 1);
|
|
if (ring->num_trbs_free < num_trbs)
|
|
return -EBUSY;
|
|
|
|
addr = req->dma;
|
|
trb = ring->enqueue;
|
|
cycle = ring->cycle_state;
|
|
length = TRB_LEN(req->length);
|
|
control = TRB_TYPE(TRB_NORMAL) | TRB_IOC;
|
|
|
|
if (cycle)
|
|
control &= cpu_to_le32(~TRB_CYCLE);
|
|
else
|
|
control |= cpu_to_le32(TRB_CYCLE);
|
|
|
|
req->trb = ring->enqueue;
|
|
req->trb_dma = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue);
|
|
xhci_dbc_queue_trb(ring,
|
|
lower_32_bits(addr),
|
|
upper_32_bits(addr),
|
|
length, control);
|
|
|
|
/*
|
|
* Add a barrier between writes of trb fields and flipping
|
|
* the cycle bit:
|
|
*/
|
|
wmb();
|
|
|
|
if (cycle)
|
|
trb->generic.field[3] |= cpu_to_le32(TRB_CYCLE);
|
|
else
|
|
trb->generic.field[3] &= cpu_to_le32(~TRB_CYCLE);
|
|
|
|
writel(DBC_DOOR_BELL_TARGET(dep->direction), &dbc->regs->doorbell);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
dbc_ep_do_queue(struct dbc_request *req)
|
|
{
|
|
int ret;
|
|
struct xhci_dbc *dbc = req->dbc;
|
|
struct device *dev = dbc->dev;
|
|
struct dbc_ep *dep = &dbc->eps[req->direction];
|
|
|
|
if (!req->length || !req->buf)
|
|
return -EINVAL;
|
|
|
|
req->actual = 0;
|
|
req->status = -EINPROGRESS;
|
|
|
|
req->dma = dma_map_single(dev,
|
|
req->buf,
|
|
req->length,
|
|
dbc_ep_dma_direction(dep));
|
|
if (dma_mapping_error(dev, req->dma)) {
|
|
dev_err(dbc->dev, "failed to map buffer\n");
|
|
return -EFAULT;
|
|
}
|
|
|
|
ret = xhci_dbc_queue_bulk_tx(dep, req);
|
|
if (ret) {
|
|
dev_err(dbc->dev, "failed to queue trbs\n");
|
|
dma_unmap_single(dev,
|
|
req->dma,
|
|
req->length,
|
|
dbc_ep_dma_direction(dep));
|
|
return -EFAULT;
|
|
}
|
|
|
|
list_add_tail(&req->list_pending, &dep->list_pending);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int dbc_ep_queue(struct dbc_request *req)
|
|
{
|
|
unsigned long flags;
|
|
struct xhci_dbc *dbc = req->dbc;
|
|
int ret = -ESHUTDOWN;
|
|
|
|
if (!dbc)
|
|
return -ENODEV;
|
|
|
|
if (req->direction != BULK_IN &&
|
|
req->direction != BULK_OUT)
|
|
return -EINVAL;
|
|
|
|
spin_lock_irqsave(&dbc->lock, flags);
|
|
if (dbc->state == DS_CONFIGURED)
|
|
ret = dbc_ep_do_queue(req);
|
|
spin_unlock_irqrestore(&dbc->lock, flags);
|
|
|
|
mod_delayed_work(system_percpu_wq, &dbc->event_work, 0);
|
|
|
|
trace_xhci_dbc_queue_request(req);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline void xhci_dbc_do_eps_init(struct xhci_dbc *dbc, bool direction)
|
|
{
|
|
struct dbc_ep *dep;
|
|
|
|
dep = &dbc->eps[direction];
|
|
dep->dbc = dbc;
|
|
dep->direction = direction;
|
|
dep->ring = direction ? dbc->ring_in : dbc->ring_out;
|
|
|
|
INIT_LIST_HEAD(&dep->list_pending);
|
|
}
|
|
|
|
static void xhci_dbc_eps_init(struct xhci_dbc *dbc)
|
|
{
|
|
xhci_dbc_do_eps_init(dbc, BULK_OUT);
|
|
xhci_dbc_do_eps_init(dbc, BULK_IN);
|
|
}
|
|
|
|
static void xhci_dbc_eps_exit(struct xhci_dbc *dbc)
|
|
{
|
|
memset(dbc->eps, 0, sizeof_field(struct xhci_dbc, eps));
|
|
}
|
|
|
|
static int dbc_erst_alloc(struct device *dev, struct xhci_ring *evt_ring,
|
|
struct xhci_erst *erst, gfp_t flags)
|
|
{
|
|
erst->entries = dma_alloc_coherent(dev, sizeof(*erst->entries),
|
|
&erst->erst_dma_addr, flags);
|
|
if (!erst->entries)
|
|
return -ENOMEM;
|
|
|
|
erst->num_entries = 1;
|
|
erst->entries[0].seg_addr = cpu_to_le64(evt_ring->first_seg->dma);
|
|
erst->entries[0].seg_size = cpu_to_le32(TRBS_PER_SEGMENT);
|
|
erst->entries[0].rsvd = 0;
|
|
return 0;
|
|
}
|
|
|
|
static void dbc_erst_free(struct device *dev, struct xhci_erst *erst)
|
|
{
|
|
dma_free_coherent(dev, sizeof(*erst->entries), erst->entries,
|
|
erst->erst_dma_addr);
|
|
erst->entries = NULL;
|
|
}
|
|
|
|
static struct xhci_container_ctx *
|
|
dbc_alloc_ctx(struct device *dev, gfp_t flags)
|
|
{
|
|
struct xhci_container_ctx *ctx;
|
|
|
|
ctx = kzalloc_obj(*ctx, flags);
|
|
if (!ctx)
|
|
return NULL;
|
|
|
|
/* xhci 7.6.9, all three contexts; info, ep-out and ep-in. Each 64 bytes*/
|
|
ctx->size = 3 * DBC_CONTEXT_SIZE;
|
|
ctx->bytes = dma_alloc_coherent(dev, ctx->size, &ctx->dma, flags);
|
|
if (!ctx->bytes) {
|
|
kfree(ctx);
|
|
return NULL;
|
|
}
|
|
return ctx;
|
|
}
|
|
|
|
static void xhci_dbc_ring_init(struct xhci_ring *ring)
|
|
{
|
|
struct xhci_segment *seg = ring->first_seg;
|
|
|
|
/* clear all trbs on ring in case of old ring */
|
|
memset(seg->trbs, 0, TRB_SEGMENT_SIZE);
|
|
|
|
/* Only event ring does not use link TRB */
|
|
if (ring->type != TYPE_EVENT) {
|
|
union xhci_trb *trb = &seg->trbs[TRBS_PER_SEGMENT - 1];
|
|
|
|
trb->link.segment_ptr = cpu_to_le64(ring->first_seg->dma);
|
|
trb->link.control = cpu_to_le32(LINK_TOGGLE | TRB_TYPE(TRB_LINK));
|
|
}
|
|
xhci_initialize_ring_info(ring);
|
|
}
|
|
|
|
static int xhci_dbc_reinit_ep_rings(struct xhci_dbc *dbc)
|
|
{
|
|
struct xhci_ring *in_ring = dbc->eps[BULK_IN].ring;
|
|
struct xhci_ring *out_ring = dbc->eps[BULK_OUT].ring;
|
|
|
|
if (!in_ring || !out_ring || !dbc->ctx) {
|
|
dev_warn(dbc->dev, "Can't re-init unallocated endpoints\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
xhci_dbc_ring_init(in_ring);
|
|
xhci_dbc_ring_init(out_ring);
|
|
|
|
/* set ep context enqueue, dequeue, and cycle to initial values */
|
|
xhci_dbc_init_ep_contexts(dbc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct xhci_ring *
|
|
xhci_dbc_ring_alloc(struct device *dev, enum xhci_ring_type type, gfp_t flags)
|
|
{
|
|
struct xhci_ring *ring;
|
|
struct xhci_segment *seg;
|
|
dma_addr_t dma;
|
|
|
|
ring = kzalloc_obj(*ring, flags);
|
|
if (!ring)
|
|
return NULL;
|
|
|
|
ring->num_segs = 1;
|
|
ring->type = type;
|
|
|
|
seg = kzalloc_obj(*seg, flags);
|
|
if (!seg)
|
|
goto seg_fail;
|
|
|
|
ring->first_seg = seg;
|
|
ring->last_seg = seg;
|
|
seg->next = seg;
|
|
|
|
seg->trbs = dma_alloc_coherent(dev, TRB_SEGMENT_SIZE, &dma, flags);
|
|
if (!seg->trbs)
|
|
goto dma_fail;
|
|
|
|
seg->dma = dma;
|
|
|
|
INIT_LIST_HEAD(&ring->td_list);
|
|
|
|
xhci_dbc_ring_init(ring);
|
|
|
|
return ring;
|
|
dma_fail:
|
|
kfree(seg);
|
|
seg_fail:
|
|
kfree(ring);
|
|
return NULL;
|
|
}
|
|
|
|
static int xhci_dbc_mem_init(struct xhci_dbc *dbc, gfp_t flags)
|
|
{
|
|
int ret;
|
|
dma_addr_t deq;
|
|
struct device *dev = dbc->dev;
|
|
|
|
/* Allocate various rings for events and transfers: */
|
|
dbc->ring_evt = xhci_dbc_ring_alloc(dev, TYPE_EVENT, flags);
|
|
if (!dbc->ring_evt)
|
|
goto evt_fail;
|
|
|
|
dbc->ring_in = xhci_dbc_ring_alloc(dev, TYPE_BULK, flags);
|
|
if (!dbc->ring_in)
|
|
goto in_fail;
|
|
|
|
dbc->ring_out = xhci_dbc_ring_alloc(dev, TYPE_BULK, flags);
|
|
if (!dbc->ring_out)
|
|
goto out_fail;
|
|
|
|
/* Allocate and populate ERST: */
|
|
ret = dbc_erst_alloc(dev, dbc->ring_evt, &dbc->erst, flags);
|
|
if (ret)
|
|
goto erst_fail;
|
|
|
|
/* Allocate context data structure: */
|
|
dbc->ctx = dbc_alloc_ctx(dev, flags); /* was sysdev, and is still */
|
|
if (!dbc->ctx)
|
|
goto ctx_fail;
|
|
|
|
/* Allocate the string table: */
|
|
dbc->str_descs_size = sizeof(*dbc->str_descs);
|
|
dbc->str_descs = dma_alloc_coherent(dev, dbc->str_descs_size,
|
|
&dbc->str_descs_dma, flags);
|
|
if (!dbc->str_descs)
|
|
goto str_descs_fail;
|
|
|
|
/* Setup ERST register: */
|
|
writel(dbc->erst.num_entries, &dbc->regs->ersts);
|
|
|
|
lo_hi_writeq(dbc->erst.erst_dma_addr, &dbc->regs->erstba);
|
|
deq = xhci_trb_virt_to_dma(dbc->ring_evt->deq_seg,
|
|
dbc->ring_evt->dequeue);
|
|
lo_hi_writeq(deq, &dbc->regs->erdp);
|
|
|
|
/* Setup string descriptors and contexts: */
|
|
xhci_dbc_populate_str_descs(dbc->str_descs, &dbc->str);
|
|
xhci_dbc_init_contexts(dbc);
|
|
|
|
xhci_dbc_eps_init(dbc);
|
|
dbc->state = DS_INITIALIZED;
|
|
|
|
return 0;
|
|
|
|
str_descs_fail:
|
|
dbc_free_ctx(dev, dbc->ctx);
|
|
dbc->ctx = NULL;
|
|
ctx_fail:
|
|
dbc_erst_free(dev, &dbc->erst);
|
|
erst_fail:
|
|
dbc_ring_free(dev, dbc->ring_out);
|
|
dbc->ring_out = NULL;
|
|
out_fail:
|
|
dbc_ring_free(dev, dbc->ring_in);
|
|
dbc->ring_in = NULL;
|
|
in_fail:
|
|
dbc_ring_free(dev, dbc->ring_evt);
|
|
dbc->ring_evt = NULL;
|
|
evt_fail:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void xhci_dbc_mem_cleanup(struct xhci_dbc *dbc)
|
|
{
|
|
if (!dbc)
|
|
return;
|
|
|
|
xhci_dbc_eps_exit(dbc);
|
|
|
|
dma_free_coherent(dbc->dev, dbc->str_descs_size, dbc->str_descs, dbc->str_descs_dma);
|
|
dbc->str_descs = NULL;
|
|
|
|
dbc_free_ctx(dbc->dev, dbc->ctx);
|
|
dbc->ctx = NULL;
|
|
|
|
dbc_erst_free(dbc->dev, &dbc->erst);
|
|
dbc_ring_free(dbc->dev, dbc->ring_out);
|
|
dbc_ring_free(dbc->dev, dbc->ring_in);
|
|
dbc_ring_free(dbc->dev, dbc->ring_evt);
|
|
dbc->ring_in = NULL;
|
|
dbc->ring_out = NULL;
|
|
dbc->ring_evt = NULL;
|
|
}
|
|
|
|
static int xhci_do_dbc_start(struct xhci_dbc *dbc)
|
|
{
|
|
int ret;
|
|
u32 ctrl;
|
|
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EINVAL;
|
|
|
|
writel(0, &dbc->regs->control);
|
|
ret = xhci_handshake(&dbc->regs->control,
|
|
DBC_CTRL_DBC_ENABLE,
|
|
0, 1000);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = xhci_dbc_mem_init(dbc, GFP_ATOMIC);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ctrl = readl(&dbc->regs->control);
|
|
writel(ctrl | DBC_CTRL_DBC_ENABLE | DBC_CTRL_PORT_ENABLE,
|
|
&dbc->regs->control);
|
|
ret = xhci_handshake(&dbc->regs->control,
|
|
DBC_CTRL_DBC_ENABLE,
|
|
DBC_CTRL_DBC_ENABLE, 1000);
|
|
if (ret)
|
|
return ret;
|
|
|
|
dbc->state = DS_ENABLED;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int xhci_do_dbc_stop(struct xhci_dbc *dbc)
|
|
{
|
|
if (dbc->state == DS_DISABLED)
|
|
return -EINVAL;
|
|
|
|
writel(0, &dbc->regs->control);
|
|
dbc->state = DS_DISABLED;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int xhci_dbc_start(struct xhci_dbc *dbc)
|
|
{
|
|
int ret;
|
|
unsigned long flags;
|
|
|
|
WARN_ON(!dbc);
|
|
|
|
pm_runtime_get_sync(dbc->dev); /* note this was self.controller */
|
|
|
|
spin_lock_irqsave(&dbc->lock, flags);
|
|
ret = xhci_do_dbc_start(dbc);
|
|
spin_unlock_irqrestore(&dbc->lock, flags);
|
|
|
|
if (ret) {
|
|
pm_runtime_put(dbc->dev); /* note this was self.controller */
|
|
return ret;
|
|
}
|
|
|
|
return mod_delayed_work(system_percpu_wq, &dbc->event_work,
|
|
msecs_to_jiffies(dbc->poll_interval));
|
|
}
|
|
|
|
static void xhci_dbc_stop(struct xhci_dbc *dbc)
|
|
{
|
|
int ret;
|
|
unsigned long flags;
|
|
|
|
WARN_ON(!dbc);
|
|
|
|
switch (dbc->state) {
|
|
case DS_DISABLED:
|
|
return;
|
|
case DS_CONFIGURED:
|
|
spin_lock(&dbc->lock);
|
|
xhci_dbc_flush_requests(dbc);
|
|
spin_unlock(&dbc->lock);
|
|
|
|
if (dbc->driver->disconnect)
|
|
dbc->driver->disconnect(dbc);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
cancel_delayed_work_sync(&dbc->event_work);
|
|
|
|
spin_lock_irqsave(&dbc->lock, flags);
|
|
ret = xhci_do_dbc_stop(dbc);
|
|
spin_unlock_irqrestore(&dbc->lock, flags);
|
|
if (ret)
|
|
return;
|
|
|
|
xhci_dbc_mem_cleanup(dbc);
|
|
pm_runtime_put_sync(dbc->dev); /* note, was self.controller */
|
|
}
|
|
|
|
static void
|
|
handle_ep_halt_changes(struct xhci_dbc *dbc, struct dbc_ep *dep, bool halted)
|
|
{
|
|
if (halted) {
|
|
dev_info(dbc->dev, "DbC Endpoint halted\n");
|
|
dep->halted = 1;
|
|
|
|
} else if (dep->halted) {
|
|
dev_info(dbc->dev, "DbC Endpoint halt cleared\n");
|
|
dep->halted = 0;
|
|
|
|
if (!list_empty(&dep->list_pending))
|
|
writel(DBC_DOOR_BELL_TARGET(dep->direction),
|
|
&dbc->regs->doorbell);
|
|
}
|
|
}
|
|
|
|
static void
|
|
dbc_handle_port_status(struct xhci_dbc *dbc, union xhci_trb *event)
|
|
{
|
|
u32 portsc;
|
|
|
|
portsc = readl(&dbc->regs->portsc);
|
|
if (portsc & DBC_PORTSC_CONN_CHANGE)
|
|
dev_info(dbc->dev, "DbC port connect change\n");
|
|
|
|
if (portsc & DBC_PORTSC_RESET_CHANGE)
|
|
dev_info(dbc->dev, "DbC port reset change\n");
|
|
|
|
if (portsc & DBC_PORTSC_LINK_CHANGE)
|
|
dev_info(dbc->dev, "DbC port link status change\n");
|
|
|
|
if (portsc & DBC_PORTSC_CONFIG_CHANGE)
|
|
dev_info(dbc->dev, "DbC config error change\n");
|
|
|
|
/* Port reset change bit will be cleared in other place: */
|
|
writel(portsc & ~DBC_PORTSC_RESET_CHANGE, &dbc->regs->portsc);
|
|
}
|
|
|
|
static void dbc_handle_xfer_event(struct xhci_dbc *dbc, union xhci_trb *event)
|
|
{
|
|
struct dbc_ep *dep;
|
|
struct xhci_ring *ring;
|
|
int ep_id;
|
|
int status;
|
|
struct xhci_ep_ctx *ep_ctx;
|
|
u32 comp_code;
|
|
size_t remain_length;
|
|
struct dbc_request *req = NULL, *r;
|
|
|
|
comp_code = GET_COMP_CODE(le32_to_cpu(event->generic.field[2]));
|
|
remain_length = EVENT_TRB_LEN(le32_to_cpu(event->generic.field[2]));
|
|
ep_id = TRB_TO_EP_ID(le32_to_cpu(event->generic.field[3]));
|
|
dep = (ep_id == EPID_OUT) ?
|
|
get_out_ep(dbc) : get_in_ep(dbc);
|
|
ep_ctx = (ep_id == EPID_OUT) ?
|
|
dbc_bulkout_ctx(dbc) : dbc_bulkin_ctx(dbc);
|
|
ring = dep->ring;
|
|
|
|
/* Match the pending request: */
|
|
list_for_each_entry(r, &dep->list_pending, list_pending) {
|
|
if (r->trb_dma == event->trans_event.buffer) {
|
|
req = r;
|
|
break;
|
|
}
|
|
if (r->status == -COMP_STALL_ERROR) {
|
|
dev_warn(dbc->dev, "Give back stale stalled req\n");
|
|
ring->num_trbs_free++;
|
|
xhci_dbc_giveback(r, 0);
|
|
}
|
|
}
|
|
|
|
if (!req) {
|
|
dev_warn(dbc->dev, "no matched request\n");
|
|
return;
|
|
}
|
|
|
|
trace_xhci_dbc_handle_transfer(ring, &req->trb->generic, req->trb_dma);
|
|
|
|
switch (comp_code) {
|
|
case COMP_SUCCESS:
|
|
remain_length = 0;
|
|
fallthrough;
|
|
case COMP_SHORT_PACKET:
|
|
status = 0;
|
|
break;
|
|
case COMP_TRB_ERROR:
|
|
case COMP_BABBLE_DETECTED_ERROR:
|
|
case COMP_USB_TRANSACTION_ERROR:
|
|
dev_warn(dbc->dev, "tx error %d detected\n", comp_code);
|
|
status = -comp_code;
|
|
break;
|
|
case COMP_STALL_ERROR:
|
|
dev_warn(dbc->dev, "Stall error at bulk TRB %llx, remaining %zu, ep deq %llx\n",
|
|
event->trans_event.buffer, remain_length, ep_ctx->deq);
|
|
status = 0;
|
|
dep->halted = 1;
|
|
|
|
/*
|
|
* xHC DbC may trigger a STALL bulk xfer event when host sends a
|
|
* ClearFeature(ENDPOINT_HALT) request even if there wasn't an
|
|
* active bulk transfer.
|
|
*
|
|
* Don't give back this transfer request as hardware will later
|
|
* start processing TRBs starting from this 'STALLED' TRB,
|
|
* causing TRBs and requests to be out of sync.
|
|
*
|
|
* If STALL event shows some bytes were transferred then assume
|
|
* it's an actual transfer issue and give back the request.
|
|
* In this case mark the TRB as No-Op to avoid hw from using the
|
|
* TRB again.
|
|
*/
|
|
|
|
if ((ep_ctx->deq & ~TRB_CYCLE) == event->trans_event.buffer) {
|
|
dev_dbg(dbc->dev, "Ep stopped on Stalled TRB\n");
|
|
if (remain_length == req->length) {
|
|
dev_dbg(dbc->dev, "Spurious stall event, keep req\n");
|
|
req->status = -COMP_STALL_ERROR;
|
|
req->actual = 0;
|
|
return;
|
|
}
|
|
dev_dbg(dbc->dev, "Give back stalled req, but turn TRB to No-op\n");
|
|
trb_to_noop(req->trb);
|
|
}
|
|
break;
|
|
|
|
default:
|
|
dev_err(dbc->dev, "unknown tx error %d\n", comp_code);
|
|
status = -comp_code;
|
|
break;
|
|
}
|
|
|
|
ring->num_trbs_free++;
|
|
req->actual = req->length - remain_length;
|
|
xhci_dbc_giveback(req, status);
|
|
}
|
|
|
|
static void inc_evt_deq(struct xhci_ring *ring)
|
|
{
|
|
/* If on the last TRB of the segment go back to the beginning */
|
|
if (ring->dequeue == &ring->deq_seg->trbs[TRBS_PER_SEGMENT - 1]) {
|
|
ring->cycle_state ^= 1;
|
|
ring->dequeue = ring->deq_seg->trbs;
|
|
return;
|
|
}
|
|
ring->dequeue++;
|
|
}
|
|
|
|
static enum evtreturn xhci_dbc_do_handle_events(struct xhci_dbc *dbc)
|
|
{
|
|
dma_addr_t deq;
|
|
union xhci_trb *evt;
|
|
enum evtreturn ret = EVT_DONE;
|
|
u32 ctrl, portsc;
|
|
bool update_erdp = false;
|
|
|
|
/* DbC state machine: */
|
|
switch (dbc->state) {
|
|
case DS_DISABLED:
|
|
case DS_INITIALIZED:
|
|
|
|
return EVT_ERR;
|
|
case DS_ENABLED:
|
|
portsc = readl(&dbc->regs->portsc);
|
|
if (portsc & DBC_PORTSC_CONN_STATUS) {
|
|
dbc->state = DS_CONNECTED;
|
|
dev_info(dbc->dev, "DbC connected\n");
|
|
}
|
|
|
|
return EVT_DONE;
|
|
case DS_CONNECTED:
|
|
ctrl = readl(&dbc->regs->control);
|
|
if (ctrl & DBC_CTRL_DBC_RUN) {
|
|
dbc->state = DS_CONFIGURED;
|
|
dev_info(dbc->dev, "DbC configured\n");
|
|
portsc = readl(&dbc->regs->portsc);
|
|
writel(portsc, &dbc->regs->portsc);
|
|
ret = EVT_GSER;
|
|
break;
|
|
}
|
|
|
|
return EVT_DONE;
|
|
case DS_CONFIGURED:
|
|
/* Handle cable unplug event: */
|
|
portsc = readl(&dbc->regs->portsc);
|
|
if (!(portsc & DBC_PORTSC_PORT_ENABLED) &&
|
|
!(portsc & DBC_PORTSC_CONN_STATUS)) {
|
|
dev_info(dbc->dev, "DbC cable unplugged\n");
|
|
dbc->state = DS_ENABLED;
|
|
xhci_dbc_flush_requests(dbc);
|
|
xhci_dbc_reinit_ep_rings(dbc);
|
|
return EVT_DISC;
|
|
}
|
|
|
|
/* Handle debug port reset event: */
|
|
if (portsc & DBC_PORTSC_RESET_CHANGE) {
|
|
dev_info(dbc->dev, "DbC port reset\n");
|
|
writel(portsc, &dbc->regs->portsc);
|
|
dbc->state = DS_ENABLED;
|
|
xhci_dbc_flush_requests(dbc);
|
|
xhci_dbc_reinit_ep_rings(dbc);
|
|
return EVT_DISC;
|
|
}
|
|
|
|
/* Check and handle changes in endpoint halt status */
|
|
ctrl = readl(&dbc->regs->control);
|
|
handle_ep_halt_changes(dbc, get_in_ep(dbc), ctrl & DBC_CTRL_HALT_IN_TR);
|
|
handle_ep_halt_changes(dbc, get_out_ep(dbc), ctrl & DBC_CTRL_HALT_OUT_TR);
|
|
|
|
/* Clear DbC run change bit: */
|
|
if (ctrl & DBC_CTRL_DBC_RUN_CHANGE) {
|
|
writel(ctrl, &dbc->regs->control);
|
|
ctrl = readl(&dbc->regs->control);
|
|
}
|
|
break;
|
|
default:
|
|
dev_err(dbc->dev, "Unknown DbC state %d\n", dbc->state);
|
|
break;
|
|
}
|
|
|
|
/* Handle the events in the event ring: */
|
|
evt = dbc->ring_evt->dequeue;
|
|
while ((le32_to_cpu(evt->event_cmd.flags) & TRB_CYCLE) ==
|
|
dbc->ring_evt->cycle_state) {
|
|
/*
|
|
* Add a barrier between reading the cycle flag and any
|
|
* reads of the event's flags/data below:
|
|
*/
|
|
rmb();
|
|
|
|
trace_xhci_dbc_handle_event(dbc->ring_evt, &evt->generic,
|
|
xhci_trb_virt_to_dma(dbc->ring_evt->deq_seg,
|
|
dbc->ring_evt->dequeue));
|
|
|
|
switch (le32_to_cpu(evt->event_cmd.flags) & TRB_TYPE_BITMASK) {
|
|
case TRB_TYPE(TRB_PORT_STATUS):
|
|
dbc_handle_port_status(dbc, evt);
|
|
break;
|
|
case TRB_TYPE(TRB_TRANSFER):
|
|
dbc_handle_xfer_event(dbc, evt);
|
|
if (ret != EVT_GSER)
|
|
ret = EVT_XFER_DONE;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
inc_evt_deq(dbc->ring_evt);
|
|
|
|
evt = dbc->ring_evt->dequeue;
|
|
update_erdp = true;
|
|
}
|
|
|
|
/* Update event ring dequeue pointer: */
|
|
if (update_erdp) {
|
|
deq = xhci_trb_virt_to_dma(dbc->ring_evt->deq_seg,
|
|
dbc->ring_evt->dequeue);
|
|
lo_hi_writeq(deq, &dbc->regs->erdp);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void xhci_dbc_handle_events(struct work_struct *work)
|
|
{
|
|
enum evtreturn evtr;
|
|
struct xhci_dbc *dbc;
|
|
unsigned long flags;
|
|
unsigned int poll_interval;
|
|
unsigned long busypoll_timelimit;
|
|
|
|
dbc = container_of(to_delayed_work(work), struct xhci_dbc, event_work);
|
|
poll_interval = dbc->poll_interval;
|
|
|
|
spin_lock_irqsave(&dbc->lock, flags);
|
|
evtr = xhci_dbc_do_handle_events(dbc);
|
|
spin_unlock_irqrestore(&dbc->lock, flags);
|
|
|
|
switch (evtr) {
|
|
case EVT_GSER:
|
|
if (dbc->driver->configure)
|
|
dbc->driver->configure(dbc);
|
|
break;
|
|
case EVT_DISC:
|
|
if (dbc->driver->disconnect)
|
|
dbc->driver->disconnect(dbc);
|
|
break;
|
|
case EVT_DONE:
|
|
/*
|
|
* Set fast poll rate if there are pending out transfers, or
|
|
* a transfer was recently processed
|
|
*/
|
|
busypoll_timelimit = dbc->xfer_timestamp +
|
|
msecs_to_jiffies(DBC_XFER_INACTIVITY_TIMEOUT);
|
|
|
|
if (!list_empty(&dbc->eps[BULK_OUT].list_pending) ||
|
|
time_is_after_jiffies(busypoll_timelimit))
|
|
poll_interval = 0;
|
|
break;
|
|
case EVT_XFER_DONE:
|
|
dbc->xfer_timestamp = jiffies;
|
|
poll_interval = 0;
|
|
break;
|
|
default:
|
|
dev_info(dbc->dev, "stop handling dbc events\n");
|
|
return;
|
|
}
|
|
|
|
mod_delayed_work(system_percpu_wq, &dbc->event_work,
|
|
msecs_to_jiffies(poll_interval));
|
|
}
|
|
|
|
static const char * const dbc_state_strings[DS_MAX] = {
|
|
[DS_DISABLED] = "disabled",
|
|
[DS_INITIALIZED] = "initialized",
|
|
[DS_ENABLED] = "enabled",
|
|
[DS_CONNECTED] = "connected",
|
|
[DS_CONFIGURED] = "configured",
|
|
};
|
|
|
|
static ssize_t dbc_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
if (dbc->state >= ARRAY_SIZE(dbc_state_strings))
|
|
return sysfs_emit(buf, "unknown\n");
|
|
|
|
return sysfs_emit(buf, "%s\n", dbc_state_strings[dbc->state]);
|
|
}
|
|
|
|
static ssize_t dbc_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct xhci_hcd *xhci;
|
|
struct xhci_dbc *dbc;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
if (sysfs_streq(buf, "enable"))
|
|
xhci_dbc_start(dbc);
|
|
else if (sysfs_streq(buf, "disable"))
|
|
xhci_dbc_stop(dbc);
|
|
else
|
|
return -EINVAL;
|
|
|
|
return count;
|
|
}
|
|
|
|
static ssize_t dbc_idVendor_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%04x\n", dbc->idVendor);
|
|
}
|
|
|
|
static ssize_t dbc_idVendor_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
void __iomem *ptr;
|
|
u16 value;
|
|
u32 dev_info;
|
|
int ret;
|
|
|
|
ret = kstrtou16(buf, 0, &value);
|
|
if (ret)
|
|
return ret;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
dbc->idVendor = value;
|
|
ptr = &dbc->regs->devinfo1;
|
|
dev_info = readl(ptr);
|
|
dev_info = (dev_info & ~(0xffffu << 16)) | (value << 16);
|
|
writel(dev_info, ptr);
|
|
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_idProduct_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%04x\n", dbc->idProduct);
|
|
}
|
|
|
|
static ssize_t dbc_idProduct_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
void __iomem *ptr;
|
|
u32 dev_info;
|
|
u16 value;
|
|
int ret;
|
|
|
|
ret = kstrtou16(buf, 0, &value);
|
|
if (ret)
|
|
return ret;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
dbc->idProduct = value;
|
|
ptr = &dbc->regs->devinfo2;
|
|
dev_info = readl(ptr);
|
|
dev_info = (dev_info & ~(0xffffu)) | value;
|
|
writel(dev_info, ptr);
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_bcdDevice_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%04x\n", dbc->bcdDevice);
|
|
}
|
|
|
|
static ssize_t dbc_bcdDevice_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
void __iomem *ptr;
|
|
u32 dev_info;
|
|
u16 value;
|
|
int ret;
|
|
|
|
ret = kstrtou16(buf, 0, &value);
|
|
if (ret)
|
|
return ret;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
dbc->bcdDevice = value;
|
|
ptr = &dbc->regs->devinfo2;
|
|
dev_info = readl(ptr);
|
|
dev_info = (dev_info & ~(0xffffu << 16)) | (value << 16);
|
|
writel(dev_info, ptr);
|
|
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_manufacturer_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_hcd *xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%s\n", dbc->str.manufacturer);
|
|
}
|
|
|
|
static ssize_t dbc_manufacturer_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_hcd *xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
size_t len;
|
|
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
len = strcspn(buf, "\n");
|
|
if (!len)
|
|
return -EINVAL;
|
|
|
|
if (len > USB_MAX_STRING_LEN)
|
|
return -E2BIG;
|
|
|
|
memcpy(dbc->str.manufacturer, buf, len);
|
|
dbc->str.manufacturer[len] = '\0';
|
|
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_product_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_hcd *xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%s\n", dbc->str.product);
|
|
}
|
|
|
|
static ssize_t dbc_product_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_hcd *xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
size_t len;
|
|
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
len = strcspn(buf, "\n");
|
|
if (!len)
|
|
return -EINVAL;
|
|
|
|
if (len > USB_MAX_STRING_LEN)
|
|
return -E2BIG;
|
|
|
|
memcpy(dbc->str.product, buf, len);
|
|
dbc->str.product[len] = '\0';
|
|
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_serial_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_hcd *xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%s\n", dbc->str.serial);
|
|
}
|
|
|
|
static ssize_t dbc_serial_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_hcd *xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
size_t len;
|
|
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
len = strcspn(buf, "\n");
|
|
if (!len)
|
|
return -EINVAL;
|
|
|
|
if (len > USB_MAX_STRING_LEN)
|
|
return -E2BIG;
|
|
|
|
memcpy(dbc->str.serial, buf, len);
|
|
dbc->str.serial[len] = '\0';
|
|
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_bInterfaceProtocol_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%02x\n", dbc->bInterfaceProtocol);
|
|
}
|
|
|
|
static ssize_t dbc_bInterfaceProtocol_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
void __iomem *ptr;
|
|
u32 dev_info;
|
|
u8 value;
|
|
int ret;
|
|
|
|
/* bInterfaceProtocol is 8 bit, but... */
|
|
ret = kstrtou8(buf, 0, &value);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* ...xhci only supports values 0 and 1 */
|
|
if (value > 1)
|
|
return -EINVAL;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
if (dbc->state != DS_DISABLED)
|
|
return -EBUSY;
|
|
|
|
dbc->bInterfaceProtocol = value;
|
|
ptr = &dbc->regs->devinfo1;
|
|
dev_info = readl(ptr);
|
|
dev_info = (dev_info & ~(0xffu)) | value;
|
|
writel(dev_info, ptr);
|
|
|
|
return size;
|
|
}
|
|
|
|
static ssize_t dbc_poll_interval_ms_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
return sysfs_emit(buf, "%u\n", dbc->poll_interval);
|
|
}
|
|
|
|
static ssize_t dbc_poll_interval_ms_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t size)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
struct xhci_hcd *xhci;
|
|
u32 value;
|
|
int ret;
|
|
|
|
ret = kstrtou32(buf, 0, &value);
|
|
if (ret || value > DBC_POLL_INTERVAL_MAX)
|
|
return -EINVAL;
|
|
|
|
xhci = hcd_to_xhci(dev_get_drvdata(dev));
|
|
dbc = xhci->dbc;
|
|
|
|
dbc->poll_interval = value;
|
|
|
|
mod_delayed_work(system_percpu_wq, &dbc->event_work, 0);
|
|
|
|
return size;
|
|
}
|
|
|
|
static DEVICE_ATTR_RW(dbc);
|
|
static DEVICE_ATTR_RW(dbc_idVendor);
|
|
static DEVICE_ATTR_RW(dbc_idProduct);
|
|
static DEVICE_ATTR_RW(dbc_bcdDevice);
|
|
static DEVICE_ATTR_RW(dbc_serial);
|
|
static DEVICE_ATTR_RW(dbc_product);
|
|
static DEVICE_ATTR_RW(dbc_manufacturer);
|
|
static DEVICE_ATTR_RW(dbc_bInterfaceProtocol);
|
|
static DEVICE_ATTR_RW(dbc_poll_interval_ms);
|
|
|
|
static struct attribute *dbc_dev_attrs[] = {
|
|
&dev_attr_dbc.attr,
|
|
&dev_attr_dbc_idVendor.attr,
|
|
&dev_attr_dbc_idProduct.attr,
|
|
&dev_attr_dbc_bcdDevice.attr,
|
|
&dev_attr_dbc_serial.attr,
|
|
&dev_attr_dbc_product.attr,
|
|
&dev_attr_dbc_manufacturer.attr,
|
|
&dev_attr_dbc_bInterfaceProtocol.attr,
|
|
&dev_attr_dbc_poll_interval_ms.attr,
|
|
NULL
|
|
};
|
|
ATTRIBUTE_GROUPS(dbc_dev);
|
|
|
|
struct xhci_dbc *
|
|
xhci_alloc_dbc(struct device *dev, void __iomem *base, const struct dbc_driver *driver)
|
|
{
|
|
struct xhci_dbc *dbc;
|
|
int ret;
|
|
|
|
dbc = kzalloc_obj(*dbc);
|
|
if (!dbc)
|
|
return NULL;
|
|
|
|
dbc->regs = base;
|
|
dbc->dev = dev;
|
|
dbc->driver = driver;
|
|
dbc->idProduct = DBC_PRODUCT_ID;
|
|
dbc->idVendor = DBC_VENDOR_ID;
|
|
dbc->bcdDevice = DBC_DEVICE_REV;
|
|
dbc->bInterfaceProtocol = DBC_PROTOCOL;
|
|
dbc->poll_interval = DBC_POLL_INTERVAL_DEFAULT;
|
|
|
|
/* initialize serial, product and manufacturer with default values */
|
|
dbc->str = dbc_str_default;
|
|
|
|
if (readl(&dbc->regs->control) & DBC_CTRL_DBC_ENABLE)
|
|
goto err;
|
|
|
|
INIT_DELAYED_WORK(&dbc->event_work, xhci_dbc_handle_events);
|
|
spin_lock_init(&dbc->lock);
|
|
|
|
ret = sysfs_create_groups(&dev->kobj, dbc_dev_groups);
|
|
if (ret)
|
|
goto err;
|
|
|
|
return dbc;
|
|
err:
|
|
kfree(dbc);
|
|
return NULL;
|
|
}
|
|
|
|
/* undo what xhci_alloc_dbc() did */
|
|
void xhci_dbc_remove(struct xhci_dbc *dbc)
|
|
{
|
|
if (!dbc)
|
|
return;
|
|
/* stop hw, stop wq and call dbc->ops->stop() */
|
|
xhci_dbc_stop(dbc);
|
|
|
|
/* remove sysfs files */
|
|
sysfs_remove_groups(&dbc->dev->kobj, dbc_dev_groups);
|
|
|
|
kfree(dbc);
|
|
}
|
|
|
|
|
|
int xhci_create_dbc_dev(struct xhci_hcd *xhci)
|
|
{
|
|
struct device *dev;
|
|
void __iomem *base;
|
|
int ret;
|
|
int dbc_cap_offs;
|
|
|
|
/* create all parameters needed resembling a dbc device */
|
|
dev = xhci_to_hcd(xhci)->self.controller;
|
|
base = &xhci->cap_regs->hc_capbase;
|
|
|
|
dbc_cap_offs = xhci_find_next_ext_cap(base, 0, XHCI_EXT_CAPS_DEBUG);
|
|
if (!dbc_cap_offs)
|
|
return -ENODEV;
|
|
|
|
/* already allocated and in use */
|
|
if (xhci->dbc)
|
|
return -EBUSY;
|
|
|
|
ret = xhci_dbc_tty_probe(dev, base + dbc_cap_offs, xhci);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void xhci_remove_dbc_dev(struct xhci_hcd *xhci)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (!xhci->dbc)
|
|
return;
|
|
|
|
xhci_dbc_tty_remove(xhci->dbc);
|
|
spin_lock_irqsave(&xhci->lock, flags);
|
|
xhci->dbc = NULL;
|
|
spin_unlock_irqrestore(&xhci->lock, flags);
|
|
}
|
|
|
|
#ifdef CONFIG_PM
|
|
int xhci_dbc_suspend(struct xhci_hcd *xhci)
|
|
{
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
|
|
if (!dbc)
|
|
return 0;
|
|
|
|
switch (dbc->state) {
|
|
case DS_ENABLED:
|
|
case DS_CONNECTED:
|
|
case DS_CONFIGURED:
|
|
dbc->resume_required = 1;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
xhci_dbc_stop(dbc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int xhci_dbc_resume(struct xhci_hcd *xhci)
|
|
{
|
|
int ret = 0;
|
|
struct xhci_dbc *dbc = xhci->dbc;
|
|
|
|
if (!dbc)
|
|
return 0;
|
|
|
|
if (dbc->resume_required) {
|
|
dbc->resume_required = 0;
|
|
xhci_dbc_start(dbc);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
#endif /* CONFIG_PM */
|
|
|
|
int xhci_dbc_init(void)
|
|
{
|
|
return dbc_tty_init();
|
|
}
|
|
|
|
void xhci_dbc_exit(void)
|
|
{
|
|
dbc_tty_exit();
|
|
}
|