|
@@ -110,7 +110,7 @@ static void xhci_scratchpad_free(struct xhci_ctrl *ctrl)
|
|
|
|
|
|
ctrl->dcbaa->dev_context_ptrs[0] = 0;
|
|
|
|
|
|
- free((void *)(uintptr_t)le64_to_cpu(ctrl->scratchpad->sp_array[0]));
|
|
|
+ free(xhci_bus_to_virt(ctrl, le64_to_cpu(ctrl->scratchpad->sp_array[0])));
|
|
|
free(ctrl->scratchpad->sp_array);
|
|
|
free(ctrl->scratchpad);
|
|
|
ctrl->scratchpad = NULL;
|
|
@@ -216,8 +216,8 @@ static void *xhci_malloc(unsigned int size)
|
|
|
* @param link_trbs flag to indicate whether to link the trbs or NOT
|
|
|
* @return none
|
|
|
*/
|
|
|
-static void xhci_link_segments(struct xhci_segment *prev,
|
|
|
- struct xhci_segment *next, bool link_trbs)
|
|
|
+static void xhci_link_segments(struct xhci_ctrl *ctrl, struct xhci_segment *prev,
|
|
|
+ struct xhci_segment *next, bool link_trbs)
|
|
|
{
|
|
|
u32 val;
|
|
|
u64 val_64 = 0;
|
|
@@ -226,7 +226,7 @@ static void xhci_link_segments(struct xhci_segment *prev,
|
|
|
return;
|
|
|
prev->next = next;
|
|
|
if (link_trbs) {
|
|
|
- val_64 = virt_to_phys(next->trbs);
|
|
|
+ val_64 = xhci_virt_to_bus(ctrl, next->trbs);
|
|
|
prev->trbs[TRBS_PER_SEGMENT-1].link.segment_ptr =
|
|
|
cpu_to_le64(val_64);
|
|
|
|
|
@@ -304,7 +304,8 @@ static struct xhci_segment *xhci_segment_alloc(void)
|
|
|
* @param link_trbs flag to indicate whether to link the trbs or NOT
|
|
|
* @return pointer to the newly created RING
|
|
|
*/
|
|
|
-struct xhci_ring *xhci_ring_alloc(unsigned int num_segs, bool link_trbs)
|
|
|
+struct xhci_ring *xhci_ring_alloc(struct xhci_ctrl *ctrl, unsigned int num_segs,
|
|
|
+ bool link_trbs)
|
|
|
{
|
|
|
struct xhci_ring *ring;
|
|
|
struct xhci_segment *prev;
|
|
@@ -327,12 +328,12 @@ struct xhci_ring *xhci_ring_alloc(unsigned int num_segs, bool link_trbs)
|
|
|
next = xhci_segment_alloc();
|
|
|
BUG_ON(!next);
|
|
|
|
|
|
- xhci_link_segments(prev, next, link_trbs);
|
|
|
+ xhci_link_segments(ctrl, prev, next, link_trbs);
|
|
|
|
|
|
prev = next;
|
|
|
num_segs--;
|
|
|
}
|
|
|
- xhci_link_segments(prev, ring->first_seg, link_trbs);
|
|
|
+ xhci_link_segments(ctrl, prev, ring->first_seg, link_trbs);
|
|
|
if (link_trbs) {
|
|
|
/* See section 4.9.2.1 and 6.4.4.1 */
|
|
|
prev->trbs[TRBS_PER_SEGMENT-1].link.control |=
|
|
@@ -354,6 +355,7 @@ static int xhci_scratchpad_alloc(struct xhci_ctrl *ctrl)
|
|
|
struct xhci_hccr *hccr = ctrl->hccr;
|
|
|
struct xhci_hcor *hcor = ctrl->hcor;
|
|
|
struct xhci_scratchpad *scratchpad;
|
|
|
+ uint64_t val_64;
|
|
|
int num_sp;
|
|
|
uint32_t page_size;
|
|
|
void *buf;
|
|
@@ -371,8 +373,9 @@ static int xhci_scratchpad_alloc(struct xhci_ctrl *ctrl)
|
|
|
scratchpad->sp_array = xhci_malloc(num_sp * sizeof(u64));
|
|
|
if (!scratchpad->sp_array)
|
|
|
goto fail_sp2;
|
|
|
- ctrl->dcbaa->dev_context_ptrs[0] =
|
|
|
- cpu_to_le64((uintptr_t)scratchpad->sp_array);
|
|
|
+
|
|
|
+ val_64 = xhci_virt_to_bus(ctrl, scratchpad->sp_array);
|
|
|
+ ctrl->dcbaa->dev_context_ptrs[0] = cpu_to_le64(val_64);
|
|
|
|
|
|
xhci_flush_cache((uintptr_t)&ctrl->dcbaa->dev_context_ptrs[0],
|
|
|
sizeof(ctrl->dcbaa->dev_context_ptrs[0]));
|
|
@@ -393,8 +396,8 @@ static int xhci_scratchpad_alloc(struct xhci_ctrl *ctrl)
|
|
|
xhci_flush_cache((uintptr_t)buf, num_sp * page_size);
|
|
|
|
|
|
for (i = 0; i < num_sp; i++) {
|
|
|
- uintptr_t ptr = (uintptr_t)buf + i * page_size;
|
|
|
- scratchpad->sp_array[i] = cpu_to_le64(ptr);
|
|
|
+ val_64 = xhci_virt_to_bus(ctrl, buf + i * page_size);
|
|
|
+ scratchpad->sp_array[i] = cpu_to_le64(val_64);
|
|
|
}
|
|
|
|
|
|
xhci_flush_cache((uintptr_t)scratchpad->sp_array,
|
|
@@ -484,9 +487,9 @@ int xhci_alloc_virt_device(struct xhci_ctrl *ctrl, unsigned int slot_id)
|
|
|
}
|
|
|
|
|
|
/* Allocate endpoint 0 ring */
|
|
|
- virt_dev->eps[0].ring = xhci_ring_alloc(1, true);
|
|
|
+ virt_dev->eps[0].ring = xhci_ring_alloc(ctrl, 1, true);
|
|
|
|
|
|
- byte_64 = virt_to_phys(virt_dev->out_ctx->bytes);
|
|
|
+ byte_64 = xhci_virt_to_bus(ctrl, virt_dev->out_ctx->bytes);
|
|
|
|
|
|
/* Point to output device context in dcbaa. */
|
|
|
ctrl->dcbaa->dev_context_ptrs[slot_id] = cpu_to_le64(byte_64);
|
|
@@ -522,15 +525,15 @@ int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
|
|
|
return -ENOMEM;
|
|
|
}
|
|
|
|
|
|
- val_64 = virt_to_phys(ctrl->dcbaa);
|
|
|
+ val_64 = xhci_virt_to_bus(ctrl, ctrl->dcbaa);
|
|
|
/* Set the pointer in DCBAA register */
|
|
|
xhci_writeq(&hcor->or_dcbaap, val_64);
|
|
|
|
|
|
/* Command ring control pointer register initialization */
|
|
|
- ctrl->cmd_ring = xhci_ring_alloc(1, true);
|
|
|
+ ctrl->cmd_ring = xhci_ring_alloc(ctrl, 1, true);
|
|
|
|
|
|
/* Set the address in the Command Ring Control register */
|
|
|
- trb_64 = virt_to_phys(ctrl->cmd_ring->first_seg->trbs);
|
|
|
+ trb_64 = xhci_virt_to_bus(ctrl, ctrl->cmd_ring->first_seg->trbs);
|
|
|
val_64 = xhci_readq(&hcor->or_crcr);
|
|
|
val_64 = (val_64 & (u64) CMD_RING_RSVD_BITS) |
|
|
|
(trb_64 & (u64) ~CMD_RING_RSVD_BITS) |
|
|
@@ -551,7 +554,7 @@ int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
|
|
|
ctrl->ir_set = &ctrl->run_regs->ir_set[0];
|
|
|
|
|
|
/* Event ring does not maintain link TRB */
|
|
|
- ctrl->event_ring = xhci_ring_alloc(ERST_NUM_SEGS, false);
|
|
|
+ ctrl->event_ring = xhci_ring_alloc(ctrl, ERST_NUM_SEGS, false);
|
|
|
ctrl->erst.entries = xhci_malloc(sizeof(struct xhci_erst_entry) *
|
|
|
ERST_NUM_SEGS);
|
|
|
|
|
@@ -560,8 +563,8 @@ int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
|
|
|
for (val = 0, seg = ctrl->event_ring->first_seg;
|
|
|
val < ERST_NUM_SEGS;
|
|
|
val++) {
|
|
|
- trb_64 = virt_to_phys(seg->trbs);
|
|
|
struct xhci_erst_entry *entry = &ctrl->erst.entries[val];
|
|
|
+ trb_64 = xhci_virt_to_bus(ctrl, seg->trbs);
|
|
|
entry->seg_addr = cpu_to_le64(trb_64);
|
|
|
entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT);
|
|
|
entry->rsvd = 0;
|
|
@@ -570,7 +573,7 @@ int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
|
|
|
xhci_flush_cache((uintptr_t)ctrl->erst.entries,
|
|
|
ERST_NUM_SEGS * sizeof(struct xhci_erst_entry));
|
|
|
|
|
|
- deq = virt_to_phys(ctrl->event_ring->dequeue);
|
|
|
+ deq = xhci_virt_to_bus(ctrl, ctrl->event_ring->dequeue);
|
|
|
|
|
|
/* Update HC event ring dequeue pointer */
|
|
|
xhci_writeq(&ctrl->ir_set->erst_dequeue,
|
|
@@ -585,7 +588,7 @@ int xhci_mem_init(struct xhci_ctrl *ctrl, struct xhci_hccr *hccr,
|
|
|
/* this is the event ring segment table pointer */
|
|
|
val_64 = xhci_readq(&ctrl->ir_set->erst_base);
|
|
|
val_64 &= ERST_PTR_MASK;
|
|
|
- val_64 |= virt_to_phys(ctrl->erst.entries) & ~ERST_PTR_MASK;
|
|
|
+ val_64 |= xhci_virt_to_bus(ctrl, ctrl->erst.entries) & ~ERST_PTR_MASK;
|
|
|
|
|
|
xhci_writeq(&ctrl->ir_set->erst_base, val_64);
|
|
|
|
|
@@ -848,7 +851,7 @@ void xhci_setup_addressable_virt_dev(struct xhci_ctrl *ctrl,
|
|
|
/* EP 0 can handle "burst" sizes of 1, so Max Burst Size field is 0 */
|
|
|
ep0_ctx->ep_info2 |= cpu_to_le32(MAX_BURST(0) | ERROR_COUNT(3));
|
|
|
|
|
|
- trb_64 = virt_to_phys(virt_dev->eps[0].ring->first_seg->trbs);
|
|
|
+ trb_64 = xhci_virt_to_bus(ctrl, virt_dev->eps[0].ring->first_seg->trbs);
|
|
|
ep0_ctx->deq = cpu_to_le64(trb_64 | virt_dev->eps[0].ring->cycle_state);
|
|
|
|
|
|
/*
|