Searched refs:TRBS_PER_SEGMENT (Results 1 – 17 of 17) sorted by relevance
126 if (TRBS_PER_SEGMENT > 40) { in cdns2_raw_ring()128 "\t\tTransfer ring %d too big\n", TRBS_PER_SEGMENT); in cdns2_raw_ring()133 for (i = 0; i < TRBS_PER_SEGMENT; ++i) { in cdns2_raw_ring()
142 link_trb = (ring->trbs + (TRBS_PER_SEGMENT - 1)); in cdns2_alloc_tr_segment()189 cdns2_ep_inc_trb(&ring->enqueue, &ring->pcs, TRBS_PER_SEGMENT); in cdns2_ep_inc_enq()195 cdns2_ep_inc_trb(&ring->dequeue, &ring->ccs, TRBS_PER_SEGMENT); in cdns2_ep_inc_deq()231 if (trb == (pep->ring.trbs + (TRBS_PER_SEGMENT - 1))) in cdns2_next_trb()333 if ((ring->enqueue + num_trbs) >= (TRBS_PER_SEGMENT - 1)) { in cdns2_prepare_ring()338 if (doorbell && dma_index == TRBS_PER_SEGMENT - 1) { in cdns2_prepare_ring()344 link_trb = ring->trbs + (TRBS_PER_SEGMENT - 1); in cdns2_prepare_ring()354 if (pep->type == USB_ENDPOINT_XFER_ISOC || TRBS_PER_SEGMENT > 2) in cdns2_prepare_ring()367 struct cdns2_trb *link_trb = pep->ring.trbs + (TRBS_PER_SEGMENT - 1); in cdns2_dbg_request_trbs()736 trb = &pep->ring.trbs[TRBS_PER_SEGMENT]; in cdns2_prepare_first_isoc_transfer()[all …]
417 #define TRBS_PER_SEGMENT 600 macro431 #if TRBS_PER_SEGMENT < 2455 #define TR_SEG_SIZE (TRB_SIZE * (TRBS_PER_SEGMENT + TRB_ISO_RESERVED))
56 for (i = 0; i < TRBS_PER_SEGMENT; i++) in cdnsp_segment_alloc()118 link = &prev->trbs[TRBS_PER_SEGMENT - 1].link; in cdnsp_link_segments()151 ring->num_trbs_free += (TRBS_PER_SEGMENT - 1) * num_segs; in cdnsp_link_rings()154 ring->last_seg->trbs[TRBS_PER_SEGMENT - 1].link.control &= in cdnsp_link_rings()156 last->trbs[TRBS_PER_SEGMENT - 1].link.control |= in cdnsp_link_rings()318 ring->num_trbs_free = ring->num_segs * (TRBS_PER_SEGMENT - 1) - 1; in cdnsp_initialize_ring_info()399 ring->last_seg->trbs[TRBS_PER_SEGMENT - 1].link.control |= in cdnsp_ring_alloc()432 num_segs_needed = (num_trbs + (TRBS_PER_SEGMENT - 1) - 1) / in cdnsp_ring_expansion()433 (TRBS_PER_SEGMENT - 1); in cdnsp_ring_expansion()1045 entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT); in cdnsp_alloc_erst()
973 #define TRBS_PER_SEGMENT 600 macro979 #if TRBS_PER_SEGMENT < 2999 TRBS_PER_ISOC_SEGMENT : TRBS_PER_SEGMENT)1015 #define TRB_RING_SIZE (TRB_SIZE * TRBS_PER_SEGMENT)
138 if (trb_per_sector > TRBS_PER_SEGMENT) { in cdns3_dbg_ring()
1158 #define TRBS_PER_SEGMENT 256 macro1161 #define TRB_SEGMENT_SIZE (TRBS_PER_SEGMENT * 16)
76 if (trb < seg->trbs || segment_offset >= TRBS_PER_SEGMENT) in cdnsp_trb_virt_to_dma()94 return trb == &seg->trbs[TRBS_PER_SEGMENT - 1]; in cdnsp_last_trb_on_seg()551 temp_trb = &cur_seg->trbs[TRBS_PER_SEGMENT - 1]; in cdnsp_trb_in_td()
1187 TRBS_PER_SEGMENT > 2) in cdns3_ep_run_transfer()1524 if (TRBS_PER_SEGMENT == 2 && priv_ep->type != USB_ENDPOINT_XFER_ISOC) { in cdns3_trb_handled()1615 TRBS_PER_SEGMENT == 2) in cdns3_transfer_completed()
1410 sizeof(union cdnsp_trb) * (TRBS_PER_SEGMENT - 1)); in cdnsp_clear_cmd_ring()
104 trb = &seg->trbs[TRBS_PER_SEGMENT - 1]; in xhci_set_link_trb()129 ring->last_seg->trbs[TRBS_PER_SEGMENT - 1].link.control |= cpu_to_le32(LINK_TOGGLE); in xhci_initialize_ring_segments()147 for (int i = 0; i < TRBS_PER_SEGMENT; i++) in xhci_link_rings()163 dst->last_seg->trbs[TRBS_PER_SEGMENT-1].link.control in xhci_link_rings()168 src->last_seg->trbs[TRBS_PER_SEGMENT-1].link.control &= ~cpu_to_le32(LINK_TOGGLE); in xhci_link_rings()325 ring->num_trbs_free = ring->num_segs * (TRBS_PER_SEGMENT - 1) - 1; in xhci_initialize_ring_info()1813 entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT); in xhci_alloc_erst()
80 if (segment_offset >= TRBS_PER_SEGMENT) in xhci_trb_virt_to_dma()114 return trb == &seg->trbs[TRBS_PER_SEGMENT - 1]; in last_trb_on_seg()165 return seg->num * TRBS_PER_SEGMENT + (trb - seg->trbs); in trb_to_pos()359 return ring->num_segs * (TRBS_PER_SEGMENT - 1); in xhci_num_trbs_free()364 last_on_seg = &enq_seg->trbs[TRBS_PER_SEGMENT - 1]; in xhci_num_trbs_free()390 trbs_past_seg = enq_used + num_trbs - (TRBS_PER_SEGMENT - 1); in xhci_ring_expansion_needed()405 new_segs = 1 + (trbs_past_seg / (TRBS_PER_SEGMENT - 1)); in xhci_ring_expansion_needed()3120 if (event_loop++ > TRBS_PER_SEGMENT / 2) { in xhci_handle_events()
432 erst->entries[0].seg_size = cpu_to_le32(TRBS_PER_SEGMENT); in dbc_erst_alloc()472 union xhci_trb *trb = &seg->trbs[TRBS_PER_SEGMENT - 1]; in xhci_dbc_ring_init()872 if (ring->dequeue == &ring->deq_seg->trbs[TRBS_PER_SEGMENT - 1]) { in inc_evt_deq()
1259 #define TRBS_PER_SEGMENT 256 macro1261 #define MAX_RSVD_CMD_TRBS (TRBS_PER_SEGMENT - 3)1262 #define TRB_SEGMENT_SIZE (TRBS_PER_SEGMENT*16)
204 for (i = 0; i < TRBS_PER_SEGMENT; i++) { in xhci_ring_dump_segment()
870 memset(seg->trbs, 0, sizeof(union xhci_trb) * (TRBS_PER_SEGMENT - 1)); in xhci_clear_command_ring()872 seg->trbs[TRBS_PER_SEGMENT - 1].link.control &= cpu_to_le32(~TRB_CYCLE); in xhci_clear_command_ring()
401 if (ring->enqueue >= &ring->segment->trbs[TRBS_PER_SEGMENT - 1]) { in xdbc_queue_trb()828 if (xdbc.evt_ring.dequeue == &xdbc.evt_seg.trbs[TRBS_PER_SEGMENT]) { in xdbc_handle_events()