Lines Matching refs:queue_size
358 req.queue_size = queue->queue_size; in mana_gd_create_hw_eq()
482 uint32_t num_cqe = cq->queue_size / GDMA_CQE_SIZE; in mana_gd_ring_cq()
493 uint32_t head = eq->head % (eq->queue_size / GDMA_EQE_SIZE); in mana_gd_process_eqe()
568 num_eqe = eq->queue_size / GDMA_EQE_SIZE; in mana_gd_process_eq_events()
797 log2_num_entries = ilog2(queue->queue_size / GDMA_EQE_SIZE); in mana_gd_create_eq()
838 uint32_t log2_num_entries = ilog2(spec->queue_size / GDMA_CQE_SIZE); in mana_gd_create_cq()
872 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_hwc_queue()
879 queue->queue_size = spec->queue_size; in mana_gd_create_hwc_queue()
1006 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_mana_eq()
1017 queue->queue_size = spec->queue_size; in mana_gd_create_mana_eq()
1051 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_mana_wq_cq()
1062 queue->queue_size = spec->queue_size; in mana_gd_create_mana_wq_cq()
1219 uint32_t wq_size = wq->queue_size; in mana_gd_wq_avail_space()
1233 (wqe_offset * GDMA_WQE_BU_SIZE) & (wq->queue_size - 1); in mana_gd_get_wqe_ptr()
1235 if ((offset + GDMA_WQE_BU_SIZE) > wq->queue_size) { in mana_gd_get_wqe_ptr()
1238 offset + GDMA_WQE_BU_SIZE, wq->queue_size); in mana_gd_get_wqe_ptr()
1301 end_ptr = base_ptr + wq->queue_size; in mana_gd_write_sgl()
1363 if (wqe_ptr >= (uint8_t *)wq->queue_mem_ptr + wq->queue_size) in mana_gd_post_work_request()
1364 wqe_ptr -= wq->queue_size; in mana_gd_post_work_request()
1396 unsigned int num_cqe = cq->queue_size / sizeof(struct gdma_cqe); in mana_gd_read_cqe()