| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_kernel_queue.c | 43 enum kfd_queue_type type, unsigned int queue_size) in kq_initialize() argument 49 pr_debug("Initializing queue type %d size %d\n", type, queue_size); in kq_initialize() 71 retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq); in kq_initialize() 74 queue_size); in kq_initialize() 111 memset(kq->pq_kernel_addr, 0, queue_size); in kq_initialize() 115 prop.queue_size = queue_size; in kq_initialize() 219 queue_size_dwords = kq->queue->properties.queue_size / 4; in kq_acquire_packet_buffer() 301 (kq->queue->properties.queue_size / 4); in kq_rollback_packet()
|
| H A D | kfd_mqd_manager_cik.c | 166 uint32_t wptr_mask = (uint32_t)((p->queue_size / 4) - 1); in load_mqd() 192 m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1; in __update_mqd() 231 m->sdma_rlc_rb_cntl = order_base_2(q->queue_size / 4) in update_mqd_sdma() 352 m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1; in update_mqd_hiq()
|
| H A D | kfd_queue.c | 37 pr_debug("Queue Size: %llu\n", q->queue_size); in print_queue_properties() 54 pr_debug("Queue Size: %llu\n", q->properties.queue_size); in print_queue() 252 properties->queue_size / 2; in kfd_queue_acquire_buffers() 255 properties->queue_size + properties->metadata_queue_size; in kfd_queue_acquire_buffers()
|
| H A D | kfd_mqd_manager_v12_1.c | 246 ffs(q->queue_size / sizeof(unsigned int)) - 1 - 1; in update_mqd() 255 if (q->metadata_queue_size == q->queue_size * 4) { in update_mqd() 262 lower_32_bits((q->queue_address + q->queue_size) >> 8); in update_mqd() 264 upper_32_bits((q->queue_address + q->queue_size) >> 8); in update_mqd() 403 m->sdmax_rlcx_rb_cntl = (ffs(q->queue_size / sizeof(unsigned int)) - 1) in update_mqd_sdma()
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | hw_channel.c | 296 enum gdma_queue_type type, u64 queue_size, in mana_hwc_create_gdma_wq() argument 306 spec.queue_size = queue_size; in mana_hwc_create_gdma_wq() 312 u64 queue_size, in mana_hwc_create_gdma_cq() argument 321 spec.queue_size = queue_size; in mana_hwc_create_gdma_cq() 330 u64 queue_size, in mana_hwc_create_gdma_eq() argument 338 spec.queue_size = queue_size; in mana_hwc_create_gdma_eq() 529 u32 queue_size; in mana_hwc_create_wq() local 535 queue_size = roundup_pow_of_two(GDMA_MAX_RQE_SIZE * q_depth); in mana_hwc_create_wq() 537 queue_size = roundup_pow_of_two(GDMA_MAX_SQE_SIZE * q_depth); in mana_hwc_create_wq() 539 if (queue_size < MANA_MIN_QSIZE) in mana_hwc_create_wq() [all …]
|
| H A D | gdma_main.c | 279 req.queue_size = queue->queue_size; in mana_gd_create_hw_eq() 393 u32 num_cqe = cq->queue_size / GDMA_CQE_SIZE; in mana_gd_ring_cq() 563 u32 head = eq->head % (eq->queue_size / GDMA_EQE_SIZE); in mana_gd_process_eqe() 670 num_eqe = eq->queue_size / GDMA_EQE_SIZE; in mana_gd_process_eq_events() 856 log2_num_entries = ilog2(queue->queue_size / GDMA_EQE_SIZE); in mana_gd_create_eq() 895 u32 log2_num_entries = ilog2(spec->queue_size / GDMA_CQE_SIZE); in mana_gd_create_cq() 931 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_hwc_queue() 934 spec->type, spec->queue_size, err); in mana_gd_create_hwc_queue() 941 queue->queue_size = spec->queue_size; in mana_gd_create_hwc_queue() 958 spec->type, spec->queue_size, err); in mana_gd_create_hwc_queue() [all …]
|
| /linux/drivers/vdpa/ifcvf/ |
| H A D | ifcvf_base.c | 74 u16 queue_size; in ifcvf_get_vq_size() local 80 queue_size = vp_ioread16(&hw->common_cfg->queue_size); in ifcvf_get_vq_size() 82 return queue_size; in ifcvf_get_vq_size() 87 u16 queue_size, max_size, qid; in ifcvf_get_max_vq_size() local 91 queue_size = ifcvf_get_vq_size(hw, qid); in ifcvf_get_max_vq_size() 93 if (!queue_size) in ifcvf_get_max_vq_size() 96 max_size = max(queue_size, max_size); in ifcvf_get_max_vq_size() 352 vp_iowrite16(num, &cfg->queue_size); in ifcvf_set_vq_num()
|
| /linux/drivers/firmware/tegra/ |
| H A D | ivc.c | 562 unsigned tegra_ivc_total_queue_size(unsigned queue_size) in tegra_ivc_total_queue_size() argument 564 if (!IS_ALIGNED(queue_size, TEGRA_IVC_ALIGN)) { in tegra_ivc_total_queue_size() 566 __func__, queue_size, TEGRA_IVC_ALIGN); in tegra_ivc_total_queue_size() 570 return queue_size + sizeof(struct tegra_ivc_header); in tegra_ivc_total_queue_size() 652 size_t queue_size; in tegra_ivc_init() local 670 queue_size = tegra_ivc_total_queue_size(num_frames * frame_size); in tegra_ivc_init() 673 ivc->rx.phys = dma_map_single(peer, iosys_map_get_vaddr(rx), queue_size, in tegra_ivc_init() 678 ivc->tx.phys = dma_map_single(peer, iosys_map_get_vaddr(tx), queue_size, in tegra_ivc_init() 681 dma_unmap_single(peer, ivc->rx.phys, queue_size, in tegra_ivc_init()
|
| H A D | bpmp-tegra186.c | 112 size_t message_size, queue_size; in tegra186_bpmp_channel_init() local 123 queue_size = tegra_ivc_total_queue_size(message_size); in tegra186_bpmp_channel_init() 124 offset = queue_size * index; in tegra186_bpmp_channel_init()
|
| /linux/drivers/media/platform/qcom/iris/ |
| H A D | iris_hfi_queue.c | 249 u32 queue_size; in iris_hfi_queues_init() local 252 queue_size = ALIGN((sizeof(*q_tbl_hdr) + (IFACEQ_QUEUE_SIZE * IFACEQ_NUMQ)), SZ_4K); in iris_hfi_queues_init() 253 core->iface_q_table_vaddr = dma_alloc_attrs(core->dev, queue_size, in iris_hfi_queues_init() 294 u32 queue_size; in iris_hfi_queues_deinit() local 309 queue_size = ALIGN(sizeof(struct iris_hfi_queue_table_header) + in iris_hfi_queues_deinit() 312 dma_free_attrs(core->dev, queue_size, core->iface_q_table_vaddr, in iris_hfi_queues_deinit()
|
| /linux/sound/firewire/ |
| H A D | amdtp-stream.c | 653 if (++s->packet_index >= s->queue_size) in queue_packet() 913 unsigned int queue_size) in compute_ohci_it_cycle() argument 916 return increment_ohci_cycle_count(cycle, queue_size); in compute_ohci_it_cycle() 926 unsigned int queue_size = s->queue_size; in generate_tx_packet_descs() local 995 packet_index = (packet_index + 1) % queue_size; in generate_tx_packet_descs() 1028 unsigned int index = (s->packet_index + i) % s->queue_size; in generate_rx_packet_descs() 1031 desc->cycle = compute_ohci_it_cycle(*ctx_header, s->queue_size); in generate_rx_packet_descs() 1242 cycle = compute_ohci_it_cycle(ctx_header[packets - 1], s->queue_size); in skip_rx_packets() 1268 const unsigned int queue_size in process_rx_packets_intermediately() local 1667 amdtp_stream_start(struct amdtp_stream * s,int channel,int speed,unsigned int queue_size,unsigned int idle_irq_interval) amdtp_stream_start() argument 2068 unsigned int queue_size; amdtp_domain_start() local [all...] |
| /linux/drivers/infiniband/hw/mana/ |
| H A D | qp.c | 186 wq_spec.queue_size = wq->wq_buf_size; in mana_ib_create_qp_rss() 189 cq_spec.queue_size = cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_rss() 337 wq_spec.queue_size = ucmd.sq_buf_size; in mana_ib_create_qp_raw() 340 cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_raw() 409 u32 queue_size; in mana_ib_queue_size() local 415 queue_size = attr->cap.max_send_wr * in mana_ib_queue_size() 418 queue_size = attr->cap.max_recv_wr * in mana_ib_queue_size() 425 return MANA_PAGE_ALIGN(roundup_pow_of_two(queue_size)); in mana_ib_queue_size() 557 err = mana_ib_create_queue(mdev, ucmd.queue_buf[j], ucmd.queue_size[j], in mana_ib_create_rc_qp() 637 u32 doorbell, queue_size; in mana_ib_create_ud_qp() local [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/ |
| H A D | rx.c | 123 WARN_ON(rxq->queue_size & (rxq->queue_size - 1)); in iwl_rxq_space() 131 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space() 273 rxq->write = (rxq->write + 1) & (rxq->queue_size - 1); in iwl_pcie_rxmq_restock() 686 free_size * rxq->queue_size, in iwl_pcie_free_rxq_dma() 697 rxq->queue_size, in iwl_pcie_free_rxq_dma() 725 rxq->queue_size = iwl_trans_get_num_rbds(trans); in iwl_pcie_alloc_rxq_dma() 727 rxq->queue_size = RX_QUEUE_SIZE; in iwl_pcie_alloc_rxq_dma() 735 rxq->bd = dma_alloc_coherent(dev, free_size * rxq->queue_size, in iwl_pcie_alloc_rxq_dma() 743 rxq->queue_size, in iwl_pcie_alloc_rxq_dma() 1085 int i, err, queue_size, allocator_pool_size, num_alloc; in _iwl_pcie_rx_init() local [all …]
|
| /linux/drivers/media/pci/intel/ipu6/ |
| H A D | ipu6-fw-isys.c | 287 input_queue_cfg[i].queue_size = IPU6_ISYS_SIZE_PROXY_SEND_QUEUE; in ipu6_isys_fwcom_cfg_init() 293 input_queue_cfg[base_dev_send + i].queue_size = max_devq_size; in ipu6_isys_fwcom_cfg_init() 299 input_queue_cfg[base_msg_send + i].queue_size = in ipu6_isys_fwcom_cfg_init() 306 output_queue_cfg[i].queue_size = in ipu6_isys_fwcom_cfg_init() 313 output_queue_cfg[base_msg_recv + i].queue_size = in ipu6_isys_fwcom_cfg_init()
|
| H A D | ipu6-fw-com.c | 199 sizeinput += size_mul(cfg->input[i].queue_size + 1, in ipu6_fw_com_prepare() 203 sizeoutput += size_mul(cfg->output[i].queue_size + 1, in ipu6_fw_com_prepare() 249 cfg->input[i].queue_size, in ipu6_fw_com_prepare() 258 cfg->output[i].queue_size, in ipu6_fw_com_prepare()
|
| /linux/kernel/bpf/ |
| H A D | queue_stack_maps.c | 69 u64 size, queue_size; in queue_stack_map_alloc() local 72 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc() 74 qs = bpf_map_area_alloc(queue_size, numa_node); in queue_stack_map_alloc()
|
| /linux/drivers/nvme/host/ |
| H A D | rdma.c | 87 int queue_size; member 266 init_attr.cap.max_send_wr = factor * queue->queue_size + 1; in nvme_rdma_create_qp() 268 init_attr.cap.max_recv_wr = queue->queue_size + 1; in nvme_rdma_create_qp() 444 nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size, in nvme_rdma_destroy_queue_ib() 505 queue->cq_size = cq_factor * queue->queue_size + 1; in nvme_rdma_create_queue_ib() 515 queue->rsp_ring = nvme_rdma_alloc_ring(ibdev, queue->queue_size, in nvme_rdma_create_queue_ib() 529 queue->queue_size, in nvme_rdma_create_queue_ib() 535 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() 541 queue->queue_size, IB_MR_TYPE_INTEGRITY, in nvme_rdma_create_queue_ib() 546 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_gc_9_4_3.c | 328 uint32_t queue_size = in kgd_gfx_v9_4_3_hqd_load() local 331 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in kgd_gfx_v9_4_3_hqd_load() 333 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in kgd_gfx_v9_4_3_hqd_load() 334 guessed_wptr += queue_size; in kgd_gfx_v9_4_3_hqd_load() 335 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in kgd_gfx_v9_4_3_hqd_load()
|
| H A D | amdgpu_amdkfd_gfx_v10_3.c | 239 uint32_t queue_size = in hqd_load_v10_3() local 242 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in hqd_load_v10_3() 244 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in hqd_load_v10_3() 245 guessed_wptr += queue_size; in hqd_load_v10_3() 246 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in hqd_load_v10_3()
|
| H A D | mes_userqueue.c | 164 queue_input.queue_size = userq_props->queue_size >> 2; in mes_userq_map() 302 userq_props->queue_size = mqd_user->queue_size; in mes_userq_mqd_create()
|
| H A D | amdgpu_amdkfd_gfx_v11.c | 224 uint32_t queue_size = in hqd_load_v11() local 227 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in hqd_load_v11() 229 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in hqd_load_v11() 230 guessed_wptr += queue_size; in hqd_load_v11() 231 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in hqd_load_v11()
|
| /linux/sound/core/ |
| H A D | timer.c | 79 int queue_size; member 1305 prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1; in snd_timer_user_interrupt() 1312 if (tu->qused >= tu->queue_size) { in snd_timer_user_interrupt() 1316 tu->qtail %= tu->queue_size; in snd_timer_user_interrupt() 1329 if (tu->qused >= tu->queue_size) { in snd_timer_user_append_to_tqueue() 1333 tu->qtail %= tu->queue_size; in snd_timer_user_append_to_tqueue() 1406 prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1; in snd_timer_user_tinterrupt() 1447 tu->queue_size = size; in realloc_user_queue() 1816 if (params.queue_size > 0 && in snd_timer_user_params() 1817 (params.queue_size < 32 || params.queue_size > 1024)) { in snd_timer_user_params() [all …]
|
| /linux/drivers/i2c/busses/ |
| H A D | i2c-qcom-cci.c | 122 u16 queue_size[NUM_QUEUES]; member 322 if (val == cci->data->queue_size[queue]) in cci_validate_queue() 677 .queue_size = { 64, 16 }, 710 .queue_size = { 64, 16 }, 743 .queue_size = { 64, 16 }, 788 .queue_size = { 64, 16 },
|
| /linux/drivers/misc/genwqe/ |
| H A D | card_ddcb.c | 1026 unsigned int queue_size; in setup_ddcb_queue() local 1032 queue_size = roundup(GENWQE_DDCB_MAX * sizeof(struct ddcb), PAGE_SIZE); in setup_ddcb_queue() 1042 queue->ddcb_vaddr = __genwqe_alloc_consistent(cd, queue_size, in setup_ddcb_queue() 1087 __genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr, in setup_ddcb_queue() 1102 unsigned int queue_size; in free_ddcb_queue() local 1104 queue_size = roundup(queue->ddcb_max * sizeof(struct ddcb), PAGE_SIZE); in free_ddcb_queue() 1110 __genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr, in free_ddcb_queue()
|
| /linux/drivers/platform/chrome/wilco_ec/ |
| H A D | event.c | 102 static int queue_size = 64; variable 103 module_param(queue_size, int, 0644); 468 dev_data->events = event_queue_new(queue_size); in event_device_add()
|