| /linux/drivers/misc/bcm-vk/ |
| H A D | bcm_vk_msg.c | 102 return (qinfo->q_size - msgq_occupied(msgq, qinfo) - 1); in msgq_avail_space() 495 qinfo->q_size = msgq_size; in bcm_vk_sync_msgq() 497 qinfo->q_low = qinfo->q_size >> 1; in bcm_vk_sync_msgq() 498 qinfo->q_mask = qinfo->q_size - 1; in bcm_vk_sync_msgq() 560 avail, qinfo->q_size); in bcm_vk_append_ib_sgl() 644 if (wr_idx >= qinfo->q_size) { in bcm_to_v_msg_enqueue() 646 wr_idx, qinfo->q_size); in bcm_to_v_msg_enqueue() 805 max_msg_to_process = BCM_VK_MSG_PROC_MAX_LOOP * qinfo->q_size; in bcm_to_h_msg_dequeue() 825 if ((rd_idx >= qinfo->q_size) || in bcm_to_h_msg_dequeue() 826 (src_size > (qinfo->q_size - 1))) { in bcm_to_h_msg_dequeue() [all …]
|
| H A D | bcm_vk_msg.h | 44 u32 q_size; member
|
| /linux/drivers/net/ethernet/amd/pds_core/ |
| H A D | core.c | 149 dma_free_coherent(dev, qcq->q_size, in pdsc_qcq_free() 235 qcq->q_size = PDS_PAGE_SIZE + in pdsc_qcq_alloc() 239 qcq->q_size + qcq->cq_size, in pdsc_qcq_alloc() 259 qcq->q_size = PDS_PAGE_SIZE + (num_descs * desc_size); in pdsc_qcq_alloc() 260 qcq->q_base = dma_alloc_coherent(dev, qcq->q_size, in pdsc_qcq_alloc() 292 dma_free_coherent(dev, qcq->q_size, qcq->q_base, qcq->q_base_pa); in pdsc_qcq_alloc()
|
| H A D | debugfs.c | 120 debugfs_create_x32("q_size", 0400, qcq_dentry, &qcq->q_size); in pdsc_debugfs_add_qcq()
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | request_manager.c | 54 u32 q_size; in octeon_init_instr_queue() local 71 q_size = (u32)conf->instr_type * num_descs; in octeon_init_instr_queue() 77 iq->base_addr = lio_dma_alloc(oct, q_size, &iq->base_addr_dma); in octeon_init_instr_queue() 94 lio_dma_free(oct, q_size, iq->base_addr, iq->base_addr_dma); in octeon_init_instr_queue() 140 lio_dma_free(oct, q_size, iq->base_addr, iq->base_addr_dma); in octeon_init_instr_queue() 158 u64 desc_size = 0, q_size; in octeon_delete_instr_queue() local 177 q_size = iq->max_count * desc_size; in octeon_delete_instr_queue() 178 lio_dma_free(oct, (u32)q_size, iq->base_addr, in octeon_delete_instr_queue()
|
| /linux/arch/sparc/kernel/ |
| H A D | pci_sun4v.c | 1066 unsigned long q_size, alloc_size, pages, order; in pci_sun4v_msiq_alloc() local 1069 q_size = pbm->msiq_ent_count * sizeof(struct pci_sun4v_msiq_entry); in pci_sun4v_msiq_alloc() 1070 alloc_size = (pbm->msiq_num * q_size); in pci_sun4v_msiq_alloc() 1082 unsigned long err, base = __pa(pages + (i * q_size)); in pci_sun4v_msiq_alloc() 1120 unsigned long q_size, alloc_size, pages, order; in pci_sun4v_msiq_free() local 1129 q_size = pbm->msiq_ent_count * sizeof(struct pci_sun4v_msiq_entry); in pci_sun4v_msiq_free() 1130 alloc_size = (pbm->msiq_num * q_size); in pci_sun4v_msiq_free()
|
| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_main.c | 206 size_t q_size; in alloc_command_queues() local 216 q_size = qlen * cqinfo->cmd_size; in alloc_command_queues() 221 size_t rem_q_size = q_size; in alloc_command_queues()
|
| /linux/drivers/net/wireless/intel/iwlwifi/mei/ |
| H A D | main.c | 129 size_t q_size[SAP_DIRECTION_MAX][SAP_QUEUE_IDX_MAX]; member 323 mem->q_size[dir][queue] = in iwl_mei_init_shared_mem() 459 q_sz = mei->shared_mem.q_size[SAP_DIRECTION_HOST_TO_ME][SAP_QUEUE_IDX_NOTIF]; in iwl_mei_send_sap_msg_payload() 510 q_sz = mei->shared_mem.q_size[SAP_DIRECTION_HOST_TO_ME][SAP_QUEUE_IDX_DATA]; in iwl_mei_add_data_to_ring() 1216 q_sz = mei->shared_mem.q_size[SAP_DIRECTION_ME_TO_HOST][SAP_QUEUE_IDX_NOTIF]; in iwl_mei_handle_check_shared_area() 1229 q_sz = mei->shared_mem.q_size[SAP_DIRECTION_ME_TO_HOST][SAP_QUEUE_IDX_DATA]; in iwl_mei_handle_check_shared_area()
|
| /linux/drivers/net/ethernet/pensando/ionic/ |
| H A D | ionic_lif.c | 425 dma_free_coherent(dev, qcq->q_size, qcq->q_base, qcq->q_base_pa); in ionic_qcq_free() 634 int q_size; in ionic_qcq_alloc() local 640 q_size = ALIGN(num_descs * desc_size, PAGE_SIZE); in ionic_qcq_alloc() 641 new->q_size = PAGE_SIZE + q_size + in ionic_qcq_alloc() 643 new->q_base = dma_alloc_coherent(dev, new->q_size, in ionic_qcq_alloc() 654 new->cq.base = PTR_ALIGN(new->q.base + q_size, PAGE_SIZE); in ionic_qcq_alloc() 655 new->cq.base_pa = ALIGN(new->q_base_pa + q_size, PAGE_SIZE); in ionic_qcq_alloc() 659 new->q_size = PAGE_SIZE + (num_descs * desc_size); in ionic_qcq_alloc() 660 new->q_base = dma_alloc_coherent(dev, new->q_size, &new->q_base_pa, in ionic_qcq_alloc() 740 dma_free_coherent(dev, new->q_size, new->q_base, new->q_base_pa); in ionic_qcq_alloc() [all …]
|
| H A D | ionic_debugfs.c | 131 debugfs_create_x32("q_size", 0400, qcq_dentry, &qcq->q_size); in ionic_debugfs_add_qcq()
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf_main.c | 193 size_t q_size, c_size, rem_q_size; in alloc_command_queues() local 202 q_size = qlen * OTX_CPT_INST_SIZE; in alloc_command_queues() 208 rem_q_size = q_size; in alloc_command_queues()
|
| /linux/drivers/media/platform/qcom/venus/ |
| H A D | hfi_venus.c | 62 u32 q_size; member 186 qsize = qhdr->q_size; in venus_write_queue() 260 qsize = qhdr->q_size; in venus_read_queue() 766 qhdr->q_size = IFACEQ_QUEUE_SIZE / 4; in venus_set_qhdr_defaults()
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs-srv.c | 991 static int post_recv_io(struct rtrs_srv_con *con, size_t q_size) in post_recv_io() argument 995 for (i = 0; i < q_size; i++) { in post_recv_io() 1008 size_t q_size; in post_recv_path() local 1013 q_size = SERVICE_CON_QUEUE_DEPTH; in post_recv_path() 1015 q_size = srv->queue_depth; in post_recv_path() 1027 err = post_recv_io(to_srv_con(srv_path->s.con[cid]), q_size); in post_recv_path()
|
| H A D | rtrs-clt.c | 695 static int post_recv_io(struct rtrs_clt_con *con, size_t q_size) in post_recv_io() argument 700 for (i = 0; i < q_size; i++) { in post_recv_io() 717 size_t q_size = 0; in post_recv_path() local 722 q_size = SERVICE_CON_QUEUE_DEPTH; in post_recv_path() 724 q_size = clt_path->queue_depth; in post_recv_path() 730 q_size *= 2; in post_recv_path() 732 err = post_recv_io(to_clt_con(clt_path->s.con[cid]), q_size); in post_recv_path()
|
| /linux/drivers/media/platform/qcom/iris/ |
| H A D | iris_hfi_queue.c | 206 iface_q->qhdr->q_size = IFACEQ_QUEUE_SIZE / sizeof(u32); in iris_hfi_queue_set_header()
|
| /linux/drivers/net/hyperv/ |
| H A D | hyperv_net.h | 735 u16 q_size; member 797 u16 q_size; member
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_process_queue_manager.c | 828 q_data->q_size = q->properties.queue_size; in criu_checkpoint_queue() 963 qp->queue_size = q_data->q_size; in set_queue_properties_from_criu()
|
| H A D | kfd_priv.h | 1275 uint64_t q_size; member
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_tx_rx.c | 3318 u32 q_size; in bna_tx_res_req() local 3328 q_size = txq_depth * BFI_TXQ_WI_SIZE; in bna_tx_res_req() 3329 q_size = ALIGN(q_size, PAGE_SIZE); in bna_tx_res_req() 3330 page_count = q_size >> PAGE_SHIFT; in bna_tx_res_req()
|
| /linux/drivers/media/platform/renesas/ |
| H A D | rcar_jpu.c | 1003 unsigned int q_size = q_data->format.plane_fmt[i].sizeimage; in jpu_queue_setup() local 1005 if (sizes[i] < q_size) in jpu_queue_setup()
|
| /linux/drivers/md/ |
| H A D | dm-cache-policy-smq.c | 288 static unsigned int q_size(struct queue *q) in q_size() function 1151 return q_size(&mq->dirty) == 0u; in clean_target_met()
|
| /linux/drivers/accel/qaic/ |
| H A D | qaic_data.c | 1093 static inline u32 fifo_space_avail(u32 head, u32 tail, u32 q_size) in fifo_space_avail() argument 1098 avail += q_size; in fifo_space_avail()
|
| /linux/drivers/scsi/ibmvscsi/ |
| H A D | ibmvfc.c | 2494 int wait, i, q_index, q_size; in ibmvfc_wait_for_ops() local 2502 q_size = vhost->scsi_scrqs.active_queues; in ibmvfc_wait_for_ops() 2505 q_size = 1; in ibmvfc_wait_for_ops() 2511 for (q_index = 0; q_index < q_size; q_index++) { in ibmvfc_wait_for_ops() 2532 for (q_index = 0; q_index < q_size; q_index++) { in ibmvfc_wait_for_ops()
|
| /linux/net/ipv4/ |
| H A D | udp.c | 1709 int q_size = 0; in __udp_enqueue_schedule_skb() local 1770 q_size += size; in __udp_enqueue_schedule_skb() 1781 atomic_add(q_size, &sk->sk_rmem_alloc); in __udp_enqueue_schedule_skb()
|