| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_queue.h | 53 enum queue_type { enum 70 enum queue_type type; 86 unsigned int elem_size, enum queue_type type); 101 enum queue_type type) in queue_get_producer() 128 enum queue_type type) in queue_get_consumer() 154 static inline int queue_empty(struct rxe_queue *q, enum queue_type type) in queue_empty() 162 static inline int queue_full(struct rxe_queue *q, enum queue_type type) in queue_full() 171 enum queue_type type) in queue_count() 180 enum queue_type type) in queue_advance_producer() 216 enum queue_type type) in queue_advance_consumer() [all …]
|
| H A D | rxe_queue.c | 56 unsigned int elem_size, enum queue_type type) in rxe_queue_init() 113 enum queue_type type = q->type; in resize_finish()
|
| H A D | rxe_cq.c | 47 enum queue_type type; in rxe_cq_from_init()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | mes_userqueue.c | 140 queue_input.queue_type = queue->queue_type; in mes_userq_map() 203 int queue_type) in mes_userq_detect_and_reset() argument 222 input.queue_type = queue_type; in mes_userq_detect_and_reset() 225 r = amdgpu_mes_detect_and_reset_hung_queues(adev, queue_type, false, in mes_userq_detect_and_reset() 232 if (queue->queue_type == queue_type) { in mes_userq_detect_and_reset() 259 struct amdgpu_mqd *mqd_hw_default = &adev->mqds[queue->queue_type]; in mes_userq_mqd_create() 287 if (queue->queue_type == AMDGPU_HW_IP_COMPUTE) { in mes_userq_mqd_create() 315 } else if (queue->queue_type == AMDGPU_HW_IP_GFX) { in mes_userq_mqd_create() 354 } else if (queue->queue_type == AMDGPU_HW_IP_DMA) { in mes_userq_mqd_create()
|
| H A D | amdgpu_mes.h | 181 int queue_type; member 187 int queue_type; member 225 uint32_t queue_type; member 246 uint32_t queue_type; member 256 uint32_t queue_type; member 277 uint32_t queue_type; member 291 uint32_t queue_type; member 421 int queue_type,
|
| H A D | mes_v11_0.c | 280 static int convert_to_mes_queue_type(int queue_type) in convert_to_mes_queue_type() argument 282 if (queue_type == AMDGPU_RING_TYPE_GFX) in convert_to_mes_queue_type() 284 else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) in convert_to_mes_queue_type() 286 else if (queue_type == AMDGPU_RING_TYPE_SDMA) in convert_to_mes_queue_type() 345 mes_add_queue_pkt.queue_type = in mes_v11_0_add_hw_queue() 346 convert_to_mes_queue_type(input->queue_type); in mes_v11_0_add_hw_queue() 391 static int mes_v11_0_reset_queue_mmio(struct amdgpu_mes *mes, uint32_t queue_type, in mes_v11_0_reset_queue_mmio() argument 401 if (queue_type == AMDGPU_RING_TYPE_GFX) { in mes_v11_0_reset_queue_mmio() 434 } else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) { in mes_v11_0_reset_queue_mmio() 454 } else if (queue_type == AMDGPU_RING_TYPE_SDMA) { in mes_v11_0_reset_queue_mmio() [all …]
|
| H A D | mes_v12_0.c | 271 static int convert_to_mes_queue_type(int queue_type) in convert_to_mes_queue_type() argument 273 if (queue_type == AMDGPU_RING_TYPE_GFX) in convert_to_mes_queue_type() 275 else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) in convert_to_mes_queue_type() 277 else if (queue_type == AMDGPU_RING_TYPE_SDMA) in convert_to_mes_queue_type() 279 else if (queue_type == AMDGPU_RING_TYPE_MES) in convert_to_mes_queue_type() 334 mes_add_queue_pkt.queue_type = in mes_v12_0_add_hw_queue() 335 convert_to_mes_queue_type(input->queue_type); in mes_v12_0_add_hw_queue() 416 static int mes_v12_0_reset_queue_mmio(struct amdgpu_mes *mes, uint32_t queue_type, in mes_v12_0_reset_queue_mmio() argument 426 if (queue_type == AMDGPU_RING_TYPE_GFX) { in mes_v12_0_reset_queue_mmio() 459 } else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) { in mes_v12_0_reset_queue_mmio() [all …]
|
| H A D | amdgpu_mes.c | 314 queue_input.queue_type = ring->funcs->type; in amdgpu_mes_map_legacy_queue() 339 queue_input.queue_type = ring->funcs->type; in amdgpu_mes_unmap_legacy_queue() 365 queue_input.queue_type = ring->funcs->type; in amdgpu_mes_reset_legacy_queue() 393 int queue_type, in amdgpu_mes_detect_and_reset_hung_queues() argument 406 if ((queue_type != AMDGPU_RING_TYPE_GFX) && in amdgpu_mes_detect_and_reset_hung_queues() 407 (queue_type != AMDGPU_RING_TYPE_COMPUTE) && in amdgpu_mes_detect_and_reset_hung_queues() 408 (queue_type != AMDGPU_RING_TYPE_SDMA)) in amdgpu_mes_detect_and_reset_hung_queues() 414 input.queue_type = queue_type; in amdgpu_mes_detect_and_reset_hung_queues()
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | hw_queue.c | 53 if (!hdev->asic_prop.max_queues || q->queue_type == QUEUE_TYPE_HW) in hl_hw_queue_update_ci() 62 if (!cs_needs_completion(cs) || q->queue_type == QUEUE_TYPE_INT) in hl_hw_queue_update_ci() 257 if (q->queue_type != QUEUE_TYPE_HW) { in hl_hw_queue_send_cb_no_cmpl() 666 switch (q->queue_type) { in hl_hw_queue_schedule_cs() 683 q->queue_type); in hl_hw_queue_schedule_cs() 695 if (q->queue_type == QUEUE_TYPE_EXT) in hl_hw_queue_schedule_cs() 773 switch (job->queue_type) { in hl_hw_queue_schedule_cs() 796 if ((q->queue_type == QUEUE_TYPE_EXT) && in hl_hw_queue_schedule_cs() 1003 switch (q->queue_type) { in queue_init() 1021 q->queue_type); in queue_init() [all …]
|
| /linux/drivers/net/ |
| H A D | virtio_net.c | 4596 u32 queue_type; in virtnet_stats_ctx_init() local 4605 queue_type = VIRTNET_Q_TYPE_RX; in virtnet_stats_ctx_init() 4608 ctx->bitmap[queue_type] |= VIRTIO_NET_STATS_TYPE_RX_BASIC; in virtnet_stats_ctx_init() 4609 ctx->desc_num[queue_type] += ARRAY_SIZE(virtnet_stats_rx_basic_desc_qstat); in virtnet_stats_ctx_init() 4610 ctx->size[queue_type] += sizeof(struct virtio_net_stats_rx_basic); in virtnet_stats_ctx_init() 4614 ctx->bitmap[queue_type] |= VIRTIO_NET_STATS_TYPE_RX_CSUM; in virtnet_stats_ctx_init() 4615 ctx->desc_num[queue_type] += ARRAY_SIZE(virtnet_stats_rx_csum_desc_qstat); in virtnet_stats_ctx_init() 4616 ctx->size[queue_type] += sizeof(struct virtio_net_stats_rx_csum); in virtnet_stats_ctx_init() 4620 ctx->bitmap[queue_type] |= VIRTIO_NET_STATS_TYPE_RX_GSO; in virtnet_stats_ctx_init() 4621 ctx->desc_num[queue_type] += ARRAY_SIZE(virtnet_stats_rx_gso_desc_qstat); in virtnet_stats_ctx_init() [all …]
|
| /linux/drivers/net/wireless/ath/ath5k/ |
| H A D | qcu.c | 203 ath5k_hw_setup_tx_queue(struct ath5k_hw *ah, enum ath5k_tx_queue queue_type, in ath5k_hw_setup_tx_queue() argument 214 switch (queue_type) { in ath5k_hw_setup_tx_queue() 226 switch (queue_type) { in ath5k_hw_setup_tx_queue() 248 ah->ah_txq[queue].tqi_type = queue_type; in ath5k_hw_setup_tx_queue() 251 queue_info->tqi_type = queue_type; in ath5k_hw_setup_tx_queue()
|
| /linux/drivers/scsi/qla2xxx/ |
| H A D | qla_tmpl.h | 135 uint8_t queue_type; member 196 uint8_t queue_type; member
|
| H A D | qla_tmpl.c | 329 uint type = ent->t263.queue_type; in qla27xx_fwdt_entry_t263() 365 ent->t263.queue_type == T263_QUEUE_TYPE_ATIO) { in qla27xx_fwdt_entry_t263() 633 ulong type = ent->t274.queue_type; in qla27xx_fwdt_entry_t274() 664 ent->t274.queue_type == T274_QUEUE_TYPE_ATIO_SHAD) { in qla27xx_fwdt_entry_t274()
|
| /linux/drivers/net/ethernet/ibm/ehea/ |
| H A D | ehea_phyp.c | 416 const u8 queue_type, const u64 resource_handle, in ehea_h_register_rpage() argument 422 | EHEA_BMASK_SET(H_REG_RPAGE_QT, queue_type); in ehea_h_register_rpage() 500 const u8 pagesize, const u8 queue_type, in ehea_h_register_rpage_mr() argument 509 queue_type, mr_handle, in ehea_h_register_rpage_mr()
|
| H A D | ehea_phyp.h | 376 const u8 queue_type, 397 const u8 pagesize, const u8 queue_type,
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_packet_manager_v9.c | 202 packet->bitfields2.queue_type = in pm_set_resources_v9() 246 packet->bitfields2.queue_type = in pm_map_queues_v9() 252 packet->bitfields2.queue_type = in pm_map_queues_v9() 256 packet->bitfields2.queue_type = in pm_map_queues_v9()
|
| H A D | kfd_pm4_headers_vi.h | 67 enum mes_set_resources_queue_type_enum queue_type:3; member 238 enum mes_map_queues_queue_type_vi_enum queue_type:3; member
|
| H A D | kfd_device_queue_manager.c | 180 static int convert_to_mes_queue_type(int queue_type) in convert_to_mes_queue_type() argument 184 switch (queue_type) { in convert_to_mes_queue_type() 192 WARN(1, "Invalid queue type %d", queue_type); in convert_to_mes_queue_type() 206 int r, queue_type; in add_queue_mes() local 247 queue_type = convert_to_mes_queue_type(q->properties.type); in add_queue_mes() 248 if (queue_type < 0) { in add_queue_mes() 254 queue_input.queue_type = (uint32_t)queue_type; in add_queue_mes() 3568 qss_entry->queue_type = set_queue_type_for_user(&q->properties); in set_queue_snapshot_entry()
|
| /linux/drivers/media/platform/qcom/iris/ |
| H A D | iris_hfi_queue.c | 67 if (queue->queue_type == IFACEQ_MSGQ_ID) in iris_hfi_queue_read() 205 iface_q->qhdr->queue_type = queue_id; in iris_hfi_queue_set_header()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_qp_ctxt.h | 150 u16 queue_type; member
|
| /linux/include/uapi/linux/ |
| H A D | kfd_ioctl.h | 76 __u32 queue_type; /* to KFD */ member 1084 __u32 queue_type; member
|
| /linux/drivers/scsi/be2iscsi/ |
| H A D | be_cmds.c | 901 int queue_type) in beiscsi_cmd_q_destroy() argument 911 "queue_type : %d\n", queue_type); in beiscsi_cmd_q_destroy() 917 switch (queue_type) { in beiscsi_cmd_q_destroy() 947 if (queue_type != QTYPE_SGL) in beiscsi_cmd_q_destroy()
|
| /linux/drivers/s390/cio/ |
| H A D | qdio_main.c | 331 if (queue_type(q) == QDIO_IQDIO_QFMT && !multicast_outbound(q)) { in qdio_siga_output() 415 if (queue_type(q) == QDIO_IQDIO_QFMT && !q->is_input_q && in process_buffer_error() 1283 if (queue_type(q) == QDIO_IQDIO_QFMT) { in handle_outbound()
|
| /linux/drivers/scsi/mpi3mr/ |
| H A D | mpi3mr.h | 334 enum queue_type { enum 478 enum queue_type qtype;
|
| /linux/drivers/net/wireless/ti/wlcore/ |
| H A D | conf.h | 188 u8 queue_type; member
|