| /linux/drivers/mailbox/ |
| H A D | ti-msgmgr.c | 74 u8 queue_count; member 650 if (qinst->queue_id > d->queue_count) { in ti_msgmgr_queue_setup() 652 idx, qinst->queue_id, d->queue_count); in ti_msgmgr_queue_setup() 774 .queue_count = 64, 789 .queue_count = 190, 819 int queue_count; in ti_msgmgr_probe() local 857 queue_count = desc->num_valid_queues; in ti_msgmgr_probe() 858 if (!queue_count || queue_count > desc->queue_count) { in ti_msgmgr_probe() 860 queue_count, desc->queue_count); in ti_msgmgr_probe() 863 inst->num_valid_queues = queue_count; in ti_msgmgr_probe() [all …]
|
| /linux/tools/testing/selftests/net/ |
| H A D | nl_netdev.py | 23 with NetdevSimDev(queue_count=100) as nsimdev: 43 with NetdevSimDev(queue_count=2) as nsimdev: 128 with NetdevSimDev(queue_count=2) as nsimdev: 168 with NetdevSimDev(queue_count=4) as nsimdev:
|
| /linux/tools/testing/selftests/net/lib/py/ |
| H A D | nsim.py | 53 def __init__(self, port_count=1, queue_count=1, ns=None): argument 63 self.ctrl_write("new_device", "%u %u %u" % (addr, port_count, queue_count))
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_sched.c | 585 u32 queue_count; member 942 for (i = 0; i < group->queue_count; i++) in group_release_work() 1017 for (u32 i = 0; i < group->queue_count; i++) in group_bind_locked() 1053 for (u32 i = 0; i < group->queue_count; i++) in group_unbind_locked() 1072 return hweight32(inactive_queues) == group->queue_count; in group_is_idle() 1336 for (i = 0; i < group->queue_count; i++) { in csg_slot_sync_queues_state_locked() 1406 for (i = 0; i < group->queue_count; i++) { in csg_slot_sync_state_locked() 1439 for (i = 0; i < group->queue_count; i++) { in csg_slot_prog_locked() 1527 struct panthor_queue *queue = group && cs_id < group->queue_count ? in cs_slot_process_fault_event_locked() 2158 group->fatal_queues |= GENMASK(group->queue_count - 1, 0); in tick_ctx_init() [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | fc.c | 2278 for (i = 1; i < ctrl->ctrl.queue_count; i++) in nvme_fc_free_io_queues() 2299 struct nvme_fc_queue *queue = &ctrl->queues[ctrl->ctrl.queue_count - 1]; in nvme_fc_delete_hw_io_queues() 2302 for (i = ctrl->ctrl.queue_count - 1; i >= 1; i--, queue--) in nvme_fc_delete_hw_io_queues() 2312 for (i = 1; i < ctrl->ctrl.queue_count; i++, queue++) { in nvme_fc_create_hw_io_queues() 2331 for (i = 1; i < ctrl->ctrl.queue_count; i++) { in nvme_fc_connect_io_queues() 2351 for (i = 1; i < ctrl->ctrl.queue_count; i++) in nvme_fc_init_io_queues() 2446 if (ctrl->ctrl.queue_count > 1) { in __nvme_fc_abort_outstanding_ios() 2447 for (q = 1; q < ctrl->ctrl.queue_count; q++) in __nvme_fc_abort_outstanding_ios() 2464 if (ctrl->ctrl.queue_count > 1) { in __nvme_fc_abort_outstanding_ios() 2875 ctrl->ctrl.queue_count = nr_io_queues + 1; in nvme_fc_create_io_queues() [all …]
|
| H A D | tcp.c | 2033 for (i = 1; i < ctrl->queue_count; i++) in nvme_tcp_free_io_queues() 2041 for (i = 1; i < ctrl->queue_count; i++) in nvme_tcp_stop_io_queues() 2043 for (i = 1; i < ctrl->queue_count; i++) in nvme_tcp_stop_io_queues() 2125 for (i = 1; i < ctrl->queue_count; i++) { in __nvme_tcp_alloc_io_queues() 2157 ctrl->queue_count = nr_io_queues + 1; in nvme_tcp_alloc_io_queues() 2188 nr_queues = min(ctrl->tagset->nr_hw_queues + 1, ctrl->queue_count); in nvme_tcp_configure_io_queues() 2207 ctrl->queue_count - 1); in nvme_tcp_configure_io_queues() 2307 if (ctrl->queue_count <= 1) in nvme_tcp_teardown_io_queues() 2407 if (ctrl->queue_count > 1) { in nvme_tcp_setup_ctrl() 2432 if (ctrl->queue_count > 1) { in nvme_tcp_setup_ctrl() [all …]
|
| H A D | pci.c | 1841 for (i = dev->ctrl.queue_count - 1; i >= lowest; i--) { in nvme_free_queues() 1842 dev->ctrl.queue_count--; in nvme_free_queues() 1868 for (i = dev->ctrl.queue_count - 1; i > 0; i--) in nvme_suspend_io_queues() 1882 for (i = dev->ctrl.queue_count - 1; i > 0; i--) { in nvme_reap_pending_cqes() 1944 if (dev->ctrl.queue_count > qid) in nvme_alloc_queue() 1964 dev->ctrl.queue_count++; in nvme_alloc_queue() 2220 for (i = dev->ctrl.queue_count; i <= dev->max_qid; i++) { in nvme_create_io_queues() 2227 max = min(dev->max_qid, dev->ctrl.queue_count - 1); in nvme_create_io_queues() 3100 if (!dead && dev->ctrl.queue_count > 0) { in nvme_dev_disable()
|
| H A D | auth.c | 1034 for (q = 1; q < ctrl->queue_count; q++) { in nvme_ctrl_auth_work() 1051 for (q = 1; q < ctrl->queue_count; q++) { in nvme_ctrl_auth_work()
|
| /linux/tools/testing/selftests/drivers/net/ |
| H A D | napi_threaded.py | 134 with NetDrvEnv(__file__, queue_count=2) as cfg:
|
| H A D | queues.py | 118 with NetDrvEnv(__file__, queue_count=100) as cfg:
|
| H A D | stats.py | 312 with NetDrvEnv(__file__, queue_count=100) as cfg:
|
| H A D | hds.py | 311 with NetDrvEnv(__file__, queue_count=3) as cfg:
|
| /linux/drivers/net/ethernet/pensando/ionic/ |
| H A D | ionic_debugfs.c | 70 (u32 *)&ionic->ident.lif.eth.config.queue_count[IONIC_QTYPE_TXQ]); in ionic_debugfs_add_sizes() 72 (u32 *)&ionic->ident.lif.eth.config.queue_count[IONIC_QTYPE_RXQ]); in ionic_debugfs_add_sizes()
|
| /linux/drivers/scsi/hisi_sas/ |
| H A D | hisi_sas_v1_hw.c | 657 (u32)((1ULL << hisi_hba->queue_count) - 1)); in init_reg_v1_hw() 699 for (i = 0; i < hisi_hba->queue_count; i++) { in init_reg_v1_hw() 1650 for (i = 0; i < hisi_hba->queue_count; i++, idx++) { in interrupt_init_v1_hw() 1664 idx = (hisi_hba->n_phy * HISI_SAS_PHY_INT_NR) + hisi_hba->queue_count; in interrupt_init_v1_hw() 1679 hisi_hba->cq_nvecs = hisi_hba->queue_count; in interrupt_init_v1_hw() 1746 if (hisi_hba->queue_count < 0 || hisi_hba->queue_count > 32) { in check_fw_info_v1_hw()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_cq.c | 28 count = queue_count(cq->queue, QUEUE_TYPE_TO_CLIENT); in rxe_cq_chk_attr()
|
| H A D | rxe_queue.c | 118 if (!queue_empty(q, q->type) && (num_elem < queue_count(q, type))) in resize_finish()
|
| H A D | rxe_queue.h | 170 static inline u32 queue_count(const struct rxe_queue *q, in queue_count() function
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_packet_manager_v9.c | 56 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_v9() 116 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_aldebaran()
|
| H A D | kfd_device_queue_manager.c | 710 qpd->queue_count++; in create_queue_nocpsch() 904 qpd->queue_count--; in destroy_queue_nocpsch_locked() 2031 qpd->queue_count++; in create_queue_cpsch() 2058 qpd->queue_count--; in create_queue_cpsch() 2577 qpd->queue_count--; in destroy_queue_cpsch() 2851 qpd->queue_count--; in process_termination_cpsch()
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| H A D | 4965-rs.c | 234 while (tl->queue_count && tl->time_stamp < oldest_time) { in il4965_rs_tl_rm_old_stats() 238 tl->queue_count--; in il4965_rs_tl_rm_old_stats() 272 if (!(tl->queue_count)) { in il4965_rs_tl_add_packet() 275 tl->queue_count = 1; in il4965_rs_tl_add_packet() 293 if ((idx + 1) > tl->queue_count) in il4965_rs_tl_add_packet() 294 tl->queue_count = idx + 1; in il4965_rs_tl_add_packet() 317 if (!(tl->queue_count)) in il4965_rs_tl_get_load()
|
| /linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
| H A D | rs.h | 309 u8 queue_count; /* number of queues that has member
|
| H A D | rs.c | 232 while (tl->queue_count && in rs_tl_rm_old_stats() 237 tl->queue_count--; in rs_tl_rm_old_stats() 271 if (!(tl->queue_count)) { in rs_tl_add_packet() 274 tl->queue_count = 1; in rs_tl_add_packet() 292 if ((index + 1) > tl->queue_count) in rs_tl_add_packet() 293 tl->queue_count = index + 1; in rs_tl_add_packet() 345 if (!(tl->queue_count)) in rs_tl_get_load()
|
| /linux/drivers/md/dm-vdo/ |
| H A D | vdo.h | 294 unsigned int queue_count, void *contexts[]);
|
| /linux/drivers/net/ethernet/atheros/atl1c/ |
| H A D | atl1c_main.c | 2619 u32 queue_count = 1; in atl1c_probe() local 2661 queue_count = 4; in atl1c_probe() 2663 netdev = alloc_etherdev_mq(sizeof(struct atl1c_adapter), queue_count); in atl1c_probe() 2682 adapter->tx_queue_count = queue_count; in atl1c_probe() 2683 adapter->rx_queue_count = queue_count; in atl1c_probe()
|
| /linux/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
| H A D | msgbuf.c | 807 u32 queue_count; in brcmf_msgbuf_tx_queue_data() local 820 queue_count = brcmf_flowring_enqueue(flow, flowid, skb); in brcmf_msgbuf_tx_queue_data() 821 force = ((queue_count % BRCMF_MSGBUF_TRICKLE_TXWORKER_THRS) == 0); in brcmf_msgbuf_tx_queue_data()
|