| /linux/drivers/net/ethernet/engleder/ |
| H A D | tsnep_xdp.c | 22 struct xsk_buff_pool *pool, u16 queue_id) in tsnep_xdp_enable_pool() argument 27 if (queue_id >= adapter->num_rx_queues || in tsnep_xdp_enable_pool() 28 queue_id >= adapter->num_tx_queues) in tsnep_xdp_enable_pool() 31 queue = &adapter->queue[queue_id]; in tsnep_xdp_enable_pool() 32 if (queue->rx->queue_index != queue_id || in tsnep_xdp_enable_pool() 33 queue->tx->queue_index != queue_id) { in tsnep_xdp_enable_pool() 58 static int tsnep_xdp_disable_pool(struct tsnep_adapter *adapter, u16 queue_id) in tsnep_xdp_disable_pool() argument 63 if (queue_id >= adapter->num_rx_queues || in tsnep_xdp_disable_pool() 64 queue_id >= adapter->num_tx_queues) in tsnep_xdp_disable_pool() 67 pool = xsk_get_pool_from_qid(adapter->netdev, queue_id); in tsnep_xdp_disable_pool() [all …]
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-cmd-queue.h | 170 cvmx_cmd_queue_result_t cvmx_cmd_queue_initialize(cvmx_cmd_queue_id_t queue_id, 183 cvmx_cmd_queue_result_t cvmx_cmd_queue_shutdown(cvmx_cmd_queue_id_t queue_id); 193 int cvmx_cmd_queue_length(cvmx_cmd_queue_id_t queue_id); 205 void *cvmx_cmd_queue_buffer(cvmx_cmd_queue_id_t queue_id); 214 static inline int __cvmx_cmd_queue_get_index(cvmx_cmd_queue_id_t queue_id) in __cvmx_cmd_queue_get_index() argument 223 int unit = queue_id >> 16; in __cvmx_cmd_queue_get_index() 224 int q = (queue_id >> 4) & 0xf; in __cvmx_cmd_queue_get_index() 225 int core = queue_id & 0xf; in __cvmx_cmd_queue_get_index() 236 static inline void __cvmx_cmd_queue_lock(cvmx_cmd_queue_id_t queue_id, in __cvmx_cmd_queue_lock() argument 278 … "=" GCC_OFF_SMALL_ASM()(__cvmx_cmd_queue_state_ptr->ticket[__cvmx_cmd_queue_get_index(queue_id)]), in __cvmx_cmd_queue_lock() [all …]
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_xdp.c | 57 struct xsk_buff_pool *pool, u16 queue_id) in igc_xdp_enable_pool() argument 67 if (queue_id >= adapter->num_rx_queues || in igc_xdp_enable_pool() 68 queue_id >= adapter->num_tx_queues) in igc_xdp_enable_pool() 89 rx_ring = adapter->rx_ring[queue_id]; in igc_xdp_enable_pool() 90 tx_ring = adapter->tx_ring[queue_id]; in igc_xdp_enable_pool() 108 err = igc_xsk_wakeup(ndev, queue_id, XDP_WAKEUP_RX); in igc_xdp_enable_pool() 118 static int igc_xdp_disable_pool(struct igc_adapter *adapter, u16 queue_id) in igc_xdp_disable_pool() argument 125 if (queue_id >= adapter->num_rx_queues || in igc_xdp_disable_pool() 126 queue_id >= adapter->num_tx_queues) in igc_xdp_disable_pool() 129 pool = xsk_get_pool_from_qid(adapter->netdev, queue_id); in igc_xdp_disable_pool() [all …]
|
| /linux/arch/mips/cavium-octeon/executive/ |
| H A D | cvmx-cmd-queue.c | 108 cvmx_cmd_queue_result_t cvmx_cmd_queue_initialize(cvmx_cmd_queue_id_t queue_id, in cvmx_cmd_queue_initialize() argument 117 qstate = __cvmx_cmd_queue_get_state(queue_id); in cvmx_cmd_queue_initialize() 188 ticket[__cvmx_cmd_queue_get_index(queue_id)] = 0; in cvmx_cmd_queue_initialize() 203 cvmx_cmd_queue_result_t cvmx_cmd_queue_shutdown(cvmx_cmd_queue_id_t queue_id) in cvmx_cmd_queue_shutdown() argument 205 __cvmx_cmd_queue_state_t *qptr = __cvmx_cmd_queue_get_state(queue_id); in cvmx_cmd_queue_shutdown() 212 if (cvmx_cmd_queue_length(queue_id) > 0) { in cvmx_cmd_queue_shutdown() 218 __cvmx_cmd_queue_lock(queue_id, qptr); in cvmx_cmd_queue_shutdown() 238 int cvmx_cmd_queue_length(cvmx_cmd_queue_id_t queue_id) in cvmx_cmd_queue_length() argument 241 if (__cvmx_cmd_queue_get_state(queue_id) == NULL) in cvmx_cmd_queue_length() 249 switch ((cvmx_cmd_queue_id_t) (queue_id & 0xff0000)) { in cvmx_cmd_queue_length() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_gfx_v10_3.c | 58 uint32_t queue_id) in acquire_queue() argument 63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 67 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument 70 queue_id; in get_queue_mask() 131 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 161 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 164 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 180 uint32_t pipe_id, uint32_t queue_id, in hqd_load_v10_3() argument 190 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in hqd_load_v10_3() 191 acquire_queue(adev, pipe_id, queue_id); in hqd_load_v10_3() [all …]
|
| H A D | amdgpu_amdkfd_gfx_v11.c | 56 uint32_t queue_id) in acquire_queue() argument 61 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 65 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument 68 queue_id; in get_queue_mask() 127 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 146 + queue_id * (regSDMA0_QUEUE1_RB_CNTL - regSDMA0_QUEUE0_RB_CNTL); in get_sdma_rlc_reg_offset() 149 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 165 uint32_t queue_id, uint32_t __user *wptr, in hqd_load_v11() argument 175 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in hqd_load_v11() 176 acquire_queue(adev, pipe_id, queue_id); in hqd_load_v11() [all …]
|
| H A D | amdgpu_amdkfd_gfx_v10.c | 58 uint32_t queue_id) in acquire_queue() argument 63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 67 uint32_t pipe_id, uint32_t queue_id) in get_queue_mask() argument 70 queue_id; in get_queue_mask() 162 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 178 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 181 queue_id, retval); in get_sdma_rlc_reg_offset() 209 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument 219 pr_debug("Load hqd of pipe %d queue %d\n", pipe_id, queue_id); in kgd_hqd_load() 220 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() [all …]
|
| H A D | amdgpu_amdkfd_gfx_v9.c | 64 uint32_t queue_id, uint32_t inst) in kgd_gfx_v9_acquire_queue() argument 69 kgd_gfx_v9_lock_srbm(adev, mec, pipe, queue_id, 0, inst); in kgd_gfx_v9_acquire_queue() 73 uint32_t pipe_id, uint32_t queue_id) in kgd_gfx_v9_get_queue_mask() argument 76 queue_id; in kgd_gfx_v9_get_queue_mask() 182 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 204 + queue_id * (mmSDMA0_RLC1_RB_CNTL - mmSDMA0_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 207 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 223 uint32_t pipe_id, uint32_t queue_id, in kgd_gfx_v9_hqd_load() argument 234 kgd_gfx_v9_acquire_queue(adev, pipe_id, queue_id, inst); in kgd_gfx_v9_hqd_load() 286 (uint32_t)kgd_gfx_v9_get_queue_mask(adev, pipe_id, queue_id)); in kgd_gfx_v9_hqd_load() [all …]
|
| H A D | amdgpu_amdkfd_gfx_v8.c | 58 uint32_t queue_id) in acquire_queue() argument 63 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 155 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument 166 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 176 mec, pipe, queue_id); in kgd_hqd_load() 179 ((mec << 5) | (pipe << 3) | queue_id | 0x80)); in kgd_hqd_load() 216 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 229 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_dump() argument 245 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_dump() 318 uint32_t engine_id, uint32_t queue_id, in kgd_hqd_sdma_dump() argument [all …]
|
| H A D | amdgpu_amdkfd_gfx_v7.c | 64 uint32_t queue_id) in acquire_queue() argument 69 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 160 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_load() argument 171 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 192 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_load() 205 uint32_t pipe_id, uint32_t queue_id, in kgd_hqd_dump() argument 221 acquire_queue(adev, pipe_id, queue_id); in kgd_hqd_dump() 295 uint32_t engine_id, uint32_t queue_id, in kgd_hqd_sdma_dump() argument 299 queue_id * KFD_CIK_SDMA_QUEUE_OFFSET; in kgd_hqd_sdma_dump() 322 uint32_t queue_id, uint32_t inst) in kgd_hqd_is_occupied() argument [all …]
|
| H A D | amdgpu_amdkfd_gfx_v12.c | 44 uint32_t queue_id) in acquire_queue() argument 49 lock_srbm(adev, mec, pipe, queue_id, 0); in acquire_queue() 78 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 97 + queue_id * (regSDMA0_QUEUE1_RB_CNTL - regSDMA0_QUEUE0_RB_CNTL); in get_sdma_rlc_reg_offset() 100 queue_id, sdma_rlc_reg_offset); in get_sdma_rlc_reg_offset() 106 uint32_t pipe_id, uint32_t queue_id, in hqd_dump_v12() argument 122 acquire_queue(adev, pipe_id, queue_id); in hqd_dump_v12() 137 uint32_t engine_id, uint32_t queue_id, in hqd_sdma_dump_v12() argument 141 engine_id, queue_id); in hqd_sdma_dump_v12()
|
| H A D | amdgpu_userq.c | 909 args->out.queue_id = qid; in amdgpu_userq_create() 1024 queue = __xa_erase(&fpriv->userq_mgr.userq_xa, args->in.queue_id); in amdgpu_userq_ioctl() 1045 unsigned long queue_id; in amdgpu_userq_restore_all() local 1049 xa_for_each(&uq_mgr->userq_xa, queue_id, queue) { in amdgpu_userq_restore_all() 1050 queue = amdgpu_userq_get(uq_mgr, queue_id); in amdgpu_userq_restore_all() 1294 unsigned long queue_id; in amdgpu_userq_evict_all() local 1299 xa_for_each(&uq_mgr->userq_xa, queue_id, queue) { in amdgpu_userq_evict_all() 1300 queue = amdgpu_userq_get(uq_mgr, queue_id); in amdgpu_userq_evict_all() 1335 unsigned long queue_id; in amdgpu_userq_wait_for_signal() local 1338 xa_for_each(&uq_mgr->userq_xa, queue_id, queue) { in amdgpu_userq_wait_for_signal() [all …]
|
| H A D | amdgpu_amdkfd_gc_9_4_3.c | 45 unsigned int queue_id) in get_sdma_rlc_reg_offset() argument 52 queue_id * (regSDMA_RLC1_RB_CNTL - regSDMA_RLC0_RB_CNTL); in get_sdma_rlc_reg_offset() 55 queue_id, retval); in get_sdma_rlc_reg_offset() 129 uint32_t engine_id, uint32_t queue_id, in kgd_gfx_v9_4_3_hqd_sdma_dump() argument 133 engine_id, queue_id); in kgd_gfx_v9_4_3_hqd_sdma_dump() 285 uint32_t pipe_id, uint32_t queue_id, in kgd_gfx_v9_4_3_hqd_load() argument 295 kgd_gfx_v9_acquire_queue(adev, pipe_id, queue_id, inst); in kgd_gfx_v9_4_3_hqd_load() 347 (uint32_t)kgd_gfx_v9_get_queue_mask(adev, pipe_id, queue_id)); in kgd_gfx_v9_4_3_hqd_load()
|
| /linux/drivers/net/wireless/st/cw1200/ |
| H A D | queue.c | 32 queue->queue_id); in __cw1200_queue_lock() 33 ieee80211_stop_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_lock() 43 queue->queue_id); in __cw1200_queue_unlock() 44 ieee80211_wake_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_unlock() 49 u8 *queue_id, u8 *item_generation, in cw1200_queue_parse_id() argument 54 *queue_id = (packet_id >> 16) & 0xFF; in cw1200_queue_parse_id() 58 static inline u32 cw1200_queue_mk_packet_id(u8 queue_generation, u8 queue_id, in cw1200_queue_mk_packet_id() argument 63 ((u32)queue_id << 16) | in cw1200_queue_mk_packet_id() 165 u8 queue_id, in cw1200_queue_init() argument 174 queue->queue_id = queue_id; in cw1200_queue_init() [all …]
|
| /linux/drivers/mailbox/ |
| H A D | ti-msgmgr.c | 47 u8 queue_id; member 108 u8 queue_id; member 473 "rx_%03d", d->is_sproxy ? qinst->proxy_id : qinst->queue_id); in ti_msgmgr_queue_rx_irq_req() 486 qinst->queue_id, qinst->proxy_id, in ti_msgmgr_queue_rx_irq_req() 614 if (req_qid == qinst->queue_id && req_pid == qinst->proxy_id) in ti_msgmgr_of_xlate() 648 qinst->queue_id = qd->queue_id; in ti_msgmgr_queue_setup() 650 if (qinst->queue_id > d->queue_count) { in ti_msgmgr_queue_setup() 652 idx, qinst->queue_id, d->queue_count); in ti_msgmgr_queue_setup() 674 Q_DATA_OFFSET(qinst->proxy_id, qinst->queue_id, in ti_msgmgr_queue_setup() 677 Q_DATA_OFFSET(qinst->proxy_id, qinst->queue_id, in ti_msgmgr_queue_setup() [all …]
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_mqd_manager.c | 211 uint32_t pipe_id, uint32_t queue_id, in kfd_hiq_load_mqd_kiq() argument 215 queue_id, p->doorbell_off, 0); in kfd_hiq_load_mqd_kiq() 220 uint32_t pipe_id, uint32_t queue_id) in kfd_destroy_mqd_cp() argument 223 pipe_id, queue_id, 0); in kfd_destroy_mqd_cp() 239 uint32_t queue_id) in kfd_is_occupied_cp() argument 242 pipe_id, queue_id, 0); in kfd_is_occupied_cp() 246 uint32_t pipe_id, uint32_t queue_id, in kfd_load_mqd_sdma() argument 261 uint32_t queue_id) in kfd_destroy_mqd_sdma() argument 268 uint32_t queue_id) in kfd_is_occupied_sdma() argument
|
| /linux/drivers/net/wireless/realtek/rtlwifi/rtl8192cu/ |
| H A D | phy.c | 381 u8 i, queue_id; in _rtl92cu_phy_set_rf_power_state() local 416 for (queue_id = 0, i = 0; in _rtl92cu_phy_set_rf_power_state() 417 queue_id < RTL_PCI_MAX_TX_QUEUE_COUNT;) { in _rtl92cu_phy_set_rf_power_state() 418 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state() 420 queue_id == BEACON_QUEUE) { in _rtl92cu_phy_set_rf_power_state() 421 queue_id++; in _rtl92cu_phy_set_rf_power_state() 427 queue_id, in _rtl92cu_phy_set_rf_power_state() 436 queue_id, in _rtl92cu_phy_set_rf_power_state() 459 for (queue_id = 0, i = 0; in _rtl92cu_phy_set_rf_power_state() 460 queue_id < RTL_PCI_MAX_TX_QUEUE_COUNT;) { in _rtl92cu_phy_set_rf_power_state() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | nfp_net_xsk.c | 111 struct xsk_buff_pool *pool, u16 queue_id) in nfp_net_xsk_setup_pool() argument 143 prev_pool = dp->xsk_pools[queue_id]; in nfp_net_xsk_setup_pool() 144 dp->xsk_pools[queue_id] = pool; in nfp_net_xsk_setup_pool() 162 int nfp_net_xsk_wakeup(struct net_device *netdev, u32 queue_id, u32 flags) in nfp_net_xsk_wakeup() argument 171 napi_schedule(&nn->r_vecs[queue_id].napi); in nfp_net_xsk_wakeup()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | bwc.c | 13 hws_bwc_get_burst_th(struct mlx5hws_context *ctx, u16 queue_id) in hws_bwc_get_burst_th() argument 15 return min(ctx->send_queue[queue_id].num_entries / 2, in hws_bwc_get_burst_th() 90 rule_attr.queue_id = mlx5hws_bwc_get_queue_id(ctx, i); in hws_bwc_matcher_move_all_simple() 110 hws_bwc_get_burst_th(ctx, rule_attr.queue_id); in hws_bwc_matcher_move_all_simple() 112 rule_attr.queue_id, in hws_bwc_matcher_move_all_simple() 132 queue = &ctx->send_queue[rule_attr.queue_id]; in hws_bwc_matcher_move_all_simple() 135 rule_attr.queue_id, in hws_bwc_matcher_move_all_simple() 407 u16 queue_id, in mlx5hws_bwc_queue_poll() argument 414 u16 burst_th = hws_bwc_get_burst_th(ctx, queue_id); in mlx5hws_bwc_queue_poll() 425 queue_full = mlx5hws_send_engine_full(&ctx->send_queue[queue_id]); in mlx5hws_bwc_queue_poll() [all …]
|
| /linux/drivers/media/platform/qcom/iris/ |
| H A D | iris_hfi_queue.c | 199 static void iris_hfi_queue_set_header(struct iris_core *core, u32 queue_id, in iris_hfi_queue_set_header() argument 205 iface_q->qhdr->queue_type = queue_id; in iris_hfi_queue_set_header() 221 if (queue_id == IFACEQ_DBGQ_ID) in iris_hfi_queue_set_header() 226 iris_hfi_queue_init(struct iris_core *core, u32 queue_id, struct iris_iface_q_info *iface_q) in iris_hfi_queue_init() argument 229 u32 offset = sizeof(*q_tbl_hdr) + (queue_id * IFACEQ_QUEUE_SIZE); in iris_hfi_queue_init() 234 iface_q->qhdr = &q_tbl_hdr->q_hdr[queue_id]; in iris_hfi_queue_init() 236 iris_hfi_queue_set_header(core, queue_id, iface_q); in iris_hfi_queue_init()
|
| /linux/net/xdp/ |
| H A D | xsk_buff_pool.c | 148 bpf.xsk.queue_id = pool->queue_id; in xp_disable_drv_zc() 158 struct net_device *netdev, u16 queue_id, u16 flags) in xp_assign_dev() argument 172 if (xsk_get_pool_from_qid(netdev, queue_id)) in xp_assign_dev() 176 pool->queue_id = queue_id; in xp_assign_dev() 177 err = xsk_reg_pool_at_qid(netdev, pool, queue_id); in xp_assign_dev() 215 bpf.xsk.queue_id = queue_id; in xp_assign_dev() 237 xsk_clear_pool_at_qid(netdev, queue_id); in xp_assign_dev() 244 struct net_device *dev, u16 queue_id) in xp_assign_dev_shared() argument 253 return xp_assign_dev(pool, dev, queue_id, flags); in xp_assign_dev_shared() 265 xsk_clear_pool_at_qid(pool->netdev, pool->queue_id); in xp_clear_dev()
|
| H A D | xsk.c | 107 u16 queue_id) in xsk_get_pool_from_qid() argument 109 if (queue_id < dev->real_num_rx_queues) in xsk_get_pool_from_qid() 110 return dev->_rx[queue_id].pool; in xsk_get_pool_from_qid() 111 if (queue_id < dev->real_num_tx_queues) in xsk_get_pool_from_qid() 112 return dev->_tx[queue_id].pool; in xsk_get_pool_from_qid() 118 void xsk_clear_pool_at_qid(struct net_device *dev, u16 queue_id) in xsk_clear_pool_at_qid() argument 120 if (queue_id < dev->num_rx_queues) in xsk_clear_pool_at_qid() 121 dev->_rx[queue_id].pool = NULL; in xsk_clear_pool_at_qid() 122 if (queue_id < dev->num_tx_queues) in xsk_clear_pool_at_qid() 123 dev->_tx[queue_id].pool = NULL; in xsk_clear_pool_at_qid() [all …]
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | xsk.c | 71 __u32 queue_id; member 459 __u32 queue_id) in xsk_get_ctx() argument 467 if (ctx->ifindex == ifindex && ctx->queue_id == queue_id) { in xsk_get_ctx() 504 __u32 queue_id, in xsk_create_ctx() argument 530 ctx->queue_id = queue_id; in xsk_create_ctx() 540 __u32 queue_id, struct xsk_umem *umem, in xsk_socket__create_shared() argument 580 ctx = xsk_get_ctx(umem, ifindex, queue_id); in xsk_socket__create_shared() 587 ctx = xsk_create_ctx(xsk, umem, ifindex, queue_id, fill, comp); in xsk_socket__create_shared() 671 sxdp.sxdp_queue_id = ctx->queue_id; in xsk_socket__create_shared() 709 __u32 queue_id, struct xsk_umem *umem, in xsk_socket__create() argument [all …]
|
| /linux/drivers/net/wireless/realtek/rtlwifi/rtl8192ce/ |
| H A D | phy.c | 407 u8 i, queue_id; in _rtl92ce_phy_set_rf_power_state() local 462 for (queue_id = 0, i = 0; in _rtl92ce_phy_set_rf_power_state() 463 queue_id < RTL_PCI_MAX_TX_QUEUE_COUNT;) { in _rtl92ce_phy_set_rf_power_state() 464 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92ce_phy_set_rf_power_state() 465 if (queue_id == BEACON_QUEUE || in _rtl92ce_phy_set_rf_power_state() 467 queue_id++; in _rtl92ce_phy_set_rf_power_state() 472 i + 1, queue_id, in _rtl92ce_phy_set_rf_power_state() 482 queue_id, in _rtl92ce_phy_set_rf_power_state()
|
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt_dcb.c | 25 static int bnxt_queue_to_tc(struct bnxt *bp, u8 queue_id) in bnxt_queue_to_tc() argument 30 if (bp->q_info[i].queue_id == queue_id) { in bnxt_queue_to_tc() 61 pri2cos[i] = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_pri2cos_cfg() 84 u8 queue_id = pri2cos[i]; in bnxt_hwrm_queue_pri2cos_qcfg() local 87 tc = bnxt_queue_to_tc(bp, queue_id); in bnxt_hwrm_queue_pri2cos_qcfg() 115 cos2bw.queue_id = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_cos2bw_cfg() 132 req->queue_id0 = cos2bw.queue_id; in bnxt_hwrm_queue_cos2bw_cfg() 167 cos2bw.queue_id = resp->queue_id0; in bnxt_hwrm_queue_cos2bw_qcfg() 177 tc = bnxt_queue_to_tc(bp, cos2bw.queue_id); in bnxt_hwrm_queue_cos2bw_qcfg()
|