/linux/block/ |
H A D | blk-mq-cpumap.c | 18 void blk_mq_map_queues(struct blk_mq_queue_map *qmap) in blk_mq_map_queues() argument 23 masks = group_cpus_evenly(qmap->nr_queues); in blk_mq_map_queues() 26 qmap->mq_map[cpu] = qmap->queue_offset; in blk_mq_map_queues() 30 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_queues() 32 qmap->mq_map[cpu] = qmap->queue_offset + queue; in blk_mq_map_queues() 46 int blk_mq_hw_queue_to_node(struct blk_mq_queue_map *qmap, unsigned int index) in blk_mq_hw_queue_to_node() argument 51 if (index == qmap->mq_map[i]) in blk_mq_hw_queue_to_node()
|
H A D | blk-mq-pci.c | 25 void blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev, in blk_mq_pci_map_queues() argument 31 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_pci_map_queues() 37 qmap->mq_map[cpu] = qmap->queue_offset + queue; in blk_mq_pci_map_queues() 43 WARN_ON_ONCE(qmap->nr_queues > 1); in blk_mq_pci_map_queues() 44 blk_mq_clear_mq_map(qmap); in blk_mq_pci_map_queues()
|
H A D | blk-mq-virtio.c | 23 void blk_mq_virtio_map_queues(struct blk_mq_queue_map *qmap, in blk_mq_virtio_map_queues() argument 32 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_virtio_map_queues() 38 qmap->mq_map[cpu] = qmap->queue_offset + queue; in blk_mq_virtio_map_queues() 44 blk_mq_map_queues(qmap); in blk_mq_virtio_map_queues()
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-class-net-qmi | 32 Write a number ranging from 1 to 254 to add a qmap mux 50 created qmap mux based network device. 52 What: /sys/class/net/<qmimux iface>/qmap/mux_id
|
/linux/include/linux/ |
H A D | blk-mq-pci.h | 8 void blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev,
|
H A D | blk-mq-virtio.h | 8 void blk_mq_virtio_map_queues(struct blk_mq_queue_map *qmap,
|
/linux/drivers/net/ipa/data/ |
H A D | ipa_data-v4.2.c | 108 .qmap = true, 133 .qmap = true,
|
H A D | ipa_data-v4.11.c | 112 .qmap = true, 136 .qmap = true,
|
H A D | ipa_data-v4.7.c | 109 .qmap = true, 132 .qmap = true,
|
H A D | ipa_data-v3.5.1.c | 118 .qmap = true, 143 .qmap = true,
|
H A D | ipa_data-v4.5.c | 121 .qmap = true, 145 .qmap = true,
|
H A D | ipa_data-v5.0.c | 125 .qmap = true, 149 .qmap = true,
|
H A D | ipa_data-v4.9.c | 113 .qmap = true, 137 .qmap = true,
|
H A D | ipa_data-v5.5.c | 125 .qmap = true, 149 .qmap = true,
|
H A D | ipa_data-v3.1.c | 127 .qmap = true, 151 .qmap = true,
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_dcb.c | 195 unsigned long qmap = 0; in bnxt_queue_remap() local 204 __set_bit(j, &qmap); in bnxt_queue_remap() 216 j = find_next_zero_bit(&qmap, max, j); in bnxt_queue_remap() 218 __set_bit(j, &qmap); in bnxt_queue_remap()
|
/linux/drivers/net/ipa/ |
H A D | ipa_endpoint.h | 113 bool qmap; member
|
H A D | ipa_endpoint.c | 793 if (endpoint->config.qmap) { in ipa_endpoint_init_hdr() 838 if (endpoint->config.qmap) { in ipa_endpoint_init_hdr_ext() 866 if (endpoint->config.qmap && !endpoint->toward_ipa) { in ipa_endpoint_init_hdr_ext() 896 if (endpoint->config.qmap) in ipa_endpoint_init_hdr_metadata_mask()
|
H A D | ipa_modem.c | 137 if (endpoint->config.qmap && skb->protocol != htons(ETH_P_MAP)) in ipa_start_xmit()
|
/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_lib.c | 988 u16 offset = 0, qmap = 0, tx_count = 0, rx_count = 0, pow = 0; in ice_vsi_setup_q_map() local 1039 qmap = FIELD_PREP(ICE_AQ_VSI_TC_Q_OFFSET_M, offset); in ice_vsi_setup_q_map() 1040 qmap |= FIELD_PREP(ICE_AQ_VSI_TC_Q_NUM_M, pow); in ice_vsi_setup_q_map() 1043 ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); in ice_vsi_setup_q_map() 1177 u16 qcount, qmap; in ice_chnl_vsi_setup_q_map() local 1184 qmap = FIELD_PREP(ICE_AQ_VSI_TC_Q_OFFSET_M, offset); in ice_chnl_vsi_setup_q_map() 1185 qmap |= FIELD_PREP(ICE_AQ_VSI_TC_Q_NUM_M, pow); in ice_chnl_vsi_setup_q_map() 1187 ctxt->info.tc_mapping[0] = cpu_to_le16(qmap); in ice_chnl_vsi_setup_q_map() 3196 u16 pow, offset = 0, qcount_tx = 0, qcount_rx = 0, qmap; in ice_vsi_setup_q_map_mqprio() local 3206 qmap = FIELD_PREP(ICE_AQ_VSI_TC_Q_OFFSET_M, tc0_offset); in ice_vsi_setup_q_map_mqprio() [all …]
|
H A D | ice_virtchnl.c | 1710 unsigned long qmap; in ice_cfg_interrupt() local 1715 qmap = map->rxq_map; in ice_cfg_interrupt() 1716 for_each_set_bit(vsi_q_id_idx, &qmap, ICE_MAX_RSS_QS_PER_VF) { in ice_cfg_interrupt() 1730 qmap = map->txq_map; in ice_cfg_interrupt() 1731 for_each_set_bit(vsi_q_id_idx, &qmap, ICE_MAX_RSS_QS_PER_VF) { in ice_cfg_interrupt()
|
/linux/tools/sched_ext/ |
H A D | scx_qmap.bpf.c | 51 struct qmap { struct 65 __array(values, struct qmap);
|
/linux/arch/powerpc/kvm/ |
H A D | book3s_xive.h | 132 u8 qmap; member
|
H A D | book3s_xive.c | 982 if (xive->qmap & (1 << prio)) in xive_check_provisioning() 1001 xive->qmap |= (1 << prio); in xive_check_provisioning() 1995 if (xive->qmap & (1 << i)) { in kvmppc_xive_connect_vcpu()
|
/linux/drivers/scsi/fnic/ |
H A D | fnic_main.c | 587 struct blk_mq_queue_map *qmap = &host->tag_set.map[HCTX_TYPE_DEFAULT]; in fnic_mq_map_queues_cpus() local 596 "qmap->nr_queues: %d\n", qmap->nr_queues); in fnic_mq_map_queues_cpus() 604 blk_mq_pci_map_queues(qmap, l_pdev, FNIC_PCI_OFFSET); in fnic_mq_map_queues_cpus()
|