| /linux/tools/perf/util/ |
| H A D | mmap.c | 232 bitmap_free(map->affinity_mask.bits); in mmap__munmap() 264 map->affinity_mask.nbits = cpu__max_cpu().cpu; in perf_mmap__setup_affinity_mask() 265 map->affinity_mask.bits = bitmap_zalloc(map->affinity_mask.nbits); in perf_mmap__setup_affinity_mask() 266 if (!map->affinity_mask.bits) in perf_mmap__setup_affinity_mask() 270 build_node_mask(cpu__get_node(map->core.cpu), &map->affinity_mask); in perf_mmap__setup_affinity_mask() 272 __set_bit(map->core.cpu.cpu, map->affinity_mask.bits); in perf_mmap__setup_affinity_mask() 293 mmap_cpu_mask__scnprintf(&map->affinity_mask, "mmap"); in mmap__mmap()
|
| H A D | mmap.h | 40 struct mmap_cpu_mask affinity_mask; member
|
| /linux/tools/tracing/rtla/sample/ |
| H A D | timerlat_load.py | |
| /linux/drivers/net/ethernet/aquantia/atlantic/ |
| H A D | aq_pci_func.h | 25 void *irq_arg, cpumask_t *affinity_mask);
|
| H A D | aq_pci_func.c | 149 void *irq_arg, cpumask_t *affinity_mask) in aq_pci_func_alloc_irq() argument 164 if (pdev->msix_enabled && affinity_mask) in aq_pci_func_alloc_irq() 166 affinity_mask); in aq_pci_func_alloc_irq()
|
| H A D | aq_vec.c | 117 &self->aq_ring_param.affinity_mask); in aq_vec_alloc() 360 return &self->aq_ring_param.affinity_mask; in aq_vec_get_affinity_mask()
|
| H A D | aq_ring.h | 160 cpumask_t affinity_mask; member
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_txrx.h | 223 cpumask_t affinity_mask; /* IRQ affinity */ member 244 return cpu_to_mem(cpumask_first(&p->affinity_mask)); in fun_irq_node()
|
| /linux/include/sound/sof/ |
| H A D | ext_manifest4.h | 102 uint32_t affinity_mask; member
|
| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_main.c | 360 free_cpumask_var(cptvf->affinity_mask[vec]); in cptvf_free_irq_affinity() 617 if (!zalloc_cpumask_var(&cptvf->affinity_mask[vec], in cptvf_set_irq_affinity() 626 cptvf->affinity_mask[vec]); in cptvf_set_irq_affinity() 628 cptvf->affinity_mask[vec]); in cptvf_set_irq_affinity()
|
| H A D | cptvf.h | 107 cpumask_var_t affinity_mask[CPT_VF_MSIX_VECTORS]; member
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf.h | 84 cpumask_var_t affinity_mask[OTX_CPT_VF_MSIX_VECTORS]; member
|
| H A D | otx_cptvf_main.c | 343 free_cpumask_var(cptvf->affinity_mask[vec]); in cptvf_free_irq_affinity() 600 if (!zalloc_cpumask_var(&cptvf->affinity_mask[vec], in cptvf_set_irq_affinity() 610 cptvf->affinity_mask[vec]); in cptvf_set_irq_affinity() 612 cptvf->affinity_mask[vec]); in cptvf_set_irq_affinity()
|
| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_irq.c | 122 irq_set_affinity_hint(irq_cfg->irq_id, &irq_cfg->affinity_mask); in hinic3_request_irq() 220 cpumask_set_cpu(local_cpu, &irq_cfg->affinity_mask); in hinic3_qps_irq_init()
|
| /linux/drivers/net/ethernet/cisco/enic/ |
| H A D | enic_main.c | 119 (cpumask_available(enic->msix[i].affinity_mask) && in enic_init_affinity_hint() 120 !cpumask_empty(enic->msix[i].affinity_mask))) in enic_init_affinity_hint() 122 if (zalloc_cpumask_var(&enic->msix[i].affinity_mask, in enic_init_affinity_hint() 125 enic->msix[i].affinity_mask); in enic_init_affinity_hint() 136 free_cpumask_var(enic->msix[i].affinity_mask); in enic_free_affinity_hint() 148 !cpumask_available(enic->msix[i].affinity_mask) || in enic_set_affinity_hint() 149 cpumask_empty(enic->msix[i].affinity_mask)) in enic_set_affinity_hint() 152 enic->msix[i].affinity_mask); in enic_set_affinity_hint() 161 if (cpumask_available(enic->msix[wq_intr].affinity_mask) && in enic_set_affinity_hint() 162 !cpumask_empty(enic->msix[wq_intr].affinity_mask)) in enic_set_affinity_hint() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_qp.h | 105 struct cpumask affinity_mask; member
|
| H A D | hinic_rx.c | 549 cpumask_set_cpu(qp->q_id % num_online_cpus(), &rq->affinity_mask); in rx_request_irq() 550 err = irq_set_affinity_and_hint(rq->irq, &rq->affinity_mask); in rx_request_irq()
|
| /linux/drivers/net/ethernet/netronome/nfp/nfdk/ |
| H A D | rings.c | 113 netif_set_xps_queue(dp->netdev, &r_vec->affinity_mask, in nfp_nfdk_tx_ring_alloc()
|
| /linux/drivers/net/ethernet/netronome/nfp/nfd3/ |
| H A D | rings.c | 148 netif_set_xps_queue(dp->netdev, &r_vec->affinity_mask, in nfp_nfd3_tx_ring_alloc()
|
| /linux/drivers/net/ethernet/cavium/thunder/ |
| H A D | nicvf_queues.h | 283 cpumask_t affinity_mask; member
|
| H A D | nic.h | 375 cpumask_var_t affinity_mask[NIC_VF_MSIX_VECTORS]; member
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | eq.c | 243 if (!cpumask_available(eq->affinity_mask) || in mlx4_set_eq_affinity_hint() 244 cpumask_empty(eq->affinity_mask)) in mlx4_set_eq_affinity_hint() 247 hint_err = irq_update_affinity_hint(eq->irq, eq->affinity_mask); in mlx4_set_eq_affinity_hint() 1125 free_cpumask_var(eq_table->eq[i].affinity_mask); in mlx4_free_irqs()
|
| /linux/drivers/net/ethernet/pensando/ionic/ |
| H A D | ionic_lif.c | 283 cpumask_copy(*intr->affinity_mask, mask); in ionic_irq_aff_notify() 327 *qcq->intr.affinity_mask); in ionic_qcq_enable() 506 cpumask_var_t *affinity_mask; in ionic_alloc_qcq_interrupt() local 531 affinity_mask = &lif->ionic->affinity_masks[qcq->intr.index]; in ionic_alloc_qcq_interrupt() 532 if (cpumask_empty(*affinity_mask)) { in ionic_alloc_qcq_interrupt() 538 cpumask_set_cpu(cpu, *affinity_mask); in ionic_alloc_qcq_interrupt() 541 qcq->intr.affinity_mask = affinity_mask; in ionic_alloc_qcq_interrupt() 3593 *qcq->intr.affinity_mask); in ionic_lif_adminq_init()
|
| /linux/drivers/net/ethernet/marvell/octeon_ep_vf/ |
| H A D | octep_vf_main.c | 212 &ioq_vector->affinity_mask); in octep_vf_request_irqs() 214 &ioq_vector->affinity_mask); in octep_vf_request_irqs()
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | octeon_device.h | 413 struct cpumask affinity_mask; member
|