Searched refs:allowed_mask (Results 1 – 9 of 9) sorted by relevance
| /linux/tools/testing/selftests/kvm/ |
| H A D | rseq_test.c | 77 cpu_set_t allowed_mask; in migration_worker() local 80 CPU_ZERO(&allowed_mask); in migration_worker() 83 CPU_SET(cpu, &allowed_mask); in migration_worker() 99 r = sched_setaffinity(rseq_tid, sizeof(allowed_mask), &allowed_mask); in migration_worker() 105 CPU_CLR(cpu, &allowed_mask); in migration_worker()
|
| /linux/include/linux/ |
| H A D | memory-tiers.h | 56 int next_demotion_node(int node, const nodemask_t *allowed_mask); 60 static inline int next_demotion_node(int node, const nodemask_t *allowed_mask) in next_demotion_node() argument 104 static inline int next_demotion_node(int node, const nodemask_t *allowed_mask) in next_demotion_node() argument
|
| /linux/io_uring/ |
| H A D | sqpoll.c | 494 cpumask_var_t allowed_mask; in io_sq_offload_create() local 501 if (!alloc_cpumask_var(&allowed_mask, GFP_KERNEL)) in io_sq_offload_create() 504 cpuset_cpus_allowed(current, allowed_mask); in io_sq_offload_create() 505 if (!cpumask_test_cpu(cpu, allowed_mask)) { in io_sq_offload_create() 506 free_cpumask_var(allowed_mask); in io_sq_offload_create() 509 free_cpumask_var(allowed_mask); in io_sq_offload_create()
|
| H A D | io-wq.c | 1447 cpumask_var_t allowed_mask; in io_wq_cpu_affinity() local 1453 if (!alloc_cpumask_var(&allowed_mask, GFP_KERNEL)) in io_wq_cpu_affinity() 1457 cpuset_cpus_allowed(tctx->io_wq->task, allowed_mask); in io_wq_cpu_affinity() 1459 if (cpumask_subset(mask, allowed_mask)) in io_wq_cpu_affinity() 1464 cpumask_copy(tctx->io_wq->cpu_mask, allowed_mask); in io_wq_cpu_affinity() 1468 free_cpumask_var(allowed_mask); in io_wq_cpu_affinity()
|
| /linux/mm/ |
| H A D | memory-tiers.c | 330 int next_demotion_node(int node, const nodemask_t *allowed_mask) in next_demotion_node() argument 349 nodes_and(mask, nd->preferred, *allowed_mask); in next_demotion_node() 372 nodes_complement(mask, *allowed_mask); in next_demotion_node()
|
| H A D | vmscan.c | 347 nodemask_t allowed_mask; in can_demote() local 354 node_get_allowed_targets(pgdat, &allowed_mask); in can_demote() 355 if (nodes_empty(allowed_mask)) in can_demote() 359 mem_cgroup_node_filter_allowed(memcg, &allowed_mask); in can_demote() 360 return !nodes_empty(allowed_mask); in can_demote() 990 nodemask_t *allowed_mask; in alloc_demote_folio() local 995 allowed_mask = mtc->nmask; in alloc_demote_folio() 1012 mtc->nmask = allowed_mask; in alloc_demote_folio() 1027 nodemask_t allowed_mask; in demote_folio_list() local 1037 .nmask = &allowed_mask, in demote_folio_list() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | kvm_util.c | 639 static uint32_t parse_pcpu(const char *cpu_str, const cpu_set_t *allowed_mask) in kvm_parse_vcpu_pinning() 643 TEST_ASSERT(CPU_ISSET(pcpu, allowed_mask), in kvm_parse_vcpu_pinning() 668 cpu_set_t allowed_mask; 676 r = sched_getaffinity(0, sizeof(allowed_mask), &allowed_mask); 684 vcpu_to_pcpu[i] = parse_pcpu(cpu, &allowed_mask); 690 pin_self_to_cpu(parse_pcpu(cpu, &allowed_mask)); in userspace_mem_region_find() 606 parse_pcpu(const char * cpu_str,const cpu_set_t * allowed_mask) parse_pcpu() argument 635 cpu_set_t allowed_mask; kvm_parse_vcpu_pinning() local
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | fs_hws.c | 1132 int allowed_mask = BIT(MLX5_SET_FTE_MODIFY_ENABLE_MASK_ACTION) | in mlx5_cmd_hws_update_fte() local 1144 if ((modify_mask & ~allowed_mask) != 0) in mlx5_cmd_hws_update_fte()
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_drv.c | 811 arg->allowed_mask |= BIT(prio); in panthor_query_group_priorities_info()
|