Searched refs:node_mask (Results 1 – 8 of 8) sorted by relevance
| /linux/tools/perf/util/ |
| H A D | mmap.c | 102 unsigned long *node_mask; in perf_mmap__aio_bind() local 110 node_mask = bitmap_zalloc(node_index + 1); in perf_mmap__aio_bind() 111 if (!node_mask) { in perf_mmap__aio_bind() 115 __set_bit(node_index, node_mask); in perf_mmap__aio_bind() 116 if (mbind(data, mmap_len, MPOL_BIND, node_mask, node_index + 1 + 1, 0)) { in perf_mmap__aio_bind() 121 bitmap_free(node_mask); in perf_mmap__aio_bind()
|
| /linux/net/netfilter/ |
| H A D | xt_cluster.c | 118 return !!((1 << hash) & info->node_mask) ^ in xt_cluster_mt() 132 if (info->node_mask >= (1ULL << info->total_nodes)) { in xt_cluster_mt_checkentry()
|
| /linux/include/uapi/linux/netfilter/ |
| H A D | xt_cluster.h | 13 __u32 node_mask; member
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | affinity.c | 981 const struct cpumask *node_mask, in hfi1_get_proc_affinity() local 1089 node_mask = cpumask_of_node(node); in hfi1_get_proc_affinity() 1091 cpumask_pr_args(node_mask)); in hfi1_get_proc_affinity() 1094 cpumask_and(available_mask, hw_thread_mask, node_mask); in hfi1_get_proc_affinity() 1122 cpumask_andnot(available_mask, available_mask, node_mask); in hfi1_get_proc_affinity()
|
| /linux/drivers/scsi/ |
| H A D | storvsc_drv.c | 1401 const struct cpumask *node_mask; in get_og_chn() local 1423 node_mask = cpumask_of_node(cpu_to_node(q_num)); in get_og_chn() 1427 if (cpumask_test_cpu(tgt_cpu, node_mask)) in get_og_chn() 1440 if (!cpumask_test_cpu(tgt_cpu, node_mask)) in get_og_chn() 1461 const struct cpumask *node_mask; in storvsc_do_io() local 1485 node_mask = cpumask_of_node(cpu_to_node(q_num)); in storvsc_do_io() 1488 if (!cpumask_test_cpu(tgt_cpu, node_mask)) in storvsc_do_io()
|
| /linux/kernel/sched/ |
| H A D | ext_idle.c | 790 const struct cpumask *node_mask = cpumask_of_node(node); in reset_idle_masks() local 792 cpumask_and(idle_cpumask(node)->cpu, cpu_online_mask, node_mask); in reset_idle_masks() 793 cpumask_and(idle_cpumask(node)->smt, cpu_online_mask, node_mask); in reset_idle_masks()
|
| H A D | ext.c | 3986 const struct cpumask *node_mask = cpumask_of_node(node); in bypass_lb_node() local 3996 for_each_cpu_and(cpu, cpu_online_mask, node_mask) { in bypass_lb_node() 4018 for_each_cpu_and(cpu, cpu_online_mask, node_mask) { in bypass_lb_node() 4025 for_each_cpu_and(cpu, cpu_online_mask, node_mask) { in bypass_lb_node() 4043 for_each_cpu_and(cpu, cpu_online_mask, node_mask) { in bypass_lb_node()
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_main.c | 482 const struct cpumask *node_mask; in gve_alloc_notify_blocks() local 550 node_mask = gve_get_node_mask(priv); in gve_alloc_notify_blocks() 551 cur_cpu = cpumask_first(node_mask); in gve_alloc_notify_blocks() 572 cur_cpu = cpumask_next(cur_cpu, node_mask); in gve_alloc_notify_blocks() 578 cur_cpu = cpumask_first(node_mask); in gve_alloc_notify_blocks()
|