| /linux/arch/x86/mm/ |
| H A D | numa.c | 115 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 119 for (node = 0; node < nr_node_ids; node++) in setup_node_to_cpumask_map() 123 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 419 if ((unsigned)node >= nr_node_ids) { in cpumask_of_node() 422 node, nr_node_ids); in cpumask_of_node()
|
| /linux/scripts/gdb/linux/ |
| H A D | slab.py | 183 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids")) 185 nr_node_ids = 1 198 for i in range(0, nr_node_ids): 278 nr_node_ids = None 294 nr_node_ids = int(gdb.parse_and_eval("nr_node_ids")) 296 nr_node_ids = 1 302 for i in range(0, nr_node_ids):
|
| /linux/lib/ |
| H A D | group_cpus.c | 50 masks = kzalloc_objs(cpumask_var_t, nr_node_ids); in alloc_node_to_cpumask() 54 for (node = 0; node < nr_node_ids; node++) { in alloc_node_to_cpumask() 72 for (node = 0; node < nr_node_ids; node++) in free_node_to_cpumask() 237 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_groups() 255 alloc_groups_to_nodes(numgrps, numcpus, node_groups, nr_node_ids); in alloc_nodes_groups() 435 node_groups = kzalloc_objs(struct node_groups, nr_node_ids); in __group_cpus_evenly() 442 for (i = 0; i < nr_node_ids; i++) { in __group_cpus_evenly()
|
| /linux/drivers/base/ |
| H A D | arch_numa.c | 50 if (WARN_ON(node < 0 || node >= nr_node_ids)) in cpumask_of_node() 105 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 109 for (node = 0; node < nr_node_ids; node++) { in setup_node_to_cpumask_map() 115 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
|
| /linux/mm/ |
| H A D | mempolicy.c | 228 new_bw = kcalloc(nr_node_ids, sizeof(unsigned int), GFP_KERNEL); in mempolicy_set_node_perf() 232 new_wi_state = kmalloc_flex(*new_wi_state, iw_table, nr_node_ids); in mempolicy_set_node_perf() 238 for (i = 0; i < nr_node_ids; i++) in mempolicy_set_node_perf() 250 memcpy(new_bw, old_bw, nr_node_ids * sizeof(*old_bw)); in mempolicy_set_node_perf() 1692 unsigned int nbytes = BITS_TO_LONGS(nr_node_ids) * sizeof(long); in copy_nodes_to_user() 1696 nbytes = BITS_TO_COMPAT_LONGS(nr_node_ids) * sizeof(compat_long_t); in copy_nodes_to_user() 1704 maxnode = nr_node_ids; in copy_nodes_to_user() 1958 if (nmask != NULL && maxnode < nr_node_ids) in kernel_get_mempolicy() 2680 weights = kzalloc(nr_node_ids, GFP_KERNEL); in alloc_pages_bulk_weighted_interleave() 2687 memcpy(weights, state->iw_table, nr_node_ids * sizeof(u8)); in alloc_pages_bulk_weighted_interleave() [all …]
|
| H A D | list_lru.c | 410 mlru = kmalloc_flex(*mlru, node, nr_node_ids, gfp); in memcg_init_list_lru_one() 588 lru->node = kzalloc_objs(*lru->node, nr_node_ids); in __list_lru_init()
|
| H A D | hugetlb_sysfs.c | 377 for (nid = 0; nid < nr_node_ids; nid++) { in kobj_to_node_hstate()
|
| H A D | ksm.c | 3589 buf = kzalloc_objs(*buf, nr_node_ids + nr_node_ids); in merge_across_nodes_store() 3595 root_unstable_tree = buf + nr_node_ids; in merge_across_nodes_store() 3602 ksm_nr_node_ids = knob ? 1 : nr_node_ids; in merge_across_nodes_store()
|
| H A D | shrinker.c | 722 size *= nr_node_ids; in shrinker_alloc()
|
| H A D | memory-tiers.c | 915 node_demotion = kzalloc_objs(struct demotion_nodes, nr_node_ids); in memory_tier_init()
|
| H A D | hugetlb_cgroup.c | 142 h_cgroup = kzalloc_flex(*h_cgroup, nodeinfo, nr_node_ids); in hugetlb_cgroup_css_alloc()
|
| H A D | slub.c | 466 for (__node = 0; __node < nr_node_ids; __node++) \ 8017 if (nr_node_ids > 1) { in setup_slab_strict_numa() 8395 nr_node_ids * sizeof(struct kmem_cache_node *), in kmem_cache_init() 8416 nr_cpu_ids, nr_node_ids); in kmem_cache_init() 8815 nodes = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); in show_slab_objects() 8865 for (node = 0; node < nr_node_ids; node++) { in show_slab_objects()
|
| /linux/arch/loongarch/kernel/ |
| H A D | numa.c | 72 if (nr_node_ids >= 8) in setup_per_cpu_areas() 255 loongson_sysconf.nr_nodes = nr_node_ids; in init_numa_memory()
|
| /linux/tools/workqueue/ |
| H A D | wq_dump.py | 223 nr_node_ids = prog['nr_node_ids'].value_() variable 250 nna = wq.node_nr_active[nr_node_ids]
|
| /linux/include/net/ |
| H A D | udp.h | 297 up->udp_prod_queue = kzalloc_objs(*up->udp_prod_queue, nr_node_ids); in udp_lib_init_sock() 300 for (int i = 0; i < nr_node_ids; i++) in udp_lib_init_sock()
|
| /linux/kernel/sched/ |
| H A D | topology.c | 1844 for (i = 0; i < nr_node_ids; i++) { in sched_numa_warn() 1846 for (j = 0; j < nr_node_ids; j++) { in sched_numa_warn() 2072 masks[i] = kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); in sched_init_numa() 2201 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set() 2218 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_clear() 2350 if (node >= nr_node_ids || hops >= sched_domains_numa_levels) in sched_numa_hop_mask()
|
| H A D | ext_idle.c | 667 scx_idle_node_masks = kzalloc_objs(*scx_idle_node_masks, nr_node_ids); in scx_idle_init_masks() 833 if (node < 0 || node >= nr_node_ids) { in validate_node()
|
| H A D | fair.c | 1691 return NR_NUMA_HINT_FAULT_TYPES * (s * nr_node_ids + nid) + priv; in task_faults_idx() 3088 nr_node_ids * sizeof(unsigned long); in task_numa_group() 3100 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_group() 3158 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) { in task_numa_group() 3201 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_free() 3216 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_free() 3252 NR_NUMA_HINT_FAULT_BUCKETS * nr_node_ids; in task_numa_fault()
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_main.c | 124 int i, num_nodes = nr_node_ids; in siw_init_cpulist()
|
| /linux/drivers/hv/ |
| H A D | hv.c | 186 hv_context.hv_numa_map = kzalloc_objs(struct cpumask, nr_node_ids); in hv_synic_alloc()
|
| H A D | channel_mgmt.c | 771 if (numa_node == nr_node_ids) { in init_vp_index()
|
| /linux/kernel/futex/ |
| H A D | core.c | 442 node = (hash >> futex_hashshift) % nr_node_ids; in __futex_hash() 445 nr_node_ids, node); in __futex_hash()
|
| /linux/arch/powerpc/sysdev/xive/ |
| H A D | common.c | 1142 ipi_domain = irq_domain_create_linear(fwnode, nr_node_ids, in xive_init_ipis() 1147 xive_ipis = kzalloc_objs(*xive_ipis, nr_node_ids, in xive_init_ipis()
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | hotplug-cpu.c | 244 if (rc && nr_node_ids > 1) { in pseries_add_processor()
|
| /linux/kernel/ |
| H A D | workqueue.c | 1578 node = nr_node_ids; in wq_node_nr_active() 4937 kfree(nna_ar[nr_node_ids]); in free_node_nr_active() 4938 nna_ar[nr_node_ids] = NULL; in free_node_nr_active() 4971 nna_ar[nr_node_ids] = nna; in alloc_node_nr_active() 5777 wq_size = struct_size(wq, node_nr_active, nr_node_ids + 1); in __alloc_workqueue()
|