| /linux/mm/ |
| H A D | hugetlb_internal.h | 83 #define for_each_node_mask_to_alloc(next_node, nr_nodes, node, mask) \ argument 84 for (nr_nodes = nodes_weight(*mask); \ 85 nr_nodes > 0 && \ 87 nr_nodes--) 89 #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ argument 90 for (nr_nodes = nodes_weight(*mask); \ 91 nr_nodes > 0 && \ 93 nr_nodes--)
|
| H A D | numa_emulation.c | 88 u64 addr, u64 max_addr, int nr_nodes) in split_nodes_interleave() argument 96 if (nr_nodes <= 0) in split_nodes_interleave() 98 if (nr_nodes > MAX_NUMNODES) { in split_nodes_interleave() 100 nr_nodes, MAX_NUMNODES); in split_nodes_interleave() 101 nr_nodes = MAX_NUMNODES; in split_nodes_interleave() 109 size = PFN_PHYS((unsigned long)(size >> PAGE_SHIFT) / nr_nodes); in split_nodes_interleave() 115 big = ((size & ~FAKE_NODE_MIN_HASH_MASK) * nr_nodes) / in split_nodes_interleave() 176 ret = emu_setup_memblk(ei, pi, nid++ % nr_nodes, in split_nodes_interleave() 204 static u64 uniform_size(u64 max_addr, u64 base, u64 hole, int nr_nodes) in uniform_size() argument 210 return PFN_PHYS((max_pfn - base_pfn - hole_pfns) / nr_nodes); in uniform_size() [all …]
|
| H A D | hugetlb.c | 1919 int nr_nodes, node; in alloc_pool_huge_folio() local 1921 for_each_node_mask_to_alloc(next_node, nr_nodes, node, nodes_allowed) { in alloc_pool_huge_folio() 1943 int nr_nodes, node; in remove_pool_hugetlb_folio() local 1947 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in remove_pool_hugetlb_folio() 3099 int nr_nodes, node = nid; in __alloc_bootmem_huge_page() local 3110 for_each_node_mask_to_alloc(&h->next_nid_to_alloc, nr_nodes, node, in __alloc_bootmem_huge_page() 3720 int nr_nodes, node; in adjust_pool_surplus() local 3726 for_each_node_mask_to_alloc(&h->next_nid_to_alloc, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus() 3731 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus() 3998 int nr_nodes, node; in demote_pool_huge_page() local [all …]
|
| H A D | page_alloc.c | 5646 unsigned nr_nodes) in build_zonelists_in_node_order() argument 5653 for (i = 0; i < nr_nodes; i++) { in build_zonelists_in_node_order() 5683 int node, nr_nodes = 0; in build_zonelists() local 5702 node_order[nr_nodes++] = node; in build_zonelists() 5706 build_zonelists_in_node_order(pgdat, node_order, nr_nodes); in build_zonelists() 5709 for (node = 0; node < nr_nodes; node++) in build_zonelists()
|
| H A D | mempolicy.c | 2207 unsigned int target, nr_nodes; in weighted_interleave_nid() local 2213 nr_nodes = read_once_policy_nodemask(pol, &nodemask); in weighted_interleave_nid() 2214 if (!nr_nodes) in weighted_interleave_nid()
|
| /linux/tools/perf/util/ |
| H A D | bpf-trace-summary.c | 111 int nr_nodes; member 146 if (max_summary == 0 || max_summary > data->nr_nodes) in print_common_stats() 147 max_summary = data->nr_nodes; in print_common_stats() 194 nodes = reallocarray(data->nodes, data->nr_nodes + 1, sizeof(*nodes)); in update_thread_stats() 199 nodes = &data->nodes[data->nr_nodes++]; in update_thread_stats() 211 qsort(data->nodes, data->nr_nodes, sizeof(*data->nodes), nodecmp); in print_thread_stat() 253 data->nr_nodes = 1; in update_total_stats() 330 nodes = reallocarray(data->nodes, data->nr_nodes + 1, sizeof(*nodes)); in update_cgroup_stats() 335 nodes = &data->nodes[data->nr_nodes++]; in update_cgroup_stats() 348 qsort(data->nodes, data->nr_nodes, sizeof(*data->nodes), nodecmp); in print_cgroup_stat()
|
| H A D | bpf_lock_contention.c | 179 skel->rodata->nr_nodes = last + 1; in init_numa_data()
|
| /linux/lib/ |
| H A D | rbtree_test.c | 172 static void check_postorder_foreach(int nr_nodes) in check_postorder_foreach() argument 179 WARN_ON_ONCE(count != nr_nodes); in check_postorder_foreach() 182 static void check_postorder(int nr_nodes) in check_postorder() argument 189 WARN_ON_ONCE(count != nr_nodes); in check_postorder() 192 static void check(int nr_nodes) in check() argument 212 WARN_ON_ONCE(count != nr_nodes); in check() 215 check_postorder(nr_nodes); in check() 216 check_postorder_foreach(nr_nodes); in check() 219 static void check_augmented(int nr_nodes) in check_augmented() argument 223 check(nr_nodes); in check_augmented()
|
| /linux/arch/loongarch/mm/ |
| H A D | cache.c | 36 int i, j, nr_nodes; in flush_cache_leaf() local 40 nr_nodes = cache_private(cdesc) ? 1 : loongson_sysconf.nr_nodes; in flush_cache_leaf() 53 } while (--nr_nodes > 0); in flush_cache_leaf()
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | topology_amd.c | 47 static void store_node(struct topo_scan *tscan, u16 nr_nodes, u16 node_id) in store_node() argument 58 tscan->amd_nodes_per_pkg = nr_nodes; in store_node()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | bootinfo.h | 35 int nr_nodes; member
|
| /linux/tools/perf/util/bpf_skel/ |
| H A D | lock_contention.bpf.c | 852 const volatile int nr_nodes; variable 889 } else if (nr_nodes > 0) { in collect_zone_lock() 892 for (int i = 0; i < nr_nodes; i++) { in collect_zone_lock()
|
| /linux/arch/loongarch/kernel/ |
| H A D | numa.c | 255 loongson_sysconf.nr_nodes = nr_node_ids; in init_numa_memory()
|
| /linux/arch/mips/include/asm/mach-loongson64/ |
| H A D | loongson.h | 24 u32 nr_nodes; member
|
| /linux/arch/mips/loongson64/ |
| H A D | env.c | 265 loongson_sysconf.nr_nodes = (loongson_sysconf.nr_cpus + in prom_lefi_init_env()
|
| /linux/Documentation/block/ |
| H A D | null_blk.rst | 39 home_node=[0--nr_nodes]: Default: NUMA_NO_NODE
|
| /linux/fs/btrfs/ |
| H A D | backref.c | 3053 cache->nr_nodes++; in btrfs_backref_alloc_node() 3067 cache->nr_nodes--; in btrfs_backref_free_node() 3169 ASSERT(!cache->nr_nodes); in btrfs_backref_release_cache()
|
| /linux/Documentation/driver-api/crypto/iaa/ |
| H A D | iaa-crypto.rst | 596 [ 60.752346] iaa_crypto: rebalance_wq_table: nr_nodes=2, nr_cpus 160, nr_iaa 8, cpus_per_iaa 20
|