Lines Matching defs:kmem_cache_node

495 struct kmem_cache_node {
507 static inline struct kmem_cache_node *get_node(struct kmem_cache *s, int node)
518 struct kmem_cache_node *n = get_node(s, numa_mem_id());
528 * a kmem_cache_node structure allocated (which is true for all online nodes)
1640 struct kmem_cache_node *n, struct slab *slab)
1649 static void remove_full(struct kmem_cache *s, struct kmem_cache_node *n, struct slab *slab)
1658 static inline unsigned long node_nr_slabs(struct kmem_cache_node *n)
1665 struct kmem_cache_node *n = get_node(s, node);
1672 struct kmem_cache_node *n = get_node(s, node);
2006 static inline void add_full(struct kmem_cache *s, struct kmem_cache_node *n,
2008 static inline void remove_full(struct kmem_cache *s, struct kmem_cache_node *n,
2018 static inline unsigned long node_nr_slabs(struct kmem_cache_node *n)
3368 __add_partial(struct kmem_cache_node *n, struct slab *slab, int tail)
3378 static inline void add_partial(struct kmem_cache_node *n,
3385 static inline void remove_partial(struct kmem_cache_node *n,
3401 struct kmem_cache_node *n, struct slab *slab, int orig_size)
3445 struct kmem_cache_node *n = get_node(s, nid);
3497 struct kmem_cache_node *n,
3594 struct kmem_cache_node *n;
3735 struct kmem_cache_node *n = get_node(s, slab_nid(slab));
3852 struct kmem_cache_node *n = NULL, *n2 = NULL;
4211 static inline unsigned long node_nr_objs(struct kmem_cache_node *n)
4276 static unsigned long count_partial(struct kmem_cache_node *n,
4294 static unsigned long count_partial_free_approx(struct kmem_cache_node *n)
4336 struct kmem_cache_node *n;
5710 * kmem_cache_node->list_lock, so kmalloc_nolock() can
5798 struct kmem_cache_node *n = get_node(s, slab_nid(slab));
5873 struct kmem_cache_node *n = NULL;
6206 struct kmem_cache_node *n;
7663 init_kmem_cache_node(struct kmem_cache_node *n, struct node_barn *barn)
7719 static struct kmem_cache *kmem_cache_node;
7726 * Note that this function only works on the kmem_cache_node
7727 * when allocating for the kmem_cache_node. This is used for bootstrapping
7733 struct kmem_cache_node *n;
7735 BUG_ON(kmem_cache_node->size < sizeof(struct kmem_cache_node));
7737 slab = new_slab(kmem_cache_node, GFP_NOWAIT, node);
7748 init_object(kmem_cache_node, n, SLUB_RED_ACTIVE);
7750 n = kasan_slab_alloc(kmem_cache_node, n, GFP_KERNEL, false);
7751 slab->freelist = get_freepointer(kmem_cache_node, n);
7753 kmem_cache_node->node[node] = n;
7755 inc_slabs_node(kmem_cache_node, node, slab->objects);
7767 struct kmem_cache_node *n;
7778 kmem_cache_free(kmem_cache_node, n);
7800 struct kmem_cache_node *n;
7815 n = kmem_cache_alloc_node(kmem_cache_node,
8037 static void free_partial(struct kmem_cache *s, struct kmem_cache_node *n)
8061 struct kmem_cache_node *n;
8075 struct kmem_cache_node *n;
8288 struct kmem_cache_node *n;
8372 struct kmem_cache_node *n;
8378 * allocate a kmem_cache_node structure in order to bring the node
8406 n = kmem_cache_alloc(kmem_cache_node, GFP_KERNEL);
8418 * Any cache created after this point will also have kmem_cache_node
8463 struct kmem_cache_node *n;
8500 kmem_cache_node = &boot_kmem_cache_node;
8510 create_boot_cache(kmem_cache_node, "kmem_cache_node",
8511 sizeof(struct kmem_cache_node),
8521 nr_node_ids * sizeof(struct kmem_cache_node *),
8525 kmem_cache_node = bootstrap(&boot_kmem_cache_node);
8719 struct kmem_cache_node *n, unsigned long *obj_map)
8759 struct kmem_cache_node *n;
9025 struct kmem_cache_node *n;
9042 struct kmem_cache_node *n;
9983 struct kmem_cache_node *n;
10093 struct kmem_cache_node *n;