Searched refs:numa_mem_id (Results 1 – 18 of 18) sorted by relevance
| /linux/include/linux/ |
| H A D | topology.h | 139 #ifndef numa_mem_id 141 static inline int numa_mem_id(void) in numa_mem_id() function 163 #ifndef numa_mem_id 165 static inline int numa_mem_id(void) in numa_mem_id() function
|
| H A D | gfp.h | 246 __alloc_pages_bulk(_gfp, numa_mem_id(), NULL, _nr_pages, _page_array) 253 nid = numa_mem_id(); in alloc_pages_bulk_node_noprof() 310 nid = numa_mem_id(); in alloc_pages_node_noprof()
|
| H A D | skbuff.h | 3606 return likely(page_to_nid(page) == numa_mem_id() && in dev_page_is_reusable()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | numa.rst | 100 以使用numa_mem_id()或cpu_to_mem()函数来定位调用或指定CPU的“本地内存节点”。同样,这是同
|
| /linux/drivers/iommu/ |
| H A D | iommu-pages.c | 64 nid = numa_mem_id(); in iommu_alloc_pages_node_sz()
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_rx.c | 54 if (c->prod_cnt - c->cons_cnt <= c->mask && buf->node == numa_mem_id()) { in cache_offer() 268 buf->node == numa_mem_id()) || !q->spare_buf.page) { in get_buf() 313 funeth_alloc_page(q, &q->spare_buf, numa_mem_id(), in fun_gather_pkt()
|
| /linux/mm/ |
| H A D | memremap.c | 186 nid = numa_mem_id(); in pagemap_range()
|
| H A D | slub.c | 518 struct kmem_cache_node *n = get_node(s, numa_mem_id()); in get_barn() 3628 searchnode = numa_mem_id(); in get_partial() 5099 !node_isset(numa_mem_id(), mpol->nodes)) in alloc_from_pcs() 5112 if (unlikely(node_requested && node != numa_mem_id())) in alloc_from_pcs() 6336 rcu_sheaf->node = numa_mem_id(); in __kfree_rcu_sheaf() 6370 int node = numa_mem_id(); in free_to_pcs_bulk() 6678 slab_nid(slab) == numa_mem_id()) in slab_free()
|
| H A D | slab_common.c | 1630 slab_nid(slab) == numa_mem_id())) in kfree_rcu_sheaf()
|
| H A D | mempolicy.c | 2142 int node = numa_mem_id(); in mempolicy_slab_node()
|
| H A D | hugetlb.c | 1925 nid = numa_mem_id(); in only_alloc_fresh_hugetlb_folio()
|
| /linux/net/core/ |
| H A D | page_pool.c | 394 pref_nid = (pool->p.nid == NUMA_NO_NODE) ? numa_mem_id() : pool->p.nid; in page_pool_refill_alloc_cache() 397 pref_nid = numa_mem_id(); /* will be zero like page_to_nid() */ in page_pool_refill_alloc_cache()
|
| H A D | skbuff.c | 664 if (unlikely(node != NUMA_NO_NODE && node != numa_mem_id())) in __alloc_skb()
|
| /linux/Documentation/mm/ |
| H A D | numa.rst | 145 architectures transparently, kernel subsystems can use the numa_mem_id()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_rx.c | 905 page_pool_nid_changed(rq->page_pool, numa_mem_id()); in mlx5e_post_rx_wqes() 1113 page_pool_nid_changed(rq->page_pool, numa_mem_id()); in mlx5e_post_rx_mpwqes() 1115 page_pool_nid_changed(rq->hd_page_pool, numa_mem_id()); in mlx5e_post_rx_mpwqes()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_rx.c | 471 page_to_nid(page) != numa_mem_id(); in mlx4_en_complete_rx_desc()
|
| /linux/net/sunrpc/ |
| H A D | svc.c | 355 return numa_mem_id(); in svc_pool_map_get_node()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 4280 nid = numa_mem_id(); in idpf_rx_splitq_clean_all()
|