Searched refs:page_to_nid (Results 1 – 21 of 21) sorted by relevance
| /linux/kernel/dma/ |
| H A D | contiguous.c | 435 if (cma_release(dma_contiguous_pernuma_area[page_to_nid(page)], in dma_free_contiguous() 438 if (cma_release(dma_contiguous_numa_area[page_to_nid(page)], in dma_free_contiguous()
|
| /linux/include/net/ |
| H A D | netmem.h | 336 return page_to_nid(netmem_to_page(netmem)) == pref_nid; in netmem_is_pref_nid()
|
| /linux/mm/ |
| H A D | hugetlb_vmemmap.c | 331 int nid = page_to_nid((struct page *)reuse); in vmemmap_remap_free() 392 int nid = page_to_nid((struct page *)start); in alloc_vmemmap_page_list()
|
| H A D | zswap.c | 619 return page_to_nid(virt_to_page(entry)); in entry_to_nid() 909 handle = zs_malloc(pool->zs_pool, dlen, gfp, page_to_nid(page)); in zswap_compress() 1422 entry = zswap_entry_cache_alloc(GFP_KERNEL, page_to_nid(page)); in zswap_store_page()
|
| H A D | mempolicy.c | 1135 ret = page_to_nid(p); in lookup_node() 2492 page_to_nid(page) == nid) { in alloc_pages_mpol()
|
| H A D | slub.c | 5134 if (page_to_nid(virt_to_page(object)) != node) { in alloc_from_pcs() 6971 nid != page_to_nid(virt_to_page(p))) in __do_krealloc() 8890 node_set(page_to_nid(virt_to_page(track)), l->nodes); in add_location() 8928 node_set(page_to_nid(virt_to_page(track)), l->nodes); in add_location()
|
| H A D | vmalloc.c | 4312 nid != page_to_nid(vmalloc_to_page(p))) in vrealloc_node_align_noprof() 5180 counters[page_to_nid(v->pages[nr])] += step; in show_numa_info()
|
| H A D | ksm.c | 2176 page_to_nid(tree_page) != nid) { in unstable_tree_search_insert()
|
| H A D | page_alloc.c | 1986 VM_BUG_ON_PAGE(page_to_nid(page) != zone_to_nid(zone), page); in __move_freepages_block()
|
| /linux/drivers/android/ |
| H A D | binder_alloc.c | 213 page_to_nid(page), in binder_lru_freelist_add() 412 page_to_nid(page), in binder_lru_freelist_del() 1012 page_to_nid(page), in binder_alloc_deferred_release()
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | octeon_network.h | 342 unlikely(page_to_nid(pg_info->page) != numa_node_id())) { in recv_buffer_recycle()
|
| /linux/drivers/block/zram/ |
| H A D | zram_drv.c | 2080 __GFP_HIGHMEM | __GFP_MOVABLE, page_to_nid(page)); in write_incompressible_page() 2143 __GFP_HIGHMEM | __GFP_MOVABLE, page_to_nid(page)); in zram_write_page() 2372 __GFP_HIGHMEM | __GFP_MOVABLE, page_to_nid(page)); in recompress_slot()
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_rx.c | 123 rb->node = page_is_pfmemalloc(p) ? -1 : page_to_nid(p); in funeth_alloc_page()
|
| /linux/include/linux/ |
| H A D | mm.h | 1776 static inline int page_to_nid(const struct page *page) in page_to_nid() function 1998 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone() 2003 return NODE_DATA(page_to_nid(page)); in page_pgdat()
|
| H A D | mmzone.h | 2283 page_to_nid(pfn_to_page(__pfn_to_nid_pfn)); \
|
| H A D | skbuff.h | 3606 return likely(page_to_nid(page) == numa_mem_id() && in dev_page_is_reusable()
|
| /linux/drivers/virt/nitro_enclaves/ |
| H A D | ne_misc_dev.c | 827 if (ne_enclave->numa_node != page_to_nid(mem_region_page)) { in ne_sanity_check_user_mem_region_page()
|
| /linux/drivers/net/ethernet/hisilicon/hns/ |
| H A D | hns_enet.c | 459 if (unlikely(page_to_nid(desc_cb->priv) != numa_node_id())) in hns_nic_reuse_page() 603 if (likely(page_to_nid(desc_cb->priv) == numa_node_id())) in hns_nic_poll_rx_skb()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_rx.c | 471 page_to_nid(page) != numa_mem_id(); in mlx4_en_complete_rx_desc()
|
| /linux/fs/proc/ |
| H A D | task_mmu.c | 3185 nid = page_to_nid(page); in can_gather_numa_stats() 3210 nid = page_to_nid(page); in can_gather_numa_stats_pmd()
|
| /linux/kernel/ |
| H A D | fork.c | 257 if (page_to_nid(page) != nid) in try_release_thread_stack_to_cache()
|