/linux/include/linux/ |
H A D | pageblock-flags.h | 80 get_pfnblock_flags_mask(page, page_to_pfn(page), \ 83 set_pfnblock_flags_mask(page, 0, page_to_pfn(page), \ 87 page_to_pfn(page), \
|
/linux/drivers/xen/ |
H A D | biomerge.c | 12 unsigned long bfn1 = pfn_to_bfn(page_to_pfn(vec1->bv_page)); in xen_biovec_phys_mergeable() 13 unsigned long bfn2 = pfn_to_bfn(page_to_pfn(page)); in xen_biovec_phys_mergeable()
|
H A D | mem-reservation.c | 37 unsigned long pfn = page_to_pfn(page); in __xenmem_reservation_va_mapping_update() 65 unsigned long pfn = page_to_pfn(page); in __xenmem_reservation_va_mapping_reset()
|
/linux/include/trace/events/ |
H A D | kmem.h | 148 __entry->pfn = page_to_pfn(page); 169 __entry->pfn = page_to_pfn(page); 192 __entry->pfn = page ? page_to_pfn(page) : -1UL; 221 __entry->pfn = page ? page_to_pfn(page) : -1UL; 256 __entry->pfn = page ? page_to_pfn(page) : -1UL; 286 __entry->pfn = page_to_pfn(page);
|
H A D | page_ref.h | 30 __entry->pfn = page_to_pfn(page); 79 __entry->pfn = page_to_pfn(page);
|
/linux/Documentation/translations/zh_CN/mm/ |
H A D | memory-model.rst | 31 每个内存模型都定义了 :c:func:`pfn_to_page` 和 :c:func:`page_to_pfn` 95 Sparse vmemmapvmemmap使用虚拟映射的内存映射来优化pfn_to_page和page_to_pfn操 116 :c:func:`page_to_pfn`, ,和 :c:func:`get_user_pages` 服务。由于页面引
|
/linux/mm/ |
H A D | page_isolation.c | 177 check_unmovable_start = max(page_to_pfn(page), start_pfn); in set_migratetype_isolate() 178 check_unmovable_end = min(pageblock_end_pfn(page_to_pfn(page)), in set_migratetype_isolate() 229 buddy = find_buddy_page_pfn(page, page_to_pfn(page), in unset_migratetype_isolate() 369 pfn = page_to_pfn(page); in isolate_single_pageblock() 401 unsigned long head_pfn = page_to_pfn(head); in isolate_single_pageblock()
|
H A D | migrate_device.c | 141 mpfn = migrate_pfn(page_to_pfn(page)) | in migrate_vma_collect_pmd() 225 page_to_pfn(page)); in migrate_vma_collect_pmd() 228 page_to_pfn(page)); in migrate_vma_collect_pmd() 231 page_to_pfn(page)); in migrate_vma_collect_pmd() 620 page_to_pfn(page)); in migrate_vma_insert_page() 623 page_to_pfn(page)); in migrate_vma_insert_page()
|
H A D | memory-failure.c | 210 num_poisoned_pages_inc(page_to_pfn(page)); in page_handle_poison() 478 page_to_pfn(p), tsk->comm); in __add_to_kill() 1028 page_to_pfn(p), action_page_types[ps->type], count); in has_extra_refcount() 1052 pr_err("%#lx: Unknown page state\n", page_to_pfn(p)); in me_unknown() 1103 ret = truncate_error_folio(folio, page_to_pfn(p), mapping); in me_pagecache_clean() 1209 res = truncate_error_folio(folio, page_to_pfn(p), mapping); in me_huge_page() 1422 pr_info("%#lx cannot catch tail\n", page_to_pfn(page)); in __get_hwpoison_page() 1487 pr_err("%#lx: unhandlable page.\n", page_to_pfn(p)); in get_any_page() 1845 unmap_and_kill(&to_kill, page_to_pfn(page), mapping, in mf_dax_kill_procs() 1917 num_poisoned_pages_sub(page_to_pfn(p->page), 1); in __folio_free_raw_hwp() [all …]
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_msg.c | 729 pdesc->statPPNs[0] = page_to_pfn(virt_to_page(pstat)); in mksstat_init_kern_id() 732 pdesc->infoPPNs[0] = page_to_pfn(virt_to_page(pinfo)); in mksstat_init_kern_id() 735 pdesc->strsPPNs[0] = page_to_pfn(virt_to_page(pstrs)); in mksstat_init_kern_id() 740 hypervisor_ppn_add((PPN64)page_to_pfn(page)); in mksstat_init_kern_id() 1028 pdesc->statPPNs[i] = page_to_pfn(pages_stat[i]); in vmw_mksstat_add_ioctl() 1035 pdesc->infoPPNs[i] = page_to_pfn(pages_info[i]); in vmw_mksstat_add_ioctl() 1042 pdesc->strsPPNs[i] = page_to_pfn(pages_strs[i]); in vmw_mksstat_add_ioctl() 1047 hypervisor_ppn_add((PPN64)page_to_pfn(page)); in vmw_mksstat_add_ioctl() 1123 hypervisor_ppn_remove((PPN64)page_to_pfn(page)); in vmw_mksstat_remove_ioctl()
|
/linux/arch/sparc/include/asm/ |
H A D | cacheflush_32.h | 22 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 27 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
H A D | cacheflush_64.h | 62 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 69 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
|
/linux/arch/arm/mm/ |
H A D | flush.c | 140 flush_pfn_alias(page_to_pfn(page), uaddr); in __flush_ptrace_access() 149 flush_icache_alias(page_to_pfn(page), uaddr, len); in __flush_ptrace_access() 393 pfn = page_to_pfn(page); in __flush_anon_page()
|
H A D | copypage-v4wb.c | 54 flush_cache_page(vma, vaddr, page_to_pfn(from)); in v4wb_copy_user_highpage()
|
H A D | copypage-xsc3.c | 71 flush_cache_page(vma, vaddr, page_to_pfn(from)); in xsc3_mc_copy_user_highpage()
|
H A D | copypage-feroceon.c | 73 flush_cache_page(vma, vaddr, page_to_pfn(from)); in feroceon_copy_user_highpage()
|
/linux/drivers/gpu/drm/gma500/ |
H A D | gtt.c | 93 pte = psb_gtt_mask_pte(page_to_pfn(pages[i]), PSB_MMU_CACHED_MEMORY); in psb_gtt_insert_pages() 114 pte = psb_gtt_mask_pte(page_to_pfn(pdev->scratch_page), PSB_MMU_CACHED_MEMORY); in psb_gtt_remove_pages() 176 pfn_base = page_to_pfn(pdev->scratch_page); in psb_gtt_clear()
|
/linux/tools/testing/scatterlist/linux/ |
H A D | mm.h | 52 #define page_to_pfn(page) ((unsigned long)(page) / PAGE_SIZE) macro 54 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n))
|
/linux/arch/arc/include/asm/ |
H A D | pgtable-levels.h | 145 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot) 180 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
|
/linux/arch/m68k/include/asm/ |
H A D | cacheflush_mm.h | 276 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 284 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
|
/linux/lib/ |
H A D | test_hmm.c | 663 page_to_pfn(spage))) in dmirror_migrate_alloc_and_copy() 685 page_to_pfn(spage), page_to_pfn(dpage)); in dmirror_migrate_alloc_and_copy() 686 *dst = migrate_pfn(page_to_pfn(dpage)); in dmirror_migrate_alloc_and_copy() 871 page_to_pfn(spage), page_to_pfn(dpage)); in dmirror_devmem_fault_alloc_and_copy() 876 *dst = migrate_pfn(page_to_pfn(dpage)); in dmirror_devmem_fault_alloc_and_copy() 1063 } else if (is_zero_pfn(page_to_pfn(page))) in dmirror_mkentry() 1244 dst_pfns[i] = migrate_pfn(page_to_pfn(dpage)); in dmirror_device_evict_chunk()
|
/linux/kernel/power/ |
H A D | snapshot.c | 1042 memory_bm_set_bit(free_pages_map, page_to_pfn(page)); in swsusp_set_page_free() 1048 memory_bm_test_bit(free_pages_map, page_to_pfn(page)) : 0; in swsusp_page_is_free() 1054 memory_bm_clear_bit(free_pages_map, page_to_pfn(page)); in swsusp_unset_page_free() 1060 memory_bm_set_bit(forbidden_pages_map, page_to_pfn(page)); in swsusp_set_page_forbidden() 1066 memory_bm_test_bit(forbidden_pages_map, page_to_pfn(page)) : 0; in swsusp_page_is_forbidden() 1072 memory_bm_clear_bit(forbidden_pages_map, page_to_pfn(page)); in swsusp_unset_page_forbidden() 1279 pfn = page_to_pfn(page); in mark_free_pages() 1659 memory_bm_set_bit(©_bm, page_to_pfn(page)); in preallocate_image_pages() 2059 memory_bm_set_bit(bm, page_to_pfn(page)); in alloc_highmem_pages() 2100 memory_bm_set_bit(copy_bm, page_to_pfn(pag in swsusp_alloc() [all...] |
/linux/scripts/gdb/linux/ |
H A D | page_owner.py | 80 pfn = self.p_ops.page_to_pfn(page) 100 pfn = self.p_ops.page_to_pfn(page)
|
/linux/include/xen/ |
H A D | page.h | 21 ((page_to_pfn(page)) << (PAGE_SHIFT - XEN_PAGE_SHIFT))
|
/linux/tools/testing/nvdimm/ |
H A D | dax-dev.c | 32 return PFN_PHYS(page_to_pfn(page)); in dax_pgoff_to_phys()
|