| /linux/arch/x86/xen/ |
| H A D | p2m.c | 127 static inline unsigned p2m_top_index(unsigned long pfn) in p2m_top_index() argument 129 BUG_ON(pfn >= MAX_P2M_PFN); in p2m_top_index() 130 return pfn / (P2M_MID_PER_PAGE * P2M_PER_PAGE); in p2m_top_index() 133 static inline unsigned p2m_mid_index(unsigned long pfn) in p2m_mid_index() argument 135 return (pfn / P2M_PER_PAGE) % P2M_MID_PER_PAGE; in p2m_mid_index() 170 static void p2m_init_identity(unsigned long *p2m, unsigned long pfn) in p2m_init_identity() argument 175 p2m[i] = IDENTITY_FRAME(pfn + i); in p2m_init_identity() 209 unsigned long pfn, mfn; in xen_build_mfn_list_list() local 232 for (pfn = 0; pfn < xen_max_p2m_pfn && pfn < MAX_P2M_PFN; in xen_build_mfn_list_list() 233 pfn += P2M_PER_PAGE) { in xen_build_mfn_list_list() [all …]
|
| /linux/arch/x86/include/asm/xen/ |
| H A D | page.h | 56 extern int xen_alloc_p2m_entry(unsigned long pfn); 58 extern unsigned long get_phys_to_machine(unsigned long pfn); 59 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn); 60 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 133 static inline unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument 137 if (pfn < xen_p2m_size) in __pfn_to_mfn() 138 mfn = xen_p2m_addr[pfn]; in __pfn_to_mfn() 139 else if (unlikely(pfn < xen_max_p2m_pfn)) in __pfn_to_mfn() 140 return get_phys_to_machine(pfn); in __pfn_to_mfn() 142 return IDENTITY_FRAME(pfn); in __pfn_to_mfn() [all …]
|
| /linux/tools/testing/scatterlist/ |
| H A D | main.c | 11 unsigned *pfn; member 28 #define pfn(...) (unsigned []){ __VA_ARGS__ } macro 42 printf(" %x", test->pfn[i]); in fail() 56 { -EINVAL, 1, pfn(0), NULL, PAGE_SIZE, 0, 1 }, in main() 57 { 0, 1, pfn(0), NULL, PAGE_SIZE, PAGE_SIZE + 1, 1 }, in main() 58 { 0, 1, pfn(0), NULL, PAGE_SIZE, sgmax, 1 }, in main() 59 { 0, 1, pfn(0), NULL, 1, sgmax, 1 }, in main() 60 { 0, 2, pfn(0, 1), NULL, 2 * PAGE_SIZE, sgmax, 1 }, in main() 61 { 0, 2, pfn(1, 0), NULL, 2 * PAGE_SIZE, sgmax, 2 }, in main() 62 { 0, 3, pfn(0, 1, 2), NULL, 3 * PAGE_SIZE, sgmax, 1 }, in main() [all …]
|
| /linux/arch/arm/xen/ |
| H A D | p2m.c | 23 unsigned long pfn; member 44 if (new->pfn == entry->pfn) in xen_add_phys_to_mach_entry() 47 if (new->pfn < entry->pfn) in xen_add_phys_to_mach_entry() 59 __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); in xen_add_phys_to_mach_entry() 64 unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument 74 if (entry->pfn <= pfn && in __pfn_to_mfn() 75 entry->pfn + entry->nr_pages > pfn) { in __pfn_to_mfn() 76 unsigned long mfn = entry->mfn + (pfn - entry->pfn); in __pfn_to_mfn() 80 if (pfn < entry->pfn) in __pfn_to_mfn() 150 bool __set_phys_to_machine_multi(unsigned long pfn, in __set_phys_to_machine_multi() argument [all …]
|
| /linux/mm/ |
| H A D | memory-failure.c | 83 void num_poisoned_pages_inc(unsigned long pfn) in num_poisoned_pages_inc() argument 86 memblk_nr_poison_inc(pfn); in num_poisoned_pages_inc() 89 void num_poisoned_pages_sub(unsigned long pfn, long i) in num_poisoned_pages_sub() argument 92 if (pfn != -1UL) in num_poisoned_pages_sub() 93 memblk_nr_poison_sub(pfn, i); in num_poisoned_pages_sub() 286 static int kill_proc(struct to_kill *tk, unsigned long pfn, int flags) in kill_proc() argument 293 pfn, t->comm, task_pid_nr(t)); in kill_proc() 463 unsigned long pfn, int flags) in kill_procs() argument 471 pfn, tk->tsk->comm, task_pid_nr(tk->tsk)); in kill_procs() 482 else if (kill_proc(tk, pfn, flags) < 0) in kill_procs() [all …]
|
| H A D | memory_hotplug.c | 28 #include <linux/pfn.h> 319 static int check_pfn_span(unsigned long pfn, unsigned long nr_pages) in check_pfn_span() argument 328 if (!IS_ALIGNED(pfn | nr_pages, PAGES_PER_SUBSECTION)) in check_pfn_span() 334 * Return page for the valid pfn only if the page is online. All pfn in check_pfn_span() 338 struct page *pfn_to_online_page(unsigned long pfn) in check_pfn_span() 340 unsigned long nr = pfn_to_section_nr(pfn); 355 if (IS_ENABLED(CONFIG_HAVE_ARCH_PFN_VALID) && !pfn_valid(pfn)) in pfn_to_online_page() 358 if (!pfn_section_valid(ms, pfn)) in pfn_to_online_page() 362 return pfn_to_page(pfn); in pfn_to_online_page() 346 pfn_to_online_page(unsigned long pfn) pfn_to_online_page() argument 389 __add_pages(int nid,unsigned long pfn,unsigned long nr_pages,struct mhp_params * params) __add_pages() argument 459 unsigned long pfn; find_biggest_section_pfn() local 482 unsigned long pfn; shrink_zone_span() local 553 unsigned long pfn, cur_nr_pages; remove_pfn_range_from_zone() local 593 __remove_pages(unsigned long pfn,unsigned long nr_pages,struct vmem_altmap * altmap) __remove_pages() argument 661 unsigned long pfn; online_pages_range() local 727 section_taint_zone_device(unsigned long pfn) section_taint_zone_device() argument 734 section_taint_zone_device(unsigned long pfn) section_taint_zone_device() argument 966 auto_movable_zone_for_pfn(int nid,struct memory_group * group,unsigned long pfn,unsigned long nr_pages) auto_movable_zone_for_pfn() argument 1090 mhp_init_memmap_on_memory(unsigned long pfn,unsigned long nr_pages,struct zone * zone) mhp_init_memmap_on_memory() argument 1121 mhp_deinit_memmap_on_memory(unsigned long pfn,unsigned long nr_pages) mhp_deinit_memmap_on_memory() argument 1144 online_pages(unsigned long pfn,unsigned long nr_pages,struct zone * zone,struct memory_group * group) online_pages() argument 1753 unsigned long pfn; scan_movable_pages() local 1795 unsigned long pfn; do_migrate_range() local 1904 unsigned long pfn, managed_pages, system_ram_pages = 0; offline_pages() local [all...] |
| H A D | page_idle.c | 32 * This function tries to get a user memory page by pfn as described above. 34 static struct folio *page_idle_get_folio(unsigned long pfn) in page_idle_get_folio() argument 36 struct page *page = pfn_to_online_page(pfn); in page_idle_get_folio() 124 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 130 pfn = pos * BITS_PER_BYTE; in page_idle_bitmap_read() 131 if (pfn >= max_pfn) in page_idle_bitmap_read() 134 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 138 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 139 bit = pfn in page_idle_bitmap_read() 169 unsigned long pfn, end_pfn; page_idle_bitmap_write() local [all...] |
| H A D | page_owner.c | 431 unsigned long pfn, block_end_pfn; in pagetypeinfo_showmixedcount_print() local 438 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print() 445 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print() 446 page = pfn_to_online_page(pfn); in pagetypeinfo_showmixedcount_print() 448 pfn = ALIGN(pfn + 1, MAX_ORDER_NR_PAGES); in pagetypeinfo_showmixedcount_print() 452 block_end_pfn = pageblock_end_pfn(pfn); in pagetypeinfo_showmixedcount_print() 457 for (; pfn < block_end_pfn; pfn++) { in pagetypeinfo_showmixedcount_print() 459 page = pfn_to_page(pfn); in pagetypeinfo_showmixedcount_print() 469 pfn += (1UL << freepage_order) - 1; in pagetypeinfo_showmixedcount_print() 491 pfn = block_end_pfn; in pagetypeinfo_showmixedcount_print() [all …]
|
| H A D | mm_init.c | 483 /* usable_startpfn is the lowest possible pfn ZONE_MOVABLE can be at */ in find_zone_movable_pfns_for_nodes() 592 void __meminit __init_single_page(struct page *page, unsigned long pfn, in __init_single_page() 596 set_page_links(page, zone, nid, pfn); in __init_single_page() 606 set_page_address(page, __va(pfn << PAGE_SHIFT)); 627 static int __meminit __early_pfn_to_nid(unsigned long pfn, in __early_pfn_to_nid() 633 if (state->last_start <= pfn && pfn < state->last_end) in __early_pfn_to_nid() 636 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in early_pfn_to_nid() 646 int __meminit early_pfn_to_nid(unsigned long pfn) in early_pfn_to_nid() 652 nid = __early_pfn_to_nid(pfn, in set_hashdist() 581 __init_single_page(struct page * page,unsigned long pfn,unsigned long zone,int nid) __init_single_page() argument 616 __early_pfn_to_nid(unsigned long pfn,struct mminit_pfnnid_cache * state) __early_pfn_to_nid() argument 635 early_pfn_to_nid(unsigned long pfn) early_pfn_to_nid() argument 669 __init_page_from_nid(unsigned long pfn,int nid) __init_page_from_nid() argument 696 early_page_initialised(unsigned long pfn,int nid) early_page_initialised() argument 709 defer_init(int nid,unsigned long pfn,unsigned long end_pfn) defer_init() argument 745 __init_deferred_page(unsigned long pfn,int nid) __init_deferred_page() argument 755 early_page_initialised(unsigned long pfn,int nid) early_page_initialised() argument 760 defer_init(int nid,unsigned long pfn,unsigned long end_pfn) defer_init() argument 765 __init_deferred_page(unsigned long pfn,int nid) __init_deferred_page() argument 770 init_deferred_page(unsigned long pfn,int nid) init_deferred_page() argument 784 unsigned long pfn; reserve_bootmem_region() local 802 overlap_memmap_init(unsigned long zone,unsigned long * pfn) overlap_memmap_init() argument 849 unsigned long pfn; init_unavailable_range() local 878 unsigned long pfn, end_pfn = start_pfn + size; memmap_init_range() local 1004 __init_zone_device_page(struct page * page,unsigned long pfn,unsigned long zone_idx,int nid,struct dev_pagemap * pgmap) __init_zone_device_page() argument 1088 unsigned long pfn, end_pfn = head_pfn + nr_pages; memmap_init_compound() local 1113 unsigned long pfn, end_pfn = start_pfn + nr_pages; memmap_init_zone_device() local 1986 deferred_free_pages(unsigned long pfn,unsigned long nr_pages) deferred_free_pages() argument 2033 deferred_init_pages(struct zone * zone,unsigned long pfn,unsigned long end_pfn) deferred_init_pages() argument 2481 memblock_free_pages(unsigned long pfn,unsigned int order) memblock_free_pages() argument [all...] |
| /linux/include/trace/events/ |
| H A D | ksm.h | 132 TP_PROTO(unsigned long pfn, void *rmap_item, void *mm, int err), 134 TP_ARGS(pfn, rmap_item, mm, err), 137 __field(unsigned long, pfn) 144 __entry->pfn = pfn; 151 __entry->pfn, __entry->rmap_item, __entry->mm, __entry->err) 167 TP_PROTO(void *ksm_page, unsigned long pfn, void *rmap_item, void *mm, int err), 169 TP_ARGS(ksm_page, pfn, rmap_item, mm, err), 173 __field(unsigned long, pfn) 181 __entry->pfn = pfn; 189 __entry->pfn, __entry->rmap_item, __entry->mm, __entry->err) [all …]
|
| H A D | cma.h | 13 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 16 TP_ARGS(name, pfn, page, count), 20 __field(unsigned long, pfn) 27 __entry->pfn = pfn; 34 __entry->pfn, 72 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 75 TP_ARGS(name, pfn, page, count, align, errorno), 79 __field(unsigned long, pfn) 88 __entry->pfn = pfn; 97 __entry->pfn, [all …]
|
| H A D | kmem.h | 146 __field( unsigned long, pfn ) 151 __entry->pfn = page_to_pfn(page); 156 pfn_to_page(__entry->pfn), 157 __entry->pfn, 168 __field( unsigned long, pfn ) 172 __entry->pfn = page_to_pfn(page); 176 pfn_to_page(__entry->pfn), 177 __entry->pfn) 188 __field( unsigned long, pfn ) 195 __entry->pfn = page ? page_to_pfn(page) : -1UL; [all …]
|
| /linux/scripts/gdb/linux/ |
| H A D | page_owner.py | 56 pfn = int(argv[1]) 57 self.read_page_owner_by_addr(self.p_ops.pfn_to_page(pfn)) 80 pfn = self.p_ops.page_to_pfn(page) 81 section = self.p_ops.pfn_to_section(pfn) 85 return self.get_entry(page_ext, pfn) 100 pfn = self.p_ops.page_to_pfn(page) 102 if pfn < self.min_pfn or pfn > self.max_pfn or (not self.p_ops.pfn_valid(pfn)): 106 page = self.p_ops.pfn_to_page(pfn) 130 gdb.write("PFN: %d, Flags: 0x%x\n" % (pfn, page['flags'])) 145 pfn = self.min_pfn [all …]
|
| H A D | mm.py | 105 def pfn_to_section_nr(self, pfn): argument 106 return pfn >> self.PFN_SECTION_SHIFT 111 def __pfn_to_section(self, pfn): argument 112 return self.__nr_to_section(self.pfn_to_section_nr(pfn)) 114 def pfn_to_section(self, pfn): argument 115 return self.__pfn_to_section(pfn) 117 def subsection_map_index(self, pfn): argument 118 return (pfn & ~(self.PAGE_SECTION_MASK)) // self.PAGES_PER_SUBSECTION 120 def pfn_section_valid(self, ms, pfn): argument 122 idx = self.subsection_map_index(pfn) [all …]
|
| /linux/include/asm-generic/ |
| H A D | memory_model.h | 18 #define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET)) argument 26 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 30 return pfn >= pfn_offset && (pfn - pfn_offset) < max_mapnr; in pfn_valid() 35 #define for_each_valid_pfn(pfn, start_pfn, end_pfn) \ argument 36 for ((pfn) = max_t(unsigned long, (start_pfn), ARCH_PFN_OFFSET); \ 37 (pfn) < min_t(unsigned long, (end_pfn), \ 39 (pfn)++) 46 #define __pfn_to_page(pfn) (vmemmap + (pfn)) argument 60 #define __pfn_to_page(pfn) \ argument 61 ({ unsigned long __pfn = (pfn); \ [all …]
|
| /linux/arch/x86/hyperv/ |
| H A D | ivm.c | 487 u64 pfn; member 498 u64 pfn; in hv_list_enc_add() local 502 pfn = pfn_list[i]; in hv_list_enc_add() 507 if ((ent->pfn <= pfn) && (ent->pfn + ent->count - 1 >= pfn)) in hv_list_enc_add() 520 if (ent->pfn + ent->count == pfn) { in hv_list_enc_add() 524 } else if (pfn + 1 == ent->pfn) { in hv_list_enc_add() 526 ent->pfn--; in hv_list_enc_add() 538 ent->pfn = pfn; in hv_list_enc_add() 556 u64 pfn; in hv_list_enc_remove() local 560 pfn = pfn_list[i]; in hv_list_enc_remove() [all …]
|
| /linux/arch/arm/mach-omap2/ |
| H A D | io.c | 68 .pfn = __phys_to_pfn(L3_24XX_PHYS), 74 .pfn = __phys_to_pfn(L4_24XX_PHYS), 84 .pfn = __phys_to_pfn(DSP_MEM_2420_PHYS), 90 .pfn = __phys_to_pfn(DSP_IPI_2420_PHYS), 96 .pfn = __phys_to_pfn(DSP_MMU_2420_PHYS), 108 .pfn = __phys_to_pfn(L4_WK_243X_PHYS), 114 .pfn = __phys_to_pfn(OMAP243X_GPMC_PHYS), 120 .pfn = __phys_to_pfn(OMAP243X_SDRC_PHYS), 126 .pfn = __phys_to_pfn(OMAP243X_SMS_PHYS), 138 .pfn = __phys_to_pfn(L3_34XX_PHYS), [all …]
|
| /linux/arch/x86/virt/svm/ |
| H A D | sev.c | 684 * others, and since the @pfn passed in cannot always be trusted, in get_raw_rmpentry() 687 static struct rmpentry_raw *get_raw_rmpentry(u64 pfn) in get_raw_rmpentry() 695 paddr = pfn << PAGE_SHIFT; in get_raw_rmpentry() 716 static int get_rmpentry(u64 pfn, struct rmpentry *e) in get_rmpentry() 726 : "a" (pfn << PAGE_SHIFT), "c" (e) in get_rmpentry() 732 e_raw = get_raw_rmpentry(pfn); in get_rmpentry() 752 static int __snp_lookup_rmpentry(u64 pfn, struct rmpentry *e, int *level) in __snp_lookup_rmpentry() 760 ret = get_rmpentry(pfn, e); in __snp_lookup_rmpentry() 769 ret = get_rmpentry(pfn & PFN_PMD_MASK, &e_large); in snp_lookup_rmpentry() 778 int snp_lookup_rmpentry(u64 pfn, boo 672 get_raw_rmpentry(u64 pfn) get_raw_rmpentry() argument 701 get_rmpentry(u64 pfn,struct rmpentry * e) get_rmpentry() argument 737 __snp_lookup_rmpentry(u64 pfn,struct rmpentry * e,int * level) __snp_lookup_rmpentry() argument 763 snp_lookup_rmpentry(u64 pfn,bool * assigned,int * level) snp_lookup_rmpentry() argument 783 dump_rmpentry(u64 pfn) dump_rmpentry() argument 863 psmash(u64 pfn) psmash() argument 910 adjust_direct_map(u64 pfn,int rmp_level) adjust_direct_map() argument 972 rmpupdate(u64 pfn,struct rmp_state * state) rmpupdate() argument 1005 rmp_make_private(u64 pfn,u64 gpa,enum pg_level level,u32 asid,bool immutable) rmp_make_private() argument 1021 rmp_make_shared(u64 pfn,enum pg_level level) rmp_make_shared() argument 1032 __snp_leak_pages(u64 pfn,unsigned int npages,bool dump_rmp) __snp_leak_pages() argument [all...] |
| /linux/include/linux/ |
| H A D | pageblock-flags.h | 76 #define pageblock_align(pfn) ALIGN((pfn), pageblock_nr_pages) argument 77 #define pageblock_aligned(pfn) IS_ALIGNED((pfn), pageblock_nr_pages) argument 78 #define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages) argument 79 #define pageblock_end_pfn(pfn) ALIGN((pfn) + 1, pageblock_nr_pages) argument 85 unsigned long pfn); 86 bool get_pfnblock_bit(const struct page *page, unsigned long pfn, 88 void set_pfnblock_bit(const struct page *page, unsigned long pfn, 90 void clear_pfnblock_bit(const struct page *page, unsigned long pfn,
|
| H A D | mmzone.h | 37 #define IS_MAX_ORDER_ALIGNED(pfn) IS_ALIGNED(pfn, MAX_ORDER_NR_PAGES) 154 * so folio_pfn() cannot be used and pfn is needed. 156 # define is_migrate_cma_folio(folio, pfn) \ 157 (get_pfnblock_migratetype(&folio->page, pfn) == MIGRATE_CMA) 161 # define is_migrate_cma_folio(folio, pfn) false 1117 /* pfn where compaction free scanner should start */ in zone_end_pfn() 1119 /* pfn where compaction migration scanner should start */ in zone_end_pfn() 1213 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) 1215 return zone->zone_start_pfn <= pfn 36 IS_MAX_ORDER_ALIGNED(pfn) global() argument 102 is_migrate_cma_folio(folio,pfn) global() argument 107 is_migrate_cma_folio(folio,pfn) global() argument 1122 zone_spans_pfn(const struct zone * zone,unsigned long pfn) zone_spans_pfn() argument 1852 pfn_to_nid(pfn) global() argument 1876 pfn_to_section_nr(unsigned long pfn) pfn_to_section_nr() argument 1885 SECTION_ALIGN_UP(pfn) global() argument 1886 SECTION_ALIGN_DOWN(pfn) global() argument 1901 SUBSECTION_ALIGN_UP(pfn) global() argument 1902 SUBSECTION_ALIGN_DOWN(pfn) global() argument 2112 __pfn_to_section(unsigned long pfn) __pfn_to_section() argument 2119 subsection_map_index(unsigned long pfn) subsection_map_index() argument 2125 pfn_section_valid(struct mem_section * ms,unsigned long pfn) pfn_section_valid() argument 2133 pfn_section_first_valid(struct mem_section * ms,unsigned long * pfn) pfn_section_first_valid() argument 2154 pfn_section_valid(struct mem_section * ms,unsigned long pfn) pfn_section_valid() argument 2159 pfn_section_first_valid(struct mem_section * ms,unsigned long * pfn) pfn_section_first_valid() argument 2180 pfn_valid(unsigned long pfn) pfn_valid() argument 2213 first_valid_pfn(unsigned long pfn,unsigned long end_pfn) first_valid_pfn() argument 2237 next_valid_pfn(unsigned long pfn,unsigned long end_pfn) next_valid_pfn() argument 2265 pfn_in_present_section(unsigned long pfn) pfn_in_present_section() argument 2293 pfn_to_nid(pfn) global() argument 2299 pfn_to_nid(pfn) global() argument [all...] |
| /linux/arch/riscv/include/asm/ |
| H A D | pgalloc.h | 21 unsigned long pfn = virt_to_pfn(pte); in pmd_populate_kernel() local 23 set_pmd(pmd, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); in pmd_populate_kernel() 29 unsigned long pfn = virt_to_pfn(page_address(pte)); in pmd_populate() local 31 set_pmd(pmd, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); in pmd_populate() 37 unsigned long pfn = virt_to_pfn(pmd); in pud_populate() local 39 set_pud(pud, __pud((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); in pud_populate() 45 unsigned long pfn = virt_to_pfn(pud); in p4d_populate() local 47 set_p4d(p4d, __p4d((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); in p4d_populate() 55 unsigned long pfn = virt_to_pfn(pud); in p4d_populate_safe() local 58 __p4d((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); in p4d_populate_safe() [all …]
|
| /linux/include/xen/arm/ |
| H A D | page.h | 15 #define phys_to_machine_mapping_valid(pfn) (1) argument 43 unsigned long __pfn_to_mfn(unsigned long pfn); 47 static inline unsigned long pfn_to_gfn(unsigned long pfn) in pfn_to_gfn() argument 49 return pfn; in pfn_to_gfn() 58 static inline unsigned long pfn_to_bfn(unsigned long pfn) in pfn_to_bfn() argument 63 mfn = __pfn_to_mfn(pfn); in pfn_to_bfn() 68 return pfn; in pfn_to_bfn() 103 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 104 bool __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn, 107 static inline bool set_phys_to_machine(unsigned long pfn, unsigned long mfn) in set_phys_to_machine() argument [all …]
|
| /linux/arch/arm/mm/ |
| H A D | flush.c | 38 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument 43 set_top_pte(to, pfn_pte(pfn, PAGE_KERNEL)); in flush_pfn_alias() 52 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument 58 set_top_pte(va, pfn_pte(pfn, PAGE_KERNEL)); in flush_icache_alias() 98 void flush_cache_pages(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn, unsi… in flush_cache_pages() argument 101 vivt_flush_cache_pages(vma, user_addr, pfn, nr); in flush_cache_pages() 106 flush_pfn_alias(pfn, user_addr); in flush_cache_pages() 115 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument 116 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument 255 unsigned long start, offset, pfn; in __flush_dcache_aliases() local [all …]
|
| /linux/kernel/power/ |
| H A D | snapshot.c | 748 static int memory_bm_find_bit(struct memory_bitmap *bm, unsigned long pfn, in memory_bm_find_bit() argument 757 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 764 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 786 ((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) in memory_bm_find_bit() 790 block_nr = (pfn - zone->start_pfn) >> BM_BLOCK_SHIFT; in memory_bm_find_bit() 805 bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; in memory_bm_find_bit() 806 bm->cur.cur_pfn = pfn; in memory_bm_find_bit() 810 *bit_nr = (pfn - zone->start_pfn) & BM_BLOCK_MASK; in memory_bm_find_bit() 815 static void memory_bm_set_bit(struct memory_bitmap *bm, unsigned long pfn) in memory_bm_set_bit() argument 821 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_set_bit() [all …]
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | mem_protect.c | 107 unsigned long nr_pages, pfn; in prepare_s2_pool() 110 pfn = hyp_virt_to_pfn(pgt_pool_base); in prepare_s2_pool() 112 ret = hyp_pool_init(&host_s2_pool, pfn, nr_pages, 0); in prepare_s2_pool() 377 * using a pfn/nr_pages pair from an untrusted source. 379 static bool pfn_range_is_valid(u64 pfn, u64 nr_pages) in pfn_range_is_valid() 383 return pfn < limit && ((limit - pfn) >= nr_pages); in pfn_range_is_valid() 936 int __pkvm_host_share_hyp(u64 pfn) in __pkvm_host_unshare_ffa() 938 u64 phys = hyp_pfn_to_phys(pfn); in __pkvm_host_unshare_ffa() 1028 int __pkvm_host_unshare_hyp(u64 pfn) in __pkvm_host_share_guest() 106 unsigned long nr_pages, pfn; prepare_s2_pool() local 378 pfn_range_is_valid(u64 pfn,u64 nr_pages) pfn_range_is_valid() argument 730 __pkvm_host_share_hyp(u64 pfn) __pkvm_host_share_hyp() argument 756 __pkvm_host_unshare_hyp(u64 pfn) __pkvm_host_unshare_hyp() argument 787 __pkvm_host_donate_hyp(u64 pfn,u64 nr_pages) __pkvm_host_donate_hyp() argument 818 __pkvm_hyp_donate_host(u64 pfn,u64 nr_pages) __pkvm_hyp_donate_host() argument 905 __pkvm_host_share_ffa(u64 pfn,u64 nr_pages) __pkvm_host_share_ffa() argument 923 __pkvm_host_unshare_ffa(u64 pfn,u64 nr_pages) __pkvm_host_unshare_ffa() argument 963 __pkvm_host_share_guest(u64 pfn,u64 gfn,u64 nr_pages,struct pkvm_hyp_vcpu * vcpu,enum kvm_pgtable_prot prot) __pkvm_host_share_guest() argument 1288 u64 phys, size, pfn, gfn; pkvm_ownership_selftest() local [all...] |