| /linux/arch/parisc/kernel/ |
| H A D | firmware.c | 162 __pa(pdc_result), 0); in set_firmware_width_unlocked() 250 retval = mem_pdc_call(PDC_INSTR, 0UL, __pa(pdc_result)); in pdc_instr() 275 __pa(pdc_result), __pa(pdc_result2), len); in pdc_chassis_info() 300 retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data)); in pdc_pat_chassis_send_log() 366 __pa(pdc_result)); in pdc_pat_get_PDC_entrypoint() 382 retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result)); in pdc_chassis_warn() 393 ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result)); in pdc_coproc_cfg_unlocked() 440 retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, in pdc_iodc_read() 441 index, __pa(pdc_result2), iodc_data_size); in pdc_iodc_read() 468 retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result), in pdc_system_map_find_mods() [all …]
|
| /linux/arch/x86/include/asm/ |
| H A D | pgalloc.h | 65 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel() 66 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel() 72 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe() 73 set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel_safe() 99 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate() 100 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate() 105 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe() 106 set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate_safe() 113 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate() 114 set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); in p4d_populate() [all …]
|
| /linux/arch/arm/mach-omap2/ |
| H A D | omap-secure.c | 81 outer_clean_range(__pa(param), __pa(param + 5)); in omap_secure_dispatcher() 82 ret = omap_smc2(idx, flag, __pa(param)); in omap_secure_dispatcher() 131 param[1] = __pa(addr); /* Physical address for saving */ in omap3_save_secure_ram() 136 ret = save_secure_ram_context(__pa(param)); in omap3_save_secure_ram() 174 outer_clean_range(__pa(param), __pa(param + 5)); in rx51_secure_dispatcher() 175 ret = omap_smc3(idx, process, flag, __pa(param)); in rx51_secure_dispatcher()
|
| /linux/arch/riscv/mm/ |
| H A D | kasan_init.c | 36 set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pte() 58 set_pud(pud, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pmd() 89 set_p4d(p4d, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pud() 120 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_p4d() 254 phys_addr = __pa((uintptr_t)kasan_early_shadow_pmd); in kasan_early_populate_pud() 292 phys_addr = __pa((uintptr_t)kasan_early_shadow_pud); in kasan_early_populate_p4d() 313 phys_addr = __pa((uintptr_t)kasan_early_shadow_p4d); in kasan_early_populate_pgd() 336 (__pa((uintptr_t)kasan_early_shadow_pte)), in kasan_early_init() 343 (__pa(((uintptr_t)kasan_early_shadow_pmd))), in kasan_early_init() 351 (__pa(((uintptr_t)kasan_early_shadow_pud))), in kasan_early_init() [all …]
|
| H A D | init.c | 268 max_mapped_addr = __pa(PAGE_OFFSET) + KERN_VIRT_SIZE; in setup_bootmem() 452 return __pa((pte_t *)ptdesc_address(ptdesc)); in alloc_pte_late() 533 return __pa((pmd_t *)ptdesc_address(ptdesc)); in alloc_pmd_late() 598 return __pa((pud_t *)ptdesc_address(ptdesc)); in alloc_pud_late() 636 return __pa((p4d_t *)ptdesc_address(ptdesc)); in alloc_p4d_late() 695 #define get_pgd_next_virt(__pa) (pgtable_l5_enabled ? \ argument 696 pt_ops.get_p4d_virt(__pa) : (pgd_next_t *)(pgtable_l4_enabled ? \ 697 pt_ops.get_pud_virt(__pa) : (pud_t *)pt_ops.get_pmd_virt(__pa))) 698 #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ argument 700 create_p4d_mapping(__nextp, __va, __pa, __sz, __prot) : \ [all …]
|
| /linux/arch/loongarch/mm/ |
| H A D | kasan_init.c | 20 (__pa(pgd_val(pgd)) == (unsigned long)__pa(kasan_early_shadow_p4d))) 27 (__pa(p4d_val(p4d)) == (unsigned long)__pa(kasan_early_shadow_pud))) 34 (__pa(pud_val(pud)) == (unsigned long)__pa(kasan_early_shadow_pmd))) 38 (__pa(pmd_val(pmd)) == (unsigned long)__pa(kasan_early_shadow_pte))) 41 ((pte_val(pte) & _PFN_MASK) == (unsigned long)__pa(kasan_early_shadow_page))) 108 __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, node); in kasan_alloc_zeroed_page() 111 __func__, PAGE_SIZE, PAGE_SIZE, node, __pa(MAX_DMA_ADDRESS)); in kasan_alloc_zeroed_page() 113 return __pa(p); in kasan_alloc_zeroed_page()
|
| /linux/arch/arm/mm/ |
| H A D | pmsa-v8.c | 255 subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup() 256 subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup() 260 subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup() 261 subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup() 277 err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup() 280 err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
|
| /linux/arch/x86/kernel/ |
| H A D | machine_kexec_32.c | 89 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one() 95 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in machine_kexec_page_table_set_one() 111 (unsigned long)control_page, __pa(control_page)); in machine_kexec_prepare_page_tables() 117 __pa(control_page), __pa(control_page)); in machine_kexec_prepare_page_tables() 196 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec() 198 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec()
|
| H A D | espfix_64.c | 168 pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap() 169 paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); in init_espfix_ap() 180 pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap() 181 paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); in init_espfix_ap() 192 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap()
|
| /linux/arch/x86/power/ |
| H A D | hibernate_64.c | 71 __pud(__pa(pmd) | pgprot_val(pgtable_prot))); in set_up_temporary_text_mapping() 73 p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping() 74 pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping() 80 pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping() 123 temp_pgt = __pa(pgd); in set_up_temporary_mappings()
|
| H A D | hibernate_32.c | 41 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in resume_one_md_table_init() 66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init() 141 __pgd(__pa(empty_zero_page) | _PAGE_PRESENT)); in resume_init_first_level_page_table() 189 temp_pgt = __pa(resume_pg_dir); in swsusp_arch_resume()
|
| /linux/arch/x86/xen/ |
| H A D | p2m.c | 336 paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list() 338 paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list() 341 pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 343 pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 376 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list() 386 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 396 set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); in xen_rebuild_p2m_list() 445 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine() 483 paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT); in alloc_p2m_pmd() 495 __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); in alloc_p2m_pmd() [all …]
|
| /linux/arch/um/kernel/ |
| H A D | physmem.c | 90 os_seek_file(physmem_fd, __pa(__syscall_stub_start)); in setup_physmem() 93 memblock_add(__pa(start), len); in setup_physmem() 94 memblock_reserve(__pa(start), reserve); in setup_physmem() 96 min_low_pfn = PFN_UP(__pa(reserve_end)); in setup_physmem()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | page.h | 190 #define __pa(x) ((phys_addr_t)(unsigned long)(x) - VIRT_PHYS_OFFSET) macro 207 #define __pa(x) \ macro 215 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START) macro 222 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
|
| /linux/drivers/pci/controller/ |
| H A D | pci-versatile.c | 94 writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem)); in versatile_pci_probe() 131 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0); in versatile_pci_probe() 132 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1); in versatile_pci_probe() 133 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2); in versatile_pci_probe()
|
| /linux/arch/xtensa/include/asm/ |
| H A D | page.h | 163 #define __pa(x) ___pa((unsigned long)(x)) macro 165 #define __pa(x) \ macro 171 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) 173 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
|
| /linux/arch/hexagon/include/asm/ |
| H A D | pgalloc.h | 39 mm->context.ptbase = __pa(pgd); in pgd_alloc() 77 set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel() 85 set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel()
|
| /linux/arch/x86/mm/ |
| H A D | kasan_init_64.c | 28 __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid); in early_alloc() 32 (void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS)); in early_alloc() 49 if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL)) in kasan_populate_pmd() 67 entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL); in kasan_populate_pmd() 85 if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL)) in kasan_populate_pud() 367 __pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE)); in kasan_init() 426 early_pfn_to_nid(__pa(_stext))); in kasan_init() 447 pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); in kasan_init()
|
| /linux/arch/alpha/include/asm/ |
| H A D | page.h | 79 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET) macro 82 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) 83 #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT))
|
| /linux/arch/csky/include/asm/ |
| H A D | page.h | 68 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + va_pa_offset) macro 71 #define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0)) 75 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
|
| /linux/arch/m68k/include/asm/ |
| H A D | page_mm.h | 80 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) macro 94 #define __pa(x) ___pa((unsigned long)(x)) macro 125 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
|
| /linux/arch/s390/include/asm/ |
| H A D | pgalloc.h | 124 set_pgd(pgd, __pgd(_REGION1_ENTRY | __pa(p4d))); in pgd_populate() 129 set_p4d(p4d, __p4d(_REGION2_ENTRY | __pa(pud))); in p4d_populate() 134 set_pud(pud, __pud(_REGION3_ENTRY | __pa(pmd))); in pud_populate() 158 set_pmd(pmd, __pmd(_SEGMENT_ENTRY | __pa(pte))); in pmd_populate()
|
| /linux/arch/microblaze/include/asm/ |
| H A D | page.h | 99 # define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)) 118 # define __pa(x) __virt_to_phys((unsigned long)(x)) macro 123 return phys_to_pfn(__pa(vaddr)); in virt_to_pfn()
|
| /linux/arch/powerpc/kernel/ |
| H A D | suspend.c | 20 unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; in pfn_is_nosave() 21 unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
|
| /linux/arch/xtensa/kernel/ |
| H A D | hibernate.c | 9 unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); in pfn_is_nosave() 10 unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end)); in pfn_is_nosave()
|