Home
last modified time | relevance | path

Searched refs:__pa (Results 1 – 25 of 259) sorted by relevance

1234567891011

/linux/arch/x86/include/asm/
H A Dpgalloc.h65 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel()
66 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel()
72 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe()
73 set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel_safe()
99 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
100 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate()
105 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe()
106 set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate_safe()
113 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate()
114 set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); in p4d_populate()
[all …]
/linux/arch/arm/mach-omap2/
H A Domap-secure.c81 outer_clean_range(__pa(param), __pa(param + 5)); in omap_secure_dispatcher()
82 ret = omap_smc2(idx, flag, __pa(param)); in omap_secure_dispatcher()
131 param[1] = __pa(addr); /* Physical address for saving */ in omap3_save_secure_ram()
136 ret = save_secure_ram_context(__pa(param)); in omap3_save_secure_ram()
174 outer_clean_range(__pa(param), __pa(param + 5)); in rx51_secure_dispatcher()
175 ret = omap_smc3(idx, process, flag, __pa(param)); in rx51_secure_dispatcher()
/linux/arch/riscv/mm/
H A Dkasan_init.c36 set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pte()
58 set_pud(pud, pfn_pud(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pmd()
89 set_p4d(p4d, pfn_p4d(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_pud()
120 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_populate_p4d()
254 phys_addr = __pa((uintptr_t)kasan_early_shadow_pmd); in kasan_early_populate_pud()
292 phys_addr = __pa((uintptr_t)kasan_early_shadow_pud); in kasan_early_populate_p4d()
313 phys_addr = __pa((uintptr_t)kasan_early_shadow_p4d); in kasan_early_populate_pgd()
336 (__pa((uintptr_t)kasan_early_shadow_pte)), in kasan_early_init()
343 (__pa(((uintptr_t)kasan_early_shadow_pmd))), in kasan_early_init()
351 (__pa(((uintptr_t)kasan_early_shadow_pud))), in kasan_early_init()
[all …]
H A Dinit.c250 max_mapped_addr = __pa(PAGE_OFFSET) + KERN_VIRT_SIZE; in setup_bootmem()
425 return __pa((pte_t *)ptdesc_address(ptdesc)); in alloc_pte_late()
488 return __pa((pmd_t *)ptdesc_address(ptdesc)); in alloc_pmd_late()
553 return __pa((pud_t *)ptdesc_address(ptdesc)); in alloc_pud_late()
591 return __pa((p4d_t *)ptdesc_address(ptdesc)); in alloc_p4d_late()
650 #define get_pgd_next_virt(__pa) (pgtable_l5_enabled ? \ argument
651 pt_ops.get_p4d_virt(__pa) : (pgd_next_t *)(pgtable_l4_enabled ? \
652 pt_ops.get_pud_virt(__pa) : (pud_t *)pt_ops.get_pmd_virt(__pa)))
653 #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ argument
655 create_p4d_mapping(__nextp, __va, __pa, __sz, __prot) : \
[all …]
/linux/arch/arm/mm/
H A Dpmsa-v8.c255 subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
256 subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
260 subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
261 subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
277 err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
280 err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
/linux/arch/x86/kernel/
H A Dmachine_kexec_32.c89 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one()
95 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in machine_kexec_page_table_set_one()
111 (unsigned long)control_page, __pa(control_page)); in machine_kexec_prepare_page_tables()
117 __pa(control_page), __pa(control_page)); in machine_kexec_prepare_page_tables()
196 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec()
198 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec()
H A Despfix_64.c168 pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
169 paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); in init_espfix_ap()
180 pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
181 paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); in init_espfix_ap()
192 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap()
/linux/arch/x86/xen/
H A Dp2m.c336 paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list()
338 paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list()
341 pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
343 pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
376 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list()
386 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
396 set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); in xen_rebuild_p2m_list()
445 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine()
483 paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT); in alloc_p2m_pmd()
495 __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); in alloc_p2m_pmd()
[all …]
H A Dmmu_pv.c803 xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_pin()
808 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_pin()
918 xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd))); in __xen_pgd_unpin()
922 PFN_DOWN(__pa(user_pgd))); in __xen_pgd_unpin()
977 if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd)) in drop_mm_ref_this_cpu()
996 if (per_cpu(xen_current_cr3, cpu) != __pa(mm->pgd)) in xen_drop_mm_ref()
1012 if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd)) in xen_drop_mm_ref()
1103 unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK; in xen_cleanmfnmap_free_pgtbl()
1249 xen_start_info->pt_base = (unsigned long)__va(__pa(xen_start_info->pt_base)); in xen_pagetable_cleanhighmap()
1416 __xen_write_cr3(false, __pa(user_pgd)); in xen_write_cr3()
[all …]
/linux/arch/um/kernel/
H A Dphysmem.c90 os_seek_file(physmem_fd, __pa(__syscall_stub_start)); in setup_physmem()
93 memblock_add(__pa(start), len); in setup_physmem()
94 memblock_reserve(__pa(start), reserve); in setup_physmem()
96 min_low_pfn = PFN_UP(__pa(reserve_end)); in setup_physmem()
/linux/drivers/pci/controller/
H A Dpci-versatile.c94 writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem)); in versatile_pci_probe()
131 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0); in versatile_pci_probe()
132 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1); in versatile_pci_probe()
133 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2); in versatile_pci_probe()
/linux/arch/hexagon/include/asm/
H A Dpgalloc.h39 mm->context.ptbase = __pa(pgd); in pgd_alloc()
77 set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel()
85 set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel()
/linux/arch/x86/power/
H A Dhibernate_32.c41 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in resume_one_md_table_init()
66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init()
141 __pgd(__pa(empty_zero_page) | _PAGE_PRESENT)); in resume_init_first_level_page_table()
189 temp_pgt = __pa(resume_pg_dir); in swsusp_arch_resume()
/linux/arch/x86/mm/
H A Dkasan_init_64.c28 __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid); in early_alloc()
32 (void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS)); in early_alloc()
49 if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL)) in kasan_populate_pmd()
67 entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL); in kasan_populate_pmd()
85 if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL)) in kasan_populate_pud()
367 __pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE)); in kasan_init()
426 early_pfn_to_nid(__pa(_stext))); in kasan_init()
447 pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); in kasan_init()
H A Dkaslr.c203 __p4d(_KERNPG_TABLE | __pa(pud_page_tramp))); in init_trampoline_kaslr()
206 __pgd(_KERNPG_TABLE | __pa(p4d_page_tramp)); in init_trampoline_kaslr()
209 __pgd(_KERNPG_TABLE | __pa(pud_page_tramp)); in init_trampoline_kaslr()
/linux/arch/csky/include/asm/
H A Dpage.h68 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + va_pa_offset) macro
71 #define __pa_symbol(x) __pa(RELOC_HIDE((unsigned long)(x), 0))
75 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
/linux/arch/m68k/include/asm/
H A Dpage_mm.h80 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) macro
94 #define __pa(x) ___pa((unsigned long)(x)) macro
125 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
/linux/arch/powerpc/kernel/
H A Dsuspend.c20 unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; in pfn_is_nosave()
21 unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
/linux/arch/xtensa/kernel/
H A Dhibernate.c9 unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); in pfn_is_nosave()
10 unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end)); in pfn_is_nosave()
/linux/arch/sh/kernel/
H A Dswsusp.c21 unsigned long begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT; in pfn_is_nosave()
22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave()
/linux/arch/mips/power/
H A Dcpu.c39 unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); in pfn_is_nosave()
40 unsigned long nosave_end_pfn = PFN_UP(__pa(&__nosave_end)); in pfn_is_nosave()
/linux/arch/arm/include/asm/
H A Dcacheflush.h393 outer_clean_range(__pa(_p), __pa(_p + size)); in __sync_cache_range_w()
415 outer_flush_range(__pa(_p), __pa(_p + size)); in __sync_cache_range_r()
/linux/arch/x86/coco/sev/
H A Dcore.c174 sev_es_wr_ghcb_msr(__pa(ghcb)); in get_jump_table_addr()
281 ghcb_set_sw_scratch(ghcb, (u64)__pa(data)); in vmgexit_psc()
344 pfn = __pa(vaddr) >> PAGE_SHIFT; in __set_pages_state()
400 return early_set_pages_state(vaddr, __pa(vaddr), npages, &d); in set_pages_state()
461 ghcb_set_sw_exit_info_2(ghcb, __pa(vmsa)); in vmgexit_ap_control()
463 sev_es_wr_ghcb_msr(__pa(ghcb)); in vmgexit_ap_control()
490 call.rcx = __pa(va); in snp_set_vmsa()
495 call.rdx = __pa(caa); in snp_set_vmsa()
665 pa = __pa(vmsa); in shutdown_all_aps()
959 address = __pa(&data->ghcb_page); in sev_es_efi_map_ghcbs_cas()
[all …]
/linux/arch/powerpc/include/asm/nohash/32/
H A Dpgalloc.h23 *pmdp = __pmd(__pa(pte) | _PMD_PRESENT); in pmd_populate_kernel()
32 *pmdp = __pmd(__pa(pte_page) | _PMD_USER | _PMD_PRESENT); in pmd_populate()
/linux/arch/arm64/mm/
H A Dkasan_init.c38 __pa(MAX_DMA_ADDRESS), in kasan_alloc_zeroed_page()
43 __pa(MAX_DMA_ADDRESS)); in kasan_alloc_zeroed_page()
45 return __pa(p); in kasan_alloc_zeroed_page()
51 __pa(MAX_DMA_ADDRESS), in kasan_alloc_raw_page()
57 __pa(MAX_DMA_ADDRESS)); in kasan_alloc_raw_page()
59 return __pa(p); in kasan_alloc_raw_page()

1234567891011