| /linux/drivers/staging/media/ipu7/ |
| H A D | ipu7-dma.c | 60 unsigned int count = PHYS_PFN(size); in __alloc_buffer() 103 unsigned int count = PHYS_PFN(size); in __free_buffer() 172 count = PHYS_PFN(size); in ipu7_dma_alloc() 175 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu7_dma_alloc() 249 struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle)); in ipu7_dma_free() 276 for (i = 0; i < PHYS_PFN(size); i++) { in ipu7_dma_free() 337 PHYS_PFN(sg_dma_address(sglist))); in ipu7_dma_unmap_sg() 417 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), in ipu7_dma_map_sg()
|
| /linux/drivers/media/pci/intel/ipu6/ |
| H A D | ipu6-dma.c | 60 int count = PHYS_PFN(size); in __alloc_buffer() 103 int count = PHYS_PFN(size); in __free_buffer() 172 count = PHYS_PFN(size); in ipu6_dma_alloc() 175 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu6_dma_alloc() 249 struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle)); in ipu6_dma_free() 276 for (i = 0; i < PHYS_PFN(size); i++) { in ipu6_dma_free() 337 PHYS_PFN(sg_dma_address(sglist))); in ipu6_dma_unmap_sg() 401 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in ipu6_dma_map_sg()
|
| H A D | ipu6-mmu.c | 425 PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0); in allocate_trash_buffer() 713 iova = find_iova(&dmap->iovad, PHYS_PFN(mmu->iova_trash_page)); in ipu6_mmu_destroy()
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | memtrace.c | 63 return remap_pfn_range(vma, vma->vm_start, PHYS_PFN(ent->start) + vma->vm_pgoff, in memtrace_mmap() 97 const unsigned long nr_pages = PHYS_PFN(size); in memtrace_alloc_node() 200 const unsigned long nr_pages = PHYS_PFN(size); in memtrace_free() 201 const unsigned long start_pfn = PHYS_PFN(start); in memtrace_free()
|
| /linux/arch/powerpc/mm/kasan/ |
| H A D | init_32.c | 28 __set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 1); in kasan_populate_pte() 75 pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL); in kasan_init_region() 109 pfn_pte(PHYS_PFN(pa), prot)); in kasan_remap_early_shadow_ro()
|
| H A D | 8xx.c | 30 pte_t pte = pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block + i * PAGE_SIZE)), PAGE_KERNEL)); in kasan_init_shadow_8M() 68 pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL); in kasan_init_region()
|
| H A D | book3s_32.c | 52 pte_t pte = pfn_pte(PHYS_PFN(phys + k_cur - k_nobat), PAGE_KERNEL); in kasan_init_region()
|
| /linux/include/asm-generic/ |
| H A D | memory_model.h | 70 #define __phys_to_pfn(paddr) PHYS_PFN(paddr) 87 #define phys_to_page(phys) pfn_to_page(PHYS_PFN(phys))
|
| /linux/scripts/gdb/linux/ |
| H A D | mm.py | 139 if self.PHYS_PFN(self.PFN_PHYS(pfn)) != pfn: 152 def PHYS_PFN(self, phys): member in x86_page_ops 180 return self.PHYS_PFN(self.__virt_to_phys(va)) 183 return self.PHYS_PFN(self.__virt_to_phys(x)) 343 if self.PHYS_PFN(self.PFN_PHYS(pfn)) != pfn: 395 def PHYS_PFN(self, phys): member in aarch64_page_ops 402 return self.PHYS_PFN(pa)
|
| /linux/drivers/nvdimm/ |
| H A D | pfn_devs.c | 658 unsigned long base_pfn = PHYS_PFN(base); in init_altmap_base() 666 unsigned long base_pfn = PHYS_PFN(base); in init_altmap_reserve() 688 .end_pfn = PHYS_PFN(end), in __nvdimm_setup_pfn() 701 nd_pfn->npfns = PHYS_PFN((range_len(range) - offset)); in __nvdimm_setup_pfn() 708 altmap->free = PHYS_PFN(offset - reserve); in __nvdimm_setup_pfn() 760 npfns = PHYS_PFN(size - SZ_8K); in nd_pfn_init() 815 npfns = PHYS_PFN(size - offset - end_trunc); in nd_pfn_init()
|
| /linux/include/linux/ |
| H A D | io-mapping.h | 77 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_atomic_wc() 98 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_local_wc()
|
| /linux/drivers/acpi/ |
| H A D | acpi_mrrm.c | 47 if (zone_intersects(zone, PHYS_PFN(e->base), PHYS_PFN(e->length))) in get_node_num()
|
| /linux/tools/include/linux/ |
| H A D | pfn.h | 10 #define PHYS_PFN(x) ((unsigned long)((x) >> PAGE_SHIFT)) macro
|
| /linux/security/integrity/ima/ |
| H A D | ima_kexec.c | 316 start_pfn = PHYS_PFN(phys); in ima_validate_range() 317 end_pfn = PHYS_PFN(end_phys); in ima_validate_range()
|
| /linux/arch/arm64/mm/ |
| H A D | ioremap.c | 55 unsigned long pfn = PHYS_PFN(offset); in arch_memremap_can_ram_remap()
|
| /linux/tools/testing/nvdimm/ |
| H A D | dax-dev.c | 20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys()
|
| H A D | pmem-dax.c | 42 *pfn = PHYS_PFN(pmem->phys_addr + offset); in __pmem_direct_access() 50 return PHYS_PFN(pmem->size - pmem->pfn_pad - offset); in __pmem_direct_access()
|
| /linux/mm/ |
| H A D | numa_emulation.c | 206 unsigned long max_pfn = PHYS_PFN(max_addr); in uniform_size() 207 unsigned long base_pfn = PHYS_PFN(base); in uniform_size() 208 unsigned long hole_pfns = PHYS_PFN(hole); in uniform_size()
|
| H A D | memory_hotplug.c | 55 return PHYS_PFN(memory_block_size_bytes()) * sizeof(struct page); in memory_block_memmap_size() 1397 if (memmap_pages == PHYS_PFN(memory_block_size_bytes())) in mhp_supports_memmap_on_memory() 1448 .base_pfn = PHYS_PFN(cur_start), in create_altmaps_and_memory_blocks() 1449 .end_pfn = PHYS_PFN(cur_start + memblock_size - 1), in create_altmaps_and_memory_blocks()
|
| H A D | vma.c | 52 .pglen = PHYS_PFN(len_), \ 200 pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start); in can_vma_merge_before() 712 pgoff = adjust->vm_pgoff + PHYS_PFN(vmg->end - adjust->vm_start); in vmg_adjust_set_range() 715 pgoff = adjust->vm_pgoff - PHYS_PFN(adjust->vm_start - vmg->end); in vmg_adjust_set_range() 968 pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start); in vma_merge_existing_range() 2909 VMG_STATE(vmg, mm, vmi, addr, addr + len, vma_flags, PHYS_PFN(addr)); in do_brk_flags()
|
| /linux/arch/xtensa/mm/ |
| H A D | kasan_init.c | 58 set_pte(pte + j, pfn_pte(PHYS_PFN(phys), PAGE_KERNEL)); in populate()
|
| /linux/kernel/liveupdate/ |
| H A D | kexec_handover.c | 403 struct page *page = pfn_to_online_page(PHYS_PFN(phys)); in kho_restore_page() 458 const unsigned long start_pfn = PHYS_PFN(phys); in kho_restore_pages() 991 unsigned long pfn = PHYS_PFN(virt_to_phys(chunk)); in kho_vmalloc_unpreserve_chunk() 996 pfn = PHYS_PFN(chunk->phys[i]); in kho_vmalloc_unpreserve_chunk() 1537 unsigned long base_pfn = PHYS_PFN(kho_scratch[i].addr); in kho_init()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_mmio_gem.c | 219 ret = vmf_insert_pfn(vma, addr, PHYS_PFN(phys_addr)); in xe_mmio_gem_vm_fault()
|
| /linux/arch/powerpc/kernel/ |
| H A D | paca.c | 85 uv_share_page(PHYS_PFN(__pa(shared_lppaca)), in alloc_shared_lppaca()
|
| /linux/drivers/vfio/pci/nvgrace-gpu/ |
| H A D | main.c | 105 start_pfn = PHYS_PFN(region->memphys); in pfn_memregion_offset() 156 pfn = PHYS_PFN(region->memphys); in nvgrace_gpu_vfio_pci_register_pfn_range() 313 pfn = PHYS_PFN(memregion->memphys) + addr_to_pgoff(vma, addr); in nvgrace_gpu_vfio_pci_huge_fault()
|