| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 23 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in mmu_wp_memory_region() 24 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in mmu_wp_memory_region() 99 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked() 100 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked() 128 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() 129 phys_addr_t size = slot->npages << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() 174 kvm_riscv_gstage_gpa_size(kvm->arch.pgd_levels) >> PAGE_SHIFT) in kvm_arch_prepare_memory_region() 178 size = new->npages << PAGE_SHIFT; in kvm_arch_prepare_memory_region() 240 kvm_riscv_gstage_unmap_range(&gstage, range->start << PAGE_SHIFT, in kvm_unmap_gfn_range() 241 (range->end - range->start) << PAGE_SHIFT, in kvm_unmap_gfn_range() [all …]
|
| /linux/arch/arc/include/asm/ |
| H A D | highmem.h | 24 #define FIXADDR_TOP (FIXMAP_BASE + (FIX_KMAP_END << PAGE_SHIFT)) 30 #define __fix_to_virt(x) (FIXADDR_TOP - ((x) << PAGE_SHIFT)) 31 #define __virt_to_fix(x) (((FIXADDR_TOP - ((x) & PAGE_MASK))) >> PAGE_SHIFT) 35 #define LAST_PKMAP (PKMAP_SIZE >> PAGE_SHIFT) 37 #define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT)) 38 #define PKMAP_NR(virt) (((virt) - PKMAP_BASE) >> PAGE_SHIFT)
|
| H A D | pgtable-levels.h | 86 #define PTRS_PER_PTE BIT(PMD_SHIFT - PAGE_SHIFT) 143 #define pmd_pfn(pmd) ((pmd_val(pmd) & PMD_MASK) >> PAGE_SHIFT) 144 #define pfn_pmd(pfn,prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 160 #define pmd_pfn(pmd) ((pmd_val(pmd) & PAGE_MASK) >> PAGE_SHIFT) 171 #define PFN_PTE_SHIFT PAGE_SHIFT 177 #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
|
| /linux/arch/arm/mm/ |
| H A D | tlb-v7.S | 39 mov r0, r0, lsr #PAGE_SHIFT @ align address 40 mov r1, r1, lsr #PAGE_SHIFT 46 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA 47 mov r1, r1, lsl #PAGE_SHIFT 73 mov r0, r0, lsr #PAGE_SHIFT @ align address 74 mov r1, r1, lsr #PAGE_SHIFT 75 mov r0, r0, lsl #PAGE_SHIFT 76 mov r1, r1, lsl #PAGE_SHIFT
|
| H A D | tlb-v6.S | 41 mov r0, r0, lsr #PAGE_SHIFT @ align address 42 mov r1, r1, lsr #PAGE_SHIFT 44 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA 45 mov r1, r1, lsl #PAGE_SHIFT 73 mov r0, r0, lsr #PAGE_SHIFT @ align address 74 mov r1, r1, lsr #PAGE_SHIFT 75 mov r0, r0, lsl #PAGE_SHIFT 76 mov r1, r1, lsl #PAGE_SHIFT
|
| H A D | mmap.c | 18 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1))) 52 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area() 76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 105 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area_topdown() 127 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown() 166 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT)); in valid_mmap_phys_addr_range()
|
| /linux/arch/x86/include/asm/ |
| H A D | pgalloc.h | 65 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel() 72 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe() 82 set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate() 99 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate() 105 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe() 113 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate() 119 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate_safe() 136 paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); in pgd_populate() 144 paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); in pgd_populate_safe()
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | iommu_api.c | 80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 100 chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) / in mm_iommu_do_alloc() 106 ret = pin_user_pages(ua + (entry << PAGE_SHIFT), n, in mm_iommu_do_alloc() 135 if ((mem2->ua < (ua + (entries << PAGE_SHIFT))) && in mm_iommu_do_alloc() 137 (mem2->entries << PAGE_SHIFT)))) { in mm_iommu_do_alloc() 150 pageshift = PAGE_SHIFT; in mm_iommu_do_alloc() 154 if ((mem->pageshift > PAGE_SHIFT) && PageHuge(page)) in mm_iommu_do_alloc() 161 mem->hpas[i] = page_to_pfn(page) << PAGE_SHIFT; in mm_iommu_do_alloc() 214 page = pfn_to_page(mem->hpas[i] >> PAGE_SHIFT); in mm_iommu_unpin() [all …]
|
| H A D | subpage_prot.c | 120 spp += (addr >> PAGE_SHIFT) & (SBP_L1_COUNT - 1); in subpage_prot_clear() 122 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in subpage_prot_clear() 124 if (addr + (nw << PAGE_SHIFT) > next) in subpage_prot_clear() 125 nw = (next - addr) >> PAGE_SHIFT; in subpage_prot_clear() 213 if (!access_ok(map, (len >> PAGE_SHIFT) * sizeof(u32))) in SYSCALL_DEFINE3() 255 spp += (addr >> PAGE_SHIFT) & (SBP_L1_COUNT - 1); in SYSCALL_DEFINE3() 261 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in SYSCALL_DEFINE3() 263 if (addr + (nw << PAGE_SHIFT) > next) in SYSCALL_DEFINE3() 264 nw = (next - addr) >> PAGE_SHIFT; in SYSCALL_DEFINE3()
|
| /linux/arch/x86/kernel/cpu/mtrr/ |
| H A D | centaur.c | 52 *base = centaur_mcr[reg].high >> PAGE_SHIFT; in centaur_get_mcr() 53 *size = -(centaur_mcr[reg].low & 0xfffff000) >> PAGE_SHIFT; in centaur_get_mcr() 74 high = base << PAGE_SHIFT; in centaur_set_mcr() 77 low = -size << PAGE_SHIFT | 0x1f; in centaur_set_mcr() 80 low = -size << PAGE_SHIFT | 0x02; /* NC */ in centaur_set_mcr() 82 low = -size << PAGE_SHIFT | 0x09; /* WWO, WC */ in centaur_set_mcr()
|
| H A D | amd.c | 20 *base = (low & 0xFFFE0000) >> PAGE_SHIFT; in amd_get_mtrr() 46 *size = (low + 4) << (15 - PAGE_SHIFT); in amd_get_mtrr() 83 regs[reg] = (-size >> (15 - PAGE_SHIFT) & 0x0001FFFC) in amd_set_mtrr() 84 | (base << PAGE_SHIFT) | (type + 1); in amd_set_mtrr() 106 if (type > MTRR_TYPE_WRCOMB || size < (1 << (17 - PAGE_SHIFT)) in amd_validate_add_page()
|
| /linux/arch/loongarch/kvm/ |
| H A D | mmu.c | 68 offset = (addr >> PAGE_SHIFT) - ctx->gfn; in kvm_mkclean_pte() 204 size = 0x1UL << (ctx->pgtable_shift + PAGE_SHIFT - 3); in kvm_ptw_leaf() 248 size = 0x1UL << (ctx->pgtable_shift + PAGE_SHIFT - 3); in kvm_ptw_dir() 305 ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, in kvm_flush_range() 306 end_gfn << PAGE_SHIFT, &ctx); in kvm_flush_range() 309 ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, in kvm_flush_range() 310 end_gfn << PAGE_SHIFT, &ctx); in kvm_flush_range() 348 return kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, end_gfn << PAGE_SHIFT, &ctx); in kvm_mkclean_gpa_pt() 376 kvm_ptw_top(kvm->arch.pgd, start << PAGE_SHIFT, end << PAGE_SHIFT, &ctx); in kvm_arch_mmu_enable_log_dirty_pt_masked() 392 if ((new->base_gfn + new->npages) > (kvm->arch.gpa_size >> PAGE_SHIFT)) in kvm_arch_prepare_memory_region() [all …]
|
| /linux/include/drm/ |
| H A D | drm_vma_manager.h | 37 #define DRM_FILE_PAGE_OFFSET_START ((0xFFFFFFFFUL >> PAGE_SHIFT) + 1) 38 #define DRM_FILE_PAGE_OFFSET_SIZE ((0xFFFFFFFFUL >> PAGE_SHIFT) * 256) 40 #define DRM_FILE_PAGE_OFFSET_START ((0xFFFFFFFUL >> PAGE_SHIFT) + 1) 41 #define DRM_FILE_PAGE_OFFSET_SIZE ((0xFFFFFFFUL >> PAGE_SHIFT) * 16) 205 return ((__u64)node->vm_node.start) << PAGE_SHIFT; in drm_vma_node_offset_addr() 226 drm_vma_node_size(node) << PAGE_SHIFT, 1); in drm_vma_node_unmap()
|
| /linux/tools/include/linux/ |
| H A D | pfn.h | 7 #define PFN_UP(x) (((x) + PAGE_SIZE - 1) >> PAGE_SHIFT) 8 #define PFN_DOWN(x) ((x) >> PAGE_SHIFT) 9 #define PFN_PHYS(x) ((phys_addr_t)(x) << PAGE_SHIFT) 10 #define PHYS_PFN(x) ((unsigned long)((x) >> PAGE_SHIFT))
|
| /linux/include/asm-generic/ |
| H A D | getorder.h | 33 return BITS_PER_LONG - PAGE_SHIFT; in get_order() 35 if (size < (1UL << PAGE_SHIFT)) in get_order() 38 return ilog2((size) - 1) - PAGE_SHIFT + 1; in get_order() 42 size >>= PAGE_SHIFT; in get_order()
|
| /linux/arch/csky/kernel/ |
| H A D | vdso.c | 21 vdso_pages = (vdso_end - vdso_start) >> PAGE_SHIFT; in vdso_init() 32 pg = virt_to_page(vdso_start + (i << PAGE_SHIFT)); in vdso_init() 51 vdso_len = vdso_pages << PAGE_SHIFT; in arch_setup_additional_pages() 69 _install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT, in arch_setup_additional_pages() 79 vdso_base += (vdso_pages << PAGE_SHIFT); in arch_setup_additional_pages()
|
| /linux/arch/powerpc/mm/ |
| H A D | mem.c | 139 update_end_of_memory_vars(start_pfn << PAGE_SHIFT, in add_pages() 140 nr_pages << PAGE_SHIFT); in add_pages() 148 unsigned long start_pfn = start >> PAGE_SHIFT; in arch_add_memory() 149 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory() 163 unsigned long start_pfn = start >> PAGE_SHIFT; in arch_remove_memory() 164 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory() 174 max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; in mem_topology_setup() 175 min_low_pfn = MEMORY_START >> PAGE_SHIFT; in mem_topology_setup() 177 max_low_pfn = lowmem_end_addr >> PAGE_SHIFT; in mem_topology_setup() 223 max_zone_pfns[ZONE_DMA] = min((zone_dma_limit >> PAGE_SHIFT) + 1, max_low_pfn); in arch_zone_limits_init()
|
| /linux/arch/sparc/include/asm/ |
| H A D | io-unit.h | 45 unsigned long bmap[(IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 3)) / sizeof(unsigned long)]; 53 #define IOUNIT_BMAP1_END (IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 1)) 55 #define IOUNIT_BMAP2_END IOUNIT_BMAP2_START + (IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 2)) 57 #define IOUNIT_BMAPM_END ((IOUNIT_DMA_SIZE - IOUNIT_DVMA_SIZE) >> PAGE_SHIFT)
|
| H A D | highmem.h | 42 #define PKMAP_SIZE (LAST_PKMAP << PAGE_SHIFT) 43 #define PKMAP_BASE PMD_ALIGN(SRMMU_NOCACHE_VADDR + (SRMMU_MAX_NOCACHE_PAGES << PAGE_SHIFT)) 46 #define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT) 47 #define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
|
| /linux/mm/ |
| H A D | hmm.c | 53 unsigned long i = (addr - range->start) >> PAGE_SHIFT; in hmm_pfns_fill() 166 i = (addr - range->start) >> PAGE_SHIFT; in hmm_vma_walk_hole() 167 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_hole() 194 hmm_pfn_flags_order(PMD_SHIFT - PAGE_SHIFT); in pmd_to_hmm_pfn_flags() 207 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_handle_pmd() 214 pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT); in hmm_vma_handle_pmd() 341 unsigned long npages = (end - start) >> PAGE_SHIFT; in hmm_vma_handle_absent_pmd() 350 hmm_pfn_flags_order(PMD_SHIFT - PAGE_SHIFT); in hmm_vma_handle_absent_pmd() 388 unsigned long npages = (end - start) >> PAGE_SHIFT; in hmm_vma_handle_absent_pmd() 404 &range->hmm_pfns[(start - range->start) >> PAGE_SHIFT]; in hmm_vma_walk_pmd() [all …]
|
| /linux/arch/hexagon/include/asm/ |
| H A D | mem-layout.h | 38 #define PHYS_PFN_OFFSET (PHYS_OFFSET >> PAGE_SHIFT) 76 #define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT) 86 #define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT) 87 #define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
|
| /linux/drivers/gpu/drm/gma500/ |
| H A D | gtt.c | 58 BUG_ON(pfn & ~(0xFFFFFFFF >> PAGE_SHIFT)); in psb_gtt_mask_pte() 67 return (pfn << PAGE_SHIFT) | mask; in psb_gtt_mask_pte() 74 return pdev->gtt_map + (offset >> PAGE_SHIFT); in psb_gtt_entry() 89 npages = resource_size(res) >> PAGE_SHIFT; in psb_gtt_insert_pages() 116 npages = resource_size(res) >> PAGE_SHIFT; in psb_gtt_remove_pages() 206 gtt_pages = pci_resource_len(pdev, PSB_GTT_RESOURCE) >> PAGE_SHIFT; in psb_gtt_init_ranges() 216 gatt_pages = pci_resource_len(pdev, PSB_GATT_RESOURCE) >> PAGE_SHIFT; in psb_gtt_init_ranges() 228 gatt_pages = (128 * 1024 * 1024) >> PAGE_SHIFT; in psb_gtt_init_ranges() 269 dev_priv->gtt_map = ioremap(pg->gtt_phys_start, pg->gtt_pages << PAGE_SHIFT); in psb_gtt_init()
|
| /linux/arch/x86/mm/pat/ |
| H A D | memtype.c | 403 unsigned long start_pfn = start >> PAGE_SHIFT; in pat_pagerange_is_ram() 404 unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; in pat_pagerange_is_ram() 414 if (start_pfn < ISA_END_ADDRESS >> PAGE_SHIFT) in pat_pagerange_is_ram() 415 start_pfn = ISA_END_ADDRESS >> PAGE_SHIFT; in pat_pagerange_is_ram() 455 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type() 473 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type() 485 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in free_ram_pages_type() 666 page = pfn_to_page(paddr >> PAGE_SHIFT); in lookup_memtype() 772 if (!phys_mem_access_encrypted(pfn << PAGE_SHIFT, size)) in phys_mem_access_prot() 818 if (!page_is_ram(base >> PAGE_SHIFT)) in memtype_kernel_map_sync() [all …]
|
| /linux/drivers/media/platform/samsung/s5p-mfc/ |
| H A D | s5p_mfc_opr.c | 37 unsigned int bits = dev->mem_size >> PAGE_SHIFT; in s5p_mfc_alloc_priv_buf() 38 unsigned int count = b->size >> PAGE_SHIFT; in s5p_mfc_alloc_priv_buf() 39 unsigned int align = (SZ_64K >> PAGE_SHIFT) - 1; in s5p_mfc_alloc_priv_buf() 50 offset = start << PAGE_SHIFT; in s5p_mfc_alloc_priv_buf() 99 unsigned int start = (b->dma - dev->mem_base) >> PAGE_SHIFT; in s5p_mfc_release_priv_buf() 100 unsigned int count = b->size >> PAGE_SHIFT; in s5p_mfc_release_priv_buf()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | pgtable.h | 31 #define PGDIR_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG)) 33 #define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG)) 36 #define PGDIR_SHIFT (PMD_SHIFT + (PAGE_SHIFT - PTRLOG)) 38 #define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG)) 41 #define PUD_SHIFT (PMD_SHIFT + (PAGE_SHIFT - PTRLOG)) 44 #define PGDIR_SHIFT (PUD_SHIFT + (PAGE_SHIFT - PTRLOG)) 53 #define VA_BITS (PGDIR_SHIFT + (PAGE_SHIFT - PTRLOG)) 92 #define PKMAP_END ((FIXADDR_START) & ~((LAST_PKMAP << PAGE_SHIFT)-1)) 188 #define p4d_page(p4d) (pfn_to_page(p4d_phys(p4d) >> PAGE_SHIFT)) 234 #define pud_page(pud) (pfn_to_page(pud_phys(pud) >> PAGE_SHIFT)) [all …]
|