Home
last modified time | relevance | path

Searched refs:PAGE_SHIFT (Results 1 – 25 of 1064) sorted by relevance

12345678910>>...43

/linux/arch/riscv/kvm/
H A Dmmu.c23 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in mmu_wp_memory_region()
24 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in mmu_wp_memory_region()
99 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked()
100 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked()
128 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()
129 phys_addr_t size = slot->npages << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()
174 kvm_riscv_gstage_gpa_size(kvm->arch.pgd_levels) >> PAGE_SHIFT) in kvm_arch_prepare_memory_region()
178 size = new->npages << PAGE_SHIFT; in kvm_arch_prepare_memory_region()
240 kvm_riscv_gstage_unmap_range(&gstage, range->start << PAGE_SHIFT, in kvm_unmap_gfn_range()
241 (range->end - range->start) << PAGE_SHIFT, in kvm_unmap_gfn_range()
[all …]
/linux/arch/arc/include/asm/
H A Dhighmem.h24 #define FIXADDR_TOP (FIXMAP_BASE + (FIX_KMAP_END << PAGE_SHIFT))
30 #define __fix_to_virt(x) (FIXADDR_TOP - ((x) << PAGE_SHIFT))
31 #define __virt_to_fix(x) (((FIXADDR_TOP - ((x) & PAGE_MASK))) >> PAGE_SHIFT)
35 #define LAST_PKMAP (PKMAP_SIZE >> PAGE_SHIFT)
37 #define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
38 #define PKMAP_NR(virt) (((virt) - PKMAP_BASE) >> PAGE_SHIFT)
H A Dpgtable-levels.h86 #define PTRS_PER_PTE BIT(PMD_SHIFT - PAGE_SHIFT)
143 #define pmd_pfn(pmd) ((pmd_val(pmd) & PMD_MASK) >> PAGE_SHIFT)
144 #define pfn_pmd(pfn,prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
160 #define pmd_pfn(pmd) ((pmd_val(pmd) & PAGE_MASK) >> PAGE_SHIFT)
171 #define PFN_PTE_SHIFT PAGE_SHIFT
177 #define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
/linux/arch/arm/mm/
H A Dtlb-v7.S39 mov r0, r0, lsr #PAGE_SHIFT @ align address
40 mov r1, r1, lsr #PAGE_SHIFT
46 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA
47 mov r1, r1, lsl #PAGE_SHIFT
73 mov r0, r0, lsr #PAGE_SHIFT @ align address
74 mov r1, r1, lsr #PAGE_SHIFT
75 mov r0, r0, lsl #PAGE_SHIFT
76 mov r1, r1, lsl #PAGE_SHIFT
H A Dtlb-v6.S41 mov r0, r0, lsr #PAGE_SHIFT @ align address
42 mov r1, r1, lsr #PAGE_SHIFT
44 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA
45 mov r1, r1, lsl #PAGE_SHIFT
73 mov r0, r0, lsr #PAGE_SHIFT @ align address
74 mov r1, r1, lsr #PAGE_SHIFT
75 mov r0, r0, lsl #PAGE_SHIFT
76 mov r1, r1, lsl #PAGE_SHIFT
H A Dmmap.c18 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
52 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area()
76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area()
105 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area_topdown()
127 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
166 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT)); in valid_mmap_phys_addr_range()
/linux/arch/x86/include/asm/
H A Dpgalloc.h65 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel()
72 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe()
82 set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
99 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
105 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe()
113 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate()
119 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate_safe()
136 paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); in pgd_populate()
144 paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); in pgd_populate_safe()
/linux/arch/powerpc/mm/book3s64/
H A Diommu_api.c80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
100 chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) / in mm_iommu_do_alloc()
106 ret = pin_user_pages(ua + (entry << PAGE_SHIFT), n, in mm_iommu_do_alloc()
135 if ((mem2->ua < (ua + (entries << PAGE_SHIFT))) && in mm_iommu_do_alloc()
137 (mem2->entries << PAGE_SHIFT)))) { in mm_iommu_do_alloc()
150 pageshift = PAGE_SHIFT; in mm_iommu_do_alloc()
154 if ((mem->pageshift > PAGE_SHIFT) && PageHuge(page)) in mm_iommu_do_alloc()
161 mem->hpas[i] = page_to_pfn(page) << PAGE_SHIFT; in mm_iommu_do_alloc()
214 page = pfn_to_page(mem->hpas[i] >> PAGE_SHIFT); in mm_iommu_unpin()
[all …]
H A Dsubpage_prot.c120 spp += (addr >> PAGE_SHIFT) & (SBP_L1_COUNT - 1); in subpage_prot_clear()
122 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in subpage_prot_clear()
124 if (addr + (nw << PAGE_SHIFT) > next) in subpage_prot_clear()
125 nw = (next - addr) >> PAGE_SHIFT; in subpage_prot_clear()
213 if (!access_ok(map, (len >> PAGE_SHIFT) * sizeof(u32))) in SYSCALL_DEFINE3()
255 spp += (addr >> PAGE_SHIFT) & (SBP_L1_COUNT - 1); in SYSCALL_DEFINE3()
261 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in SYSCALL_DEFINE3()
263 if (addr + (nw << PAGE_SHIFT) > next) in SYSCALL_DEFINE3()
264 nw = (next - addr) >> PAGE_SHIFT; in SYSCALL_DEFINE3()
/linux/arch/x86/kernel/cpu/mtrr/
H A Dcentaur.c52 *base = centaur_mcr[reg].high >> PAGE_SHIFT; in centaur_get_mcr()
53 *size = -(centaur_mcr[reg].low & 0xfffff000) >> PAGE_SHIFT; in centaur_get_mcr()
74 high = base << PAGE_SHIFT; in centaur_set_mcr()
77 low = -size << PAGE_SHIFT | 0x1f; in centaur_set_mcr()
80 low = -size << PAGE_SHIFT | 0x02; /* NC */ in centaur_set_mcr()
82 low = -size << PAGE_SHIFT | 0x09; /* WWO, WC */ in centaur_set_mcr()
H A Damd.c20 *base = (low & 0xFFFE0000) >> PAGE_SHIFT; in amd_get_mtrr()
46 *size = (low + 4) << (15 - PAGE_SHIFT); in amd_get_mtrr()
83 regs[reg] = (-size >> (15 - PAGE_SHIFT) & 0x0001FFFC) in amd_set_mtrr()
84 | (base << PAGE_SHIFT) | (type + 1); in amd_set_mtrr()
106 if (type > MTRR_TYPE_WRCOMB || size < (1 << (17 - PAGE_SHIFT)) in amd_validate_add_page()
/linux/arch/loongarch/kvm/
H A Dmmu.c68 offset = (addr >> PAGE_SHIFT) - ctx->gfn; in kvm_mkclean_pte()
204 size = 0x1UL << (ctx->pgtable_shift + PAGE_SHIFT - 3); in kvm_ptw_leaf()
248 size = 0x1UL << (ctx->pgtable_shift + PAGE_SHIFT - 3); in kvm_ptw_dir()
305 ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, in kvm_flush_range()
306 end_gfn << PAGE_SHIFT, &ctx); in kvm_flush_range()
309 ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, in kvm_flush_range()
310 end_gfn << PAGE_SHIFT, &ctx); in kvm_flush_range()
348 return kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, end_gfn << PAGE_SHIFT, &ctx); in kvm_mkclean_gpa_pt()
376 kvm_ptw_top(kvm->arch.pgd, start << PAGE_SHIFT, end << PAGE_SHIFT, &ctx); in kvm_arch_mmu_enable_log_dirty_pt_masked()
392 if ((new->base_gfn + new->npages) > (kvm->arch.gpa_size >> PAGE_SHIFT)) in kvm_arch_prepare_memory_region()
[all …]
/linux/include/drm/
H A Ddrm_vma_manager.h37 #define DRM_FILE_PAGE_OFFSET_START ((0xFFFFFFFFUL >> PAGE_SHIFT) + 1)
38 #define DRM_FILE_PAGE_OFFSET_SIZE ((0xFFFFFFFFUL >> PAGE_SHIFT) * 256)
40 #define DRM_FILE_PAGE_OFFSET_START ((0xFFFFFFFUL >> PAGE_SHIFT) + 1)
41 #define DRM_FILE_PAGE_OFFSET_SIZE ((0xFFFFFFFUL >> PAGE_SHIFT) * 16)
205 return ((__u64)node->vm_node.start) << PAGE_SHIFT; in drm_vma_node_offset_addr()
226 drm_vma_node_size(node) << PAGE_SHIFT, 1); in drm_vma_node_unmap()
/linux/tools/include/linux/
H A Dpfn.h7 #define PFN_UP(x) (((x) + PAGE_SIZE - 1) >> PAGE_SHIFT)
8 #define PFN_DOWN(x) ((x) >> PAGE_SHIFT)
9 #define PFN_PHYS(x) ((phys_addr_t)(x) << PAGE_SHIFT)
10 #define PHYS_PFN(x) ((unsigned long)((x) >> PAGE_SHIFT))
/linux/include/asm-generic/
H A Dgetorder.h33 return BITS_PER_LONG - PAGE_SHIFT; in get_order()
35 if (size < (1UL << PAGE_SHIFT)) in get_order()
38 return ilog2((size) - 1) - PAGE_SHIFT + 1; in get_order()
42 size >>= PAGE_SHIFT; in get_order()
/linux/arch/csky/kernel/
H A Dvdso.c21 vdso_pages = (vdso_end - vdso_start) >> PAGE_SHIFT; in vdso_init()
32 pg = virt_to_page(vdso_start + (i << PAGE_SHIFT)); in vdso_init()
51 vdso_len = vdso_pages << PAGE_SHIFT; in arch_setup_additional_pages()
69 _install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT, in arch_setup_additional_pages()
79 vdso_base += (vdso_pages << PAGE_SHIFT); in arch_setup_additional_pages()
/linux/arch/powerpc/mm/
H A Dmem.c139 update_end_of_memory_vars(start_pfn << PAGE_SHIFT, in add_pages()
140 nr_pages << PAGE_SHIFT); in add_pages()
148 unsigned long start_pfn = start >> PAGE_SHIFT; in arch_add_memory()
149 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory()
163 unsigned long start_pfn = start >> PAGE_SHIFT; in arch_remove_memory()
164 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory()
174 max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; in mem_topology_setup()
175 min_low_pfn = MEMORY_START >> PAGE_SHIFT; in mem_topology_setup()
177 max_low_pfn = lowmem_end_addr >> PAGE_SHIFT; in mem_topology_setup()
223 max_zone_pfns[ZONE_DMA] = min((zone_dma_limit >> PAGE_SHIFT) + 1, max_low_pfn); in arch_zone_limits_init()
/linux/arch/sparc/include/asm/
H A Dio-unit.h45 unsigned long bmap[(IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 3)) / sizeof(unsigned long)];
53 #define IOUNIT_BMAP1_END (IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 1))
55 #define IOUNIT_BMAP2_END IOUNIT_BMAP2_START + (IOUNIT_DMA_SIZE >> (PAGE_SHIFT + 2))
57 #define IOUNIT_BMAPM_END ((IOUNIT_DMA_SIZE - IOUNIT_DVMA_SIZE) >> PAGE_SHIFT)
H A Dhighmem.h42 #define PKMAP_SIZE (LAST_PKMAP << PAGE_SHIFT)
43 #define PKMAP_BASE PMD_ALIGN(SRMMU_NOCACHE_VADDR + (SRMMU_MAX_NOCACHE_PAGES << PAGE_SHIFT))
46 #define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT)
47 #define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
/linux/mm/
H A Dhmm.c53 unsigned long i = (addr - range->start) >> PAGE_SHIFT; in hmm_pfns_fill()
166 i = (addr - range->start) >> PAGE_SHIFT; in hmm_vma_walk_hole()
167 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_hole()
194 hmm_pfn_flags_order(PMD_SHIFT - PAGE_SHIFT); in pmd_to_hmm_pfn_flags()
207 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_handle_pmd()
214 pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT); in hmm_vma_handle_pmd()
341 unsigned long npages = (end - start) >> PAGE_SHIFT; in hmm_vma_handle_absent_pmd()
350 hmm_pfn_flags_order(PMD_SHIFT - PAGE_SHIFT); in hmm_vma_handle_absent_pmd()
388 unsigned long npages = (end - start) >> PAGE_SHIFT; in hmm_vma_handle_absent_pmd()
404 &range->hmm_pfns[(start - range->start) >> PAGE_SHIFT]; in hmm_vma_walk_pmd()
[all …]
/linux/arch/hexagon/include/asm/
H A Dmem-layout.h38 #define PHYS_PFN_OFFSET (PHYS_OFFSET >> PAGE_SHIFT)
76 #define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT)
86 #define PKMAP_NR(virt) ((virt - PKMAP_BASE) >> PAGE_SHIFT)
87 #define PKMAP_ADDR(nr) (PKMAP_BASE + ((nr) << PAGE_SHIFT))
/linux/drivers/gpu/drm/gma500/
H A Dgtt.c58 BUG_ON(pfn & ~(0xFFFFFFFF >> PAGE_SHIFT)); in psb_gtt_mask_pte()
67 return (pfn << PAGE_SHIFT) | mask; in psb_gtt_mask_pte()
74 return pdev->gtt_map + (offset >> PAGE_SHIFT); in psb_gtt_entry()
89 npages = resource_size(res) >> PAGE_SHIFT; in psb_gtt_insert_pages()
116 npages = resource_size(res) >> PAGE_SHIFT; in psb_gtt_remove_pages()
206 gtt_pages = pci_resource_len(pdev, PSB_GTT_RESOURCE) >> PAGE_SHIFT; in psb_gtt_init_ranges()
216 gatt_pages = pci_resource_len(pdev, PSB_GATT_RESOURCE) >> PAGE_SHIFT; in psb_gtt_init_ranges()
228 gatt_pages = (128 * 1024 * 1024) >> PAGE_SHIFT; in psb_gtt_init_ranges()
269 dev_priv->gtt_map = ioremap(pg->gtt_phys_start, pg->gtt_pages << PAGE_SHIFT); in psb_gtt_init()
/linux/arch/x86/mm/pat/
H A Dmemtype.c403 unsigned long start_pfn = start >> PAGE_SHIFT; in pat_pagerange_is_ram()
404 unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; in pat_pagerange_is_ram()
414 if (start_pfn < ISA_END_ADDRESS >> PAGE_SHIFT) in pat_pagerange_is_ram()
415 start_pfn = ISA_END_ADDRESS >> PAGE_SHIFT; in pat_pagerange_is_ram()
455 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type()
473 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type()
485 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in free_ram_pages_type()
666 page = pfn_to_page(paddr >> PAGE_SHIFT); in lookup_memtype()
772 if (!phys_mem_access_encrypted(pfn << PAGE_SHIFT, size)) in phys_mem_access_prot()
818 if (!page_is_ram(base >> PAGE_SHIFT)) in memtype_kernel_map_sync()
[all …]
/linux/drivers/media/platform/samsung/s5p-mfc/
H A Ds5p_mfc_opr.c37 unsigned int bits = dev->mem_size >> PAGE_SHIFT; in s5p_mfc_alloc_priv_buf()
38 unsigned int count = b->size >> PAGE_SHIFT; in s5p_mfc_alloc_priv_buf()
39 unsigned int align = (SZ_64K >> PAGE_SHIFT) - 1; in s5p_mfc_alloc_priv_buf()
50 offset = start << PAGE_SHIFT; in s5p_mfc_alloc_priv_buf()
99 unsigned int start = (b->dma - dev->mem_base) >> PAGE_SHIFT; in s5p_mfc_release_priv_buf()
100 unsigned int count = b->size >> PAGE_SHIFT; in s5p_mfc_release_priv_buf()
/linux/arch/loongarch/include/asm/
H A Dpgtable.h31 #define PGDIR_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG))
33 #define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG))
36 #define PGDIR_SHIFT (PMD_SHIFT + (PAGE_SHIFT - PTRLOG))
38 #define PMD_SHIFT (PAGE_SHIFT + (PAGE_SHIFT - PTRLOG))
41 #define PUD_SHIFT (PMD_SHIFT + (PAGE_SHIFT - PTRLOG))
44 #define PGDIR_SHIFT (PUD_SHIFT + (PAGE_SHIFT - PTRLOG))
53 #define VA_BITS (PGDIR_SHIFT + (PAGE_SHIFT - PTRLOG))
92 #define PKMAP_END ((FIXADDR_START) & ~((LAST_PKMAP << PAGE_SHIFT)-1))
188 #define p4d_page(p4d) (pfn_to_page(p4d_phys(p4d) >> PAGE_SHIFT))
234 #define pud_page(pud) (pfn_to_page(pud_phys(pud) >> PAGE_SHIFT))
[all …]

12345678910>>...43