Home
last modified time | relevance | path

Searched refs:PAGE_MASK (Results 1 – 25 of 428) sorted by relevance

12345678910>>...18

/linux/arch/microblaze/mm/
H A Dpgtable.c62 p = addr & PAGE_MASK; in __ioremap()
119 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap()
132 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap()
204 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr()
206 p4d = p4d_offset(pgd, addr & PAGE_MASK); in get_pteptr()
207 pud = pud_offset(p4d, addr & PAGE_MASK); in get_pteptr()
208 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr()
210 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr()
240 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa()
/linux/arch/arc/include/asm/
H A Dpgtable-levels.h108 #define p4d_bad(x) ((p4d_val(x) & ~PAGE_MASK))
111 #define p4d_pgtable(p4d) ((pud_t *)(p4d_val(p4d) & PAGE_MASK))
130 #define pud_bad(x) ((pud_val(x) & ~PAGE_MASK))
133 #define pud_pgtable(pud) ((pmd_t *)(pud_val(pud) & PAGE_MASK))
156 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK))
159 #define pmd_page_vaddr(pmd) (pmd_val(pmd) & PAGE_MASK)
160 #define pmd_pfn(pmd) ((pmd_val(pmd) & PAGE_MASK) >> PAGE_SHIFT)
/linux/arch/um/include/asm/
H A Dpgtable-4level.h59 #define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
67 #define p4d_bad(x) ((p4d_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
94 #define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK)
95 #define pud_pgtable(pud) ((pmd_t *) __va(pud_val(pud) & PAGE_MASK))
97 #define p4d_page(p4d) phys_to_page(p4d_val(p4d) & PAGE_MASK)
98 #define p4d_pgtable(p4d) ((pud_t *) __va(p4d_val(p4d) & PAGE_MASK))
/linux/arch/m68k/kernel/
H A Dsys_m68k.c63 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \
104 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040()
107 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040()
125 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040()
180 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040()
268 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060()
286 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060()
319 addr &= PAGE_MASK; in cache_flush_060()
341 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060()
342 addr &= PAGE_MASK; /* Workaround for bug in some in cache_flush_060()
/linux/arch/sh/mm/
H A Dtlbflush_32.c25 page &= PAGE_MASK; in local_flush_tlb_page()
60 start &= PAGE_MASK; in local_flush_tlb_range()
62 end &= PAGE_MASK; in local_flush_tlb_range()
93 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
/linux/arch/mips/mm/
H A Dtlb-r3k.c87 start &= PAGE_MASK; in local_flush_tlb_range()
89 end &= PAGE_MASK; in local_flush_tlb_range()
123 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
125 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
160 page &= PAGE_MASK; in local_flush_tlb_page()
201 address &= PAGE_MASK; in __update_tlb()
/linux/drivers/infiniband/hw/mlx4/
H A Ddoorbell.c56 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user()
65 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user()
67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx4_ib_db_map_user()
79 (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
/linux/drivers/infiniband/hw/mlx5/
H A Ddoorbell.c58 (page->user_virt == (virt & PAGE_MASK))) in mlx5_ib_db_map_user()
67 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user()
69 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx5_ib_db_map_user()
83 (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
/linux/arch/powerpc/kernel/
H A Disa-bridge.c44 WARN_ON_ONCE(ISA_IO_BASE & ~PAGE_MASK); in remap_isa_base()
45 WARN_ON_ONCE(pa & ~PAGE_MASK); in remap_isa_base()
46 WARN_ON_ONCE(size & ~PAGE_MASK); in remap_isa_base()
78 if ((range.bus_addr & ~PAGE_MASK) || (range.cpu_addr & ~PAGE_MASK)) { in process_ISA_OF_ranges()
/linux/arch/powerpc/sysdev/
H A Dindirect_pci.c162 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci()
166 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci()
167 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci()
168 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci()
169 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
/linux/arch/mips/lib/
H A Dr3k_dump_tlb.c44 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb()
54 entryhi & PAGE_MASK, in dump_tlb()
56 entrylo0 & PAGE_MASK, in dump_tlb()
/linux/arch/sparc/mm/
H A Diommu.c165 start &= PAGE_MASK; in iommu_flush_iotlb()
211 for (p = vaddr & PAGE_MASK; p < vaddr + len; p += PAGE_SIZE) in __sbus_iommu_map_phys()
286 unsigned int busa = dma_addr & PAGE_MASK; in sbus_iommu_unmap_phys()
287 unsigned long off = dma_addr & ~PAGE_MASK; in sbus_iommu_unmap_phys()
337 BUG_ON((va & ~PAGE_MASK) != 0); in sbus_iommu_alloc()
338 BUG_ON((addr & ~PAGE_MASK) != 0); in sbus_iommu_alloc()
339 BUG_ON((len & ~PAGE_MASK) != 0); in sbus_iommu_alloc()
409 BUG_ON((busa & ~PAGE_MASK) != 0); in sbus_iommu_free()
410 BUG_ON((len & ~PAGE_MASK) != 0); in sbus_iommu_free()
H A Dfault_32.c356 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_overflow_fault()
365 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_underflow_fault()
377 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_ret_fault()
/linux/arch/m68k/mm/
H A Dmemory.c110 pushcl040(paddr & PAGE_MASK); in cache_clear()
116 paddr &= PAGE_MASK; in cache_clear()
164 paddr &= PAGE_MASK; in cache_push()
H A Dcache.c50 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
71 address &= PAGE_MASK; in flush_icache_user_range()
/linux/drivers/uio/
H A Duio_dfl.c31 uiomem->addr = r->start & PAGE_MASK; in uio_dfl_probe()
32 uiomem->offs = r->start & ~PAGE_MASK; in uio_dfl_probe()
34 + PAGE_SIZE - 1) & PAGE_MASK; in uio_dfl_probe()
H A Duio_mf624.c123 mem->addr = start & PAGE_MASK; in mf624_setup_mem()
124 mem->offs = start & ~PAGE_MASK; in mf624_setup_mem()
127 mem->size = ((start & ~PAGE_MASK) + len + PAGE_SIZE - 1) & PAGE_MASK; in mf624_setup_mem()
H A Duio_pci_generic.c122 uiomem->addr = r->start & PAGE_MASK; in probe()
123 uiomem->offs = r->start & ~PAGE_MASK; in probe()
126 PAGE_MASK; in probe()
/linux/include/vdso/
H A Dpage.h26 #define PAGE_MASK (~((1 << CONFIG_PAGE_SHIFT) - 1)) macro
28 #define PAGE_MASK (~(PAGE_SIZE - 1)) macro
/linux/arch/powerpc/include/asm/nohash/32/
H A Dpte-44x.h80 #define _PMD_PRESENT_MASK (PAGE_MASK)
81 #define _PMD_BAD (~PAGE_MASK)
/linux/arch/powerpc/mm/
H A Dcacheflush.c88 unsigned long addr = physaddr & PAGE_MASK; in flush_dcache_icache_phys()
133 unsigned long addr = (unsigned long)p & PAGE_MASK;
218 maddr = kmap_local_page(page) + (addr & ~PAGE_MASK); in flush_icache_user_page()
/linux/arch/mips/include/asm/
H A Dginvt.h44 addr &= PAGE_MASK << 1; in ginvt_va()
55 addr &= PAGE_MASK << 1; in ginvt_va_mmid()
/linux/arch/arm/mm/
H A Dmmap.c75 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area()
126 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area_topdown()
136 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
/linux/tools/virtio/linux/
H A Dkernel.h28 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
29 #define PAGE_ALIGN(x) ((x + PAGE_SIZE - 1) & PAGE_MASK)
48 #define virt_to_page(p) ((struct page *)((unsigned long)p & PAGE_MASK))
/linux/arch/sparc/kernel/
H A Dioport.c127 unsigned long vaddr = (unsigned long) virtual & PAGE_MASK; in iounmap()
207 unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); in _sparc_ioremap()
210 (offset + sz + PAGE_SIZE-1) & PAGE_MASK, in _sparc_ioremap()
218 pa &= PAGE_MASK; in _sparc_ioremap()

12345678910>>...18