| /linux/arch/microblaze/mm/ |
| H A D | pgtable.c | 62 p = addr & PAGE_MASK; in __ioremap() 119 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap() 132 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap() 204 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr() 206 p4d = p4d_offset(pgd, addr & PAGE_MASK); in get_pteptr() 207 pud = pud_offset(p4d, addr & PAGE_MASK); in get_pteptr() 208 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr() 210 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr() 240 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa()
|
| /linux/arch/arc/include/asm/ |
| H A D | pgtable-levels.h | 108 #define p4d_bad(x) ((p4d_val(x) & ~PAGE_MASK)) 111 #define p4d_pgtable(p4d) ((pud_t *)(p4d_val(p4d) & PAGE_MASK)) 130 #define pud_bad(x) ((pud_val(x) & ~PAGE_MASK)) 133 #define pud_pgtable(pud) ((pmd_t *)(pud_val(pud) & PAGE_MASK)) 156 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK)) 159 #define pmd_page_vaddr(pmd) (pmd_val(pmd) & PAGE_MASK) 160 #define pmd_pfn(pmd) ((pmd_val(pmd) & PAGE_MASK) >> PAGE_SHIFT)
|
| /linux/arch/um/include/asm/ |
| H A D | pgtable-4level.h | 59 #define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) 67 #define p4d_bad(x) ((p4d_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) 94 #define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK) 95 #define pud_pgtable(pud) ((pmd_t *) __va(pud_val(pud) & PAGE_MASK)) 97 #define p4d_page(p4d) phys_to_page(p4d_val(p4d) & PAGE_MASK) 98 #define p4d_pgtable(p4d) ((pud_t *) __va(p4d_val(p4d) & PAGE_MASK))
|
| /linux/arch/m68k/kernel/ |
| H A D | sys_m68k.c | 63 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \ 104 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040() 107 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040() 125 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040() 180 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040() 268 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060() 286 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060() 319 addr &= PAGE_MASK; in cache_flush_060() 341 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060() 342 addr &= PAGE_MASK; /* Workaround for bug in some in cache_flush_060()
|
| /linux/arch/sh/mm/ |
| H A D | tlbflush_32.c | 25 page &= PAGE_MASK; in local_flush_tlb_page() 60 start &= PAGE_MASK; in local_flush_tlb_range() 62 end &= PAGE_MASK; in local_flush_tlb_range() 93 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
|
| /linux/arch/mips/mm/ |
| H A D | tlb-r3k.c | 87 start &= PAGE_MASK; in local_flush_tlb_range() 89 end &= PAGE_MASK; in local_flush_tlb_range() 123 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 125 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 160 page &= PAGE_MASK; in local_flush_tlb_page() 201 address &= PAGE_MASK; in __update_tlb()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | doorbell.c | 56 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user() 65 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user() 67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx4_ib_db_map_user() 79 (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | doorbell.c | 58 (page->user_virt == (virt & PAGE_MASK))) in mlx5_ib_db_map_user() 67 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user() 69 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx5_ib_db_map_user() 83 (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
|
| /linux/arch/powerpc/kernel/ |
| H A D | isa-bridge.c | 44 WARN_ON_ONCE(ISA_IO_BASE & ~PAGE_MASK); in remap_isa_base() 45 WARN_ON_ONCE(pa & ~PAGE_MASK); in remap_isa_base() 46 WARN_ON_ONCE(size & ~PAGE_MASK); in remap_isa_base() 78 if ((range.bus_addr & ~PAGE_MASK) || (range.cpu_addr & ~PAGE_MASK)) { in process_ISA_OF_ranges()
|
| /linux/arch/powerpc/sysdev/ |
| H A D | indirect_pci.c | 162 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 166 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 167 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 168 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 169 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
| /linux/arch/mips/lib/ |
| H A D | r3k_dump_tlb.c | 44 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb() 54 entryhi & PAGE_MASK, in dump_tlb() 56 entrylo0 & PAGE_MASK, in dump_tlb()
|
| /linux/arch/sparc/mm/ |
| H A D | iommu.c | 165 start &= PAGE_MASK; in iommu_flush_iotlb() 211 for (p = vaddr & PAGE_MASK; p < vaddr + len; p += PAGE_SIZE) in __sbus_iommu_map_phys() 286 unsigned int busa = dma_addr & PAGE_MASK; in sbus_iommu_unmap_phys() 287 unsigned long off = dma_addr & ~PAGE_MASK; in sbus_iommu_unmap_phys() 337 BUG_ON((va & ~PAGE_MASK) != 0); in sbus_iommu_alloc() 338 BUG_ON((addr & ~PAGE_MASK) != 0); in sbus_iommu_alloc() 339 BUG_ON((len & ~PAGE_MASK) != 0); in sbus_iommu_alloc() 409 BUG_ON((busa & ~PAGE_MASK) != 0); in sbus_iommu_free() 410 BUG_ON((len & ~PAGE_MASK) != 0); in sbus_iommu_free()
|
| H A D | fault_32.c | 356 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_overflow_fault() 365 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_underflow_fault() 377 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_ret_fault()
|
| /linux/arch/m68k/mm/ |
| H A D | memory.c | 110 pushcl040(paddr & PAGE_MASK); in cache_clear() 116 paddr &= PAGE_MASK; in cache_clear() 164 paddr &= PAGE_MASK; in cache_push()
|
| H A D | cache.c | 50 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 71 address &= PAGE_MASK; in flush_icache_user_range()
|
| /linux/drivers/uio/ |
| H A D | uio_dfl.c | 31 uiomem->addr = r->start & PAGE_MASK; in uio_dfl_probe() 32 uiomem->offs = r->start & ~PAGE_MASK; in uio_dfl_probe() 34 + PAGE_SIZE - 1) & PAGE_MASK; in uio_dfl_probe()
|
| H A D | uio_mf624.c | 123 mem->addr = start & PAGE_MASK; in mf624_setup_mem() 124 mem->offs = start & ~PAGE_MASK; in mf624_setup_mem() 127 mem->size = ((start & ~PAGE_MASK) + len + PAGE_SIZE - 1) & PAGE_MASK; in mf624_setup_mem()
|
| H A D | uio_pci_generic.c | 122 uiomem->addr = r->start & PAGE_MASK; in probe() 123 uiomem->offs = r->start & ~PAGE_MASK; in probe() 126 PAGE_MASK; in probe()
|
| /linux/include/vdso/ |
| H A D | page.h | 26 #define PAGE_MASK (~((1 << CONFIG_PAGE_SHIFT) - 1)) macro 28 #define PAGE_MASK (~(PAGE_SIZE - 1)) macro
|
| /linux/arch/powerpc/include/asm/nohash/32/ |
| H A D | pte-44x.h | 80 #define _PMD_PRESENT_MASK (PAGE_MASK) 81 #define _PMD_BAD (~PAGE_MASK)
|
| /linux/arch/powerpc/mm/ |
| H A D | cacheflush.c | 88 unsigned long addr = physaddr & PAGE_MASK; in flush_dcache_icache_phys() 133 unsigned long addr = (unsigned long)p & PAGE_MASK; 218 maddr = kmap_local_page(page) + (addr & ~PAGE_MASK); in flush_icache_user_page()
|
| /linux/arch/mips/include/asm/ |
| H A D | ginvt.h | 44 addr &= PAGE_MASK << 1; in ginvt_va() 55 addr &= PAGE_MASK << 1; in ginvt_va_mmid()
|
| /linux/arch/arm/mm/ |
| H A D | mmap.c | 75 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area() 126 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area_topdown() 136 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
|
| /linux/tools/virtio/linux/ |
| H A D | kernel.h | 28 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 29 #define PAGE_ALIGN(x) ((x + PAGE_SIZE - 1) & PAGE_MASK) 48 #define virt_to_page(p) ((struct page *)((unsigned long)p & PAGE_MASK))
|
| /linux/arch/sparc/kernel/ |
| H A D | ioport.c | 127 unsigned long vaddr = (unsigned long) virtual & PAGE_MASK; in iounmap() 207 unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); in _sparc_ioremap() 210 (offset + sz + PAGE_SIZE-1) & PAGE_MASK, in _sparc_ioremap() 218 pa &= PAGE_MASK; in _sparc_ioremap()
|