/linux/arch/mips/mm/ |
H A D | mmap.c | 23 #define COLOUR_ALIGN(addr, pgoff) \ argument 25 (((pgoff) << PAGE_SHIFT) & shm_align_mask)) 30 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument 52 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area_common() 64 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_common() 76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_common() 101 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 105 addr0, len, pgoff, flags, UP); in arch_get_unmapped_area() 113 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 117 addr0, len, pgoff, flags, DOWN); in arch_get_unmapped_area_topdown()
|
/linux/arch/loongarch/mm/ |
H A D | mmap.c | 15 #define COLOUR_ALIGN(addr, pgoff) \ argument 17 + (((pgoff) << PAGE_SHIFT) & SHM_ALIGN_MASK)) 22 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument 44 ((addr - (pgoff << PAGE_SHIFT)) & SHM_ALIGN_MASK)) in arch_get_unmapped_area_common() 56 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_common() 67 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_common() 96 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 100 addr0, len, pgoff, flags, UP); in arch_get_unmapped_area() 108 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 112 addr0, len, pgoff, flags, DOWN); in arch_get_unmapped_area_topdown()
|
/linux/drivers/dax/ |
H A D | device.c | 54 __weak phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, in dax_pgoff_to_phys() argument 65 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 66 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys() 68 phys = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys() 82 pgoff_t pgoff; in dax_set_mapping() local 88 pgoff = linear_page_index(vmf->vma, in dax_set_mapping() 99 page->index = pgoff + i; in dax_set_mapping() 123 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE); in __dev_dax_pte_fault() 125 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff); in __dev_dax_pte_fault() 142 pgoff_t pgoff; in __dev_dax_pmd_fault() local [all …]
|
/linux/tools/testing/nvdimm/ |
H A D | dax-dev.c | 9 phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, in dax_pgoff_to_phys() argument 20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys() 21 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys() 23 addr = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys()
|
H A D | pmem-dax.c | 11 long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, in __pmem_direct_access() argument 15 resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; in __pmem_direct_access() 17 if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512, in __pmem_direct_access() 34 __func__, pmem, pgoff, page_to_pfn(page)); in __pmem_direct_access()
|
/linux/arch/arm/mm/ |
H A D | mmap.c | 16 #define COLOUR_ALIGN(addr,pgoff) \ argument 18 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1))) 31 unsigned long len, unsigned long pgoff, in arch_get_unmapped_area() argument 52 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area() 62 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 82 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 105 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area_topdown() 113 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown() 127 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
|
/linux/drivers/md/ |
H A D | dm-linear.c | 160 static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in linear_dax_pgoff() argument 163 sector_t sector = linear_map_sector(ti, *pgoff << PAGE_SECTORS_SHIFT); in linear_dax_pgoff() 165 *pgoff = (get_start_sect(lc->dev->bdev) + sector) >> PAGE_SECTORS_SHIFT; in linear_dax_pgoff() 169 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() argument 173 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_direct_access() 175 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access() 178 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() argument 181 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_zero_page_range() 183 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in linear_dax_zero_page_range() 186 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write() argument [all...] |
H A D | dm-stripe.c | 302 static struct dax_device *stripe_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in stripe_dax_pgoff() argument 309 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); in stripe_dax_pgoff() 313 *pgoff = (get_start_sect(bdev) + dev_sector) >> PAGE_SECTORS_SHIFT; in stripe_dax_pgoff() 317 static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_direct_access() argument 321 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_direct_access() 323 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in stripe_dax_direct_access() 326 static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_zero_page_range() argument 329 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_zero_page_range() 331 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in stripe_dax_zero_page_range() 334 static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_recovery_write() argument [all...] |
/linux/arch/sh/mm/ |
H A D | mmap.c | 46 unsigned long pgoff) in COLOUR_ALIGN() argument 49 unsigned long off = (pgoff << PAGE_SHIFT) & shm_align_mask; in COLOUR_ALIGN() 55 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 68 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area() 82 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 96 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area() 102 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument 116 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area_topdown() 131 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown() 146 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
|
/linux/tools/testing/radix-tree/ |
H A D | iteration_check.c | 51 unsigned long pgoff; in add_entries_fn() local 53 for (pgoff = 0; pgoff < MAX_IDX; pgoff++) { in add_entries_fn() 54 my_item_insert(&array, pgoff); in add_entries_fn() 138 int pgoff; in remove_entries_fn() local 141 pgoff = rand_r(&seeds[2]) % MAX_IDX; in remove_entries_fn() 143 item = xa_erase(&array, pgoff); in remove_entries_fn() 145 item_free(item, pgoff); in remove_entries_fn()
|
/linux/arch/parisc/kernel/ |
H A D | sys_parisc.c | 38 unsigned long pgoff) in shared_align_offset() argument 40 return (filp_pgoff + pgoff) << PAGE_SHIFT; in shared_align_offset() 44 unsigned long filp_pgoff, unsigned long pgoff) in COLOR_ALIGN() argument 48 shared_align_offset(filp_pgoff, pgoff); in COLOR_ALIGN() 100 unsigned long addr, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument 125 (addr - shared_align_offset(filp_pgoff, pgoff)) in arch_get_unmapped_area_common() 133 addr = COLOR_ALIGN(addr, filp_pgoff, pgoff); in arch_get_unmapped_area_common() 145 info.align_offset = shared_align_offset(filp_pgoff, pgoff); in arch_get_unmapped_area_common() 170 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 174 addr, len, pgoff, flags, UP); in arch_get_unmapped_area() [all …]
|
/linux/arch/csky/abiv1/ |
H A D | mmap.c | 12 #define COLOUR_ALIGN(addr,pgoff) \ argument 14 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1))) 26 unsigned long len, unsigned long pgoff, in arch_get_unmapped_area() argument 36 .align_offset = pgoff << PAGE_SHIFT in arch_get_unmapped_area() 50 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area() 60 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
|
/linux/mm/ |
H A D | mmap.c | 268 unsigned long pgoff, unsigned long len) in file_mmap_ok() argument 275 if (pgoff > maxsize >> PAGE_SHIFT) in file_mmap_ok() 340 unsigned long pgoff, unsigned long *populate, in do_mmap() argument 376 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) in do_mmap() 409 addr = __get_unmapped_area(file, addr, len, pgoff, flags, vm_flags); in do_mmap() 430 if (!file_mmap_ok(file, inode, pgoff, len)) in do_mmap() 503 pgoff = 0; in do_mmap() 540 pgoff = addr >> PAGE_SHIFT; in do_mmap() 561 addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); in do_mmap() 571 unsigned long fd, unsigned long pgoff) in ksys_mmap_pgoff() argument [all …]
|
/linux/include/trace/events/ |
H A D | fs_dax.h | 20 __field(pgoff_t, pgoff) 34 __entry->pgoff = vmf->pgoff; 39 "%#lx vm_end %#lx pgoff %#lx max_pgoff %#lx %s", 48 __entry->pgoff, 160 __field(pgoff_t, pgoff) 171 __entry->pgoff = vmf->pgoff; 174 TP_printk("dev %d:%d ino %#lx %s %s address %#lx pgoff %#lx %s", 181 __entry->pgoff, [all...] |
/linux/arch/xtensa/kernel/ |
H A D | syscall.c | 35 #define COLOUR_ALIGN(addr, pgoff) \ argument 37 (((pgoff) << PAGE_SHIFT) & (SHMLBA - 1))) 58 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument 69 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) in arch_get_unmapped_area() 80 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area() 92 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
|
/linux/kernel/bpf/ |
H A D | arena.c | 285 ret = range_tree_clear(&arena->rt, vmf->pgoff, 1); in arena_vm_fault() 292 range_tree_set(&arena->rt, vmf->pgoff, 1); in arena_vm_fault() 298 range_tree_set(&arena->rt, vmf->pgoff, 1); in arena_vm_fault() 315 unsigned long len, unsigned long pgoff, in arena_get_unmapped_area() argument 322 if (pgoff) in arena_get_unmapped_area() 432 long pgoff = 0; in arena_alloc_pages() local 442 pgoff = compute_pgoff(arena, uaddr); in arena_alloc_pages() 443 if (pgoff > page_cnt_max - page_cnt) in arena_alloc_pages() 456 ret = is_range_tree_set(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() 459 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() [all …]
|
/linux/drivers/w1/slaves/ |
H A D | w1_ds250x.c | 49 int pgoff = pageno * W1_PAGE_SIZE; in w1_ds2502_read_page() local 63 buf[1] = pgoff & 0xff; in w1_ds2502_read_page() 64 buf[2] = pgoff >> 8; in w1_ds2502_read_page() 71 w1_read_block(sl->master, &data->eprom[pgoff], W1_PAGE_SIZE); in w1_ds2502_read_page() 74 if (w1_calc_crc8(&data->eprom[pgoff], W1_PAGE_SIZE) != crc8) in w1_ds2502_read_page() 88 int pgoff, epoff; in w1_ds2505_read_page() local 97 epoff = pgoff = pageno * W1_PAGE_SIZE; in w1_ds2505_read_page() 105 buf[1] = pgoff & 0xff; in w1_ds2505_read_page() 106 buf[2] = pgoff >> 8; in w1_ds2505_read_page() 121 pgoff = (redir ^ 0xff) * W1_PAGE_SIZE; in w1_ds2505_read_page()
|
/linux/fs/romfs/ |
H A D | mmap-nommu.c | 21 unsigned long pgoff, in romfs_get_unmapped_area() argument 35 offset = pgoff << PAGE_SHIFT; in romfs_get_unmapped_area() 38 if ((pgoff >= maxpages) || (maxpages - pgoff < lpages)) in romfs_get_unmapped_area() 44 if (len > mtd->size || pgoff >= (mtd->size >> PAGE_SHIFT)) in romfs_get_unmapped_area()
|
/linux/drivers/dma-buf/ |
H A D | udmabuf.c | 51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault() local 55 if (pgoff >= ubuf->pagecount) in udmabuf_vm_fault() 58 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault() 59 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault() 66 pgoff = vma->vm_pgoff; in udmabuf_vm_fault() 69 for (; addr < vma->vm_end; pgoff++, addr += PAGE_SIZE) { in udmabuf_vm_fault() 73 if (WARN_ON(pgoff >= ubuf->pagecount)) in udmabuf_vm_fault() 76 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault() 77 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault() 340 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios() local [all …]
|
/linux/arch/sh/kernel/ |
H A D | sys_sh.c | 43 unsigned long fd, unsigned long pgoff) in sys_mmap2() argument 49 if (pgoff & ((1 << (PAGE_SHIFT - 12)) - 1)) in sys_mmap2() 52 pgoff >>= PAGE_SHIFT - 12; in sys_mmap2() 54 return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); in sys_mmap2()
|
/linux/virt/kvm/ |
H A D | guest_memfd.c | 33 gfn_t gfn = slot->base_gfn + index - slot->gmem.pgoff; in __kvm_gmem_prepare_folio() 73 * the base pgoff of memslots is naturally aligned with the in kvm_gmem_prepare_folio() 80 WARN_ON(!IS_ALIGNED(slot->gmem.pgoff, 1 << folio_order(folio))); in kvm_gmem_prepare_folio() 81 index = gfn - slot->base_gfn + slot->gmem.pgoff; in kvm_gmem_prepare_folio() 114 pgoff_t pgoff = slot->gmem.pgoff; in kvm_gmem_invalidate_begin() local 117 .start = slot->base_gfn + max(pgoff, start) - pgoff, in kvm_gmem_invalidate_begin() 118 .end = slot->base_gfn + min(pgoff + slot->npages, end) - pgoff, in kvm_gmem_invalidate_begin() [all...] |
/linux/tools/perf/arch/x86/util/ |
H A D | machine.c | 26 u64 end, u64 pgoff, const char *name) in add_extra_kernel_map() argument 42 mi->maps[mi->cnt].pgoff = pgoff; in add_extra_kernel_map() 93 xm->pgoff = mi.entry_trampoline; in machine__create_extra_kernel_maps()
|
/linux/fs/cramfs/ |
H A D | inode.c | 288 static u32 cramfs_get_block_range(struct inode *inode, u32 pgoff, u32 *pages) in cramfs_get_block_range() argument 299 blockptrs = (u32 *)(sbi->linear_virt_addr + OFFSET(inode) + pgoff * 4); in cramfs_get_block_range() 309 pgoff+i, pgoff + *pages - 1, in cramfs_get_block_range() 351 unsigned long address, pgoff = vma->vm_pgoff; in cramfs_physmem_mmap() local 371 if (pgoff >= max_pages) in cramfs_physmem_mmap() 373 pages = min(vma_pages(vma), max_pages - pgoff); in cramfs_physmem_mmap() 375 offset = cramfs_get_block_range(inode, pgoff, &pages); in cramfs_physmem_mmap() 385 if (pgoff + pages == max_pages && cramfs_last_page_is_shared(inode)) { in cramfs_physmem_mmap() 425 pgoff, address, pages, vma_pages(vma), vma->vm_start, in cramfs_physmem_mmap() 431 file, pgoff, bailout_reason); in cramfs_physmem_mmap() [all …]
|
/linux/include/linux/sched/ |
H A D | mm.h | 185 unsigned long len, unsigned long pgoff, 189 unsigned long len, unsigned long pgoff, 194 unsigned long pgoff, unsigned long flags); 200 unsigned long pgoff, 206 unsigned long len, unsigned long pgoff, 210 unsigned long len, unsigned long pgoff,
|
/linux/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 50 unsigned long pgoff = offset >> PAGE_SHIFT; in fb_deferred_io_pageref_lookup() local 53 if (fb_WARN_ON_ONCE(info, pgoff >= info->npagerefs)) in fb_deferred_io_pageref_lookup() 57 pageref = &info->pagerefs[pgoff]; in fb_deferred_io_pageref_lookup() 63 pageref->offset = pgoff << PAGE_SHIFT; in fb_deferred_io_pageref_lookup() 135 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault() 149 page->index = vmf->pgoff; /* for folio_mkclean() */ in fb_deferred_io_fault() 232 unsigned long offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_page_mkwrite()
|