Home
last modified time | relevance | path

Searched refs:pgoff (Results 1 – 25 of 170) sorted by relevance

1234567

/linux/arch/mips/mm/
H A Dmmap.c23 #define COLOUR_ALIGN(addr, pgoff) \ argument
25 (((pgoff) << PAGE_SHIFT) & shm_align_mask))
30 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument
52 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area_common()
64 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_common()
76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_common()
101 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument
105 addr0, len, pgoff, flags, UP); in arch_get_unmapped_area()
113 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_topdown() argument
117 addr0, len, pgoff, flags, DOWN); in arch_get_unmapped_area_topdown()
/linux/arch/loongarch/mm/
H A Dmmap.c15 #define COLOUR_ALIGN(addr, pgoff) \ argument
17 + (((pgoff) << PAGE_SHIFT) & SHM_ALIGN_MASK))
22 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument
44 ((addr - (pgoff << PAGE_SHIFT)) & SHM_ALIGN_MASK)) in arch_get_unmapped_area_common()
56 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_common()
67 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_common()
96 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument
100 addr0, len, pgoff, flags, UP); in arch_get_unmapped_area()
108 unsigned long addr0, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_topdown() argument
112 addr0, len, pgoff, flags, DOWN); in arch_get_unmapped_area_topdown()
/linux/drivers/dax/
H A Ddevice.c54 __weak phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, in dax_pgoff_to_phys() argument
65 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys()
66 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys()
68 phys = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys()
82 pgoff_t pgoff; in dax_set_mapping() local
88 pgoff = linear_page_index(vmf->vma, in dax_set_mapping()
99 page->index = pgoff + i; in dax_set_mapping()
123 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE); in __dev_dax_pte_fault()
125 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff); in __dev_dax_pte_fault()
142 pgoff_t pgoff; in __dev_dax_pmd_fault() local
[all …]
/linux/tools/testing/nvdimm/
H A Ddax-dev.c9 phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, in dax_pgoff_to_phys() argument
20 pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1; in dax_pgoff_to_phys()
21 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys()
23 addr = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys()
H A Dpmem-dax.c11 long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, in __pmem_direct_access() argument
15 resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; in __pmem_direct_access()
17 if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512, in __pmem_direct_access()
34 __func__, pmem, pgoff, page_to_pfn(page)); in __pmem_direct_access()
/linux/arch/arm/mm/
H A Dmmap.c16 #define COLOUR_ALIGN(addr,pgoff) \ argument
18 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
31 unsigned long len, unsigned long pgoff, in arch_get_unmapped_area() argument
52 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area()
62 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
76 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area()
82 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument
105 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area_topdown()
113 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown()
127 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
/linux/drivers/md/
H A Ddm-linear.c160 static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in linear_dax_pgoff() argument
163 sector_t sector = linear_map_sector(ti, *pgoff << PAGE_SECTORS_SHIFT); in linear_dax_pgoff()
165 *pgoff = (get_start_sect(lc->dev->bdev) + sector) >> PAGE_SECTORS_SHIFT; in linear_dax_pgoff()
169 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() argument
173 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_direct_access()
175 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access()
178 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() argument
181 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_zero_page_range()
183 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in linear_dax_zero_page_range()
186 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write() argument
[all...]
H A Ddm-stripe.c302 static struct dax_device *stripe_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in stripe_dax_pgoff() argument
309 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); in stripe_dax_pgoff()
313 *pgoff = (get_start_sect(bdev) + dev_sector) >> PAGE_SECTORS_SHIFT; in stripe_dax_pgoff()
317 static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_direct_access() argument
321 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_direct_access()
323 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in stripe_dax_direct_access()
326 static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_zero_page_range() argument
329 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_zero_page_range()
331 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in stripe_dax_zero_page_range()
334 static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_recovery_write() argument
[all...]
/linux/arch/sh/mm/
H A Dmmap.c46 unsigned long pgoff) in COLOUR_ALIGN() argument
49 unsigned long off = (pgoff << PAGE_SHIFT) & shm_align_mask; in COLOUR_ALIGN()
55 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument
68 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area()
82 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
96 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area()
102 const unsigned long len, const unsigned long pgoff, in arch_get_unmapped_area_topdown() argument
116 ((addr - (pgoff << PAGE_SHIFT)) & shm_align_mask)) in arch_get_unmapped_area_topdown()
131 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area_topdown()
146 info.align_offset = pgoff << PAGE_SHIFT; in arch_get_unmapped_area_topdown()
/linux/tools/testing/radix-tree/
H A Diteration_check.c51 unsigned long pgoff; in add_entries_fn() local
53 for (pgoff = 0; pgoff < MAX_IDX; pgoff++) { in add_entries_fn()
54 my_item_insert(&array, pgoff); in add_entries_fn()
138 int pgoff; in remove_entries_fn() local
141 pgoff = rand_r(&seeds[2]) % MAX_IDX; in remove_entries_fn()
143 item = xa_erase(&array, pgoff); in remove_entries_fn()
145 item_free(item, pgoff); in remove_entries_fn()
/linux/arch/parisc/kernel/
H A Dsys_parisc.c38 unsigned long pgoff) in shared_align_offset() argument
40 return (filp_pgoff + pgoff) << PAGE_SHIFT; in shared_align_offset()
44 unsigned long filp_pgoff, unsigned long pgoff) in COLOR_ALIGN() argument
48 shared_align_offset(filp_pgoff, pgoff); in COLOR_ALIGN()
100 unsigned long addr, unsigned long len, unsigned long pgoff, in arch_get_unmapped_area_common() argument
125 (addr - shared_align_offset(filp_pgoff, pgoff)) in arch_get_unmapped_area_common()
133 addr = COLOR_ALIGN(addr, filp_pgoff, pgoff); in arch_get_unmapped_area_common()
145 info.align_offset = shared_align_offset(filp_pgoff, pgoff); in arch_get_unmapped_area_common()
170 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument
174 addr, len, pgoff, flags, UP); in arch_get_unmapped_area()
[all …]
/linux/arch/csky/abiv1/
H A Dmmap.c12 #define COLOUR_ALIGN(addr,pgoff) \ argument
14 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
26 unsigned long len, unsigned long pgoff, in arch_get_unmapped_area() argument
36 .align_offset = pgoff << PAGE_SHIFT in arch_get_unmapped_area()
50 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) in arch_get_unmapped_area()
60 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
/linux/mm/
H A Dmmap.c268 unsigned long pgoff, unsigned long len) in file_mmap_ok() argument
275 if (pgoff > maxsize >> PAGE_SHIFT) in file_mmap_ok()
340 unsigned long pgoff, unsigned long *populate, in do_mmap() argument
376 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) in do_mmap()
409 addr = __get_unmapped_area(file, addr, len, pgoff, flags, vm_flags); in do_mmap()
430 if (!file_mmap_ok(file, inode, pgoff, len)) in do_mmap()
503 pgoff = 0; in do_mmap()
540 pgoff = addr >> PAGE_SHIFT; in do_mmap()
561 addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); in do_mmap()
571 unsigned long fd, unsigned long pgoff) in ksys_mmap_pgoff() argument
[all …]
/linux/include/trace/events/
H A Dfs_dax.h20 __field(pgoff_t, pgoff)
34 __entry->pgoff = vmf->pgoff;
39 "%#lx vm_end %#lx pgoff %#lx max_pgoff %#lx %s",
48 __entry->pgoff,
160 __field(pgoff_t, pgoff)
171 __entry->pgoff = vmf->pgoff;
174 TP_printk("dev %d:%d ino %#lx %s %s address %#lx pgoff %#lx %s",
181 __entry->pgoff,
[all...]
/linux/arch/xtensa/kernel/
H A Dsyscall.c35 #define COLOUR_ALIGN(addr, pgoff) \ argument
37 (((pgoff) << PAGE_SHIFT) & (SHMLBA - 1)))
58 unsigned long len, unsigned long pgoff, unsigned long flags, in arch_get_unmapped_area() argument
69 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) in arch_get_unmapped_area()
80 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
92 addr = COLOUR_ALIGN(addr, pgoff); in arch_get_unmapped_area()
/linux/kernel/bpf/
H A Darena.c285 ret = range_tree_clear(&arena->rt, vmf->pgoff, 1); in arena_vm_fault()
292 range_tree_set(&arena->rt, vmf->pgoff, 1); in arena_vm_fault()
298 range_tree_set(&arena->rt, vmf->pgoff, 1); in arena_vm_fault()
315 unsigned long len, unsigned long pgoff, in arena_get_unmapped_area() argument
322 if (pgoff) in arena_get_unmapped_area()
432 long pgoff = 0; in arena_alloc_pages() local
442 pgoff = compute_pgoff(arena, uaddr); in arena_alloc_pages()
443 if (pgoff > page_cnt_max - page_cnt) in arena_alloc_pages()
456 ret = is_range_tree_set(&arena->rt, pgoff, page_cnt); in arena_alloc_pages()
459 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages()
[all …]
/linux/drivers/w1/slaves/
H A Dw1_ds250x.c49 int pgoff = pageno * W1_PAGE_SIZE; in w1_ds2502_read_page() local
63 buf[1] = pgoff & 0xff; in w1_ds2502_read_page()
64 buf[2] = pgoff >> 8; in w1_ds2502_read_page()
71 w1_read_block(sl->master, &data->eprom[pgoff], W1_PAGE_SIZE); in w1_ds2502_read_page()
74 if (w1_calc_crc8(&data->eprom[pgoff], W1_PAGE_SIZE) != crc8) in w1_ds2502_read_page()
88 int pgoff, epoff; in w1_ds2505_read_page() local
97 epoff = pgoff = pageno * W1_PAGE_SIZE; in w1_ds2505_read_page()
105 buf[1] = pgoff & 0xff; in w1_ds2505_read_page()
106 buf[2] = pgoff >> 8; in w1_ds2505_read_page()
121 pgoff = (redir ^ 0xff) * W1_PAGE_SIZE; in w1_ds2505_read_page()
/linux/fs/romfs/
H A Dmmap-nommu.c21 unsigned long pgoff, in romfs_get_unmapped_area() argument
35 offset = pgoff << PAGE_SHIFT; in romfs_get_unmapped_area()
38 if ((pgoff >= maxpages) || (maxpages - pgoff < lpages)) in romfs_get_unmapped_area()
44 if (len > mtd->size || pgoff >= (mtd->size >> PAGE_SHIFT)) in romfs_get_unmapped_area()
/linux/drivers/dma-buf/
H A Dudmabuf.c51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault() local
55 if (pgoff >= ubuf->pagecount) in udmabuf_vm_fault()
58 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault()
59 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault()
66 pgoff = vma->vm_pgoff; in udmabuf_vm_fault()
69 for (; addr < vma->vm_end; pgoff++, addr += PAGE_SIZE) { in udmabuf_vm_fault()
73 if (WARN_ON(pgoff >= ubuf->pagecount)) in udmabuf_vm_fault()
76 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault()
77 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault()
340 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios() local
[all …]
/linux/arch/sh/kernel/
H A Dsys_sh.c43 unsigned long fd, unsigned long pgoff) in sys_mmap2() argument
49 if (pgoff & ((1 << (PAGE_SHIFT - 12)) - 1)) in sys_mmap2()
52 pgoff >>= PAGE_SHIFT - 12; in sys_mmap2()
54 return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); in sys_mmap2()
/linux/virt/kvm/
H A Dguest_memfd.c33 gfn_t gfn = slot->base_gfn + index - slot->gmem.pgoff; in __kvm_gmem_prepare_folio()
73 * the base pgoff of memslots is naturally aligned with the in kvm_gmem_prepare_folio()
80 WARN_ON(!IS_ALIGNED(slot->gmem.pgoff, 1 << folio_order(folio))); in kvm_gmem_prepare_folio()
81 index = gfn - slot->base_gfn + slot->gmem.pgoff; in kvm_gmem_prepare_folio()
114 pgoff_t pgoff = slot->gmem.pgoff; in kvm_gmem_invalidate_begin() local
117 .start = slot->base_gfn + max(pgoff, start) - pgoff, in kvm_gmem_invalidate_begin()
118 .end = slot->base_gfn + min(pgoff + slot->npages, end) - pgoff, in kvm_gmem_invalidate_begin()
[all...]
/linux/tools/perf/arch/x86/util/
H A Dmachine.c26 u64 end, u64 pgoff, const char *name) in add_extra_kernel_map() argument
42 mi->maps[mi->cnt].pgoff = pgoff; in add_extra_kernel_map()
93 xm->pgoff = mi.entry_trampoline; in machine__create_extra_kernel_maps()
/linux/fs/cramfs/
H A Dinode.c288 static u32 cramfs_get_block_range(struct inode *inode, u32 pgoff, u32 *pages) in cramfs_get_block_range() argument
299 blockptrs = (u32 *)(sbi->linear_virt_addr + OFFSET(inode) + pgoff * 4); in cramfs_get_block_range()
309 pgoff+i, pgoff + *pages - 1, in cramfs_get_block_range()
351 unsigned long address, pgoff = vma->vm_pgoff; in cramfs_physmem_mmap() local
371 if (pgoff >= max_pages) in cramfs_physmem_mmap()
373 pages = min(vma_pages(vma), max_pages - pgoff); in cramfs_physmem_mmap()
375 offset = cramfs_get_block_range(inode, pgoff, &pages); in cramfs_physmem_mmap()
385 if (pgoff + pages == max_pages && cramfs_last_page_is_shared(inode)) { in cramfs_physmem_mmap()
425 pgoff, address, pages, vma_pages(vma), vma->vm_start, in cramfs_physmem_mmap()
431 file, pgoff, bailout_reason); in cramfs_physmem_mmap()
[all …]
/linux/include/linux/sched/
H A Dmm.h185 unsigned long len, unsigned long pgoff,
189 unsigned long len, unsigned long pgoff,
194 unsigned long pgoff, unsigned long flags);
200 unsigned long pgoff,
206 unsigned long len, unsigned long pgoff,
210 unsigned long len, unsigned long pgoff,
/linux/drivers/video/fbdev/core/
H A Dfb_defio.c50 unsigned long pgoff = offset >> PAGE_SHIFT; in fb_deferred_io_pageref_lookup() local
53 if (fb_WARN_ON_ONCE(info, pgoff >= info->npagerefs)) in fb_deferred_io_pageref_lookup()
57 pageref = &info->pagerefs[pgoff]; in fb_deferred_io_pageref_lookup()
63 pageref->offset = pgoff << PAGE_SHIFT; in fb_deferred_io_pageref_lookup()
135 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault()
149 page->index = vmf->pgoff; /* for folio_mkclean() */ in fb_deferred_io_fault()
232 unsigned long offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_page_mkwrite()

1234567