Home
last modified time | relevance | path

Searched refs:pgoff_t (Results 1 – 25 of 193) sorted by relevance

12345678

/linux/include/linux/
H A Dpagemap.h22 pgoff_t start, pgoff_t end);
32 pgoff_t start, pgoff_t end);
508 static inline pgoff_t mapping_align_index(const struct address_space *mapping, in mapping_align_index()
509 pgoff_t index) in mapping_align_index()
690 pgoff_t page_cache_next_miss(struct address_space *mapping,
691 pgoff_t index, unsigned long max_scan);
692 pgoff_t page_cache_prev_miss(struct address_space *mapping,
693 pgoff_t index, unsigned long max_scan);
765 void *filemap_get_entry(struct address_space *mapping, pgoff_t index);
767 pgoff_t index, fgf_t fgf_flags, gfp_t gfp, struct mempolicy *policy);
[all …]
H A Dswapops.h84 static inline swp_entry_t swp_entry(unsigned long type, pgoff_t offset) in swp_entry()
105 static inline pgoff_t swp_offset(swp_entry_t entry) in swp_offset()
136 static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset) in make_readable_device_private_entry()
141 static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset) in make_writable_device_private_entry()
146 static inline swp_entry_t make_device_exclusive_entry(pgoff_t offset) in make_device_exclusive_entry()
152 static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset) in make_readable_device_private_entry()
157 static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset) in make_writable_device_private_entry()
162 static inline swp_entry_t make_device_exclusive_entry(pgoff_t offset) in make_device_exclusive_entry()
171 static inline swp_entry_t make_readable_migration_entry(pgoff_t offset) in make_readable_migration_entry()
176 static inline swp_entry_t make_readable_exclusive_migration_entry(pgoff_t offset) in make_readable_exclusive_migration_entry()
[all …]
H A Dshmem_fs.h52 pgoff_t fallocend; /* highest fallocate endindex */
126 pgoff_t index, gfp_t gfp_mask);
134 struct vm_area_struct *vma, pgoff_t index,
139 struct vm_area_struct *vma, pgoff_t index, in shmem_allowable_huge_orders()
165 pgoff_t start, pgoff_t end);
176 int shmem_get_folio(struct inode *inode, pgoff_t index, loff_t write_end,
179 pgoff_t index, gfp_t gfp);
182 pgoff_t index) in shmem_read_folio()
188 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page()
219 static inline pgoff_t shmem_fallocend(struct inode *inode, pgoff_t eof) in shmem_fallocend()
H A Dmempolicy.h118 pgoff_t start, end;
128 pgoff_t idx);
132 unsigned long addr, pgoff_t *ilx);
134 unsigned long addr, int order, pgoff_t *ilx);
223 mpol_shared_policy_lookup(struct shared_policy *sp, pgoff_t idx) in mpol_shared_policy_lookup()
229 unsigned long addr, int order, pgoff_t *ilx) in get_vma_policy()
H A Dmemfd.h11 struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx);
26 static inline struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx) in memfd_alloc_folio()
/linux/include/trace/events/
H A Dfs_dax.h12 pgoff_t max_pgoff, int result),
20 __field(pgoff_t, pgoff)
21 __field(pgoff_t, max_pgoff)
57 pgoff_t max_pgoff, int result), \
112 __field(pgoff_t, pgoff)
150 TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index),
154 __field(pgoff_t, start_index)
155 __field(pgoff_t, end_index)
175 TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index),\
182 TP_PROTO(struct inode *inode, pgoff_t pgoff, pgoff_t pglen),
[all …]
H A Dreadahead.h15 TP_PROTO(struct inode *inode, pgoff_t index, unsigned long nr_to_read,
23 __field(pgoff_t, index)
44 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra),
51 __field(pgoff_t, index)
77 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
85 __field(pgoff_t, index)
118 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
124 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
H A Dfilemap.h63 pgoff_t index,
64 pgoff_t last_index
99 pgoff_t index,
100 pgoff_t last_index
108 pgoff_t index,
109 pgoff_t last_index
115 TP_PROTO(struct address_space *mapping, pgoff_t index),
/linux/mm/
H A Dmapping_dirty_helpers.c67 pgoff_t bitmap_pgoff;
69 pgoff_t start;
70 pgoff_t end;
97 pgoff_t pgoff = ((addr - walk->vma->vm_start) >> PAGE_SHIFT) + in clean_record_pte()
264 pgoff_t first_index, pgoff_t nr) in wp_shared_mapping_range()
314 pgoff_t first_index, pgoff_t nr, in clean_record_shared_mapping_range()
315 pgoff_t bitmap_pgoff, in clean_record_shared_mapping_range()
317 pgoff_t *start, in clean_record_shared_mapping_range()
318 pgoff_t *end) in clean_record_shared_mapping_range()
H A Dtruncate.c61 struct folio_batch *fbatch, pgoff_t *indices) in truncate_folio_batch_exceptionals()
369 pgoff_t start; /* inclusive */ in truncate_inode_pages_range()
370 pgoff_t end; /* exclusive */ in truncate_inode_pages_range()
372 pgoff_t indices[PAGEVEC_SIZE]; in truncate_inode_pages_range()
373 pgoff_t index; in truncate_inode_pages_range()
535 pgoff_t start, pgoff_t end, unsigned long *nr_failed) in mapping_try_invalidate()
537 pgoff_t indices[PAGEVEC_SIZE]; in mapping_try_invalidate()
539 pgoff_t index = start; in mapping_try_invalidate()
600 pgoff_t start, pgoff_t end) in invalidate_mapping_pages()
673 pgoff_t start, pgoff_t end) in invalidate_inode_pages2_range()
[all …]
H A Dfilemap.c484 pgoff_t max = end_byte >> PAGE_SHIFT; in filemap_range_has_page()
513 pgoff_t index = start_byte >> PAGE_SHIFT; in __filemap_fdatawait_range()
514 pgoff_t end = end_byte >> PAGE_SHIFT; in __filemap_fdatawait_range()
641 pgoff_t max = end_byte >> PAGE_SHIFT; in filemap_range_has_writeback()
815 pgoff_t offset = old->index; in replace_page_cache_folio()
849 struct folio *folio, pgoff_t index, gfp_t gfp, void **shadowp) in __filemap_add_folio()
950 pgoff_t index, gfp_t gfp) in filemap_add_folio()
1801 pgoff_t page_cache_next_miss(struct address_space *mapping, in page_cache_next_miss()
1802 pgoff_t index, unsigned long max_scan) in page_cache_next_miss()
1838 pgoff_t page_cache_prev_miss(struct address_space *mapping, in page_cache_prev_miss()
[all …]
H A Dreadahead.c320 pgoff_t end_index; /* The last page we want to read */ in do_page_cache_ra()
442 static inline int ra_alloc_folio(struct readahead_control *ractl, pgoff_t index, in ra_alloc_folio()
443 pgoff_t mark, unsigned int order, gfp_t gfp) in ra_alloc_folio()
468 pgoff_t start = readahead_index(ractl); in page_cache_ra_order()
469 pgoff_t index = start; in page_cache_ra_order()
471 pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT; in page_cache_ra_order()
472 pgoff_t mark = index + ra->size - ra->async_size; in page_cache_ra_order()
557 pgoff_t index = readahead_index(ractl); in page_cache_sync_ra()
561 pgoff_t prev_index, miss; in page_cache_sync_ra()
635 pgoff_t index = readahead_index(ractl); in page_cache_async_ra()
[all …]
H A Dshmem.c107 pgoff_t start; /* start of range currently being fallocated */
108 pgoff_t next; /* the next page offset to be fallocated */
109 pgoff_t nr_falloced; /* how many new pages have been fallocated */
110 pgoff_t nr_unswapped; /* how often writeout refused to swap out */
160 static int shmem_swapin_folio(struct inode *inode, pgoff_t index,
498 pgoff_t index, void *expected, void *replacement) in shmem_replace_entry()
520 static int shmem_confirm_swap(struct address_space *mapping, pgoff_t index, in shmem_confirm_swap()
605 unsigned long within_size_orders, pgoff_t index, in shmem_get_orders_within_size()
608 pgoff_t aligned_index; in shmem_get_orders_within_size()
626 static unsigned int shmem_huge_global_enabled(struct inode *inode, pgoff_t index, in shmem_huge_global_enabled()
[all …]
H A Dswap.h84 struct swap_info_struct *si, pgoff_t offset) in __swap_offset_to_cluster()
265 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated,
268 struct mempolicy *mpol, pgoff_t ilx);
307 pgoff_t offset = swp_offset(entry); in non_swapcache_batch()
326 struct swap_info_struct *si, pgoff_t offset, bool irq) in swap_cluster_lock()
378 gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) in swap_cluster_readahead()
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_page_dirty.c71 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_pagetable()
73 pgoff_t num_marked; in vmw_bo_dirty_scan_pagetable()
111 pgoff_t num_marked; in vmw_bo_dirty_scan_mkwrite()
127 pgoff_t start = 0; in vmw_bo_dirty_scan_mkwrite()
128 pgoff_t end = dirty->bitmap_size; in vmw_bo_dirty_scan_mkwrite()
171 pgoff_t start, pgoff_t end) in vmw_bo_dirty_pre_unmap()
196 pgoff_t start, pgoff_t end) in vmw_bo_dirty_unmap()
219 pgoff_t num_pages = PFN_UP(vbo->tbo.resource->size); in vmw_bo_dirty_add()
243 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_add()
295 pgoff_t start, cur, end; in vmw_bo_dirty_transfer_to_res()
[all …]
/linux/virt/kvm/
H A Dguest_memfd.c52 static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index) in folio_file_pfn()
57 static pgoff_t kvm_gmem_get_index(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_gmem_get_index()
63 pgoff_t index, struct folio *folio) in __kvm_gmem_prepare_folio()
94 pgoff_t index; in kvm_gmem_prepare_folio()
133 static struct folio *kvm_gmem_get_folio(struct inode *inode, pgoff_t index) in kvm_gmem_get_folio()
165 static void __kvm_gmem_invalidate_begin(struct gmem_file *f, pgoff_t start, in __kvm_gmem_invalidate_begin()
166 pgoff_t end, in __kvm_gmem_invalidate_begin()
175 pgoff_t pgoff = slot->gmem.pgoff; in __kvm_gmem_invalidate_begin()
202 static void kvm_gmem_invalidate_begin(struct inode *inode, pgoff_t start, in kvm_gmem_invalidate_begin()
203 pgoff_t end) in kvm_gmem_invalidate_begin()
[all …]
/linux/drivers/dma-buf/
H A Dudmabuf.c28 pgoff_t pagecount;
39 pgoff_t nr_pinned;
44 pgoff_t *offsets;
51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault()
114 pgoff_t pg; in vmap_udmabuf()
200 pgoff_t i; in unpin_all_folios()
208 static __always_inline int init_udmabuf(struct udmabuf *ubuf, pgoff_t pgcnt) in init_udmabuf()
328 pgoff_t nr_pinned = ubuf->nr_pinned; in udmabuf_pin_folios()
329 pgoff_t upgcnt = ubuf->pagecount; in udmabuf_pin_folios()
331 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios()
[all …]
/linux/include/drm/ttm/
H A Dttm_backup.h56 void ttm_backup_drop(struct file *backup, pgoff_t handle);
59 pgoff_t handle, bool intr);
63 bool writeback, pgoff_t idx, gfp_t page_gfp,
/linux/drivers/gpu/drm/xe/
H A Dxe_pt_walk.h63 typedef int (*xe_pt_entry_fn)(struct xe_ptw *parent, pgoff_t offset,
124 static inline pgoff_t
143 static inline pgoff_t
H A Dxe_pt_walk.c36 static bool xe_pt_next(pgoff_t *offset, u64 *addr, u64 next, u64 end, in xe_pt_next()
39 pgoff_t step = 1; in xe_pt_next()
76 pgoff_t offset = xe_pt_offset(addr, level, walk); in xe_pt_walk_range()
/linux/fs/nilfs2/
H A Dpage.c29 unsigned long block, pgoff_t index, int blkbits, in __nilfs_get_folio_block()
52 pgoff_t index = blkoff >> (PAGE_SHIFT - blkbits); in nilfs_grab_buffer()
248 pgoff_t index = 0; in nilfs_copy_dirty_pages()
253 if (!filemap_get_folios_tag(smap, &index, (pgoff_t)-1, in nilfs_copy_dirty_pages()
303 pgoff_t start = 0; in nilfs_copy_back_pages()
313 pgoff_t index = folio->index; in nilfs_copy_back_pages()
365 pgoff_t index = 0; in nilfs_clear_dirty_pages()
369 while (filemap_get_folios_tag(mapping, &index, (pgoff_t)-1, in nilfs_clear_dirty_pages()
511 pgoff_t index; in nilfs_find_uncommitted_extent()
/linux/fs/erofs/
H A Ddir.c54 pgoff_t ra_pages = DIV_ROUND_UP_POW2( in erofs_readdir()
56 pgoff_t nr_pages = DIV_ROUND_UP_POW2(dir->i_size, PAGE_SIZE); in erofs_readdir()
73 pgoff_t idx = DIV_ROUND_UP_POW2(ctx->pos, PAGE_SIZE); in erofs_readdir()
74 pgoff_t pages = min(nr_pages - idx, ra_pages); in erofs_readdir()
/linux/drivers/md/
H A Ddm-linear.c162 static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in linear_dax_pgoff()
171 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access()
180 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range()
188 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write()
/linux/fs/f2fs/
H A Dnode.h204 static inline pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start) in current_nat_addr()
207 pgoff_t block_off; in current_nat_addr()
208 pgoff_t block_addr; in current_nat_addr()
217 block_addr = (pgoff_t)(nm_i->nat_blkaddr + in current_nat_addr()
227 static inline pgoff_t next_nat_addr(struct f2fs_sb_info *sbi, in next_nat_addr()
228 pgoff_t block_addr) in next_nat_addr()
H A Dfile.c421 struct dnode_of_data *dn, pgoff_t index, int whence) in __found_offset()
459 pgoff_t pgofs, end_offset; in f2fs_seek_block()
481 pgofs = (pgoff_t)(offset >> PAGE_SHIFT); in f2fs_seek_block()
715 pgoff_t fofs; in f2fs_truncate_data_blocks_range()
737 pgoff_t index = from >> PAGE_SHIFT; in truncate_partial_data_page()
773 pgoff_t free_from; in f2fs_do_truncate_blocks()
785 free_from = (pgoff_t)F2FS_BLK_ALIGN(from); in f2fs_do_truncate_blocks()
1210 static int fill_zero(struct inode *inode, pgoff_t index, in fill_zero()
1235 int f2fs_truncate_hole(struct inode *inode, pgoff_t pg_start, pgoff_t pg_end) in f2fs_truncate_hole()
1241 pgoff_t end_offset, count; in f2fs_truncate_hole()
[all …]

12345678