| /linux/include/linux/ |
| H A D | pagemap.h | 22 pgoff_t start, pgoff_t end); 32 pgoff_t start, pgoff_t end); 508 static inline pgoff_t mapping_align_index(const struct address_space *mapping, in mapping_align_index() 509 pgoff_t index) in mapping_align_index() 690 pgoff_t page_cache_next_miss(struct address_space *mapping, 691 pgoff_t index, unsigned long max_scan); 692 pgoff_t page_cache_prev_miss(struct address_space *mapping, 693 pgoff_t index, unsigned long max_scan); 765 void *filemap_get_entry(struct address_space *mapping, pgoff_t index); 767 pgoff_t index, fgf_t fgf_flags, gfp_t gfp, struct mempolicy *policy); [all …]
|
| H A D | swapops.h | 84 static inline swp_entry_t swp_entry(unsigned long type, pgoff_t offset) in swp_entry() 105 static inline pgoff_t swp_offset(swp_entry_t entry) in swp_offset() 136 static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset) in make_readable_device_private_entry() 141 static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset) in make_writable_device_private_entry() 146 static inline swp_entry_t make_device_exclusive_entry(pgoff_t offset) in make_device_exclusive_entry() 152 static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset) in make_readable_device_private_entry() 157 static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset) in make_writable_device_private_entry() 162 static inline swp_entry_t make_device_exclusive_entry(pgoff_t offset) in make_device_exclusive_entry() 171 static inline swp_entry_t make_readable_migration_entry(pgoff_t offset) in make_readable_migration_entry() 176 static inline swp_entry_t make_readable_exclusive_migration_entry(pgoff_t offset) in make_readable_exclusive_migration_entry() [all …]
|
| H A D | shmem_fs.h | 52 pgoff_t fallocend; /* highest fallocate endindex */ 126 pgoff_t index, gfp_t gfp_mask); 134 struct vm_area_struct *vma, pgoff_t index, 139 struct vm_area_struct *vma, pgoff_t index, in shmem_allowable_huge_orders() 165 pgoff_t start, pgoff_t end); 176 int shmem_get_folio(struct inode *inode, pgoff_t index, loff_t write_end, 179 pgoff_t index, gfp_t gfp); 182 pgoff_t index) in shmem_read_folio() 188 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page() 219 static inline pgoff_t shmem_fallocend(struct inode *inode, pgoff_t eof) in shmem_fallocend()
|
| H A D | mempolicy.h | 118 pgoff_t start, end; 128 pgoff_t idx); 132 unsigned long addr, pgoff_t *ilx); 134 unsigned long addr, int order, pgoff_t *ilx); 223 mpol_shared_policy_lookup(struct shared_policy *sp, pgoff_t idx) in mpol_shared_policy_lookup() 229 unsigned long addr, int order, pgoff_t *ilx) in get_vma_policy()
|
| H A D | memfd.h | 11 struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx); 26 static inline struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx) in memfd_alloc_folio()
|
| /linux/include/trace/events/ |
| H A D | fs_dax.h | 12 pgoff_t max_pgoff, int result), 20 __field(pgoff_t, pgoff) 21 __field(pgoff_t, max_pgoff) 57 pgoff_t max_pgoff, int result), \ 112 __field(pgoff_t, pgoff) 150 TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index), 154 __field(pgoff_t, start_index) 155 __field(pgoff_t, end_index) 175 TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index),\ 182 TP_PROTO(struct inode *inode, pgoff_t pgoff, pgoff_t pglen), [all …]
|
| H A D | readahead.h | 15 TP_PROTO(struct inode *inode, pgoff_t index, unsigned long nr_to_read, 23 __field(pgoff_t, index) 44 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra), 51 __field(pgoff_t, index) 77 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra, 85 __field(pgoff_t, index) 118 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra, 124 TP_PROTO(struct inode *inode, pgoff_t index, struct file_ra_state *ra,
|
| H A D | filemap.h | 63 pgoff_t index, 64 pgoff_t last_index 99 pgoff_t index, 100 pgoff_t last_index 108 pgoff_t index, 109 pgoff_t last_index 115 TP_PROTO(struct address_space *mapping, pgoff_t index),
|
| /linux/mm/ |
| H A D | mapping_dirty_helpers.c | 67 pgoff_t bitmap_pgoff; 69 pgoff_t start; 70 pgoff_t end; 97 pgoff_t pgoff = ((addr - walk->vma->vm_start) >> PAGE_SHIFT) + in clean_record_pte() 264 pgoff_t first_index, pgoff_t nr) in wp_shared_mapping_range() 314 pgoff_t first_index, pgoff_t nr, in clean_record_shared_mapping_range() 315 pgoff_t bitmap_pgoff, in clean_record_shared_mapping_range() 317 pgoff_t *start, in clean_record_shared_mapping_range() 318 pgoff_t *end) in clean_record_shared_mapping_range()
|
| H A D | truncate.c | 61 struct folio_batch *fbatch, pgoff_t *indices) in truncate_folio_batch_exceptionals() 369 pgoff_t start; /* inclusive */ in truncate_inode_pages_range() 370 pgoff_t end; /* exclusive */ in truncate_inode_pages_range() 372 pgoff_t indices[PAGEVEC_SIZE]; in truncate_inode_pages_range() 373 pgoff_t index; in truncate_inode_pages_range() 535 pgoff_t start, pgoff_t end, unsigned long *nr_failed) in mapping_try_invalidate() 537 pgoff_t indices[PAGEVEC_SIZE]; in mapping_try_invalidate() 539 pgoff_t index = start; in mapping_try_invalidate() 600 pgoff_t start, pgoff_t end) in invalidate_mapping_pages() 673 pgoff_t start, pgoff_t end) in invalidate_inode_pages2_range() [all …]
|
| H A D | filemap.c | 484 pgoff_t max = end_byte >> PAGE_SHIFT; in filemap_range_has_page() 513 pgoff_t index = start_byte >> PAGE_SHIFT; in __filemap_fdatawait_range() 514 pgoff_t end = end_byte >> PAGE_SHIFT; in __filemap_fdatawait_range() 641 pgoff_t max = end_byte >> PAGE_SHIFT; in filemap_range_has_writeback() 815 pgoff_t offset = old->index; in replace_page_cache_folio() 849 struct folio *folio, pgoff_t index, gfp_t gfp, void **shadowp) in __filemap_add_folio() 950 pgoff_t index, gfp_t gfp) in filemap_add_folio() 1801 pgoff_t page_cache_next_miss(struct address_space *mapping, in page_cache_next_miss() 1802 pgoff_t index, unsigned long max_scan) in page_cache_next_miss() 1838 pgoff_t page_cache_prev_miss(struct address_space *mapping, in page_cache_prev_miss() [all …]
|
| H A D | readahead.c | 320 pgoff_t end_index; /* The last page we want to read */ in do_page_cache_ra() 442 static inline int ra_alloc_folio(struct readahead_control *ractl, pgoff_t index, in ra_alloc_folio() 443 pgoff_t mark, unsigned int order, gfp_t gfp) in ra_alloc_folio() 468 pgoff_t start = readahead_index(ractl); in page_cache_ra_order() 469 pgoff_t index = start; in page_cache_ra_order() 471 pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT; in page_cache_ra_order() 472 pgoff_t mark = index + ra->size - ra->async_size; in page_cache_ra_order() 557 pgoff_t index = readahead_index(ractl); in page_cache_sync_ra() 561 pgoff_t prev_index, miss; in page_cache_sync_ra() 635 pgoff_t index = readahead_index(ractl); in page_cache_async_ra() [all …]
|
| H A D | shmem.c | 107 pgoff_t start; /* start of range currently being fallocated */ 108 pgoff_t next; /* the next page offset to be fallocated */ 109 pgoff_t nr_falloced; /* how many new pages have been fallocated */ 110 pgoff_t nr_unswapped; /* how often writeout refused to swap out */ 160 static int shmem_swapin_folio(struct inode *inode, pgoff_t index, 498 pgoff_t index, void *expected, void *replacement) in shmem_replace_entry() 520 static int shmem_confirm_swap(struct address_space *mapping, pgoff_t index, in shmem_confirm_swap() 605 unsigned long within_size_orders, pgoff_t index, in shmem_get_orders_within_size() 608 pgoff_t aligned_index; in shmem_get_orders_within_size() 626 static unsigned int shmem_huge_global_enabled(struct inode *inode, pgoff_t index, in shmem_huge_global_enabled() [all …]
|
| H A D | swap.h | 84 struct swap_info_struct *si, pgoff_t offset) in __swap_offset_to_cluster() 265 struct mempolicy *mpol, pgoff_t ilx, bool *new_page_allocated, 268 struct mempolicy *mpol, pgoff_t ilx); 307 pgoff_t offset = swp_offset(entry); in non_swapcache_batch() 326 struct swap_info_struct *si, pgoff_t offset, bool irq) in swap_cluster_lock() 378 gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) in swap_cluster_readahead()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_page_dirty.c | 71 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_pagetable() 73 pgoff_t num_marked; in vmw_bo_dirty_scan_pagetable() 111 pgoff_t num_marked; in vmw_bo_dirty_scan_mkwrite() 127 pgoff_t start = 0; in vmw_bo_dirty_scan_mkwrite() 128 pgoff_t end = dirty->bitmap_size; in vmw_bo_dirty_scan_mkwrite() 171 pgoff_t start, pgoff_t end) in vmw_bo_dirty_pre_unmap() 196 pgoff_t start, pgoff_t end) in vmw_bo_dirty_unmap() 219 pgoff_t num_pages = PFN_UP(vbo->tbo.resource->size); in vmw_bo_dirty_add() 243 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_add() 295 pgoff_t start, cur, end; in vmw_bo_dirty_transfer_to_res() [all …]
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 52 static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index) in folio_file_pfn() 57 static pgoff_t kvm_gmem_get_index(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_gmem_get_index() 63 pgoff_t index, struct folio *folio) in __kvm_gmem_prepare_folio() 94 pgoff_t index; in kvm_gmem_prepare_folio() 133 static struct folio *kvm_gmem_get_folio(struct inode *inode, pgoff_t index) in kvm_gmem_get_folio() 165 static void __kvm_gmem_invalidate_begin(struct gmem_file *f, pgoff_t start, in __kvm_gmem_invalidate_begin() 166 pgoff_t end, in __kvm_gmem_invalidate_begin() 175 pgoff_t pgoff = slot->gmem.pgoff; in __kvm_gmem_invalidate_begin() 202 static void kvm_gmem_invalidate_begin(struct inode *inode, pgoff_t start, in kvm_gmem_invalidate_begin() 203 pgoff_t end) in kvm_gmem_invalidate_begin() [all …]
|
| /linux/drivers/dma-buf/ |
| H A D | udmabuf.c | 28 pgoff_t pagecount; 39 pgoff_t nr_pinned; 44 pgoff_t *offsets; 51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault() 114 pgoff_t pg; in vmap_udmabuf() 200 pgoff_t i; in unpin_all_folios() 208 static __always_inline int init_udmabuf(struct udmabuf *ubuf, pgoff_t pgcnt) in init_udmabuf() 328 pgoff_t nr_pinned = ubuf->nr_pinned; in udmabuf_pin_folios() 329 pgoff_t upgcnt = ubuf->pagecount; in udmabuf_pin_folios() 331 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios() [all …]
|
| /linux/include/drm/ttm/ |
| H A D | ttm_backup.h | 56 void ttm_backup_drop(struct file *backup, pgoff_t handle); 59 pgoff_t handle, bool intr); 63 bool writeback, pgoff_t idx, gfp_t page_gfp,
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_pt_walk.h | 63 typedef int (*xe_pt_entry_fn)(struct xe_ptw *parent, pgoff_t offset, 124 static inline pgoff_t 143 static inline pgoff_t
|
| H A D | xe_pt_walk.c | 36 static bool xe_pt_next(pgoff_t *offset, u64 *addr, u64 next, u64 end, in xe_pt_next() 39 pgoff_t step = 1; in xe_pt_next() 76 pgoff_t offset = xe_pt_offset(addr, level, walk); in xe_pt_walk_range()
|
| /linux/fs/nilfs2/ |
| H A D | page.c | 29 unsigned long block, pgoff_t index, int blkbits, in __nilfs_get_folio_block() 52 pgoff_t index = blkoff >> (PAGE_SHIFT - blkbits); in nilfs_grab_buffer() 248 pgoff_t index = 0; in nilfs_copy_dirty_pages() 253 if (!filemap_get_folios_tag(smap, &index, (pgoff_t)-1, in nilfs_copy_dirty_pages() 303 pgoff_t start = 0; in nilfs_copy_back_pages() 313 pgoff_t index = folio->index; in nilfs_copy_back_pages() 365 pgoff_t index = 0; in nilfs_clear_dirty_pages() 369 while (filemap_get_folios_tag(mapping, &index, (pgoff_t)-1, in nilfs_clear_dirty_pages() 511 pgoff_t index; in nilfs_find_uncommitted_extent()
|
| /linux/fs/erofs/ |
| H A D | dir.c | 54 pgoff_t ra_pages = DIV_ROUND_UP_POW2( in erofs_readdir() 56 pgoff_t nr_pages = DIV_ROUND_UP_POW2(dir->i_size, PAGE_SIZE); in erofs_readdir() 73 pgoff_t idx = DIV_ROUND_UP_POW2(ctx->pos, PAGE_SIZE); in erofs_readdir() 74 pgoff_t pages = min(nr_pages - idx, ra_pages); in erofs_readdir()
|
| /linux/drivers/md/ |
| H A D | dm-linear.c | 162 static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in linear_dax_pgoff() 171 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() 180 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() 188 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write()
|
| /linux/fs/f2fs/ |
| H A D | node.h | 204 static inline pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start) in current_nat_addr() 207 pgoff_t block_off; in current_nat_addr() 208 pgoff_t block_addr; in current_nat_addr() 217 block_addr = (pgoff_t)(nm_i->nat_blkaddr + in current_nat_addr() 227 static inline pgoff_t next_nat_addr(struct f2fs_sb_info *sbi, in next_nat_addr() 228 pgoff_t block_addr) in next_nat_addr()
|
| H A D | file.c | 421 struct dnode_of_data *dn, pgoff_t index, int whence) in __found_offset() 459 pgoff_t pgofs, end_offset; in f2fs_seek_block() 481 pgofs = (pgoff_t)(offset >> PAGE_SHIFT); in f2fs_seek_block() 715 pgoff_t fofs; in f2fs_truncate_data_blocks_range() 737 pgoff_t index = from >> PAGE_SHIFT; in truncate_partial_data_page() 773 pgoff_t free_from; in f2fs_do_truncate_blocks() 785 free_from = (pgoff_t)F2FS_BLK_ALIGN(from); in f2fs_do_truncate_blocks() 1210 static int fill_zero(struct inode *inode, pgoff_t index, in fill_zero() 1235 int f2fs_truncate_hole(struct inode *inode, pgoff_t pg_start, pgoff_t pg_end) in f2fs_truncate_hole() 1241 pgoff_t end_offset, count; in f2fs_truncate_hole() [all …]
|