| /linux/mm/ |
| H A D | truncate.c | 367 loff_t lstart, uoff_t lend) in truncate_inode_pages_range() argument 387 start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT; in truncate_inode_pages_range() 412 same_folio = (lstart >> PAGE_SHIFT) == (lend >> PAGE_SHIFT); in truncate_inode_pages_range() 413 folio = __filemap_get_folio(mapping, lstart >> PAGE_SHIFT, FGP_LOCK, 0); in truncate_inode_pages_range() 416 if (!truncate_inode_partial_folio(folio, lstart, lend)) { in truncate_inode_pages_range() 430 if (!truncate_inode_partial_folio(folio, lstart, lend)) in truncate_inode_pages_range() 483 void truncate_inode_pages(struct address_space *mapping, loff_t lstart) in truncate_inode_pages() argument 485 truncate_inode_pages_range(mapping, lstart, (loff_t)-1); in truncate_inode_pages() 900 void truncate_pagecache_range(struct inode *inode, loff_t lstart, loff_t lend) in truncate_pagecache_range() argument 903 loff_t unmap_start = round_up(lstart, PAGE_SIZE); in truncate_pagecache_range() [all …]
|
| H A D | filemap.c | 676 loff_t lstart, loff_t lend) in filemap_write_and_wait_range() argument 680 if (lend < lstart) in filemap_write_and_wait_range() 684 err = filemap_fdatawrite_range(mapping, lstart, lend); in filemap_write_and_wait_range() 692 __filemap_fdatawait_range(mapping, lstart, lend); in filemap_write_and_wait_range() 777 int file_write_and_wait_range(struct file *file, loff_t lstart, loff_t lend) in file_write_and_wait_range() argument 782 if (lend < lstart) in file_write_and_wait_range() 786 err = filemap_fdatawrite_range(mapping, lstart, lend); in file_write_and_wait_range() 789 __filemap_fdatawait_range(mapping, lstart, lend); in file_write_and_wait_range()
|
| H A D | shmem.c | 1108 static void shmem_undo_range(struct inode *inode, loff_t lstart, uoff_t lend, in shmem_undo_range() argument 1113 pgoff_t start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT; in shmem_undo_range() 1162 same_folio = (lstart >> PAGE_SHIFT) == (lend >> PAGE_SHIFT); in shmem_undo_range() 1163 folio = shmem_get_partial_folio(inode, lstart >> PAGE_SHIFT); in shmem_undo_range() 1167 if (!truncate_inode_partial_folio(folio, lstart, lend)) { in shmem_undo_range() 1181 if (!truncate_inode_partial_folio(folio, lstart, lend)) in shmem_undo_range() 1250 } else if (truncate_inode_partial_folio(folio, lstart, lend)) { in shmem_undo_range() 1275 void shmem_truncate_range(struct inode *inode, loff_t lstart, uoff_t lend) in shmem_truncate_range() argument 1277 shmem_undo_range(inode, lstart, lend, false); in shmem_truncate_range() 5800 void shmem_truncate_range(struct inode *inode, loff_t lstart, uoff_t lend) in shmem_truncate_range() argument [all …]
|
| H A D | khugepaged.c | 2814 loff_t lstart = (loff_t)pgoff << PAGE_SHIFT; in madvise_collapse() local 2815 loff_t lend = lstart + HPAGE_PMD_SIZE - 1; in madvise_collapse() 2817 filemap_write_and_wait_range(file->f_mapping, lstart, lend); in madvise_collapse()
|
| /linux/fs/erofs/ |
| H A D | zmap.c | 518 erofs_off_t l, r, mid, pa, la, lstart; in z_erofs_map_blocks_ext() local 531 lstart = 0; in z_erofs_map_blocks_ext() 533 lstart = round_down(map->m_la, 1 << vi->z_lclusterbits); in z_erofs_map_blocks_ext() 534 pos += (lstart >> vi->z_lclusterbits) * recsz; in z_erofs_map_blocks_ext() 538 for (; lstart <= map->m_la; lstart += 1 << vi->z_lclusterbits) { in z_erofs_map_blocks_ext() 551 last = (lstart >= round_up(lend, 1 << vi->z_lclusterbits)); in z_erofs_map_blocks_ext() 552 lend = min(lstart, lend); in z_erofs_map_blocks_ext() 553 lstart -= 1 << vi->z_lclusterbits; in z_erofs_map_blocks_ext() 555 lstart = lend; in z_erofs_map_blocks_ext() 580 lstart = la; in z_erofs_map_blocks_ext() [all …]
|
| /linux/drivers/rapidio/devices/ |
| H A D | tsi721.c | 1110 static int tsi721_rio_map_inb_mem(struct rio_mport *mport, dma_addr_t lstart, in tsi721_rio_map_inb_mem() argument 1117 bool direct = (lstart == rstart); in tsi721_rio_map_inb_mem() 1132 ibw_start = lstart & ~(ibw_size - 1); in tsi721_rio_map_inb_mem() 1136 rstart, &lstart, size, ibw_start); in tsi721_rio_map_inb_mem() 1138 while ((lstart + size) > (ibw_start + ibw_size)) { in tsi721_rio_map_inb_mem() 1140 ibw_start = lstart & ~(ibw_size - 1); in tsi721_rio_map_inb_mem() 1155 rstart, &lstart, size); in tsi721_rio_map_inb_mem() 1158 ((u64)lstart & (size - 1)) || (rstart & (size - 1))) in tsi721_rio_map_inb_mem() 1164 loc_start = lstart; in tsi721_rio_map_inb_mem() 1196 map->lstart = lstart; in tsi721_rio_map_inb_mem() [all …]
|
| H A D | tsi721.h | 840 dma_addr_t lstart; member 846 dma_addr_t lstart; member
|
| /linux/arch/powerpc/sysdev/ |
| H A D | fsl_rio.c | 307 static int fsl_map_inb_mem(struct rio_mport *mport, dma_addr_t lstart, in fsl_map_inb_mem() argument 324 if (lstart & (base_size - 1)) in fsl_map_inb_mem() 350 out_be32(&priv->inb_atmu_regs[i].riwtar, lstart >> RIWTAR_TRAD_VAL_SHIFT); in fsl_map_inb_mem() 358 static void fsl_unmap_inb_mem(struct rio_mport *mport, dma_addr_t lstart) in fsl_unmap_inb_mem() argument 366 base_start_shift = lstart >> RIWTAR_TRAD_VAL_SHIFT; in fsl_unmap_inb_mem()
|
| /linux/fs/udf/ |
| H A D | udf_i.h | 15 loff_t lstart; member
|
| H A D | inode.c | 74 if (iinfo->cached_extent.lstart != -1) { in __udf_clear_extent_cache() 76 iinfo->cached_extent.lstart = -1; in __udf_clear_extent_cache() 98 if ((iinfo->cached_extent.lstart <= bcount) && in udf_read_extent_cache() 99 (iinfo->cached_extent.lstart != -1)) { in udf_read_extent_cache() 101 *lbcount = iinfo->cached_extent.lstart; in udf_read_extent_cache() 124 iinfo->cached_extent.lstart = estart; in udf_update_extent_cache()
|
| H A D | super.c | 166 ei->cached_extent.lstart = -1; in udf_alloc_inode()
|
| /linux/lib/zstd/compress/ |
| H A D | zstd_compress_superblock.c | 494 const BYTE* const lstart = seqStorePtr->litStart; in ZSTD_compressSubBlock_multi() local 496 const BYTE* lp = lstart; in ZSTD_compressSubBlock_multi() 497 size_t const nbLiterals = (size_t)(lend - lstart); in ZSTD_compressSubBlock_multi() 512 (unsigned)srcSize, (unsigned)(lend-lstart), (unsigned)(send-sstart)); in ZSTD_compressSubBlock_multi()
|
| /linux/include/linux/ |
| H A D | rio.h | 420 int (*map_inb)(struct rio_mport *mport, dma_addr_t lstart, 422 void (*unmap_inb)(struct rio_mport *mport, dma_addr_t lstart);
|
| H A D | pagemap.h | 43 int filemap_fdatawait_range(struct address_space *, loff_t lstart, loff_t lend); 54 bool filemap_range_has_page(struct address_space *, loff_t lstart, loff_t lend); 56 loff_t lstart, loff_t lend);
|
| H A D | rio_drv.h | 367 extern void rio_unmap_inb_region(struct rio_mport *mport, dma_addr_t lstart);
|
| H A D | mm.h | 3926 void truncate_inode_pages(struct address_space *mapping, loff_t lstart); 3927 void truncate_inode_pages_range(struct address_space *mapping, loff_t lstart,
|
| H A D | fs.h | 2606 extern int __must_check file_fdatawait_range(struct file *file, loff_t lstart,
|
| /linux/fs/jfs/ |
| H A D | jfs_xtree.h | 100 extern int xtLookup(struct inode *ip, s64 lstart, s64 llen,
|
| /linux/fs/hugetlbfs/ |
| H A D | inode.c | 577 static void remove_inode_hugepages(struct inode *inode, loff_t lstart, in remove_inode_hugepages() argument 589 next = lstart >> PAGE_SHIFT; in remove_inode_hugepages() 614 lstart >> huge_page_shift(h), in remove_inode_hugepages()
|
| /linux/arch/arm64/mm/ |
| H A D | mmu.c | 913 unsigned long lstart = _PAGE_OFFSET(vabits_actual); in linear_map_split_to_ptes() local 930 ret = range_split_to_ptes(lstart, kstart, GFP_ATOMIC); in linear_map_split_to_ptes() 935 flush_tlb_kernel_range(lstart, lend); in linear_map_split_to_ptes()
|
| /linux/drivers/rapidio/ |
| H A D | rio.c | 709 void rio_unmap_inb_region(struct rio_mport *mport, dma_addr_t lstart) in rio_unmap_inb_region() argument 715 mport->ops->unmap_inb(mport, lstart); in rio_unmap_inb_region()
|
| /linux/fs/nfs/ |
| H A D | write.c | 1966 loff_t lstart, loff_t lend) in nfs_filemap_write_and_wait_range() argument 1970 ret = filemap_write_and_wait_range(mapping, lstart, lend); in nfs_filemap_write_and_wait_range()
|