| /linux/fs/btrfs/ |
| H A D | accessors.c | 59 char *kaddr = folio_address(eb->folios[idx]) + oif; \ 73 kaddr = folio_address(eb->folios[idx + 1]); \ 77 folio_address(eb->folios[idx + 1]), \ 89 char *kaddr = folio_address(eb->folios[idx]) + oif; \ 105 kaddr = folio_address(eb->folios[idx + 1]); \ 109 kaddr = folio_address(eb->folios[idx + 1]); \
|
| H A D | zlib.c | 183 cfolio_out = folio_address(out_folio); in zlib_compress_bio() 266 cfolio_out = folio_address(out_folio); in zlib_compress_bio() 304 cfolio_out = folio_address(out_folio); in zlib_compress_bio()
|
| H A D | zstd.c | 447 workspace->out_buf.dst = folio_address(out_folio); in zstd_compress_bio() 490 workspace->out_buf.dst = folio_address(out_folio); in zstd_compress_bio() 563 workspace->out_buf.dst = folio_address(out_folio); in zstd_compress_bio()
|
| H A D | extent_io.c | 3537 eb->addr = folio_address(eb->folios[0]) + offset_in_page(eb->start); in alloc_extent_buffer() 3996 kaddr = folio_address(eb->folios[i]); in read_extent_buffer() 4032 kaddr = folio_address(eb->folios[i]); in read_extent_buffer_to_user_nofault() 4068 kaddr = folio_address(eb->folios[i]); in memcmp_extent_buffer() 4147 kaddr = folio_address(eb->folios[i]); in __write_extent_buffer() 4183 memset(folio_address(eb->folios[index]) + offset, c, cur_len); in memset_extent_buffer() 4209 void *addr = folio_address(src->folios[index]) + offset; in copy_extent_buffer_full() 4242 kaddr = folio_address(dst->folios[i]); in copy_extent_buffer() 4300 kaddr = folio_address(eb->folios[i]); in extent_buffer_test_bit() 4310 return folio_address(eb->folios[index]) + get_eb_offset_in_folio(eb, bytenr); in extent_buffer_get_byte() [all …]
|
| /linux/lib/ |
| H A D | test_kho.c | 75 folios_info[i] = virt_to_phys(folio_address(folio)) | order; in kho_test_preserve_data() 101 void *fdt = folio_address(state->fdt); in kho_test_prepare_fdt() 146 err = kho_add_subtree(KHO_TEST_FDT, folio_address(state->fdt)); in kho_test_preserve() 188 addr = folio_address(folio); in kho_test_generate_data() 274 csum = csum_partial(folio_address(folio), size, csum); in kho_test_restore_data() 357 kho_remove_subtree(folio_address(kho_test_state.fdt)); in kho_test_exit()
|
| /linux/drivers/iommu/ |
| H A D | iommu-pages.c | 86 return folio_address(folio); in iommu_alloc_pages_node_sz() 200 folio_address(ioptdesc_folio(cur)), dma_dev); in iommu_pages_start_incoherent_list() 227 dma_unmap_single(dma_dev, virt_to_phys(folio_address(folio)), in iommu_pages_stop_incoherent_list()
|
| /linux/arch/arm64/mm/ |
| H A D | flush.c | 57 sync_icache_aliases((unsigned long)folio_address(folio), in __sync_icache_dcache() 58 (unsigned long)folio_address(folio) + in __sync_icache_dcache()
|
| /linux/fs/coda/ |
| H A D | symlink.c | 29 char *p = folio_address(folio); in coda_symlink_filler()
|
| /linux/fs/xfs/scrub/ |
| H A D | xfile.c | 150 memcpy(buf, folio_address(folio) + offset, len); in xfile_load() 211 memcpy(folio_address(folio) + offset, buf, len); in xfile_store()
|
| /linux/fs/jfs/ |
| H A D | jfs_metapage.c | 194 int page_offset = mp->data - folio_address(src); in __metapage_migrate_folio() 196 mp->data = folio_address(dst) + page_offset; in __metapage_migrate_folio() 251 page_offset = mp->data - folio_address(src); in __metapage_migrate_folio() 252 mp->data = folio_address(dst) + page_offset; in __metapage_migrate_folio() 770 mp->data = folio_address(folio) + page_offset; in __get_metapage()
|
| /linux/fs/efs/ |
| H A D | symlink.c | 17 char *link = folio_address(folio); in efs_symlink_read_folio()
|
| /linux/arch/nios2/mm/ |
| H A D | cacheflush.c | 170 unsigned long start = (unsigned long)folio_address(folio); in __flush_dcache_folio() 194 unsigned long start = (unsigned long)folio_address(folio); in flush_dcache_folio()
|
| /linux/fs/affs/ |
| H A D | symlink.c | 18 char *link = folio_address(folio); in affs_symlink_read_folio()
|
| /linux/arch/sh/mm/ |
| H A D | cache.c | 153 __flush_purge_region(folio_address(folio), in __update_cache() 173 __flush_purge_region(folio_address(folio), in __flush_anon_page()
|
| H A D | cache-sh7705.c | 187 __flush_purge_region(folio_address(folio), folio_size(folio)); in sh7705_flush_icache_folio()
|
| H A D | cache-sh4.c | 122 unsigned long addr = (unsigned long)folio_address(folio); in sh4_flush_dcache_folio()
|
| /linux/fs/ |
| H A D | aio.c | 566 ring = folio_address(ctx->ring_folios[0]); in aio_setup_ring() 687 ring = folio_address(ctx->ring_folios[0]); in ioctx_add_table() 1029 ring = folio_address(ctx->ring_folios[0]); in user_refill_reqs_available() 1141 ev_page = folio_address(ctx->ring_folios[pos / AIO_EVENTS_PER_PAGE]); in aio_complete() 1159 ring = folio_address(ctx->ring_folios[0]); in aio_complete() 1235 ring = folio_address(ctx->ring_folios[0]); in aio_read_events_ring() 1269 ev = folio_address(folio); in aio_read_events_ring() 1283 ring = folio_address(ctx->ring_folios[0]); in aio_read_events_ring()
|
| /linux/arch/powerpc/mm/ |
| H A D | cacheflush.c | 159 void *addr = folio_address(folio); in flush_dcache_icache_folio()
|
| /linux/arch/m68k/include/asm/ |
| H A D | cacheflush_mm.h | 260 __flush_pages_to_ram(folio_address(folio), folio_nr_pages(folio))
|
| /linux/fs/xfs/ |
| H A D | xfs_buf_mem.c | 168 bp->b_addr = folio_address(folio) + offset_in_folio(folio, pos); in xmbuf_map_backing_mem()
|
| /linux/include/linux/ |
| H A D | highmem-internal.h | 198 return folio_address(folio) + offset; in kmap_local_folio()
|
| /linux/fs/ramfs/ |
| H A D | file-nommu.c | 235 ret = (unsigned long) folio_address(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area()
|
| /linux/mm/ |
| H A D | secretmem.c | 99 addr = (unsigned long)folio_address(folio); in secretmem_fault()
|
| /linux/arch/sparc/kernel/ |
| H A D | smp_64.c | 927 __flush_dcache_page(folio_address(folio) + i * PAGE_SIZE, in __local_flush_dcache_folio() 956 void *pg_addr = folio_address(folio); in smp_flush_dcache_folio_impl() 999 pg_addr = folio_address(folio); in flush_dcache_folio_all()
|
| /linux/drivers/misc/lkdtm/ |
| H A D | usercopy.c | 417 addr = folio_address(folio); in lkdtm_USERCOPY_FOLIO()
|