Searched refs:swp_offset (Results 1 – 17 of 17) sorted by relevance
| /linux/include/linux/ |
| H A D | swapops.h | 105 static inline pgoff_t swp_offset(swp_entry_t entry) in swp_offset() function 118 arch_entry = __swp_entry(swp_type(entry), swp_offset(entry)); in swp_entry_to_pte() 204 swp_offset(entry) | SWP_MIG_YOUNG); in make_migration_entry_young() 212 swp_offset(entry) | SWP_MIG_DIRTY); in make_migration_entry_dirty() 337 arch_entry = __swp_entry(swp_type(entry), swp_offset(entry)); in swp_entry_to_pmd()
|
| H A D | leafops.h | 321 return swp_offset(entry) & PTE_MARKER_MASK; in softleaf_to_marker() 363 return swp_offset(entry) & SWP_PFN_MASK; in softleaf_to_pfn() 472 return swp_offset(entry) & SWP_MIG_YOUNG; in softleaf_is_migration_young() 491 return swp_offset(entry) & SWP_MIG_DIRTY; in softleaf_is_migration_dirty()
|
| /linux/mm/ |
| H A D | swap.h | 57 #include <linux/swapops.h> /* for swp_offset */ 62 return swp_offset(entry) % SWAPFILE_CLUSTER; in swp_cluster_offset() 94 swp_offset(entry)); in __swap_entry_to_cluster() 140 swp_offset(folio->swap), irq); in __swap_cluster_get_and_lock() 239 return ((loff_t)swp_offset(entry)) << PAGE_SHIFT; 318 unsigned long start = swp_offset(entry); in non_swapcache_batch()
|
| H A D | zswap.c | 233 return &zswap_trees[swp_type(swp)][swp_offset(swp) in swap_zswap_tree() 967 swp_offset(entry->swpentry), in zswap_decompress() 992 pgoff_t offset = swp_offset(swpentry); in zswap_writeback_entry() 1420 swp_offset(page_swpentry), in zswap_store_page() 1553 pgoff_t offset = swp_offset(swp); in zswap_store() 1591 pgoff_t offset = swp_offset(swp); in zswap_load() 1650 pgoff_t offset = swp_offset(swp); in zswap_invalidate()
|
| H A D | swapfile.c | 237 offset = swp_offset(folio->swap); in __try_to_reclaim_swap() 342 offset = swp_offset(folio->swap); in swap_folio_sector() 1598 swap_put_entries_cluster(si, swp_offset(entry), nr_pages, false); in folio_put_swap() 1673 offset = swp_offset(entry); in get_swap_device() 1724 pgoff_t offset = swp_offset(entry); in __swap_count() 1736 pgoff_t offset = swp_offset(entry); in swap_entry_swapped() 1764 offset = swp_offset(entry); in swp_swapcount() 1800 unsigned long roffset = swp_offset(entry); in swap_page_trans_huge_swapped() 1903 const unsigned long start_offset = swp_offset(entry); in swap_put_entries_direct() 1960 pgoff_t offset = swp_offset(entry); in swap_free_hibernation_slot() [all …]
|
| H A D | swap_state.c | 193 offset = swp_offset(entry); in __swap_cache_del_folio() 194 ci = swap_cluster_lock(si, swp_offset(entry)); in __swap_cache_del_folio() 262 swp_offset(entry) + ci_off - ci_start))) in __swap_cache_replace_folio() 274 swap_entries_free(si, ci, swp_offset(entry), nr_pages); in __swap_cache_replace_folio() 278 swap_entries_free(si, ci, swp_offset(entry), 1); in __swap_cache_replace_folio() 299 ci = swap_cluster_lock(__swap_entry_to_info(entry), swp_offset(entry)); in __swap_cache_clear_shadow() 600 pgoff_t offset = swp_offset(entry); 725 unsigned long entry_offset = swp_offset(entry); in swap_vma_readahead()
|
| H A D | mprotect.c | 332 swp_offset(entry)); in change_pte_range() 334 entry = make_readable_migration_entry(swp_offset(entry)); in change_pte_range() 344 swp_offset(entry)); in change_pte_range()
|
| H A D | page_io.c | 214 set_bit(swp_offset(entry), sis->zeromap); in swap_zeromap_folio_set() 232 clear_bit(swp_offset(entry), sis->zeromap); in swap_zeromap_folio_clear()
|
| H A D | debug_vm_pgtable.c | 874 entry = make_readable_migration_entry(swp_offset(entry)); in swap_migration_tests() 1229 max_swap_offset = swp_offset(softleaf_from_pte(softleaf_to_pte(swp_entry(0, ~0UL)))); in init_args()
|
| H A D | huge_memory.c | 1806 entry = make_readable_migration_entry(swp_offset(entry)); in copy_huge_non_present_pmd() 1819 entry = make_readable_device_private_entry(swp_offset(entry)); in copy_huge_non_present_pmd() 2530 entry = make_readable_exclusive_migration_entry(swp_offset(entry)); in change_non_present_huge_pmd() 2532 entry = make_readable_migration_entry(swp_offset(entry)); in change_non_present_huge_pmd() 2537 entry = make_readable_device_private_entry(swp_offset(entry)); in change_non_present_huge_pmd()
|
| H A D | memory.c | 980 swp_offset(entry)); in copy_nonpresent_pte() 1016 swp_offset(entry)); in copy_nonpresent_pte() 4578 static inline unsigned long thp_swap_suitable_orders(pgoff_t swp_offset, in thp_swap_suitable_orders() argument 4593 if ((addr >> PAGE_SHIFT) % nr == swp_offset % nr) in thp_swap_suitable_orders() 4636 orders = thp_swap_suitable_orders(swp_offset(entry), in alloc_swap_folio() 4829 page = folio_file_page(folio, swp_offset(entry)); in do_swap_page()
|
| H A D | internal.h | 409 (swp_offset(entry) + delta))); in pte_move_swp_offset()
|
| H A D | hugetlb.c | 4964 swp_offset(softleaf)); in copy_hugetlb_page_range() 6500 swp_offset(entry)); in hugetlb_change_protection() 6503 swp_offset(entry)); in hugetlb_change_protection()
|
| H A D | userfaultfd.c | |
| H A D | shmem.c | 2265 swp_offset(swap) + swap_offset + in shmem_split_large_entry() 2332 swap = swp_entry(swp_type(swap), swp_offset(swap) + offset); in shmem_swapin_folio()
|
| /linux/kernel/power/ |
| H A D | swap.c | 177 offset = swp_offset(swap_alloc_hibernation_slot(swap)); in alloc_swapdev_block()
|
| /linux/fs/proc/ |
| H A D | task_mmu.c | 1971 offset = swp_offset(entry); in pte_to_pagemap_entry() 2036 offset = swp_offset(entry) + idx; in pagemap_pmd_range_thp()
|