| /linux/include/linux/ |
| H A D | page_ref.h | 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *page, int v, int u); 34 extern void __page_ref_freeze(struct page *page, int v, int ret); 35 extern void __page_ref_unfreeze(struct page *page, int v); 41 static inline void __page_ref_set(struct page *page, int v) in __page_ref_set() argument 44 static inline void __page_ref_mod(struct page *page, int v) in __page_ref_mod() argument 47 static inline void __page_ref_mod_and_test(struct page *page, int v, int ret) in __page_ref_mod_and_test() argument [all …]
|
| H A D | page-flags.h | 208 static __always_inline const struct page *page_fixed_fake_head(const struct page *page) in page_fixed_fake_head() argument 211 return page; in page_fixed_fake_head() 219 if (IS_ALIGNED((unsigned long)page, PAGE_SIZE) && in page_fixed_fake_head() 220 test_bit(PG_head, &page->flags.f)) { in page_fixed_fake_head() 226 unsigned long head = READ_ONCE(page[1].compound_head); in page_fixed_fake_head() 229 return (const struct page *)(head - 1); in page_fixed_fake_head() 231 return page; in page_fixed_fake_head() 234 static __always_inline bool page_count_writable(const struct page *page, int u) in page_count_writable() argument 260 if (atomic_read_acquire(&page->_refcount) == u) in page_count_writable() 263 return page_fixed_fake_head(page) == page; in page_count_writable() [all …]
|
| H A D | highmem-internal.h | 10 void *__kmap_local_page_prot(const struct page *page, pgprot_t prot); 35 void *kmap_high(struct page *page); 36 void kunmap_high(const struct page *page); 38 struct page *__kmap_to_page(void *addr); 40 static inline void *kmap(struct page *page) in kmap() argument 45 if (!PageHighMem(page)) in kmap() 46 addr = page_address(page); in kmap() 48 addr = kmap_high(page); in kmap() 53 static inline void kunmap(const struct page *page) in kunmap() argument 56 if (!PageHighMem(page)) in kunmap() [all …]
|
| H A D | pageblock-flags.h | 82 struct page; 84 enum migratetype get_pfnblock_migratetype(const struct page *page, 86 bool get_pfnblock_bit(const struct page *page, unsigned long pfn, 88 void set_pfnblock_bit(const struct page *page, unsigned long pfn, 90 void clear_pfnblock_bit(const struct page *page, unsigned long pfn, 95 #define get_pageblock_skip(page) \ argument 96 get_pfnblock_bit(page, page_to_pfn(page), PB_compact_skip) 97 #define clear_pageblock_skip(page) \ argument 98 clear_pfnblock_bit(page, page_to_pfn(page), PB_compact_skip) 99 #define set_pageblock_skip(page) \ argument [all …]
|
| H A D | mm.h | 153 static inline void __mm_zero_struct_page(struct page *page) in __mm_zero_struct_page() argument 155 unsigned long *_pp = (void *)page; in __mm_zero_struct_page() 158 BUILD_BUG_ON(sizeof(struct page) & 7); in __mm_zero_struct_page() 159 BUILD_BUG_ON(sizeof(struct page) < 56); in __mm_zero_struct_page() 160 BUILD_BUG_ON(sizeof(struct page) > 96); in __mm_zero_struct_page() 162 switch (sizeof(struct page)) { in __mm_zero_struct_page() 189 #define mm_zero_struct_page(pp) ((void)memset((pp), 0, sizeof(struct page))) 217 bool page_range_contiguous(const struct page *page, unsigned long nr_pages); 219 static inline bool page_range_contiguous(const struct page *page, in page_range_contiguous() argument 244 const struct page *page) in folio_page_idx() argument [all …]
|
| H A D | pgalloc_tag.h | 25 struct page *page; /* reference in page flags */ member 101 static inline bool get_page_tag_ref(struct page *page, union codetag_ref *ref, in get_page_tag_ref() argument 104 if (!page) in get_page_tag_ref() 110 idx = (page->flags.f >> alloc_tag_ref_offs) & in get_page_tag_ref() 113 handle->page = page; in get_page_tag_ref() 118 page_ext = page_ext_get(page); in get_page_tag_ref() 142 struct page *page = handle.page; in update_page_tag_ref() local 147 if (WARN_ON(!page || !ref)) in update_page_tag_ref() 153 old_flags = READ_ONCE(page->flags.f); in update_page_tag_ref() 157 } while (unlikely(!try_cmpxchg(&page->flags.f, &old_flags, flags))); in update_page_tag_ref() [all …]
|
| /linux/mm/ |
| H A D | page_alloc.c | 244 static void __free_pages_ok(struct page *page, unsigned int order, 319 static bool page_contains_unaccepted(struct page *page, unsigned int order); 322 static bool __free_unaccepted(struct page *page); 363 static inline unsigned long *get_pageblock_bitmap(const struct page *page, in get_pageblock_bitmap() argument 369 return page_zone(page)->pageblock_flags; in get_pageblock_bitmap() 373 static inline int pfn_to_bitidx(const struct page *page, unsigned long pfn) in pfn_to_bitidx() argument 378 pfn = pfn - pageblock_start_pfn(page_zone(page)->zone_start_pfn); in pfn_to_bitidx() 389 get_pfnblock_bitmap_bitidx(const struct page *page, unsigned long pfn, in get_pfnblock_bitmap_bitidx() argument 401 VM_BUG_ON_PAGE(!zone_spans_pfn(page_zone(page), pfn), page); in get_pfnblock_bitmap_bitidx() 403 bitmap = get_pageblock_bitmap(page, pfn); in get_pfnblock_bitmap_bitidx() [all …]
|
| H A D | balloon_compaction.c | |
| H A D | folio-compat.c | 14 void unlock_page(struct page *page) in unlock_page() 16 return folio_unlock(page_folio(page)); in unlock_page() 20 void end_page_writeback(struct page *page) in end_page_writeback() 22 return folio_end_writeback(page_folio(page)); in end_page_writeback() 26 void wait_on_page_writeback(struct page *page) in wait_on_page_writeback() 28 return folio_wait_writeback(page_folio(page)); in wait_on_page_writeback() 32 void mark_page_accessed(struct page *pag in mark_page_accessed() 13 unlock_page(struct page * page) unlock_page() argument 19 end_page_writeback(struct page * page) end_page_writeback() argument 25 wait_on_page_writeback(struct page * page) wait_on_page_writeback() argument 31 mark_page_accessed(struct page * page) mark_page_accessed() argument 37 set_page_writeback(struct page * page) set_page_writeback() argument 43 set_page_dirty(struct page * page) set_page_dirty() argument 49 set_page_dirty_lock(struct page * page) set_page_dirty_lock() argument 55 clear_page_dirty_for_io(struct page * page) clear_page_dirty_for_io() argument 62 redirty_page_for_writepage(struct writeback_control * wbc,struct page * page) redirty_page_for_writepage() argument 68 add_to_page_cache_lru(struct page * page,struct address_space * mapping,pgoff_t index,gfp_t gfp) add_to_page_cache_lru() argument [all...] |
| H A D | debug_page_ref.c | 8 void __page_ref_set(struct page *page, int v) in __page_ref_set() argument 10 trace_page_ref_set(page, v); in __page_ref_set() 15 void __page_ref_mod(struct page *page, int v) in __page_ref_mod() argument 17 trace_page_ref_mod(page, v); in __page_ref_mod() 22 void __page_ref_mod_and_test(struct page *page, int v, int ret) in __page_ref_mod_and_test() argument 24 trace_page_ref_mod_and_test(page, v, ret); in __page_ref_mod_and_test() 29 void __page_ref_mod_and_return(struct page *page, int v, int ret) in __page_ref_mod_and_return() argument 31 trace_page_ref_mod_and_return(page, v, ret); in __page_ref_mod_and_return() 36 void __page_ref_mod_unless(struct page *page, int v, int u) in __page_ref_mod_unless() argument 38 trace_page_ref_mod_unless(page, v, u); in __page_ref_mod_unless() [all …]
|
| H A D | highmem.c | 13 * Rewrote high memory support to move the page cache into 62 * Determine color of virtual address where the page should be mapped. 64 static inline unsigned int get_pkmap_color(const struct page *page) in get_pkmap_color() argument 71 * Get next index for mapping inside PKMAP region for page with given color. 82 * Determine if page index inside PKMAP region (pkmap_nr) of given color 165 struct page *__kmap_to_page(void *vaddr) in __kmap_to_page() 205 struct page *page; in flush_all_zero_pkmaps() 224 * no-one has the page mappe in flush_all_zero_pkmaps() 204 struct page *page; flush_all_zero_pkmaps() local 245 map_new_virtual(struct page * page) map_new_virtual() argument 307 kmap_high(struct page * page) kmap_high() argument 337 kmap_high_get(const struct page * page) kmap_high_get() argument 359 kunmap_high(const struct page * page) kunmap_high() argument 403 zero_user_segments(struct page * page,unsigned start1,unsigned end1,unsigned start2,unsigned end2) zero_user_segments() argument 511 arch_kmap_local_high_get(const struct page * page) arch_kmap_local_high_get() argument 575 __kmap_local_page_prot(const struct page * page,pgprot_t prot) __kmap_local_page_prot() argument 727 struct page *page; global() member 742 page_slot(const struct page * page) page_slot() argument 753 page_address(const struct page * page) page_address() argument 786 set_page_address(struct page * page,void * virtual) set_page_address() argument [all...] |
| /linux/include/trace/events/ |
| H A D | page_ref.h | 15 TP_PROTO(struct page *page, int v), 17 TP_ARGS(page, v), 30 __entry->pfn = page_to_pfn(page); 31 __entry->flags = page->flags.f; 32 __entry->count = page_ref_count(page); 33 __entry->mapcount = atomic_read(&page->_mapcount); 34 __entry->mapping = page->mapping; 35 __entry->mt = get_pageblock_migratetype(page); 49 TP_PROTO(struct page *page, int v), 51 TP_ARGS(page, v) [all …]
|
| H A D | cma.h | 13 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 16 TP_ARGS(name, pfn, page, count), 21 __field(const struct page *, page) 28 __entry->page = page; 35 __entry->page, 72 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 75 TP_ARGS(name, pfn, page, count, align, errorno), 80 __field(const struct page *, page) 89 __entry->page = page; 98 __entry->page, [all …]
|
| /linux/Documentation/trace/ |
| H A D | ring-buffer-design.rst | 40 - A page outside the ring buffer used solely (for the most part) 44 - a pointer to the page that the reader will use next 47 - a pointer to the page that will be written to next 50 - a pointer to the page with the last finished non-nested write. 110 At initialization a reader page is allocated for the reader that is not 114 to the same page. 116 The reader page is initialized to have its next pointer pointing to 117 the head page, and its previous pointer pointing to a page before 118 the head page. 120 The reader has its own page to use. At start up time, this page is [all …]
|
| /linux/mm/kmsan/ |
| H A D | shadow.c | 22 #define shadow_page_for(page) ((page)->kmsan_shadow) argument 24 #define origin_page_for(page) ((page)->kmsan_origin) argument 26 static void *shadow_ptr_for(struct page *page) in shadow_ptr_for() argument 28 return page_address(shadow_page_for(page)); in shadow_ptr_for() 31 static void *origin_ptr_for(struct page *page) in origin_ptr_for() argument 33 return page_address(origin_page_for(page)); in origin_ptr_for() 36 static bool page_has_metadata(struct page *page) in page_has_metadata() argument 38 return shadow_page_for(page) && origin_page_for(page); in page_has_metadata() 41 static void set_no_shadow_origin_page(struct page *page) in set_no_shadow_origin_page() argument 43 shadow_page_for(page) = NULL; in set_no_shadow_origin_page() [all …]
|
| /linux/arch/openrisc/include/asm/ |
| H A D | cacheflush.h | 24 extern void local_dcache_page_flush(struct page *page); 25 extern void local_icache_page_inv(struct page *page); 36 #define dcache_page_flush(page) local_dcache_page_flush(page) argument 37 #define icache_page_inv(page) local_icache_page_inv(page) argument 39 #define dcache_page_flush(page) local_dcache_page_flush(page) argument 40 #define icache_page_inv(page) smp_icache_page_inv(page) argument 41 extern void smp_icache_page_inv(struct page *page); 62 static inline void sync_icache_dcache(struct page *page) in sync_icache_dcache() argument 65 dcache_page_flush(page); in sync_icache_dcache() 66 icache_page_inv(page); in sync_icache_dcache() [all …]
|
| /linux/drivers/target/iscsi/ |
| H A D | iscsi_target_stat.c | 51 char *page) in iscsi_stat_instance_inst_show() argument 53 return snprintf(page, PAGE_SIZE, "%u\n", in iscsi_stat_instance_inst_show() 58 char *page) in iscsi_stat_instance_min_ver_show() argument 60 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); in iscsi_stat_instance_min_ver_show() 64 char *page) in iscsi_stat_instance_max_ver_show() argument 66 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_DRAFT20_VERSION); in iscsi_stat_instance_max_ver_show() 70 char *page) in iscsi_stat_instance_portals_show() argument 72 return snprintf(page, PAGE_SIZE, "%u\n", in iscsi_stat_instance_portals_show() 77 char *page) in iscsi_stat_instance_nodes_show() argument 79 return snprintf(page, PAGE_SIZE, "%u\n", ISCSI_INST_NUM_NODES); in iscsi_stat_instance_nodes_show() [all …]
|
| /linux/drivers/xen/ |
| H A D | balloon.c | 147 static void balloon_append(struct page *page) in balloon_append() argument 149 if (!PageOffline(page)) in balloon_append() 150 __SetPageOffline(page); in balloon_append() 153 if (PageHighMem(page)) { in balloon_append() 154 list_add_tail(&page->lru, &ballooned_pages); in balloon_append() 157 list_add(&page->lru, &ballooned_pages); in balloon_append() 160 inc_node_page_state(page, NR_BALLOON_PAGES); in balloon_append() 166 static struct page *balloon_retrieve(bool require_lowmem) in balloon_retrieve() 168 struct page *page; in balloon_retrieve() local 173 page = list_entry(ballooned_pages.next, struct page, lru); in balloon_retrieve() [all …]
|
| /linux/fs/ceph/ |
| H A D | crypto.h | 147 struct page *page, unsigned int len, 150 struct page *page, unsigned int len, 152 int ceph_fscrypt_decrypt_pages(struct inode *inode, struct page **page, 154 int ceph_fscrypt_decrypt_extents(struct inode *inode, struct page **page, 157 int ceph_fscrypt_encrypt_pages(struct inode *inode, struct page **page, u64 off, 160 static inline struct page *ceph_fscrypt_pagecache_page(struct page *page) in ceph_fscrypt_pagecache_page() argument 162 return fscrypt_is_bounce_page(page) ? fscrypt_pagecache_page(page) : page; in ceph_fscrypt_pagecache_page() 226 struct page *page, unsigned int len, in ceph_fscrypt_decrypt_block_inplace() argument 233 struct page *page, unsigned int len, in ceph_fscrypt_encrypt_block_inplace() argument 240 struct page **page, u64 off, in ceph_fscrypt_decrypt_pages() argument [all …]
|
| /linux/drivers/nvme/target/ |
| H A D | configfs.c | 67 static ssize_t nvmet_addr_adrfam_show(struct config_item *item, char *page) in nvmet_addr_adrfam_show() argument 74 return snprintf(page, PAGE_SIZE, "%s\n", in nvmet_addr_adrfam_show() 78 return snprintf(page, PAGE_SIZE, "\n"); in nvmet_addr_adrfam_show() 82 const char *page, size_t count) in nvmet_addr_adrfam_store() argument 91 if (sysfs_streq(page, nvmet_addr_family[i].name)) in nvmet_addr_adrfam_store() 95 pr_err("Invalid value '%s' for adrfam\n", page); in nvmet_addr_adrfam_store() 106 char *page) in nvmet_addr_portid_show() argument 110 return snprintf(page, PAGE_SIZE, "%d\n", le16_to_cpu(portid)); in nvmet_addr_portid_show() 114 const char *page, size_t count) in nvmet_addr_portid_store() argument 119 if (kstrtou16(page, 0, &portid)) { in nvmet_addr_portid_store() [all …]
|
| /linux/drivers/target/ |
| H A D | target_core_stat.c | 48 static ssize_t target_stat_inst_show(struct config_item *item, char *page) in target_stat_inst_show() argument 52 return snprintf(page, PAGE_SIZE, "%u\n", hba->hba_index); in target_stat_inst_show() 55 static ssize_t target_stat_indx_show(struct config_item *item, char *page) in target_stat_indx_show() argument 57 return snprintf(page, PAGE_SIZE, "%u\n", to_stat_dev(item)->dev_index); in target_stat_indx_show() 60 static ssize_t target_stat_role_show(struct config_item *item, char *page) in target_stat_role_show() argument 62 return snprintf(page, PAGE_SIZE, "Target\n"); in target_stat_role_show() 65 static ssize_t target_stat_ports_show(struct config_item *item, char *page) in target_stat_ports_show() argument 67 return snprintf(page, PAGE_SIZE, "%u\n", to_stat_dev(item)->export_count); in target_stat_ports_show() 98 static ssize_t target_stat_tgt_inst_show(struct config_item *item, char *page) in target_stat_tgt_inst_show() argument 102 return snprintf(page, PAGE_SIZE, "%u\n", hba->hba_index); in target_stat_tgt_inst_show() [all …]
|
| /linux/net/ceph/ |
| H A D | pagelist.c | 32 struct page *page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_unmap_tail() local 33 kunmap(page); in ceph_pagelist_unmap_tail() 44 struct page *page = list_first_entry(&pl->head, struct page, in ceph_pagelist_release() local 46 list_del(&page->lru); in ceph_pagelist_release() 47 __free_page(page); in ceph_pagelist_release() 56 struct page *page; in ceph_pagelist_addpage() local 59 page = __page_cache_alloc(GFP_NOFS); in ceph_pagelist_addpage() 61 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage() 62 list_del(&page->lru); in ceph_pagelist_addpage() 65 if (!page) in ceph_pagelist_addpage() [all …]
|
| /linux/drivers/block/ |
| H A D | brd.c | 49 static struct page *brd_lookup_page(struct brd_device *brd, sector_t sector) in brd_lookup_page() 51 struct page *page; in brd_lookup_page() local 56 page = xas_load(&xas); in brd_lookup_page() 57 if (xas_retry(&xas, page)) { in brd_lookup_page() 62 if (!page) in brd_lookup_page() 65 if (!get_page_unless_zero(page)) { in brd_lookup_page() 70 if (unlikely(page != xas_reload(&xas))) { in brd_lookup_page() 71 put_page(page); in brd_lookup_page() 78 return page; in brd_lookup_page() 85 static struct page *brd_insert_page(struct brd_device *brd, sector_t sector, in brd_insert_page() [all …]
|
| /linux/kernel/dma/ |
| H A D | direct.c | 34 static inline struct page *dma_direct_to_page(struct device *dev, in dma_direct_to_page() 99 static void __dma_direct_free_pages(struct device *dev, struct page *page, in __dma_direct_free_pages() argument 102 if (swiotlb_free(dev, page, size)) in __dma_direct_free_pages() 104 dma_free_contiguous(dev, page, size); in __dma_direct_free_pages() 107 static struct page *dma_direct_alloc_swiotlb(struct device *dev, size_t size) in dma_direct_alloc_swiotlb() 109 struct page *page = swiotlb_alloc(dev, size); in dma_direct_alloc_swiotlb() local 111 if (page && !dma_coherent_ok(dev, page_to_phys(page), size)) { in dma_direct_alloc_swiotlb() 112 swiotlb_free(dev, page, size); in dma_direct_alloc_swiotlb() 116 return page; in dma_direct_alloc_swiotlb() 119 static struct page *__dma_direct_alloc_pages(struct device *dev, size_t size, in __dma_direct_alloc_pages() [all …]
|
| /linux/drivers/hwmon/pmbus/ |
| H A D | mp2891.c | 83 int page) in mp2891_identify_vout_scale() argument 88 ret = i2c_smbus_write_byte_data(client, PMBUS_PAGE, page); in mp2891_identify_vout_scale() 107 data->vout_scale[page] = 250; in mp2891_identify_vout_scale() 111 data->vout_scale[page] = 625; in mp2891_identify_vout_scale() 113 data->vout_scale[page] = 500; in mp2891_identify_vout_scale() 115 data->vout_scale[page] = 200; in mp2891_identify_vout_scale() 117 data->vout_scale[page] = 100; in mp2891_identify_vout_scale() 125 int page) in mp2891_identify_iout_scale() argument 130 ret = i2c_smbus_write_byte_data(client, PMBUS_PAGE, page); in mp2891_identify_iout_scale() 150 data->iout_scale[page] = 32; in mp2891_identify_iout_scale() [all …]
|