Home
last modified time | relevance | path

Searched refs:page_array (Results 1 – 24 of 24) sorted by relevance

/linux/arch/x86/kernel/cpu/sgx/
H A Dvirt.c23 struct xarray page_array; member
46 epc_page = xa_load(&vepc->page_array, index); in __sgx_vepc_fault()
54 ret = xa_err(xa_store(&vepc->page_array, index, epc_page, GFP_KERNEL)); in __sgx_vepc_fault()
69 xa_erase(&vepc->page_array, index); in __sgx_vepc_fault()
160 xa_for_each(&vepc->page_array, index, entry) { in sgx_vepc_remove_all()
197 xa_for_each(&vepc->page_array, index, entry) { in sgx_vepc_release()
207 xa_erase(&vepc->page_array, index); in sgx_vepc_release()
215 xa_for_each(&vepc->page_array, index, entry) { in sgx_vepc_release()
226 xa_erase(&vepc->page_array, index); in sgx_vepc_release()
256 xa_destroy(&vepc->page_array); in sgx_vepc_release()
[all …]
H A Dencl.c86 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in reclaimer_writing_to_pcmd()
287 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in sgx_encl_load_page_in_vma()
307 entry = xa_load(&encl->page_array, PFN_DOWN(addr)); in sgx_encl_load_page()
379 ret = xa_insert(&encl->page_array, PFN_DOWN(encl_page->desc), in sgx_encl_eaug_page()
417 xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc)); in sgx_encl_eaug_page()
457 (!xa_load(&encl->page_array, PFN_DOWN(addr)))) in sgx_vma_fault()
530 XA_STATE(xas, &encl->page_array, PFN_DOWN(start)); in sgx_encl_may_map()
708 XA_STATE(xas, &encl->page_array, PFN_DOWN(encl->base)); in sgx_encl_release()
741 xa_destroy(&encl->page_array); in sgx_encl_release()
H A Dioctl.c307 ret = xa_insert(&encl->page_array, PFN_DOWN(encl_page->desc), in sgx_encl_add_page()
339 xa_erase(&encl->page_array, PFN_DOWN(encl_page->desc)); in sgx_encl_add_page()
1135 xa_erase(&encl->page_array, PFN_DOWN(entry->desc)); in sgx_encl_remove_pages()
H A Ddriver.c27 xa_init(&encl->page_array); in __sgx_open()
/linux/fs/orangefs/
H A Dorangefs-bufmap.c142 struct page **page_array; /* array of mapped pages */ member
154 struct page **page_array; member
171 unpin_user_pages(bufmap->page_array, bufmap->page_count); in orangefs_bufmap_unmap()
177 kfree(bufmap->page_array); in orangefs_bufmap_free()
229 bufmap->page_array = in orangefs_bufmap_alloc()
231 if (!bufmap->page_array) in orangefs_bufmap_alloc()
255 bufmap->page_count, FOLL_WRITE, bufmap->page_array); in orangefs_bufmap_map()
265 unpin_user_page(bufmap->page_array[i]); in orangefs_bufmap_map()
276 flush_dcache_page(bufmap->page_array[i]); in orangefs_bufmap_map()
280 bufmap->desc_array[i].page_array = &bufmap->page_array[offset]; in orangefs_bufmap_map()
[all …]
/linux/drivers/s390/cio/
H A Dvfio_ccw_cp.c22 struct page_array { struct
39 struct page_array *ch_pa; argument
58 static int page_array_alloc(struct page_array *pa, unsigned int len) in page_array_alloc()
91 static void page_array_unpin(struct page_array *pa, in page_array_unpin()
131 static int page_array_pin(struct page_array *pa, struct vfio_device *vdev, bool unaligned) in page_array_pin()
169 static void page_array_unpin_free(struct page_array *pa, struct vfio_device *vdev, bool unaligned) in page_array_unpin_free()
176 static bool page_array_iova_pinned(struct page_array *pa, u64 iova, u64 length) in page_array_iova_pinned()
192 static inline void page_array_idal_create_words(struct page_array *pa, in page_array_idal_create_words()
608 struct page_array *pa, in ccwchain_fetch_ccw()
685 struct page_array *pa, in ccwchain_fetch_one()
[all …]
/linux/drivers/firewire/
H A Dcore-iso.c33 struct page **page_array __free(kfree) = kzalloc_objs(page_array[0], in fw_iso_buffer_alloc()
36 if (!page_array) in fw_iso_buffer_alloc()
43 page_count, page_array); in fw_iso_buffer_alloc()
46 release_pages(page_array, nr_populated); in fw_iso_buffer_alloc()
51 buffer->pages = no_free_ptr(page_array); in fw_iso_buffer_alloc()
/linux/kernel/
H A Drelay.c116 buf->page_array = relay_alloc_page_array(n_pages); in relay_alloc_buf()
117 if (!buf->page_array) in relay_alloc_buf()
121 buf->page_array[i] = alloc_page(GFP_KERNEL | __GFP_ZERO); in relay_alloc_buf()
122 if (unlikely(!buf->page_array[i])) in relay_alloc_buf()
124 set_page_private(buf->page_array[i], (unsigned long)buf); in relay_alloc_buf()
126 mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL); in relay_alloc_buf()
135 __free_page(buf->page_array[j]); in relay_alloc_buf()
136 relay_free_page_array(buf->page_array); in relay_alloc_buf()
199 __free_page(buf->page_array[i]); in relay_destroy_buf()
200 relay_free_page_array(buf->page_array); in relay_destroy_buf()
/linux/io_uring/
H A Drsrc.c680 struct page **page_array = *pages, **new_array = NULL; in io_coalesce_buffer() local
691 struct page *p = compound_head(page_array[j]); in io_coalesce_buffer()
695 WARN_ON_ONCE(i > 0 && p != page_array[j]); in io_coalesce_buffer()
709 kvfree(page_array); in io_coalesce_buffer()
715 bool io_check_coalesce_buffer(struct page **page_array, int nr_pages, in io_check_coalesce_buffer() argument
718 struct folio *folio = page_folio(page_array[0]); in io_check_coalesce_buffer()
724 data->first_folio_page_idx = folio_page_idx(folio, page_array[0]); in io_check_coalesce_buffer()
731 if (page_folio(page_array[i]) == folio && in io_check_coalesce_buffer()
732 page_array[i] == page_array[i-1] + 1) { in io_check_coalesce_buffer()
738 if (folio_page_idx(folio, page_array[i-1]) != in io_check_coalesce_buffer()
[all …]
H A Drsrc.h93 bool io_check_coalesce_buffer(struct page **page_array, int nr_pages, in io_rsrc_node_lookup()
/linux/include/linux/
H A Dgfp.h239 struct page **page_array);
244 struct page **page_array);
254 struct page **page_array) in alloc_pages_bulk_node_noprof() argument
259 return alloc_pages_bulk_noprof(gfp, nid, NULL, nr_pages, page_array); in alloc_pages_bulk_node_noprof()
H A Drelay.h63 struct page **page_array; /* array of current buffer pages */ member
H A Dbpf.h2654 unsigned long nr_pages, struct page **page_array);
/linux/fs/nfs/
H A Dpagelist.c659 if (hdr->page_array.pagevec != hdr->page_array.page_array) in nfs_pgio_data_destroy()
660 kfree(hdr->page_array.pagevec); in nfs_pgio_data_destroy()
696 hdr->args.pages = hdr->page_array.pagevec; in nfs_pgio_rpcsetup()
885 struct nfs_page_array *pg_array = &hdr->page_array; in nfs_generic_pgio()
893 if (pagecount <= ARRAY_SIZE(pg_array->page_array)) in nfs_generic_pgio()
894 pg_array->pagevec = pg_array->page_array; in nfs_generic_pgio()
907 pages = hdr->page_array.pagevec; in nfs_generic_pgio()
H A Dlocalio.c323 iocb->bvec = kmalloc_objs(struct bio_vec, hdr->page_array.npages, flags); in nfs_local_iocb_alloc()
473 struct page **pagevec = hdr->page_array.pagevec; in nfs_local_iters_init()
483 while (total && v < hdr->page_array.npages) { in nfs_local_iters_init()
/linux/mm/
H A Dmempolicy.c2591 struct page **page_array) in alloc_pages_bulk_interleave() argument
2609 page_array); in alloc_pages_bulk_interleave()
2614 nr_pages_per_node, page_array); in alloc_pages_bulk_interleave()
2617 page_array += nr_allocated; in alloc_pages_bulk_interleave()
2626 struct page **page_array) in alloc_pages_bulk_weighted_interleave() argument
2664 page_array); in alloc_pages_bulk_weighted_interleave()
2665 page_array += nr_allocated; in alloc_pages_bulk_weighted_interleave()
2729 page_array); in alloc_pages_bulk_weighted_interleave()
2730 page_array += nr_allocated; in alloc_pages_bulk_weighted_interleave()
2744 struct page **page_array) in alloc_pages_bulk_preferred_many() argument
[all …]
H A Dpage_alloc.c5067 struct page **page_array) in alloc_pages_bulk_noprof() argument
5084 while (nr_populated < nr_pages && page_array[nr_populated]) in alloc_pages_bulk_noprof()
5173 if (page_array[nr_populated]) { in alloc_pages_bulk_noprof()
5192 page_array[nr_populated++] = page; in alloc_pages_bulk_noprof()
5206 page_array[nr_populated++] = page; in alloc_pages_bulk_noprof()
/linux/fs/nfs/blocklayout/
H A Dblocklayout.c248 header->page_array.npages, f_offset, in bl_read_pagelist()
260 for (i = pg_index; i < header->page_array.npages; i++) { in bl_read_pagelist()
293 header->page_array.npages - i, in bl_read_pagelist()
408 for (i = pg_index; i < header->page_array.npages; i++) { in bl_write_pagelist()
422 bio = do_add_page_to_bio(bio, header->page_array.npages - i, in bl_write_pagelist()
/linux/mm/kasan/
H A Dshadow.c342 struct page **page_array = pages; in ___alloc_pages_bulk() local
347 ___free_pages_bulk(page_array, nr_total - nr_pages); in ___alloc_pages_bulk()
/linux/drivers/infiniband/sw/siw/
H A Dsiw_qp_tx.c443 struct page *page_array[MAX_ARRAY]; in siw_tx_hdt() local
512 page_array[seg] = p; in siw_tx_hdt()
540 page_array[seg] = ib_virt_dma_to_page(va); in siw_tx_hdt()
593 rv = siw_0copy_tx(s, page_array, &wqe->sqe.sge[c_tx->sge_idx], in siw_tx_hdt()
/linux/Documentation/filesystems/
H A Dorangefs.rst273 * page_array - a pointer to ``page_count * (sizeof(struct page*))`` bytes
287 bufmap->desc_array[0].page_array = &bufmap->page_array[offset]
294 bufmap->desc_array[9].page_array = &bufmap->page_array[offset]
/linux/fs/btrfs/
H A Dextent_io.h365 int btrfs_alloc_page_array(unsigned int nr_pages, struct page **page_array,
H A Dextent_io.c672 int btrfs_alloc_page_array(unsigned int nr_pages, struct page **page_array, in btrfs_alloc_page_array() argument
681 allocated = alloc_pages_bulk(gfp, nr_pages, page_array); in btrfs_alloc_page_array()
685 __free_page(page_array[i]); in btrfs_alloc_page_array()
686 page_array[i] = NULL; in btrfs_alloc_page_array()
701 struct page *page_array[INLINE_EXTENT_BUFFER_PAGES] = { 0 }; in alloc_eb_folio_array() local
705 ret = btrfs_alloc_page_array(num_pages, page_array, nofail); in alloc_eb_folio_array()
710 eb->folios[i] = page_folio(page_array[i]); in alloc_eb_folio_array()
/linux/drivers/staging/media/atomisp/pci/hmm/
H A Dhmm_bo.c606 static void free_pages_bulk_array(unsigned long nr_pages, struct page **page_array) in free_pages_bulk_array() argument
611 __free_pages(page_array[i], 0); in free_pages_bulk_array()