| /linux/mm/ |
| H A D | readahead.c | 348 unsigned long max_pages; in force_page_cache_ra() 357 max_pages = max_t(unsigned long, bdi->io_pages, ra->ra_pages); in force_page_cache_ra() 358 nr_to_read = min_t(unsigned long, nr_to_read, max_pages); in force_page_cache_ra() 546 unsigned long max_pages = ractl->ra->ra_pages; in ractl_max_pages() 552 if (req_size > max_pages && bdi->io_pages > max_pages) in ractl_max_pages() 553 max_pages = min(req_size, bdi->io_pages); 554 return max_pages; in page_cache_sync_ra() 563 unsigned long max_pages, contig_count; in page_cache_sync_ra() 586 max_pages in page_cache_sync_ra() 345 unsigned long max_pages; force_page_cache_ra() local 543 unsigned long max_pages = ractl->ra->ra_pages; ractl_max_pages() local 560 unsigned long max_pages, contig_count; page_cache_sync_ra() local 633 unsigned long max_pages; page_cache_async_ra() local [all...] |
| H A D | swap_state.c | 645 int max_pages, in swap_cluster_readahead() 671 if (pages > max_pages) in swap_vma_ra_win() 672 pages = max_pages; in swap_vma_ra_win() 685 unsigned int hits, pages, max_pages; in swap_vma_ra_win() 688 max_pages = 1 << READ_ONCE(page_cluster); in swap_vma_ra_win() 689 if (max_pages <= 1) in swap_vma_ra_win() 694 max_pages, in swap_vma_ra_win() 541 __swapin_nr_pages(unsigned long prev_offset,unsigned long offset,int hits,int max_pages,int prev_win) __swapin_nr_pages() argument 581 unsigned int hits, pages, max_pages; swapin_nr_pages() local
|
| /linux/fs/netfs/ |
| H A D | iterator.c | 43 unsigned int max_pages; in netfs_extract_user_iter() local 53 max_pages = iov_iter_npages(orig, INT_MAX); in netfs_extract_user_iter() 54 bv_size = array_size(max_pages, sizeof(*bv)); in netfs_extract_user_iter() 63 pg_size = array_size(max_pages, sizeof(*pages)); in netfs_extract_user_iter() 66 while (count && npages < max_pages) { in netfs_extract_user_iter() 68 max_pages - npages, extraction_flags, in netfs_extract_user_iter() 84 if (npages + cur_npages > max_pages) { in netfs_extract_user_iter() 86 npages + cur_npages, max_pages); in netfs_extract_user_iter()
|
| /linux/fs/fuse/ |
| H A D | ioctl.c | 78 u32 max = fc->max_pages << PAGE_SHIFT; in fuse_verify_ioctl_iov() 231 unsigned int in_iovs = 0, out_iovs = 0, max_pages; in fuse_do_ioctl() local 254 ap.folios = fuse_folios_alloc(fm->fc->max_pages, GFP_KERNEL, &ap.descs); in fuse_do_ioctl() 259 fuse_folio_descs_length_init(ap.descs, 0, fm->fc->max_pages); in fuse_do_ioctl() 303 max_pages = DIV_ROUND_UP(max(in_size, out_size), PAGE_SIZE); in fuse_do_ioctl() 307 if (max_pages > fm->fc->max_pages) in fuse_do_ioctl() 309 while (ap.num_folios < max_pages) { in fuse_do_ioctl()
|
| H A D | file.c | 907 nr_pages = min(fc->max_pages, readahead_count(rac)); in fuse_handle_readahead() 1324 unsigned int max_pages) in fuse_wr_pages() argument 1329 return min(pages, max_pages); in fuse_wr_pages() 1350 fc->max_pages); in fuse_perform_write() 1545 unsigned int max_pages, in fuse_get_user_pages() argument 1584 struct page **pages = kzalloc(max_pages * sizeof(struct page *), in fuse_get_user_pages() 1591 while (nbytes < *nbytesp && nr_pages < max_pages) { in fuse_get_user_pages() 1597 max_pages - nr_pages, in fuse_get_user_pages() 1659 unsigned int max_pages; in fuse_direct_io() local 1662 max_pages = iov_iter_npages(iter, fc->max_pages); in fuse_direct_io() [all …]
|
| /linux/include/linux/ |
| H A D | swap_cgroup.h | 12 extern int swap_cgroup_swapon(int type, unsigned long max_pages); 35 swap_cgroup_swapon(int type, unsigned long max_pages) in swap_cgroup_swapon() argument
|
| H A D | memory.h | 59 unsigned long max_pages; member 166 extern int memory_group_register_static(int nid, unsigned long max_pages);
|
| /linux/kernel/module/ |
| H A D | decompress.c | 20 unsigned int new_max = info->max_pages + extent; in module_extend_max_pages() 29 info->max_pages = new_max; in module_extend_max_pages() 39 if (info->max_pages == info->used_pages) { in module_get_next_page() 344 info->max_pages = info->used_pages = 0; in module_decompress_cleanup()
|
| /linux/fs/nfs/ |
| H A D | pnfs_dev.c | 105 int max_pages; in nfs4_get_device_info() local 113 max_pages = nfs_page_array_len(0, max_resp_sz); in nfs4_get_device_info() 115 __func__, server, max_resp_sz, max_pages); in nfs4_get_device_info() 121 pages = kzalloc_objs(struct page *, max_pages, gfp_flags); in nfs4_get_device_info() 125 for (i = 0; i < max_pages; i++) { in nfs4_get_device_info()
|
| /linux/fs/iomap/ |
| H A D | swapfile.c | 31 unsigned long max_pages; in iomap_swapfile_add_extent() local 39 max_pages = isi->sis->max - isi->nr_pages; in iomap_swapfile_add_extent() 53 nr_pages = min(nr_pages, max_pages); in iomap_swapfile_add_extent()
|
| /linux/fs/cramfs/ |
| H A D | inode.c | 359 unsigned int pages, max_pages, offset; in cramfs_physmem_mmap() local 378 max_pages = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT; in cramfs_physmem_mmap() 380 if (pgoff >= max_pages) in cramfs_physmem_mmap() 382 pages = min(vma_pages(vma), max_pages - pgoff); in cramfs_physmem_mmap() 394 if (pgoff + pages == max_pages && cramfs_last_page_is_shared(inode)) { in cramfs_physmem_mmap() 459 unsigned int pages, block_pages, max_pages, offset; in cramfs_physmem_get_unmapped_area() local 462 max_pages = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT; in cramfs_physmem_get_unmapped_area() 463 if (pgoff >= max_pages || pages > max_pages - pgoff) in cramfs_physmem_get_unmapped_area()
|
| /linux/drivers/gpu/drm/imagination/ |
| H A D | pvr_free_list.c | 91 free_list->max_pages = args->max_num_pages; in free_list_create_kernel_structure() 179 fw_data->max_pages = free_list->max_pages; in free_list_fw_init() 188 ((fw_data->max_pages - fw_data->current_pages) * in free_list_fw_init() 279 start_page = free_list->max_pages - free_list->current_pages - in pvr_free_list_insert_node_locked()
|
| H A D | pvr_free_list.h | 76 u32 max_pages; member
|
| /linux/fs/squashfs/ |
| H A D | page_actor.c | 70 int max_pages = (actor->length + PAGE_SIZE - 1) >> PAGE_SHIFT; in handle_next_page() local 72 if (actor->returned_pages == max_pages) in handle_next_page()
|
| H A D | file.c | 587 unsigned int max_pages = 1UL << shift; in squashfs_readahead() local 591 pages = kmalloc_array(max_pages, sizeof(void *), GFP_KERNEL); in squashfs_readahead() 605 max_pages = (expected + PAGE_SIZE - 1) >> PAGE_SHIFT; in squashfs_readahead() 607 nr_pages = __readahead_batch(ractl, pages, max_pages); in squashfs_readahead()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | mr.c | 293 int max_pages) in mlx4_alloc_priv_pages() argument 302 mr->page_map_size = roundup(max_pages * sizeof(u64), in mlx4_alloc_priv_pages() 410 mr->max_pages = max_num_sg; in mlx4_ib_alloc_mr() 434 if (unlikely(mr->npages == mr->max_pages)) in mlx4_set_page()
|
| /linux/net/rds/ |
| H A D | ib_rdma.c | 180 iinfo->rdma_mr_size = pool_1m->max_pages; in rds_ib_get_mr_info() 190 iinfo6->rdma_mr_size = pool_1m->max_pages; in rds6_ib_get_mr_info() 659 pool->max_pages = RDS_MR_1M_MSG_SIZE + 1; in rds_ib_create_mr_pool() 663 pool->max_pages = RDS_MR_8K_MSG_SIZE + 1; in rds_ib_create_mr_pool() 667 pool->max_free_pinned = pool->max_items * pool->max_pages / 4; in rds_ib_create_mr_pool()
|
| /linux/drivers/md/bcache/ |
| H A D | bcache.h | 780 unsigned int n, max_pages; in meta_bucket_pages() local 782 max_pages = min_t(unsigned int, in meta_bucket_pages() 787 if (n > max_pages) in meta_bucket_pages() 788 n = max_pages; in meta_bucket_pages()
|
| /linux/crypto/ |
| H A D | algif_hash.c | 70 size_t len, max_pages, npages; in hash_sendmsg() local 74 max_pages = min_t(size_t, ALG_MAX_PAGES, in hash_sendmsg() 111 npages = iov_iter_npages(&msg->msg_iter, max_pages); in hash_sendmsg()
|
| /linux/fs/ceph/ |
| H A D | addr.c | 173 unsigned long max_pages = inode->i_sb->s_bdi->ra_pages; in ceph_netfs_expand_readahead() local 182 max_pages = 0; in ceph_netfs_expand_readahead() 184 max_pages = priv->file_ra_pages; in ceph_netfs_expand_readahead() 189 if (!max_pages) in ceph_netfs_expand_readahead() 192 max_len = max_pages << PAGE_SHIFT; in ceph_netfs_expand_readahead() 601 unsigned int max_pages; member 1038 ceph_wbc->max_pages = 0; in ceph_init_writeback_ctl() 1130 ceph_wbc->locked_pages < ceph_wbc->max_pages; in can_next_page_be_processed() 1187 unsigned int max_pages) in __ceph_allocate_page_array() argument 1189 ceph_wbc->pages = kmalloc_objs(*ceph_wbc->pages, max_pages, GFP_NOFS); in __ceph_allocate_page_array() [all …]
|
| /linux/drivers/gpu/drm/amd/ras/rascore/ |
| H A D | ras_core.c | 606 uint64_t soc_pa, uint64_t *page_pfn, uint32_t max_pages) in ras_core_convert_soc_pa_to_cur_nps_pages() argument 612 if (!ras_core || !page_pfn || !max_pages) in ras_core_convert_soc_pa_to_cur_nps_pages() 623 &record, cur_nps_mode, page_pfn, max_pages); in ras_core_convert_soc_pa_to_cur_nps_pages()
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_mr.c | 257 mr->max_pages = max_num_sg; in pvrdma_alloc_mr() 311 if (mr->npages == mr->max_pages) in pvrdma_set_page()
|
| /linux/drivers/base/ |
| H A D | memory.c | 1113 int memory_group_register_static(int nid, unsigned long max_pages) in memory_group_register_static() argument 1118 .max_pages = max_pages, in memory_group_register_static() 1122 if (!max_pages) in memory_group_register_static()
|
| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_eqs.c | 508 u32 max_pages, min_page_size, page_size, total_size; in eq_calc_page_size_and_num() local 514 max_pages = HINIC3_EQ_MAX_PAGES(eq); in eq_calc_page_size_and_num() 518 if (total_size <= max_pages * min_page_size) in eq_calc_page_size_and_num() 521 page_size = total_size / max_pages; in eq_calc_page_size_and_num()
|
| /linux/io_uring/ |
| H A D | memmap.c | 287 unsigned max_pages) in io_region_mmap() argument 289 unsigned long nr_pages = min(mr->nr_pages, max_pages); in io_region_mmap()
|