Home
last modified time | relevance | path

Searched refs:page_count (Results 1 – 25 of 115) sorted by relevance

12345

/linux/drivers/hv/
H A Dmshv_regions.c35 u64 gfn, u64 page_count) in mshv_chunk_stride() argument
46 !IS_ALIGNED(page_count, PTRS_PER_PMD)) in mshv_chunk_stride()
79 u64 page_offset, u64 page_count, in mshv_region_process_chunk() argument
83 u64 page_count, in mshv_region_process_chunk()
95 stride = mshv_chunk_stride(page, gfn, page_count); in mshv_region_process_chunk()
100 for (count = stride; count < page_count; count += stride) { in mshv_region_process_chunk()
109 page_count - count)) in mshv_region_process_chunk()
141 u64 page_offset, u64 page_count, in mshv_region_process_range() argument
145 u64 page_count, in mshv_region_process_range()
150 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range()
[all …]
H A Dmshv_root_hv_call.c190 u64 page_count = page_struct_count; in hv_do_map_gpa_hcall() local
192 if (page_count == 0 || (pages && mmio_spa)) in hv_do_map_gpa_hcall()
199 if (!HV_PAGE_COUNT_2M_ALIGNED(page_count)) in hv_do_map_gpa_hcall()
203 page_count >>= large_shift; in hv_do_map_gpa_hcall()
206 while (done < page_count) { in hv_do_map_gpa_hcall()
207 ulong i, completed, remain = page_count - done; in hv_do_map_gpa_hcall()
267 int hv_call_map_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count, in hv_call_map_gpa_pages() argument
270 return hv_do_map_gpa_hcall(partition_id, gpa_target, page_count, in hv_call_map_gpa_pages()
293 u64 status, page_count = page_count_4k; in hv_call_unmap_gpa_pages() local
297 if (page_count == 0) in hv_call_unmap_gpa_pages()
[all …]
H A Dhv_proc.c24 int i, j, page_count; in hv_call_deposit_pages() local
82 for (i = 0, page_count = 0; i < num_allocations; ++i) { in hv_call_deposit_pages()
84 for (j = 0; j < counts[i]; ++j, ++page_count) in hv_call_deposit_pages()
85 input_page->gpa_page_list[page_count] = base_pfn + j; in hv_call_deposit_pages()
88 page_count, 0, input_page, NULL); in hv_call_deposit_pages()
/linux/tools/testing/selftests/mm/
H A Dksm_tests.c321 long page_count, int timeout, size_t page_size) in check_ksm_merge() argument
332 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_merge()
336 if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_merge()
340 if (assert_ksm_pages_count(page_count)) { in check_ksm_merge()
342 munmap(map_ptr, page_size * page_count); in check_ksm_merge()
350 munmap(map_ptr, page_size * page_count); in check_ksm_merge()
358 int page_count = 2; in check_ksm_unmerge() local
366 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_unmerge()
370 if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_unmerge()
384 munmap(map_ptr, page_size * page_count); in check_ksm_unmerge()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_bo85b5.c47 u32 page_count = PFN_UP(new_reg->size); in nva3_bo_move_copy() local
50 page_count = PFN_UP(new_reg->size); in nva3_bo_move_copy()
51 while (page_count) { in nva3_bo_move_copy()
52 int line_count = (page_count > 8191) ? 8191 : page_count; in nva3_bo_move_copy()
68 page_count -= line_count; in nva3_bo_move_copy()
H A Dnouveau_bo90b5.c40 u32 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_copy() local
43 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_copy()
44 while (page_count) { in nvc0_bo_move_copy()
45 int line_count = (page_count > 8191) ? 8191 : page_count; in nvc0_bo_move_copy()
61 page_count -= line_count; in nvc0_bo_move_copy()
H A Dnouveau_bo9039.c45 u32 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_m2mf() local
48 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_m2mf()
49 while (page_count) { in nvc0_bo_move_m2mf()
50 int line_count = (page_count > 2047) ? 2047 : page_count; in nvc0_bo_move_m2mf()
78 page_count -= line_count; in nvc0_bo_move_m2mf()
H A Dnouveau_bo0039.c55 u32 page_count = PFN_UP(new_reg->size); in nv04_bo_move_m2mf() local
65 page_count = PFN_UP(new_reg->size); in nv04_bo_move_m2mf()
66 while (page_count) { in nv04_bo_move_m2mf()
67 int line_count = (page_count > 2047) ? 2047 : page_count; in nv04_bo_move_m2mf()
88 page_count -= line_count; in nv04_bo_move_m2mf()
/linux/drivers/firewire/
H A Dcore-iso.c31 int fw_iso_buffer_alloc(struct fw_iso_buffer *buffer, int page_count) in fw_iso_buffer_alloc() argument
34 page_count); in fw_iso_buffer_alloc()
43 page_count, page_array); in fw_iso_buffer_alloc()
44 if (nr_populated != page_count) { in fw_iso_buffer_alloc()
50 buffer->page_count = page_count; in fw_iso_buffer_alloc()
60 buffer->page_count); in fw_iso_buffer_map_dma()
68 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_map_dma()
77 if (i < buffer->page_count) { in fw_iso_buffer_map_dma()
90 int page_count, enum dma_data_direction direction) in fw_iso_buffer_init() argument
94 ret = fw_iso_buffer_alloc(buffer, page_count); in fw_iso_buffer_init()
[all …]
/linux/drivers/char/agp/
H A Dgeneric.c181 if (curr->page_count != 0) { in agp_free_memory()
186 for (i = 0; i < curr->page_count; i++) { in agp_free_memory()
191 for (i = 0; i < curr->page_count; i++) { in agp_free_memory()
219 size_t page_count, u32 type) in agp_allocate_memory() argument
230 if ((cur_memory + page_count > bridge->max_memory_agp) || in agp_allocate_memory()
231 (cur_memory + page_count < page_count)) in agp_allocate_memory()
235 new = agp_generic_alloc_user(page_count, type); in agp_allocate_memory()
242 new = bridge->driver->alloc_by_type(page_count, type); in agp_allocate_memory()
248 scratch_pages = (page_count + ENTRIES_PER_PAGE - 1) / ENTRIES_PER_PAGE; in agp_allocate_memory()
256 if (bridge->driver->agp_alloc_pages(bridge, new, page_count)) { in agp_allocate_memory()
[all …]
H A Dati-agp.c272 if (mem->page_count == 0) in ati_insert_memory()
275 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory()
279 while (j < (pg_start + mem->page_count)) { in ati_insert_memory()
293 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in ati_insert_memory()
318 if (mem->page_count == 0) in ati_remove_memory()
321 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in ati_remove_memory()
H A Duninorth-agp.c166 if (mem->page_count == 0) in uninorth_insert_memory()
172 if ((pg_start + mem->page_count) > num_entries) in uninorth_insert_memory()
176 for (i = 0; i < mem->page_count; ++i) { in uninorth_insert_memory()
185 for (i = 0; i < mem->page_count; i++) { in uninorth_insert_memory()
215 if (mem->page_count == 0) in uninorth_remove_memory()
219 for (i = 0; i < mem->page_count; ++i) { in uninorth_remove_memory()
H A Defficeon-agp.c238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory()
246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory()
287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory()
293 if ((pg_start + mem->page_count) > num_entries) in efficeon_remove_memory()
H A Dagp.h194 struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type);
198 struct agp_memory *memory, size_t page_count);
215 struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
/linux/drivers/vfio/pci/pds/
H A Ddirty.c65 le32_to_cpu(region_info[i].page_count), in pds_vfio_print_guest_region_info()
143 u32 page_count) in pds_vfio_dirty_alloc_sgl() argument
152 max_sge = DIV_ROUND_UP(page_count, PAGE_SIZE * 8); in pds_vfio_dirty_alloc_sgl()
197 u32 page_count; in pds_vfio_dirty_alloc_regions() local
200 page_count = le32_to_cpu(ri->page_count); in pds_vfio_dirty_alloc_regions()
202 region_size = page_count * region_page_size; in pds_vfio_dirty_alloc_regions()
205 page_count / BITS_PER_BYTE); in pds_vfio_dirty_alloc_regions()
212 err = pds_vfio_dirty_alloc_sgl(pds_vfio, region, page_count); in pds_vfio_dirty_alloc_regions()
224 dev_bmp_offset_byte += page_count / BITS_PER_BYTE; in pds_vfio_dirty_alloc_regions()
302 u32 page_count; in pds_vfio_dirty_enable() local
[all …]
/linux/fs/squashfs/
H A Dblock.c81 u64 read_start, u64 read_end, int page_count) in squashfs_bio_read_cached() argument
91 struct folio **cache_folios = kmalloc_objs(*cache_folios, page_count, in squashfs_bio_read_cached()
114 else if (idx == page_count - 1 && index + length != read_end) in squashfs_bio_read_cached()
178 for (idx = 0; idx < page_count; idx++) { in squashfs_bio_read_cached()
227 const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE); in squashfs_bio_read() local
231 bio = bio_kmalloc(page_count, GFP_NOIO); in squashfs_bio_read()
234 bio_init_inline(bio, sb->s_bdev, page_count, REQ_OP_READ); in squashfs_bio_read()
237 for (i = 0; i < page_count; ++i) { in squashfs_bio_read()
264 page_count); in squashfs_bio_read()
/linux/sound/soc/sof/amd/
H A Dacp-loader.c55 u32 dma_size, page_count; in acp_dsp_block_write() local
64 page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT; in acp_dsp_block_write()
65 dma_size = page_count * ACP_PAGE_SIZE; in acp_dsp_block_write()
171 u32 page_count, dma_size; in acp_dsp_pre_fw_run() local
181 page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT; in acp_dsp_pre_fw_run()
182 adata->fw_bin_page_count = page_count; in acp_dsp_pre_fw_run()
184 configure_pte_for_fw_loading(FW_BIN, page_count, adata); in acp_dsp_pre_fw_run()
193 src_addr = ACP_SYSTEM_MEMORY_WINDOW + (page_count * ACP_PAGE_SIZE); in acp_dsp_pre_fw_run()
208 (page_count * ACP_PAGE_SIZE); in acp_dsp_pre_fw_run()
/linux/fs/orangefs/
H A Dorangefs-bufmap.c152 int page_count; member
171 unpin_user_pages(bufmap->page_array, bufmap->page_count); in orangefs_bufmap_unmap()
226 bufmap->page_count = bufmap->total_size / PAGE_SIZE; in orangefs_bufmap_alloc()
230 kzalloc_objs(struct page *, bufmap->page_count); in orangefs_bufmap_alloc()
255 bufmap->page_count, FOLL_WRITE, bufmap->page_array); in orangefs_bufmap_map()
260 if (ret != bufmap->page_count) { in orangefs_bufmap_map()
262 bufmap->page_count, ret); in orangefs_bufmap_map()
275 for (i = 0; i < bufmap->page_count; i++) in orangefs_bufmap_map()
/linux/drivers/target/
H A Dtarget_core_rd.c68 u32 i, j, page_count = 0, sg_per_table; in rd_release_sgl_table() local
78 page_count++; in rd_release_sgl_table()
85 return page_count; in rd_release_sgl_table()
90 u32 page_count; in rd_release_device_space() local
95 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_table_array, in rd_release_device_space()
100 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_device_space()
101 rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE); in rd_release_device_space()
215 u32 page_count; in rd_release_prot_space() local
220 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_prot_array, in rd_release_prot_space()
225 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_prot_space()
[all …]
/linux/drivers/w1/slaves/
H A Dw1_ds2433.c44 unsigned int page_count; /* number of 256 bits pages */ member
50 .page_count = 16,
56 .page_count = 80,
94 bitmap_zero(data->validcrc, data->cfg->page_count); in w1_f23_refresh_block()
337 if (data->cfg->page_count > W1_VALIDCRC_MAX) { in w1_f23_add_slave()
347 bitmap_zero(data->validcrc, data->cfg->page_count); in w1_f23_add_slave()
/linux/drivers/tee/
H A Dtee_shm.c26 static void shm_put_kernel_pages(struct page **pages, size_t page_count) in shm_put_kernel_pages() argument
30 for (n = 0; n < page_count; n++) in shm_put_kernel_pages()
34 static void shm_get_kernel_pages(struct page **pages, size_t page_count) in shm_get_kernel_pages() argument
38 for (n = 0; n < page_count; n++) in shm_get_kernel_pages()
303 size_t page_count) in tee_shm_alloc_dma_mem() argument
313 page = dma_alloc_pages(&teedev->dev, page_count * PAGE_SIZE, in tee_shm_alloc_dma_mem()
327 dma_mem->shm.size = page_count * PAGE_SIZE; in tee_shm_alloc_dma_mem()
335 dma_free_pages(&teedev->dev, page_count * PAGE_SIZE, page, dma_addr, in tee_shm_alloc_dma_mem()
345 size_t page_count) in tee_shm_alloc_dma_mem() argument
/linux/drivers/virt/vboxguest/
H A Dvboxguest_utils.c199 u32 page_count; in hgcm_call_add_pagelist_size() local
201 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_add_pagelist_size()
202 *extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_add_pagelist_size()
340 u32 i, page_count; in hgcm_call_init_linaddr() local
351 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_init_linaddr()
359 dst_pg_lst->page_count = page_count; in hgcm_call_init_linaddr()
361 for (i = 0; i < page_count; i++) { in hgcm_call_init_linaddr()
371 *off_extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_init_linaddr()
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_shmem.c70 unsigned int page_count; /* restricted by sg_alloc_table */ in shmem_sg_alloc_table() local
77 if (overflows_type(size / PAGE_SIZE, page_count)) in shmem_sg_alloc_table()
80 page_count = size / PAGE_SIZE; in shmem_sg_alloc_table()
88 if (sg_alloc_table(st, page_count, GFP_KERNEL | __GFP_NOWARN)) in shmem_sg_alloc_table()
103 for (i = 0; i < page_count; i++) { in shmem_sg_alloc_table()
123 i915_gem_shrink(NULL, i915, 2 * page_count, NULL, *s++); in shmem_sg_alloc_table()
157 folio_nr_pages(folio), page_count - i); in shmem_sg_alloc_table()
/linux/tools/testing/radix-tree/
H A Dregression2.c61 unsigned long page_count = 0; variable
71 p->index = page_count++; in page_alloc()
/linux/drivers/gpu/drm/ttm/
H A Dttm_agp_backend.c66 mem->page_count = 0; in ttm_agp_bind()
73 mem->pages[mem->page_count++] = page; in ttm_agp_bind()

12345