| /linux/drivers/hv/ |
| H A D | mshv_regions.c | 44 u64 page_offset, u64 page_count, in mshv_region_process_chunk() argument 48 u64 page_count)) in mshv_region_process_chunk() 67 for (count = stride; count < page_count; count += stride) { in mshv_region_process_chunk() 107 u64 page_offset, u64 page_count, in mshv_region_process_range() argument 111 u64 page_count)) in mshv_region_process_range() 115 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range() 118 while (page_count) { in mshv_region_process_range() 122 page_count--; in mshv_region_process_range() 128 page_count, in mshv_region_process_range() 134 page_count -= ret; in mshv_region_process_range() [all …]
|
| H A D | mshv_root_hv_call.c | 191 u64 page_count = page_struct_count; in hv_do_map_gpa_hcall() local 193 if (page_count == 0 || (pages && mmio_spa)) in hv_do_map_gpa_hcall() 200 if (!HV_PAGE_COUNT_2M_ALIGNED(page_count)) in hv_do_map_gpa_hcall() 204 page_count >>= large_shift; in hv_do_map_gpa_hcall() 207 while (done < page_count) { in hv_do_map_gpa_hcall() 208 ulong i, completed, remain = page_count - done; in hv_do_map_gpa_hcall() 268 int hv_call_map_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count, in hv_call_map_gpa_pages() argument 271 return hv_do_map_gpa_hcall(partition_id, gpa_target, page_count, in hv_call_map_gpa_pages() 294 u64 status, page_count = page_count_4k; in hv_call_unmap_gpa_pages() local 298 if (page_count == 0) in hv_call_unmap_gpa_pages() [all …]
|
| H A D | hv_proc.c | 24 int i, j, page_count; in hv_call_deposit_pages() local 82 for (i = 0, page_count = 0; i < num_allocations; ++i) { in hv_call_deposit_pages() 84 for (j = 0; j < counts[i]; ++j, ++page_count) in hv_call_deposit_pages() 85 input_page->gpa_page_list[page_count] = base_pfn + j; in hv_call_deposit_pages() 88 page_count, 0, input_page, NULL); in hv_call_deposit_pages()
|
| H A D | mshv_root_main.c | 706 unsigned long user_pfn, size_t page_count, in mshv_vp_ioctl_get_set_state_pfn() argument 713 if (page_count > INT_MAX) in mshv_vp_ioctl_get_set_state_pfn() 720 if (check_add_overflow(user_pfn, (page_count - 1), &check)) in mshv_vp_ioctl_get_set_state_pfn() 726 pages = kcalloc(page_count, sizeof(struct page *), GFP_KERNEL); in mshv_vp_ioctl_get_set_state_pfn() 730 for (completed = 0; completed < page_count; completed += ret) { in mshv_vp_ioctl_get_set_state_pfn() 732 int remaining = page_count - completed; in mshv_vp_ioctl_get_set_state_pfn() 746 state_data, page_count, pages, in mshv_vp_ioctl_get_set_state_pfn() 751 state_data, page_count, pages, in mshv_vp_ioctl_get_set_state_pfn() 836 size_t page_count = PFN_DOWN(args.buf_sz); in mshv_vp_ioctl_get_set_state() local 839 page_count, is_set); in mshv_vp_ioctl_get_set_state() [all …]
|
| H A D | mshv_root.h | 268 int hv_call_map_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count, 270 int hv_call_unmap_gpa_pages(u64 partition_id, u64 gpa_target, u64 page_count, 284 u64 page_count, struct page **pages, 288 struct hv_vp_state_data state_data, u64 page_count,
|
| /linux/tools/testing/selftests/mm/ |
| H A D | ksm_tests.c | 321 long page_count, int timeout, size_t page_size) in check_ksm_merge() argument 332 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_merge() 336 if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_merge() 340 if (assert_ksm_pages_count(page_count)) { in check_ksm_merge() 342 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 350 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 358 int page_count = 2; in check_ksm_unmerge() local 366 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_unmerge() 370 if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_unmerge() 384 munmap(map_ptr, page_size * page_count); in check_ksm_unmerge() [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_bo85b5.c | 47 u32 page_count = PFN_UP(new_reg->size); in nva3_bo_move_copy() local 50 page_count = PFN_UP(new_reg->size); in nva3_bo_move_copy() 51 while (page_count) { in nva3_bo_move_copy() 52 int line_count = (page_count > 8191) ? 8191 : page_count; in nva3_bo_move_copy() 68 page_count -= line_count; in nva3_bo_move_copy()
|
| H A D | nouveau_bo90b5.c | 40 u32 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_copy() local 43 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_copy() 44 while (page_count) { in nvc0_bo_move_copy() 45 int line_count = (page_count > 8191) ? 8191 : page_count; in nvc0_bo_move_copy() 61 page_count -= line_count; in nvc0_bo_move_copy()
|
| H A D | nouveau_bo9039.c | 45 u32 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_m2mf() local 48 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_m2mf() 49 while (page_count) { in nvc0_bo_move_m2mf() 50 int line_count = (page_count > 2047) ? 2047 : page_count; in nvc0_bo_move_m2mf() 78 page_count -= line_count; in nvc0_bo_move_m2mf()
|
| H A D | nouveau_bo0039.c | 55 u32 page_count = PFN_UP(new_reg->size); in nv04_bo_move_m2mf() local 65 page_count = PFN_UP(new_reg->size); in nv04_bo_move_m2mf() 66 while (page_count) { in nv04_bo_move_m2mf() 67 int line_count = (page_count > 2047) ? 2047 : page_count; in nv04_bo_move_m2mf() 88 page_count -= line_count; in nv04_bo_move_m2mf()
|
| /linux/drivers/firewire/ |
| H A D | core-iso.c | 31 int fw_iso_buffer_alloc(struct fw_iso_buffer *buffer, int page_count) in fw_iso_buffer_alloc() argument 35 buffer->page_count = 0; in fw_iso_buffer_alloc() 37 buffer->pages = kmalloc_array(page_count, sizeof(buffer->pages[0]), in fw_iso_buffer_alloc() 42 for (i = 0; i < page_count; i++) { in fw_iso_buffer_alloc() 47 buffer->page_count = i; in fw_iso_buffer_alloc() 48 if (i < page_count) { in fw_iso_buffer_alloc() 64 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_map_dma() 73 if (i < buffer->page_count) in fw_iso_buffer_map_dma() 80 int page_count, enum dma_data_direction direction) in fw_iso_buffer_init() argument 84 ret = fw_iso_buffer_alloc(buffer, page_count); in fw_iso_buffer_init() [all …]
|
| /linux/drivers/char/agp/ |
| H A D | generic.c | 181 if (curr->page_count != 0) { in agp_free_memory() 186 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 191 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 219 size_t page_count, u32 type) in agp_allocate_memory() argument 230 if ((cur_memory + page_count > bridge->max_memory_agp) || in agp_allocate_memory() 231 (cur_memory + page_count < page_count)) in agp_allocate_memory() 235 new = agp_generic_alloc_user(page_count, type); in agp_allocate_memory() 242 new = bridge->driver->alloc_by_type(page_count, type); in agp_allocate_memory() 248 scratch_pages = (page_count + ENTRIES_PER_PAGE - 1) / ENTRIES_PER_PAGE; in agp_allocate_memory() 256 if (bridge->driver->agp_alloc_pages(bridge, new, page_count)) { in agp_allocate_memory() [all …]
|
| H A D | ati-agp.c | 273 if (mem->page_count == 0) in ati_insert_memory() 276 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory() 280 while (j < (pg_start + mem->page_count)) { in ati_insert_memory() 294 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in ati_insert_memory() 319 if (mem->page_count == 0) in ati_remove_memory() 322 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in ati_remove_memory()
|
| H A D | uninorth-agp.c | 166 if (mem->page_count == 0) in uninorth_insert_memory() 172 if ((pg_start + mem->page_count) > num_entries) in uninorth_insert_memory() 176 for (i = 0; i < mem->page_count; ++i) { in uninorth_insert_memory() 185 for (i = 0; i < mem->page_count; i++) { in uninorth_insert_memory() 215 if (mem->page_count == 0) in uninorth_remove_memory() 219 for (i = 0; i < mem->page_count; ++i) { in uninorth_remove_memory()
|
| H A D | efficeon-agp.c | 238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() 246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() 293 if ((pg_start + mem->page_count) > num_entries) in efficeon_remove_memory()
|
| H A D | agp.h | 194 struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type); 198 struct agp_memory *memory, size_t page_count); 215 struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | dirty.c | 67 le32_to_cpu(region_info[i].page_count), in pds_vfio_print_guest_region_info() 145 u32 page_count) in pds_vfio_dirty_alloc_sgl() argument 154 max_sge = DIV_ROUND_UP(page_count, PAGE_SIZE * 8); in pds_vfio_dirty_alloc_sgl() 199 u32 page_count; in pds_vfio_dirty_alloc_regions() local 202 page_count = le32_to_cpu(ri->page_count); in pds_vfio_dirty_alloc_regions() 204 region_size = page_count * region_page_size; in pds_vfio_dirty_alloc_regions() 207 page_count / BITS_PER_BYTE); in pds_vfio_dirty_alloc_regions() 214 err = pds_vfio_dirty_alloc_sgl(pds_vfio, region, page_count); in pds_vfio_dirty_alloc_regions() 226 dev_bmp_offset_byte += page_count / BITS_PER_BYTE; in pds_vfio_dirty_alloc_regions() 304 u32 page_count; in pds_vfio_dirty_enable() local [all …]
|
| /linux/sound/soc/sof/amd/ |
| H A D | acp-loader.c | 55 u32 dma_size, page_count; in acp_dsp_block_write() local 64 page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT; in acp_dsp_block_write() 65 dma_size = page_count * ACP_PAGE_SIZE; in acp_dsp_block_write() 171 u32 page_count, dma_size; in acp_dsp_pre_fw_run() local 181 page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT; in acp_dsp_pre_fw_run() 182 adata->fw_bin_page_count = page_count; in acp_dsp_pre_fw_run() 184 configure_pte_for_fw_loading(FW_BIN, page_count, adata); in acp_dsp_pre_fw_run() 193 src_addr = ACP_SYSTEM_MEMORY_WINDOW + (page_count * ACP_PAGE_SIZE); in acp_dsp_pre_fw_run() 208 (page_count * ACP_PAGE_SIZE); in acp_dsp_pre_fw_run()
|
| /linux/fs/orangefs/ |
| H A D | orangefs-bufmap.c | 152 int page_count; member 171 unpin_user_pages(bufmap->page_array, bufmap->page_count); in orangefs_bufmap_unmap() 227 bufmap->page_count = bufmap->total_size / PAGE_SIZE; in orangefs_bufmap_alloc() 231 kcalloc(bufmap->page_count, sizeof(struct page *), GFP_KERNEL); in orangefs_bufmap_alloc() 256 bufmap->page_count, FOLL_WRITE, bufmap->page_array); in orangefs_bufmap_map() 261 if (ret != bufmap->page_count) { in orangefs_bufmap_map() 263 bufmap->page_count, ret); in orangefs_bufmap_map() 276 for (i = 0; i < bufmap->page_count; i++) in orangefs_bufmap_map()
|
| /linux/drivers/target/ |
| H A D | target_core_rd.c | 68 u32 i, j, page_count = 0, sg_per_table; in rd_release_sgl_table() local 78 page_count++; in rd_release_sgl_table() 85 return page_count; in rd_release_sgl_table() 90 u32 page_count; in rd_release_device_space() local 95 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_table_array, in rd_release_device_space() 100 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_device_space() 101 rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE); in rd_release_device_space() 216 u32 page_count; in rd_release_prot_space() local 221 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_prot_array, in rd_release_prot_space() 226 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_prot_space() [all …]
|
| /linux/drivers/w1/slaves/ |
| H A D | w1_ds2433.c | 44 unsigned int page_count; /* number of 256 bits pages */ member 50 .page_count = 16, 56 .page_count = 80, 94 bitmap_zero(data->validcrc, data->cfg->page_count); in w1_f23_refresh_block() 337 if (data->cfg->page_count > W1_VALIDCRC_MAX) { in w1_f23_add_slave() 347 bitmap_zero(data->validcrc, data->cfg->page_count); in w1_f23_add_slave()
|
| /linux/drivers/tee/ |
| H A D | tee_shm.c | 26 static void shm_put_kernel_pages(struct page **pages, size_t page_count) in shm_put_kernel_pages() argument 30 for (n = 0; n < page_count; n++) in shm_put_kernel_pages() 34 static void shm_get_kernel_pages(struct page **pages, size_t page_count) in shm_get_kernel_pages() argument 38 for (n = 0; n < page_count; n++) in shm_get_kernel_pages() 303 size_t page_count) in tee_shm_alloc_dma_mem() argument 313 page = dma_alloc_pages(&teedev->dev, page_count * PAGE_SIZE, in tee_shm_alloc_dma_mem() 327 dma_mem->shm.size = page_count * PAGE_SIZE; in tee_shm_alloc_dma_mem() 335 dma_free_pages(&teedev->dev, page_count * PAGE_SIZE, page, dma_addr, in tee_shm_alloc_dma_mem() 345 size_t page_count) in tee_shm_alloc_dma_mem() argument
|
| /linux/drivers/virt/vboxguest/ |
| H A D | vboxguest_utils.c | 199 u32 page_count; in hgcm_call_add_pagelist_size() local 201 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_add_pagelist_size() 202 *extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_add_pagelist_size() 340 u32 i, page_count; in hgcm_call_init_linaddr() local 351 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_init_linaddr() 359 dst_pg_lst->page_count = page_count; in hgcm_call_init_linaddr() 361 for (i = 0; i < page_count; i++) { in hgcm_call_init_linaddr() 371 *off_extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_init_linaddr()
|
| /linux/tools/testing/radix-tree/ |
| H A D | regression2.c | 61 unsigned long page_count = 0; variable 71 p->index = page_count++; in page_alloc()
|
| /linux/fs/pstore/ |
| H A D | ram_core.c | 408 unsigned int page_count; in persistent_ram_vmap() local 414 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); in persistent_ram_vmap() 431 pages = kmalloc_array(page_count, sizeof(struct page *), GFP_KERNEL); in persistent_ram_vmap() 434 __func__, page_count); in persistent_ram_vmap() 438 for (i = 0; i < page_count; i++) { in persistent_ram_vmap() 446 vaddr = vmap(pages, page_count, VM_MAP | VM_IOREMAP, prot); in persistent_ram_vmap()
|