| /linux/tools/testing/selftests/kvm/s390/ |
| H A D | cmma_test.c | 42 " llilf 1,%[start_gfn]\n" in guest_do_one_essa() 51 : [start_gfn] "L"(TEST_DATA_START_GFN) in guest_do_one_essa() 65 " llilf 1,%[start_gfn]\n" in guest_dirty_test_data() 82 : [start_gfn] "L"(TEST_DATA_START_GFN), in guest_dirty_test_data() 216 .start_gfn = 0, in vm_get_cmma_bits() 385 .start_gfn = 0, in assert_all_slots_cmma_dirty() 394 TEST_ASSERT_EQ(args.start_gfn, 0); in assert_all_slots_cmma_dirty() 398 .start_gfn = MAIN_PAGE_COUNT, in assert_all_slots_cmma_dirty() 406 TEST_ASSERT_EQ(args.start_gfn, TEST_DATA_START_GFN); in assert_all_slots_cmma_dirty() 411 .start_gfn in assert_all_slots_cmma_dirty() 474 query_cmma_range(struct kvm_vm * vm,u64 start_gfn,u64 gfn_count,struct kvm_s390_cmma_log * res_out) query_cmma_range() argument [all...] |
| /linux/arch/x86/kvm/ |
| H A D | kvm_onhyperv.c | 14 u64 start_gfn; member 23 return hyperv_fill_flush_guest_mapping_list(flush, range->start_gfn, in kvm_fill_hv_flush_list_func() 95 int hv_flush_remote_tlbs_range(struct kvm *kvm, gfn_t start_gfn, gfn_t nr_pages) in hv_flush_remote_tlbs_range() argument 98 .start_gfn = start_gfn, in hv_flush_remote_tlbs_range()
|
| /linux/drivers/hv/ |
| H A D | mshv_regions.c | 86 u64 gfn = region->start_gfn + page_offset; in mshv_region_process_chunk() 185 region->start_gfn = guest_pfn; in mshv_region_create() 255 region->start_gfn + page_offset, in mshv_region_chunk_remap() 339 region->start_gfn + page_offset, in mshv_region_chunk_unmap() 483 page_offset = ALIGN_DOWN(gfn - region->start_gfn, in mshv_region_handle_gfn_fault()
|
| H A D | mshv_trace.h | 394 TP_PROTO(u64 partition_id, u64 start_uaddr, u64 start_gfn, u64 nr_pages, u32 map_flags, 396 TP_ARGS(partition_id, start_uaddr, start_gfn, nr_pages, map_flags, ret), 400 __field(u64, start_gfn) 408 __entry->start_gfn = start_gfn; 416 __entry->start_gfn,
|
| H A D | mshv_root_main.c | 620 if (gfn >= region->start_gfn && in mshv_partition_region_by_gfn() 621 gfn < region->start_gfn + region->nr_pages) in mshv_partition_region_by_gfn() 1228 if (mem->guest_pfn + nr_pages <= rg->start_gfn || in mshv_partition_create_region() 1229 rg->start_gfn + rg->nr_pages <= mem->guest_pfn) in mshv_partition_create_region() 1293 region->start_gfn, ret); in mshv_prepare_pinned_region() 1308 region->start_gfn, shrc); in mshv_prepare_pinned_region() 1375 region->start_gfn, in mshv_map_user_memory() 1381 region->start_gfn, in mshv_map_user_memory() 1388 region->start_gfn, region->nr_pages, in mshv_map_user_memory() 1425 region->start_gfn != mem->guest_pfn || in mshv_unmap_user_memory()
|
| H A D | mshv_root.h | 88 u64 start_gfn; member
|
| /linux/arch/x86/hyperv/ |
| H A D | nested.c | 58 u64 start_gfn, u64 pages) in hyperv_fill_flush_guest_mapping_list() argument 60 u64 cur = start_gfn; in hyperv_fill_flush_guest_mapping_list()
|
| /linux/arch/loongarch/kvm/ |
| H A D | mmu.c | 292 static void kvm_flush_range(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn, int lock) in kvm_flush_range() argument 305 ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, in kvm_flush_range() 309 ret = kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, in kvm_flush_range() 341 static int kvm_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in kvm_mkclean_gpa_pt() argument 348 return kvm_ptw_top(kvm->arch.pgd, start_gfn << PAGE_SHIFT, end_gfn << PAGE_SHIFT, &ctx); in kvm_mkclean_gpa_pt()
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 861 long kvm_gmem_populate(struct kvm *kvm, gfn_t start_gfn, void __user *src, long npages, in kvm_gmem_populate() argument 876 slot = gfn_to_memslot(kvm, start_gfn); in kvm_gmem_populate() 884 npages = min_t(ulong, slot->npages - (start_gfn - slot->base_gfn), npages); in kvm_gmem_populate() 905 ret = __kvm_gmem_populate(kvm, slot, file, start_gfn + i, src_page, in kvm_gmem_populate()
|
| H A D | kvm_main.c | 3374 gfn_t start_gfn = gpa >> PAGE_SHIFT; in __kvm_gfn_to_hva_cache_init() local 3376 gfn_t nr_pages_needed = end_gfn - start_gfn + 1; in __kvm_gfn_to_hva_cache_init() 3382 if (start_gfn > end_gfn) { in __kvm_gfn_to_hva_cache_init() 3391 for ( ; start_gfn <= end_gfn; start_gfn += nr_pages_avail) { in __kvm_gfn_to_hva_cache_init() 3392 ghc->memslot = __gfn_to_memslot(slots, start_gfn); in __kvm_gfn_to_hva_cache_init() 3393 ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn, in __kvm_gfn_to_hva_cache_init()
|
| /linux/drivers/gpu/drm/i915/gvt/ |
| H A D | gtt.c | 1095 unsigned long start_gfn; in split_2MB_gtt_entry() local 1102 start_gfn = ops->get_pfn(se); in split_2MB_gtt_entry() 1109 ret = intel_gvt_dma_map_guest_page(vgpu, start_gfn + sub_index, in split_2MB_gtt_entry() 1150 unsigned long start_gfn; in split_64KB_gtt_entry() local 1158 start_gfn = ops->get_pfn(se); in split_64KB_gtt_entry() 1164 ret = intel_gvt_dma_map_guest_page(vgpu, start_gfn + i, in split_64KB_gtt_entry()
|
| /linux/arch/s390/kvm/ |
| H A D | kvm-s390.c | 984 gfn_t start_gfn = 0; in kvm_s390_set_mem_control() local 995 start_gfn = dat_reset_cmma(kvm->arch.gmap->asce, start_gfn); in kvm_s390_set_mem_control() 997 } while (start_gfn); in kvm_s390_set_mem_control() 2120 args->start_gfn + i, keys + i); in kvm_s390_get_skeys() 2187 args->start_gfn + i, keys[i], 0); in kvm_s390_set_skeys() 2240 ret = dat_peek_cmma(args->start_gfn, kvm->arch.gmap->asce, &args->count, in kvm_s390_get_cmma_bits() 2243 ret = dat_get_cmma(kvm->arch.gmap->asce, &args->start_gfn, &args->count, in kvm_s390_get_cmma_bits() 2301 r = dat_set_cmma_bits(mc, kvm->arch.gmap->asce, args->start_gfn, in kvm_s390_set_cmma_bits()
|
| H A D | dat.h | 550 long dat_reset_cmma(union asce asce, gfn_t start_gfn);
|
| /linux/Documentation/virt/kvm/ |
| H A D | api.rst | 4017 __u64 start_gfn; 4024 The start_gfn field is the number of the first guest frame whose storage keys 4027 The count field is the number of consecutive frames (starting from start_gfn) 4048 The start_gfn field is the number of the first guest frame whose storage keys 4051 The count field is the number of consecutive frames (starting from start_gfn) 4058 single frame starting at start_gfn for count frames. 4655 __u64 start_gfn; 4665 start_gfn is the number of the first guest frame whose CMMA values are 4683 start_gfn will indicate the first page frame whose CMMA bits were dirty. 4703 the output start_gfn will be equal to the input start_gfn, [all …]
|