H A D | kvm_util.c | 182 void vm_enable_dirty_ring(struct kvm_vm *vm, uint32_t ring_size) in vm_enable_dirty_ring() argument 184 if (vm_check_cap(vm, KVM_CAP_DIRTY_LOG_RING_ACQ_REL)) in vm_enable_dirty_ring() 185 vm_enable_cap(vm, KVM_CAP_DIRTY_LOG_RING_ACQ_REL, ring_size); in vm_enable_dirty_ring() 187 vm_enable_cap(vm, KVM_CAP_DIRTY_LOG_RING, ring_size); in vm_enable_dirty_ring() 188 vm->dirty_ring_size = ring_size; in vm_enable_dirty_ring() 191 static void vm_open(struct kvm_vm *vm) in vm_open() argument 193 vm->kvm_fd = _open_kvm_dev_path_or_exit(O_RDWR); in vm_open() 197 vm->fd = __kvm_ioctl(vm->kvm_fd, KVM_CREATE_VM, (void *)vm in vm_open() 259 vm_vaddr_populate_bitmap(struct kvm_vm * vm) vm_vaddr_populate_bitmap() argument 270 struct kvm_vm *vm; ____vm_create() local 415 struct kvm_vm *vm; __vm_create() local 473 struct kvm_vm *vm; __vm_create_with_vcpus() local 492 struct kvm_vm *vm; __vm_create_shape_with_one_vcpu() local 535 vm_arch_vcpu_recreate(struct kvm_vm * vm,uint32_t vcpu_id) vm_arch_vcpu_recreate() argument 541 vm_recreate_with_one_vcpu(struct kvm_vm * vm) vm_recreate_with_one_vcpu() argument 638 userspace_mem_region_find(struct kvm_vm * vm,uint64_t start,uint64_t end) userspace_mem_region_find() argument 677 vm_vcpu_rm(struct kvm_vm * vm,struct kvm_vcpu * vcpu) vm_vcpu_rm() argument 714 __vm_mem_region_delete(struct kvm_vm * vm,struct userspace_mem_region * region) __vm_mem_region_delete() argument 846 __vm_set_user_memory_region(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva) __vm_set_user_memory_region() argument 860 vm_set_user_memory_region(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva) vm_set_user_memory_region() argument 873 __vm_set_user_memory_region2(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva,uint32_t guest_memfd,uint64_t guest_memfd_offset) __vm_set_user_memory_region2() argument 892 vm_set_user_memory_region2(struct kvm_vm * vm,uint32_t slot,uint32_t flags,uint64_t gpa,uint64_t size,void * hva,uint32_t guest_memfd,uint64_t guest_memfd_offset) vm_set_user_memory_region2() argument 905 vm_mem_add(struct kvm_vm * vm,enum vm_mem_backing_src_type src_type,uint64_t guest_paddr,uint32_t slot,uint64_t npages,uint32_t flags,int guest_memfd,uint64_t guest_memfd_offset) vm_mem_add() argument 1084 vm_userspace_mem_region_add(struct kvm_vm * vm,enum vm_mem_backing_src_type src_type,uint64_t guest_paddr,uint32_t slot,uint64_t npages,uint32_t flags) vm_userspace_mem_region_add() argument 1108 memslot2region(struct kvm_vm * vm,uint32_t memslot) memslot2region() argument 1139 vm_mem_region_set_flags(struct kvm_vm * vm,uint32_t slot,uint32_t flags) vm_mem_region_set_flags() argument 1169 vm_mem_region_move(struct kvm_vm * vm,uint32_t slot,uint64_t new_gpa) vm_mem_region_move() argument 1198 vm_mem_region_delete(struct kvm_vm * vm,uint32_t slot) vm_mem_region_delete() argument 1203 vm_guest_mem_fallocate(struct kvm_vm * vm,uint64_t base,uint64_t size,bool punch_hole) vm_guest_mem_fallocate() argument 1247 vcpu_exists(struct kvm_vm * vm,uint32_t vcpu_id) vcpu_exists() argument 1263 __vm_vcpu_add(struct kvm_vm * vm,uint32_t vcpu_id) __vm_vcpu_add() argument 1313 vm_vaddr_unused_gap(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min) vm_vaddr_unused_gap() argument 1379 ____vm_vaddr_alloc(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min,enum kvm_mem_region_type type,bool protected) ____vm_vaddr_alloc() argument 1409 __vm_vaddr_alloc(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min,enum kvm_mem_region_type type) __vm_vaddr_alloc() argument 1416 vm_vaddr_alloc_shared(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min,enum kvm_mem_region_type type) vm_vaddr_alloc_shared() argument 1442 vm_vaddr_alloc(struct kvm_vm * vm,size_t sz,vm_vaddr_t vaddr_min) vm_vaddr_alloc() argument 1461 vm_vaddr_alloc_pages(struct kvm_vm * vm,int nr_pages) vm_vaddr_alloc_pages() argument 1466 __vm_vaddr_alloc_page(struct kvm_vm * vm,enum kvm_mem_region_type type) __vm_vaddr_alloc_page() argument 1485 vm_vaddr_alloc_page(struct kvm_vm * vm) vm_vaddr_alloc_page() argument 1506 virt_map(struct kvm_vm * vm,uint64_t vaddr,uint64_t paddr,unsigned int npages) virt_map() argument 1541 addr_gpa2hva(struct kvm_vm * vm,vm_paddr_t gpa) addr_gpa2hva() argument 1574 addr_hva2gpa(struct kvm_vm * vm,void * hva) addr_hva2gpa() argument 1617 addr_gpa2alias(struct kvm_vm * vm,vm_paddr_t gpa) addr_gpa2alias() argument 1634 vm_create_irqchip(struct kvm_vm * vm) vm_create_irqchip() argument 1741 __kvm_test_create_device(struct kvm_vm * vm,uint64_t type) __kvm_test_create_device() argument 1751 __kvm_create_device(struct kvm_vm * vm,uint64_t type) __kvm_create_device() argument 1793 _kvm_irq_line(struct kvm_vm * vm,uint32_t irq,int level) _kvm_irq_line() argument 1803 kvm_irq_line(struct kvm_vm * vm,uint32_t irq,int level) kvm_irq_line() argument 1841 _kvm_gsi_routing_write(struct kvm_vm * vm,struct kvm_irq_routing * routing) _kvm_gsi_routing_write() argument 1852 kvm_gsi_routing_write(struct kvm_vm * vm,struct kvm_irq_routing * routing) kvm_gsi_routing_write() argument 1875 vm_dump(FILE * stream,struct kvm_vm * vm,uint8_t indent) vm_dump() argument 2010 __vm_phy_pages_alloc(struct kvm_vm * vm,size_t num,vm_paddr_t paddr_min,uint32_t memslot,bool protected) __vm_phy_pages_alloc() argument 2056 vm_phy_page_alloc(struct kvm_vm * vm,vm_paddr_t paddr_min,uint32_t memslot) vm_phy_page_alloc() argument 2062 vm_alloc_page_table(struct kvm_vm * vm) vm_alloc_page_table() argument 2080 addr_gva2hva(struct kvm_vm * vm,vm_vaddr_t gva) addr_gva2hva() argument 2085 vm_compute_max_gfn(struct kvm_vm * vm) vm_compute_max_gfn() argument 2211 __vm_get_stat(struct kvm_vm * vm,const char * stat_name,uint64_t * data,size_t max_elements) __vm_get_stat() argument 2240 kvm_arch_vm_post_create(struct kvm_vm * vm) kvm_arch_vm_post_create() argument 2259 vm_is_gpa_protected(struct kvm_vm * vm,vm_paddr_t paddr) vm_is_gpa_protected() argument [all...] |