Lines Matching refs:kvm_vm
63 struct kvm_vm *vm;
99 struct kvm_vm { struct
167 memslot2region(struct kvm_vm *vm, uint32_t memslot);
169 static inline struct userspace_mem_region *vm_get_mem_region(struct kvm_vm *vm, in vm_get_mem_region()
336 static __always_inline void static_assert_is_vm(struct kvm_vm *vm) { } in static_assert_is_vm()
399 static inline int vm_check_cap(struct kvm_vm *vm, long cap) in vm_check_cap()
407 static inline int __vm_enable_cap(struct kvm_vm *vm, uint32_t cap, uint64_t arg0) in __vm_enable_cap()
413 static inline void vm_enable_cap(struct kvm_vm *vm, uint32_t cap, uint64_t arg0) in vm_enable_cap()
420 static inline void vm_set_memory_attributes(struct kvm_vm *vm, uint64_t gpa, in vm_set_memory_attributes()
441 static inline void vm_mem_set_private(struct kvm_vm *vm, uint64_t gpa, in vm_mem_set_private()
447 static inline void vm_mem_set_shared(struct kvm_vm *vm, uint64_t gpa, in vm_mem_set_shared()
453 void vm_guest_mem_fallocate(struct kvm_vm *vm, uint64_t gpa, uint64_t size,
456 static inline void vm_guest_mem_punch_hole(struct kvm_vm *vm, uint64_t gpa, in vm_guest_mem_punch_hole()
462 static inline void vm_guest_mem_allocate(struct kvm_vm *vm, uint64_t gpa, in vm_guest_mem_allocate()
468 void vm_enable_dirty_ring(struct kvm_vm *vm, uint32_t ring_size);
471 void kvm_vm_free(struct kvm_vm *vmp);
472 void kvm_vm_restart(struct kvm_vm *vmp);
473 void kvm_vm_release(struct kvm_vm *vmp);
474 void kvm_vm_elf_load(struct kvm_vm *vm, const char *filename);
477 void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent);
479 static inline void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log) in kvm_vm_get_dirty_log()
486 static inline void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log, in kvm_vm_clear_dirty_log()
499 static inline uint32_t kvm_vm_reset_dirty_ring(struct kvm_vm *vm) in kvm_vm_reset_dirty_ring()
504 static inline void kvm_vm_register_coalesced_io(struct kvm_vm *vm, in kvm_vm_register_coalesced_io()
517 static inline void kvm_vm_unregister_coalesced_io(struct kvm_vm *vm, in kvm_vm_unregister_coalesced_io()
530 static inline int vm_get_stats_fd(struct kvm_vm *vm) in vm_get_stats_fd()
538 static inline int __kvm_irqfd(struct kvm_vm *vm, uint32_t gsi, int eventfd, in __kvm_irqfd()
551 static inline void kvm_irqfd(struct kvm_vm *vm, uint32_t gsi, int eventfd, in kvm_irqfd()
559 static inline void kvm_assign_irqfd(struct kvm_vm *vm, uint32_t gsi, int eventfd) in kvm_assign_irqfd()
564 static inline void kvm_deassign_irqfd(struct kvm_vm *vm, uint32_t gsi, int eventfd) in kvm_deassign_irqfd()
665 void vm_create_irqchip(struct kvm_vm *vm);
667 static inline int __vm_create_guest_memfd(struct kvm_vm *vm, uint64_t size, in __vm_create_guest_memfd()
678 static inline int vm_create_guest_memfd(struct kvm_vm *vm, uint64_t size, in vm_create_guest_memfd()
687 void vm_set_user_memory_region(struct kvm_vm *vm, uint32_t slot, uint32_t flags,
689 int __vm_set_user_memory_region(struct kvm_vm *vm, uint32_t slot, uint32_t flags,
691 void vm_set_user_memory_region2(struct kvm_vm *vm, uint32_t slot, uint32_t flags,
694 int __vm_set_user_memory_region2(struct kvm_vm *vm, uint32_t slot, uint32_t flags,
698 void vm_userspace_mem_region_add(struct kvm_vm *vm,
702 void vm_mem_add(struct kvm_vm *vm, enum vm_mem_backing_src_type src_type,
707 static inline bool vm_arch_has_protected_memory(struct kvm_vm *vm) in vm_arch_has_protected_memory()
713 void vm_mem_region_set_flags(struct kvm_vm *vm, uint32_t slot, uint32_t flags);
714 void vm_mem_region_reload(struct kvm_vm *vm, uint32_t slot);
715 void vm_mem_region_move(struct kvm_vm *vm, uint32_t slot, uint64_t new_gpa);
716 void vm_mem_region_delete(struct kvm_vm *vm, uint32_t slot);
717 struct kvm_vcpu *__vm_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id);
718 void vm_populate_vaddr_bitmap(struct kvm_vm *vm);
719 vm_vaddr_t vm_vaddr_unused_gap(struct kvm_vm *vm, size_t sz, vm_vaddr_t vaddr_min);
720 vm_vaddr_t vm_vaddr_alloc(struct kvm_vm *vm, size_t sz, vm_vaddr_t vaddr_min);
721 vm_vaddr_t __vm_vaddr_alloc(struct kvm_vm *vm, size_t sz, vm_vaddr_t vaddr_min,
723 vm_vaddr_t vm_vaddr_alloc_shared(struct kvm_vm *vm, size_t sz,
726 vm_vaddr_t vm_vaddr_alloc_pages(struct kvm_vm *vm, int nr_pages);
727 vm_vaddr_t __vm_vaddr_alloc_page(struct kvm_vm *vm,
729 vm_vaddr_t vm_vaddr_alloc_page(struct kvm_vm *vm);
731 void virt_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr,
733 void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa);
734 void *addr_gva2hva(struct kvm_vm *vm, vm_vaddr_t gva);
735 vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva);
736 void *addr_gpa2alias(struct kvm_vm *vm, vm_paddr_t gpa);
742 static inline vm_paddr_t vm_untag_gpa(struct kvm_vm *vm, vm_paddr_t gpa) in vm_untag_gpa()
948 int __kvm_test_create_device(struct kvm_vm *vm, uint64_t type);
949 int __kvm_create_device(struct kvm_vm *vm, uint64_t type);
951 static inline int kvm_create_device(struct kvm_vm *vm, uint64_t type) in kvm_create_device()
980 void kvm_irq_line(struct kvm_vm *vm, uint32_t irq, int level);
981 int _kvm_irq_line(struct kvm_vm *vm, uint32_t irq, int level);
988 int _kvm_gsi_routing_write(struct kvm_vm *vm, struct kvm_irq_routing *routing);
989 void kvm_gsi_routing_write(struct kvm_vm *vm, struct kvm_irq_routing *routing);
993 vm_paddr_t vm_phy_page_alloc(struct kvm_vm *vm, vm_paddr_t paddr_min,
995 vm_paddr_t __vm_phy_pages_alloc(struct kvm_vm *vm, size_t num,
998 vm_paddr_t vm_alloc_page_table(struct kvm_vm *vm);
1000 static inline vm_paddr_t vm_phy_pages_alloc(struct kvm_vm *vm, size_t num, in vm_phy_pages_alloc()
1018 struct kvm_vm *____vm_create(struct vm_shape shape);
1019 struct kvm_vm *__vm_create(struct vm_shape shape, uint32_t nr_runnable_vcpus,
1022 static inline struct kvm_vm *vm_create_barebones(void) in vm_create_barebones()
1027 static inline struct kvm_vm *vm_create_barebones_type(unsigned long type) in vm_create_barebones_type()
1037 static inline struct kvm_vm *vm_create(uint32_t nr_runnable_vcpus) in vm_create()
1042 struct kvm_vm *__vm_create_with_vcpus(struct vm_shape shape, uint32_t nr_vcpus,
1046 static inline struct kvm_vm *vm_create_with_vcpus(uint32_t nr_vcpus, in vm_create_with_vcpus()
1055 struct kvm_vm *__vm_create_shape_with_one_vcpu(struct vm_shape shape,
1064 static inline struct kvm_vm *__vm_create_with_one_vcpu(struct kvm_vcpu **vcpu, in __vm_create_with_one_vcpu()
1072 static inline struct kvm_vm *vm_create_with_one_vcpu(struct kvm_vcpu **vcpu, in vm_create_with_one_vcpu()
1078 static inline struct kvm_vm *vm_create_shape_with_one_vcpu(struct vm_shape shape, in vm_create_shape_with_one_vcpu()
1085 struct kvm_vcpu *vm_recreate_with_one_vcpu(struct kvm_vm *vm);
1121 unsigned long vm_compute_max_gfn(struct kvm_vm *vm);
1178 struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id);
1181 static inline struct kvm_vcpu *vm_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, in vm_vcpu_add()
1192 struct kvm_vcpu *vm_arch_vcpu_recreate(struct kvm_vm *vm, uint32_t vcpu_id);
1194 static inline struct kvm_vcpu *vm_vcpu_recreate(struct kvm_vm *vm, in vm_vcpu_recreate()
1202 void virt_arch_pgd_alloc(struct kvm_vm *vm);
1204 static inline void virt_pgd_alloc(struct kvm_vm *vm) in virt_pgd_alloc()
1225 void virt_arch_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr);
1227 static inline void virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr) in virt_pg_map()
1249 vm_paddr_t addr_arch_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva);
1251 static inline vm_paddr_t addr_gva2gpa(struct kvm_vm *vm, vm_vaddr_t gva) in addr_gva2gpa()
1271 void virt_arch_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent);
1273 static inline void virt_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) in virt_dump()
1279 static inline int __vm_disable_nx_huge_pages(struct kvm_vm *vm) in __vm_disable_nx_huge_pages()
1284 static inline uint64_t vm_page_align(struct kvm_vm *vm, uint64_t v) in vm_page_align()
1296 void kvm_arch_vm_post_create(struct kvm_vm *vm, unsigned int nr_vcpus);
1297 void kvm_arch_vm_finalize_vcpus(struct kvm_vm *vm);
1298 void kvm_arch_vm_release(struct kvm_vm *vm);
1300 bool vm_is_gpa_protected(struct kvm_vm *vm, vm_paddr_t paddr);