int kvm_init(int smp_cpus);
int kvm_init_vcpu(CPUState *env);
-int kvm_sync_vcpus(void);
int kvm_cpu_exec(CPUState *env);
ram_addr_t size,
ram_addr_t phys_offset);
-void kvm_physical_sync_dirty_bitmap(target_phys_addr_t start_addr, target_phys_addr_t end_addr);
+int kvm_physical_sync_dirty_bitmap(target_phys_addr_t start_addr,
+ target_phys_addr_t end_addr);
-int kvm_log_start(target_phys_addr_t phys_addr, target_phys_addr_t len);
-int kvm_log_stop(target_phys_addr_t phys_addr, target_phys_addr_t len);
+int kvm_log_start(target_phys_addr_t phys_addr, ram_addr_t size);
+int kvm_log_stop(target_phys_addr_t phys_addr, ram_addr_t size);
+int kvm_set_migration_log(int enable);
int kvm_has_sync_mmu(void);
+void kvm_setup_guest_memory(void *start, size_t size);
+
int kvm_coalesce_mmio_region(target_phys_addr_t start, ram_addr_t size);
int kvm_uncoalesce_mmio_region(target_phys_addr_t start, ram_addr_t size);
void kvm_remove_all_breakpoints(CPUState *current_env);
int kvm_update_guest_debug(CPUState *env, unsigned long reinject_trap);
+int kvm_pit_in_kernel(void);
+int kvm_irqchip_in_kernel(void);
+
/* internal API */
struct KVMState;
int kvm_vcpu_ioctl(CPUState *env, int type, ...);
+int kvm_get_mp_state(CPUState *env);
+int kvm_put_mp_state(CPUState *env);
+
/* Arch specific hooks */
int kvm_arch_post_run(CPUState *env, struct kvm_run *run);
void kvm_arch_update_guest_debug(CPUState *env, struct kvm_guest_debug *dbg);
+int kvm_check_extension(KVMState *s, unsigned int extension);
+
+uint32_t kvm_arch_get_supported_cpuid(CPUState *env, uint32_t function,
+ int reg);
+void kvm_cpu_synchronize_state(CPUState *env);
+
/* generic hooks - to be moved/refactored once there are more users */
-static inline void cpu_synchronize_state(CPUState *env, int modified)
+static inline void cpu_synchronize_state(CPUState *env)
{
if (kvm_enabled()) {
- if (modified)
- kvm_arch_put_registers(env);
- else
- kvm_arch_get_registers(env);
+ kvm_cpu_synchronize_state(env);
}
}