Loading drivers/kvm/kvm_main.c +14 −14 Original line number Original line Diff line number Diff line Loading @@ -62,7 +62,7 @@ static struct kvm_stats_debugfs_item { { "halt_exits", &kvm_stat.halt_exits }, { "halt_exits", &kvm_stat.halt_exits }, { "request_irq", &kvm_stat.request_irq_exits }, { "request_irq", &kvm_stat.request_irq_exits }, { "irq_exits", &kvm_stat.irq_exits }, { "irq_exits", &kvm_stat.irq_exits }, { 0, 0 } { NULL, NULL } }; }; static struct dentry *debugfs_dir; static struct dentry *debugfs_dir; Loading Loading @@ -205,7 +205,7 @@ static struct kvm_vcpu *vcpu_load(struct kvm *kvm, int vcpu_slot) mutex_lock(&vcpu->mutex); mutex_lock(&vcpu->mutex); if (unlikely(!vcpu->vmcs)) { if (unlikely(!vcpu->vmcs)) { mutex_unlock(&vcpu->mutex); mutex_unlock(&vcpu->mutex); return 0; return NULL; } } return kvm_arch_ops->vcpu_load(vcpu); return kvm_arch_ops->vcpu_load(vcpu); } } Loading Loading @@ -257,9 +257,9 @@ static void kvm_free_physmem_slot(struct kvm_memory_slot *free, if (!dont || free->dirty_bitmap != dont->dirty_bitmap) if (!dont || free->dirty_bitmap != dont->dirty_bitmap) vfree(free->dirty_bitmap); vfree(free->dirty_bitmap); free->phys_mem = 0; free->phys_mem = NULL; free->npages = 0; free->npages = 0; free->dirty_bitmap = 0; free->dirty_bitmap = NULL; } } static void kvm_free_physmem(struct kvm *kvm) static void kvm_free_physmem(struct kvm *kvm) Loading @@ -267,7 +267,7 @@ static void kvm_free_physmem(struct kvm *kvm) int i; int i; for (i = 0; i < kvm->nmemslots; ++i) for (i = 0; i < kvm->nmemslots; ++i) kvm_free_physmem_slot(&kvm->memslots[i], 0); kvm_free_physmem_slot(&kvm->memslots[i], NULL); } } static void kvm_free_vcpu(struct kvm_vcpu *vcpu) static void kvm_free_vcpu(struct kvm_vcpu *vcpu) Loading Loading @@ -640,11 +640,11 @@ static int kvm_dev_ioctl_set_memory_region(struct kvm *kvm, /* Deallocate if slot is being removed */ /* Deallocate if slot is being removed */ if (!npages) if (!npages) new.phys_mem = 0; new.phys_mem = NULL; /* Free page dirty bitmap if unneeded */ /* Free page dirty bitmap if unneeded */ if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES)) if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES)) new.dirty_bitmap = 0; new.dirty_bitmap = NULL; r = -ENOMEM; r = -ENOMEM; Loading Loading @@ -799,14 +799,14 @@ struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) && gfn < memslot->base_gfn + memslot->npages) && gfn < memslot->base_gfn + memslot->npages) return memslot; return memslot; } } return 0; return NULL; } } EXPORT_SYMBOL_GPL(gfn_to_memslot); EXPORT_SYMBOL_GPL(gfn_to_memslot); void mark_page_dirty(struct kvm *kvm, gfn_t gfn) void mark_page_dirty(struct kvm *kvm, gfn_t gfn) { { int i; int i; struct kvm_memory_slot *memslot = 0; struct kvm_memory_slot *memslot = NULL; unsigned long rel_gfn; unsigned long rel_gfn; for (i = 0; i < kvm->nmemslots; ++i) { for (i = 0; i < kvm->nmemslots; ++i) { Loading Loading @@ -2015,7 +2015,7 @@ static int kvm_reboot(struct notifier_block *notifier, unsigned long val, * in vmx root mode. * in vmx root mode. */ */ printk(KERN_INFO "kvm: exiting hardware virtualization\n"); printk(KERN_INFO "kvm: exiting hardware virtualization\n"); on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1); } } return NOTIFY_OK; return NOTIFY_OK; } } Loading @@ -2029,7 +2029,7 @@ static __init void kvm_init_debug(void) { { struct kvm_stats_debugfs_item *p; struct kvm_stats_debugfs_item *p; debugfs_dir = debugfs_create_dir("kvm", 0); debugfs_dir = debugfs_create_dir("kvm", NULL); for (p = debugfs_entries; p->name; ++p) for (p = debugfs_entries; p->name; ++p) p->dentry = debugfs_create_u32(p->name, 0444, debugfs_dir, p->dentry = debugfs_create_u32(p->name, 0444, debugfs_dir, p->data); p->data); Loading Loading @@ -2070,7 +2070,7 @@ int kvm_init_arch(struct kvm_arch_ops *ops, struct module *module) if (r < 0) if (r < 0) return r; return r; on_each_cpu(kvm_arch_ops->hardware_enable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_enable, NULL, 0, 1); register_reboot_notifier(&kvm_reboot_notifier); register_reboot_notifier(&kvm_reboot_notifier); kvm_chardev_ops.owner = module; kvm_chardev_ops.owner = module; Loading @@ -2085,7 +2085,7 @@ int kvm_init_arch(struct kvm_arch_ops *ops, struct module *module) out_free: out_free: unregister_reboot_notifier(&kvm_reboot_notifier); unregister_reboot_notifier(&kvm_reboot_notifier); on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1); kvm_arch_ops->hardware_unsetup(); kvm_arch_ops->hardware_unsetup(); return r; return r; } } Loading @@ -2095,7 +2095,7 @@ void kvm_exit_arch(void) misc_deregister(&kvm_dev); misc_deregister(&kvm_dev); unregister_reboot_notifier(&kvm_reboot_notifier); unregister_reboot_notifier(&kvm_reboot_notifier); on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1); kvm_arch_ops->hardware_unsetup(); kvm_arch_ops->hardware_unsetup(); kvm_arch_ops = NULL; kvm_arch_ops = NULL; } } Loading drivers/kvm/svm.c +4 −4 Original line number Original line Diff line number Diff line Loading @@ -274,7 +274,7 @@ static void svm_hardware_disable(void *garbage) wrmsrl(MSR_VM_HSAVE_PA, 0); wrmsrl(MSR_VM_HSAVE_PA, 0); rdmsrl(MSR_EFER, efer); rdmsrl(MSR_EFER, efer); wrmsrl(MSR_EFER, efer & ~MSR_EFER_SVME_MASK); wrmsrl(MSR_EFER, efer & ~MSR_EFER_SVME_MASK); per_cpu(svm_data, raw_smp_processor_id()) = 0; per_cpu(svm_data, raw_smp_processor_id()) = NULL; __free_page(svm_data->save_area); __free_page(svm_data->save_area); kfree(svm_data); kfree(svm_data); } } Loading Loading @@ -642,7 +642,7 @@ static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg) case VCPU_SREG_LDTR: return &save->ldtr; case VCPU_SREG_LDTR: return &save->ldtr; } } BUG(); BUG(); return 0; return NULL; } } static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg) static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg) Loading Loading @@ -934,7 +934,7 @@ static int io_get_override(struct kvm_vcpu *vcpu, return 0; return 0; *addr_override = 0; *addr_override = 0; *seg = 0; *seg = NULL; for (i = 0; i < ins_length; i++) for (i = 0; i < ins_length; i++) switch (inst[i]) { switch (inst[i]) { case 0xf0: case 0xf0: Loading Loading @@ -1087,7 +1087,7 @@ static int cpuid_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) { { if (emulate_instruction(vcpu, 0, 0, 0) != EMULATE_DONE) if (emulate_instruction(vcpu, NULL, 0, 0) != EMULATE_DONE) printk(KERN_ERR "%s: failed\n", __FUNCTION__); printk(KERN_ERR "%s: failed\n", __FUNCTION__); return 1; return 1; } } Loading drivers/kvm/vmx.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -98,7 +98,7 @@ static struct vmx_msr_entry *find_msr_entry(struct kvm_vcpu *vcpu, u32 msr) for (i = 0; i < vcpu->nmsrs; ++i) for (i = 0; i < vcpu->nmsrs; ++i) if (vcpu->guest_msrs[i].index == msr) if (vcpu->guest_msrs[i].index == msr) return &vcpu->guest_msrs[i]; return &vcpu->guest_msrs[i]; return 0; return NULL; } } static void vmcs_clear(struct vmcs *vmcs) static void vmcs_clear(struct vmcs *vmcs) Loading Loading
drivers/kvm/kvm_main.c +14 −14 Original line number Original line Diff line number Diff line Loading @@ -62,7 +62,7 @@ static struct kvm_stats_debugfs_item { { "halt_exits", &kvm_stat.halt_exits }, { "halt_exits", &kvm_stat.halt_exits }, { "request_irq", &kvm_stat.request_irq_exits }, { "request_irq", &kvm_stat.request_irq_exits }, { "irq_exits", &kvm_stat.irq_exits }, { "irq_exits", &kvm_stat.irq_exits }, { 0, 0 } { NULL, NULL } }; }; static struct dentry *debugfs_dir; static struct dentry *debugfs_dir; Loading Loading @@ -205,7 +205,7 @@ static struct kvm_vcpu *vcpu_load(struct kvm *kvm, int vcpu_slot) mutex_lock(&vcpu->mutex); mutex_lock(&vcpu->mutex); if (unlikely(!vcpu->vmcs)) { if (unlikely(!vcpu->vmcs)) { mutex_unlock(&vcpu->mutex); mutex_unlock(&vcpu->mutex); return 0; return NULL; } } return kvm_arch_ops->vcpu_load(vcpu); return kvm_arch_ops->vcpu_load(vcpu); } } Loading Loading @@ -257,9 +257,9 @@ static void kvm_free_physmem_slot(struct kvm_memory_slot *free, if (!dont || free->dirty_bitmap != dont->dirty_bitmap) if (!dont || free->dirty_bitmap != dont->dirty_bitmap) vfree(free->dirty_bitmap); vfree(free->dirty_bitmap); free->phys_mem = 0; free->phys_mem = NULL; free->npages = 0; free->npages = 0; free->dirty_bitmap = 0; free->dirty_bitmap = NULL; } } static void kvm_free_physmem(struct kvm *kvm) static void kvm_free_physmem(struct kvm *kvm) Loading @@ -267,7 +267,7 @@ static void kvm_free_physmem(struct kvm *kvm) int i; int i; for (i = 0; i < kvm->nmemslots; ++i) for (i = 0; i < kvm->nmemslots; ++i) kvm_free_physmem_slot(&kvm->memslots[i], 0); kvm_free_physmem_slot(&kvm->memslots[i], NULL); } } static void kvm_free_vcpu(struct kvm_vcpu *vcpu) static void kvm_free_vcpu(struct kvm_vcpu *vcpu) Loading Loading @@ -640,11 +640,11 @@ static int kvm_dev_ioctl_set_memory_region(struct kvm *kvm, /* Deallocate if slot is being removed */ /* Deallocate if slot is being removed */ if (!npages) if (!npages) new.phys_mem = 0; new.phys_mem = NULL; /* Free page dirty bitmap if unneeded */ /* Free page dirty bitmap if unneeded */ if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES)) if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES)) new.dirty_bitmap = 0; new.dirty_bitmap = NULL; r = -ENOMEM; r = -ENOMEM; Loading Loading @@ -799,14 +799,14 @@ struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) && gfn < memslot->base_gfn + memslot->npages) && gfn < memslot->base_gfn + memslot->npages) return memslot; return memslot; } } return 0; return NULL; } } EXPORT_SYMBOL_GPL(gfn_to_memslot); EXPORT_SYMBOL_GPL(gfn_to_memslot); void mark_page_dirty(struct kvm *kvm, gfn_t gfn) void mark_page_dirty(struct kvm *kvm, gfn_t gfn) { { int i; int i; struct kvm_memory_slot *memslot = 0; struct kvm_memory_slot *memslot = NULL; unsigned long rel_gfn; unsigned long rel_gfn; for (i = 0; i < kvm->nmemslots; ++i) { for (i = 0; i < kvm->nmemslots; ++i) { Loading Loading @@ -2015,7 +2015,7 @@ static int kvm_reboot(struct notifier_block *notifier, unsigned long val, * in vmx root mode. * in vmx root mode. */ */ printk(KERN_INFO "kvm: exiting hardware virtualization\n"); printk(KERN_INFO "kvm: exiting hardware virtualization\n"); on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1); } } return NOTIFY_OK; return NOTIFY_OK; } } Loading @@ -2029,7 +2029,7 @@ static __init void kvm_init_debug(void) { { struct kvm_stats_debugfs_item *p; struct kvm_stats_debugfs_item *p; debugfs_dir = debugfs_create_dir("kvm", 0); debugfs_dir = debugfs_create_dir("kvm", NULL); for (p = debugfs_entries; p->name; ++p) for (p = debugfs_entries; p->name; ++p) p->dentry = debugfs_create_u32(p->name, 0444, debugfs_dir, p->dentry = debugfs_create_u32(p->name, 0444, debugfs_dir, p->data); p->data); Loading Loading @@ -2070,7 +2070,7 @@ int kvm_init_arch(struct kvm_arch_ops *ops, struct module *module) if (r < 0) if (r < 0) return r; return r; on_each_cpu(kvm_arch_ops->hardware_enable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_enable, NULL, 0, 1); register_reboot_notifier(&kvm_reboot_notifier); register_reboot_notifier(&kvm_reboot_notifier); kvm_chardev_ops.owner = module; kvm_chardev_ops.owner = module; Loading @@ -2085,7 +2085,7 @@ int kvm_init_arch(struct kvm_arch_ops *ops, struct module *module) out_free: out_free: unregister_reboot_notifier(&kvm_reboot_notifier); unregister_reboot_notifier(&kvm_reboot_notifier); on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1); kvm_arch_ops->hardware_unsetup(); kvm_arch_ops->hardware_unsetup(); return r; return r; } } Loading @@ -2095,7 +2095,7 @@ void kvm_exit_arch(void) misc_deregister(&kvm_dev); misc_deregister(&kvm_dev); unregister_reboot_notifier(&kvm_reboot_notifier); unregister_reboot_notifier(&kvm_reboot_notifier); on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1); on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1); kvm_arch_ops->hardware_unsetup(); kvm_arch_ops->hardware_unsetup(); kvm_arch_ops = NULL; kvm_arch_ops = NULL; } } Loading
drivers/kvm/svm.c +4 −4 Original line number Original line Diff line number Diff line Loading @@ -274,7 +274,7 @@ static void svm_hardware_disable(void *garbage) wrmsrl(MSR_VM_HSAVE_PA, 0); wrmsrl(MSR_VM_HSAVE_PA, 0); rdmsrl(MSR_EFER, efer); rdmsrl(MSR_EFER, efer); wrmsrl(MSR_EFER, efer & ~MSR_EFER_SVME_MASK); wrmsrl(MSR_EFER, efer & ~MSR_EFER_SVME_MASK); per_cpu(svm_data, raw_smp_processor_id()) = 0; per_cpu(svm_data, raw_smp_processor_id()) = NULL; __free_page(svm_data->save_area); __free_page(svm_data->save_area); kfree(svm_data); kfree(svm_data); } } Loading Loading @@ -642,7 +642,7 @@ static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg) case VCPU_SREG_LDTR: return &save->ldtr; case VCPU_SREG_LDTR: return &save->ldtr; } } BUG(); BUG(); return 0; return NULL; } } static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg) static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg) Loading Loading @@ -934,7 +934,7 @@ static int io_get_override(struct kvm_vcpu *vcpu, return 0; return 0; *addr_override = 0; *addr_override = 0; *seg = 0; *seg = NULL; for (i = 0; i < ins_length; i++) for (i = 0; i < ins_length; i++) switch (inst[i]) { switch (inst[i]) { case 0xf0: case 0xf0: Loading Loading @@ -1087,7 +1087,7 @@ static int cpuid_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) { { if (emulate_instruction(vcpu, 0, 0, 0) != EMULATE_DONE) if (emulate_instruction(vcpu, NULL, 0, 0) != EMULATE_DONE) printk(KERN_ERR "%s: failed\n", __FUNCTION__); printk(KERN_ERR "%s: failed\n", __FUNCTION__); return 1; return 1; } } Loading
drivers/kvm/vmx.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -98,7 +98,7 @@ static struct vmx_msr_entry *find_msr_entry(struct kvm_vcpu *vcpu, u32 msr) for (i = 0; i < vcpu->nmsrs; ++i) for (i = 0; i < vcpu->nmsrs; ++i) if (vcpu->guest_msrs[i].index == msr) if (vcpu->guest_msrs[i].index == msr) return &vcpu->guest_msrs[i]; return &vcpu->guest_msrs[i]; return 0; return NULL; } } static void vmcs_clear(struct vmcs *vmcs) static void vmcs_clear(struct vmcs *vmcs) Loading