Loading arch/arm/kvm/arm.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -376,7 +376,7 @@ void force_vm_exit(const cpumask_t *mask) /** /** * need_new_vmid_gen - check that the VMID is still valid * need_new_vmid_gen - check that the VMID is still valid * @kvm: The VM's VMID to checkt * @kvm: The VM's VMID to check * * * return true if there is a new generation of VMIDs being used * return true if there is a new generation of VMIDs being used * * Loading arch/arm/kvm/emulate.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -210,7 +210,7 @@ bool kvm_condition_valid(struct kvm_vcpu *vcpu) * @vcpu: The VCPU pointer * @vcpu: The VCPU pointer * * * When exceptions occur while instructions are executed in Thumb IF-THEN * When exceptions occur while instructions are executed in Thumb IF-THEN * blocks, the ITSTATE field of the CPSR is not advanved (updated), so we have * blocks, the ITSTATE field of the CPSR is not advanced (updated), so we have * to do this little bit of work manually. The fields map like this: * to do this little bit of work manually. The fields map like this: * * * IT[7:0] -> CPSR[26:25],CPSR[15:10] * IT[7:0] -> CPSR[26:25],CPSR[15:10] Loading arch/arm/kvm/guest.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -182,7 +182,7 @@ unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu) /** /** * kvm_arm_copy_reg_indices - get indices of all registers. * kvm_arm_copy_reg_indices - get indices of all registers. * * * We do core registers right here, then we apppend coproc regs. * We do core registers right here, then we append coproc regs. */ */ int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) { { Loading arch/arm/kvm/reset.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -52,7 +52,7 @@ static const struct kvm_irq_level cortexa_vtimer_irq = { * @vcpu: The VCPU pointer * @vcpu: The VCPU pointer * * * This function finds the right table above and sets the registers on the * This function finds the right table above and sets the registers on the * virtual CPU struct to their architectually defined reset values. * virtual CPU struct to their architecturally defined reset values. */ */ int kvm_reset_vcpu(struct kvm_vcpu *vcpu) int kvm_reset_vcpu(struct kvm_vcpu *vcpu) { { Loading Loading
arch/arm/kvm/arm.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -376,7 +376,7 @@ void force_vm_exit(const cpumask_t *mask) /** /** * need_new_vmid_gen - check that the VMID is still valid * need_new_vmid_gen - check that the VMID is still valid * @kvm: The VM's VMID to checkt * @kvm: The VM's VMID to check * * * return true if there is a new generation of VMIDs being used * return true if there is a new generation of VMIDs being used * * Loading
arch/arm/kvm/emulate.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -210,7 +210,7 @@ bool kvm_condition_valid(struct kvm_vcpu *vcpu) * @vcpu: The VCPU pointer * @vcpu: The VCPU pointer * * * When exceptions occur while instructions are executed in Thumb IF-THEN * When exceptions occur while instructions are executed in Thumb IF-THEN * blocks, the ITSTATE field of the CPSR is not advanved (updated), so we have * blocks, the ITSTATE field of the CPSR is not advanced (updated), so we have * to do this little bit of work manually. The fields map like this: * to do this little bit of work manually. The fields map like this: * * * IT[7:0] -> CPSR[26:25],CPSR[15:10] * IT[7:0] -> CPSR[26:25],CPSR[15:10] Loading
arch/arm/kvm/guest.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -182,7 +182,7 @@ unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu) /** /** * kvm_arm_copy_reg_indices - get indices of all registers. * kvm_arm_copy_reg_indices - get indices of all registers. * * * We do core registers right here, then we apppend coproc regs. * We do core registers right here, then we append coproc regs. */ */ int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) { { Loading
arch/arm/kvm/reset.c +1 −1 Original line number Original line Diff line number Diff line Loading @@ -52,7 +52,7 @@ static const struct kvm_irq_level cortexa_vtimer_irq = { * @vcpu: The VCPU pointer * @vcpu: The VCPU pointer * * * This function finds the right table above and sets the registers on the * This function finds the right table above and sets the registers on the * virtual CPU struct to their architectually defined reset values. * virtual CPU struct to their architecturally defined reset values. */ */ int kvm_reset_vcpu(struct kvm_vcpu *vcpu) int kvm_reset_vcpu(struct kvm_vcpu *vcpu) { { Loading