cregit-Linux how code gets into the kernel

Release 4.8 arch/arm/kvm/emulate.c

Directory: arch/arm/kvm
/*
 * Copyright (C) 2012 - Virtual Open Systems and Columbia University
 * Author: Christoffer Dall <c.dall@virtualopensystems.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License, version 2, as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
 */

#include <linux/mm.h>
#include <linux/kvm_host.h>
#include <asm/kvm_arm.h>
#include <asm/kvm_emulate.h>
#include <asm/opcodes.h>
#include <trace/events/kvm.h>

#include "trace.h"


#define VCPU_NR_MODES		6

#define VCPU_REG_OFFSET_USR	0

#define VCPU_REG_OFFSET_FIQ	1

#define VCPU_REG_OFFSET_IRQ	2

#define VCPU_REG_OFFSET_SVC	3

#define VCPU_REG_OFFSET_ABT	4

#define VCPU_REG_OFFSET_UND	5

#define REG_OFFSET(_reg) \
	(offsetof(struct kvm_regs, _reg) / sizeof(u32))


#define USR_REG_OFFSET(_num) REG_OFFSET(usr_regs.uregs[_num])


static const unsigned long vcpu_reg_offsets[VCPU_NR_MODES][15] = {
	/* USR/SYS Registers */
	[VCPU_REG_OFFSET_USR] = {
		USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
		USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
		USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
		USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
		USR_REG_OFFSET(12), USR_REG_OFFSET(13),	USR_REG_OFFSET(14),
        },

	/* FIQ Registers */
	[VCPU_REG_OFFSET_FIQ] = {
		USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
		USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
		USR_REG_OFFSET(6), USR_REG_OFFSET(7),
		REG_OFFSET(fiq_regs[0]), /* r8 */
		REG_OFFSET(fiq_regs[1]), /* r9 */
		REG_OFFSET(fiq_regs[2]), /* r10 */
		REG_OFFSET(fiq_regs[3]), /* r11 */
		REG_OFFSET(fiq_regs[4]), /* r12 */
		REG_OFFSET(fiq_regs[5]), /* r13 */
		REG_OFFSET(fiq_regs[6]), /* r14 */
	},

	/* IRQ Registers */
	[VCPU_REG_OFFSET_IRQ] = {
		USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
		USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
		USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
		USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
		USR_REG_OFFSET(12),
		REG_OFFSET(irq_regs[0]), /* r13 */
		REG_OFFSET(irq_regs[1]), /* r14 */
	},

	/* SVC Registers */
	[VCPU_REG_OFFSET_SVC] = {
		USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
		USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
		USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
		USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
		USR_REG_OFFSET(12),
		REG_OFFSET(svc_regs[0]), /* r13 */
		REG_OFFSET(svc_regs[1]), /* r14 */
	},

	/* ABT Registers */
	[VCPU_REG_OFFSET_ABT] = {
		USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
		USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
		USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
		USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
		USR_REG_OFFSET(12),
		REG_OFFSET(abt_regs[0]), /* r13 */
		REG_OFFSET(abt_regs[1]), /* r14 */
	},

	/* UND Registers */
	[VCPU_REG_OFFSET_UND] = {
		USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
		USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
		USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
		USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
		USR_REG_OFFSET(12),
		REG_OFFSET(und_regs[0]), /* r13 */
		REG_OFFSET(und_regs[1]), /* r14 */
	},
};

/*
 * Return a pointer to the register number valid in the current mode of
 * the virtual CPU.
 */

unsigned long *vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num) { unsigned long *reg_array = (unsigned long *)&vcpu->arch.ctxt.gp_regs; unsigned long mode = *vcpu_cpsr(vcpu) & MODE_MASK; switch (mode) { case USR_MODE...SVC_MODE: mode &= ~MODE32_BIT; /* 0 ... 3 */ break; case ABT_MODE: mode = VCPU_REG_OFFSET_ABT; break; case UND_MODE: mode = VCPU_REG_OFFSET_UND; break; case SYSTEM_MODE: mode = VCPU_REG_OFFSET_USR; break; default: BUG(); } return reg_array + vcpu_reg_offsets[mode][reg_num]; }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall9289.32%133.33%
marc zyngiermarc zyngier1110.68%266.67%
Total103100.00%3100.00%

/* * Return the SPSR for the current mode of the virtual CPU. */
unsigned long *vcpu_spsr(struct kvm_vcpu *vcpu) { unsigned long mode = *vcpu_cpsr(vcpu) & MODE_MASK; switch (mode) { case SVC_MODE: return &vcpu->arch.ctxt.gp_regs.KVM_ARM_SVC_spsr; case ABT_MODE: return &vcpu->arch.ctxt.gp_regs.KVM_ARM_ABT_spsr; case UND_MODE: return &vcpu->arch.ctxt.gp_regs.KVM_ARM_UND_spsr; case IRQ_MODE: return &vcpu->arch.ctxt.gp_regs.KVM_ARM_IRQ_spsr; case FIQ_MODE: return &vcpu->arch.ctxt.gp_regs.KVM_ARM_FIQ_spsr; default: BUG(); } }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall9082.57%133.33%
marc zyngiermarc zyngier1917.43%266.67%
Total109100.00%3100.00%

/* * A conditional instruction is allowed to trap, even though it * wouldn't be executed. So let's re-implement the hardware, in * software! */
bool kvm_condition_valid(struct kvm_vcpu *vcpu) { unsigned long cpsr, cond, insn; /* * Exception Code 0 can only happen if we set HCR.TGE to 1, to * catch undefined instructions, and then we won't get past * the arm_exit_handlers test anyway. */ BUG_ON(!kvm_vcpu_trap_get_class(vcpu)); /* Top two bits non-zero? Unconditional. */ if (kvm_vcpu_get_hsr(vcpu) >> 30) return true; cpsr = *vcpu_cpsr(vcpu); /* Is condition field valid? */ if ((kvm_vcpu_get_hsr(vcpu) & HSR_CV) >> HSR_CV_SHIFT) cond = (kvm_vcpu_get_hsr(vcpu) & HSR_COND) >> HSR_COND_SHIFT; else { /* This can happen in Thumb mode: examine IT state. */ unsigned long it; it = ((cpsr >> 8) & 0xFC) | ((cpsr >> 25) & 0x3); /* it == 0 => unconditional. */ if (it == 0) return true; /* The cond for this insn works out as the top 4 bits. */ cond = (it >> 4); } /* Shift makes it look like an ARM-mode instruction */ insn = cond << 28; return arm_check_condition(insn, cpsr) != ARM_OPCODE_CONDTEST_FAIL; }

Contributors

PersonTokensPropCommitsCommitProp
marc zyngiermarc zyngier142100.00%1100.00%
Total142100.00%1100.00%

/** * adjust_itstate - adjust ITSTATE when emulating instructions in IT-block * @vcpu: The VCPU pointer * * When exceptions occur while instructions are executed in Thumb IF-THEN * blocks, the ITSTATE field of the CPSR is not advanced (updated), so we have * to do this little bit of work manually. The fields map like this: * * IT[7:0] -> CPSR[26:25],CPSR[15:10] */
static void kvm_adjust_itstate(struct kvm_vcpu *vcpu) { unsigned long itbits, cond; unsigned long cpsr = *vcpu_cpsr(vcpu); bool is_arm = !(cpsr & PSR_T_BIT); BUG_ON(is_arm && (cpsr & PSR_IT_MASK)); if (!(cpsr & PSR_IT_MASK)) return; cond = (cpsr & 0xe000) >> 13; itbits = (cpsr & 0x1c00) >> (10 - 2); itbits |= (cpsr & (0x3 << 25)) >> 25; /* Perform ITAdvance (see page A-52 in ARM DDI 0406C) */ if ((itbits & 0x7) == 0) itbits = cond = 0; else itbits = (itbits << 1) & 0x1f; cpsr &= ~PSR_IT_MASK; cpsr |= cond << 13; cpsr |= (itbits & 0x1c) << (10 - 2); cpsr |= (itbits & 0x3) << 25; *vcpu_cpsr(vcpu) = cpsr; }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall167100.00%1100.00%
Total167100.00%1100.00%

/** * kvm_skip_instr - skip a trapped instruction and proceed to the next * @vcpu: The vcpu pointer */
void kvm_skip_instr(struct kvm_vcpu *vcpu, bool is_wide_instr) { bool is_thumb; is_thumb = !!(*vcpu_cpsr(vcpu) & PSR_T_BIT); if (is_thumb && !is_wide_instr) *vcpu_pc(vcpu) += 2; else *vcpu_pc(vcpu) += 4; kvm_adjust_itstate(vcpu); }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall59100.00%1100.00%
Total59100.00%1100.00%

/****************************************************************************** * Inject exceptions into the guest */
static u32 exc_vector_base(struct kvm_vcpu *vcpu) { u32 sctlr = vcpu_cp15(vcpu, c1_SCTLR); u32 vbar = vcpu_cp15(vcpu, c12_VBAR); if (sctlr & SCTLR_V) return 0xffff0000; else /* always have security exceptions */ return vbar; }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall3782.22%150.00%
marc zyngiermarc zyngier817.78%150.00%
Total45100.00%2100.00%

/* * Switch to an exception mode, updating both CPSR and SPSR. Follow * the logic described in AArch32.EnterMode() from the ARMv8 ARM. */
static void kvm_update_psr(struct kvm_vcpu *vcpu, unsigned long mode) { unsigned long cpsr = *vcpu_cpsr(vcpu); u32 sctlr = vcpu_cp15(vcpu, c1_SCTLR); *vcpu_cpsr(vcpu) = (cpsr & ~MODE_MASK) | mode; switch (mode) { case FIQ_MODE: *vcpu_cpsr(vcpu) |= PSR_F_BIT; /* Fall through */ case ABT_MODE: case IRQ_MODE: *vcpu_cpsr(vcpu) |= PSR_A_BIT; /* Fall through */ default: *vcpu_cpsr(vcpu) |= PSR_I_BIT; } *vcpu_cpsr(vcpu) &= ~(PSR_IT_MASK | PSR_J_BIT | PSR_E_BIT | PSR_T_BIT); if (sctlr & SCTLR_TE) *vcpu_cpsr(vcpu) |= PSR_T_BIT; if (sctlr & SCTLR_EE) *vcpu_cpsr(vcpu) |= PSR_E_BIT; /* Note: These now point to the mode banked copies */ *vcpu_spsr(vcpu) = cpsr; }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall9766.44%125.00%
marc zyngiermarc zyngier4933.56%375.00%
Total146100.00%4100.00%

/** * kvm_inject_undefined - inject an undefined exception into the guest * @vcpu: The VCPU to receive the undefined exception * * It is assumed that this code is called from the VCPU thread and that the * VCPU therefore is not currently executing guest code. * * Modelled after TakeUndefInstrException() pseudocode. */
void kvm_inject_undefined(struct kvm_vcpu *vcpu) { unsigned long cpsr = *vcpu_cpsr(vcpu); bool is_thumb = (cpsr & PSR_T_BIT); u32 vect_offset = 4; u32 return_offset = (is_thumb) ? 2 : 4; kvm_update_psr(vcpu, UND_MODE); *vcpu_reg(vcpu, 14) = *vcpu_pc(vcpu) - return_offset; /* Branch to exception vector */ *vcpu_pc(vcpu) = exc_vector_base(vcpu) + vect_offset; }

Contributors

PersonTokensPropCommitsCommitProp
marc zyngiermarc zyngier5870.73%150.00%
christoffer dallchristoffer dall2429.27%150.00%
Total82100.00%2100.00%

/* * Modelled after TakeDataAbortException() and TakePrefetchAbortException * pseudocode. */
static void inject_abt(struct kvm_vcpu *vcpu, bool is_pabt, unsigned long addr) { unsigned long cpsr = *vcpu_cpsr(vcpu); bool is_thumb = (cpsr & PSR_T_BIT); u32 vect_offset; u32 return_offset = (is_thumb) ? 4 : 0; bool is_lpae; kvm_update_psr(vcpu, ABT_MODE); *vcpu_reg(vcpu, 14) = *vcpu_pc(vcpu) + return_offset; if (is_pabt) vect_offset = 12; else vect_offset = 16; /* Branch to exception vector */ *vcpu_pc(vcpu) = exc_vector_base(vcpu) + vect_offset; if (is_pabt) { /* Set IFAR and IFSR */ vcpu_cp15(vcpu, c6_IFAR) = addr; is_lpae = (vcpu_cp15(vcpu, c2_TTBCR) >> 31); /* Always give debug fault for now - should give guest a clue */ if (is_lpae) vcpu_cp15(vcpu, c5_IFSR) = 1 << 9 | 0x22; else vcpu_cp15(vcpu, c5_IFSR) = 2; } else { /* !iabt */ /* Set DFAR and DFSR */ vcpu_cp15(vcpu, c6_DFAR) = addr; is_lpae = (vcpu_cp15(vcpu, c2_TTBCR) >> 31); /* Always give debug fault for now - should give guest a clue */ if (is_lpae) vcpu_cp15(vcpu, c5_DFSR) = 1 << 9 | 0x22; else vcpu_cp15(vcpu, c5_DFSR) = 2; } }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall17380.09%120.00%
marc zyngiermarc zyngier4219.44%360.00%
anup patelanup patel10.46%120.00%
Total216100.00%5100.00%

/** * kvm_inject_dabt - inject a data abort into the guest * @vcpu: The VCPU to receive the undefined exception * @addr: The address to report in the DFAR * * It is assumed that this code is called from the VCPU thread and that the * VCPU therefore is not currently executing guest code. */
void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr) { inject_abt(vcpu, false, addr); }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall23100.00%1100.00%
Total23100.00%1100.00%

/** * kvm_inject_pabt - inject a prefetch abort into the guest * @vcpu: The VCPU to receive the undefined exception * @addr: The address to report in the DFAR * * It is assumed that this code is called from the VCPU thread and that the * VCPU therefore is not currently executing guest code. */
void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr) { inject_abt(vcpu, true, addr); }

Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall23100.00%1100.00%
Total23100.00%1100.00%


Overall Contributors

PersonTokensPropCommitsCommitProp
christoffer dallchristoffer dall142580.87%222.22%
marc zyngiermarc zyngier33519.01%555.56%
anup patelanup patel10.06%111.11%
andrea gelminiandrea gelmini10.06%111.11%
Total1762100.00%9100.00%
Directory: arch/arm/kvm
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.