1 /*
2 * Debug and Guest Debug support
3 *
4 * Copyright (C) 2015 - Linaro Ltd
5 * Author: Alex Bennée <alex.bennee@linaro.org>
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program. If not, see <http://www.gnu.org/licenses/>.
18 */
19
20 #include <linux/kvm_host.h>
21 #include <linux/hw_breakpoint.h>
22
23 #include <asm/debug-monitors.h>
24 #include <asm/kvm_asm.h>
25 #include <asm/kvm_arm.h>
26 #include <asm/kvm_emulate.h>
27
28 #include "trace.h"
29
30 /* These are the bits of MDSCR_EL1 we may manipulate */
31 #define MDSCR_EL1_DEBUG_MASK (DBG_MDSCR_SS | \
32 DBG_MDSCR_KDE | \
33 DBG_MDSCR_MDE)
34
35 static DEFINE_PER_CPU(u32, mdcr_el2);
36
37 /**
38 * save/restore_guest_debug_regs
39 *
40 * For some debug operations we need to tweak some guest registers. As
41 * a result we need to save the state of those registers before we
42 * make those modifications.
43 *
44 * Guest access to MDSCR_EL1 is trapped by the hypervisor and handled
45 * after we have restored the preserved value to the main context.
46 */
save_guest_debug_regs(struct kvm_vcpu * vcpu)47 static void save_guest_debug_regs(struct kvm_vcpu *vcpu)
48 {
49 u64 val = vcpu_read_sys_reg(vcpu, MDSCR_EL1);
50
51 vcpu->arch.guest_debug_preserved.mdscr_el1 = val;
52
53 trace_kvm_arm_set_dreg32("Saved MDSCR_EL1",
54 vcpu->arch.guest_debug_preserved.mdscr_el1);
55 }
56
restore_guest_debug_regs(struct kvm_vcpu * vcpu)57 static void restore_guest_debug_regs(struct kvm_vcpu *vcpu)
58 {
59 u64 val = vcpu->arch.guest_debug_preserved.mdscr_el1;
60
61 vcpu_write_sys_reg(vcpu, val, MDSCR_EL1);
62
63 trace_kvm_arm_set_dreg32("Restored MDSCR_EL1",
64 vcpu_read_sys_reg(vcpu, MDSCR_EL1));
65 }
66
67 /**
68 * kvm_arm_init_debug - grab what we need for debug
69 *
70 * Currently the sole task of this function is to retrieve the initial
71 * value of mdcr_el2 so we can preserve MDCR_EL2.HPMN which has
72 * presumably been set-up by some knowledgeable bootcode.
73 *
74 * It is called once per-cpu during CPU hyp initialisation.
75 */
76
kvm_arm_init_debug(void)77 void kvm_arm_init_debug(void)
78 {
79 __this_cpu_write(mdcr_el2, kvm_call_hyp(__kvm_get_mdcr_el2));
80 }
81
82 /**
83 * kvm_arm_reset_debug_ptr - reset the debug ptr to point to the vcpu state
84 */
85
kvm_arm_reset_debug_ptr(struct kvm_vcpu * vcpu)86 void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu)
87 {
88 vcpu->arch.debug_ptr = &vcpu->arch.vcpu_debug_state;
89 }
90
91 /**
92 * kvm_arm_setup_debug - set up debug related stuff
93 *
94 * @vcpu: the vcpu pointer
95 *
96 * This is called before each entry into the hypervisor to setup any
97 * debug related registers. Currently this just ensures we will trap
98 * access to:
99 * - Performance monitors (MDCR_EL2_TPM/MDCR_EL2_TPMCR)
100 * - Debug ROM Address (MDCR_EL2_TDRA)
101 * - OS related registers (MDCR_EL2_TDOSA)
102 * - Statistical profiler (MDCR_EL2_TPMS/MDCR_EL2_E2PB)
103 *
104 * Additionally, KVM only traps guest accesses to the debug registers if
105 * the guest is not actively using them (see the KVM_ARM64_DEBUG_DIRTY
106 * flag on vcpu->arch.flags). Since the guest must not interfere
107 * with the hardware state when debugging the guest, we must ensure that
108 * trapping is enabled whenever we are debugging the guest using the
109 * debug registers.
110 */
111
kvm_arm_setup_debug(struct kvm_vcpu * vcpu)112 void kvm_arm_setup_debug(struct kvm_vcpu *vcpu)
113 {
114 bool trap_debug = !(vcpu->arch.flags & KVM_ARM64_DEBUG_DIRTY);
115 unsigned long mdscr, orig_mdcr_el2 = vcpu->arch.mdcr_el2;
116
117 trace_kvm_arm_setup_debug(vcpu, vcpu->guest_debug);
118
119 /*
120 * This also clears MDCR_EL2_E2PB_MASK to disable guest access
121 * to the profiling buffer.
122 */
123 vcpu->arch.mdcr_el2 = __this_cpu_read(mdcr_el2) & MDCR_EL2_HPMN_MASK;
124 vcpu->arch.mdcr_el2 |= (MDCR_EL2_TPM |
125 MDCR_EL2_TPMS |
126 MDCR_EL2_TPMCR |
127 MDCR_EL2_TDRA |
128 MDCR_EL2_TDOSA);
129
130 /* Is Guest debugging in effect? */
131 if (vcpu->guest_debug) {
132 /* Route all software debug exceptions to EL2 */
133 vcpu->arch.mdcr_el2 |= MDCR_EL2_TDE;
134
135 /* Save guest debug state */
136 save_guest_debug_regs(vcpu);
137
138 /*
139 * Single Step (ARM ARM D2.12.3 The software step state
140 * machine)
141 *
142 * If we are doing Single Step we need to manipulate
143 * the guest's MDSCR_EL1.SS and PSTATE.SS. Once the
144 * step has occurred the hypervisor will trap the
145 * debug exception and we return to userspace.
146 *
147 * If the guest attempts to single step its userspace
148 * we would have to deal with a trapped exception
149 * while in the guest kernel. Because this would be
150 * hard to unwind we suppress the guest's ability to
151 * do so by masking MDSCR_EL.SS.
152 *
153 * This confuses guest debuggers which use
154 * single-step behind the scenes but everything
155 * returns to normal once the host is no longer
156 * debugging the system.
157 */
158 if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP) {
159 *vcpu_cpsr(vcpu) |= DBG_SPSR_SS;
160 mdscr = vcpu_read_sys_reg(vcpu, MDSCR_EL1);
161 mdscr |= DBG_MDSCR_SS;
162 vcpu_write_sys_reg(vcpu, mdscr, MDSCR_EL1);
163 } else {
164 mdscr = vcpu_read_sys_reg(vcpu, MDSCR_EL1);
165 mdscr &= ~DBG_MDSCR_SS;
166 vcpu_write_sys_reg(vcpu, mdscr, MDSCR_EL1);
167 }
168
169 trace_kvm_arm_set_dreg32("SPSR_EL2", *vcpu_cpsr(vcpu));
170
171 /*
172 * HW Breakpoints and watchpoints
173 *
174 * We simply switch the debug_ptr to point to our new
175 * external_debug_state which has been populated by the
176 * debug ioctl. The existing KVM_ARM64_DEBUG_DIRTY
177 * mechanism ensures the registers are updated on the
178 * world switch.
179 */
180 if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW) {
181 /* Enable breakpoints/watchpoints */
182 mdscr = vcpu_read_sys_reg(vcpu, MDSCR_EL1);
183 mdscr |= DBG_MDSCR_MDE;
184 vcpu_write_sys_reg(vcpu, mdscr, MDSCR_EL1);
185
186 vcpu->arch.debug_ptr = &vcpu->arch.external_debug_state;
187 vcpu->arch.flags |= KVM_ARM64_DEBUG_DIRTY;
188 trap_debug = true;
189
190 trace_kvm_arm_set_regset("BKPTS", get_num_brps(),
191 &vcpu->arch.debug_ptr->dbg_bcr[0],
192 &vcpu->arch.debug_ptr->dbg_bvr[0]);
193
194 trace_kvm_arm_set_regset("WAPTS", get_num_wrps(),
195 &vcpu->arch.debug_ptr->dbg_wcr[0],
196 &vcpu->arch.debug_ptr->dbg_wvr[0]);
197 }
198 }
199
200 BUG_ON(!vcpu->guest_debug &&
201 vcpu->arch.debug_ptr != &vcpu->arch.vcpu_debug_state);
202
203 /* Trap debug register access */
204 if (trap_debug)
205 vcpu->arch.mdcr_el2 |= MDCR_EL2_TDA;
206
207 /* If KDE or MDE are set, perform a full save/restore cycle. */
208 if (vcpu_read_sys_reg(vcpu, MDSCR_EL1) & (DBG_MDSCR_KDE | DBG_MDSCR_MDE))
209 vcpu->arch.flags |= KVM_ARM64_DEBUG_DIRTY;
210
211 /* Write mdcr_el2 changes since vcpu_load on VHE systems */
212 if (has_vhe() && orig_mdcr_el2 != vcpu->arch.mdcr_el2)
213 write_sysreg(vcpu->arch.mdcr_el2, mdcr_el2);
214
215 trace_kvm_arm_set_dreg32("MDCR_EL2", vcpu->arch.mdcr_el2);
216 trace_kvm_arm_set_dreg32("MDSCR_EL1", vcpu_read_sys_reg(vcpu, MDSCR_EL1));
217 }
218
kvm_arm_clear_debug(struct kvm_vcpu * vcpu)219 void kvm_arm_clear_debug(struct kvm_vcpu *vcpu)
220 {
221 trace_kvm_arm_clear_debug(vcpu->guest_debug);
222
223 if (vcpu->guest_debug) {
224 restore_guest_debug_regs(vcpu);
225
226 /*
227 * If we were using HW debug we need to restore the
228 * debug_ptr to the guest debug state.
229 */
230 if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW) {
231 kvm_arm_reset_debug_ptr(vcpu);
232
233 trace_kvm_arm_set_regset("BKPTS", get_num_brps(),
234 &vcpu->arch.debug_ptr->dbg_bcr[0],
235 &vcpu->arch.debug_ptr->dbg_bvr[0]);
236
237 trace_kvm_arm_set_regset("WAPTS", get_num_wrps(),
238 &vcpu->arch.debug_ptr->dbg_wcr[0],
239 &vcpu->arch.debug_ptr->dbg_wvr[0]);
240 }
241 }
242 }
243
244
245 /*
246 * After successfully emulating an instruction, we might want to
247 * return to user space with a KVM_EXIT_DEBUG. We can only do this
248 * once the emulation is complete, though, so for userspace emulations
249 * we have to wait until we have re-entered KVM before calling this
250 * helper.
251 *
252 * Return true (and set exit_reason) to return to userspace or false
253 * if no further action is required.
254 */
kvm_arm_handle_step_debug(struct kvm_vcpu * vcpu,struct kvm_run * run)255 bool kvm_arm_handle_step_debug(struct kvm_vcpu *vcpu, struct kvm_run *run)
256 {
257 if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP) {
258 run->exit_reason = KVM_EXIT_DEBUG;
259 run->debug.arch.hsr = ESR_ELx_EC_SOFTSTP_LOW << ESR_ELx_EC_SHIFT;
260 return true;
261 }
262 return false;
263 }
264