KVM: stats: Separate generic stats from architecture specific ones
[linux-2.6-microblaze.git] / arch / powerpc / kvm / e500_emulate.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
4  *
5  * Author: Yu Liu, <yu.liu@freescale.com>
6  *
7  * Description:
8  * This file is derived from arch/powerpc/kvm/44x_emulate.c,
9  * by Hollis Blanchard <hollisb@us.ibm.com>.
10  */
11
12 #include <asm/kvm_ppc.h>
13 #include <asm/disassemble.h>
14 #include <asm/dbell.h>
15 #include <asm/reg_booke.h>
16
17 #include "booke.h"
18 #include "e500.h"
19
20 #define XOP_DCBTLS  166
21 #define XOP_MSGSND  206
22 #define XOP_MSGCLR  238
23 #define XOP_MFTMR   366
24 #define XOP_TLBIVAX 786
25 #define XOP_TLBSX   914
26 #define XOP_TLBRE   946
27 #define XOP_TLBWE   978
28 #define XOP_TLBILX  18
29 #define XOP_EHPRIV  270
30
31 #ifdef CONFIG_KVM_E500MC
32 static int dbell2prio(ulong param)
33 {
34         int msg = param & PPC_DBELL_TYPE_MASK;
35         int prio = -1;
36
37         switch (msg) {
38         case PPC_DBELL_TYPE(PPC_DBELL):
39                 prio = BOOKE_IRQPRIO_DBELL;
40                 break;
41         case PPC_DBELL_TYPE(PPC_DBELL_CRIT):
42                 prio = BOOKE_IRQPRIO_DBELL_CRIT;
43                 break;
44         default:
45                 break;
46         }
47
48         return prio;
49 }
50
51 static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb)
52 {
53         ulong param = vcpu->arch.regs.gpr[rb];
54         int prio = dbell2prio(param);
55
56         if (prio < 0)
57                 return EMULATE_FAIL;
58
59         clear_bit(prio, &vcpu->arch.pending_exceptions);
60         return EMULATE_DONE;
61 }
62
63 static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb)
64 {
65         ulong param = vcpu->arch.regs.gpr[rb];
66         int prio = dbell2prio(rb);
67         int pir = param & PPC_DBELL_PIR_MASK;
68         int i;
69         struct kvm_vcpu *cvcpu;
70
71         if (prio < 0)
72                 return EMULATE_FAIL;
73
74         kvm_for_each_vcpu(i, cvcpu, vcpu->kvm) {
75                 int cpir = cvcpu->arch.shared->pir;
76                 if ((param & PPC_DBELL_MSG_BRDCAST) || (cpir == pir)) {
77                         set_bit(prio, &cvcpu->arch.pending_exceptions);
78                         kvm_vcpu_kick(cvcpu);
79                 }
80         }
81
82         return EMULATE_DONE;
83 }
84 #endif
85
86 static int kvmppc_e500_emul_ehpriv(struct kvm_vcpu *vcpu,
87                                    unsigned int inst, int *advance)
88 {
89         int emulated = EMULATE_DONE;
90
91         switch (get_oc(inst)) {
92         case EHPRIV_OC_DEBUG:
93                 vcpu->run->exit_reason = KVM_EXIT_DEBUG;
94                 vcpu->run->debug.arch.address = vcpu->arch.regs.nip;
95                 vcpu->run->debug.arch.status = 0;
96                 kvmppc_account_exit(vcpu, DEBUG_EXITS);
97                 emulated = EMULATE_EXIT_USER;
98                 *advance = 0;
99                 break;
100         default:
101                 emulated = EMULATE_FAIL;
102         }
103         return emulated;
104 }
105
106 static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu)
107 {
108         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
109
110         /* Always fail to lock the cache */
111         vcpu_e500->l1csr0 |= L1CSR0_CUL;
112         return EMULATE_DONE;
113 }
114
115 static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst,
116                                   int rt)
117 {
118         /* Expose one thread per vcpu */
119         if (get_tmrn(inst) == TMRN_TMCFG0) {
120                 kvmppc_set_gpr(vcpu, rt,
121                                1 | (1 << TMRN_TMCFG0_NATHRD_SHIFT));
122                 return EMULATE_DONE;
123         }
124
125         return EMULATE_FAIL;
126 }
127
128 int kvmppc_core_emulate_op_e500(struct kvm_vcpu *vcpu,
129                                 unsigned int inst, int *advance)
130 {
131         int emulated = EMULATE_DONE;
132         int ra = get_ra(inst);
133         int rb = get_rb(inst);
134         int rt = get_rt(inst);
135         gva_t ea;
136
137         switch (get_op(inst)) {
138         case 31:
139                 switch (get_xop(inst)) {
140
141                 case XOP_DCBTLS:
142                         emulated = kvmppc_e500_emul_dcbtls(vcpu);
143                         break;
144
145 #ifdef CONFIG_KVM_E500MC
146                 case XOP_MSGSND:
147                         emulated = kvmppc_e500_emul_msgsnd(vcpu, rb);
148                         break;
149
150                 case XOP_MSGCLR:
151                         emulated = kvmppc_e500_emul_msgclr(vcpu, rb);
152                         break;
153 #endif
154
155                 case XOP_TLBRE:
156                         emulated = kvmppc_e500_emul_tlbre(vcpu);
157                         break;
158
159                 case XOP_TLBWE:
160                         emulated = kvmppc_e500_emul_tlbwe(vcpu);
161                         break;
162
163                 case XOP_TLBSX:
164                         ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
165                         emulated = kvmppc_e500_emul_tlbsx(vcpu, ea);
166                         break;
167
168                 case XOP_TLBILX: {
169                         int type = rt & 0x3;
170                         ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
171                         emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea);
172                         break;
173                 }
174
175                 case XOP_TLBIVAX:
176                         ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
177                         emulated = kvmppc_e500_emul_tlbivax(vcpu, ea);
178                         break;
179
180                 case XOP_MFTMR:
181                         emulated = kvmppc_e500_emul_mftmr(vcpu, inst, rt);
182                         break;
183
184                 case XOP_EHPRIV:
185                         emulated = kvmppc_e500_emul_ehpriv(vcpu, inst, advance);
186                         break;
187
188                 default:
189                         emulated = EMULATE_FAIL;
190                 }
191
192                 break;
193
194         default:
195                 emulated = EMULATE_FAIL;
196         }
197
198         if (emulated == EMULATE_FAIL)
199                 emulated = kvmppc_booke_emulate_op(vcpu, inst, advance);
200
201         return emulated;
202 }
203
204 int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
205 {
206         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
207         int emulated = EMULATE_DONE;
208
209         switch (sprn) {
210 #ifndef CONFIG_KVM_BOOKE_HV
211         case SPRN_PID:
212                 kvmppc_set_pid(vcpu, spr_val);
213                 break;
214         case SPRN_PID1:
215                 if (spr_val != 0)
216                         return EMULATE_FAIL;
217                 vcpu_e500->pid[1] = spr_val;
218                 break;
219         case SPRN_PID2:
220                 if (spr_val != 0)
221                         return EMULATE_FAIL;
222                 vcpu_e500->pid[2] = spr_val;
223                 break;
224         case SPRN_MAS0:
225                 vcpu->arch.shared->mas0 = spr_val;
226                 break;
227         case SPRN_MAS1:
228                 vcpu->arch.shared->mas1 = spr_val;
229                 break;
230         case SPRN_MAS2:
231                 vcpu->arch.shared->mas2 = spr_val;
232                 break;
233         case SPRN_MAS3:
234                 vcpu->arch.shared->mas7_3 &= ~(u64)0xffffffff;
235                 vcpu->arch.shared->mas7_3 |= spr_val;
236                 break;
237         case SPRN_MAS4:
238                 vcpu->arch.shared->mas4 = spr_val;
239                 break;
240         case SPRN_MAS6:
241                 vcpu->arch.shared->mas6 = spr_val;
242                 break;
243         case SPRN_MAS7:
244                 vcpu->arch.shared->mas7_3 &= (u64)0xffffffff;
245                 vcpu->arch.shared->mas7_3 |= (u64)spr_val << 32;
246                 break;
247 #endif
248         case SPRN_L1CSR0:
249                 vcpu_e500->l1csr0 = spr_val;
250                 vcpu_e500->l1csr0 &= ~(L1CSR0_DCFI | L1CSR0_CLFC);
251                 break;
252         case SPRN_L1CSR1:
253                 vcpu_e500->l1csr1 = spr_val;
254                 vcpu_e500->l1csr1 &= ~(L1CSR1_ICFI | L1CSR1_ICLFR);
255                 break;
256         case SPRN_HID0:
257                 vcpu_e500->hid0 = spr_val;
258                 break;
259         case SPRN_HID1:
260                 vcpu_e500->hid1 = spr_val;
261                 break;
262
263         case SPRN_MMUCSR0:
264                 emulated = kvmppc_e500_emul_mt_mmucsr0(vcpu_e500,
265                                 spr_val);
266                 break;
267
268         case SPRN_PWRMGTCR0:
269                 /*
270                  * Guest relies on host power management configurations
271                  * Treat the request as a general store
272                  */
273                 vcpu->arch.pwrmgtcr0 = spr_val;
274                 break;
275
276         case SPRN_BUCSR:
277                 /*
278                  * If we are here, it means that we have already flushed the
279                  * branch predictor, so just return to guest.
280                  */
281                 break;
282
283         /* extra exceptions */
284 #ifdef CONFIG_SPE_POSSIBLE
285         case SPRN_IVOR32:
286                 vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] = spr_val;
287                 break;
288         case SPRN_IVOR33:
289                 vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] = spr_val;
290                 break;
291         case SPRN_IVOR34:
292                 vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] = spr_val;
293                 break;
294 #endif
295 #ifdef CONFIG_ALTIVEC
296         case SPRN_IVOR32:
297                 vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL] = spr_val;
298                 break;
299         case SPRN_IVOR33:
300                 vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST] = spr_val;
301                 break;
302 #endif
303         case SPRN_IVOR35:
304                 vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = spr_val;
305                 break;
306 #ifdef CONFIG_KVM_BOOKE_HV
307         case SPRN_IVOR36:
308                 vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = spr_val;
309                 break;
310         case SPRN_IVOR37:
311                 vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = spr_val;
312                 break;
313 #endif
314         default:
315                 emulated = kvmppc_booke_emulate_mtspr(vcpu, sprn, spr_val);
316         }
317
318         return emulated;
319 }
320
321 int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
322 {
323         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
324         int emulated = EMULATE_DONE;
325
326         switch (sprn) {
327 #ifndef CONFIG_KVM_BOOKE_HV
328         case SPRN_PID:
329                 *spr_val = vcpu_e500->pid[0];
330                 break;
331         case SPRN_PID1:
332                 *spr_val = vcpu_e500->pid[1];
333                 break;
334         case SPRN_PID2:
335                 *spr_val = vcpu_e500->pid[2];
336                 break;
337         case SPRN_MAS0:
338                 *spr_val = vcpu->arch.shared->mas0;
339                 break;
340         case SPRN_MAS1:
341                 *spr_val = vcpu->arch.shared->mas1;
342                 break;
343         case SPRN_MAS2:
344                 *spr_val = vcpu->arch.shared->mas2;
345                 break;
346         case SPRN_MAS3:
347                 *spr_val = (u32)vcpu->arch.shared->mas7_3;
348                 break;
349         case SPRN_MAS4:
350                 *spr_val = vcpu->arch.shared->mas4;
351                 break;
352         case SPRN_MAS6:
353                 *spr_val = vcpu->arch.shared->mas6;
354                 break;
355         case SPRN_MAS7:
356                 *spr_val = vcpu->arch.shared->mas7_3 >> 32;
357                 break;
358 #endif
359         case SPRN_DECAR:
360                 *spr_val = vcpu->arch.decar;
361                 break;
362         case SPRN_TLB0CFG:
363                 *spr_val = vcpu->arch.tlbcfg[0];
364                 break;
365         case SPRN_TLB1CFG:
366                 *spr_val = vcpu->arch.tlbcfg[1];
367                 break;
368         case SPRN_TLB0PS:
369                 if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
370                         return EMULATE_FAIL;
371                 *spr_val = vcpu->arch.tlbps[0];
372                 break;
373         case SPRN_TLB1PS:
374                 if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
375                         return EMULATE_FAIL;
376                 *spr_val = vcpu->arch.tlbps[1];
377                 break;
378         case SPRN_L1CSR0:
379                 *spr_val = vcpu_e500->l1csr0;
380                 break;
381         case SPRN_L1CSR1:
382                 *spr_val = vcpu_e500->l1csr1;
383                 break;
384         case SPRN_HID0:
385                 *spr_val = vcpu_e500->hid0;
386                 break;
387         case SPRN_HID1:
388                 *spr_val = vcpu_e500->hid1;
389                 break;
390         case SPRN_SVR:
391                 *spr_val = vcpu_e500->svr;
392                 break;
393
394         case SPRN_MMUCSR0:
395                 *spr_val = 0;
396                 break;
397
398         case SPRN_MMUCFG:
399                 *spr_val = vcpu->arch.mmucfg;
400                 break;
401         case SPRN_EPTCFG:
402                 if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
403                         return EMULATE_FAIL;
404                 /*
405                  * Legacy Linux guests access EPTCFG register even if the E.PT
406                  * category is disabled in the VM. Give them a chance to live.
407                  */
408                 *spr_val = vcpu->arch.eptcfg;
409                 break;
410
411         case SPRN_PWRMGTCR0:
412                 *spr_val = vcpu->arch.pwrmgtcr0;
413                 break;
414
415         /* extra exceptions */
416 #ifdef CONFIG_SPE_POSSIBLE
417         case SPRN_IVOR32:
418                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
419                 break;
420         case SPRN_IVOR33:
421                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
422                 break;
423         case SPRN_IVOR34:
424                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
425                 break;
426 #endif
427 #ifdef CONFIG_ALTIVEC
428         case SPRN_IVOR32:
429                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL];
430                 break;
431         case SPRN_IVOR33:
432                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST];
433                 break;
434 #endif
435         case SPRN_IVOR35:
436                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
437                 break;
438 #ifdef CONFIG_KVM_BOOKE_HV
439         case SPRN_IVOR36:
440                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
441                 break;
442         case SPRN_IVOR37:
443                 *spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
444                 break;
445 #endif
446         default:
447                 emulated = kvmppc_booke_emulate_mfspr(vcpu, sprn, spr_val);
448         }
449
450         return emulated;
451 }
452