1 // SPDX-License-Identifier: GPL-2.0-only
5 * Copyright (C) 2021, Intel, Inc.
7 * Tests for amx #NM exception and save/restore.
10 #define _GNU_SOURCE /* for program_invocation_short_name */
15 #include <sys/ioctl.h>
16 #include <sys/syscall.h>
18 #include "test_util.h"
21 #include "processor.h"
25 # error This test is 64-bit only
29 #define X86_FEATURE_XSAVE (1 << 26)
30 #define X86_FEATURE_OSXSAVE (1 << 27)
32 #define PAGE_SIZE (1 << 12)
34 #define TILE_SIZE 1024
35 #define XSAVE_SIZE ((NUM_TILES * TILE_SIZE) + PAGE_SIZE)
37 /* Tile configuration associated: */
39 #define RESERVED_BYTES 14
41 #define XFEATURE_XTILECFG 17
42 #define XFEATURE_XTILEDATA 18
43 #define XFEATURE_MASK_XTILECFG (1 << XFEATURE_XTILECFG)
44 #define XFEATURE_MASK_XTILEDATA (1 << XFEATURE_XTILEDATA)
45 #define XFEATURE_MASK_XTILE (XFEATURE_MASK_XTILECFG | XFEATURE_MASK_XTILEDATA)
47 #define TILE_CPUID 0x1d
48 #define XSTATE_CPUID 0xd
49 #define TILE_PALETTE_CPUID_SUBLEAVE 0x1
50 #define XSTATE_USER_STATE_SUBLEAVE 0x0
52 #define XSAVE_HDR_OFFSET 512
61 u8 reserved[RESERVED_BYTES];
67 u8 data[NUM_TILES * TILE_SIZE];
79 static struct xtile_info xtile;
81 static inline u64 __xgetbv(u32 index)
85 asm volatile("xgetbv;"
86 : "=a" (eax), "=d" (edx)
88 return eax + ((u64)edx << 32);
91 static inline void __xsetbv(u32 index, u64 value)
94 u32 edx = value >> 32;
96 asm volatile("xsetbv" :: "a" (eax), "d" (edx), "c" (index));
99 static inline void __ldtilecfg(void *cfg)
101 asm volatile(".byte 0xc4,0xe2,0x78,0x49,0x00"
105 static inline void __tileloadd(void *tile)
107 asm volatile(".byte 0xc4,0xe2,0x7b,0x4b,0x04,0x10"
108 : : "a"(tile), "d"(0));
111 static inline void __tilerelease(void)
113 asm volatile(".byte 0xc4, 0xe2, 0x78, 0x49, 0xc0" ::);
116 static inline void __xsavec(struct xsave_data *data, uint64_t rfbm)
118 uint32_t rfbm_lo = rfbm;
119 uint32_t rfbm_hi = rfbm >> 32;
121 asm volatile("xsavec (%%rdi)"
122 : : "D" (data), "a" (rfbm_lo), "d" (rfbm_hi)
126 static inline void check_cpuid_xsave(void)
128 uint32_t eax, ebx, ecx, edx;
132 cpuid(&eax, &ebx, &ecx, &edx);
133 if (!(ecx & X86_FEATURE_XSAVE))
134 GUEST_ASSERT(!"cpuid: no CPU xsave support!");
135 if (!(ecx & X86_FEATURE_OSXSAVE))
136 GUEST_ASSERT(!"cpuid: no OS xsave support!");
139 static bool check_xsave_supports_xtile(void)
141 return __xgetbv(0) & XFEATURE_MASK_XTILE;
144 static bool enum_xtile_config(void)
146 u32 eax, ebx, ecx, edx;
149 ecx = TILE_PALETTE_CPUID_SUBLEAVE;
151 cpuid(&eax, &ebx, &ecx, &edx);
152 if (!eax || !ebx || !ecx)
155 xtile.max_names = ebx >> 16;
156 if (xtile.max_names < NUM_TILES)
159 xtile.bytes_per_tile = eax >> 16;
160 if (xtile.bytes_per_tile < TILE_SIZE)
163 xtile.bytes_per_row = ebx;
164 xtile.max_rows = ecx;
169 static bool enum_xsave_tile(void)
171 u32 eax, ebx, ecx, edx;
174 ecx = XFEATURE_XTILEDATA;
176 cpuid(&eax, &ebx, &ecx, &edx);
180 xtile.xsave_offset = ebx;
181 xtile.xsave_size = eax;
186 static bool check_xsave_size(void)
188 u32 eax, ebx, ecx, edx;
192 ecx = XSTATE_USER_STATE_SUBLEAVE;
194 cpuid(&eax, &ebx, &ecx, &edx);
195 if (ebx && ebx <= XSAVE_SIZE)
201 static bool check_xtile_info(void)
205 if (!check_xsave_size())
208 if (!enum_xsave_tile())
211 if (!enum_xtile_config())
214 if (sizeof(struct tile_data) >= xtile.xsave_size)
220 static void set_tilecfg(struct tile_config *cfg)
224 /* Only palette id 1 */
226 for (i = 0; i < xtile.max_names; i++) {
227 cfg->colsb[i] = xtile.bytes_per_row;
228 cfg->rows[i] = xtile.max_rows;
232 static void set_xstatebv(void *data, uint64_t bv)
234 *(uint64_t *)(data + XSAVE_HDR_OFFSET) = bv;
237 static u64 get_xstatebv(void *data)
239 return *(u64 *)(data + XSAVE_HDR_OFFSET);
242 static void init_regs(void)
246 /* turn on CR4.OSXSAVE */
248 cr4 |= X86_CR4_OSXSAVE;
252 xcr0 |= XFEATURE_MASK_XTILE;
256 static void __attribute__((__flatten__)) guest_code(struct tile_config *amx_cfg,
257 struct tile_data *tiledata,
258 struct xsave_data *xsave_data)
262 GUEST_ASSERT(check_xsave_supports_xtile());
263 GUEST_ASSERT(check_xtile_info());
265 /* check xtile configs */
266 GUEST_ASSERT(xtile.xsave_offset == 2816);
267 GUEST_ASSERT(xtile.xsave_size == 8192);
268 GUEST_ASSERT(xtile.max_names == 8);
269 GUEST_ASSERT(xtile.bytes_per_tile == 1024);
270 GUEST_ASSERT(xtile.bytes_per_row == 64);
271 GUEST_ASSERT(xtile.max_rows == 16);
274 /* xfd=0, enable amx */
275 wrmsr(MSR_IA32_XFD, 0);
277 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == 0);
278 set_tilecfg(amx_cfg);
279 __ldtilecfg(amx_cfg);
281 /* Check save/restore when trap to userspace */
282 __tileloadd(tiledata);
286 /* bit 18 not in the XCOMP_BV after xsavec() */
287 set_xstatebv(xsave_data, XFEATURE_MASK_XTILEDATA);
288 __xsavec(xsave_data, XFEATURE_MASK_XTILEDATA);
289 GUEST_ASSERT((get_xstatebv(xsave_data) & XFEATURE_MASK_XTILEDATA) == 0);
291 /* xfd=0x40000, disable amx tiledata */
292 wrmsr(MSR_IA32_XFD, XFEATURE_MASK_XTILEDATA);
294 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == XFEATURE_MASK_XTILEDATA);
295 set_tilecfg(amx_cfg);
296 __ldtilecfg(amx_cfg);
297 /* Trigger #NM exception */
298 __tileloadd(tiledata);
304 void guest_nm_handler(struct ex_regs *regs)
306 /* Check if #NM is triggered by XFEATURE_MASK_XTILEDATA */
308 GUEST_ASSERT(rdmsr(MSR_IA32_XFD_ERR) == XFEATURE_MASK_XTILEDATA);
310 GUEST_ASSERT(rdmsr(MSR_IA32_XFD_ERR) == XFEATURE_MASK_XTILEDATA);
312 wrmsr(MSR_IA32_XFD_ERR, 0);
313 /* xfd=0, enable amx */
314 wrmsr(MSR_IA32_XFD, 0);
318 int main(int argc, char *argv[])
320 struct kvm_cpuid_entry2 *entry;
321 struct kvm_regs regs1, regs2;
322 bool amx_supported = false;
325 struct kvm_x86_state *state;
326 int xsave_restore_size = 0;
327 vm_vaddr_t amx_cfg, tiledata, xsavedata;
332 vm_xsave_req_perm(XSTATE_XTILE_DATA_BIT);
335 vm = vm_create_default(VCPU_ID, 0, guest_code);
337 entry = kvm_get_supported_cpuid_entry(1);
338 if (!(entry->ecx & X86_FEATURE_XSAVE)) {
339 print_skip("XSAVE feature not supported");
343 if (kvm_get_cpuid_max_basic() >= 0xd) {
344 entry = kvm_get_supported_cpuid_index(0xd, 0);
345 amx_supported = entry && !!(entry->eax & XFEATURE_MASK_XTILE);
346 if (!amx_supported) {
347 print_skip("AMX is not supported by the vCPU (eax=0x%x)", entry->eax);
350 /* Get xsave/restore max size */
351 xsave_restore_size = entry->ecx;
354 run = vcpu_state(vm, VCPU_ID);
355 vcpu_regs_get(vm, VCPU_ID, ®s1);
357 /* Register #NM handler */
358 vm_init_descriptor_tables(vm);
359 vcpu_init_descriptor_tables(vm, VCPU_ID);
360 vm_install_exception_handler(vm, NM_VECTOR, guest_nm_handler);
362 /* amx cfg for guest_code */
363 amx_cfg = vm_vaddr_alloc_page(vm);
364 memset(addr_gva2hva(vm, amx_cfg), 0x0, getpagesize());
366 /* amx tiledata for guest_code */
367 tiledata = vm_vaddr_alloc_pages(vm, 2);
368 memset(addr_gva2hva(vm, tiledata), rand() | 1, 2 * getpagesize());
370 /* xsave data for guest_code */
371 xsavedata = vm_vaddr_alloc_pages(vm, 3);
372 memset(addr_gva2hva(vm, xsavedata), 0, 3 * getpagesize());
373 vcpu_args_set(vm, VCPU_ID, 3, amx_cfg, tiledata, xsavedata);
375 for (stage = 1; ; stage++) {
376 _vcpu_run(vm, VCPU_ID);
377 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
378 "Stage %d: unexpected exit reason: %u (%s),\n",
379 stage, run->exit_reason,
380 exit_reason_str(run->exit_reason));
382 switch (get_ucall(vm, VCPU_ID, &uc)) {
384 TEST_FAIL("%s at %s:%ld", (const char *)uc.args[0],
385 __FILE__, uc.args[1]);
388 switch (uc.args[1]) {
396 fprintf(stderr, "GUEST_SYNC(%ld)\n", uc.args[1]);
401 "GUEST_SYNC(%ld), check save/restore status\n", uc.args[1]);
403 /* Compacted mode, get amx offset by xsave area
404 * size subtract 8K amx size.
406 amx_offset = xsave_restore_size - NUM_TILES*TILE_SIZE;
407 state = vcpu_save_state(vm, VCPU_ID);
408 void *amx_start = (void *)state->xsave + amx_offset;
409 void *tiles_data = (void *)addr_gva2hva(vm, tiledata);
410 /* Only check TMM0 register, 1 tile */
411 ret = memcmp(amx_start, tiles_data, TILE_SIZE);
412 TEST_ASSERT(ret == 0, "memcmp failed, ret=%d\n", ret);
413 kvm_x86_state_cleanup(state);
417 "GUEST_SYNC(%ld), #NM exception and enable amx\n", uc.args[1]);
422 fprintf(stderr, "UCALL_DONE\n");
425 TEST_FAIL("Unknown ucall %lu", uc.cmd);
428 state = vcpu_save_state(vm, VCPU_ID);
429 memset(®s1, 0, sizeof(regs1));
430 vcpu_regs_get(vm, VCPU_ID, ®s1);
434 /* Restore state in a new VM. */
435 kvm_vm_restart(vm, O_RDWR);
436 vm_vcpu_add(vm, VCPU_ID);
437 vcpu_set_cpuid(vm, VCPU_ID, kvm_get_supported_cpuid());
438 vcpu_load_state(vm, VCPU_ID, state);
439 run = vcpu_state(vm, VCPU_ID);
440 kvm_x86_state_cleanup(state);
442 memset(®s2, 0, sizeof(regs2));
443 vcpu_regs_get(vm, VCPU_ID, ®s2);
444 TEST_ASSERT(!memcmp(®s1, ®s2, sizeof(regs2)),
445 "Unexpected register values after vcpu_load_state; rdi: %lx rsi: %lx",
446 (ulong) regs2.rdi, (ulong) regs2.rsi);