Files
linux/arch/x86/kvm/fpu.h
Paolo Bonzini 4cb21be4c3 KVM: x86: Add AVX support to the emulator's register fetch and writeback
Prepare struct operand for hosting AVX registers.  Remove the
existing, incomplete code that placed the Avx flag in the operand
alignment field, and repurpose the name for a separate bit that
indicates:

- after decode, whether an instruction supports the VEX prefix;

- before writeback, that the instruction did have the VEX prefix and
therefore 1) it can have op_bytes == 32; 2) t should clear high
bytes of XMM registers.

Right now the bit will never be set and the patch has no intended
functional change.  However, this is actually more vexing than the
decoder changes itself, and therefore worth separating.

Co-developed-by: Keith Busch <kbusch@kernel.org>
Signed-off-by: Keith Busch <kbusch@kernel.org>
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
Link: https://patch.msgid.link/20251114003633.60689-8-pbonzini@redhat.com
[sean: guard ymm[8-15] accesses with #ifdef CONFIG_X86_64]
Signed-off-by: Sean Christopherson <seanjc@google.com>
2025-11-19 14:30:54 -08:00

207 lines
7.0 KiB
C

/* SPDX-License-Identifier: GPL-2.0 */
#ifndef __KVM_FPU_H_
#define __KVM_FPU_H_
#include <asm/fpu/api.h>
typedef u32 __attribute__((vector_size(16))) sse128_t;
#define __sse128_u union { sse128_t vec; u64 as_u64[2]; u32 as_u32[4]; }
#define sse128_lo(x) ({ __sse128_u t; t.vec = x; t.as_u64[0]; })
#define sse128_hi(x) ({ __sse128_u t; t.vec = x; t.as_u64[1]; })
#define sse128_l0(x) ({ __sse128_u t; t.vec = x; t.as_u32[0]; })
#define sse128_l1(x) ({ __sse128_u t; t.vec = x; t.as_u32[1]; })
#define sse128_l2(x) ({ __sse128_u t; t.vec = x; t.as_u32[2]; })
#define sse128_l3(x) ({ __sse128_u t; t.vec = x; t.as_u32[3]; })
#define sse128(lo, hi) ({ __sse128_u t; t.as_u64[0] = lo; t.as_u64[1] = hi; t.vec; })
typedef u32 __attribute__((vector_size(32))) avx256_t;
static inline void _kvm_read_avx_reg(int reg, avx256_t *data)
{
switch (reg) {
case 0: asm("vmovdqa %%ymm0, %0" : "=m"(*data)); break;
case 1: asm("vmovdqa %%ymm1, %0" : "=m"(*data)); break;
case 2: asm("vmovdqa %%ymm2, %0" : "=m"(*data)); break;
case 3: asm("vmovdqa %%ymm3, %0" : "=m"(*data)); break;
case 4: asm("vmovdqa %%ymm4, %0" : "=m"(*data)); break;
case 5: asm("vmovdqa %%ymm5, %0" : "=m"(*data)); break;
case 6: asm("vmovdqa %%ymm6, %0" : "=m"(*data)); break;
case 7: asm("vmovdqa %%ymm7, %0" : "=m"(*data)); break;
#ifdef CONFIG_X86_64
case 8: asm("vmovdqa %%ymm8, %0" : "=m"(*data)); break;
case 9: asm("vmovdqa %%ymm9, %0" : "=m"(*data)); break;
case 10: asm("vmovdqa %%ymm10, %0" : "=m"(*data)); break;
case 11: asm("vmovdqa %%ymm11, %0" : "=m"(*data)); break;
case 12: asm("vmovdqa %%ymm12, %0" : "=m"(*data)); break;
case 13: asm("vmovdqa %%ymm13, %0" : "=m"(*data)); break;
case 14: asm("vmovdqa %%ymm14, %0" : "=m"(*data)); break;
case 15: asm("vmovdqa %%ymm15, %0" : "=m"(*data)); break;
#endif
default: BUG();
}
}
static inline void _kvm_write_avx_reg(int reg, const avx256_t *data)
{
switch (reg) {
case 0: asm("vmovdqa %0, %%ymm0" : : "m"(*data)); break;
case 1: asm("vmovdqa %0, %%ymm1" : : "m"(*data)); break;
case 2: asm("vmovdqa %0, %%ymm2" : : "m"(*data)); break;
case 3: asm("vmovdqa %0, %%ymm3" : : "m"(*data)); break;
case 4: asm("vmovdqa %0, %%ymm4" : : "m"(*data)); break;
case 5: asm("vmovdqa %0, %%ymm5" : : "m"(*data)); break;
case 6: asm("vmovdqa %0, %%ymm6" : : "m"(*data)); break;
case 7: asm("vmovdqa %0, %%ymm7" : : "m"(*data)); break;
#ifdef CONFIG_X86_64
case 8: asm("vmovdqa %0, %%ymm8" : : "m"(*data)); break;
case 9: asm("vmovdqa %0, %%ymm9" : : "m"(*data)); break;
case 10: asm("vmovdqa %0, %%ymm10" : : "m"(*data)); break;
case 11: asm("vmovdqa %0, %%ymm11" : : "m"(*data)); break;
case 12: asm("vmovdqa %0, %%ymm12" : : "m"(*data)); break;
case 13: asm("vmovdqa %0, %%ymm13" : : "m"(*data)); break;
case 14: asm("vmovdqa %0, %%ymm14" : : "m"(*data)); break;
case 15: asm("vmovdqa %0, %%ymm15" : : "m"(*data)); break;
#endif
default: BUG();
}
}
static inline void _kvm_read_sse_reg(int reg, sse128_t *data)
{
switch (reg) {
case 0: asm("movdqa %%xmm0, %0" : "=m"(*data)); break;
case 1: asm("movdqa %%xmm1, %0" : "=m"(*data)); break;
case 2: asm("movdqa %%xmm2, %0" : "=m"(*data)); break;
case 3: asm("movdqa %%xmm3, %0" : "=m"(*data)); break;
case 4: asm("movdqa %%xmm4, %0" : "=m"(*data)); break;
case 5: asm("movdqa %%xmm5, %0" : "=m"(*data)); break;
case 6: asm("movdqa %%xmm6, %0" : "=m"(*data)); break;
case 7: asm("movdqa %%xmm7, %0" : "=m"(*data)); break;
#ifdef CONFIG_X86_64
case 8: asm("movdqa %%xmm8, %0" : "=m"(*data)); break;
case 9: asm("movdqa %%xmm9, %0" : "=m"(*data)); break;
case 10: asm("movdqa %%xmm10, %0" : "=m"(*data)); break;
case 11: asm("movdqa %%xmm11, %0" : "=m"(*data)); break;
case 12: asm("movdqa %%xmm12, %0" : "=m"(*data)); break;
case 13: asm("movdqa %%xmm13, %0" : "=m"(*data)); break;
case 14: asm("movdqa %%xmm14, %0" : "=m"(*data)); break;
case 15: asm("movdqa %%xmm15, %0" : "=m"(*data)); break;
#endif
default: BUG();
}
}
static inline void _kvm_write_sse_reg(int reg, const sse128_t *data)
{
switch (reg) {
case 0: asm("movdqa %0, %%xmm0" : : "m"(*data)); break;
case 1: asm("movdqa %0, %%xmm1" : : "m"(*data)); break;
case 2: asm("movdqa %0, %%xmm2" : : "m"(*data)); break;
case 3: asm("movdqa %0, %%xmm3" : : "m"(*data)); break;
case 4: asm("movdqa %0, %%xmm4" : : "m"(*data)); break;
case 5: asm("movdqa %0, %%xmm5" : : "m"(*data)); break;
case 6: asm("movdqa %0, %%xmm6" : : "m"(*data)); break;
case 7: asm("movdqa %0, %%xmm7" : : "m"(*data)); break;
#ifdef CONFIG_X86_64
case 8: asm("movdqa %0, %%xmm8" : : "m"(*data)); break;
case 9: asm("movdqa %0, %%xmm9" : : "m"(*data)); break;
case 10: asm("movdqa %0, %%xmm10" : : "m"(*data)); break;
case 11: asm("movdqa %0, %%xmm11" : : "m"(*data)); break;
case 12: asm("movdqa %0, %%xmm12" : : "m"(*data)); break;
case 13: asm("movdqa %0, %%xmm13" : : "m"(*data)); break;
case 14: asm("movdqa %0, %%xmm14" : : "m"(*data)); break;
case 15: asm("movdqa %0, %%xmm15" : : "m"(*data)); break;
#endif
default: BUG();
}
}
static inline void _kvm_read_mmx_reg(int reg, u64 *data)
{
switch (reg) {
case 0: asm("movq %%mm0, %0" : "=m"(*data)); break;
case 1: asm("movq %%mm1, %0" : "=m"(*data)); break;
case 2: asm("movq %%mm2, %0" : "=m"(*data)); break;
case 3: asm("movq %%mm3, %0" : "=m"(*data)); break;
case 4: asm("movq %%mm4, %0" : "=m"(*data)); break;
case 5: asm("movq %%mm5, %0" : "=m"(*data)); break;
case 6: asm("movq %%mm6, %0" : "=m"(*data)); break;
case 7: asm("movq %%mm7, %0" : "=m"(*data)); break;
default: BUG();
}
}
static inline void _kvm_write_mmx_reg(int reg, const u64 *data)
{
switch (reg) {
case 0: asm("movq %0, %%mm0" : : "m"(*data)); break;
case 1: asm("movq %0, %%mm1" : : "m"(*data)); break;
case 2: asm("movq %0, %%mm2" : : "m"(*data)); break;
case 3: asm("movq %0, %%mm3" : : "m"(*data)); break;
case 4: asm("movq %0, %%mm4" : : "m"(*data)); break;
case 5: asm("movq %0, %%mm5" : : "m"(*data)); break;
case 6: asm("movq %0, %%mm6" : : "m"(*data)); break;
case 7: asm("movq %0, %%mm7" : : "m"(*data)); break;
default: BUG();
}
}
static inline void kvm_fpu_get(void)
{
fpregs_lock();
fpregs_assert_state_consistent();
if (test_thread_flag(TIF_NEED_FPU_LOAD))
switch_fpu_return();
}
static inline void kvm_fpu_put(void)
{
fpregs_unlock();
}
static inline void kvm_read_avx_reg(int reg, avx256_t *data)
{
kvm_fpu_get();
_kvm_read_avx_reg(reg, data);
kvm_fpu_put();
}
static inline void kvm_write_avx_reg(int reg, const avx256_t *data)
{
kvm_fpu_get();
_kvm_write_avx_reg(reg, data);
kvm_fpu_put();
}
static inline void kvm_read_sse_reg(int reg, sse128_t *data)
{
kvm_fpu_get();
_kvm_read_sse_reg(reg, data);
kvm_fpu_put();
}
static inline void kvm_write_sse_reg(int reg, const sse128_t *data)
{
kvm_fpu_get();
_kvm_write_sse_reg(reg, data);
kvm_fpu_put();
}
static inline void kvm_read_mmx_reg(int reg, u64 *data)
{
kvm_fpu_get();
_kvm_read_mmx_reg(reg, data);
kvm_fpu_put();
}
static inline void kvm_write_mmx_reg(int reg, const u64 *data)
{
kvm_fpu_get();
_kvm_write_mmx_reg(reg, data);
kvm_fpu_put();
}
#endif