mirror of
https://github.com/torvalds/linux.git
synced 2025-12-07 20:06:24 +00:00
Prepare struct operand for hosting AVX registers. Remove the existing, incomplete code that placed the Avx flag in the operand alignment field, and repurpose the name for a separate bit that indicates: - after decode, whether an instruction supports the VEX prefix; - before writeback, that the instruction did have the VEX prefix and therefore 1) it can have op_bytes == 32; 2) t should clear high bytes of XMM registers. Right now the bit will never be set and the patch has no intended functional change. However, this is actually more vexing than the decoder changes itself, and therefore worth separating. Co-developed-by: Keith Busch <kbusch@kernel.org> Signed-off-by: Keith Busch <kbusch@kernel.org> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> Link: https://patch.msgid.link/20251114003633.60689-8-pbonzini@redhat.com [sean: guard ymm[8-15] accesses with #ifdef CONFIG_X86_64] Signed-off-by: Sean Christopherson <seanjc@google.com>
207 lines
7.0 KiB
C
207 lines
7.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
|
|
#ifndef __KVM_FPU_H_
|
|
#define __KVM_FPU_H_
|
|
|
|
#include <asm/fpu/api.h>
|
|
|
|
typedef u32 __attribute__((vector_size(16))) sse128_t;
|
|
#define __sse128_u union { sse128_t vec; u64 as_u64[2]; u32 as_u32[4]; }
|
|
#define sse128_lo(x) ({ __sse128_u t; t.vec = x; t.as_u64[0]; })
|
|
#define sse128_hi(x) ({ __sse128_u t; t.vec = x; t.as_u64[1]; })
|
|
#define sse128_l0(x) ({ __sse128_u t; t.vec = x; t.as_u32[0]; })
|
|
#define sse128_l1(x) ({ __sse128_u t; t.vec = x; t.as_u32[1]; })
|
|
#define sse128_l2(x) ({ __sse128_u t; t.vec = x; t.as_u32[2]; })
|
|
#define sse128_l3(x) ({ __sse128_u t; t.vec = x; t.as_u32[3]; })
|
|
#define sse128(lo, hi) ({ __sse128_u t; t.as_u64[0] = lo; t.as_u64[1] = hi; t.vec; })
|
|
|
|
typedef u32 __attribute__((vector_size(32))) avx256_t;
|
|
|
|
static inline void _kvm_read_avx_reg(int reg, avx256_t *data)
|
|
{
|
|
switch (reg) {
|
|
case 0: asm("vmovdqa %%ymm0, %0" : "=m"(*data)); break;
|
|
case 1: asm("vmovdqa %%ymm1, %0" : "=m"(*data)); break;
|
|
case 2: asm("vmovdqa %%ymm2, %0" : "=m"(*data)); break;
|
|
case 3: asm("vmovdqa %%ymm3, %0" : "=m"(*data)); break;
|
|
case 4: asm("vmovdqa %%ymm4, %0" : "=m"(*data)); break;
|
|
case 5: asm("vmovdqa %%ymm5, %0" : "=m"(*data)); break;
|
|
case 6: asm("vmovdqa %%ymm6, %0" : "=m"(*data)); break;
|
|
case 7: asm("vmovdqa %%ymm7, %0" : "=m"(*data)); break;
|
|
#ifdef CONFIG_X86_64
|
|
case 8: asm("vmovdqa %%ymm8, %0" : "=m"(*data)); break;
|
|
case 9: asm("vmovdqa %%ymm9, %0" : "=m"(*data)); break;
|
|
case 10: asm("vmovdqa %%ymm10, %0" : "=m"(*data)); break;
|
|
case 11: asm("vmovdqa %%ymm11, %0" : "=m"(*data)); break;
|
|
case 12: asm("vmovdqa %%ymm12, %0" : "=m"(*data)); break;
|
|
case 13: asm("vmovdqa %%ymm13, %0" : "=m"(*data)); break;
|
|
case 14: asm("vmovdqa %%ymm14, %0" : "=m"(*data)); break;
|
|
case 15: asm("vmovdqa %%ymm15, %0" : "=m"(*data)); break;
|
|
#endif
|
|
default: BUG();
|
|
}
|
|
}
|
|
|
|
static inline void _kvm_write_avx_reg(int reg, const avx256_t *data)
|
|
{
|
|
switch (reg) {
|
|
case 0: asm("vmovdqa %0, %%ymm0" : : "m"(*data)); break;
|
|
case 1: asm("vmovdqa %0, %%ymm1" : : "m"(*data)); break;
|
|
case 2: asm("vmovdqa %0, %%ymm2" : : "m"(*data)); break;
|
|
case 3: asm("vmovdqa %0, %%ymm3" : : "m"(*data)); break;
|
|
case 4: asm("vmovdqa %0, %%ymm4" : : "m"(*data)); break;
|
|
case 5: asm("vmovdqa %0, %%ymm5" : : "m"(*data)); break;
|
|
case 6: asm("vmovdqa %0, %%ymm6" : : "m"(*data)); break;
|
|
case 7: asm("vmovdqa %0, %%ymm7" : : "m"(*data)); break;
|
|
#ifdef CONFIG_X86_64
|
|
case 8: asm("vmovdqa %0, %%ymm8" : : "m"(*data)); break;
|
|
case 9: asm("vmovdqa %0, %%ymm9" : : "m"(*data)); break;
|
|
case 10: asm("vmovdqa %0, %%ymm10" : : "m"(*data)); break;
|
|
case 11: asm("vmovdqa %0, %%ymm11" : : "m"(*data)); break;
|
|
case 12: asm("vmovdqa %0, %%ymm12" : : "m"(*data)); break;
|
|
case 13: asm("vmovdqa %0, %%ymm13" : : "m"(*data)); break;
|
|
case 14: asm("vmovdqa %0, %%ymm14" : : "m"(*data)); break;
|
|
case 15: asm("vmovdqa %0, %%ymm15" : : "m"(*data)); break;
|
|
#endif
|
|
default: BUG();
|
|
}
|
|
}
|
|
|
|
static inline void _kvm_read_sse_reg(int reg, sse128_t *data)
|
|
{
|
|
switch (reg) {
|
|
case 0: asm("movdqa %%xmm0, %0" : "=m"(*data)); break;
|
|
case 1: asm("movdqa %%xmm1, %0" : "=m"(*data)); break;
|
|
case 2: asm("movdqa %%xmm2, %0" : "=m"(*data)); break;
|
|
case 3: asm("movdqa %%xmm3, %0" : "=m"(*data)); break;
|
|
case 4: asm("movdqa %%xmm4, %0" : "=m"(*data)); break;
|
|
case 5: asm("movdqa %%xmm5, %0" : "=m"(*data)); break;
|
|
case 6: asm("movdqa %%xmm6, %0" : "=m"(*data)); break;
|
|
case 7: asm("movdqa %%xmm7, %0" : "=m"(*data)); break;
|
|
#ifdef CONFIG_X86_64
|
|
case 8: asm("movdqa %%xmm8, %0" : "=m"(*data)); break;
|
|
case 9: asm("movdqa %%xmm9, %0" : "=m"(*data)); break;
|
|
case 10: asm("movdqa %%xmm10, %0" : "=m"(*data)); break;
|
|
case 11: asm("movdqa %%xmm11, %0" : "=m"(*data)); break;
|
|
case 12: asm("movdqa %%xmm12, %0" : "=m"(*data)); break;
|
|
case 13: asm("movdqa %%xmm13, %0" : "=m"(*data)); break;
|
|
case 14: asm("movdqa %%xmm14, %0" : "=m"(*data)); break;
|
|
case 15: asm("movdqa %%xmm15, %0" : "=m"(*data)); break;
|
|
#endif
|
|
default: BUG();
|
|
}
|
|
}
|
|
|
|
static inline void _kvm_write_sse_reg(int reg, const sse128_t *data)
|
|
{
|
|
switch (reg) {
|
|
case 0: asm("movdqa %0, %%xmm0" : : "m"(*data)); break;
|
|
case 1: asm("movdqa %0, %%xmm1" : : "m"(*data)); break;
|
|
case 2: asm("movdqa %0, %%xmm2" : : "m"(*data)); break;
|
|
case 3: asm("movdqa %0, %%xmm3" : : "m"(*data)); break;
|
|
case 4: asm("movdqa %0, %%xmm4" : : "m"(*data)); break;
|
|
case 5: asm("movdqa %0, %%xmm5" : : "m"(*data)); break;
|
|
case 6: asm("movdqa %0, %%xmm6" : : "m"(*data)); break;
|
|
case 7: asm("movdqa %0, %%xmm7" : : "m"(*data)); break;
|
|
#ifdef CONFIG_X86_64
|
|
case 8: asm("movdqa %0, %%xmm8" : : "m"(*data)); break;
|
|
case 9: asm("movdqa %0, %%xmm9" : : "m"(*data)); break;
|
|
case 10: asm("movdqa %0, %%xmm10" : : "m"(*data)); break;
|
|
case 11: asm("movdqa %0, %%xmm11" : : "m"(*data)); break;
|
|
case 12: asm("movdqa %0, %%xmm12" : : "m"(*data)); break;
|
|
case 13: asm("movdqa %0, %%xmm13" : : "m"(*data)); break;
|
|
case 14: asm("movdqa %0, %%xmm14" : : "m"(*data)); break;
|
|
case 15: asm("movdqa %0, %%xmm15" : : "m"(*data)); break;
|
|
#endif
|
|
default: BUG();
|
|
}
|
|
}
|
|
|
|
static inline void _kvm_read_mmx_reg(int reg, u64 *data)
|
|
{
|
|
switch (reg) {
|
|
case 0: asm("movq %%mm0, %0" : "=m"(*data)); break;
|
|
case 1: asm("movq %%mm1, %0" : "=m"(*data)); break;
|
|
case 2: asm("movq %%mm2, %0" : "=m"(*data)); break;
|
|
case 3: asm("movq %%mm3, %0" : "=m"(*data)); break;
|
|
case 4: asm("movq %%mm4, %0" : "=m"(*data)); break;
|
|
case 5: asm("movq %%mm5, %0" : "=m"(*data)); break;
|
|
case 6: asm("movq %%mm6, %0" : "=m"(*data)); break;
|
|
case 7: asm("movq %%mm7, %0" : "=m"(*data)); break;
|
|
default: BUG();
|
|
}
|
|
}
|
|
|
|
static inline void _kvm_write_mmx_reg(int reg, const u64 *data)
|
|
{
|
|
switch (reg) {
|
|
case 0: asm("movq %0, %%mm0" : : "m"(*data)); break;
|
|
case 1: asm("movq %0, %%mm1" : : "m"(*data)); break;
|
|
case 2: asm("movq %0, %%mm2" : : "m"(*data)); break;
|
|
case 3: asm("movq %0, %%mm3" : : "m"(*data)); break;
|
|
case 4: asm("movq %0, %%mm4" : : "m"(*data)); break;
|
|
case 5: asm("movq %0, %%mm5" : : "m"(*data)); break;
|
|
case 6: asm("movq %0, %%mm6" : : "m"(*data)); break;
|
|
case 7: asm("movq %0, %%mm7" : : "m"(*data)); break;
|
|
default: BUG();
|
|
}
|
|
}
|
|
|
|
static inline void kvm_fpu_get(void)
|
|
{
|
|
fpregs_lock();
|
|
|
|
fpregs_assert_state_consistent();
|
|
if (test_thread_flag(TIF_NEED_FPU_LOAD))
|
|
switch_fpu_return();
|
|
}
|
|
|
|
static inline void kvm_fpu_put(void)
|
|
{
|
|
fpregs_unlock();
|
|
}
|
|
|
|
static inline void kvm_read_avx_reg(int reg, avx256_t *data)
|
|
{
|
|
kvm_fpu_get();
|
|
_kvm_read_avx_reg(reg, data);
|
|
kvm_fpu_put();
|
|
}
|
|
|
|
static inline void kvm_write_avx_reg(int reg, const avx256_t *data)
|
|
{
|
|
kvm_fpu_get();
|
|
_kvm_write_avx_reg(reg, data);
|
|
kvm_fpu_put();
|
|
}
|
|
|
|
static inline void kvm_read_sse_reg(int reg, sse128_t *data)
|
|
{
|
|
kvm_fpu_get();
|
|
_kvm_read_sse_reg(reg, data);
|
|
kvm_fpu_put();
|
|
}
|
|
|
|
static inline void kvm_write_sse_reg(int reg, const sse128_t *data)
|
|
{
|
|
kvm_fpu_get();
|
|
_kvm_write_sse_reg(reg, data);
|
|
kvm_fpu_put();
|
|
}
|
|
|
|
static inline void kvm_read_mmx_reg(int reg, u64 *data)
|
|
{
|
|
kvm_fpu_get();
|
|
_kvm_read_mmx_reg(reg, data);
|
|
kvm_fpu_put();
|
|
}
|
|
|
|
static inline void kvm_write_mmx_reg(int reg, const u64 *data)
|
|
{
|
|
kvm_fpu_get();
|
|
_kvm_write_mmx_reg(reg, data);
|
|
kvm_fpu_put();
|
|
}
|
|
|
|
#endif
|