mirror of
https://github.com/xemu-project/xemu.git
synced 2024-12-11 21:44:04 +00:00
i386: hvf: unify register enums between HVF and the rest
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
This commit is contained in:
parent
f9fea77740
commit
6701d81d74
@ -56,30 +56,44 @@
|
||||
#include "fpu/softfloat.h"
|
||||
#endif
|
||||
|
||||
#define R_EAX 0
|
||||
#define R_ECX 1
|
||||
#define R_EDX 2
|
||||
#define R_EBX 3
|
||||
#define R_ESP 4
|
||||
#define R_EBP 5
|
||||
#define R_ESI 6
|
||||
#define R_EDI 7
|
||||
enum {
|
||||
R_EAX = 0,
|
||||
R_ECX = 1,
|
||||
R_EDX = 2,
|
||||
R_EBX = 3,
|
||||
R_ESP = 4,
|
||||
R_EBP = 5,
|
||||
R_ESI = 6,
|
||||
R_EDI = 7,
|
||||
R_R8 = 8,
|
||||
R_R9 = 9,
|
||||
R_R10 = 10,
|
||||
R_R11 = 11,
|
||||
R_R12 = 12,
|
||||
R_R13 = 13,
|
||||
R_R14 = 14,
|
||||
R_R15 = 15,
|
||||
|
||||
#define R_AL 0
|
||||
#define R_CL 1
|
||||
#define R_DL 2
|
||||
#define R_BL 3
|
||||
#define R_AH 4
|
||||
#define R_CH 5
|
||||
#define R_DH 6
|
||||
#define R_BH 7
|
||||
R_AL = 0,
|
||||
R_CL = 1,
|
||||
R_DL = 2,
|
||||
R_BL = 3,
|
||||
R_AH = 4,
|
||||
R_CH = 5,
|
||||
R_DH = 6,
|
||||
R_BH = 7,
|
||||
};
|
||||
|
||||
#define R_ES 0
|
||||
#define R_CS 1
|
||||
#define R_SS 2
|
||||
#define R_DS 3
|
||||
#define R_FS 4
|
||||
#define R_GS 5
|
||||
typedef enum X86Seg {
|
||||
R_ES = 0,
|
||||
R_CS = 1,
|
||||
R_SS = 2,
|
||||
R_DS = 3,
|
||||
R_FS = 4,
|
||||
R_GS = 5,
|
||||
R_LDTR = 6,
|
||||
R_TR = 7,
|
||||
} X86Seg;
|
||||
|
||||
/* segment descriptor fields */
|
||||
#define DESC_G_SHIFT 23
|
||||
|
@ -88,14 +88,14 @@ static void enter_long_mode(hv_vcpuid_t vcpu, uint64_t cr0, uint64_t efer)
|
||||
{
|
||||
uint64_t entry_ctls;
|
||||
|
||||
efer |= EFER_LMA;
|
||||
efer |= MSR_EFER_LMA;
|
||||
wvmcs(vcpu, VMCS_GUEST_IA32_EFER, efer);
|
||||
entry_ctls = rvmcs(vcpu, VMCS_ENTRY_CTLS);
|
||||
wvmcs(vcpu, VMCS_ENTRY_CTLS, rvmcs(vcpu, VMCS_ENTRY_CTLS) |
|
||||
VM_ENTRY_GUEST_LMA);
|
||||
|
||||
uint64_t guest_tr_ar = rvmcs(vcpu, VMCS_GUEST_TR_ACCESS_RIGHTS);
|
||||
if ((efer & EFER_LME) &&
|
||||
if ((efer & MSR_EFER_LME) &&
|
||||
(guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) {
|
||||
wvmcs(vcpu, VMCS_GUEST_TR_ACCESS_RIGHTS,
|
||||
(guest_tr_ar & ~AR_TYPE_MASK) | AR_TYPE_BUSY_64_TSS);
|
||||
@ -109,7 +109,7 @@ static void exit_long_mode(hv_vcpuid_t vcpu, uint64_t cr0, uint64_t efer)
|
||||
entry_ctls = rvmcs(vcpu, VMCS_ENTRY_CTLS);
|
||||
wvmcs(vcpu, VMCS_ENTRY_CTLS, entry_ctls & ~VM_ENTRY_GUEST_LMA);
|
||||
|
||||
efer &= ~EFER_LMA;
|
||||
efer &= ~MSR_EFER_LMA;
|
||||
wvmcs(vcpu, VMCS_GUEST_IA32_EFER, efer);
|
||||
}
|
||||
|
||||
@ -121,7 +121,7 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
|
||||
uint64_t old_cr0 = rvmcs(vcpu, VMCS_GUEST_CR0);
|
||||
|
||||
if ((cr0 & CR0_PG) && (rvmcs(vcpu, VMCS_GUEST_CR4) & CR4_PAE) &&
|
||||
!(efer & EFER_LME)) {
|
||||
!(efer & MSR_EFER_LME)) {
|
||||
address_space_rw(&address_space_memory,
|
||||
rvmcs(vcpu, VMCS_GUEST_CR3) & ~0x1f,
|
||||
MEMTXATTRS_UNSPECIFIED,
|
||||
@ -138,7 +138,7 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
|
||||
cr0 &= ~CR0_CD;
|
||||
wvmcs(vcpu, VMCS_GUEST_CR0, cr0 | CR0_NE | CR0_ET);
|
||||
|
||||
if (efer & EFER_LME) {
|
||||
if (efer & MSR_EFER_LME) {
|
||||
if (!(old_cr0 & CR0_PG) && (cr0 & CR0_PG)) {
|
||||
enter_long_mode(vcpu, cr0, efer);
|
||||
}
|
||||
|
@ -134,13 +134,13 @@ bool x86_is_v8086(struct CPUState *cpu)
|
||||
|
||||
bool x86_is_long_mode(struct CPUState *cpu)
|
||||
{
|
||||
return rvmcs(cpu->hvf_fd, VMCS_GUEST_IA32_EFER) & EFER_LMA;
|
||||
return rvmcs(cpu->hvf_fd, VMCS_GUEST_IA32_EFER) & MSR_EFER_LMA;
|
||||
}
|
||||
|
||||
bool x86_is_long64_mode(struct CPUState *cpu)
|
||||
{
|
||||
struct vmx_segment desc;
|
||||
vmx_read_segment_descriptor(cpu, &desc, REG_SEG_CS);
|
||||
vmx_read_segment_descriptor(cpu, &desc, R_CS);
|
||||
|
||||
return x86_is_long_mode(cpu) && ((desc.ar >> 13) & 1);
|
||||
}
|
||||
@ -157,13 +157,13 @@ bool x86_is_pae_enabled(struct CPUState *cpu)
|
||||
return cr4 & CR4_PAE;
|
||||
}
|
||||
|
||||
addr_t linear_addr(struct CPUState *cpu, addr_t addr, x86_reg_segment seg)
|
||||
addr_t linear_addr(struct CPUState *cpu, addr_t addr, X86Seg seg)
|
||||
{
|
||||
return vmx_read_segment_base(cpu, seg) + addr;
|
||||
}
|
||||
|
||||
addr_t linear_addr_size(struct CPUState *cpu, addr_t addr, int size,
|
||||
x86_reg_segment seg)
|
||||
X86Seg seg)
|
||||
{
|
||||
switch (size) {
|
||||
case 2:
|
||||
@ -180,5 +180,5 @@ addr_t linear_addr_size(struct CPUState *cpu, addr_t addr, int size,
|
||||
|
||||
addr_t linear_rip(struct CPUState *cpu, addr_t rip)
|
||||
{
|
||||
return linear_addr(cpu, rip, REG_SEG_CS);
|
||||
return linear_addr(cpu, rip, R_CS);
|
||||
}
|
||||
|
@ -21,57 +21,6 @@
|
||||
|
||||
#include "x86_gen.h"
|
||||
|
||||
/* exceptions */
|
||||
typedef enum x86_exception {
|
||||
EXCEPTION_DE, /* divide error */
|
||||
EXCEPTION_DB, /* debug fault */
|
||||
EXCEPTION_NMI, /* non-maskable interrupt */
|
||||
EXCEPTION_BP, /* breakpoint trap */
|
||||
EXCEPTION_OF, /* overflow trap */
|
||||
EXCEPTION_BR, /* boundary range exceeded fault */
|
||||
EXCEPTION_UD, /* undefined opcode */
|
||||
EXCEPTION_NM, /* device not available */
|
||||
EXCEPTION_DF, /* double fault */
|
||||
EXCEPTION_RSVD, /* not defined */
|
||||
EXCEPTION_TS, /* invalid TSS fault */
|
||||
EXCEPTION_NP, /* not present fault */
|
||||
EXCEPTION_GP, /* general protection fault */
|
||||
EXCEPTION_PF, /* page fault */
|
||||
EXCEPTION_RSVD2, /* not defined */
|
||||
} x86_exception;
|
||||
|
||||
/* general purpose regs */
|
||||
typedef enum x86_reg_name {
|
||||
REG_RAX = 0,
|
||||
REG_RCX = 1,
|
||||
REG_RDX = 2,
|
||||
REG_RBX = 3,
|
||||
REG_RSP = 4,
|
||||
REG_RBP = 5,
|
||||
REG_RSI = 6,
|
||||
REG_RDI = 7,
|
||||
REG_R8 = 8,
|
||||
REG_R9 = 9,
|
||||
REG_R10 = 10,
|
||||
REG_R11 = 11,
|
||||
REG_R12 = 12,
|
||||
REG_R13 = 13,
|
||||
REG_R14 = 14,
|
||||
REG_R15 = 15,
|
||||
} x86_reg_name;
|
||||
|
||||
/* segment regs */
|
||||
typedef enum x86_reg_segment {
|
||||
REG_SEG_ES = 0,
|
||||
REG_SEG_CS = 1,
|
||||
REG_SEG_SS = 2,
|
||||
REG_SEG_DS = 3,
|
||||
REG_SEG_FS = 4,
|
||||
REG_SEG_GS = 5,
|
||||
REG_SEG_LDTR = 6,
|
||||
REG_SEG_TR = 7,
|
||||
} x86_reg_segment;
|
||||
|
||||
typedef struct x86_register {
|
||||
union {
|
||||
struct {
|
||||
@ -153,15 +102,6 @@ typedef struct x86_reg_flags {
|
||||
};
|
||||
} __attribute__ ((__packed__)) x86_reg_flags;
|
||||
|
||||
typedef enum x86_reg_efer {
|
||||
EFER_SCE = (1L << 0),
|
||||
EFER_LME = (1L << 8),
|
||||
EFER_LMA = (1L << 10),
|
||||
EFER_NXE = (1L << 11),
|
||||
EFER_SVME = (1L << 12),
|
||||
EFER_FXSR = (1L << 14),
|
||||
} x86_reg_efer;
|
||||
|
||||
typedef struct x86_efer {
|
||||
uint64_t efer;
|
||||
} __attribute__ ((__packed__)) x86_efer;
|
||||
@ -376,54 +316,54 @@ struct HVFX86EmulatorState {
|
||||
#define EFLAGS(cpu) (cpu->hvf_emul->rflags.eflags)
|
||||
|
||||
#define RRX(cpu, reg) (cpu->hvf_emul->regs[reg].rrx)
|
||||
#define RAX(cpu) RRX(cpu, REG_RAX)
|
||||
#define RCX(cpu) RRX(cpu, REG_RCX)
|
||||
#define RDX(cpu) RRX(cpu, REG_RDX)
|
||||
#define RBX(cpu) RRX(cpu, REG_RBX)
|
||||
#define RSP(cpu) RRX(cpu, REG_RSP)
|
||||
#define RBP(cpu) RRX(cpu, REG_RBP)
|
||||
#define RSI(cpu) RRX(cpu, REG_RSI)
|
||||
#define RDI(cpu) RRX(cpu, REG_RDI)
|
||||
#define R8(cpu) RRX(cpu, REG_R8)
|
||||
#define R9(cpu) RRX(cpu, REG_R9)
|
||||
#define R10(cpu) RRX(cpu, REG_R10)
|
||||
#define R11(cpu) RRX(cpu, REG_R11)
|
||||
#define R12(cpu) RRX(cpu, REG_R12)
|
||||
#define R13(cpu) RRX(cpu, REG_R13)
|
||||
#define R14(cpu) RRX(cpu, REG_R14)
|
||||
#define R15(cpu) RRX(cpu, REG_R15)
|
||||
#define RAX(cpu) RRX(cpu, R_EAX)
|
||||
#define RCX(cpu) RRX(cpu, R_ECX)
|
||||
#define RDX(cpu) RRX(cpu, R_EDX)
|
||||
#define RBX(cpu) RRX(cpu, R_EBX)
|
||||
#define RSP(cpu) RRX(cpu, R_ESP)
|
||||
#define RBP(cpu) RRX(cpu, R_EBP)
|
||||
#define RSI(cpu) RRX(cpu, R_ESI)
|
||||
#define RDI(cpu) RRX(cpu, R_EDI)
|
||||
#define R8(cpu) RRX(cpu, R_R8)
|
||||
#define R9(cpu) RRX(cpu, R_R9)
|
||||
#define R10(cpu) RRX(cpu, R_R10)
|
||||
#define R11(cpu) RRX(cpu, R_R11)
|
||||
#define R12(cpu) RRX(cpu, R_R12)
|
||||
#define R13(cpu) RRX(cpu, R_R13)
|
||||
#define R14(cpu) RRX(cpu, R_R14)
|
||||
#define R15(cpu) RRX(cpu, R_R15)
|
||||
|
||||
#define ERX(cpu, reg) (cpu->hvf_emul->regs[reg].erx)
|
||||
#define EAX(cpu) ERX(cpu, REG_RAX)
|
||||
#define ECX(cpu) ERX(cpu, REG_RCX)
|
||||
#define EDX(cpu) ERX(cpu, REG_RDX)
|
||||
#define EBX(cpu) ERX(cpu, REG_RBX)
|
||||
#define ESP(cpu) ERX(cpu, REG_RSP)
|
||||
#define EBP(cpu) ERX(cpu, REG_RBP)
|
||||
#define ESI(cpu) ERX(cpu, REG_RSI)
|
||||
#define EDI(cpu) ERX(cpu, REG_RDI)
|
||||
#define EAX(cpu) ERX(cpu, R_EAX)
|
||||
#define ECX(cpu) ERX(cpu, R_ECX)
|
||||
#define EDX(cpu) ERX(cpu, R_EDX)
|
||||
#define EBX(cpu) ERX(cpu, R_EBX)
|
||||
#define ESP(cpu) ERX(cpu, R_ESP)
|
||||
#define EBP(cpu) ERX(cpu, R_EBP)
|
||||
#define ESI(cpu) ERX(cpu, R_ESI)
|
||||
#define EDI(cpu) ERX(cpu, R_EDI)
|
||||
|
||||
#define RX(cpu, reg) (cpu->hvf_emul->regs[reg].rx)
|
||||
#define AX(cpu) RX(cpu, REG_RAX)
|
||||
#define CX(cpu) RX(cpu, REG_RCX)
|
||||
#define DX(cpu) RX(cpu, REG_RDX)
|
||||
#define BP(cpu) RX(cpu, REG_RBP)
|
||||
#define SP(cpu) RX(cpu, REG_RSP)
|
||||
#define BX(cpu) RX(cpu, REG_RBX)
|
||||
#define SI(cpu) RX(cpu, REG_RSI)
|
||||
#define DI(cpu) RX(cpu, REG_RDI)
|
||||
#define AX(cpu) RX(cpu, R_EAX)
|
||||
#define CX(cpu) RX(cpu, R_ECX)
|
||||
#define DX(cpu) RX(cpu, R_EDX)
|
||||
#define BP(cpu) RX(cpu, R_EBP)
|
||||
#define SP(cpu) RX(cpu, R_ESP)
|
||||
#define BX(cpu) RX(cpu, R_EBX)
|
||||
#define SI(cpu) RX(cpu, R_ESI)
|
||||
#define DI(cpu) RX(cpu, R_EDI)
|
||||
|
||||
#define RL(cpu, reg) (cpu->hvf_emul->regs[reg].lx)
|
||||
#define AL(cpu) RL(cpu, REG_RAX)
|
||||
#define CL(cpu) RL(cpu, REG_RCX)
|
||||
#define DL(cpu) RL(cpu, REG_RDX)
|
||||
#define BL(cpu) RL(cpu, REG_RBX)
|
||||
#define AL(cpu) RL(cpu, R_EAX)
|
||||
#define CL(cpu) RL(cpu, R_ECX)
|
||||
#define DL(cpu) RL(cpu, R_EDX)
|
||||
#define BL(cpu) RL(cpu, R_EBX)
|
||||
|
||||
#define RH(cpu, reg) (cpu->hvf_emul->regs[reg].hx)
|
||||
#define AH(cpu) RH(cpu, REG_RAX)
|
||||
#define CH(cpu) RH(cpu, REG_RCX)
|
||||
#define DH(cpu) RH(cpu, REG_RDX)
|
||||
#define BH(cpu) RH(cpu, REG_RBX)
|
||||
#define AH(cpu) RH(cpu, R_EAX)
|
||||
#define CH(cpu) RH(cpu, R_ECX)
|
||||
#define DH(cpu) RH(cpu, R_EDX)
|
||||
#define BH(cpu) RH(cpu, R_EBX)
|
||||
|
||||
/* deal with GDT/LDT descriptors in memory */
|
||||
bool x86_read_segment_descriptor(struct CPUState *cpu,
|
||||
@ -445,9 +385,10 @@ bool x86_is_long64_mode(struct CPUState *cpu);
|
||||
bool x86_is_paging_mode(struct CPUState *cpu);
|
||||
bool x86_is_pae_enabled(struct CPUState *cpu);
|
||||
|
||||
addr_t linear_addr(struct CPUState *cpu, addr_t addr, x86_reg_segment seg);
|
||||
enum X86Seg;
|
||||
addr_t linear_addr(struct CPUState *cpu, addr_t addr, enum X86Seg seg);
|
||||
addr_t linear_addr_size(struct CPUState *cpu, addr_t addr, int size,
|
||||
x86_reg_segment seg);
|
||||
enum X86Seg seg);
|
||||
addr_t linear_rip(struct CPUState *cpu, addr_t rip);
|
||||
|
||||
static inline uint64_t rdtscp(void)
|
||||
|
@ -121,7 +121,7 @@ static void decode_rax(CPUX86State *env, struct x86_decode *decode,
|
||||
struct x86_decode_op *op)
|
||||
{
|
||||
op->type = X86_VAR_REG;
|
||||
op->reg = REG_RAX;
|
||||
op->reg = R_EAX;
|
||||
op->ptr = get_reg_ref(env, op->reg, 0, decode->operand_size);
|
||||
}
|
||||
|
||||
@ -213,22 +213,22 @@ static void decode_pushseg(CPUX86State *env, struct x86_decode *decode)
|
||||
decode->op[0].type = X86_VAR_REG;
|
||||
switch (op) {
|
||||
case 0xe:
|
||||
decode->op[0].reg = REG_SEG_CS;
|
||||
decode->op[0].reg = R_CS;
|
||||
break;
|
||||
case 0x16:
|
||||
decode->op[0].reg = REG_SEG_SS;
|
||||
decode->op[0].reg = R_SS;
|
||||
break;
|
||||
case 0x1e:
|
||||
decode->op[0].reg = REG_SEG_DS;
|
||||
decode->op[0].reg = R_DS;
|
||||
break;
|
||||
case 0x06:
|
||||
decode->op[0].reg = REG_SEG_ES;
|
||||
decode->op[0].reg = R_ES;
|
||||
break;
|
||||
case 0xa0:
|
||||
decode->op[0].reg = REG_SEG_FS;
|
||||
decode->op[0].reg = R_FS;
|
||||
break;
|
||||
case 0xa8:
|
||||
decode->op[0].reg = REG_SEG_GS;
|
||||
decode->op[0].reg = R_GS;
|
||||
break;
|
||||
}
|
||||
}
|
||||
@ -240,22 +240,22 @@ static void decode_popseg(CPUX86State *env, struct x86_decode *decode)
|
||||
decode->op[0].type = X86_VAR_REG;
|
||||
switch (op) {
|
||||
case 0xf:
|
||||
decode->op[0].reg = REG_SEG_CS;
|
||||
decode->op[0].reg = R_CS;
|
||||
break;
|
||||
case 0x17:
|
||||
decode->op[0].reg = REG_SEG_SS;
|
||||
decode->op[0].reg = R_SS;
|
||||
break;
|
||||
case 0x1f:
|
||||
decode->op[0].reg = REG_SEG_DS;
|
||||
decode->op[0].reg = R_DS;
|
||||
break;
|
||||
case 0x07:
|
||||
decode->op[0].reg = REG_SEG_ES;
|
||||
decode->op[0].reg = R_ES;
|
||||
break;
|
||||
case 0xa1:
|
||||
decode->op[0].reg = REG_SEG_FS;
|
||||
decode->op[0].reg = R_FS;
|
||||
break;
|
||||
case 0xa9:
|
||||
decode->op[0].reg = REG_SEG_GS;
|
||||
decode->op[0].reg = R_GS;
|
||||
break;
|
||||
}
|
||||
}
|
||||
@ -412,7 +412,7 @@ static void decode_rcx(CPUX86State *env, struct x86_decode *decode,
|
||||
struct x86_decode_op *op)
|
||||
{
|
||||
op->type = X86_VAR_REG;
|
||||
op->reg = REG_RCX;
|
||||
op->reg = R_ECX;
|
||||
op->ptr = get_reg_ref(env, op->reg, decode->rex.b, decode->operand_size);
|
||||
}
|
||||
|
||||
@ -1639,7 +1639,7 @@ void calc_modrm_operand16(CPUX86State *env, struct x86_decode *decode,
|
||||
struct x86_decode_op *op)
|
||||
{
|
||||
addr_t ptr = 0;
|
||||
x86_reg_segment seg = REG_SEG_DS;
|
||||
X86Seg seg = R_DS;
|
||||
|
||||
if (!decode->modrm.mod && 6 == decode->modrm.rm) {
|
||||
op->ptr = (uint16_t)decode->displacement;
|
||||
@ -1659,11 +1659,11 @@ void calc_modrm_operand16(CPUX86State *env, struct x86_decode *decode,
|
||||
break;
|
||||
case 2:
|
||||
ptr += BP(env) + SI(env);
|
||||
seg = REG_SEG_SS;
|
||||
seg = R_SS;
|
||||
break;
|
||||
case 3:
|
||||
ptr += BP(env) + DI(env);
|
||||
seg = REG_SEG_SS;
|
||||
seg = R_SS;
|
||||
break;
|
||||
case 4:
|
||||
ptr += SI(env);
|
||||
@ -1673,7 +1673,7 @@ void calc_modrm_operand16(CPUX86State *env, struct x86_decode *decode,
|
||||
break;
|
||||
case 6:
|
||||
ptr += BP(env);
|
||||
seg = REG_SEG_SS;
|
||||
seg = R_SS;
|
||||
break;
|
||||
case 7:
|
||||
ptr += BX(env);
|
||||
@ -1693,7 +1693,7 @@ addr_t get_reg_ref(CPUX86State *env, int reg, int is_extended, int size)
|
||||
int which = 0;
|
||||
|
||||
if (is_extended) {
|
||||
reg |= REG_R8;
|
||||
reg |= R_R8;
|
||||
}
|
||||
|
||||
|
||||
@ -1723,7 +1723,7 @@ addr_t get_reg_val(CPUX86State *env, int reg, int is_extended, int size)
|
||||
}
|
||||
|
||||
static addr_t get_sib_val(CPUX86State *env, struct x86_decode *decode,
|
||||
x86_reg_segment *sel)
|
||||
X86Seg *sel)
|
||||
{
|
||||
addr_t base = 0;
|
||||
addr_t scaled_index = 0;
|
||||
@ -1731,23 +1731,23 @@ static addr_t get_sib_val(CPUX86State *env, struct x86_decode *decode,
|
||||
int base_reg = decode->sib.base;
|
||||
int index_reg = decode->sib.index;
|
||||
|
||||
*sel = REG_SEG_DS;
|
||||
*sel = R_DS;
|
||||
|
||||
if (decode->modrm.mod || base_reg != REG_RBP) {
|
||||
if (decode->modrm.mod || base_reg != R_EBP) {
|
||||
if (decode->rex.b) {
|
||||
base_reg |= REG_R8;
|
||||
base_reg |= R_R8;
|
||||
}
|
||||
if (REG_RSP == base_reg || REG_RBP == base_reg) {
|
||||
*sel = REG_SEG_SS;
|
||||
if (base_reg == R_ESP || base_reg == R_EBP) {
|
||||
*sel = R_SS;
|
||||
}
|
||||
base = get_reg_val(env, decode->sib.base, decode->rex.b, addr_size);
|
||||
}
|
||||
|
||||
if (decode->rex.x) {
|
||||
index_reg |= REG_R8;
|
||||
index_reg |= R_R8;
|
||||
}
|
||||
|
||||
if (index_reg != REG_RSP) {
|
||||
if (index_reg != R_ESP) {
|
||||
scaled_index = get_reg_val(env, index_reg, decode->rex.x, addr_size) <<
|
||||
decode->sib.scale;
|
||||
}
|
||||
@ -1757,7 +1757,7 @@ static addr_t get_sib_val(CPUX86State *env, struct x86_decode *decode,
|
||||
void calc_modrm_operand32(CPUX86State *env, struct x86_decode *decode,
|
||||
struct x86_decode_op *op)
|
||||
{
|
||||
x86_reg_segment seg = REG_SEG_DS;
|
||||
X86Seg seg = R_DS;
|
||||
addr_t ptr = 0;
|
||||
int addr_size = decode->addressing_size;
|
||||
|
||||
@ -1774,8 +1774,8 @@ void calc_modrm_operand32(CPUX86State *env, struct x86_decode *decode,
|
||||
ptr = decode->displacement;
|
||||
}
|
||||
} else {
|
||||
if (REG_RBP == decode->modrm.rm || REG_RSP == decode->modrm.rm) {
|
||||
seg = REG_SEG_SS;
|
||||
if (decode->modrm.rm == R_EBP || decode->modrm.rm == R_ESP) {
|
||||
seg = R_SS;
|
||||
}
|
||||
ptr += get_reg_val(env, decode->modrm.rm, decode->rex.b, addr_size);
|
||||
}
|
||||
@ -1790,7 +1790,7 @@ void calc_modrm_operand32(CPUX86State *env, struct x86_decode *decode,
|
||||
void calc_modrm_operand64(CPUX86State *env, struct x86_decode *decode,
|
||||
struct x86_decode_op *op)
|
||||
{
|
||||
x86_reg_segment seg = REG_SEG_DS;
|
||||
X86Seg seg = R_DS;
|
||||
int32_t offset = 0;
|
||||
int mod = decode->modrm.mod;
|
||||
int rm = decode->modrm.rm;
|
||||
@ -1895,7 +1895,7 @@ void set_addressing_size(CPUX86State *env, struct x86_decode *decode)
|
||||
} else if (!x86_is_long_mode(ENV_GET_CPU(env))) {
|
||||
/* protected */
|
||||
struct vmx_segment cs;
|
||||
vmx_read_segment_descriptor(ENV_GET_CPU(env), &cs, REG_SEG_CS);
|
||||
vmx_read_segment_descriptor(ENV_GET_CPU(env), &cs, R_CS);
|
||||
/* check db */
|
||||
if ((cs.ar >> 14) & 1) {
|
||||
if (decode->addr_size_override) {
|
||||
@ -1932,7 +1932,7 @@ void set_operand_size(CPUX86State *env, struct x86_decode *decode)
|
||||
} else if (!x86_is_long_mode(ENV_GET_CPU(env))) {
|
||||
/* protected */
|
||||
struct vmx_segment cs;
|
||||
vmx_read_segment_descriptor(ENV_GET_CPU(env), &cs, REG_SEG_CS);
|
||||
vmx_read_segment_descriptor(ENV_GET_CPU(env), &cs, R_CS);
|
||||
/* check db */
|
||||
if ((cs.ar >> 14) & 1) {
|
||||
if (decode->op_size_override) {
|
||||
@ -2159,26 +2159,26 @@ const char *decode_cmd_to_string(enum x86_decode_cmd cmd)
|
||||
}
|
||||
|
||||
addr_t decode_linear_addr(CPUX86State *env, struct x86_decode *decode,
|
||||
addr_t addr, x86_reg_segment seg)
|
||||
addr_t addr, X86Seg seg)
|
||||
{
|
||||
switch (decode->segment_override) {
|
||||
case PREFIX_CS_SEG_OVEERIDE:
|
||||
seg = REG_SEG_CS;
|
||||
seg = R_CS;
|
||||
break;
|
||||
case PREFIX_SS_SEG_OVEERIDE:
|
||||
seg = REG_SEG_SS;
|
||||
seg = R_SS;
|
||||
break;
|
||||
case PREFIX_DS_SEG_OVEERIDE:
|
||||
seg = REG_SEG_DS;
|
||||
seg = R_DS;
|
||||
break;
|
||||
case PREFIX_ES_SEG_OVEERIDE:
|
||||
seg = REG_SEG_ES;
|
||||
seg = R_ES;
|
||||
break;
|
||||
case PREFIX_FS_SEG_OVEERIDE:
|
||||
seg = REG_SEG_FS;
|
||||
seg = R_FS;
|
||||
break;
|
||||
case PREFIX_GS_SEG_OVEERIDE:
|
||||
seg = REG_SEG_GS;
|
||||
seg = R_GS;
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
|
@ -308,7 +308,7 @@ addr_t get_reg_val(CPUX86State *env, int reg, int is_extended, int size);
|
||||
void calc_modrm_operand(CPUX86State *env, struct x86_decode *decode,
|
||||
struct x86_decode_op *op);
|
||||
addr_t decode_linear_addr(CPUX86State *env, struct x86_decode *decode,
|
||||
addr_t addr, x86_reg_segment seg);
|
||||
addr_t addr, enum X86Seg seg);
|
||||
|
||||
void init_decoder(void);
|
||||
void calc_modrm_operand16(CPUX86State *env, struct x86_decode *decode,
|
||||
|
@ -22,12 +22,12 @@
|
||||
#include "vmx.h"
|
||||
#include "x86_descr.h"
|
||||
|
||||
#define VMX_SEGMENT_FIELD(seg) \
|
||||
[REG_SEG_##seg] = { \
|
||||
.selector = VMCS_GUEST_##seg##_SELECTOR, \
|
||||
.base = VMCS_GUEST_##seg##_BASE, \
|
||||
.limit = VMCS_GUEST_##seg##_LIMIT, \
|
||||
.ar_bytes = VMCS_GUEST_##seg##_ACCESS_RIGHTS, \
|
||||
#define VMX_SEGMENT_FIELD(seg) \
|
||||
[R_##seg] = { \
|
||||
.selector = VMCS_GUEST_##seg##_SELECTOR, \
|
||||
.base = VMCS_GUEST_##seg##_BASE, \
|
||||
.limit = VMCS_GUEST_##seg##_LIMIT, \
|
||||
.ar_bytes = VMCS_GUEST_##seg##_ACCESS_RIGHTS, \
|
||||
}
|
||||
|
||||
static const struct vmx_segment_field {
|
||||
@ -46,34 +46,34 @@ static const struct vmx_segment_field {
|
||||
VMX_SEGMENT_FIELD(TR),
|
||||
};
|
||||
|
||||
uint32_t vmx_read_segment_limit(CPUState *cpu, x86_reg_segment seg)
|
||||
uint32_t vmx_read_segment_limit(CPUState *cpu, X86Seg seg)
|
||||
{
|
||||
return (uint32_t)rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].limit);
|
||||
}
|
||||
|
||||
uint32_t vmx_read_segment_ar(CPUState *cpu, x86_reg_segment seg)
|
||||
uint32_t vmx_read_segment_ar(CPUState *cpu, X86Seg seg)
|
||||
{
|
||||
return (uint32_t)rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].ar_bytes);
|
||||
}
|
||||
|
||||
uint64_t vmx_read_segment_base(CPUState *cpu, x86_reg_segment seg)
|
||||
uint64_t vmx_read_segment_base(CPUState *cpu, X86Seg seg)
|
||||
{
|
||||
return rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].base);
|
||||
}
|
||||
|
||||
x68_segment_selector vmx_read_segment_selector(CPUState *cpu, x86_reg_segment seg)
|
||||
x68_segment_selector vmx_read_segment_selector(CPUState *cpu, X86Seg seg)
|
||||
{
|
||||
x68_segment_selector sel;
|
||||
sel.sel = rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].selector);
|
||||
return sel;
|
||||
}
|
||||
|
||||
void vmx_write_segment_selector(struct CPUState *cpu, x68_segment_selector selector, x86_reg_segment seg)
|
||||
void vmx_write_segment_selector(struct CPUState *cpu, x68_segment_selector selector, X86Seg seg)
|
||||
{
|
||||
wvmcs(cpu->hvf_fd, vmx_segment_fields[seg].selector, selector.sel);
|
||||
}
|
||||
|
||||
void vmx_read_segment_descriptor(struct CPUState *cpu, struct vmx_segment *desc, x86_reg_segment seg)
|
||||
void vmx_read_segment_descriptor(struct CPUState *cpu, struct vmx_segment *desc, X86Seg seg)
|
||||
{
|
||||
desc->sel = rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].selector);
|
||||
desc->base = rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].base);
|
||||
@ -81,7 +81,7 @@ void vmx_read_segment_descriptor(struct CPUState *cpu, struct vmx_segment *desc,
|
||||
desc->ar = rvmcs(cpu->hvf_fd, vmx_segment_fields[seg].ar_bytes);
|
||||
}
|
||||
|
||||
void vmx_write_segment_descriptor(CPUState *cpu, struct vmx_segment *desc, x86_reg_segment seg)
|
||||
void vmx_write_segment_descriptor(CPUState *cpu, struct vmx_segment *desc, X86Seg seg)
|
||||
{
|
||||
const struct vmx_segment_field *sf = &vmx_segment_fields[seg];
|
||||
|
||||
|
@ -30,18 +30,18 @@ typedef struct vmx_segment {
|
||||
|
||||
/* deal with vmstate descriptors */
|
||||
void vmx_read_segment_descriptor(struct CPUState *cpu,
|
||||
struct vmx_segment *desc, x86_reg_segment seg);
|
||||
struct vmx_segment *desc, enum X86Seg seg);
|
||||
void vmx_write_segment_descriptor(CPUState *cpu, struct vmx_segment *desc,
|
||||
x86_reg_segment seg);
|
||||
enum X86Seg seg);
|
||||
|
||||
x68_segment_selector vmx_read_segment_selector(struct CPUState *cpu,
|
||||
x86_reg_segment seg);
|
||||
enum X86Seg seg);
|
||||
void vmx_write_segment_selector(struct CPUState *cpu,
|
||||
x68_segment_selector selector,
|
||||
x86_reg_segment seg);
|
||||
enum X86Seg seg);
|
||||
|
||||
uint64_t vmx_read_segment_base(struct CPUState *cpu, x86_reg_segment seg);
|
||||
void vmx_write_segment_base(struct CPUState *cpu, x86_reg_segment seg,
|
||||
uint64_t vmx_read_segment_base(struct CPUState *cpu, enum X86Seg seg);
|
||||
void vmx_write_segment_base(struct CPUState *cpu, enum X86Seg seg,
|
||||
uint64_t base);
|
||||
|
||||
void x86_segment_descriptor_to_vmx(struct CPUState *cpu,
|
||||
@ -49,8 +49,8 @@ void x86_segment_descriptor_to_vmx(struct CPUState *cpu,
|
||||
struct x86_segment_descriptor *desc,
|
||||
struct vmx_segment *vmx_desc);
|
||||
|
||||
uint32_t vmx_read_segment_limit(CPUState *cpu, x86_reg_segment seg);
|
||||
uint32_t vmx_read_segment_ar(CPUState *cpu, x86_reg_segment seg);
|
||||
uint32_t vmx_read_segment_limit(CPUState *cpu, enum X86Seg seg);
|
||||
uint32_t vmx_read_segment_ar(CPUState *cpu, enum X86Seg seg);
|
||||
void vmx_segment_to_x86_descriptor(struct CPUState *cpu,
|
||||
struct vmx_segment *vmx_desc,
|
||||
struct x86_segment_descriptor *desc);
|
||||
|
@ -294,7 +294,7 @@ static void fetch_operands(struct CPUX86State *env, struct x86_decode *decode,
|
||||
case X86_VAR_OFFSET:
|
||||
decode->op[i].ptr = decode_linear_addr(env, decode,
|
||||
decode->op[i].ptr,
|
||||
REG_SEG_DS);
|
||||
R_DS);
|
||||
if (calc_val[i]) {
|
||||
decode->op[i].val = read_val_ext(env, decode->op[i].ptr,
|
||||
decode->operand_size);
|
||||
@ -514,10 +514,10 @@ static inline void string_rep(struct CPUX86State *env, struct x86_decode *decode
|
||||
void (*func)(struct CPUX86State *env,
|
||||
struct x86_decode *ins), int rep)
|
||||
{
|
||||
addr_t rcx = read_reg(env, REG_RCX, decode->addressing_size);
|
||||
addr_t rcx = read_reg(env, R_ECX, decode->addressing_size);
|
||||
while (rcx--) {
|
||||
func(env, decode);
|
||||
write_reg(env, REG_RCX, rcx, decode->addressing_size);
|
||||
write_reg(env, R_ECX, rcx, decode->addressing_size);
|
||||
if ((PREFIX_REP == rep) && !get_ZF(env)) {
|
||||
break;
|
||||
}
|
||||
@ -530,13 +530,13 @@ static inline void string_rep(struct CPUX86State *env, struct x86_decode *decode
|
||||
static void exec_ins_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
{
|
||||
addr_t addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size,
|
||||
REG_SEG_ES);
|
||||
R_ES);
|
||||
|
||||
hvf_handle_io(ENV_GET_CPU(env), DX(env), env->hvf_emul->mmio_buf, 0,
|
||||
decode->operand_size, 1);
|
||||
vmx_write_mem(ENV_GET_CPU(env), addr, env->hvf_emul->mmio_buf, decode->operand_size);
|
||||
|
||||
string_increment_reg(env, REG_RDI, decode);
|
||||
string_increment_reg(env, R_EDI, decode);
|
||||
}
|
||||
|
||||
static void exec_ins(struct CPUX86State *env, struct x86_decode *decode)
|
||||
@ -552,13 +552,13 @@ static void exec_ins(struct CPUX86State *env, struct x86_decode *decode)
|
||||
|
||||
static void exec_outs_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
{
|
||||
addr_t addr = decode_linear_addr(env, decode, RSI(env), REG_SEG_DS);
|
||||
addr_t addr = decode_linear_addr(env, decode, RSI(env), R_DS);
|
||||
|
||||
vmx_read_mem(ENV_GET_CPU(env), env->hvf_emul->mmio_buf, addr, decode->operand_size);
|
||||
hvf_handle_io(ENV_GET_CPU(env), DX(env), env->hvf_emul->mmio_buf, 1,
|
||||
decode->operand_size, 1);
|
||||
|
||||
string_increment_reg(env, REG_RSI, decode);
|
||||
string_increment_reg(env, R_ESI, decode);
|
||||
}
|
||||
|
||||
static void exec_outs(struct CPUX86State *env, struct x86_decode *decode)
|
||||
@ -578,15 +578,15 @@ static void exec_movs_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
addr_t dst_addr;
|
||||
addr_t val;
|
||||
|
||||
src_addr = decode_linear_addr(env, decode, RSI(env), REG_SEG_DS);
|
||||
src_addr = decode_linear_addr(env, decode, RSI(env), R_DS);
|
||||
dst_addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size,
|
||||
REG_SEG_ES);
|
||||
R_ES);
|
||||
|
||||
val = read_val_ext(env, src_addr, decode->operand_size);
|
||||
write_val_ext(env, dst_addr, val, decode->operand_size);
|
||||
|
||||
string_increment_reg(env, REG_RSI, decode);
|
||||
string_increment_reg(env, REG_RDI, decode);
|
||||
string_increment_reg(env, R_ESI, decode);
|
||||
string_increment_reg(env, R_EDI, decode);
|
||||
}
|
||||
|
||||
static void exec_movs(struct CPUX86State *env, struct x86_decode *decode)
|
||||
@ -605,9 +605,9 @@ static void exec_cmps_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
addr_t src_addr;
|
||||
addr_t dst_addr;
|
||||
|
||||
src_addr = decode_linear_addr(env, decode, RSI(env), REG_SEG_DS);
|
||||
src_addr = decode_linear_addr(env, decode, RSI(env), R_DS);
|
||||
dst_addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size,
|
||||
REG_SEG_ES);
|
||||
R_ES);
|
||||
|
||||
decode->op[0].type = X86_VAR_IMMEDIATE;
|
||||
decode->op[0].val = read_val_ext(env, src_addr, decode->operand_size);
|
||||
@ -616,8 +616,8 @@ static void exec_cmps_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
|
||||
EXEC_2OP_ARITH_CMD(env, decode, -, SET_FLAGS_OSZAPC_SUB, false);
|
||||
|
||||
string_increment_reg(env, REG_RSI, decode);
|
||||
string_increment_reg(env, REG_RDI, decode);
|
||||
string_increment_reg(env, R_ESI, decode);
|
||||
string_increment_reg(env, R_EDI, decode);
|
||||
}
|
||||
|
||||
static void exec_cmps(struct CPUX86State *env, struct x86_decode *decode)
|
||||
@ -636,11 +636,11 @@ static void exec_stos_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
addr_t addr;
|
||||
addr_t val;
|
||||
|
||||
addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size, REG_SEG_ES);
|
||||
val = read_reg(env, REG_RAX, decode->operand_size);
|
||||
addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size, R_ES);
|
||||
val = read_reg(env, R_EAX, decode->operand_size);
|
||||
vmx_write_mem(ENV_GET_CPU(env), addr, &val, decode->operand_size);
|
||||
|
||||
string_increment_reg(env, REG_RDI, decode);
|
||||
string_increment_reg(env, R_EDI, decode);
|
||||
}
|
||||
|
||||
|
||||
@ -659,18 +659,18 @@ static void exec_scas_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
{
|
||||
addr_t addr;
|
||||
|
||||
addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size, REG_SEG_ES);
|
||||
addr = linear_addr_size(ENV_GET_CPU(env), RDI(env), decode->addressing_size, R_ES);
|
||||
decode->op[1].type = X86_VAR_IMMEDIATE;
|
||||
vmx_read_mem(ENV_GET_CPU(env), &decode->op[1].val, addr, decode->operand_size);
|
||||
|
||||
EXEC_2OP_ARITH_CMD(env, decode, -, SET_FLAGS_OSZAPC_SUB, false);
|
||||
string_increment_reg(env, REG_RDI, decode);
|
||||
string_increment_reg(env, R_EDI, decode);
|
||||
}
|
||||
|
||||
static void exec_scas(struct CPUX86State *env, struct x86_decode *decode)
|
||||
{
|
||||
decode->op[0].type = X86_VAR_REG;
|
||||
decode->op[0].reg = REG_RAX;
|
||||
decode->op[0].reg = R_EAX;
|
||||
if (decode->rep) {
|
||||
string_rep(env, decode, exec_scas_single, decode->rep);
|
||||
} else {
|
||||
@ -685,11 +685,11 @@ static void exec_lods_single(struct CPUX86State *env, struct x86_decode *decode)
|
||||
addr_t addr;
|
||||
addr_t val = 0;
|
||||
|
||||
addr = decode_linear_addr(env, decode, RSI(env), REG_SEG_DS);
|
||||
addr = decode_linear_addr(env, decode, RSI(env), R_DS);
|
||||
vmx_read_mem(ENV_GET_CPU(env), &val, addr, decode->operand_size);
|
||||
write_reg(env, REG_RAX, val, decode->operand_size);
|
||||
write_reg(env, R_EAX, val, decode->operand_size);
|
||||
|
||||
string_increment_reg(env, REG_RSI, decode);
|
||||
string_increment_reg(env, R_ESI, decode);
|
||||
}
|
||||
|
||||
static void exec_lods(struct CPUX86State *env, struct x86_decode *decode)
|
||||
@ -840,7 +840,7 @@ void simulate_wrmsr(struct CPUState *cpu)
|
||||
env->hvf_emul->efer.efer = data;
|
||||
/*printf("new efer %llx\n", EFER(cpu));*/
|
||||
wvmcs(cpu->hvf_fd, VMCS_GUEST_IA32_EFER, data);
|
||||
if (data & EFER_NXE) {
|
||||
if (data & MSR_EFER_NXE) {
|
||||
hv_vcpu_invalidate_tlb(cpu->hvf_fd);
|
||||
}
|
||||
break;
|
||||
@ -1465,14 +1465,14 @@ void load_regs(struct CPUState *cpu)
|
||||
CPUX86State *env = &x86_cpu->env;
|
||||
|
||||
int i = 0;
|
||||
RRX(env, REG_RAX) = rreg(cpu->hvf_fd, HV_X86_RAX);
|
||||
RRX(env, REG_RBX) = rreg(cpu->hvf_fd, HV_X86_RBX);
|
||||
RRX(env, REG_RCX) = rreg(cpu->hvf_fd, HV_X86_RCX);
|
||||
RRX(env, REG_RDX) = rreg(cpu->hvf_fd, HV_X86_RDX);
|
||||
RRX(env, REG_RSI) = rreg(cpu->hvf_fd, HV_X86_RSI);
|
||||
RRX(env, REG_RDI) = rreg(cpu->hvf_fd, HV_X86_RDI);
|
||||
RRX(env, REG_RSP) = rreg(cpu->hvf_fd, HV_X86_RSP);
|
||||
RRX(env, REG_RBP) = rreg(cpu->hvf_fd, HV_X86_RBP);
|
||||
RRX(env, R_EAX) = rreg(cpu->hvf_fd, HV_X86_RAX);
|
||||
RRX(env, R_EBX) = rreg(cpu->hvf_fd, HV_X86_RBX);
|
||||
RRX(env, R_ECX) = rreg(cpu->hvf_fd, HV_X86_RCX);
|
||||
RRX(env, R_EDX) = rreg(cpu->hvf_fd, HV_X86_RDX);
|
||||
RRX(env, R_ESI) = rreg(cpu->hvf_fd, HV_X86_RSI);
|
||||
RRX(env, R_EDI) = rreg(cpu->hvf_fd, HV_X86_RDI);
|
||||
RRX(env, R_ESP) = rreg(cpu->hvf_fd, HV_X86_RSP);
|
||||
RRX(env, R_EBP) = rreg(cpu->hvf_fd, HV_X86_RBP);
|
||||
for (i = 8; i < 16; i++) {
|
||||
RRX(env, i) = rreg(cpu->hvf_fd, HV_X86_RAX + i);
|
||||
}
|
||||
|
@ -54,12 +54,12 @@ static void save_state_to_tss32(CPUState *cpu, struct x86_tss_segment32 *tss)
|
||||
tss->esi = ESI(env);
|
||||
tss->edi = EDI(env);
|
||||
|
||||
tss->es = vmx_read_segment_selector(cpu, REG_SEG_ES).sel;
|
||||
tss->cs = vmx_read_segment_selector(cpu, REG_SEG_CS).sel;
|
||||
tss->ss = vmx_read_segment_selector(cpu, REG_SEG_SS).sel;
|
||||
tss->ds = vmx_read_segment_selector(cpu, REG_SEG_DS).sel;
|
||||
tss->fs = vmx_read_segment_selector(cpu, REG_SEG_FS).sel;
|
||||
tss->gs = vmx_read_segment_selector(cpu, REG_SEG_GS).sel;
|
||||
tss->es = vmx_read_segment_selector(cpu, R_ES).sel;
|
||||
tss->cs = vmx_read_segment_selector(cpu, R_CS).sel;
|
||||
tss->ss = vmx_read_segment_selector(cpu, R_SS).sel;
|
||||
tss->ds = vmx_read_segment_selector(cpu, R_DS).sel;
|
||||
tss->fs = vmx_read_segment_selector(cpu, R_FS).sel;
|
||||
tss->gs = vmx_read_segment_selector(cpu, R_GS).sel;
|
||||
}
|
||||
|
||||
static void load_state_from_tss32(CPUState *cpu, struct x86_tss_segment32 *tss)
|
||||
@ -82,22 +82,22 @@ static void load_state_from_tss32(CPUState *cpu, struct x86_tss_segment32 *tss)
|
||||
RSI(env) = tss->esi;
|
||||
RDI(env) = tss->edi;
|
||||
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->ldt}}, REG_SEG_LDTR);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->es}}, REG_SEG_ES);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->cs}}, REG_SEG_CS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->ss}}, REG_SEG_SS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->ds}}, REG_SEG_DS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->fs}}, REG_SEG_FS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->gs}}, REG_SEG_GS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->ldt}}, R_LDTR);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->es}}, R_ES);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->cs}}, R_CS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->ss}}, R_SS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->ds}}, R_DS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->fs}}, R_FS);
|
||||
vmx_write_segment_selector(cpu, (x68_segment_selector){{tss->gs}}, R_GS);
|
||||
|
||||
#if 0
|
||||
load_segment(cpu, REG_SEG_LDTR, tss->ldt);
|
||||
load_segment(cpu, REG_SEG_ES, tss->es);
|
||||
load_segment(cpu, REG_SEG_CS, tss->cs);
|
||||
load_segment(cpu, REG_SEG_SS, tss->ss);
|
||||
load_segment(cpu, REG_SEG_DS, tss->ds);
|
||||
load_segment(cpu, REG_SEG_FS, tss->fs);
|
||||
load_segment(cpu, REG_SEG_GS, tss->gs);
|
||||
load_segment(cpu, R_LDTR, tss->ldt);
|
||||
load_segment(cpu, R_ES, tss->es);
|
||||
load_segment(cpu, R_CS, tss->cs);
|
||||
load_segment(cpu, R_SS, tss->ss);
|
||||
load_segment(cpu, R_DS, tss->ds);
|
||||
load_segment(cpu, R_FS, tss->fs);
|
||||
load_segment(cpu, R_GS, tss->gs);
|
||||
#endif
|
||||
}
|
||||
|
||||
@ -139,8 +139,8 @@ void vmx_handle_task_switch(CPUState *cpu, x68_segment_selector tss_sel, int rea
|
||||
|
||||
struct x86_segment_descriptor curr_tss_desc, next_tss_desc;
|
||||
int ret;
|
||||
x68_segment_selector old_tss_sel = vmx_read_segment_selector(cpu, REG_SEG_TR);
|
||||
uint64_t old_tss_base = vmx_read_segment_base(cpu, REG_SEG_TR);
|
||||
x68_segment_selector old_tss_sel = vmx_read_segment_selector(cpu, R_TR);
|
||||
uint64_t old_tss_base = vmx_read_segment_base(cpu, R_TR);
|
||||
uint32_t desc_limit;
|
||||
struct x86_call_gate task_gate_desc;
|
||||
struct vmx_segment vmx_seg;
|
||||
@ -157,7 +157,7 @@ void vmx_handle_task_switch(CPUState *cpu, x68_segment_selector tss_sel, int rea
|
||||
ret = x86_read_call_gate(cpu, &task_gate_desc, gate);
|
||||
|
||||
dpl = task_gate_desc.dpl;
|
||||
x68_segment_selector cs = vmx_read_segment_selector(cpu, REG_SEG_CS);
|
||||
x68_segment_selector cs = vmx_read_segment_selector(cpu, R_CS);
|
||||
if (tss_sel.rpl > dpl || cs.rpl > dpl)
|
||||
;//DPRINTF("emulate_gp");
|
||||
}
|
||||
@ -191,7 +191,7 @@ void vmx_handle_task_switch(CPUState *cpu, x68_segment_selector tss_sel, int rea
|
||||
|
||||
macvm_set_cr0(cpu->hvf_fd, rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0) | CR0_TS);
|
||||
x86_segment_descriptor_to_vmx(cpu, tss_sel, &next_tss_desc, &vmx_seg);
|
||||
vmx_write_segment_descriptor(cpu, &vmx_seg, REG_SEG_TR);
|
||||
vmx_write_segment_descriptor(cpu, &vmx_seg, R_TR);
|
||||
|
||||
store_regs(cpu);
|
||||
|
||||
|
@ -107,28 +107,28 @@ void hvf_put_segments(CPUState *cpu_state)
|
||||
macvm_set_cr0(cpu_state->hvf_fd, env->cr[0]);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->segs[R_CS], false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_CS);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_CS);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->segs[R_DS], false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_DS);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_DS);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->segs[R_ES], false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_ES);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_ES);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->segs[R_SS], false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_SS);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_SS);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->segs[R_FS], false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_FS);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_FS);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->segs[R_GS], false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_GS);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_GS);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->tr, true);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_TR);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_TR);
|
||||
|
||||
hvf_set_segment(cpu_state, &seg, &env->ldt, false);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, REG_SEG_LDTR);
|
||||
vmx_write_segment_descriptor(cpu_state, &seg, R_LDTR);
|
||||
|
||||
hv_vcpu_flush(cpu_state->hvf_fd);
|
||||
}
|
||||
@ -183,28 +183,28 @@ void hvf_get_segments(CPUState *cpu_state)
|
||||
|
||||
env->interrupt_injected = -1;
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_CS);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_CS);
|
||||
hvf_get_segment(&env->segs[R_CS], &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_DS);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_DS);
|
||||
hvf_get_segment(&env->segs[R_DS], &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_ES);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_ES);
|
||||
hvf_get_segment(&env->segs[R_ES], &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_FS);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_FS);
|
||||
hvf_get_segment(&env->segs[R_FS], &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_GS);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_GS);
|
||||
hvf_get_segment(&env->segs[R_GS], &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_SS);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_SS);
|
||||
hvf_get_segment(&env->segs[R_SS], &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_TR);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_TR);
|
||||
hvf_get_segment(&env->tr, &seg);
|
||||
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, REG_SEG_LDTR);
|
||||
vmx_read_segment_descriptor(cpu_state, &seg, R_LDTR);
|
||||
hvf_get_segment(&env->ldt, &seg);
|
||||
|
||||
env->idt.limit = rvmcs(cpu_state->hvf_fd, VMCS_GUEST_IDTR_LIMIT);
|
||||
|
Loading…
Reference in New Issue
Block a user