mirror of
https://github.com/RPCS3/llvm.git
synced 2024-12-26 22:26:16 +00:00
continue moving stuff out to X86InstrSystem.td. Move
control flow stuff out to X86InstrControl.td. Move some compiler pseudo instructions and Pat<> patterns out to X86InstrCompiler.td git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@115596 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
a0b991be3f
commit
87be16a9e1
@ -117,117 +117,6 @@ def ADJCALLSTACKUP64 : I<0, Pseudo, (outs), (ins i32imm:$amt1, i32imm:$amt2),
|
||||
}
|
||||
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Call Instructions...
|
||||
//
|
||||
let isCall = 1 in
|
||||
// All calls clobber the non-callee saved registers. RSP is marked as
|
||||
// a use to prevent stack-pointer assignments that appear immediately
|
||||
// before calls from potentially appearing dead. Uses for argument
|
||||
// registers are added manually.
|
||||
let Defs = [RAX, RCX, RDX, RSI, RDI, R8, R9, R10, R11,
|
||||
FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0, ST1,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [RSP] in {
|
||||
|
||||
// NOTE: this pattern doesn't match "X86call imm", because we do not know
|
||||
// that the offset between an arbitrary immediate and the call will fit in
|
||||
// the 32-bit pcrel field that we have.
|
||||
def CALL64pcrel32 : Ii32PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i64i32imm_pcrel:$dst, variable_ops),
|
||||
"call{q}\t$dst", []>,
|
||||
Requires<[In64BitMode, NotWin64]>;
|
||||
def CALL64r : I<0xFF, MRM2r, (outs), (ins GR64:$dst, variable_ops),
|
||||
"call{q}\t{*}$dst", [(X86call GR64:$dst)]>,
|
||||
Requires<[In64BitMode, NotWin64]>;
|
||||
def CALL64m : I<0xFF, MRM2m, (outs), (ins i64mem:$dst, variable_ops),
|
||||
"call{q}\t{*}$dst", [(X86call (loadi64 addr:$dst))]>,
|
||||
Requires<[In64BitMode, NotWin64]>;
|
||||
|
||||
def FARCALL64 : RI<0xFF, MRM3m, (outs), (ins opaque80mem:$dst),
|
||||
"lcall{q}\t{*}$dst", []>;
|
||||
}
|
||||
|
||||
// FIXME: We need to teach codegen about single list of call-clobbered
|
||||
// registers.
|
||||
let isCall = 1, isCodeGenOnly = 1 in
|
||||
// All calls clobber the non-callee saved registers. RSP is marked as
|
||||
// a use to prevent stack-pointer assignments that appear immediately
|
||||
// before calls from potentially appearing dead. Uses for argument
|
||||
// registers are added manually.
|
||||
let Defs = [RAX, RCX, RDX, R8, R9, R10, R11,
|
||||
FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0, ST1,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, EFLAGS],
|
||||
Uses = [RSP] in {
|
||||
def WINCALL64pcrel32 : Ii32PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i64i32imm_pcrel:$dst, variable_ops),
|
||||
"call{q}\t$dst", []>,
|
||||
Requires<[IsWin64]>;
|
||||
def WINCALL64r : I<0xFF, MRM2r, (outs), (ins GR64:$dst, variable_ops),
|
||||
"call{q}\t{*}$dst",
|
||||
[(X86call GR64:$dst)]>, Requires<[IsWin64]>;
|
||||
def WINCALL64m : I<0xFF, MRM2m, (outs), (ins i64mem:$dst,variable_ops),
|
||||
"call{q}\t{*}$dst",
|
||||
[(X86call (loadi64 addr:$dst))]>,
|
||||
Requires<[IsWin64]>;
|
||||
}
|
||||
|
||||
|
||||
let isCall = 1, isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
isCodeGenOnly = 1 in
|
||||
let Defs = [RAX, RCX, RDX, RSI, RDI, R8, R9, R10, R11,
|
||||
FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0, ST1,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [RSP] in {
|
||||
def TCRETURNdi64 : I<0, Pseudo, (outs),
|
||||
(ins i64i32imm_pcrel:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
def TCRETURNri64 : I<0, Pseudo, (outs), (ins GR64_TC:$dst, i32imm:$offset,
|
||||
variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
let mayLoad = 1 in
|
||||
def TCRETURNmi64 : I<0, Pseudo, (outs),
|
||||
(ins i64mem_TC:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
|
||||
def TAILJMPd64 : Ii32PCRel<0xE9, RawFrm, (outs),
|
||||
(ins i64i32imm_pcrel:$dst, variable_ops),
|
||||
"jmp\t$dst # TAILCALL", []>;
|
||||
def TAILJMPr64 : I<0xFF, MRM4r, (outs), (ins GR64_TC:$dst, variable_ops),
|
||||
"jmp{q}\t{*}$dst # TAILCALL", []>;
|
||||
|
||||
let mayLoad = 1 in
|
||||
def TAILJMPm64 : I<0xFF, MRM4m, (outs), (ins i64mem_TC:$dst, variable_ops),
|
||||
"jmp{q}\t{*}$dst # TAILCALL", []>;
|
||||
}
|
||||
|
||||
// Branches
|
||||
let isBranch = 1, isTerminator = 1, isBarrier = 1, isIndirectBranch = 1 in {
|
||||
def JMP64pcrel32 : I<0xE9, RawFrm, (outs), (ins brtarget:$dst),
|
||||
"jmp{q}\t$dst", []>;
|
||||
def JMP64r : I<0xFF, MRM4r, (outs), (ins GR64:$dst), "jmp{q}\t{*}$dst",
|
||||
[(brind GR64:$dst)]>, Requires<[In64BitMode]>;
|
||||
def JMP64m : I<0xFF, MRM4m, (outs), (ins i64mem:$dst), "jmp{q}\t{*}$dst",
|
||||
[(brind (loadi64 addr:$dst))]>, Requires<[In64BitMode]>;
|
||||
def FARJMP64 : RI<0xFF, MRM5m, (outs), (ins opaque80mem:$dst),
|
||||
"ljmp{q}\t{*}$dst", []>;
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// EH Pseudo Instructions
|
||||
//
|
||||
let isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
hasCtrlDep = 1, isCodeGenOnly = 1 in {
|
||||
def EH_RETURN64 : I<0xC3, RawFrm, (outs), (ins GR64:$addr),
|
||||
"ret\t#eh_return, addr: $addr",
|
||||
[(X86ehret GR64:$addr)]>;
|
||||
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Miscellaneous Instructions...
|
||||
@ -398,27 +287,6 @@ def MOV64ao64 : RIi32<0xA3, RawFrm, (outs offset64:$dst), (ins),
|
||||
"mov{q}\t{%rax, $dst|$dst, %rax}", []>;
|
||||
*/
|
||||
|
||||
// Moves to and from segment registers
|
||||
def MOV64rs : RI<0x8C, MRMDestReg, (outs GR64:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64ms : RI<0x8C, MRMDestMem, (outs i64mem:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64sr : RI<0x8E, MRMSrcReg, (outs SEGMENT_REG:$dst), (ins GR64:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64sm : RI<0x8E, MRMSrcMem, (outs SEGMENT_REG:$dst), (ins i64mem:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
// Moves to and from debug registers
|
||||
def MOV64rd : I<0x21, MRMDestReg, (outs GR64:$dst), (ins DEBUG_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV64dr : I<0x23, MRMSrcReg, (outs DEBUG_REG:$dst), (ins GR64:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
// Moves to and from control registers
|
||||
def MOV64rc : I<0x20, MRMDestReg, (outs GR64:$dst), (ins CONTROL_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV64cr : I<0x22, MRMSrcReg, (outs CONTROL_REG:$dst), (ins GR64:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
// Sign/Zero extenders
|
||||
|
||||
@ -1717,470 +1585,11 @@ def ATOMUMAX64: I<0, Pseudo, (outs GR64:$dst),(ins i64mem:$ptr, GR64:$val),
|
||||
[(set GR64:$dst, (atomic_load_umax_64 addr:$ptr, GR64:$val))]>;
|
||||
}
|
||||
|
||||
// Segmentation support instructions
|
||||
|
||||
// i16mem operand in LAR64rm and GR32 operand in LAR32rr is not a typo.
|
||||
def LAR64rm : RI<0x02, MRMSrcMem, (outs GR64:$dst), (ins i16mem:$src),
|
||||
"lar{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LAR64rr : RI<0x02, MRMSrcReg, (outs GR64:$dst), (ins GR32:$src),
|
||||
"lar{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def LSL64rm : RI<0x03, MRMSrcMem, (outs GR64:$dst), (ins i64mem:$src),
|
||||
"lsl{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LSL64rr : RI<0x03, MRMSrcReg, (outs GR64:$dst), (ins GR64:$src),
|
||||
"lsl{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def SWAPGS : I<0x01, MRM_F8, (outs), (ins), "swapgs", []>, TB;
|
||||
|
||||
|
||||
// String manipulation instructions
|
||||
|
||||
def LODSQ : RI<0xAD, RawFrm, (outs), (ins), "lodsq", []>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Non-Instruction Patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// ConstantPool GlobalAddress, ExternalSymbol, and JumpTable when not in small
|
||||
// code model mode, should use 'movabs'. FIXME: This is really a hack, the
|
||||
// 'movabs' predicate should handle this sort of thing.
|
||||
def : Pat<(i64 (X86Wrapper tconstpool :$dst)),
|
||||
(MOV64ri tconstpool :$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper tjumptable :$dst)),
|
||||
(MOV64ri tjumptable :$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper tglobaladdr :$dst)),
|
||||
(MOV64ri tglobaladdr :$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper texternalsym:$dst)),
|
||||
(MOV64ri texternalsym:$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper tblockaddress:$dst)),
|
||||
(MOV64ri tblockaddress:$dst)>, Requires<[FarData]>;
|
||||
|
||||
// In static codegen with small code model, we can get the address of a label
|
||||
// into a register with 'movl'. FIXME: This is a hack, the 'imm' predicate of
|
||||
// the MOV64ri64i32 should accept these.
|
||||
def : Pat<(i64 (X86Wrapper tconstpool :$dst)),
|
||||
(MOV64ri64i32 tconstpool :$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tjumptable :$dst)),
|
||||
(MOV64ri64i32 tjumptable :$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tglobaladdr :$dst)),
|
||||
(MOV64ri64i32 tglobaladdr :$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper texternalsym:$dst)),
|
||||
(MOV64ri64i32 texternalsym:$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tblockaddress:$dst)),
|
||||
(MOV64ri64i32 tblockaddress:$dst)>, Requires<[SmallCode]>;
|
||||
|
||||
// In kernel code model, we can get the address of a label
|
||||
// into a register with 'movq'. FIXME: This is a hack, the 'imm' predicate of
|
||||
// the MOV64ri32 should accept these.
|
||||
def : Pat<(i64 (X86Wrapper tconstpool :$dst)),
|
||||
(MOV64ri32 tconstpool :$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tjumptable :$dst)),
|
||||
(MOV64ri32 tjumptable :$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tglobaladdr :$dst)),
|
||||
(MOV64ri32 tglobaladdr :$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper texternalsym:$dst)),
|
||||
(MOV64ri32 texternalsym:$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tblockaddress:$dst)),
|
||||
(MOV64ri32 tblockaddress:$dst)>, Requires<[KernelCode]>;
|
||||
|
||||
// If we have small model and -static mode, it is safe to store global addresses
|
||||
// directly as immediates. FIXME: This is really a hack, the 'imm' predicate
|
||||
// for MOV64mi32 should handle this sort of thing.
|
||||
def : Pat<(store (i64 (X86Wrapper tconstpool:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tconstpool:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper tjumptable:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tjumptable:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper tglobaladdr:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tglobaladdr:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper texternalsym:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, texternalsym:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper tblockaddress:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tblockaddress:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
|
||||
// Calls
|
||||
// Direct PC relative function call for small code model. 32-bit displacement
|
||||
// sign extended to 64-bit.
|
||||
def : Pat<(X86call (i64 tglobaladdr:$dst)),
|
||||
(CALL64pcrel32 tglobaladdr:$dst)>, Requires<[NotWin64]>;
|
||||
def : Pat<(X86call (i64 texternalsym:$dst)),
|
||||
(CALL64pcrel32 texternalsym:$dst)>, Requires<[NotWin64]>;
|
||||
|
||||
def : Pat<(X86call (i64 tglobaladdr:$dst)),
|
||||
(WINCALL64pcrel32 tglobaladdr:$dst)>, Requires<[IsWin64]>;
|
||||
def : Pat<(X86call (i64 texternalsym:$dst)),
|
||||
(WINCALL64pcrel32 texternalsym:$dst)>, Requires<[IsWin64]>;
|
||||
|
||||
// tailcall stuff
|
||||
def : Pat<(X86tcret GR64_TC:$dst, imm:$off),
|
||||
(TCRETURNri64 GR64_TC:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (load addr:$dst), imm:$off),
|
||||
(TCRETURNmi64 addr:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (i64 tglobaladdr:$dst), imm:$off),
|
||||
(TCRETURNdi64 tglobaladdr:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (i64 texternalsym:$dst), imm:$off),
|
||||
(TCRETURNdi64 texternalsym:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// tls has some funny stuff here...
|
||||
// This corresponds to movabs $foo@tpoff, %rax
|
||||
def : Pat<(i64 (X86Wrapper tglobaltlsaddr :$dst)),
|
||||
(MOV64ri tglobaltlsaddr :$dst)>;
|
||||
// This corresponds to add $foo@tpoff, %rax
|
||||
def : Pat<(add GR64:$src1, (X86Wrapper tglobaltlsaddr :$dst)),
|
||||
(ADD64ri32 GR64:$src1, tglobaltlsaddr :$dst)>;
|
||||
// This corresponds to mov foo@tpoff(%rbx), %eax
|
||||
def : Pat<(load (i64 (X86Wrapper tglobaltlsaddr :$dst))),
|
||||
(MOV64rm tglobaltlsaddr :$dst)>;
|
||||
|
||||
// Comparisons.
|
||||
|
||||
// TEST R,R is smaller than CMP R,0
|
||||
def : Pat<(X86cmp GR64:$src1, 0),
|
||||
(TEST64rr GR64:$src1, GR64:$src1)>;
|
||||
|
||||
// Conditional moves with folded loads with operands swapped and conditions
|
||||
// inverted.
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_B, EFLAGS),
|
||||
(CMOVAE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_AE, EFLAGS),
|
||||
(CMOVB64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_E, EFLAGS),
|
||||
(CMOVNE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NE, EFLAGS),
|
||||
(CMOVE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_BE, EFLAGS),
|
||||
(CMOVA64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_A, EFLAGS),
|
||||
(CMOVBE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_L, EFLAGS),
|
||||
(CMOVGE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_GE, EFLAGS),
|
||||
(CMOVL64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_LE, EFLAGS),
|
||||
(CMOVG64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_G, EFLAGS),
|
||||
(CMOVLE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_P, EFLAGS),
|
||||
(CMOVNP64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NP, EFLAGS),
|
||||
(CMOVP64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_S, EFLAGS),
|
||||
(CMOVNS64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NS, EFLAGS),
|
||||
(CMOVS64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_O, EFLAGS),
|
||||
(CMOVNO64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NO, EFLAGS),
|
||||
(CMOVO64rm GR64:$src2, addr:$src1)>;
|
||||
|
||||
// zextload bool -> zextload byte
|
||||
def : Pat<(zextloadi64i1 addr:$src), (MOVZX64rm8 addr:$src)>;
|
||||
|
||||
// extload
|
||||
// When extloading from 16-bit and smaller memory locations into 64-bit
|
||||
// registers, use zero-extending loads so that the entire 64-bit register is
|
||||
// defined, avoiding partial-register updates.
|
||||
def : Pat<(extloadi64i1 addr:$src), (MOVZX64rm8 addr:$src)>;
|
||||
def : Pat<(extloadi64i8 addr:$src), (MOVZX64rm8 addr:$src)>;
|
||||
def : Pat<(extloadi64i16 addr:$src), (MOVZX64rm16 addr:$src)>;
|
||||
// For other extloads, use subregs, since the high contents of the register are
|
||||
// defined after an extload.
|
||||
def : Pat<(extloadi64i32 addr:$src),
|
||||
(SUBREG_TO_REG (i64 0), (MOV32rm addr:$src),
|
||||
sub_32bit)>;
|
||||
|
||||
// anyext. Define these to do an explicit zero-extend to
|
||||
// avoid partial-register updates.
|
||||
def : Pat<(i64 (anyext GR8 :$src)), (MOVZX64rr8 GR8 :$src)>;
|
||||
def : Pat<(i64 (anyext GR16:$src)), (MOVZX64rr16 GR16 :$src)>;
|
||||
def : Pat<(i64 (anyext GR32:$src)),
|
||||
(SUBREG_TO_REG (i64 0), GR32:$src, sub_32bit)>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Some peepholes
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// Odd encoding trick: -128 fits into an 8-bit immediate field while
|
||||
// +128 doesn't, so in this special case use a sub instead of an add.
|
||||
def : Pat<(add GR64:$src1, 128),
|
||||
(SUB64ri8 GR64:$src1, -128)>;
|
||||
def : Pat<(store (add (loadi64 addr:$dst), 128), addr:$dst),
|
||||
(SUB64mi8 addr:$dst, -128)>;
|
||||
|
||||
// The same trick applies for 32-bit immediate fields in 64-bit
|
||||
// instructions.
|
||||
def : Pat<(add GR64:$src1, 0x0000000080000000),
|
||||
(SUB64ri32 GR64:$src1, 0xffffffff80000000)>;
|
||||
def : Pat<(store (add (loadi64 addr:$dst), 0x00000000800000000), addr:$dst),
|
||||
(SUB64mi32 addr:$dst, 0xffffffff80000000)>;
|
||||
|
||||
// Use a 32-bit and with implicit zero-extension instead of a 64-bit and if it
|
||||
// has an immediate with at least 32 bits of leading zeros, to avoid needing to
|
||||
// materialize that immediate in a register first.
|
||||
def : Pat<(and GR64:$src, i64immZExt32:$imm),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(AND32ri
|
||||
(EXTRACT_SUBREG GR64:$src, sub_32bit),
|
||||
(i32 (GetLo32XForm imm:$imm))),
|
||||
sub_32bit)>;
|
||||
|
||||
// r & (2^32-1) ==> movz
|
||||
def : Pat<(and GR64:$src, 0x00000000FFFFFFFF),
|
||||
(MOVZX64rr32 (EXTRACT_SUBREG GR64:$src, sub_32bit))>;
|
||||
// r & (2^16-1) ==> movz
|
||||
def : Pat<(and GR64:$src, 0xffff),
|
||||
(MOVZX64rr16 (i16 (EXTRACT_SUBREG GR64:$src, sub_16bit)))>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR64:$src, 0xff),
|
||||
(MOVZX64rr8 (i8 (EXTRACT_SUBREG GR64:$src, sub_8bit)))>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR32:$src1, 0xff),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG GR32:$src1, sub_8bit))>,
|
||||
Requires<[In64BitMode]>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR16:$src1, 0xff),
|
||||
(MOVZX16rr8 (i8 (EXTRACT_SUBREG GR16:$src1, sub_8bit)))>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// sext_inreg patterns
|
||||
def : Pat<(sext_inreg GR64:$src, i32),
|
||||
(MOVSX64rr32 (EXTRACT_SUBREG GR64:$src, sub_32bit))>;
|
||||
def : Pat<(sext_inreg GR64:$src, i16),
|
||||
(MOVSX64rr16 (EXTRACT_SUBREG GR64:$src, sub_16bit))>;
|
||||
def : Pat<(sext_inreg GR64:$src, i8),
|
||||
(MOVSX64rr8 (EXTRACT_SUBREG GR64:$src, sub_8bit))>;
|
||||
def : Pat<(sext_inreg GR32:$src, i8),
|
||||
(MOVSX32rr8 (EXTRACT_SUBREG GR32:$src, sub_8bit))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(sext_inreg GR16:$src, i8),
|
||||
(MOVSX16rr8 (i8 (EXTRACT_SUBREG GR16:$src, sub_8bit)))>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// trunc patterns
|
||||
def : Pat<(i32 (trunc GR64:$src)),
|
||||
(EXTRACT_SUBREG GR64:$src, sub_32bit)>;
|
||||
def : Pat<(i16 (trunc GR64:$src)),
|
||||
(EXTRACT_SUBREG GR64:$src, sub_16bit)>;
|
||||
def : Pat<(i8 (trunc GR64:$src)),
|
||||
(EXTRACT_SUBREG GR64:$src, sub_8bit)>;
|
||||
def : Pat<(i8 (trunc GR32:$src)),
|
||||
(EXTRACT_SUBREG GR32:$src, sub_8bit)>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i8 (trunc GR16:$src)),
|
||||
(EXTRACT_SUBREG GR16:$src, sub_8bit)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// h-register tricks.
|
||||
// For now, be conservative on x86-64 and use an h-register extract only if the
|
||||
// value is immediately zero-extended or stored, which are somewhat common
|
||||
// cases. This uses a bunch of code to prevent a register requiring a REX prefix
|
||||
// from being allocated in the same instruction as the h register, as there's
|
||||
// currently no way to describe this requirement to the register allocator.
|
||||
|
||||
// h-register extract and zero-extend.
|
||||
def : Pat<(and (srl_su GR64:$src, (i8 8)), (i64 255)),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i64 (COPY_TO_REGCLASS GR64:$src, GR64_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_32bit)>;
|
||||
def : Pat<(and (srl_su GR32:$src, (i8 8)), (i32 255)),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(srl (and_su GR32:$src, 0xff00), (i8 8)),
|
||||
(MOVZX32_NOREXrr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(srl GR16:$src, (i8 8)),
|
||||
(EXTRACT_SUBREG
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_16bit)>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i32 (zext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i32 (anyext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i64 (zext (srl_su GR16:$src, (i8 8)))),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_32bit)>;
|
||||
def : Pat<(i64 (anyext (srl_su GR16:$src, (i8 8)))),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_32bit)>;
|
||||
|
||||
// h-register extract and store.
|
||||
def : Pat<(store (i8 (trunc_su (srl_su GR64:$src, (i8 8)))), addr:$dst),
|
||||
(MOV8mr_NOREX
|
||||
addr:$dst,
|
||||
(EXTRACT_SUBREG (i64 (COPY_TO_REGCLASS GR64:$src, GR64_ABCD)),
|
||||
sub_8bit_hi))>;
|
||||
def : Pat<(store (i8 (trunc_su (srl_su GR32:$src, (i8 8)))), addr:$dst),
|
||||
(MOV8mr_NOREX
|
||||
addr:$dst,
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(store (i8 (trunc_su (srl_su GR16:$src, (i8 8)))), addr:$dst),
|
||||
(MOV8mr_NOREX
|
||||
addr:$dst,
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// (shl x, 1) ==> (add x, x)
|
||||
def : Pat<(shl GR64:$src1, (i8 1)), (ADD64rr GR64:$src1, GR64:$src1)>;
|
||||
|
||||
// (shl x (and y, 63)) ==> (shl x, y)
|
||||
def : Pat<(shl GR64:$src1, (and CL, 63)),
|
||||
(SHL64rCL GR64:$src1)>;
|
||||
def : Pat<(store (shl (loadi64 addr:$dst), (and CL, 63)), addr:$dst),
|
||||
(SHL64mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(srl GR64:$src1, (and CL, 63)),
|
||||
(SHR64rCL GR64:$src1)>;
|
||||
def : Pat<(store (srl (loadi64 addr:$dst), (and CL, 63)), addr:$dst),
|
||||
(SHR64mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(sra GR64:$src1, (and CL, 63)),
|
||||
(SAR64rCL GR64:$src1)>;
|
||||
def : Pat<(store (sra (loadi64 addr:$dst), (and CL, 63)), addr:$dst),
|
||||
(SAR64mCL addr:$dst)>;
|
||||
|
||||
// (or x1, x2) -> (add x1, x2) if two operands are known not to share bits.
|
||||
let AddedComplexity = 5 in { // Try this before the selecting to OR
|
||||
def : Pat<(or_is_add GR64:$src1, i64immSExt8:$src2),
|
||||
(ADD64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(or_is_add GR64:$src1, i64immSExt32:$src2),
|
||||
(ADD64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(or_is_add GR64:$src1, GR64:$src2),
|
||||
(ADD64rr GR64:$src1, GR64:$src2)>;
|
||||
} // AddedComplexity
|
||||
|
||||
// X86 specific add which produces a flag.
|
||||
def : Pat<(addc GR64:$src1, GR64:$src2),
|
||||
(ADD64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(addc GR64:$src1, (load addr:$src2)),
|
||||
(ADD64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(addc GR64:$src1, i64immSExt8:$src2),
|
||||
(ADD64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(addc GR64:$src1, i64immSExt32:$src2),
|
||||
(ADD64ri32 GR64:$src1, imm:$src2)>;
|
||||
|
||||
def : Pat<(subc GR64:$src1, GR64:$src2),
|
||||
(SUB64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(subc GR64:$src1, (load addr:$src2)),
|
||||
(SUB64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(subc GR64:$src1, i64immSExt8:$src2),
|
||||
(SUB64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(subc GR64:$src1, imm:$src2),
|
||||
(SUB64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// EFLAGS-defining Patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// addition
|
||||
def : Pat<(add GR64:$src1, GR64:$src2),
|
||||
(ADD64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(add GR64:$src1, i64immSExt8:$src2),
|
||||
(ADD64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(add GR64:$src1, i64immSExt32:$src2),
|
||||
(ADD64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(add GR64:$src1, (loadi64 addr:$src2)),
|
||||
(ADD64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// subtraction
|
||||
def : Pat<(sub GR64:$src1, GR64:$src2),
|
||||
(SUB64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(sub GR64:$src1, (loadi64 addr:$src2)),
|
||||
(SUB64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(sub GR64:$src1, i64immSExt8:$src2),
|
||||
(SUB64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(sub GR64:$src1, i64immSExt32:$src2),
|
||||
(SUB64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// Multiply
|
||||
def : Pat<(mul GR64:$src1, GR64:$src2),
|
||||
(IMUL64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(mul GR64:$src1, (loadi64 addr:$src2)),
|
||||
(IMUL64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(mul GR64:$src1, i64immSExt8:$src2),
|
||||
(IMUL64rri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(mul GR64:$src1, i64immSExt32:$src2),
|
||||
(IMUL64rri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(mul (loadi64 addr:$src1), i64immSExt8:$src2),
|
||||
(IMUL64rmi8 addr:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(mul (loadi64 addr:$src1), i64immSExt32:$src2),
|
||||
(IMUL64rmi32 addr:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// inc/dec
|
||||
def : Pat<(add GR16:$src, 1), (INC64_16r GR16:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR16:$src, -1), (DEC64_16r GR16:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR32:$src, 1), (INC64_32r GR32:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR32:$src, -1), (DEC64_32r GR32:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR64:$src, 1), (INC64r GR64:$src)>;
|
||||
def : Pat<(add GR64:$src, -1), (DEC64r GR64:$src)>;
|
||||
|
||||
// or
|
||||
def : Pat<(or GR64:$src1, GR64:$src2),
|
||||
(OR64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(or GR64:$src1, i64immSExt8:$src2),
|
||||
(OR64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(or GR64:$src1, i64immSExt32:$src2),
|
||||
(OR64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(or GR64:$src1, (loadi64 addr:$src2)),
|
||||
(OR64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// xor
|
||||
def : Pat<(xor GR64:$src1, GR64:$src2),
|
||||
(XOR64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, i64immSExt8:$src2),
|
||||
(XOR64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, i64immSExt32:$src2),
|
||||
(XOR64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, (loadi64 addr:$src2)),
|
||||
(XOR64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// and
|
||||
def : Pat<(and GR64:$src1, GR64:$src2),
|
||||
(AND64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(and GR64:$src1, i64immSExt8:$src2),
|
||||
(AND64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(and GR64:$src1, i64immSExt32:$src2),
|
||||
(AND64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(and GR64:$src1, (loadi64 addr:$src2)),
|
||||
(AND64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// X86-64 SSE Instructions
|
||||
|
961
lib/Target/X86/X86InstrCompiler.td
Normal file
961
lib/Target/X86/X86InstrCompiler.td
Normal file
@ -0,0 +1,961 @@
|
||||
//===- X86InstrCompiler.td - Compiler Pseudos and Patterns -*- tablegen -*-===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// This file describes the various pseudo instructions used by the compiler,
|
||||
// as well as Pat patterns used during instruction selection.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// EH Pseudo Instructions
|
||||
//
|
||||
let isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
hasCtrlDep = 1, isCodeGenOnly = 1 in {
|
||||
def EH_RETURN : I<0xC3, RawFrm, (outs), (ins GR32:$addr),
|
||||
"ret\t#eh_return, addr: $addr",
|
||||
[(X86ehret GR32:$addr)]>;
|
||||
|
||||
}
|
||||
|
||||
let isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
hasCtrlDep = 1, isCodeGenOnly = 1 in {
|
||||
def EH_RETURN64 : I<0xC3, RawFrm, (outs), (ins GR64:$addr),
|
||||
"ret\t#eh_return, addr: $addr",
|
||||
[(X86ehret GR64:$addr)]>;
|
||||
|
||||
}
|
||||
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Non-Instruction Patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// ConstantPool GlobalAddress, ExternalSymbol, and JumpTable
|
||||
def : Pat<(i32 (X86Wrapper tconstpool :$dst)), (MOV32ri tconstpool :$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tjumptable :$dst)), (MOV32ri tjumptable :$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tglobaltlsaddr:$dst)),(MOV32ri tglobaltlsaddr:$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tglobaladdr :$dst)), (MOV32ri tglobaladdr :$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper texternalsym:$dst)), (MOV32ri texternalsym:$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tblockaddress:$dst)), (MOV32ri tblockaddress:$dst)>;
|
||||
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tconstpool:$src2)),
|
||||
(ADD32ri GR32:$src1, tconstpool:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tjumptable:$src2)),
|
||||
(ADD32ri GR32:$src1, tjumptable:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tglobaladdr :$src2)),
|
||||
(ADD32ri GR32:$src1, tglobaladdr:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper texternalsym:$src2)),
|
||||
(ADD32ri GR32:$src1, texternalsym:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tblockaddress:$src2)),
|
||||
(ADD32ri GR32:$src1, tblockaddress:$src2)>;
|
||||
|
||||
def : Pat<(store (i32 (X86Wrapper tglobaladdr:$src)), addr:$dst),
|
||||
(MOV32mi addr:$dst, tglobaladdr:$src)>;
|
||||
def : Pat<(store (i32 (X86Wrapper texternalsym:$src)), addr:$dst),
|
||||
(MOV32mi addr:$dst, texternalsym:$src)>;
|
||||
def : Pat<(store (i32 (X86Wrapper tblockaddress:$src)), addr:$dst),
|
||||
(MOV32mi addr:$dst, tblockaddress:$src)>;
|
||||
|
||||
|
||||
|
||||
// ConstantPool GlobalAddress, ExternalSymbol, and JumpTable when not in small
|
||||
// code model mode, should use 'movabs'. FIXME: This is really a hack, the
|
||||
// 'movabs' predicate should handle this sort of thing.
|
||||
def : Pat<(i64 (X86Wrapper tconstpool :$dst)),
|
||||
(MOV64ri tconstpool :$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper tjumptable :$dst)),
|
||||
(MOV64ri tjumptable :$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper tglobaladdr :$dst)),
|
||||
(MOV64ri tglobaladdr :$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper texternalsym:$dst)),
|
||||
(MOV64ri texternalsym:$dst)>, Requires<[FarData]>;
|
||||
def : Pat<(i64 (X86Wrapper tblockaddress:$dst)),
|
||||
(MOV64ri tblockaddress:$dst)>, Requires<[FarData]>;
|
||||
|
||||
// In static codegen with small code model, we can get the address of a label
|
||||
// into a register with 'movl'. FIXME: This is a hack, the 'imm' predicate of
|
||||
// the MOV64ri64i32 should accept these.
|
||||
def : Pat<(i64 (X86Wrapper tconstpool :$dst)),
|
||||
(MOV64ri64i32 tconstpool :$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tjumptable :$dst)),
|
||||
(MOV64ri64i32 tjumptable :$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tglobaladdr :$dst)),
|
||||
(MOV64ri64i32 tglobaladdr :$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper texternalsym:$dst)),
|
||||
(MOV64ri64i32 texternalsym:$dst)>, Requires<[SmallCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tblockaddress:$dst)),
|
||||
(MOV64ri64i32 tblockaddress:$dst)>, Requires<[SmallCode]>;
|
||||
|
||||
// In kernel code model, we can get the address of a label
|
||||
// into a register with 'movq'. FIXME: This is a hack, the 'imm' predicate of
|
||||
// the MOV64ri32 should accept these.
|
||||
def : Pat<(i64 (X86Wrapper tconstpool :$dst)),
|
||||
(MOV64ri32 tconstpool :$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tjumptable :$dst)),
|
||||
(MOV64ri32 tjumptable :$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tglobaladdr :$dst)),
|
||||
(MOV64ri32 tglobaladdr :$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper texternalsym:$dst)),
|
||||
(MOV64ri32 texternalsym:$dst)>, Requires<[KernelCode]>;
|
||||
def : Pat<(i64 (X86Wrapper tblockaddress:$dst)),
|
||||
(MOV64ri32 tblockaddress:$dst)>, Requires<[KernelCode]>;
|
||||
|
||||
// If we have small model and -static mode, it is safe to store global addresses
|
||||
// directly as immediates. FIXME: This is really a hack, the 'imm' predicate
|
||||
// for MOV64mi32 should handle this sort of thing.
|
||||
def : Pat<(store (i64 (X86Wrapper tconstpool:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tconstpool:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper tjumptable:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tjumptable:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper tglobaladdr:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tglobaladdr:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper texternalsym:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, texternalsym:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
def : Pat<(store (i64 (X86Wrapper tblockaddress:$src)), addr:$dst),
|
||||
(MOV64mi32 addr:$dst, tblockaddress:$src)>,
|
||||
Requires<[NearData, IsStatic]>;
|
||||
|
||||
|
||||
|
||||
// Calls
|
||||
|
||||
// tls has some funny stuff here...
|
||||
// This corresponds to movabs $foo@tpoff, %rax
|
||||
def : Pat<(i64 (X86Wrapper tglobaltlsaddr :$dst)),
|
||||
(MOV64ri tglobaltlsaddr :$dst)>;
|
||||
// This corresponds to add $foo@tpoff, %rax
|
||||
def : Pat<(add GR64:$src1, (X86Wrapper tglobaltlsaddr :$dst)),
|
||||
(ADD64ri32 GR64:$src1, tglobaltlsaddr :$dst)>;
|
||||
// This corresponds to mov foo@tpoff(%rbx), %eax
|
||||
def : Pat<(load (i64 (X86Wrapper tglobaltlsaddr :$dst))),
|
||||
(MOV64rm tglobaltlsaddr :$dst)>;
|
||||
|
||||
|
||||
// Direct PC relative function call for small code model. 32-bit displacement
|
||||
// sign extended to 64-bit.
|
||||
def : Pat<(X86call (i64 tglobaladdr:$dst)),
|
||||
(CALL64pcrel32 tglobaladdr:$dst)>, Requires<[NotWin64]>;
|
||||
def : Pat<(X86call (i64 texternalsym:$dst)),
|
||||
(CALL64pcrel32 texternalsym:$dst)>, Requires<[NotWin64]>;
|
||||
|
||||
def : Pat<(X86call (i64 tglobaladdr:$dst)),
|
||||
(WINCALL64pcrel32 tglobaladdr:$dst)>, Requires<[IsWin64]>;
|
||||
def : Pat<(X86call (i64 texternalsym:$dst)),
|
||||
(WINCALL64pcrel32 texternalsym:$dst)>, Requires<[IsWin64]>;
|
||||
|
||||
// tailcall stuff
|
||||
def : Pat<(X86tcret GR32_TC:$dst, imm:$off),
|
||||
(TCRETURNri GR32_TC:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// FIXME: This is disabled for 32-bit PIC mode because the global base
|
||||
// register which is part of the address mode may be assigned a
|
||||
// callee-saved register.
|
||||
def : Pat<(X86tcret (load addr:$dst), imm:$off),
|
||||
(TCRETURNmi addr:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode, IsNotPIC]>;
|
||||
|
||||
def : Pat<(X86tcret (i32 tglobaladdr:$dst), imm:$off),
|
||||
(TCRETURNdi texternalsym:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (i32 texternalsym:$dst), imm:$off),
|
||||
(TCRETURNdi texternalsym:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret GR64_TC:$dst, imm:$off),
|
||||
(TCRETURNri64 GR64_TC:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (load addr:$dst), imm:$off),
|
||||
(TCRETURNmi64 addr:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (i64 tglobaladdr:$dst), imm:$off),
|
||||
(TCRETURNdi64 tglobaladdr:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (i64 texternalsym:$dst), imm:$off),
|
||||
(TCRETURNdi64 texternalsym:$dst, imm:$off)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// Normal calls, with various flavors of addresses.
|
||||
def : Pat<(X86call (i32 tglobaladdr:$dst)),
|
||||
(CALLpcrel32 tglobaladdr:$dst)>;
|
||||
def : Pat<(X86call (i32 texternalsym:$dst)),
|
||||
(CALLpcrel32 texternalsym:$dst)>;
|
||||
def : Pat<(X86call (i32 imm:$dst)),
|
||||
(CALLpcrel32 imm:$dst)>, Requires<[CallImmAddr]>;
|
||||
|
||||
// X86 specific add which produces a flag.
|
||||
def : Pat<(addc GR32:$src1, GR32:$src2),
|
||||
(ADD32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(addc GR32:$src1, (load addr:$src2)),
|
||||
(ADD32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(addc GR32:$src1, imm:$src2),
|
||||
(ADD32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(addc GR32:$src1, i32immSExt8:$src2),
|
||||
(ADD32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
def : Pat<(addc GR64:$src1, GR64:$src2),
|
||||
(ADD64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(addc GR64:$src1, (load addr:$src2)),
|
||||
(ADD64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(addc GR64:$src1, i64immSExt8:$src2),
|
||||
(ADD64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(addc GR64:$src1, i64immSExt32:$src2),
|
||||
(ADD64ri32 GR64:$src1, imm:$src2)>;
|
||||
|
||||
def : Pat<(subc GR32:$src1, GR32:$src2),
|
||||
(SUB32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(subc GR32:$src1, (load addr:$src2)),
|
||||
(SUB32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(subc GR32:$src1, imm:$src2),
|
||||
(SUB32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(subc GR32:$src1, i32immSExt8:$src2),
|
||||
(SUB32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
def : Pat<(subc GR64:$src1, GR64:$src2),
|
||||
(SUB64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(subc GR64:$src1, (load addr:$src2)),
|
||||
(SUB64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(subc GR64:$src1, i64immSExt8:$src2),
|
||||
(SUB64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(subc GR64:$src1, imm:$src2),
|
||||
(SUB64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// Comparisons.
|
||||
|
||||
// TEST R,R is smaller than CMP R,0
|
||||
def : Pat<(X86cmp GR8:$src1, 0),
|
||||
(TEST8rr GR8:$src1, GR8:$src1)>;
|
||||
def : Pat<(X86cmp GR16:$src1, 0),
|
||||
(TEST16rr GR16:$src1, GR16:$src1)>;
|
||||
def : Pat<(X86cmp GR32:$src1, 0),
|
||||
(TEST32rr GR32:$src1, GR32:$src1)>;
|
||||
def : Pat<(X86cmp GR64:$src1, 0),
|
||||
(TEST64rr GR64:$src1, GR64:$src1)>;
|
||||
|
||||
// Conditional moves with folded loads with operands swapped and conditions
|
||||
// inverted.
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_B, EFLAGS),
|
||||
(CMOVAE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_B, EFLAGS),
|
||||
(CMOVAE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_AE, EFLAGS),
|
||||
(CMOVB16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_AE, EFLAGS),
|
||||
(CMOVB32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_E, EFLAGS),
|
||||
(CMOVNE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_E, EFLAGS),
|
||||
(CMOVNE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NE, EFLAGS),
|
||||
(CMOVE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NE, EFLAGS),
|
||||
(CMOVE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_BE, EFLAGS),
|
||||
(CMOVA16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_BE, EFLAGS),
|
||||
(CMOVA32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_A, EFLAGS),
|
||||
(CMOVBE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_A, EFLAGS),
|
||||
(CMOVBE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_L, EFLAGS),
|
||||
(CMOVGE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_L, EFLAGS),
|
||||
(CMOVGE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_GE, EFLAGS),
|
||||
(CMOVL16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_GE, EFLAGS),
|
||||
(CMOVL32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_LE, EFLAGS),
|
||||
(CMOVG16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_LE, EFLAGS),
|
||||
(CMOVG32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_G, EFLAGS),
|
||||
(CMOVLE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_G, EFLAGS),
|
||||
(CMOVLE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_P, EFLAGS),
|
||||
(CMOVNP16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_P, EFLAGS),
|
||||
(CMOVNP32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NP, EFLAGS),
|
||||
(CMOVP16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NP, EFLAGS),
|
||||
(CMOVP32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_S, EFLAGS),
|
||||
(CMOVNS16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_S, EFLAGS),
|
||||
(CMOVNS32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NS, EFLAGS),
|
||||
(CMOVS16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NS, EFLAGS),
|
||||
(CMOVS32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_O, EFLAGS),
|
||||
(CMOVNO16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_O, EFLAGS),
|
||||
(CMOVNO32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NO, EFLAGS),
|
||||
(CMOVO16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NO, EFLAGS),
|
||||
(CMOVO32rm GR32:$src2, addr:$src1)>;
|
||||
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_B, EFLAGS),
|
||||
(CMOVAE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_AE, EFLAGS),
|
||||
(CMOVB64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_E, EFLAGS),
|
||||
(CMOVNE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NE, EFLAGS),
|
||||
(CMOVE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_BE, EFLAGS),
|
||||
(CMOVA64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_A, EFLAGS),
|
||||
(CMOVBE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_L, EFLAGS),
|
||||
(CMOVGE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_GE, EFLAGS),
|
||||
(CMOVL64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_LE, EFLAGS),
|
||||
(CMOVG64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_G, EFLAGS),
|
||||
(CMOVLE64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_P, EFLAGS),
|
||||
(CMOVNP64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NP, EFLAGS),
|
||||
(CMOVP64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_S, EFLAGS),
|
||||
(CMOVNS64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NS, EFLAGS),
|
||||
(CMOVS64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_O, EFLAGS),
|
||||
(CMOVNO64rm GR64:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi64 addr:$src1), GR64:$src2, X86_COND_NO, EFLAGS),
|
||||
(CMOVO64rm GR64:$src2, addr:$src1)>;
|
||||
|
||||
|
||||
// zextload bool -> zextload byte
|
||||
def : Pat<(zextloadi8i1 addr:$src), (MOV8rm addr:$src)>;
|
||||
def : Pat<(zextloadi16i1 addr:$src), (MOVZX16rm8 addr:$src)>;
|
||||
def : Pat<(zextloadi32i1 addr:$src), (MOVZX32rm8 addr:$src)>;
|
||||
def : Pat<(zextloadi64i1 addr:$src), (MOVZX64rm8 addr:$src)>;
|
||||
|
||||
// extload bool -> extload byte
|
||||
// When extloading from 16-bit and smaller memory locations into 64-bit
|
||||
// registers, use zero-extending loads so that the entire 64-bit register is
|
||||
// defined, avoiding partial-register updates.
|
||||
|
||||
def : Pat<(extloadi8i1 addr:$src), (MOV8rm addr:$src)>;
|
||||
def : Pat<(extloadi16i1 addr:$src), (MOVZX16rm8 addr:$src)>;
|
||||
def : Pat<(extloadi32i1 addr:$src), (MOVZX32rm8 addr:$src)>;
|
||||
def : Pat<(extloadi16i8 addr:$src), (MOVZX16rm8 addr:$src)>;
|
||||
def : Pat<(extloadi32i8 addr:$src), (MOVZX32rm8 addr:$src)>;
|
||||
def : Pat<(extloadi32i16 addr:$src), (MOVZX32rm16 addr:$src)>;
|
||||
|
||||
def : Pat<(extloadi64i1 addr:$src), (MOVZX64rm8 addr:$src)>;
|
||||
def : Pat<(extloadi64i8 addr:$src), (MOVZX64rm8 addr:$src)>;
|
||||
def : Pat<(extloadi64i16 addr:$src), (MOVZX64rm16 addr:$src)>;
|
||||
// For other extloads, use subregs, since the high contents of the register are
|
||||
// defined after an extload.
|
||||
def : Pat<(extloadi64i32 addr:$src),
|
||||
(SUBREG_TO_REG (i64 0), (MOV32rm addr:$src),
|
||||
sub_32bit)>;
|
||||
|
||||
// anyext. Define these to do an explicit zero-extend to
|
||||
// avoid partial-register updates.
|
||||
def : Pat<(i16 (anyext GR8 :$src)), (MOVZX16rr8 GR8 :$src)>;
|
||||
def : Pat<(i32 (anyext GR8 :$src)), (MOVZX32rr8 GR8 :$src)>;
|
||||
|
||||
// Except for i16 -> i32 since isel expect i16 ops to be promoted to i32.
|
||||
def : Pat<(i32 (anyext GR16:$src)),
|
||||
(INSERT_SUBREG (i32 (IMPLICIT_DEF)), GR16:$src, sub_16bit)>;
|
||||
|
||||
def : Pat<(i64 (anyext GR8 :$src)), (MOVZX64rr8 GR8 :$src)>;
|
||||
def : Pat<(i64 (anyext GR16:$src)), (MOVZX64rr16 GR16 :$src)>;
|
||||
def : Pat<(i64 (anyext GR32:$src)),
|
||||
(SUBREG_TO_REG (i64 0), GR32:$src, sub_32bit)>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Some peepholes
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// Odd encoding trick: -128 fits into an 8-bit immediate field while
|
||||
// +128 doesn't, so in this special case use a sub instead of an add.
|
||||
def : Pat<(add GR16:$src1, 128),
|
||||
(SUB16ri8 GR16:$src1, -128)>;
|
||||
def : Pat<(store (add (loadi16 addr:$dst), 128), addr:$dst),
|
||||
(SUB16mi8 addr:$dst, -128)>;
|
||||
|
||||
def : Pat<(add GR32:$src1, 128),
|
||||
(SUB32ri8 GR32:$src1, -128)>;
|
||||
def : Pat<(store (add (loadi32 addr:$dst), 128), addr:$dst),
|
||||
(SUB32mi8 addr:$dst, -128)>;
|
||||
|
||||
def : Pat<(add GR64:$src1, 128),
|
||||
(SUB64ri8 GR64:$src1, -128)>;
|
||||
def : Pat<(store (add (loadi64 addr:$dst), 128), addr:$dst),
|
||||
(SUB64mi8 addr:$dst, -128)>;
|
||||
|
||||
// The same trick applies for 32-bit immediate fields in 64-bit
|
||||
// instructions.
|
||||
def : Pat<(add GR64:$src1, 0x0000000080000000),
|
||||
(SUB64ri32 GR64:$src1, 0xffffffff80000000)>;
|
||||
def : Pat<(store (add (loadi64 addr:$dst), 0x00000000800000000), addr:$dst),
|
||||
(SUB64mi32 addr:$dst, 0xffffffff80000000)>;
|
||||
|
||||
// Use a 32-bit and with implicit zero-extension instead of a 64-bit and if it
|
||||
// has an immediate with at least 32 bits of leading zeros, to avoid needing to
|
||||
// materialize that immediate in a register first.
|
||||
def : Pat<(and GR64:$src, i64immZExt32:$imm),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(AND32ri
|
||||
(EXTRACT_SUBREG GR64:$src, sub_32bit),
|
||||
(i32 (GetLo32XForm imm:$imm))),
|
||||
sub_32bit)>;
|
||||
|
||||
|
||||
// r & (2^16-1) ==> movz
|
||||
def : Pat<(and GR32:$src1, 0xffff),
|
||||
(MOVZX32rr16 (EXTRACT_SUBREG GR32:$src1, sub_16bit))>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR32:$src1, 0xff),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src1,
|
||||
GR32_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR16:$src1, 0xff),
|
||||
(MOVZX16rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src1,
|
||||
GR16_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// r & (2^32-1) ==> movz
|
||||
def : Pat<(and GR64:$src, 0x00000000FFFFFFFF),
|
||||
(MOVZX64rr32 (EXTRACT_SUBREG GR64:$src, sub_32bit))>;
|
||||
// r & (2^16-1) ==> movz
|
||||
def : Pat<(and GR64:$src, 0xffff),
|
||||
(MOVZX64rr16 (i16 (EXTRACT_SUBREG GR64:$src, sub_16bit)))>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR64:$src, 0xff),
|
||||
(MOVZX64rr8 (i8 (EXTRACT_SUBREG GR64:$src, sub_8bit)))>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR32:$src1, 0xff),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG GR32:$src1, sub_8bit))>,
|
||||
Requires<[In64BitMode]>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR16:$src1, 0xff),
|
||||
(MOVZX16rr8 (i8 (EXTRACT_SUBREG GR16:$src1, sub_8bit)))>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
|
||||
// sext_inreg patterns
|
||||
def : Pat<(sext_inreg GR32:$src, i16),
|
||||
(MOVSX32rr16 (EXTRACT_SUBREG GR32:$src, sub_16bit))>;
|
||||
def : Pat<(sext_inreg GR32:$src, i8),
|
||||
(MOVSX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(sext_inreg GR16:$src, i8),
|
||||
(MOVSX16rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src,
|
||||
GR16_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
def : Pat<(sext_inreg GR64:$src, i32),
|
||||
(MOVSX64rr32 (EXTRACT_SUBREG GR64:$src, sub_32bit))>;
|
||||
def : Pat<(sext_inreg GR64:$src, i16),
|
||||
(MOVSX64rr16 (EXTRACT_SUBREG GR64:$src, sub_16bit))>;
|
||||
def : Pat<(sext_inreg GR64:$src, i8),
|
||||
(MOVSX64rr8 (EXTRACT_SUBREG GR64:$src, sub_8bit))>;
|
||||
def : Pat<(sext_inreg GR32:$src, i8),
|
||||
(MOVSX32rr8 (EXTRACT_SUBREG GR32:$src, sub_8bit))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(sext_inreg GR16:$src, i8),
|
||||
(MOVSX16rr8 (i8 (EXTRACT_SUBREG GR16:$src, sub_8bit)))>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
|
||||
// trunc patterns
|
||||
def : Pat<(i16 (trunc GR32:$src)),
|
||||
(EXTRACT_SUBREG GR32:$src, sub_16bit)>;
|
||||
def : Pat<(i8 (trunc GR32:$src)),
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i8 (trunc GR16:$src)),
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i32 (trunc GR64:$src)),
|
||||
(EXTRACT_SUBREG GR64:$src, sub_32bit)>;
|
||||
def : Pat<(i16 (trunc GR64:$src)),
|
||||
(EXTRACT_SUBREG GR64:$src, sub_16bit)>;
|
||||
def : Pat<(i8 (trunc GR64:$src)),
|
||||
(EXTRACT_SUBREG GR64:$src, sub_8bit)>;
|
||||
def : Pat<(i8 (trunc GR32:$src)),
|
||||
(EXTRACT_SUBREG GR32:$src, sub_8bit)>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i8 (trunc GR16:$src)),
|
||||
(EXTRACT_SUBREG GR16:$src, sub_8bit)>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
// h-register tricks
|
||||
def : Pat<(i8 (trunc (srl_su GR16:$src, (i8 8)))),
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i8 (trunc (srl_su GR32:$src, (i8 8)))),
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit_hi)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(srl GR16:$src, (i8 8)),
|
||||
(EXTRACT_SUBREG
|
||||
(MOVZX32rr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_16bit)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i32 (zext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src,
|
||||
GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i32 (anyext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src,
|
||||
GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(and (srl_su GR32:$src, (i8 8)), (i32 255)),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(srl (and_su GR32:$src, 0xff00), (i8 8)),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// h-register tricks.
|
||||
// For now, be conservative on x86-64 and use an h-register extract only if the
|
||||
// value is immediately zero-extended or stored, which are somewhat common
|
||||
// cases. This uses a bunch of code to prevent a register requiring a REX prefix
|
||||
// from being allocated in the same instruction as the h register, as there's
|
||||
// currently no way to describe this requirement to the register allocator.
|
||||
|
||||
// h-register extract and zero-extend.
|
||||
def : Pat<(and (srl_su GR64:$src, (i8 8)), (i64 255)),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i64 (COPY_TO_REGCLASS GR64:$src, GR64_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_32bit)>;
|
||||
def : Pat<(and (srl_su GR32:$src, (i8 8)), (i32 255)),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(srl (and_su GR32:$src, 0xff00), (i8 8)),
|
||||
(MOVZX32_NOREXrr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(srl GR16:$src, (i8 8)),
|
||||
(EXTRACT_SUBREG
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_16bit)>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i32 (zext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i32 (anyext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(i64 (zext (srl_su GR16:$src, (i8 8)))),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_32bit)>;
|
||||
def : Pat<(i64 (anyext (srl_su GR16:$src, (i8 8)))),
|
||||
(SUBREG_TO_REG
|
||||
(i64 0),
|
||||
(MOVZX32_NOREXrr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_32bit)>;
|
||||
|
||||
// h-register extract and store.
|
||||
def : Pat<(store (i8 (trunc_su (srl_su GR64:$src, (i8 8)))), addr:$dst),
|
||||
(MOV8mr_NOREX
|
||||
addr:$dst,
|
||||
(EXTRACT_SUBREG (i64 (COPY_TO_REGCLASS GR64:$src, GR64_ABCD)),
|
||||
sub_8bit_hi))>;
|
||||
def : Pat<(store (i8 (trunc_su (srl_su GR32:$src, (i8 8)))), addr:$dst),
|
||||
(MOV8mr_NOREX
|
||||
addr:$dst,
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
def : Pat<(store (i8 (trunc_su (srl_su GR16:$src, (i8 8)))), addr:$dst),
|
||||
(MOV8mr_NOREX
|
||||
addr:$dst,
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
|
||||
// (shl x, 1) ==> (add x, x)
|
||||
def : Pat<(shl GR8 :$src1, (i8 1)), (ADD8rr GR8 :$src1, GR8 :$src1)>;
|
||||
def : Pat<(shl GR16:$src1, (i8 1)), (ADD16rr GR16:$src1, GR16:$src1)>;
|
||||
def : Pat<(shl GR32:$src1, (i8 1)), (ADD32rr GR32:$src1, GR32:$src1)>;
|
||||
def : Pat<(shl GR64:$src1, (i8 1)), (ADD64rr GR64:$src1, GR64:$src1)>;
|
||||
|
||||
// (shl x (and y, 31)) ==> (shl x, y)
|
||||
def : Pat<(shl GR8:$src1, (and CL, 31)),
|
||||
(SHL8rCL GR8:$src1)>;
|
||||
def : Pat<(shl GR16:$src1, (and CL, 31)),
|
||||
(SHL16rCL GR16:$src1)>;
|
||||
def : Pat<(shl GR32:$src1, (and CL, 31)),
|
||||
(SHL32rCL GR32:$src1)>;
|
||||
def : Pat<(store (shl (loadi8 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHL8mCL addr:$dst)>;
|
||||
def : Pat<(store (shl (loadi16 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHL16mCL addr:$dst)>;
|
||||
def : Pat<(store (shl (loadi32 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHL32mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(srl GR8:$src1, (and CL, 31)),
|
||||
(SHR8rCL GR8:$src1)>;
|
||||
def : Pat<(srl GR16:$src1, (and CL, 31)),
|
||||
(SHR16rCL GR16:$src1)>;
|
||||
def : Pat<(srl GR32:$src1, (and CL, 31)),
|
||||
(SHR32rCL GR32:$src1)>;
|
||||
def : Pat<(store (srl (loadi8 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHR8mCL addr:$dst)>;
|
||||
def : Pat<(store (srl (loadi16 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHR16mCL addr:$dst)>;
|
||||
def : Pat<(store (srl (loadi32 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHR32mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(sra GR8:$src1, (and CL, 31)),
|
||||
(SAR8rCL GR8:$src1)>;
|
||||
def : Pat<(sra GR16:$src1, (and CL, 31)),
|
||||
(SAR16rCL GR16:$src1)>;
|
||||
def : Pat<(sra GR32:$src1, (and CL, 31)),
|
||||
(SAR32rCL GR32:$src1)>;
|
||||
def : Pat<(store (sra (loadi8 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SAR8mCL addr:$dst)>;
|
||||
def : Pat<(store (sra (loadi16 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SAR16mCL addr:$dst)>;
|
||||
def : Pat<(store (sra (loadi32 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SAR32mCL addr:$dst)>;
|
||||
|
||||
// (shl x (and y, 63)) ==> (shl x, y)
|
||||
def : Pat<(shl GR64:$src1, (and CL, 63)),
|
||||
(SHL64rCL GR64:$src1)>;
|
||||
def : Pat<(store (shl (loadi64 addr:$dst), (and CL, 63)), addr:$dst),
|
||||
(SHL64mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(srl GR64:$src1, (and CL, 63)),
|
||||
(SHR64rCL GR64:$src1)>;
|
||||
def : Pat<(store (srl (loadi64 addr:$dst), (and CL, 63)), addr:$dst),
|
||||
(SHR64mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(sra GR64:$src1, (and CL, 63)),
|
||||
(SAR64rCL GR64:$src1)>;
|
||||
def : Pat<(store (sra (loadi64 addr:$dst), (and CL, 63)), addr:$dst),
|
||||
(SAR64mCL addr:$dst)>;
|
||||
|
||||
|
||||
// (anyext (setcc_carry)) -> (setcc_carry)
|
||||
def : Pat<(i16 (anyext (i8 (X86setcc_c X86_COND_B, EFLAGS)))),
|
||||
(SETB_C16r)>;
|
||||
def : Pat<(i32 (anyext (i8 (X86setcc_c X86_COND_B, EFLAGS)))),
|
||||
(SETB_C32r)>;
|
||||
def : Pat<(i32 (anyext (i16 (X86setcc_c X86_COND_B, EFLAGS)))),
|
||||
(SETB_C32r)>;
|
||||
|
||||
// (or x1, x2) -> (add x1, x2) if two operands are known not to share bits.
|
||||
let AddedComplexity = 5 in { // Try this before the selecting to OR
|
||||
def : Pat<(or_is_add GR16:$src1, imm:$src2),
|
||||
(ADD16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(or_is_add GR32:$src1, imm:$src2),
|
||||
(ADD32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(or_is_add GR16:$src1, i16immSExt8:$src2),
|
||||
(ADD16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(or_is_add GR32:$src1, i32immSExt8:$src2),
|
||||
(ADD32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
def : Pat<(or_is_add GR16:$src1, GR16:$src2),
|
||||
(ADD16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(or_is_add GR32:$src1, GR32:$src2),
|
||||
(ADD32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(or_is_add GR64:$src1, i64immSExt8:$src2),
|
||||
(ADD64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(or_is_add GR64:$src1, i64immSExt32:$src2),
|
||||
(ADD64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(or_is_add GR64:$src1, GR64:$src2),
|
||||
(ADD64rr GR64:$src1, GR64:$src2)>;
|
||||
} // AddedComplexity
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// EFLAGS-defining Patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// add reg, reg
|
||||
def : Pat<(add GR8 :$src1, GR8 :$src2), (ADD8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(add GR16:$src1, GR16:$src2), (ADD16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(add GR32:$src1, GR32:$src2), (ADD32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// add reg, mem
|
||||
def : Pat<(add GR8:$src1, (loadi8 addr:$src2)),
|
||||
(ADD8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(add GR16:$src1, (loadi16 addr:$src2)),
|
||||
(ADD16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (loadi32 addr:$src2)),
|
||||
(ADD32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// add reg, imm
|
||||
def : Pat<(add GR8 :$src1, imm:$src2), (ADD8ri GR8:$src1 , imm:$src2)>;
|
||||
def : Pat<(add GR16:$src1, imm:$src2), (ADD16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(add GR32:$src1, imm:$src2), (ADD32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(add GR16:$src1, i16immSExt8:$src2),
|
||||
(ADD16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(add GR32:$src1, i32immSExt8:$src2),
|
||||
(ADD32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// sub reg, reg
|
||||
def : Pat<(sub GR8 :$src1, GR8 :$src2), (SUB8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(sub GR16:$src1, GR16:$src2), (SUB16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, GR32:$src2), (SUB32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// sub reg, mem
|
||||
def : Pat<(sub GR8:$src1, (loadi8 addr:$src2)),
|
||||
(SUB8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(sub GR16:$src1, (loadi16 addr:$src2)),
|
||||
(SUB16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, (loadi32 addr:$src2)),
|
||||
(SUB32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// sub reg, imm
|
||||
def : Pat<(sub GR8:$src1, imm:$src2),
|
||||
(SUB8ri GR8:$src1, imm:$src2)>;
|
||||
def : Pat<(sub GR16:$src1, imm:$src2),
|
||||
(SUB16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, imm:$src2),
|
||||
(SUB32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(sub GR16:$src1, i16immSExt8:$src2),
|
||||
(SUB16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, i32immSExt8:$src2),
|
||||
(SUB32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// mul reg, reg
|
||||
def : Pat<(mul GR16:$src1, GR16:$src2),
|
||||
(IMUL16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, GR32:$src2),
|
||||
(IMUL32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// mul reg, mem
|
||||
def : Pat<(mul GR16:$src1, (loadi16 addr:$src2)),
|
||||
(IMUL16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, (loadi32 addr:$src2)),
|
||||
(IMUL32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// mul reg, imm
|
||||
def : Pat<(mul GR16:$src1, imm:$src2),
|
||||
(IMUL16rri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, imm:$src2),
|
||||
(IMUL32rri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(mul GR16:$src1, i16immSExt8:$src2),
|
||||
(IMUL16rri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, i32immSExt8:$src2),
|
||||
(IMUL32rri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// reg = mul mem, imm
|
||||
def : Pat<(mul (loadi16 addr:$src1), imm:$src2),
|
||||
(IMUL16rmi addr:$src1, imm:$src2)>;
|
||||
def : Pat<(mul (loadi32 addr:$src1), imm:$src2),
|
||||
(IMUL32rmi addr:$src1, imm:$src2)>;
|
||||
def : Pat<(mul (loadi16 addr:$src1), i16immSExt8:$src2),
|
||||
(IMUL16rmi8 addr:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(mul (loadi32 addr:$src1), i32immSExt8:$src2),
|
||||
(IMUL32rmi8 addr:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// Optimize multiply by 2 with EFLAGS result.
|
||||
let AddedComplexity = 2 in {
|
||||
def : Pat<(X86smul_flag GR16:$src1, 2), (ADD16rr GR16:$src1, GR16:$src1)>;
|
||||
def : Pat<(X86smul_flag GR32:$src1, 2), (ADD32rr GR32:$src1, GR32:$src1)>;
|
||||
}
|
||||
|
||||
// Patterns for nodes that do not produce flags, for instructions that do.
|
||||
|
||||
// addition
|
||||
def : Pat<(add GR64:$src1, GR64:$src2),
|
||||
(ADD64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(add GR64:$src1, i64immSExt8:$src2),
|
||||
(ADD64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(add GR64:$src1, i64immSExt32:$src2),
|
||||
(ADD64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(add GR64:$src1, (loadi64 addr:$src2)),
|
||||
(ADD64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// subtraction
|
||||
def : Pat<(sub GR64:$src1, GR64:$src2),
|
||||
(SUB64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(sub GR64:$src1, (loadi64 addr:$src2)),
|
||||
(SUB64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(sub GR64:$src1, i64immSExt8:$src2),
|
||||
(SUB64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(sub GR64:$src1, i64immSExt32:$src2),
|
||||
(SUB64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// Multiply
|
||||
def : Pat<(mul GR64:$src1, GR64:$src2),
|
||||
(IMUL64rr GR64:$src1, GR64:$src2)>;
|
||||
def : Pat<(mul GR64:$src1, (loadi64 addr:$src2)),
|
||||
(IMUL64rm GR64:$src1, addr:$src2)>;
|
||||
def : Pat<(mul GR64:$src1, i64immSExt8:$src2),
|
||||
(IMUL64rri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(mul GR64:$src1, i64immSExt32:$src2),
|
||||
(IMUL64rri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
def : Pat<(mul (loadi64 addr:$src1), i64immSExt8:$src2),
|
||||
(IMUL64rmi8 addr:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(mul (loadi64 addr:$src1), i64immSExt32:$src2),
|
||||
(IMUL64rmi32 addr:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// Increment reg.
|
||||
def : Pat<(add GR8 :$src, 1), (INC8r GR8 :$src)>;
|
||||
def : Pat<(add GR16:$src, 1), (INC16r GR16:$src)>, Requires<[In32BitMode]>;
|
||||
def : Pat<(add GR16:$src, 1), (INC64_16r GR16:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR32:$src, 1), (INC32r GR32:$src)>, Requires<[In32BitMode]>;
|
||||
def : Pat<(add GR32:$src, 1), (INC64_32r GR32:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR64:$src, 1), (INC64r GR64:$src)>;
|
||||
|
||||
// Decrement reg.
|
||||
def : Pat<(add GR8 :$src, -1), (DEC8r GR8 :$src)>;
|
||||
def : Pat<(add GR16:$src, -1), (DEC16r GR16:$src)>, Requires<[In32BitMode]>;
|
||||
def : Pat<(add GR16:$src, -1), (DEC64_16r GR16:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR32:$src, -1), (DEC32r GR32:$src)>, Requires<[In32BitMode]>;
|
||||
def : Pat<(add GR32:$src, -1), (DEC64_32r GR32:$src)>, Requires<[In64BitMode]>;
|
||||
def : Pat<(add GR64:$src, -1), (DEC64r GR64:$src)>;
|
||||
|
||||
// or reg/reg.
|
||||
def : Pat<(or GR8 :$src1, GR8 :$src2), (OR8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(or GR16:$src1, GR16:$src2), (OR16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(or GR32:$src1, GR32:$src2), (OR32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(or GR64:$src1, GR64:$src2), (OR64rr GR64:$src1, GR64:$src2)>;
|
||||
|
||||
// or reg/mem
|
||||
def : Pat<(or GR8:$src1, (loadi8 addr:$src2)),
|
||||
(OR8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(or GR16:$src1, (loadi16 addr:$src2)),
|
||||
(OR16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(or GR32:$src1, (loadi32 addr:$src2)),
|
||||
(OR32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(or GR64:$src1, (loadi64 addr:$src2)),
|
||||
(OR64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// or reg/imm
|
||||
def : Pat<(or GR8:$src1 , imm:$src2), (OR8ri GR8 :$src1, imm:$src2)>;
|
||||
def : Pat<(or GR16:$src1, imm:$src2), (OR16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(or GR32:$src1, imm:$src2), (OR32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(or GR16:$src1, i16immSExt8:$src2),
|
||||
(OR16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(or GR32:$src1, i32immSExt8:$src2),
|
||||
(OR32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
def : Pat<(or GR64:$src1, i64immSExt8:$src2),
|
||||
(OR64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(or GR64:$src1, i64immSExt32:$src2),
|
||||
(OR64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// xor reg/reg
|
||||
def : Pat<(xor GR8 :$src1, GR8 :$src2), (XOR8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(xor GR16:$src1, GR16:$src2), (XOR16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, GR32:$src2), (XOR32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, GR64:$src2), (XOR64rr GR64:$src1, GR64:$src2)>;
|
||||
|
||||
// xor reg/mem
|
||||
def : Pat<(xor GR8:$src1, (loadi8 addr:$src2)),
|
||||
(XOR8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(xor GR16:$src1, (loadi16 addr:$src2)),
|
||||
(XOR16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, (loadi32 addr:$src2)),
|
||||
(XOR32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, (loadi64 addr:$src2)),
|
||||
(XOR64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// xor reg/imm
|
||||
def : Pat<(xor GR8:$src1, imm:$src2),
|
||||
(XOR8ri GR8:$src1, imm:$src2)>;
|
||||
def : Pat<(xor GR16:$src1, imm:$src2),
|
||||
(XOR16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, imm:$src2),
|
||||
(XOR32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(xor GR16:$src1, i16immSExt8:$src2),
|
||||
(XOR16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, i32immSExt8:$src2),
|
||||
(XOR32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, i64immSExt8:$src2),
|
||||
(XOR64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(xor GR64:$src1, i64immSExt32:$src2),
|
||||
(XOR64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
// and reg/reg
|
||||
def : Pat<(and GR8 :$src1, GR8 :$src2), (AND8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(and GR16:$src1, GR16:$src2), (AND16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(and GR32:$src1, GR32:$src2), (AND32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(and GR64:$src1, GR64:$src2), (AND64rr GR64:$src1, GR64:$src2)>;
|
||||
|
||||
// and reg/mem
|
||||
def : Pat<(and GR8:$src1, (loadi8 addr:$src2)),
|
||||
(AND8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(and GR16:$src1, (loadi16 addr:$src2)),
|
||||
(AND16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(and GR32:$src1, (loadi32 addr:$src2)),
|
||||
(AND32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(and GR64:$src1, (loadi64 addr:$src2)),
|
||||
(AND64rm GR64:$src1, addr:$src2)>;
|
||||
|
||||
// and reg/imm
|
||||
def : Pat<(and GR8:$src1, imm:$src2),
|
||||
(AND8ri GR8:$src1, imm:$src2)>;
|
||||
def : Pat<(and GR16:$src1, imm:$src2),
|
||||
(AND16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(and GR32:$src1, imm:$src2),
|
||||
(AND32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(and GR16:$src1, i16immSExt8:$src2),
|
||||
(AND16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(and GR32:$src1, i32immSExt8:$src2),
|
||||
(AND32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
def : Pat<(and GR64:$src1, i64immSExt8:$src2),
|
||||
(AND64ri8 GR64:$src1, i64immSExt8:$src2)>;
|
||||
def : Pat<(and GR64:$src1, i64immSExt32:$src2),
|
||||
(AND64ri32 GR64:$src1, i64immSExt32:$src2)>;
|
||||
|
||||
|
||||
|
291
lib/Target/X86/X86InstrControl.td
Normal file
291
lib/Target/X86/X86InstrControl.td
Normal file
@ -0,0 +1,291 @@
|
||||
//===- X86InstrControl.td - Control Flow Instructions ------*- tablegen -*-===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// This file describes the X86 jump, return, call, and related instructions.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Control Flow Instructions.
|
||||
//
|
||||
|
||||
// Return instructions.
|
||||
let isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
hasCtrlDep = 1, FPForm = SpecialFP in {
|
||||
def RET : I <0xC3, RawFrm, (outs), (ins variable_ops),
|
||||
"ret",
|
||||
[(X86retflag 0)]>;
|
||||
def RETI : Ii16<0xC2, RawFrm, (outs), (ins i16imm:$amt, variable_ops),
|
||||
"ret\t$amt",
|
||||
[(X86retflag timm:$amt)]>;
|
||||
def LRET : I <0xCB, RawFrm, (outs), (ins),
|
||||
"lret", []>;
|
||||
def LRETI : Ii16<0xCA, RawFrm, (outs), (ins i16imm:$amt),
|
||||
"lret\t$amt", []>;
|
||||
}
|
||||
|
||||
// Unconditional branches.
|
||||
let isBarrier = 1, isBranch = 1, isTerminator = 1 in {
|
||||
def JMP_4 : Ii32PCRel<0xE9, RawFrm, (outs), (ins brtarget:$dst),
|
||||
"jmp\t$dst", [(br bb:$dst)]>;
|
||||
def JMP_1 : Ii8PCRel<0xEB, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jmp\t$dst", []>;
|
||||
def JMP64pcrel32 : I<0xE9, RawFrm, (outs), (ins brtarget:$dst),
|
||||
"jmp{q}\t$dst", []>;
|
||||
}
|
||||
|
||||
// Conditional Branches.
|
||||
let isBranch = 1, isTerminator = 1, Uses = [EFLAGS] in {
|
||||
multiclass ICBr<bits<8> opc1, bits<8> opc4, string asm, PatFrag Cond> {
|
||||
def _1 : Ii8PCRel <opc1, RawFrm, (outs), (ins brtarget8:$dst), asm, []>;
|
||||
def _4 : Ii32PCRel<opc4, RawFrm, (outs), (ins brtarget:$dst), asm,
|
||||
[(X86brcond bb:$dst, Cond, EFLAGS)]>, TB;
|
||||
}
|
||||
}
|
||||
|
||||
defm JO : ICBr<0x70, 0x80, "jo\t$dst" , X86_COND_O>;
|
||||
defm JNO : ICBr<0x71, 0x81, "jno\t$dst" , X86_COND_NO>;
|
||||
defm JB : ICBr<0x72, 0x82, "jb\t$dst" , X86_COND_B>;
|
||||
defm JAE : ICBr<0x73, 0x83, "jae\t$dst", X86_COND_AE>;
|
||||
defm JE : ICBr<0x74, 0x84, "je\t$dst" , X86_COND_E>;
|
||||
defm JNE : ICBr<0x75, 0x85, "jne\t$dst", X86_COND_NE>;
|
||||
defm JBE : ICBr<0x76, 0x86, "jbe\t$dst", X86_COND_BE>;
|
||||
defm JA : ICBr<0x77, 0x87, "ja\t$dst" , X86_COND_A>;
|
||||
defm JS : ICBr<0x78, 0x88, "js\t$dst" , X86_COND_S>;
|
||||
defm JNS : ICBr<0x79, 0x89, "jns\t$dst", X86_COND_NS>;
|
||||
defm JP : ICBr<0x7A, 0x8A, "jp\t$dst" , X86_COND_P>;
|
||||
defm JNP : ICBr<0x7B, 0x8B, "jnp\t$dst", X86_COND_NP>;
|
||||
defm JL : ICBr<0x7C, 0x8C, "jl\t$dst" , X86_COND_L>;
|
||||
defm JGE : ICBr<0x7D, 0x8D, "jge\t$dst", X86_COND_GE>;
|
||||
defm JLE : ICBr<0x7E, 0x8E, "jle\t$dst", X86_COND_LE>;
|
||||
defm JG : ICBr<0x7F, 0x8F, "jg\t$dst" , X86_COND_G>;
|
||||
|
||||
// jcx/jecx/jrcx instructions.
|
||||
let isAsmParserOnly = 1, isBranch = 1, isTerminator = 1 in {
|
||||
// These are the 32-bit versions of this instruction for the asmparser. In
|
||||
// 32-bit mode, the address size prefix is jcxz and the unprefixed version is
|
||||
// jecxz.
|
||||
let Uses = [CX] in
|
||||
def JCXZ : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jcxz\t$dst", []>, AdSize, Requires<[In32BitMode]>;
|
||||
let Uses = [ECX] in
|
||||
def JECXZ_32 : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jecxz\t$dst", []>, Requires<[In32BitMode]>;
|
||||
|
||||
// J*CXZ instruction: 64-bit versions of this instruction for the asmparser.
|
||||
// In 64-bit mode, the address size prefix is jecxz and the unprefixed version
|
||||
// is jrcxz.
|
||||
let Uses = [ECX] in
|
||||
def JECXZ_64 : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jecxz\t$dst", []>, AdSize, Requires<[In64BitMode]>;
|
||||
let Uses = [RCX] in
|
||||
def JRCXZ : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jrcxz\t$dst", []>, Requires<[In64BitMode]>;
|
||||
}
|
||||
|
||||
// Indirect branches
|
||||
let isBranch = 1, isTerminator = 1, isBarrier = 1, isIndirectBranch = 1 in {
|
||||
def JMP32r : I<0xFF, MRM4r, (outs), (ins GR32:$dst), "jmp{l}\t{*}$dst",
|
||||
[(brind GR32:$dst)]>, Requires<[In32BitMode]>;
|
||||
def JMP32m : I<0xFF, MRM4m, (outs), (ins i32mem:$dst), "jmp{l}\t{*}$dst",
|
||||
[(brind (loadi32 addr:$dst))]>, Requires<[In32BitMode]>;
|
||||
|
||||
def JMP64r : I<0xFF, MRM4r, (outs), (ins GR64:$dst), "jmp{q}\t{*}$dst",
|
||||
[(brind GR64:$dst)]>, Requires<[In64BitMode]>;
|
||||
def JMP64m : I<0xFF, MRM4m, (outs), (ins i64mem:$dst), "jmp{q}\t{*}$dst",
|
||||
[(brind (loadi64 addr:$dst))]>, Requires<[In64BitMode]>;
|
||||
|
||||
def FARJMP16i : Iseg16<0xEA, RawFrmImm16, (outs),
|
||||
(ins i16imm:$off, i16imm:$seg),
|
||||
"ljmp{w}\t{$seg, $off|$off, $seg}", []>, OpSize;
|
||||
def FARJMP32i : Iseg32<0xEA, RawFrmImm16, (outs),
|
||||
(ins i32imm:$off, i16imm:$seg),
|
||||
"ljmp{l}\t{$seg, $off|$off, $seg}", []>;
|
||||
def FARJMP64 : RI<0xFF, MRM5m, (outs), (ins opaque80mem:$dst),
|
||||
"ljmp{q}\t{*}$dst", []>;
|
||||
|
||||
def FARJMP16m : I<0xFF, MRM5m, (outs), (ins opaque32mem:$dst),
|
||||
"ljmp{w}\t{*}$dst", []>, OpSize;
|
||||
def FARJMP32m : I<0xFF, MRM5m, (outs), (ins opaque48mem:$dst),
|
||||
"ljmp{l}\t{*}$dst", []>;
|
||||
}
|
||||
|
||||
|
||||
// Loop instructions
|
||||
|
||||
def LOOP : Ii8PCRel<0xE2, RawFrm, (outs), (ins brtarget8:$dst), "loop\t$dst", []>;
|
||||
def LOOPE : Ii8PCRel<0xE1, RawFrm, (outs), (ins brtarget8:$dst), "loope\t$dst", []>;
|
||||
def LOOPNE : Ii8PCRel<0xE0, RawFrm, (outs), (ins brtarget8:$dst), "loopne\t$dst", []>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Call Instructions...
|
||||
//
|
||||
let isCall = 1 in
|
||||
// All calls clobber the non-callee saved registers. ESP is marked as
|
||||
// a use to prevent stack-pointer assignments that appear immediately
|
||||
// before calls from potentially appearing dead. Uses for argument
|
||||
// registers are added manually.
|
||||
let Defs = [EAX, ECX, EDX, FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [ESP] in {
|
||||
def CALLpcrel32 : Ii32PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i32imm_pcrel:$dst,variable_ops),
|
||||
"call{l}\t$dst", []>, Requires<[In32BitMode]>;
|
||||
def CALL32r : I<0xFF, MRM2r, (outs), (ins GR32:$dst, variable_ops),
|
||||
"call{l}\t{*}$dst", [(X86call GR32:$dst)]>,
|
||||
Requires<[In32BitMode]>;
|
||||
def CALL32m : I<0xFF, MRM2m, (outs), (ins i32mem:$dst, variable_ops),
|
||||
"call{l}\t{*}$dst", [(X86call (loadi32 addr:$dst))]>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
def FARCALL16i : Iseg16<0x9A, RawFrmImm16, (outs),
|
||||
(ins i16imm:$off, i16imm:$seg),
|
||||
"lcall{w}\t{$seg, $off|$off, $seg}", []>, OpSize;
|
||||
def FARCALL32i : Iseg32<0x9A, RawFrmImm16, (outs),
|
||||
(ins i32imm:$off, i16imm:$seg),
|
||||
"lcall{l}\t{$seg, $off|$off, $seg}", []>;
|
||||
|
||||
def FARCALL16m : I<0xFF, MRM3m, (outs), (ins opaque32mem:$dst),
|
||||
"lcall{w}\t{*}$dst", []>, OpSize;
|
||||
def FARCALL32m : I<0xFF, MRM3m, (outs), (ins opaque48mem:$dst),
|
||||
"lcall{l}\t{*}$dst", []>;
|
||||
|
||||
// callw for 16 bit code for the assembler.
|
||||
let isAsmParserOnly = 1 in
|
||||
def CALLpcrel16 : Ii16PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i16imm_pcrel:$dst, variable_ops),
|
||||
"callw\t$dst", []>, OpSize;
|
||||
}
|
||||
|
||||
|
||||
// Tail call stuff.
|
||||
|
||||
let isCall = 1, isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
isCodeGenOnly = 1 in
|
||||
let Defs = [EAX, ECX, EDX, FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [ESP] in {
|
||||
def TCRETURNdi : I<0, Pseudo, (outs),
|
||||
(ins i32imm_pcrel:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
def TCRETURNri : I<0, Pseudo, (outs),
|
||||
(ins GR32_TC:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
let mayLoad = 1 in
|
||||
def TCRETURNmi : I<0, Pseudo, (outs),
|
||||
(ins i32mem_TC:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
|
||||
// FIXME: The should be pseudo instructions that are lowered when going to
|
||||
// mcinst.
|
||||
def TAILJMPd : Ii32PCRel<0xE9, RawFrm, (outs),
|
||||
(ins i32imm_pcrel:$dst, variable_ops),
|
||||
"jmp\t$dst # TAILCALL",
|
||||
[]>;
|
||||
def TAILJMPr : I<0xFF, MRM4r, (outs), (ins GR32_TC:$dst, variable_ops),
|
||||
"", []>; // FIXME: Remove encoding when JIT is dead.
|
||||
let mayLoad = 1 in
|
||||
def TAILJMPm : I<0xFF, MRM4m, (outs), (ins i32mem_TC:$dst, variable_ops),
|
||||
"jmp{l}\t{*}$dst # TAILCALL", []>;
|
||||
}
|
||||
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Call Instructions...
|
||||
//
|
||||
let isCall = 1 in
|
||||
// All calls clobber the non-callee saved registers. RSP is marked as
|
||||
// a use to prevent stack-pointer assignments that appear immediately
|
||||
// before calls from potentially appearing dead. Uses for argument
|
||||
// registers are added manually.
|
||||
let Defs = [RAX, RCX, RDX, RSI, RDI, R8, R9, R10, R11,
|
||||
FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0, ST1,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [RSP] in {
|
||||
|
||||
// NOTE: this pattern doesn't match "X86call imm", because we do not know
|
||||
// that the offset between an arbitrary immediate and the call will fit in
|
||||
// the 32-bit pcrel field that we have.
|
||||
def CALL64pcrel32 : Ii32PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i64i32imm_pcrel:$dst, variable_ops),
|
||||
"call{q}\t$dst", []>,
|
||||
Requires<[In64BitMode, NotWin64]>;
|
||||
def CALL64r : I<0xFF, MRM2r, (outs), (ins GR64:$dst, variable_ops),
|
||||
"call{q}\t{*}$dst", [(X86call GR64:$dst)]>,
|
||||
Requires<[In64BitMode, NotWin64]>;
|
||||
def CALL64m : I<0xFF, MRM2m, (outs), (ins i64mem:$dst, variable_ops),
|
||||
"call{q}\t{*}$dst", [(X86call (loadi64 addr:$dst))]>,
|
||||
Requires<[In64BitMode, NotWin64]>;
|
||||
|
||||
def FARCALL64 : RI<0xFF, MRM3m, (outs), (ins opaque80mem:$dst),
|
||||
"lcall{q}\t{*}$dst", []>;
|
||||
}
|
||||
|
||||
// FIXME: We need to teach codegen about single list of call-clobbered
|
||||
// registers.
|
||||
let isCall = 1, isCodeGenOnly = 1 in
|
||||
// All calls clobber the non-callee saved registers. RSP is marked as
|
||||
// a use to prevent stack-pointer assignments that appear immediately
|
||||
// before calls from potentially appearing dead. Uses for argument
|
||||
// registers are added manually.
|
||||
let Defs = [RAX, RCX, RDX, R8, R9, R10, R11,
|
||||
FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0, ST1,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, EFLAGS],
|
||||
Uses = [RSP] in {
|
||||
def WINCALL64pcrel32 : Ii32PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i64i32imm_pcrel:$dst, variable_ops),
|
||||
"call{q}\t$dst", []>,
|
||||
Requires<[IsWin64]>;
|
||||
def WINCALL64r : I<0xFF, MRM2r, (outs), (ins GR64:$dst, variable_ops),
|
||||
"call{q}\t{*}$dst",
|
||||
[(X86call GR64:$dst)]>, Requires<[IsWin64]>;
|
||||
def WINCALL64m : I<0xFF, MRM2m, (outs), (ins i64mem:$dst,variable_ops),
|
||||
"call{q}\t{*}$dst",
|
||||
[(X86call (loadi64 addr:$dst))]>,
|
||||
Requires<[IsWin64]>;
|
||||
}
|
||||
|
||||
|
||||
let isCall = 1, isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
isCodeGenOnly = 1 in
|
||||
let Defs = [RAX, RCX, RDX, RSI, RDI, R8, R9, R10, R11,
|
||||
FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0, ST1,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [RSP] in {
|
||||
def TCRETURNdi64 : I<0, Pseudo, (outs),
|
||||
(ins i64i32imm_pcrel:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
def TCRETURNri64 : I<0, Pseudo, (outs), (ins GR64_TC:$dst, i32imm:$offset,
|
||||
variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
let mayLoad = 1 in
|
||||
def TCRETURNmi64 : I<0, Pseudo, (outs),
|
||||
(ins i64mem_TC:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
|
||||
def TAILJMPd64 : Ii32PCRel<0xE9, RawFrm, (outs),
|
||||
(ins i64i32imm_pcrel:$dst, variable_ops),
|
||||
"jmp\t$dst # TAILCALL", []>;
|
||||
def TAILJMPr64 : I<0xFF, MRM4r, (outs), (ins GR64_TC:$dst, variable_ops),
|
||||
"jmp{q}\t{*}$dst # TAILCALL", []>;
|
||||
|
||||
let mayLoad = 1 in
|
||||
def TAILJMPm64 : I<0xFF, MRM4m, (outs), (ins i64mem_TC:$dst, variable_ops),
|
||||
"jmp{q}\t{*}$dst # TAILCALL", []>;
|
||||
}
|
||||
|
@ -562,197 +562,18 @@ let neverHasSideEffects = 1, isNotDuplicable = 1, Uses = [ESP] in
|
||||
def MOVPC32r : Ii32<0xE8, Pseudo, (outs GR32:$reg), (ins i32imm:$label),
|
||||
"", []>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Control Flow Instructions.
|
||||
//
|
||||
|
||||
// Return instructions.
|
||||
let isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
hasCtrlDep = 1, FPForm = SpecialFP in {
|
||||
def RET : I <0xC3, RawFrm, (outs), (ins variable_ops),
|
||||
"ret",
|
||||
[(X86retflag 0)]>;
|
||||
def RETI : Ii16<0xC2, RawFrm, (outs), (ins i16imm:$amt, variable_ops),
|
||||
"ret\t$amt",
|
||||
[(X86retflag timm:$amt)]>;
|
||||
def LRET : I <0xCB, RawFrm, (outs), (ins),
|
||||
"lret", []>;
|
||||
def LRETI : Ii16<0xCA, RawFrm, (outs), (ins i16imm:$amt),
|
||||
"lret\t$amt", []>;
|
||||
}
|
||||
|
||||
// Unconditional branches.
|
||||
let isBarrier = 1, isBranch = 1, isTerminator = 1 in {
|
||||
def JMP_4 : Ii32PCRel<0xE9, RawFrm, (outs), (ins brtarget:$dst),
|
||||
"jmp\t$dst", [(br bb:$dst)]>;
|
||||
def JMP_1 : Ii8PCRel<0xEB, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jmp\t$dst", []>;
|
||||
}
|
||||
|
||||
// Conditional Branches.
|
||||
let isBranch = 1, isTerminator = 1, Uses = [EFLAGS] in {
|
||||
multiclass ICBr<bits<8> opc1, bits<8> opc4, string asm, PatFrag Cond> {
|
||||
def _1 : Ii8PCRel <opc1, RawFrm, (outs), (ins brtarget8:$dst), asm, []>;
|
||||
def _4 : Ii32PCRel<opc4, RawFrm, (outs), (ins brtarget:$dst), asm,
|
||||
[(X86brcond bb:$dst, Cond, EFLAGS)]>, TB;
|
||||
}
|
||||
}
|
||||
|
||||
defm JO : ICBr<0x70, 0x80, "jo\t$dst" , X86_COND_O>;
|
||||
defm JNO : ICBr<0x71, 0x81, "jno\t$dst" , X86_COND_NO>;
|
||||
defm JB : ICBr<0x72, 0x82, "jb\t$dst" , X86_COND_B>;
|
||||
defm JAE : ICBr<0x73, 0x83, "jae\t$dst", X86_COND_AE>;
|
||||
defm JE : ICBr<0x74, 0x84, "je\t$dst" , X86_COND_E>;
|
||||
defm JNE : ICBr<0x75, 0x85, "jne\t$dst", X86_COND_NE>;
|
||||
defm JBE : ICBr<0x76, 0x86, "jbe\t$dst", X86_COND_BE>;
|
||||
defm JA : ICBr<0x77, 0x87, "ja\t$dst" , X86_COND_A>;
|
||||
defm JS : ICBr<0x78, 0x88, "js\t$dst" , X86_COND_S>;
|
||||
defm JNS : ICBr<0x79, 0x89, "jns\t$dst", X86_COND_NS>;
|
||||
defm JP : ICBr<0x7A, 0x8A, "jp\t$dst" , X86_COND_P>;
|
||||
defm JNP : ICBr<0x7B, 0x8B, "jnp\t$dst", X86_COND_NP>;
|
||||
defm JL : ICBr<0x7C, 0x8C, "jl\t$dst" , X86_COND_L>;
|
||||
defm JGE : ICBr<0x7D, 0x8D, "jge\t$dst", X86_COND_GE>;
|
||||
defm JLE : ICBr<0x7E, 0x8E, "jle\t$dst", X86_COND_LE>;
|
||||
defm JG : ICBr<0x7F, 0x8F, "jg\t$dst" , X86_COND_G>;
|
||||
|
||||
// jcx/jecx/jrcx instructions.
|
||||
let isAsmParserOnly = 1, isBranch = 1, isTerminator = 1 in {
|
||||
// These are the 32-bit versions of this instruction for the asmparser. In
|
||||
// 32-bit mode, the address size prefix is jcxz and the unprefixed version is
|
||||
// jecxz.
|
||||
let Uses = [CX] in
|
||||
def JCXZ : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jcxz\t$dst", []>, AdSize, Requires<[In32BitMode]>;
|
||||
let Uses = [ECX] in
|
||||
def JECXZ_32 : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jecxz\t$dst", []>, Requires<[In32BitMode]>;
|
||||
|
||||
// J*CXZ instruction: 64-bit versions of this instruction for the asmparser.
|
||||
// In 64-bit mode, the address size prefix is jecxz and the unprefixed version
|
||||
// is jrcxz.
|
||||
let Uses = [ECX] in
|
||||
def JECXZ_64 : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jecxz\t$dst", []>, AdSize, Requires<[In64BitMode]>;
|
||||
let Uses = [RCX] in
|
||||
def JRCXZ : Ii8PCRel<0xE3, RawFrm, (outs), (ins brtarget8:$dst),
|
||||
"jrcxz\t$dst", []>, Requires<[In64BitMode]>;
|
||||
}
|
||||
|
||||
|
||||
// Indirect branches
|
||||
let isBranch = 1, isTerminator = 1, isBarrier = 1, isIndirectBranch = 1 in {
|
||||
def JMP32r : I<0xFF, MRM4r, (outs), (ins GR32:$dst), "jmp{l}\t{*}$dst",
|
||||
[(brind GR32:$dst)]>, Requires<[In32BitMode]>;
|
||||
def JMP32m : I<0xFF, MRM4m, (outs), (ins i32mem:$dst), "jmp{l}\t{*}$dst",
|
||||
[(brind (loadi32 addr:$dst))]>, Requires<[In32BitMode]>;
|
||||
|
||||
def FARJMP16i : Iseg16<0xEA, RawFrmImm16, (outs),
|
||||
(ins i16imm:$off, i16imm:$seg),
|
||||
"ljmp{w}\t{$seg, $off|$off, $seg}", []>, OpSize;
|
||||
def FARJMP32i : Iseg32<0xEA, RawFrmImm16, (outs),
|
||||
(ins i32imm:$off, i16imm:$seg),
|
||||
"ljmp{l}\t{$seg, $off|$off, $seg}", []>;
|
||||
|
||||
def FARJMP16m : I<0xFF, MRM5m, (outs), (ins opaque32mem:$dst),
|
||||
"ljmp{w}\t{*}$dst", []>, OpSize;
|
||||
def FARJMP32m : I<0xFF, MRM5m, (outs), (ins opaque48mem:$dst),
|
||||
"ljmp{l}\t{*}$dst", []>;
|
||||
}
|
||||
|
||||
|
||||
// Loop instructions
|
||||
|
||||
def LOOP : Ii8PCRel<0xE2, RawFrm, (outs), (ins brtarget8:$dst), "loop\t$dst", []>;
|
||||
def LOOPE : Ii8PCRel<0xE1, RawFrm, (outs), (ins brtarget8:$dst), "loope\t$dst", []>;
|
||||
def LOOPNE : Ii8PCRel<0xE0, RawFrm, (outs), (ins brtarget8:$dst), "loopne\t$dst", []>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Call Instructions...
|
||||
//
|
||||
let isCall = 1 in
|
||||
// All calls clobber the non-callee saved registers. ESP is marked as
|
||||
// a use to prevent stack-pointer assignments that appear immediately
|
||||
// before calls from potentially appearing dead. Uses for argument
|
||||
// registers are added manually.
|
||||
let Defs = [EAX, ECX, EDX, FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [ESP] in {
|
||||
def CALLpcrel32 : Ii32PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i32imm_pcrel:$dst,variable_ops),
|
||||
"call{l}\t$dst", []>, Requires<[In32BitMode]>;
|
||||
def CALL32r : I<0xFF, MRM2r, (outs), (ins GR32:$dst, variable_ops),
|
||||
"call{l}\t{*}$dst", [(X86call GR32:$dst)]>,
|
||||
Requires<[In32BitMode]>;
|
||||
def CALL32m : I<0xFF, MRM2m, (outs), (ins i32mem:$dst, variable_ops),
|
||||
"call{l}\t{*}$dst", [(X86call (loadi32 addr:$dst))]>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
def FARCALL16i : Iseg16<0x9A, RawFrmImm16, (outs),
|
||||
(ins i16imm:$off, i16imm:$seg),
|
||||
"lcall{w}\t{$seg, $off|$off, $seg}", []>, OpSize;
|
||||
def FARCALL32i : Iseg32<0x9A, RawFrmImm16, (outs),
|
||||
(ins i32imm:$off, i16imm:$seg),
|
||||
"lcall{l}\t{$seg, $off|$off, $seg}", []>;
|
||||
|
||||
def FARCALL16m : I<0xFF, MRM3m, (outs), (ins opaque32mem:$dst),
|
||||
"lcall{w}\t{*}$dst", []>, OpSize;
|
||||
def FARCALL32m : I<0xFF, MRM3m, (outs), (ins opaque48mem:$dst),
|
||||
"lcall{l}\t{*}$dst", []>;
|
||||
|
||||
// callw for 16 bit code for the assembler.
|
||||
let isAsmParserOnly = 1 in
|
||||
def CALLpcrel16 : Ii16PCRel<0xE8, RawFrm,
|
||||
(outs), (ins i16imm_pcrel:$dst, variable_ops),
|
||||
"callw\t$dst", []>, OpSize;
|
||||
}
|
||||
|
||||
// Constructing a stack frame.
|
||||
|
||||
def ENTER : Ii16<0xC8, RawFrmImm8, (outs), (ins i16imm:$len, i8imm:$lvl),
|
||||
"enter\t$len, $lvl", []>;
|
||||
|
||||
// Tail call stuff.
|
||||
|
||||
let isCall = 1, isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
isCodeGenOnly = 1 in
|
||||
let Defs = [EAX, ECX, EDX, FP0, FP1, FP2, FP3, FP4, FP5, FP6, ST0,
|
||||
MM0, MM1, MM2, MM3, MM4, MM5, MM6, MM7,
|
||||
XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
||||
XMM8, XMM9, XMM10, XMM11, XMM12, XMM13, XMM14, XMM15, EFLAGS],
|
||||
Uses = [ESP] in {
|
||||
def TCRETURNdi : I<0, Pseudo, (outs),
|
||||
(ins i32imm_pcrel:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
def TCRETURNri : I<0, Pseudo, (outs),
|
||||
(ins GR32_TC:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
let mayLoad = 1 in
|
||||
def TCRETURNmi : I<0, Pseudo, (outs),
|
||||
(ins i32mem_TC:$dst, i32imm:$offset, variable_ops),
|
||||
"#TC_RETURN $dst $offset", []>;
|
||||
|
||||
// FIXME: The should be pseudo instructions that are lowered when going to
|
||||
// mcinst.
|
||||
def TAILJMPd : Ii32PCRel<0xE9, RawFrm, (outs),
|
||||
(ins i32imm_pcrel:$dst, variable_ops),
|
||||
"jmp\t$dst # TAILCALL",
|
||||
[]>;
|
||||
def TAILJMPr : I<0xFF, MRM4r, (outs), (ins GR32_TC:$dst, variable_ops),
|
||||
"", []>; // FIXME: Remove encoding when JIT is dead.
|
||||
let mayLoad = 1 in
|
||||
def TAILJMPm : I<0xFF, MRM4m, (outs), (ins i32mem_TC:$dst, variable_ops),
|
||||
"jmp{l}\t{*}$dst # TAILCALL", []>;
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Miscellaneous Instructions...
|
||||
//
|
||||
let Defs = [EBP, ESP], Uses = [EBP, ESP], mayLoad = 1, neverHasSideEffects=1 in
|
||||
def LEAVE : I<0xC9, RawFrm,
|
||||
(outs), (ins), "leave", []>, Requires<[In32BitMode]>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Miscellaneous Instructions...
|
||||
//
|
||||
def POPCNT16rr : I<0xB8, MRMSrcReg, (outs GR16:$dst), (ins GR16:$src),
|
||||
"popcnt{w}\t{$src, $dst|$dst, $src}", []>, OpSize, XS;
|
||||
let mayLoad = 1 in
|
||||
@ -966,23 +787,6 @@ def MOV32ao32 : Ii32 <0xA3, RawFrm, (outs offset32:$dst), (ins),
|
||||
"mov{l}\t{%eax, $dst|$dst, %eax}", []>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// Moves to and from segment registers
|
||||
def MOV16rs : I<0x8C, MRMDestReg, (outs GR16:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32rs : I<0x8C, MRMDestReg, (outs GR32:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV16ms : I<0x8C, MRMDestMem, (outs i16mem:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32ms : I<0x8C, MRMDestMem, (outs i32mem:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV16sr : I<0x8E, MRMSrcReg, (outs SEGMENT_REG:$dst), (ins GR16:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32sr : I<0x8E, MRMSrcReg, (outs SEGMENT_REG:$dst), (ins GR32:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV16sm : I<0x8E, MRMSrcMem, (outs SEGMENT_REG:$dst), (ins i16mem:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32sm : I<0x8E, MRMSrcMem, (outs SEGMENT_REG:$dst), (ins i32mem:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
let isCodeGenOnly = 1 in {
|
||||
def MOV8rr_REV : I<0x8A, MRMSrcReg, (outs GR8:$dst), (ins GR8:$src),
|
||||
@ -1052,18 +856,6 @@ def MOV8rm_NOREX : I<0x8A, MRMSrcMem,
|
||||
"mov{b}\t{$src, $dst|$dst, $src} # NOREX", []>;
|
||||
}
|
||||
|
||||
// Moves to and from debug registers
|
||||
def MOV32rd : I<0x21, MRMDestReg, (outs GR32:$dst), (ins DEBUG_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV32dr : I<0x23, MRMSrcReg, (outs DEBUG_REG:$dst), (ins GR32:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
// Moves to and from control registers
|
||||
def MOV32rc : I<0x20, MRMDestReg, (outs GR32:$dst), (ins CONTROL_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV32cr : I<0x22, MRMSrcReg, (outs CONTROL_REG:$dst), (ins GR32:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Fixed-Register Multiplication and Division Instructions...
|
||||
//
|
||||
@ -3803,16 +3595,6 @@ def TLSCall_32 : I<0, Pseudo, (outs), (ins i32mem:$sym),
|
||||
[(X86TLSCall addr:$sym)]>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// EH Pseudo Instructions
|
||||
//
|
||||
let isTerminator = 1, isReturn = 1, isBarrier = 1,
|
||||
hasCtrlDep = 1, isCodeGenOnly = 1 in {
|
||||
def EH_RETURN : I<0xC3, RawFrm, (outs), (ins GR32:$addr),
|
||||
"ret\t#eh_return, addr: $addr",
|
||||
[(X86ehret GR32:$addr)]>;
|
||||
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Atomic support
|
||||
@ -4190,519 +3972,7 @@ def VMXOFF : I<0x01, MRM_C4, (outs), (ins), "vmxoff", []>, TB;
|
||||
def VMXON : I<0xC7, MRM6m, (outs), (ins i64mem:$vmxon),
|
||||
"vmxon\t{$vmxon}", []>, XS;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Non-Instruction Patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// ConstantPool GlobalAddress, ExternalSymbol, and JumpTable
|
||||
def : Pat<(i32 (X86Wrapper tconstpool :$dst)), (MOV32ri tconstpool :$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tjumptable :$dst)), (MOV32ri tjumptable :$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tglobaltlsaddr:$dst)),(MOV32ri tglobaltlsaddr:$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tglobaladdr :$dst)), (MOV32ri tglobaladdr :$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper texternalsym:$dst)), (MOV32ri texternalsym:$dst)>;
|
||||
def : Pat<(i32 (X86Wrapper tblockaddress:$dst)), (MOV32ri tblockaddress:$dst)>;
|
||||
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tconstpool:$src2)),
|
||||
(ADD32ri GR32:$src1, tconstpool:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tjumptable:$src2)),
|
||||
(ADD32ri GR32:$src1, tjumptable:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tglobaladdr :$src2)),
|
||||
(ADD32ri GR32:$src1, tglobaladdr:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper texternalsym:$src2)),
|
||||
(ADD32ri GR32:$src1, texternalsym:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (X86Wrapper tblockaddress:$src2)),
|
||||
(ADD32ri GR32:$src1, tblockaddress:$src2)>;
|
||||
|
||||
def : Pat<(store (i32 (X86Wrapper tglobaladdr:$src)), addr:$dst),
|
||||
(MOV32mi addr:$dst, tglobaladdr:$src)>;
|
||||
def : Pat<(store (i32 (X86Wrapper texternalsym:$src)), addr:$dst),
|
||||
(MOV32mi addr:$dst, texternalsym:$src)>;
|
||||
def : Pat<(store (i32 (X86Wrapper tblockaddress:$src)), addr:$dst),
|
||||
(MOV32mi addr:$dst, tblockaddress:$src)>;
|
||||
|
||||
// Calls
|
||||
// tailcall stuff
|
||||
def : Pat<(X86tcret GR32_TC:$dst, imm:$off),
|
||||
(TCRETURNri GR32_TC:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// FIXME: This is disabled for 32-bit PIC mode because the global base
|
||||
// register which is part of the address mode may be assigned a
|
||||
// callee-saved register.
|
||||
def : Pat<(X86tcret (load addr:$dst), imm:$off),
|
||||
(TCRETURNmi addr:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode, IsNotPIC]>;
|
||||
|
||||
def : Pat<(X86tcret (i32 tglobaladdr:$dst), imm:$off),
|
||||
(TCRETURNdi texternalsym:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
def : Pat<(X86tcret (i32 texternalsym:$dst), imm:$off),
|
||||
(TCRETURNdi texternalsym:$dst, imm:$off)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// Normal calls, with various flavors of addresses.
|
||||
def : Pat<(X86call (i32 tglobaladdr:$dst)),
|
||||
(CALLpcrel32 tglobaladdr:$dst)>;
|
||||
def : Pat<(X86call (i32 texternalsym:$dst)),
|
||||
(CALLpcrel32 texternalsym:$dst)>;
|
||||
def : Pat<(X86call (i32 imm:$dst)),
|
||||
(CALLpcrel32 imm:$dst)>, Requires<[CallImmAddr]>;
|
||||
|
||||
// X86 specific add which produces a flag.
|
||||
def : Pat<(addc GR32:$src1, GR32:$src2),
|
||||
(ADD32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(addc GR32:$src1, (load addr:$src2)),
|
||||
(ADD32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(addc GR32:$src1, imm:$src2),
|
||||
(ADD32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(addc GR32:$src1, i32immSExt8:$src2),
|
||||
(ADD32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
def : Pat<(subc GR32:$src1, GR32:$src2),
|
||||
(SUB32rr GR32:$src1, GR32:$src2)>;
|
||||
def : Pat<(subc GR32:$src1, (load addr:$src2)),
|
||||
(SUB32rm GR32:$src1, addr:$src2)>;
|
||||
def : Pat<(subc GR32:$src1, imm:$src2),
|
||||
(SUB32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(subc GR32:$src1, i32immSExt8:$src2),
|
||||
(SUB32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// Comparisons.
|
||||
|
||||
// TEST R,R is smaller than CMP R,0
|
||||
def : Pat<(X86cmp GR8:$src1, 0),
|
||||
(TEST8rr GR8:$src1, GR8:$src1)>;
|
||||
def : Pat<(X86cmp GR16:$src1, 0),
|
||||
(TEST16rr GR16:$src1, GR16:$src1)>;
|
||||
def : Pat<(X86cmp GR32:$src1, 0),
|
||||
(TEST32rr GR32:$src1, GR32:$src1)>;
|
||||
|
||||
// Conditional moves with folded loads with operands swapped and conditions
|
||||
// inverted.
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_B, EFLAGS),
|
||||
(CMOVAE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_B, EFLAGS),
|
||||
(CMOVAE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_AE, EFLAGS),
|
||||
(CMOVB16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_AE, EFLAGS),
|
||||
(CMOVB32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_E, EFLAGS),
|
||||
(CMOVNE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_E, EFLAGS),
|
||||
(CMOVNE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NE, EFLAGS),
|
||||
(CMOVE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NE, EFLAGS),
|
||||
(CMOVE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_BE, EFLAGS),
|
||||
(CMOVA16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_BE, EFLAGS),
|
||||
(CMOVA32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_A, EFLAGS),
|
||||
(CMOVBE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_A, EFLAGS),
|
||||
(CMOVBE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_L, EFLAGS),
|
||||
(CMOVGE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_L, EFLAGS),
|
||||
(CMOVGE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_GE, EFLAGS),
|
||||
(CMOVL16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_GE, EFLAGS),
|
||||
(CMOVL32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_LE, EFLAGS),
|
||||
(CMOVG16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_LE, EFLAGS),
|
||||
(CMOVG32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_G, EFLAGS),
|
||||
(CMOVLE16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_G, EFLAGS),
|
||||
(CMOVLE32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_P, EFLAGS),
|
||||
(CMOVNP16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_P, EFLAGS),
|
||||
(CMOVNP32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NP, EFLAGS),
|
||||
(CMOVP16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NP, EFLAGS),
|
||||
(CMOVP32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_S, EFLAGS),
|
||||
(CMOVNS16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_S, EFLAGS),
|
||||
(CMOVNS32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NS, EFLAGS),
|
||||
(CMOVS16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NS, EFLAGS),
|
||||
(CMOVS32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_O, EFLAGS),
|
||||
(CMOVNO16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_O, EFLAGS),
|
||||
(CMOVNO32rm GR32:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi16 addr:$src1), GR16:$src2, X86_COND_NO, EFLAGS),
|
||||
(CMOVO16rm GR16:$src2, addr:$src1)>;
|
||||
def : Pat<(X86cmov (loadi32 addr:$src1), GR32:$src2, X86_COND_NO, EFLAGS),
|
||||
(CMOVO32rm GR32:$src2, addr:$src1)>;
|
||||
|
||||
// zextload bool -> zextload byte
|
||||
def : Pat<(zextloadi8i1 addr:$src), (MOV8rm addr:$src)>;
|
||||
def : Pat<(zextloadi16i1 addr:$src), (MOVZX16rm8 addr:$src)>;
|
||||
def : Pat<(zextloadi32i1 addr:$src), (MOVZX32rm8 addr:$src)>;
|
||||
|
||||
// extload bool -> extload byte
|
||||
def : Pat<(extloadi8i1 addr:$src), (MOV8rm addr:$src)>;
|
||||
def : Pat<(extloadi16i1 addr:$src), (MOVZX16rm8 addr:$src)>;
|
||||
def : Pat<(extloadi32i1 addr:$src), (MOVZX32rm8 addr:$src)>;
|
||||
def : Pat<(extloadi16i8 addr:$src), (MOVZX16rm8 addr:$src)>;
|
||||
def : Pat<(extloadi32i8 addr:$src), (MOVZX32rm8 addr:$src)>;
|
||||
def : Pat<(extloadi32i16 addr:$src), (MOVZX32rm16 addr:$src)>;
|
||||
|
||||
// anyext. Define these to do an explicit zero-extend to
|
||||
// avoid partial-register updates.
|
||||
def : Pat<(i16 (anyext GR8 :$src)), (MOVZX16rr8 GR8 :$src)>;
|
||||
def : Pat<(i32 (anyext GR8 :$src)), (MOVZX32rr8 GR8 :$src)>;
|
||||
|
||||
// Except for i16 -> i32 since isel expect i16 ops to be promoted to i32.
|
||||
def : Pat<(i32 (anyext GR16:$src)),
|
||||
(INSERT_SUBREG (i32 (IMPLICIT_DEF)), GR16:$src, sub_16bit)>;
|
||||
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Some peepholes
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// Odd encoding trick: -128 fits into an 8-bit immediate field while
|
||||
// +128 doesn't, so in this special case use a sub instead of an add.
|
||||
def : Pat<(add GR16:$src1, 128),
|
||||
(SUB16ri8 GR16:$src1, -128)>;
|
||||
def : Pat<(store (add (loadi16 addr:$dst), 128), addr:$dst),
|
||||
(SUB16mi8 addr:$dst, -128)>;
|
||||
def : Pat<(add GR32:$src1, 128),
|
||||
(SUB32ri8 GR32:$src1, -128)>;
|
||||
def : Pat<(store (add (loadi32 addr:$dst), 128), addr:$dst),
|
||||
(SUB32mi8 addr:$dst, -128)>;
|
||||
|
||||
// r & (2^16-1) ==> movz
|
||||
def : Pat<(and GR32:$src1, 0xffff),
|
||||
(MOVZX32rr16 (EXTRACT_SUBREG GR32:$src1, sub_16bit))>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR32:$src1, 0xff),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src1,
|
||||
GR32_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
// r & (2^8-1) ==> movz
|
||||
def : Pat<(and GR16:$src1, 0xff),
|
||||
(MOVZX16rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src1,
|
||||
GR16_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// sext_inreg patterns
|
||||
def : Pat<(sext_inreg GR32:$src, i16),
|
||||
(MOVSX32rr16 (EXTRACT_SUBREG GR32:$src, sub_16bit))>;
|
||||
def : Pat<(sext_inreg GR32:$src, i8),
|
||||
(MOVSX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(sext_inreg GR16:$src, i8),
|
||||
(MOVSX16rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src,
|
||||
GR16_ABCD)),
|
||||
sub_8bit))>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// trunc patterns
|
||||
def : Pat<(i16 (trunc GR32:$src)),
|
||||
(EXTRACT_SUBREG GR32:$src, sub_16bit)>;
|
||||
def : Pat<(i8 (trunc GR32:$src)),
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i8 (trunc GR16:$src)),
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit)>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// h-register tricks
|
||||
def : Pat<(i8 (trunc (srl_su GR16:$src, (i8 8)))),
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i8 (trunc (srl_su GR32:$src, (i8 8)))),
|
||||
(EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src, GR32_ABCD)),
|
||||
sub_8bit_hi)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(srl GR16:$src, (i8 8)),
|
||||
(EXTRACT_SUBREG
|
||||
(MOVZX32rr8
|
||||
(EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src, GR16_ABCD)),
|
||||
sub_8bit_hi)),
|
||||
sub_16bit)>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i32 (zext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src,
|
||||
GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(i32 (anyext (srl_su GR16:$src, (i8 8)))),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i16 (COPY_TO_REGCLASS GR16:$src,
|
||||
GR16_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(and (srl_su GR32:$src, (i8 8)), (i32 255)),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
def : Pat<(srl (and_su GR32:$src, 0xff00), (i8 8)),
|
||||
(MOVZX32rr8 (EXTRACT_SUBREG (i32 (COPY_TO_REGCLASS GR32:$src,
|
||||
GR32_ABCD)),
|
||||
sub_8bit_hi))>,
|
||||
Requires<[In32BitMode]>;
|
||||
|
||||
// (shl x, 1) ==> (add x, x)
|
||||
def : Pat<(shl GR8 :$src1, (i8 1)), (ADD8rr GR8 :$src1, GR8 :$src1)>;
|
||||
def : Pat<(shl GR16:$src1, (i8 1)), (ADD16rr GR16:$src1, GR16:$src1)>;
|
||||
def : Pat<(shl GR32:$src1, (i8 1)), (ADD32rr GR32:$src1, GR32:$src1)>;
|
||||
|
||||
// (shl x (and y, 31)) ==> (shl x, y)
|
||||
def : Pat<(shl GR8:$src1, (and CL, 31)),
|
||||
(SHL8rCL GR8:$src1)>;
|
||||
def : Pat<(shl GR16:$src1, (and CL, 31)),
|
||||
(SHL16rCL GR16:$src1)>;
|
||||
def : Pat<(shl GR32:$src1, (and CL, 31)),
|
||||
(SHL32rCL GR32:$src1)>;
|
||||
def : Pat<(store (shl (loadi8 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHL8mCL addr:$dst)>;
|
||||
def : Pat<(store (shl (loadi16 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHL16mCL addr:$dst)>;
|
||||
def : Pat<(store (shl (loadi32 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHL32mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(srl GR8:$src1, (and CL, 31)),
|
||||
(SHR8rCL GR8:$src1)>;
|
||||
def : Pat<(srl GR16:$src1, (and CL, 31)),
|
||||
(SHR16rCL GR16:$src1)>;
|
||||
def : Pat<(srl GR32:$src1, (and CL, 31)),
|
||||
(SHR32rCL GR32:$src1)>;
|
||||
def : Pat<(store (srl (loadi8 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHR8mCL addr:$dst)>;
|
||||
def : Pat<(store (srl (loadi16 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHR16mCL addr:$dst)>;
|
||||
def : Pat<(store (srl (loadi32 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SHR32mCL addr:$dst)>;
|
||||
|
||||
def : Pat<(sra GR8:$src1, (and CL, 31)),
|
||||
(SAR8rCL GR8:$src1)>;
|
||||
def : Pat<(sra GR16:$src1, (and CL, 31)),
|
||||
(SAR16rCL GR16:$src1)>;
|
||||
def : Pat<(sra GR32:$src1, (and CL, 31)),
|
||||
(SAR32rCL GR32:$src1)>;
|
||||
def : Pat<(store (sra (loadi8 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SAR8mCL addr:$dst)>;
|
||||
def : Pat<(store (sra (loadi16 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SAR16mCL addr:$dst)>;
|
||||
def : Pat<(store (sra (loadi32 addr:$dst), (and CL, 31)), addr:$dst),
|
||||
(SAR32mCL addr:$dst)>;
|
||||
|
||||
// (anyext (setcc_carry)) -> (setcc_carry)
|
||||
def : Pat<(i16 (anyext (i8 (X86setcc_c X86_COND_B, EFLAGS)))),
|
||||
(SETB_C16r)>;
|
||||
def : Pat<(i32 (anyext (i8 (X86setcc_c X86_COND_B, EFLAGS)))),
|
||||
(SETB_C32r)>;
|
||||
def : Pat<(i32 (anyext (i16 (X86setcc_c X86_COND_B, EFLAGS)))),
|
||||
(SETB_C32r)>;
|
||||
|
||||
// (or x1, x2) -> (add x1, x2) if two operands are known not to share bits.
|
||||
let AddedComplexity = 5 in { // Try this before the selecting to OR
|
||||
def : Pat<(or_is_add GR16:$src1, imm:$src2),
|
||||
(ADD16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(or_is_add GR32:$src1, imm:$src2),
|
||||
(ADD32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(or_is_add GR16:$src1, i16immSExt8:$src2),
|
||||
(ADD16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(or_is_add GR32:$src1, i32immSExt8:$src2),
|
||||
(ADD32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
def : Pat<(or_is_add GR16:$src1, GR16:$src2),
|
||||
(ADD16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(or_is_add GR32:$src1, GR32:$src2),
|
||||
(ADD32rr GR32:$src1, GR32:$src2)>;
|
||||
} // AddedComplexity
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// EFLAGS-defining Patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// add reg, reg
|
||||
def : Pat<(add GR8 :$src1, GR8 :$src2), (ADD8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(add GR16:$src1, GR16:$src2), (ADD16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(add GR32:$src1, GR32:$src2), (ADD32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// add reg, mem
|
||||
def : Pat<(add GR8:$src1, (loadi8 addr:$src2)),
|
||||
(ADD8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(add GR16:$src1, (loadi16 addr:$src2)),
|
||||
(ADD16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(add GR32:$src1, (loadi32 addr:$src2)),
|
||||
(ADD32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// add reg, imm
|
||||
def : Pat<(add GR8 :$src1, imm:$src2), (ADD8ri GR8:$src1 , imm:$src2)>;
|
||||
def : Pat<(add GR16:$src1, imm:$src2), (ADD16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(add GR32:$src1, imm:$src2), (ADD32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(add GR16:$src1, i16immSExt8:$src2),
|
||||
(ADD16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(add GR32:$src1, i32immSExt8:$src2),
|
||||
(ADD32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// sub reg, reg
|
||||
def : Pat<(sub GR8 :$src1, GR8 :$src2), (SUB8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(sub GR16:$src1, GR16:$src2), (SUB16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, GR32:$src2), (SUB32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// sub reg, mem
|
||||
def : Pat<(sub GR8:$src1, (loadi8 addr:$src2)),
|
||||
(SUB8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(sub GR16:$src1, (loadi16 addr:$src2)),
|
||||
(SUB16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, (loadi32 addr:$src2)),
|
||||
(SUB32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// sub reg, imm
|
||||
def : Pat<(sub GR8:$src1, imm:$src2),
|
||||
(SUB8ri GR8:$src1, imm:$src2)>;
|
||||
def : Pat<(sub GR16:$src1, imm:$src2),
|
||||
(SUB16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, imm:$src2),
|
||||
(SUB32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(sub GR16:$src1, i16immSExt8:$src2),
|
||||
(SUB16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(sub GR32:$src1, i32immSExt8:$src2),
|
||||
(SUB32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// mul reg, reg
|
||||
def : Pat<(mul GR16:$src1, GR16:$src2),
|
||||
(IMUL16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, GR32:$src2),
|
||||
(IMUL32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// mul reg, mem
|
||||
def : Pat<(mul GR16:$src1, (loadi16 addr:$src2)),
|
||||
(IMUL16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, (loadi32 addr:$src2)),
|
||||
(IMUL32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// mul reg, imm
|
||||
def : Pat<(mul GR16:$src1, imm:$src2),
|
||||
(IMUL16rri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, imm:$src2),
|
||||
(IMUL32rri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(mul GR16:$src1, i16immSExt8:$src2),
|
||||
(IMUL16rri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(mul GR32:$src1, i32immSExt8:$src2),
|
||||
(IMUL32rri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// reg = mul mem, imm
|
||||
def : Pat<(mul (loadi16 addr:$src1), imm:$src2),
|
||||
(IMUL16rmi addr:$src1, imm:$src2)>;
|
||||
def : Pat<(mul (loadi32 addr:$src1), imm:$src2),
|
||||
(IMUL32rmi addr:$src1, imm:$src2)>;
|
||||
def : Pat<(mul (loadi16 addr:$src1), i16immSExt8:$src2),
|
||||
(IMUL16rmi8 addr:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(mul (loadi32 addr:$src1), i32immSExt8:$src2),
|
||||
(IMUL32rmi8 addr:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// Optimize multiply by 2 with EFLAGS result.
|
||||
let AddedComplexity = 2 in {
|
||||
def : Pat<(X86smul_flag GR16:$src1, 2), (ADD16rr GR16:$src1, GR16:$src1)>;
|
||||
def : Pat<(X86smul_flag GR32:$src1, 2), (ADD32rr GR32:$src1, GR32:$src1)>;
|
||||
}
|
||||
|
||||
// Patterns for nodes that do not produce flags, for instructions that do.
|
||||
|
||||
// Increment reg.
|
||||
def : Pat<(add GR8:$src1 , 1), (INC8r GR8:$src1)>;
|
||||
def : Pat<(add GR16:$src1, 1), (INC16r GR16:$src1)>, Requires<[In32BitMode]>;
|
||||
def : Pat<(add GR32:$src1, 1), (INC32r GR32:$src1)>, Requires<[In32BitMode]>;
|
||||
|
||||
// Decrement reg.
|
||||
def : Pat<(add GR8:$src1 , -1), (DEC8r GR8:$src1)>;
|
||||
def : Pat<(add GR16:$src1, -1), (DEC16r GR16:$src1)>, Requires<[In32BitMode]>;
|
||||
def : Pat<(add GR32:$src1, -1), (DEC32r GR32:$src1)>, Requires<[In32BitMode]>;
|
||||
|
||||
// or reg/reg.
|
||||
def : Pat<(or GR8 :$src1, GR8 :$src2), (OR8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(or GR16:$src1, GR16:$src2), (OR16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(or GR32:$src1, GR32:$src2), (OR32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// or reg/mem
|
||||
def : Pat<(or GR8:$src1, (loadi8 addr:$src2)),
|
||||
(OR8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(or GR16:$src1, (loadi16 addr:$src2)),
|
||||
(OR16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(or GR32:$src1, (loadi32 addr:$src2)),
|
||||
(OR32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// or reg/imm
|
||||
def : Pat<(or GR8:$src1 , imm:$src2), (OR8ri GR8 :$src1, imm:$src2)>;
|
||||
def : Pat<(or GR16:$src1, imm:$src2), (OR16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(or GR32:$src1, imm:$src2), (OR32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(or GR16:$src1, i16immSExt8:$src2),
|
||||
(OR16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(or GR32:$src1, i32immSExt8:$src2),
|
||||
(OR32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// xor reg/reg
|
||||
def : Pat<(xor GR8 :$src1, GR8 :$src2), (XOR8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(xor GR16:$src1, GR16:$src2), (XOR16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, GR32:$src2), (XOR32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// xor reg/mem
|
||||
def : Pat<(xor GR8:$src1, (loadi8 addr:$src2)),
|
||||
(XOR8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(xor GR16:$src1, (loadi16 addr:$src2)),
|
||||
(XOR16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, (loadi32 addr:$src2)),
|
||||
(XOR32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// xor reg/imm
|
||||
def : Pat<(xor GR8:$src1, imm:$src2),
|
||||
(XOR8ri GR8:$src1, imm:$src2)>;
|
||||
def : Pat<(xor GR16:$src1, imm:$src2),
|
||||
(XOR16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, imm:$src2),
|
||||
(XOR32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(xor GR16:$src1, i16immSExt8:$src2),
|
||||
(XOR16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(xor GR32:$src1, i32immSExt8:$src2),
|
||||
(XOR32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
// and reg/reg
|
||||
def : Pat<(and GR8 :$src1, GR8 :$src2), (AND8rr GR8 :$src1, GR8 :$src2)>;
|
||||
def : Pat<(and GR16:$src1, GR16:$src2), (AND16rr GR16:$src1, GR16:$src2)>;
|
||||
def : Pat<(and GR32:$src1, GR32:$src2), (AND32rr GR32:$src1, GR32:$src2)>;
|
||||
|
||||
// and reg/mem
|
||||
def : Pat<(and GR8:$src1, (loadi8 addr:$src2)),
|
||||
(AND8rm GR8:$src1, addr:$src2)>;
|
||||
def : Pat<(and GR16:$src1, (loadi16 addr:$src2)),
|
||||
(AND16rm GR16:$src1, addr:$src2)>;
|
||||
def : Pat<(and GR32:$src1, (loadi32 addr:$src2)),
|
||||
(AND32rm GR32:$src1, addr:$src2)>;
|
||||
|
||||
// and reg/imm
|
||||
def : Pat<(and GR8:$src1, imm:$src2),
|
||||
(AND8ri GR8:$src1, imm:$src2)>;
|
||||
def : Pat<(and GR16:$src1, imm:$src2),
|
||||
(AND16ri GR16:$src1, imm:$src2)>;
|
||||
def : Pat<(and GR32:$src1, imm:$src2),
|
||||
(AND32ri GR32:$src1, imm:$src2)>;
|
||||
def : Pat<(and GR16:$src1, i16immSExt8:$src2),
|
||||
(AND16ri8 GR16:$src1, i16immSExt8:$src2)>;
|
||||
def : Pat<(and GR32:$src1, i32immSExt8:$src2),
|
||||
(AND32ri8 GR32:$src1, i32immSExt8:$src2)>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Subsystems.
|
||||
@ -4714,6 +3984,8 @@ include "X86InstrFPStack.td"
|
||||
// X86-64 Support
|
||||
include "X86Instr64bit.td"
|
||||
|
||||
include "X86InstrControl.td"
|
||||
|
||||
// SIMD support (SSE, MMX and AVX)
|
||||
include "X86InstrFragmentsSIMD.td"
|
||||
|
||||
@ -4727,3 +3999,7 @@ include "X86Instr3DNow.td"
|
||||
|
||||
// System instructions.
|
||||
include "X86InstrSystem.td"
|
||||
|
||||
// Compiler Pseudo Instructions and Pat Patterns
|
||||
include "X86InstrCompiler.td"
|
||||
|
||||
|
@ -14,17 +14,15 @@
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
let Defs = [RAX, RDX] in
|
||||
def RDTSC : I<0x31, RawFrm, (outs), (ins), "rdtsc", [(X86rdtsc)]>,
|
||||
TB;
|
||||
def RDTSC : I<0x31, RawFrm, (outs), (ins), "rdtsc", [(X86rdtsc)]>, TB;
|
||||
|
||||
let Defs = [RAX, RCX, RDX] in
|
||||
def RDTSCP : I<0x01, MRM_F9, (outs), (ins), "rdtscp", []>, TB;
|
||||
def RDTSCP : I<0x01, MRM_F9, (outs), (ins), "rdtscp", []>, TB;
|
||||
|
||||
// CPU flow control instructions
|
||||
|
||||
let isTerminator = 1, isBarrier = 1, hasCtrlDep = 1 in {
|
||||
def TRAP : I<0x0B, RawFrm, (outs), (ins), "ud2", [(trap)]>, TB;
|
||||
}
|
||||
let isTerminator = 1, isBarrier = 1, hasCtrlDep = 1 in
|
||||
def TRAP : I<0x0B, RawFrm, (outs), (ins), "ud2", [(trap)]>, TB;
|
||||
|
||||
def HLT : I<0xF4, RawFrm, (outs), (ins), "hlt", []>;
|
||||
def RSM : I<0xAA, RawFrm, (outs), (ins), "rsm", []>, TB;
|
||||
@ -37,20 +35,17 @@ def INT3 : I<0xcc, RawFrm, (outs), (ins), "int3",
|
||||
def INT : Ii8<0xcd, RawFrm, (outs), (ins i8imm:$trap), "int\t$trap",
|
||||
[(int_x86_int imm:$trap)]>;
|
||||
|
||||
def SYSCALL : I<0x05, RawFrm,
|
||||
(outs), (ins), "syscall", []>, TB;
|
||||
def SYSRETL : I<0x07, RawFrm,
|
||||
(outs), (ins), "sysretl", []>, TB;
|
||||
def SYSRETQ : RI<0x07, RawFrm,
|
||||
(outs), (ins), "sysretq", []>, TB, Requires<[In64BitMode]>;
|
||||
def SYSCALL : I<0x05, RawFrm, (outs), (ins), "syscall", []>, TB;
|
||||
def SYSRETL : I<0x07, RawFrm, (outs), (ins), "sysretl", []>, TB;
|
||||
def SYSRETQ :RI<0x07, RawFrm, (outs), (ins), "sysretq", []>, TB,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def SYSENTER : I<0x34, RawFrm,
|
||||
(outs), (ins), "sysenter", []>, TB;
|
||||
def SYSEXIT : I<0x35, RawFrm,
|
||||
(outs), (ins), "sysexit", []>, TB, Requires<[In32BitMode]>;
|
||||
|
||||
def SYSEXIT64 : RI<0x35, RawFrm,
|
||||
(outs), (ins), "sysexit", []>, TB, Requires<[In64BitMode]>;
|
||||
def SYSENTER : I<0x34, RawFrm, (outs), (ins), "sysenter", []>, TB;
|
||||
|
||||
def SYSEXIT : I<0x35, RawFrm, (outs), (ins), "sysexit", []>, TB,
|
||||
Requires<[In32BitMode]>;
|
||||
def SYSEXIT64 :RI<0x35, RawFrm, (outs), (ins), "sysexit", []>, TB,
|
||||
Requires<[In64BitMode]>;
|
||||
|
||||
def IRET16 : I<0xcf, RawFrm, (outs), (ins), "iretw", []>, OpSize;
|
||||
def IRET32 : I<0xcf, RawFrm, (outs), (ins), "iret{l|d}", []>;
|
||||
@ -101,27 +96,84 @@ let Uses = [EAX] in
|
||||
def OUT32ir : Ii8<0xE7, RawFrm, (outs), (ins i8imm:$port),
|
||||
"out{l}\t{%eax, $port|$port, %EAX}", []>;
|
||||
|
||||
def IN8 : I<0x6C, RawFrm, (outs), (ins),
|
||||
"ins{b}", []>;
|
||||
def IN16 : I<0x6D, RawFrm, (outs), (ins),
|
||||
"ins{w}", []>, OpSize;
|
||||
def IN32 : I<0x6D, RawFrm, (outs), (ins),
|
||||
"ins{l}", []>;
|
||||
def IN8 : I<0x6C, RawFrm, (outs), (ins), "ins{b}", []>;
|
||||
def IN16 : I<0x6D, RawFrm, (outs), (ins), "ins{w}", []>, OpSize;
|
||||
def IN32 : I<0x6D, RawFrm, (outs), (ins), "ins{l}", []>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Moves to and from debug registers
|
||||
|
||||
def MOV32rd : I<0x21, MRMDestReg, (outs GR32:$dst), (ins DEBUG_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV64rd : I<0x21, MRMDestReg, (outs GR64:$dst), (ins DEBUG_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def MOV32dr : I<0x23, MRMSrcReg, (outs DEBUG_REG:$dst), (ins GR32:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV64dr : I<0x23, MRMSrcReg, (outs DEBUG_REG:$dst), (ins GR64:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Moves to and from control registers
|
||||
|
||||
def MOV32rc : I<0x20, MRMDestReg, (outs GR32:$dst), (ins CONTROL_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV64rc : I<0x20, MRMDestReg, (outs GR64:$dst), (ins CONTROL_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def MOV32cr : I<0x22, MRMSrcReg, (outs CONTROL_REG:$dst), (ins GR32:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def MOV64cr : I<0x22, MRMSrcReg, (outs CONTROL_REG:$dst), (ins GR64:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Segment override instruction prefixes
|
||||
|
||||
def CS_PREFIX : I<0x2E, RawFrm, (outs), (ins), "cs", []>;
|
||||
def SS_PREFIX : I<0x36, RawFrm, (outs), (ins), "ss", []>;
|
||||
def DS_PREFIX : I<0x3E, RawFrm, (outs), (ins), "ds", []>;
|
||||
def ES_PREFIX : I<0x26, RawFrm, (outs), (ins), "es", []>;
|
||||
def FS_PREFIX : I<0x64, RawFrm, (outs), (ins), "fs", []>;
|
||||
def GS_PREFIX : I<0x65, RawFrm, (outs), (ins), "gs", []>;
|
||||
def CS_PREFIX : I<0x2E, RawFrm, (outs), (ins), "cs", []>;
|
||||
def SS_PREFIX : I<0x36, RawFrm, (outs), (ins), "ss", []>;
|
||||
def DS_PREFIX : I<0x3E, RawFrm, (outs), (ins), "ds", []>;
|
||||
def ES_PREFIX : I<0x26, RawFrm, (outs), (ins), "es", []>;
|
||||
def FS_PREFIX : I<0x64, RawFrm, (outs), (ins), "fs", []>;
|
||||
def GS_PREFIX : I<0x65, RawFrm, (outs), (ins), "gs", []>;
|
||||
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Moves to and from segment registers.
|
||||
//
|
||||
|
||||
def MOV16rs : I<0x8C, MRMDestReg, (outs GR16:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32rs : I<0x8C, MRMDestReg, (outs GR32:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64rs : RI<0x8C, MRMDestReg, (outs GR64:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
def MOV16ms : I<0x8C, MRMDestMem, (outs i16mem:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32ms : I<0x8C, MRMDestMem, (outs i32mem:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64ms : RI<0x8C, MRMDestMem, (outs i64mem:$dst), (ins SEGMENT_REG:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
def MOV16sr : I<0x8E, MRMSrcReg, (outs SEGMENT_REG:$dst), (ins GR16:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32sr : I<0x8E, MRMSrcReg, (outs SEGMENT_REG:$dst), (ins GR32:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64sr : RI<0x8E, MRMSrcReg, (outs SEGMENT_REG:$dst), (ins GR64:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
def MOV16sm : I<0x8E, MRMSrcMem, (outs SEGMENT_REG:$dst), (ins i16mem:$src),
|
||||
"mov{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def MOV32sm : I<0x8E, MRMSrcMem, (outs SEGMENT_REG:$dst), (ins i32mem:$src),
|
||||
"mov{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
def MOV64sm : RI<0x8E, MRMSrcMem, (outs SEGMENT_REG:$dst), (ins i64mem:$src),
|
||||
"mov{q}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Segmentation support instructions.
|
||||
|
||||
def SWAPGS : I<0x01, MRM_F8, (outs), (ins), "swapgs", []>, TB;
|
||||
|
||||
def LAR16rm : I<0x02, MRMSrcMem, (outs GR16:$dst), (ins i16mem:$src),
|
||||
"lar{w}\t{$src, $dst|$dst, $src}", []>, TB, OpSize;
|
||||
def LAR16rr : I<0x02, MRMSrcReg, (outs GR16:$dst), (ins GR16:$src),
|
||||
@ -132,6 +184,11 @@ def LAR32rm : I<0x02, MRMSrcMem, (outs GR32:$dst), (ins i16mem:$src),
|
||||
"lar{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LAR32rr : I<0x02, MRMSrcReg, (outs GR32:$dst), (ins GR32:$src),
|
||||
"lar{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
// i16mem operand in LAR64rm and GR32 operand in LAR32rr is not a typo.
|
||||
def LAR64rm : RI<0x02, MRMSrcMem, (outs GR64:$dst), (ins i16mem:$src),
|
||||
"lar{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LAR64rr : RI<0x02, MRMSrcReg, (outs GR64:$dst), (ins GR32:$src),
|
||||
"lar{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def LSL16rm : I<0x03, MRMSrcMem, (outs GR16:$dst), (ins i16mem:$src),
|
||||
"lsl{w}\t{$src, $dst|$dst, $src}", []>, TB, OpSize;
|
||||
@ -141,7 +198,11 @@ def LSL32rm : I<0x03, MRMSrcMem, (outs GR32:$dst), (ins i32mem:$src),
|
||||
"lsl{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LSL32rr : I<0x03, MRMSrcReg, (outs GR32:$dst), (ins GR32:$src),
|
||||
"lsl{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def LSL64rm : RI<0x03, MRMSrcMem, (outs GR64:$dst), (ins i64mem:$src),
|
||||
"lsl{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LSL64rr : RI<0x03, MRMSrcReg, (outs GR64:$dst), (ins GR64:$src),
|
||||
"lsl{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def INVLPG : I<0x01, MRM7m, (outs), (ins i8mem:$addr), "invlpg\t$addr", []>, TB;
|
||||
|
||||
def STRr : I<0x00, MRM1r, (outs GR16:$dst), (ins),
|
||||
@ -189,24 +250,28 @@ def POPSS16 : I<0x17, RawFrm, (outs), (ins),
|
||||
"pop{w}\t%ss", []>, OpSize, Requires<[In32BitMode]>;
|
||||
def POPSS32 : I<0x17, RawFrm, (outs), (ins),
|
||||
"pop{l}\t%ss", []> , Requires<[In32BitMode]>;
|
||||
|
||||
def POPDS16 : I<0x1F, RawFrm, (outs), (ins),
|
||||
"pop{w}\t%ds", []>, OpSize, Requires<[In32BitMode]>;
|
||||
def POPDS32 : I<0x1F, RawFrm, (outs), (ins),
|
||||
"pop{l}\t%ds", []> , Requires<[In32BitMode]>;
|
||||
|
||||
def POPES16 : I<0x07, RawFrm, (outs), (ins),
|
||||
"pop{w}\t%es", []>, OpSize, Requires<[In32BitMode]>;
|
||||
def POPES32 : I<0x07, RawFrm, (outs), (ins),
|
||||
"pop{l}\t%es", []> , Requires<[In32BitMode]>;
|
||||
|
||||
def POPFS16 : I<0xa1, RawFrm, (outs), (ins),
|
||||
"pop{w}\t%fs", []>, OpSize, TB;
|
||||
def POPFS32 : I<0xa1, RawFrm, (outs), (ins),
|
||||
"pop{l}\t%fs", []>, TB , Requires<[In32BitMode]>;
|
||||
def POPFS64 : I<0xa1, RawFrm, (outs), (ins),
|
||||
"pop{q}\t%fs", []>, TB;
|
||||
|
||||
def POPGS16 : I<0xa9, RawFrm, (outs), (ins),
|
||||
"pop{w}\t%gs", []>, OpSize, TB;
|
||||
def POPGS32 : I<0xa9, RawFrm, (outs), (ins),
|
||||
"pop{l}\t%gs", []>, TB , Requires<[In32BitMode]>;
|
||||
def POPFS64 : I<0xa1, RawFrm, (outs), (ins),
|
||||
"pop{q}\t%fs", []>, TB;
|
||||
def POPGS64 : I<0xa9, RawFrm, (outs), (ins),
|
||||
"pop{q}\t%gs", []>, TB;
|
||||
|
||||
@ -215,26 +280,31 @@ def LDS16rm : I<0xc5, MRMSrcMem, (outs GR16:$dst), (ins opaque32mem:$src),
|
||||
"lds{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def LDS32rm : I<0xc5, MRMSrcMem, (outs GR32:$dst), (ins opaque48mem:$src),
|
||||
"lds{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
def LSS16rm : I<0xb2, MRMSrcMem, (outs GR16:$dst), (ins opaque32mem:$src),
|
||||
"lss{w}\t{$src, $dst|$dst, $src}", []>, TB, OpSize;
|
||||
def LSS32rm : I<0xb2, MRMSrcMem, (outs GR32:$dst), (ins opaque48mem:$src),
|
||||
"lss{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LSS64rm : RI<0xb2, MRMSrcMem, (outs GR64:$dst), (ins opaque80mem:$src),
|
||||
"lss{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def LES16rm : I<0xc4, MRMSrcMem, (outs GR16:$dst), (ins opaque32mem:$src),
|
||||
"les{w}\t{$src, $dst|$dst, $src}", []>, OpSize;
|
||||
def LES32rm : I<0xc4, MRMSrcMem, (outs GR32:$dst), (ins opaque48mem:$src),
|
||||
"les{l}\t{$src, $dst|$dst, $src}", []>;
|
||||
|
||||
def LFS16rm : I<0xb4, MRMSrcMem, (outs GR16:$dst), (ins opaque32mem:$src),
|
||||
"lfs{w}\t{$src, $dst|$dst, $src}", []>, TB, OpSize;
|
||||
def LFS32rm : I<0xb4, MRMSrcMem, (outs GR32:$dst), (ins opaque48mem:$src),
|
||||
"lfs{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LFS64rm : RI<0xb4, MRMSrcMem, (outs GR64:$dst), (ins opaque80mem:$src),
|
||||
"lfs{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def LGS16rm : I<0xb5, MRMSrcMem, (outs GR16:$dst), (ins opaque32mem:$src),
|
||||
"lgs{w}\t{$src, $dst|$dst, $src}", []>, TB, OpSize;
|
||||
def LGS32rm : I<0xb5, MRMSrcMem, (outs GR32:$dst), (ins opaque48mem:$src),
|
||||
"lgs{l}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LSS64rm : RI<0xb2, MRMSrcMem, (outs GR64:$dst), (ins opaque80mem:$src),
|
||||
"lss{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
def LFS64rm : RI<0xb4, MRMSrcMem, (outs GR64:$dst), (ins opaque80mem:$src),
|
||||
"lfs{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
def LGS64rm : RI<0xb5, MRMSrcMem, (outs GR64:$dst), (ins opaque80mem:$src),
|
||||
"lgs{q}\t{$src, $dst|$dst, $src}", []>, TB;
|
||||
|
||||
@ -280,7 +350,6 @@ def SMSW16r : I<0x01, MRM4r, (outs GR16:$dst), (ins),
|
||||
"smsw{w}\t$dst", []>, OpSize, TB;
|
||||
def SMSW32r : I<0x01, MRM4r, (outs GR32:$dst), (ins),
|
||||
"smsw{l}\t$dst", []>, TB;
|
||||
|
||||
// no m form encodable; use SMSW16m
|
||||
def SMSW64r : RI<0x01, MRM4r, (outs GR64:$dst), (ins),
|
||||
"smsw{q}\t$dst", []>, TB;
|
||||
|
Loading…
Reference in New Issue
Block a user