mirror of
https://github.com/RPCSX/llvm.git
synced 2025-03-01 09:26:22 +00:00
Merge current work back to tree to minimize diffs and drift. Major highlights
for CellSPU modifications: - SPUInstrInfo.td refactoring: "multiclass" really is _your_ friend. - Other improvements based on refactoring effort in SPUISelLowering.cpp, esp. in SPUISelLowering::PerformDAGCombine(), where zero amount shifts and rotates are now eliminiated, other scalar-to-vector-to-scalar silliness is also eliminated. - 64-bit operations are being implemented, _muldi3.c gcc runtime now compiles and generates the right code. More work still needs to be done. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@47532 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
f65a0f7860
commit
a59d469e9b
@ -20,6 +20,9 @@
|
||||
def cell_i7_ty: LLVMType<i8>;
|
||||
def cell_i8_ty: LLVMType<i8>;
|
||||
|
||||
// Keep this here until it's actually supported:
|
||||
def llvm_i128_ty : LLVMType<i128>;
|
||||
|
||||
class v16i8_u7imm<string builtin_suffix> :
|
||||
GCCBuiltin<!strconcat("__builtin_si_", builtin_suffix)>,
|
||||
Intrinsic<[llvm_v16i8_ty, llvm_v16i8_ty, cell_i7_ty],
|
||||
@ -158,9 +161,17 @@ let TargetPrefix = "spu" in {
|
||||
[IntrNoMem]>;
|
||||
|
||||
def int_spu_si_shli: v4i32_u7imm<"shli">;
|
||||
def int_spu_si_shlqbi: v16i8_rr<"shlqbi">;
|
||||
|
||||
def int_spu_si_shlqbi:
|
||||
GCCBuiltin<!strconcat("__builtin_si_", "shlqbi")>,
|
||||
Intrinsic<[llvm_v8i16_ty, llvm_v8i16_ty, llvm_i32_ty],
|
||||
[IntrNoMem]>;
|
||||
|
||||
def int_spu_si_shlqbii: v16i8_u7imm<"shlqbii">;
|
||||
def int_spu_si_shlqby: v16i8_rr<"shlqby">;
|
||||
def int_spu_si_shlqby:
|
||||
GCCBuiltin<!strconcat("__builtin_si_", "shlqby")>,
|
||||
Intrinsic<[llvm_v8i16_ty, llvm_v8i16_ty, llvm_i32_ty],
|
||||
[IntrNoMem]>;
|
||||
def int_spu_si_shlqbyi: v16i8_u7imm<"shlqbyi">;
|
||||
|
||||
def int_spu_si_ceq: v4i32_rr<"ceq">;
|
||||
|
@ -210,20 +210,20 @@ def CellSDKshli:
|
||||
(SHLIv4i32 VECREG:$rA, uimm7:$val)>;
|
||||
|
||||
def CellSDKshlqbi:
|
||||
Pat<(int_spu_si_shlqbi VECREG:$rA, VECREG:$rB),
|
||||
(SHLQBIvec VECREG:$rA, VECREG:$rB)>;
|
||||
Pat<(int_spu_si_shlqbi VECREG:$rA, R32C:$rB),
|
||||
(SHLQBIv16i8 VECREG:$rA, R32C:$rB)>;
|
||||
|
||||
def CellSDKshlqii:
|
||||
Pat<(int_spu_si_shlqbii VECREG:$rA, uimm7:$val),
|
||||
(SHLQBIIvec VECREG:$rA, uimm7:$val)>;
|
||||
(SHLQBIIv16i8 VECREG:$rA, uimm7:$val)>;
|
||||
|
||||
def CellSDKshlqby:
|
||||
Pat<(int_spu_si_shlqby VECREG:$rA, VECREG:$rB),
|
||||
(SHLQBYvec VECREG:$rA, VECREG:$rB)>;
|
||||
Pat<(int_spu_si_shlqby VECREG:$rA, R32C:$rB),
|
||||
(SHLQBYv16i8 VECREG:$rA, R32C:$rB)>;
|
||||
|
||||
def CellSDKshlqbyi:
|
||||
Pat<(int_spu_si_shlqbyi VECREG:$rA, uimm7:$val),
|
||||
(SHLQBYIvec VECREG:$rA, uimm7:$val)>;
|
||||
(SHLQBYIv16i8 VECREG:$rA, uimm7:$val)>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Branch/compare intrinsics:
|
||||
|
@ -166,24 +166,25 @@ namespace {
|
||||
struct valtype_map_s {
|
||||
MVT::ValueType VT;
|
||||
unsigned ldresult_ins; /// LDRESULT instruction (0 = undefined)
|
||||
bool ldresult_imm; /// LDRESULT instruction requires immediate?
|
||||
int prefslot_byte; /// Byte offset of the "preferred" slot
|
||||
};
|
||||
|
||||
const valtype_map_s valtype_map[] = {
|
||||
{ MVT::i1, 0, 3 },
|
||||
{ MVT::i8, SPU::ORBIr8, 3 },
|
||||
{ MVT::i16, SPU::ORHIr16, 2 },
|
||||
{ MVT::i32, SPU::ORIr32, 0 },
|
||||
{ MVT::i64, SPU::ORIr64, 0 },
|
||||
{ MVT::f32, 0, 0 },
|
||||
{ MVT::f64, 0, 0 },
|
||||
{ MVT::i1, 0, false, 3 },
|
||||
{ MVT::i8, SPU::ORBIr8, true, 3 },
|
||||
{ MVT::i16, SPU::ORHIr16, true, 2 },
|
||||
{ MVT::i32, SPU::ORIr32, true, 0 },
|
||||
{ MVT::i64, SPU::ORr64, false, 0 },
|
||||
{ MVT::f32, SPU::ORf32, false, 0 },
|
||||
{ MVT::f64, SPU::ORf64, false, 0 },
|
||||
// vector types... (sigh!)
|
||||
{ MVT::v16i8, 0, 0 },
|
||||
{ MVT::v8i16, 0, 0 },
|
||||
{ MVT::v4i32, 0, 0 },
|
||||
{ MVT::v2i64, 0, 0 },
|
||||
{ MVT::v4f32, 0, 0 },
|
||||
{ MVT::v2f64, 0, 0 }
|
||||
{ MVT::v16i8, 0, false, 0 },
|
||||
{ MVT::v8i16, 0, false, 0 },
|
||||
{ MVT::v4i32, 0, false, 0 },
|
||||
{ MVT::v2i64, 0, false, 0 },
|
||||
{ MVT::v4f32, 0, false, 0 },
|
||||
{ MVT::v2f64, 0, false, 0 }
|
||||
};
|
||||
|
||||
const size_t n_valtype_map = sizeof(valtype_map) / sizeof(valtype_map[0]);
|
||||
@ -603,7 +604,7 @@ SPUDAGToDAGISel::Select(SDOperand Op) {
|
||||
// to i8, then i8 to i16 in logical/branching operations.
|
||||
DEBUG(cerr << "CellSPU: Coalescing (zero_extend:i16 (and:i8 "
|
||||
"<arg>, <const>))\n");
|
||||
NewOpc = SPU::ANDHI1To2;
|
||||
NewOpc = SPU::ANDHIi8i16;
|
||||
Ops[0] = Op1.getOperand(0);
|
||||
Ops[1] = Op1.getOperand(1);
|
||||
n_ops = 2;
|
||||
@ -615,24 +616,23 @@ SPUDAGToDAGISel::Select(SDOperand Op) {
|
||||
SDOperand Arg = N->getOperand(0);
|
||||
SDOperand Chain = N->getOperand(1);
|
||||
SDNode *Result;
|
||||
const valtype_map_s *vtm = getValueTypeMapEntry(VT);
|
||||
|
||||
if (vtm->ldresult_ins == 0) {
|
||||
cerr << "LDRESULT for unsupported type: "
|
||||
<< MVT::getValueTypeString(VT)
|
||||
<< "\n";
|
||||
abort();
|
||||
}
|
||||
|
||||
AddToISelQueue(Arg);
|
||||
if (!MVT::isFloatingPoint(VT)) {
|
||||
Opc = vtm->ldresult_ins;
|
||||
if (vtm->ldresult_imm) {
|
||||
SDOperand Zero = CurDAG->getTargetConstant(0, VT);
|
||||
const valtype_map_s *vtm = getValueTypeMapEntry(VT);
|
||||
|
||||
if (vtm->ldresult_ins == 0) {
|
||||
cerr << "LDRESULT for unsupported type: "
|
||||
<< MVT::getValueTypeString(VT)
|
||||
<< "\n";
|
||||
abort();
|
||||
} else
|
||||
Opc = vtm->ldresult_ins;
|
||||
|
||||
AddToISelQueue(Zero);
|
||||
Result = CurDAG->getTargetNode(Opc, VT, MVT::Other, Arg, Zero, Chain);
|
||||
} else {
|
||||
Opc = (VT == MVT::f32 ? SPU::ORf32 : SPU::ORf64);
|
||||
Result = CurDAG->getTargetNode(Opc, MVT::Other, Arg, Arg, Chain);
|
||||
}
|
||||
|
||||
|
@ -216,6 +216,10 @@ SPUTargetLowering::SPUTargetLowering(SPUTargetMachine &TM)
|
||||
setOperationAction(ISD::SHL, MVT::i8, Custom);
|
||||
setOperationAction(ISD::SRL, MVT::i8, Custom);
|
||||
setOperationAction(ISD::SRA, MVT::i8, Custom);
|
||||
// And SPU needs custom lowering for shift left/right for i64
|
||||
setOperationAction(ISD::SHL, MVT::i64, Custom);
|
||||
setOperationAction(ISD::SRL, MVT::i64, Custom);
|
||||
setOperationAction(ISD::SRA, MVT::i64, Custom);
|
||||
|
||||
// Custom lower i32 multiplications
|
||||
setOperationAction(ISD::MUL, MVT::i32, Custom);
|
||||
@ -248,13 +252,11 @@ SPUTargetLowering::SPUTargetLowering(SPUTargetMachine &TM)
|
||||
setOperationAction(ISD::SELECT, MVT::f32, Expand);
|
||||
setOperationAction(ISD::SELECT, MVT::f64, Expand);
|
||||
|
||||
setOperationAction(ISD::SETCC, MVT::i1, Expand);
|
||||
setOperationAction(ISD::SETCC, MVT::i8, Expand);
|
||||
setOperationAction(ISD::SETCC, MVT::i16, Expand);
|
||||
setOperationAction(ISD::SETCC, MVT::i32, Expand);
|
||||
setOperationAction(ISD::SETCC, MVT::i64, Expand);
|
||||
setOperationAction(ISD::SETCC, MVT::f32, Expand);
|
||||
setOperationAction(ISD::SETCC, MVT::f64, Expand);
|
||||
// Zero extension and sign extension for i64 have to be
|
||||
// custom legalized
|
||||
setOperationAction(ISD::ZERO_EXTEND, MVT::i64, Custom);
|
||||
setOperationAction(ISD::SIGN_EXTEND, MVT::i64, Custom);
|
||||
setOperationAction(ISD::ANY_EXTEND, MVT::i64, Custom);
|
||||
|
||||
// SPU has a legal FP -> signed INT instruction
|
||||
setOperationAction(ISD::FP_TO_SINT, MVT::i32, Legal);
|
||||
@ -377,6 +379,9 @@ SPUTargetLowering::SPUTargetLowering(SPUTargetMachine &TM)
|
||||
|
||||
// We have target-specific dag combine patterns for the following nodes:
|
||||
setTargetDAGCombine(ISD::ADD);
|
||||
setTargetDAGCombine(ISD::ZERO_EXTEND);
|
||||
setTargetDAGCombine(ISD::SIGN_EXTEND);
|
||||
setTargetDAGCombine(ISD::ANY_EXTEND);
|
||||
|
||||
computeRegisterProperties();
|
||||
}
|
||||
@ -407,13 +412,17 @@ SPUTargetLowering::getTargetNodeName(unsigned Opcode) const
|
||||
node_names[(unsigned) SPUISD::MPYU] = "SPUISD::MPYU";
|
||||
node_names[(unsigned) SPUISD::MPYH] = "SPUISD::MPYH";
|
||||
node_names[(unsigned) SPUISD::MPYHH] = "SPUISD::MPYHH";
|
||||
node_names[(unsigned) SPUISD::SHLQUAD_L_BITS] = "SPUISD::SHLQUAD_L_BITS";
|
||||
node_names[(unsigned) SPUISD::SHLQUAD_L_BYTES] = "SPUISD::SHLQUAD_L_BYTES";
|
||||
node_names[(unsigned) SPUISD::VEC_SHL] = "SPUISD::VEC_SHL";
|
||||
node_names[(unsigned) SPUISD::VEC_SRL] = "SPUISD::VEC_SRL";
|
||||
node_names[(unsigned) SPUISD::VEC_SRA] = "SPUISD::VEC_SRA";
|
||||
node_names[(unsigned) SPUISD::VEC_ROTL] = "SPUISD::VEC_ROTL";
|
||||
node_names[(unsigned) SPUISD::VEC_ROTR] = "SPUISD::VEC_ROTR";
|
||||
node_names[(unsigned) SPUISD::ROTBYTES_RIGHT_Z] =
|
||||
"SPUISD::ROTBYTES_RIGHT_Z";
|
||||
node_names[(unsigned) SPUISD::ROTQUAD_RZ_BYTES] =
|
||||
"SPUISD::ROTQUAD_RZ_BYTES";
|
||||
node_names[(unsigned) SPUISD::ROTQUAD_RZ_BITS] =
|
||||
"SPUISD::ROTQUAD_RZ_BITS";
|
||||
node_names[(unsigned) SPUISD::ROTBYTES_RIGHT_S] =
|
||||
"SPUISD::ROTBYTES_RIGHT_S";
|
||||
node_names[(unsigned) SPUISD::ROTBYTES_LEFT] = "SPUISD::ROTBYTES_LEFT";
|
||||
@ -708,15 +717,7 @@ LowerSTORE(SDOperand Op, SelectionDAG &DAG, const SPUSubtarget *ST) {
|
||||
&& basePtr.getOperand(0).getOpcode() == SPUISD::IndirectAddr)) {
|
||||
insertEltPtr = basePtr;
|
||||
} else {
|
||||
#if 0
|
||||
// $sp is always aligned, so use it when necessary to avoid loading
|
||||
// an address
|
||||
SDOperand ptrP =
|
||||
basePtr.Val->hasOneUse() ? DAG.getRegister(SPU::R1, PtrVT) : basePtr;
|
||||
insertEltPtr = DAG.getNode(ISD::ADD, PtrVT, ptrP, insertEltOffs);
|
||||
#else
|
||||
insertEltPtr = DAG.getNode(ISD::ADD, PtrVT, basePtr, insertEltOffs);
|
||||
#endif
|
||||
}
|
||||
|
||||
insertEltOp = DAG.getNode(SPUISD::INSERT_MASK, stVecVT, insertEltPtr);
|
||||
@ -762,14 +763,9 @@ LowerConstantPool(SDOperand Op, SelectionDAG &DAG, const SPUSubtarget *ST) {
|
||||
// Just return the SDOperand with the constant pool address in it.
|
||||
return DAG.getNode(SPUISD::AFormAddr, PtrVT, CPI, Zero);
|
||||
} else {
|
||||
#if 1
|
||||
SDOperand Hi = DAG.getNode(SPUISD::Hi, PtrVT, CPI, Zero);
|
||||
SDOperand Lo = DAG.getNode(SPUISD::Lo, PtrVT, CPI, Zero);
|
||||
|
||||
return DAG.getNode(ISD::ADD, PtrVT, Lo, Hi);
|
||||
#else
|
||||
return DAG.getNode(SPUISD::IndirectAddr, PtrVT, CPI, Zero);
|
||||
#endif
|
||||
return DAG.getNode(SPUISD::IndirectAddr, PtrVT, Hi, Lo);
|
||||
}
|
||||
}
|
||||
|
||||
@ -787,10 +783,13 @@ LowerJumpTable(SDOperand Op, SelectionDAG &DAG, const SPUSubtarget *ST) {
|
||||
const TargetMachine &TM = DAG.getTarget();
|
||||
|
||||
if (TM.getRelocationModel() == Reloc::Static) {
|
||||
SDOperand JmpAForm = DAG.getNode(SPUISD::AFormAddr, PtrVT, JTI, Zero);
|
||||
return (!ST->usingLargeMem()
|
||||
? JmpAForm
|
||||
: DAG.getNode(SPUISD::IndirectAddr, PtrVT, JmpAForm, Zero));
|
||||
if (!ST->usingLargeMem()) {
|
||||
return DAG.getNode(SPUISD::AFormAddr, PtrVT, JTI, Zero);
|
||||
} else {
|
||||
SDOperand Hi = DAG.getNode(SPUISD::Hi, PtrVT, JTI, Zero);
|
||||
SDOperand Lo = DAG.getNode(SPUISD::Lo, PtrVT, JTI, Zero);
|
||||
return DAG.getNode(SPUISD::IndirectAddr, PtrVT, Hi, Lo);
|
||||
}
|
||||
}
|
||||
|
||||
assert(0 &&
|
||||
@ -1807,7 +1806,7 @@ static SDOperand LowerVECTOR_SHUFFLE(SDOperand Op, SelectionDAG &DAG) {
|
||||
else
|
||||
SrcElt = cast<ConstantSDNode>(PermMask.getOperand(i))->getValue();
|
||||
|
||||
for (unsigned j = 0; j != BytesPerElement; ++j) {
|
||||
for (unsigned j = 0; j < BytesPerElement; ++j) {
|
||||
ResultMask.push_back(DAG.getConstant(SrcElt*BytesPerElement+j,
|
||||
MVT::i8));
|
||||
}
|
||||
@ -1925,17 +1924,10 @@ static SDOperand LowerVectorMUL(SDOperand Op, SelectionDAG &DAG) {
|
||||
// is to break it all apart, sign extend, and reassemble the various
|
||||
// intermediate products.
|
||||
case MVT::v16i8: {
|
||||
MachineFunction &MF = DAG.getMachineFunction();
|
||||
MachineRegisterInfo &RegInfo = MF.getRegInfo();
|
||||
SDOperand Chain = Op.getOperand(0);
|
||||
SDOperand rA = Op.getOperand(0);
|
||||
SDOperand rB = Op.getOperand(1);
|
||||
SDOperand c8 = DAG.getConstant(8, MVT::i8);
|
||||
SDOperand c16 = DAG.getConstant(16, MVT::i8);
|
||||
|
||||
unsigned FSMBreg_2222 = RegInfo.createVirtualRegister(&SPU::VECREGRegClass);
|
||||
unsigned LoProd_reg = RegInfo.createVirtualRegister(&SPU::VECREGRegClass);
|
||||
unsigned HiProd_reg = RegInfo.createVirtualRegister(&SPU::VECREGRegClass);
|
||||
SDOperand c8 = DAG.getConstant(8, MVT::i32);
|
||||
SDOperand c16 = DAG.getConstant(16, MVT::i32);
|
||||
|
||||
SDOperand LLProd =
|
||||
DAG.getNode(SPUISD::MPY, MVT::v8i16,
|
||||
@ -1950,24 +1942,19 @@ static SDOperand LowerVectorMUL(SDOperand Op, SelectionDAG &DAG) {
|
||||
DAG.getNode(SPUISD::VEC_SHL, MVT::v8i16,
|
||||
DAG.getNode(SPUISD::MPY, MVT::v8i16, rALH, rBLH), c8);
|
||||
|
||||
SDOperand FSMBdef_2222 =
|
||||
DAG.getCopyToReg(Chain, FSMBreg_2222,
|
||||
DAG.getNode(SPUISD::FSMBI, MVT::v8i16,
|
||||
DAG.getConstant(0x2222, MVT::i32)));
|
||||
SDOperand FSMBmask = DAG.getNode(SPUISD::FSMBI, MVT::v8i16,
|
||||
DAG.getConstant(0x2222, MVT::i32));
|
||||
|
||||
SDOperand FSMBuse_2222 =
|
||||
DAG.getCopyFromReg(FSMBdef_2222, FSMBreg_2222, MVT::v4i32);
|
||||
|
||||
SDOperand LoProd_1 =
|
||||
DAG.getCopyToReg(Chain, LoProd_reg,
|
||||
DAG.getNode(SPUISD::SELB, MVT::v8i16, LLProd, LHProd,
|
||||
FSMBuse_2222));
|
||||
SDOperand LoProdParts =
|
||||
DAG.getNode(ISD::BIT_CONVERT, MVT::v4i32,
|
||||
DAG.getNode(SPUISD::SELB, MVT::v8i16,
|
||||
LLProd, LHProd, FSMBmask));
|
||||
|
||||
SDOperand LoProdMask = DAG.getConstant(0xffff, MVT::i32);
|
||||
|
||||
SDOperand LoProd =
|
||||
DAG.getNode(ISD::AND, MVT::v4i32,
|
||||
DAG.getCopyFromReg(LoProd_1, LoProd_reg, MVT::v4i32),
|
||||
LoProdParts,
|
||||
DAG.getNode(ISD::BUILD_VECTOR, MVT::v4i32,
|
||||
LoProdMask, LoProdMask,
|
||||
LoProdMask, LoProdMask));
|
||||
@ -1993,15 +1980,13 @@ static SDOperand LowerVectorMUL(SDOperand Op, SelectionDAG &DAG) {
|
||||
DAG.getNode(SPUISD::VEC_SRA, MVT::v4i32, rBH, c8)));
|
||||
|
||||
SDOperand HHProd =
|
||||
DAG.getCopyToReg(Chain, HiProd_reg,
|
||||
DAG.getNode(SPUISD::SELB, MVT::v8i16,
|
||||
HLProd,
|
||||
DAG.getNode(SPUISD::VEC_SHL, MVT::v8i16, HHProd_1, c8),
|
||||
FSMBuse_2222));
|
||||
DAG.getNode(SPUISD::SELB, MVT::v8i16,
|
||||
HLProd,
|
||||
DAG.getNode(SPUISD::VEC_SHL, MVT::v8i16, HHProd_1, c8),
|
||||
FSMBmask);
|
||||
|
||||
SDOperand HiProd =
|
||||
DAG.getNode(SPUISD::VEC_SHL, MVT::v4i32,
|
||||
DAG.getCopyFromReg(HHProd, HiProd_reg, MVT::v4i32), c16);
|
||||
DAG.getNode(SPUISD::VEC_SHL, MVT::v4i32, HHProd, c16);
|
||||
|
||||
return DAG.getNode(ISD::BIT_CONVERT, MVT::v16i8,
|
||||
DAG.getNode(ISD::OR, MVT::v4i32,
|
||||
@ -2168,7 +2153,8 @@ static SDOperand LowerINSERT_VECTOR_ELT(SDOperand Op, SelectionDAG &DAG) {
|
||||
return result;
|
||||
}
|
||||
|
||||
static SDOperand LowerI8Math(SDOperand Op, SelectionDAG &DAG, unsigned Opc) {
|
||||
static SDOperand LowerI8Math(SDOperand Op, SelectionDAG &DAG, unsigned Opc)
|
||||
{
|
||||
SDOperand N0 = Op.getOperand(0); // Everything has at least one operand
|
||||
|
||||
assert(Op.getValueType() == MVT::i8);
|
||||
@ -2254,6 +2240,87 @@ static SDOperand LowerI8Math(SDOperand Op, SelectionDAG &DAG, unsigned Opc) {
|
||||
return SDOperand();
|
||||
}
|
||||
|
||||
static SDOperand LowerI64Math(SDOperand Op, SelectionDAG &DAG, unsigned Opc)
|
||||
{
|
||||
MVT::ValueType VT = Op.getValueType();
|
||||
unsigned VecVT =
|
||||
MVT::getVectorType(VT, (128 / MVT::getSizeInBits(VT)));
|
||||
|
||||
SDOperand Op0 = Op.getOperand(0);
|
||||
|
||||
switch (Opc) {
|
||||
case ISD::ZERO_EXTEND:
|
||||
case ISD::SIGN_EXTEND:
|
||||
case ISD::ANY_EXTEND: {
|
||||
MVT::ValueType Op0VT = Op0.getValueType();
|
||||
unsigned Op0VecVT =
|
||||
MVT::getVectorType(Op0VT, (128 / MVT::getSizeInBits(Op0VT)));
|
||||
|
||||
assert(Op0VT == MVT::i32
|
||||
&& "CellSPU: Zero/sign extending something other than i32");
|
||||
|
||||
unsigned NewOpc = (Opc == ISD::SIGN_EXTEND
|
||||
? SPUISD::ROTBYTES_RIGHT_S
|
||||
: SPUISD::ROTQUAD_RZ_BYTES);
|
||||
SDOperand PromoteScalar =
|
||||
DAG.getNode(SPUISD::PROMOTE_SCALAR, Op0VecVT, Op0);
|
||||
|
||||
return DAG.getNode(SPUISD::EXTRACT_ELT0, VT,
|
||||
DAG.getNode(ISD::BIT_CONVERT, VecVT,
|
||||
DAG.getNode(NewOpc, Op0VecVT,
|
||||
PromoteScalar,
|
||||
DAG.getConstant(4, MVT::i32))));
|
||||
}
|
||||
|
||||
case ISD::SHL: {
|
||||
SDOperand ShiftAmt = Op.getOperand(1);
|
||||
unsigned ShiftAmtVT = unsigned(ShiftAmt.getValueType());
|
||||
SDOperand Op0Vec = DAG.getNode(SPUISD::PROMOTE_SCALAR, VecVT, Op0);
|
||||
SDOperand MaskLower =
|
||||
DAG.getNode(SPUISD::SELB, VecVT,
|
||||
Op0Vec,
|
||||
DAG.getConstant(0, VecVT),
|
||||
DAG.getNode(SPUISD::FSMBI, VecVT,
|
||||
DAG.getConstant(0xff00ULL, MVT::i16)));
|
||||
SDOperand ShiftAmtBytes =
|
||||
DAG.getNode(ISD::SRL, ShiftAmtVT,
|
||||
ShiftAmt,
|
||||
DAG.getConstant(3, ShiftAmtVT));
|
||||
SDOperand ShiftAmtBits =
|
||||
DAG.getNode(ISD::AND, ShiftAmtVT,
|
||||
ShiftAmt,
|
||||
DAG.getConstant(7, ShiftAmtVT));
|
||||
|
||||
return DAG.getNode(SPUISD::EXTRACT_ELT0, VT,
|
||||
DAG.getNode(SPUISD::SHLQUAD_L_BITS, VecVT,
|
||||
DAG.getNode(SPUISD::SHLQUAD_L_BYTES, VecVT,
|
||||
MaskLower, ShiftAmtBytes),
|
||||
ShiftAmtBits));
|
||||
}
|
||||
|
||||
case ISD::SRL: {
|
||||
unsigned VT = unsigned(Op.getValueType());
|
||||
SDOperand ShiftAmt = Op.getOperand(1);
|
||||
unsigned ShiftAmtVT = unsigned(ShiftAmt.getValueType());
|
||||
SDOperand ShiftAmtBytes =
|
||||
DAG.getNode(ISD::SRL, ShiftAmtVT,
|
||||
ShiftAmt,
|
||||
DAG.getConstant(3, ShiftAmtVT));
|
||||
SDOperand ShiftAmtBits =
|
||||
DAG.getNode(ISD::AND, ShiftAmtVT,
|
||||
ShiftAmt,
|
||||
DAG.getConstant(7, ShiftAmtVT));
|
||||
|
||||
return DAG.getNode(SPUISD::ROTQUAD_RZ_BITS, VT,
|
||||
DAG.getNode(SPUISD::ROTQUAD_RZ_BYTES, VT,
|
||||
Op0, ShiftAmtBytes),
|
||||
ShiftAmtBits);
|
||||
}
|
||||
}
|
||||
|
||||
return SDOperand();
|
||||
}
|
||||
|
||||
//! Lower byte immediate operations for v16i8 vectors:
|
||||
static SDOperand
|
||||
LowerByteImmed(SDOperand Op, SelectionDAG &DAG) {
|
||||
@ -2438,10 +2505,13 @@ static SDOperand LowerCTPOP(SDOperand Op, SelectionDAG &DAG) {
|
||||
SDOperand
|
||||
SPUTargetLowering::LowerOperation(SDOperand Op, SelectionDAG &DAG)
|
||||
{
|
||||
switch (Op.getOpcode()) {
|
||||
unsigned Opc = (unsigned) Op.getOpcode();
|
||||
unsigned VT = (unsigned) Op.getValueType();
|
||||
|
||||
switch (Opc) {
|
||||
default: {
|
||||
cerr << "SPUTargetLowering::LowerOperation(): need to lower this!\n";
|
||||
cerr << "Op.getOpcode() = " << Op.getOpcode() << "\n";
|
||||
cerr << "Op.getOpcode() = " << Opc << "\n";
|
||||
cerr << "*Op.Val:\n";
|
||||
Op.Val->dump();
|
||||
abort();
|
||||
@ -2471,14 +2541,22 @@ SPUTargetLowering::LowerOperation(SDOperand Op, SelectionDAG &DAG)
|
||||
case ISD::RET:
|
||||
return LowerRET(Op, DAG, getTargetMachine());
|
||||
|
||||
// i8 math ops:
|
||||
|
||||
// i8, i64 math ops:
|
||||
case ISD::ZERO_EXTEND:
|
||||
case ISD::SIGN_EXTEND:
|
||||
case ISD::ANY_EXTEND:
|
||||
case ISD::SUB:
|
||||
case ISD::ROTR:
|
||||
case ISD::ROTL:
|
||||
case ISD::SRL:
|
||||
case ISD::SHL:
|
||||
case ISD::SRA:
|
||||
return LowerI8Math(Op, DAG, Op.getOpcode());
|
||||
if (VT == MVT::i8)
|
||||
return LowerI8Math(Op, DAG, Opc);
|
||||
else if (VT == MVT::i64)
|
||||
return LowerI64Math(Op, DAG, Opc);
|
||||
break;
|
||||
|
||||
// Vector-related lowering.
|
||||
case ISD::BUILD_VECTOR:
|
||||
@ -2500,15 +2578,15 @@ SPUTargetLowering::LowerOperation(SDOperand Op, SelectionDAG &DAG)
|
||||
|
||||
// Vector and i8 multiply:
|
||||
case ISD::MUL:
|
||||
if (MVT::isVector(Op.getValueType()))
|
||||
if (MVT::isVector(VT))
|
||||
return LowerVectorMUL(Op, DAG);
|
||||
else if (Op.getValueType() == MVT::i8)
|
||||
return LowerI8Math(Op, DAG, Op.getOpcode());
|
||||
else if (VT == MVT::i8)
|
||||
return LowerI8Math(Op, DAG, Opc);
|
||||
else
|
||||
return LowerMUL(Op, DAG, Op.getValueType(), Op.getOpcode());
|
||||
return LowerMUL(Op, DAG, VT, Opc);
|
||||
|
||||
case ISD::FDIV:
|
||||
if (Op.getValueType() == MVT::f32 || Op.getValueType() == MVT::v4f32)
|
||||
if (VT == MVT::f32 || VT == MVT::v4f32)
|
||||
return LowerFDIVf32(Op, DAG);
|
||||
// else if (Op.getValueType() == MVT::f64)
|
||||
// return LowerFDIVf64(Op, DAG);
|
||||
@ -2534,29 +2612,12 @@ SPUTargetLowering::PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const
|
||||
#endif
|
||||
const SPUSubtarget *ST = SPUTM.getSubtargetImpl();
|
||||
SelectionDAG &DAG = DCI.DAG;
|
||||
SDOperand N0 = N->getOperand(0); // everything has at least one operand
|
||||
SDOperand Op0 = N->getOperand(0); // everything has at least one operand
|
||||
SDOperand Result; // Initially, NULL result
|
||||
|
||||
switch (N->getOpcode()) {
|
||||
default: break;
|
||||
case SPUISD::IndirectAddr: {
|
||||
if (!ST->usingLargeMem() && N0.getOpcode() == SPUISD::AFormAddr) {
|
||||
ConstantSDNode *CN = cast<ConstantSDNode>(N->getOperand(1));
|
||||
if (CN->getValue() == 0) {
|
||||
// (SPUindirect (SPUaform <addr>, 0), 0) ->
|
||||
// (SPUaform <addr>, 0)
|
||||
|
||||
DEBUG(cerr << "Replace: ");
|
||||
DEBUG(N->dump(&DAG));
|
||||
DEBUG(cerr << "\nWith: ");
|
||||
DEBUG(N0.Val->dump(&DAG));
|
||||
DEBUG(cerr << "\n");
|
||||
|
||||
return N0;
|
||||
}
|
||||
}
|
||||
}
|
||||
case ISD::ADD: {
|
||||
SDOperand Op0 = N->getOperand(0);
|
||||
SDOperand Op1 = N->getOperand(1);
|
||||
|
||||
if ((Op1.getOpcode() == ISD::Constant
|
||||
@ -2603,10 +2664,104 @@ SPUTargetLowering::PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const
|
||||
Op1.getOperand(0), combinedConst);
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
case ISD::SIGN_EXTEND:
|
||||
case ISD::ZERO_EXTEND:
|
||||
case ISD::ANY_EXTEND: {
|
||||
if (Op0.getOpcode() == SPUISD::EXTRACT_ELT0 &&
|
||||
N->getValueType(0) == Op0.getValueType()) {
|
||||
// (any_extend (SPUextract_elt0 <arg>)) ->
|
||||
// (SPUextract_elt0 <arg>)
|
||||
// Types must match, however...
|
||||
DEBUG(cerr << "Replace: ");
|
||||
DEBUG(N->dump(&DAG));
|
||||
DEBUG(cerr << "\nWith: ");
|
||||
DEBUG(Op0.Val->dump(&DAG));
|
||||
DEBUG(cerr << "\n");
|
||||
|
||||
return Op0;
|
||||
}
|
||||
break;
|
||||
}
|
||||
case SPUISD::IndirectAddr: {
|
||||
if (!ST->usingLargeMem() && Op0.getOpcode() == SPUISD::AFormAddr) {
|
||||
ConstantSDNode *CN = cast<ConstantSDNode>(N->getOperand(1));
|
||||
if (CN->getValue() == 0) {
|
||||
// (SPUindirect (SPUaform <addr>, 0), 0) ->
|
||||
// (SPUaform <addr>, 0)
|
||||
|
||||
DEBUG(cerr << "Replace: ");
|
||||
DEBUG(N->dump(&DAG));
|
||||
DEBUG(cerr << "\nWith: ");
|
||||
DEBUG(Op0.Val->dump(&DAG));
|
||||
DEBUG(cerr << "\n");
|
||||
|
||||
return Op0;
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
case SPUISD::SHLQUAD_L_BITS:
|
||||
case SPUISD::SHLQUAD_L_BYTES:
|
||||
case SPUISD::VEC_SHL:
|
||||
case SPUISD::VEC_SRL:
|
||||
case SPUISD::VEC_SRA:
|
||||
case SPUISD::ROTQUAD_RZ_BYTES:
|
||||
case SPUISD::ROTQUAD_RZ_BITS: {
|
||||
SDOperand Op1 = N->getOperand(1);
|
||||
|
||||
if (isa<ConstantSDNode>(Op1)) {
|
||||
// Kill degenerate vector shifts:
|
||||
ConstantSDNode *CN = cast<ConstantSDNode>(Op1);
|
||||
|
||||
if (CN->getValue() == 0) {
|
||||
Result = Op0;
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
case SPUISD::PROMOTE_SCALAR: {
|
||||
switch (Op0.getOpcode()) {
|
||||
default:
|
||||
break;
|
||||
case ISD::ANY_EXTEND:
|
||||
case ISD::ZERO_EXTEND:
|
||||
case ISD::SIGN_EXTEND: {
|
||||
// (SPUpromote_scalar (any|sign|zero_extend (SPUextract_elt0 <arg>))) ->
|
||||
// <arg>
|
||||
// but only if the SPUpromote_scalar and <arg> types match.
|
||||
SDOperand Op00 = Op0.getOperand(0);
|
||||
if (Op00.getOpcode() == SPUISD::EXTRACT_ELT0) {
|
||||
SDOperand Op000 = Op00.getOperand(0);
|
||||
if (Op000.getValueType() == N->getValueType(0)) {
|
||||
Result = Op000;
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
case SPUISD::EXTRACT_ELT0: {
|
||||
// (SPUpromote_scalar (SPUextract_elt0 <arg>)) ->
|
||||
// <arg>
|
||||
Result = Op0.getOperand(0);
|
||||
break;
|
||||
}
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
// Otherwise, return unchanged.
|
||||
return SDOperand();
|
||||
#if 0
|
||||
if (Result.Val) {
|
||||
DEBUG(cerr << "\nReplace.SPU: ");
|
||||
DEBUG(N->dump(&DAG));
|
||||
DEBUG(cerr << "\nWith: ");
|
||||
DEBUG(Result.Val->dump(&DAG));
|
||||
DEBUG(cerr << "\n");
|
||||
}
|
||||
#endif
|
||||
|
||||
return Result;
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
@ -2657,6 +2812,7 @@ SPUTargetLowering::getRegForInlineAsmConstraint(const std::string &Constraint,
|
||||
return TargetLowering::getRegForInlineAsmConstraint(Constraint, VT);
|
||||
}
|
||||
|
||||
//! Compute used/known bits for a SPU operand
|
||||
void
|
||||
SPUTargetLowering::computeMaskedBitsForTargetNode(const SDOperand Op,
|
||||
const APInt &Mask,
|
||||
@ -2664,7 +2820,66 @@ SPUTargetLowering::computeMaskedBitsForTargetNode(const SDOperand Op,
|
||||
APInt &KnownOne,
|
||||
const SelectionDAG &DAG,
|
||||
unsigned Depth ) const {
|
||||
KnownZero = KnownOne = APInt(Mask.getBitWidth(), 0);
|
||||
const uint64_t uint64_sizebits = sizeof(uint64_t) * 8;
|
||||
|
||||
switch (Op.getOpcode()) {
|
||||
default:
|
||||
// KnownZero = KnownOne = APInt(Mask.getBitWidth(), 0);
|
||||
break;
|
||||
|
||||
#if 0
|
||||
case CALL:
|
||||
case SHUFB:
|
||||
case INSERT_MASK:
|
||||
case CNTB:
|
||||
#endif
|
||||
|
||||
case SPUISD::PROMOTE_SCALAR: {
|
||||
SDOperand Op0 = Op.getOperand(0);
|
||||
uint64_t InMask = MVT::getIntVTBitMask(Op0.getValueType());
|
||||
KnownZero |= APInt(uint64_sizebits, ~InMask, false);
|
||||
KnownOne |= APInt(uint64_sizebits, InMask, false);
|
||||
break;
|
||||
}
|
||||
|
||||
case SPUISD::LDRESULT:
|
||||
case SPUISD::EXTRACT_ELT0:
|
||||
case SPUISD::EXTRACT_ELT0_CHAINED: {
|
||||
uint64_t InMask = MVT::getIntVTBitMask(Op.getValueType());
|
||||
KnownZero |= APInt(uint64_sizebits, ~InMask, false);
|
||||
KnownOne |= APInt(uint64_sizebits, InMask, false);
|
||||
break;
|
||||
}
|
||||
|
||||
#if 0
|
||||
case EXTRACT_I1_ZEXT:
|
||||
case EXTRACT_I1_SEXT:
|
||||
case EXTRACT_I8_ZEXT:
|
||||
case EXTRACT_I8_SEXT:
|
||||
case MPY:
|
||||
case MPYU:
|
||||
case MPYH:
|
||||
case MPYHH:
|
||||
case SHLQUAD_L_BITS:
|
||||
case SHLQUAD_L_BYTES:
|
||||
case VEC_SHL:
|
||||
case VEC_SRL:
|
||||
case VEC_SRA:
|
||||
case VEC_ROTL:
|
||||
case VEC_ROTR:
|
||||
case ROTQUAD_RZ_BYTES:
|
||||
case ROTQUAD_RZ_BITS:
|
||||
case ROTBYTES_RIGHT_S:
|
||||
case ROTBYTES_LEFT:
|
||||
case ROTBYTES_LEFT_CHAINED:
|
||||
case FSMBI:
|
||||
case SELB:
|
||||
case SFPConstant:
|
||||
case FPInterp:
|
||||
case FPRecipEst:
|
||||
case SEXT32TO64:
|
||||
#endif
|
||||
}
|
||||
}
|
||||
|
||||
// LowerAsmOperandForConstraint
|
||||
|
@ -50,12 +50,15 @@ namespace llvm {
|
||||
MPYU, ///< Multiply Unsigned
|
||||
MPYH, ///< Multiply High
|
||||
MPYHH, ///< Multiply High-High
|
||||
SHLQUAD_L_BITS, ///< Rotate quad left, by bits
|
||||
SHLQUAD_L_BYTES, ///< Rotate quad left, by bytes
|
||||
VEC_SHL, ///< Vector shift left
|
||||
VEC_SRL, ///< Vector shift right (logical)
|
||||
VEC_SRA, ///< Vector shift right (arithmetic)
|
||||
VEC_ROTL, ///< Vector rotate left
|
||||
VEC_ROTR, ///< Vector rotate right
|
||||
ROTBYTES_RIGHT_Z, ///< Vector rotate right, by bytes, zero fill
|
||||
ROTQUAD_RZ_BYTES, ///< Rotate quad right, by bytes, zero fill
|
||||
ROTQUAD_RZ_BITS, ///< Rotate quad right, by bits, zero fill
|
||||
ROTBYTES_RIGHT_S, ///< Vector rotate right, by bytes, sign fill
|
||||
ROTBYTES_LEFT, ///< Rotate bytes (loads -> ROTQBYI)
|
||||
ROTBYTES_LEFT_CHAINED, ///< Rotate bytes (loads -> ROTQBYI), with chain
|
||||
|
@ -14,7 +14,7 @@
|
||||
|
||||
// This was kiped from the PPC instruction formats (seemed like a good idea...)
|
||||
|
||||
class I<dag OOL, dag IOL, string asmstr, InstrItinClass itin>
|
||||
class SPUInstr<dag OOL, dag IOL, string asmstr, InstrItinClass itin>
|
||||
: Instruction {
|
||||
field bits<32> Inst;
|
||||
|
||||
@ -28,7 +28,7 @@ class I<dag OOL, dag IOL, string asmstr, InstrItinClass itin>
|
||||
// RR Format
|
||||
class RRForm<bits<11> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin> {
|
||||
: SPUInstr<OOL, IOL, asmstr, itin> {
|
||||
bits<7> RA;
|
||||
bits<7> RB;
|
||||
bits<7> RT;
|
||||
@ -70,7 +70,7 @@ let RT = 0 in {
|
||||
// RRR Format
|
||||
class RRRForm<bits<4> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
: SPUInstr<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<7> RA;
|
||||
bits<7> RB;
|
||||
@ -89,7 +89,7 @@ class RRRForm<bits<4> opcode, dag OOL, dag IOL, string asmstr,
|
||||
// RI7 Format
|
||||
class RI7Form<bits<11> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
: SPUInstr<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<7> i7;
|
||||
bits<7> RA;
|
||||
@ -106,7 +106,7 @@ class RI7Form<bits<11> opcode, dag OOL, dag IOL, string asmstr,
|
||||
// CVTIntFp Format
|
||||
class CVTIntFPForm<bits<10> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
: SPUInstr<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<7> RA;
|
||||
bits<7> RT;
|
||||
@ -149,7 +149,7 @@ let RA = 0 in {
|
||||
|
||||
// Branch indirect external data forms:
|
||||
class BISLEDForm<bits<2> DE_flag, string asmstr, list<dag> pattern>
|
||||
: I<(outs), (ins indcalltarget:$func), asmstr, BranchResolv>
|
||||
: SPUInstr<(outs), (ins indcalltarget:$func), asmstr, BranchResolv>
|
||||
{
|
||||
bits<7> Rcalldest;
|
||||
|
||||
@ -166,7 +166,7 @@ class BISLEDForm<bits<2> DE_flag, string asmstr, list<dag> pattern>
|
||||
// RI10 Format
|
||||
class RI10Form<bits<8> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
: SPUInstr<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<10> i10;
|
||||
bits<7> RA;
|
||||
@ -182,43 +182,27 @@ class RI10Form<bits<8> opcode, dag OOL, dag IOL, string asmstr,
|
||||
|
||||
// RI10 Format, where the constant is zero (or effectively ignored by the
|
||||
// SPU)
|
||||
class RI10Form_1<bits<8> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<7> RA;
|
||||
bits<7> RT;
|
||||
|
||||
let Pattern = pattern;
|
||||
|
||||
let Inst{0-7} = opcode;
|
||||
let Inst{8-17} = 0;
|
||||
let Inst{18-24} = RA;
|
||||
let Inst{25-31} = RT;
|
||||
let i10 = 0 in {
|
||||
class RI10Form_1<bits<8> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: RI10Form<opcode, OOL, IOL, asmstr, itin, pattern>
|
||||
{ }
|
||||
}
|
||||
|
||||
// RI10 Format, where RT is ignored.
|
||||
// This format is used primarily by the Halt If ... Immediate set of
|
||||
// instructions
|
||||
class RI10Form_2<bits<8> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<10> i10;
|
||||
bits<7> RA;
|
||||
|
||||
let Pattern = pattern;
|
||||
|
||||
let Inst{0-7} = opcode;
|
||||
let Inst{8-17} = i10;
|
||||
let Inst{18-24} = RA;
|
||||
let Inst{25-31} = 0;
|
||||
let RT = 0 in {
|
||||
class RI10Form_2<bits<8> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: RI10Form<opcode, OOL, IOL, asmstr, itin, pattern>
|
||||
{ }
|
||||
}
|
||||
|
||||
// RI16 Format
|
||||
class RI16Form<bits<9> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
: SPUInstr<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<16> i16;
|
||||
bits<7> RT;
|
||||
@ -254,7 +238,7 @@ let RT = 0 in {
|
||||
// RI18 Format
|
||||
class RI18Form<bits<7> opcode, dag OOL, dag IOL, string asmstr,
|
||||
InstrItinClass itin, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, itin>
|
||||
: SPUInstr<OOL, IOL, asmstr, itin>
|
||||
{
|
||||
bits<18> i18;
|
||||
bits<7> RT;
|
||||
@ -306,7 +290,7 @@ class RR_Int_v4i32<bits<11> opcode, string opc, InstrItinClass itin,
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
class Pseudo<dag OOL, dag IOL, string asmstr, list<dag> pattern>
|
||||
: I<OOL, IOL, asmstr, NoItinerary> {
|
||||
: SPUInstr<OOL, IOL, asmstr, NoItinerary> {
|
||||
let Pattern = pattern;
|
||||
let Inst{31-0} = 0;
|
||||
}
|
||||
|
@ -49,14 +49,13 @@ SPUInstrInfo::isMoveInstr(const MachineInstr& MI,
|
||||
break;
|
||||
case SPU::ORIv4i32:
|
||||
case SPU::ORIr32:
|
||||
case SPU::ORIr64:
|
||||
case SPU::ORHIv8i16:
|
||||
case SPU::ORHIr16:
|
||||
case SPU::ORHI1To2:
|
||||
case SPU::ORHIi8i16:
|
||||
case SPU::ORBIv16i8:
|
||||
case SPU::ORBIr8:
|
||||
case SPU::ORI2To4:
|
||||
case SPU::ORI1To4:
|
||||
case SPU::ORIi16i32:
|
||||
case SPU::ORIi8i32:
|
||||
case SPU::AHIvec:
|
||||
case SPU::AHIr16:
|
||||
case SPU::AIvec:
|
||||
@ -103,7 +102,6 @@ SPUInstrInfo::isMoveInstr(const MachineInstr& MI,
|
||||
case SPU::ORr64:
|
||||
case SPU::ORf32:
|
||||
case SPU::ORf64:
|
||||
case SPU::ORgprc:
|
||||
assert(MI.getNumOperands() == 3 &&
|
||||
MI.getOperand(0).isRegister() &&
|
||||
MI.getOperand(1).isRegister() &&
|
||||
@ -203,14 +201,15 @@ void SPUInstrInfo::copyRegToReg(MachineBasicBlock &MBB,
|
||||
BuildMI(MBB, MI, get(SPU::ORf32), DestReg).addReg(SrcReg)
|
||||
.addReg(SrcReg);
|
||||
} else if (DestRC == SPU::R64CRegisterClass) {
|
||||
BuildMI(MBB, MI, get(SPU::ORIr64), DestReg).addReg(SrcReg).addImm(0);
|
||||
BuildMI(MBB, MI, get(SPU::ORr64), DestReg).addReg(SrcReg)
|
||||
.addReg(SrcReg);
|
||||
} else if (DestRC == SPU::R64FPRegisterClass) {
|
||||
BuildMI(MBB, MI, get(SPU::ORf64), DestReg).addReg(SrcReg)
|
||||
.addReg(SrcReg);
|
||||
} else if (DestRC == SPU::GPRCRegisterClass) {
|
||||
} /* else if (DestRC == SPU::GPRCRegisterClass) {
|
||||
BuildMI(MBB, MI, get(SPU::ORgprc), DestReg).addReg(SrcReg)
|
||||
.addReg(SrcReg);
|
||||
} else if (DestRC == SPU::VECREGRegisterClass) {
|
||||
} */ else if (DestRC == SPU::VECREGRegisterClass) {
|
||||
BuildMI(MBB, MI, get(SPU::ORv4i32), DestReg).addReg(SrcReg)
|
||||
.addReg(SrcReg);
|
||||
} else {
|
||||
|
File diff suppressed because it is too large
Load Diff
@ -32,7 +32,7 @@ def SPUcall : SDNode<"SPUISD::CALL", SDT_SPUCall,
|
||||
|
||||
// Operand type constraints for vector shuffle/permute operations
|
||||
def SDT_SPUshuffle : SDTypeProfile<1, 3, [
|
||||
SDTCisVT<3, v16i8>, SDTCisSameAs<0, 1>, SDTCisSameAs<0, 2>
|
||||
SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>
|
||||
]>;
|
||||
|
||||
// Unary, binary v16i8 operator type constraints:
|
||||
@ -62,27 +62,12 @@ def SPUfsmbi_type: SDTypeProfile<1, 1, [
|
||||
SDTCisVT<1, i32>]>;
|
||||
|
||||
// SELB type constraints:
|
||||
def SPUselb_type_v16i8: SDTypeProfile<1, 3, [
|
||||
SDTCisVT<0, v16i8>, SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>,
|
||||
SDTCisSameAs<0, 3> ]>;
|
||||
|
||||
def SPUselb_type_v8i16: SDTypeProfile<1, 3, [
|
||||
SDTCisVT<0, v8i16>, SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>,
|
||||
SDTCisSameAs<0, 3> ]>;
|
||||
|
||||
def SPUselb_type_v4i32: SDTypeProfile<1, 3, [
|
||||
SDTCisVT<0, v4i32>, SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>,
|
||||
SDTCisSameAs<0, 3> ]>;
|
||||
def SPUselb_type: SDTypeProfile<1, 3, [
|
||||
SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>, SDTCisSameAs<0, 3> ]>;
|
||||
|
||||
// SPU Vector shift pseudo-instruction type constraints
|
||||
def SPUvecshift_type_v16i8: SDTypeProfile<1, 2, [
|
||||
SDTCisVT<0, v16i8>, SDTCisSameAs<0, 1>, SDTCisInt<2>]>;
|
||||
|
||||
def SPUvecshift_type_v8i16: SDTypeProfile<1, 2, [
|
||||
SDTCisVT<0, v8i16>, SDTCisSameAs<0, 1>, SDTCisInt<2>]>;
|
||||
|
||||
def SPUvecshift_type_v4i32: SDTypeProfile<1, 2, [
|
||||
SDTCisVT<0, v4i32>, SDTCisSameAs<0, 1>, SDTCisInt<2>]>;
|
||||
def SPUvecshift_type: SDTypeProfile<1, 2, [
|
||||
SDTCisSameAs<0, 1>, SDTCisInt<2>]>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Synthetic/pseudo-instructions
|
||||
@ -116,41 +101,37 @@ def SPUmpyh_i32: SDNode<"SPUISD::MPYH", SDTIntBinOp, []>;
|
||||
// Used to compute intermediate products for 16-bit multiplies
|
||||
def SPUmpyhh_v8i16: SDNode<"SPUISD::MPYHH", SPUv8i16_binop, []>;
|
||||
|
||||
// Shift left quadword by bits and bytes
|
||||
def SPUshlquad_l_bits: SDNode<"SPUISD::SHLQUAD_L_BITS", SPUvecshift_type, []>;
|
||||
def SPUshlquad_l_bytes: SDNode<"SPUISD::SHLQUAD_L_BYTES", SPUvecshift_type, []>;
|
||||
|
||||
// Vector shifts (ISD::SHL,SRL,SRA are for _integers_ only):
|
||||
def SPUvec_shl_v8i16: SDNode<"SPUISD::VEC_SHL", SPUvecshift_type_v8i16, []>;
|
||||
def SPUvec_srl_v8i16: SDNode<"SPUISD::VEC_SRL", SPUvecshift_type_v8i16, []>;
|
||||
def SPUvec_sra_v8i16: SDNode<"SPUISD::VEC_SRA", SPUvecshift_type_v8i16, []>;
|
||||
def SPUvec_shl: SDNode<"SPUISD::VEC_SHL", SPUvecshift_type, []>;
|
||||
def SPUvec_srl: SDNode<"SPUISD::VEC_SRL", SPUvecshift_type, []>;
|
||||
def SPUvec_sra: SDNode<"SPUISD::VEC_SRA", SPUvecshift_type, []>;
|
||||
|
||||
def SPUvec_shl_v4i32: SDNode<"SPUISD::VEC_SHL", SPUvecshift_type_v4i32, []>;
|
||||
def SPUvec_srl_v4i32: SDNode<"SPUISD::VEC_SRL", SPUvecshift_type_v4i32, []>;
|
||||
def SPUvec_sra_v4i32: SDNode<"SPUISD::VEC_SRA", SPUvecshift_type_v4i32, []>;
|
||||
def SPUvec_rotl: SDNode<"SPUISD::VEC_ROTL", SPUvecshift_type, []>;
|
||||
def SPUvec_rotr: SDNode<"SPUISD::VEC_ROTR", SPUvecshift_type, []>;
|
||||
|
||||
def SPUvec_rotl_v8i16: SDNode<"SPUISD::VEC_ROTL", SPUvecshift_type_v8i16, []>;
|
||||
def SPUvec_rotl_v4i32: SDNode<"SPUISD::VEC_ROTL", SPUvecshift_type_v4i32, []>;
|
||||
def SPUrotquad_rz_bytes: SDNode<"SPUISD::ROTQUAD_RZ_BYTES",
|
||||
SPUvecshift_type, []>;
|
||||
def SPUrotquad_rz_bits: SDNode<"SPUISD::ROTQUAD_RZ_BITS",
|
||||
SPUvecshift_type, []>;
|
||||
|
||||
def SPUvec_rotr_v8i16: SDNode<"SPUISD::VEC_ROTR", SPUvecshift_type_v8i16, []>;
|
||||
def SPUvec_rotr_v4i32: SDNode<"SPUISD::VEC_ROTR", SPUvecshift_type_v4i32, []>;
|
||||
|
||||
def SPUrotbytes_right_zfill: SDNode<"SPUISD::ROTBYTES_RIGHT_Z",
|
||||
SPUvecshift_type_v16i8, []>;
|
||||
def SPUrotbytes_right_sfill: SDNode<"SPUISD::ROTBYTES_RIGHT_S",
|
||||
SPUvecshift_type_v16i8, []>;
|
||||
SPUvecshift_type, []>;
|
||||
|
||||
def SPUrotbytes_left: SDNode<"SPUISD::ROTBYTES_LEFT",
|
||||
SPUvecshift_type_v16i8, []>;
|
||||
SPUvecshift_type, []>;
|
||||
|
||||
def SPUrotbytes_left_chained : SDNode<"SPUISD::ROTBYTES_LEFT_CHAINED",
|
||||
SPUvecshift_type_v16i8, [SDNPHasChain]>;
|
||||
SPUvecshift_type, [SDNPHasChain]>;
|
||||
|
||||
// SPU form select mask for bytes, immediate
|
||||
def SPUfsmbi: SDNode<"SPUISD::FSMBI", SPUfsmbi_type, []>;
|
||||
|
||||
// SPU select bits instruction
|
||||
def SPUselb_v16i8: SDNode<"SPUISD::SELB", SPUselb_type_v16i8, []>;
|
||||
def SPUselb_v8i16: SDNode<"SPUISD::SELB", SPUselb_type_v8i16, []>;
|
||||
def SPUselb_v4i32: SDNode<"SPUISD::SELB", SPUselb_type_v4i32, []>;
|
||||
|
||||
// SPU single precision floating point constant load
|
||||
def SPUFPconstant: SDNode<"SPUISD::SFPConstant", SDTFPUnaryOp, []>;
|
||||
def SPUselb: SDNode<"SPUISD::SELB", SPUselb_type, []>;
|
||||
|
||||
// SPU floating point interpolate
|
||||
def SPUinterpolate : SDNode<"SPUISD::FPInterp", SDTFPBinOp, []>;
|
||||
@ -158,8 +139,8 @@ def SPUinterpolate : SDNode<"SPUISD::FPInterp", SDTFPBinOp, []>;
|
||||
// SPU floating point reciprocal estimate (used for fdiv)
|
||||
def SPUreciprocalEst: SDNode<"SPUISD::FPRecipEst", SDTFPUnaryOp, []>;
|
||||
|
||||
def SDT_vec_promote : SDTypeProfile<1, 1, []>;
|
||||
def SPUpromote_scalar: SDNode<"SPUISD::PROMOTE_SCALAR", SDT_vec_promote, []>;
|
||||
def SDTpromote_scalar: SDTypeProfile<1, 1, []>;
|
||||
def SPUpromote_scalar: SDNode<"SPUISD::PROMOTE_SCALAR", SDTpromote_scalar, []>;
|
||||
|
||||
def SPU_vec_demote : SDTypeProfile<1, 1, []>;
|
||||
def SPUextract_elt0: SDNode<"SPUISD::EXTRACT_ELT0", SPU_vec_demote, []>;
|
||||
|
@ -63,7 +63,7 @@ def HI16_vec : SDNodeXForm<scalar_to_vector, [{
|
||||
// simm7 predicate - True if the immediate fits in an 7-bit signed
|
||||
// field.
|
||||
def simm7: PatLeaf<(imm), [{
|
||||
int sextVal = ((((int) N->getValue()) << 25) >> 25);
|
||||
int sextVal = int(N->getSignExtended());
|
||||
return (sextVal >= -64 && sextVal <= 63);
|
||||
}]>;
|
||||
|
||||
@ -162,6 +162,13 @@ def hi16 : PatLeaf<(imm), [{
|
||||
return false;
|
||||
}], HI16>;
|
||||
|
||||
def bitshift : PatLeaf<(imm), [{
|
||||
// bitshift predicate - returns true if 0 < imm <= 7 for SHLQBII
|
||||
// (shift left quadword by bits immediate)
|
||||
int64_t Val = N->getValue();
|
||||
return (Val > 0 && Val <= 7);
|
||||
}]>;
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Floating point operands:
|
||||
//===----------------------------------------------------------------------===//
|
||||
@ -447,6 +454,10 @@ def s10imm : Operand<i16> {
|
||||
let PrintMethod = "printS10ImmOperand";
|
||||
}
|
||||
|
||||
def s10imm_i8: Operand<i8> {
|
||||
let PrintMethod = "printS10ImmOperand";
|
||||
}
|
||||
|
||||
def s10imm_i32: Operand<i32> {
|
||||
let PrintMethod = "printS10ImmOperand";
|
||||
}
|
||||
|
@ -8,11 +8,10 @@
|
||||
; RUN: grep and %t1.s | count 2
|
||||
; RUN: grep selb %t1.s | count 6
|
||||
; RUN: grep fsmbi %t1.s | count 4
|
||||
; RUN: grep shli %t1.s | count 4
|
||||
; RUN: grep shli %t1.s | count 2
|
||||
; RUN: grep shlhi %t1.s | count 4
|
||||
; RUN: grep ila %t1.s | count 2
|
||||
; RUN: grep xsbh %t1.s | count 8
|
||||
; RUN: grep xshw %t1.s | count 4
|
||||
; RUN: grep xsbh %t1.s | count 4
|
||||
target datalayout = "E-p:32:32:128-f64:64:128-f32:32:128-i64:32:128-i32:32:128-i16:16:128-i8:8:128-i1:8:128-a0:0:128-v128:128:128-s0:128:128"
|
||||
target triple = "spu"
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user