mirror of
https://github.com/RPCSX/llvm.git
synced 2024-11-27 13:40:30 +00:00
e3e43d9d57
I did this a long time ago with a janky python script, but now clang-format has built-in support for this. I fed clang-format every line with a #include and let it re-sort things according to the precise LLVM rules for include ordering baked into clang-format these days. I've reverted a number of files where the results of sorting includes isn't healthy. Either places where we have legacy code relying on particular include ordering (where possible, I'll fix these separately) or where we have particular formatting around #include lines that I didn't want to disturb in this patch. This patch is *entirely* mechanical. If you get merge conflicts or anything, just ignore the changes in this patch and run clang-format over your #include lines in the files. Sorry for any noise here, but it is important to keep these things stable. I was seeing an increasing number of patches with irrelevant re-ordering of #include lines because clang-format was used. This patch at least isolates that churn, makes it easy to skip when resolving conflicts, and gets us to a clean baseline (again). git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@304787 91177308-0d34-0410-b5e6-96231b3b80d8
162 lines
5.2 KiB
C++
162 lines
5.2 KiB
C++
//===-- SILowerI1Copies.cpp - Lower I1 Copies -----------------------------===//
|
|
//
|
|
// The LLVM Compiler Infrastructure
|
|
//
|
|
// This file is distributed under the University of Illinois Open Source
|
|
// License. See LICENSE.TXT for details.
|
|
//
|
|
/// i1 values are usually inserted by the CFG Structurize pass and they are
|
|
/// unique in that they can be copied from VALU to SALU registers.
|
|
/// This is not possible for any other value type. Since there are no
|
|
/// MOV instructions for i1, we to use V_CMP_* and V_CNDMASK to move the i1.
|
|
///
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
|
|
#define DEBUG_TYPE "si-i1-copies"
|
|
#include "AMDGPU.h"
|
|
#include "AMDGPUSubtarget.h"
|
|
#include "SIInstrInfo.h"
|
|
#include "llvm/CodeGen/LiveIntervalAnalysis.h"
|
|
#include "llvm/CodeGen/MachineFunctionPass.h"
|
|
#include "llvm/CodeGen/MachineInstrBuilder.h"
|
|
#include "llvm/CodeGen/MachineRegisterInfo.h"
|
|
#include "llvm/IR/Function.h"
|
|
#include "llvm/IR/LLVMContext.h"
|
|
#include "llvm/Support/Debug.h"
|
|
#include "llvm/Target/TargetMachine.h"
|
|
|
|
using namespace llvm;
|
|
|
|
namespace {
|
|
|
|
class SILowerI1Copies : public MachineFunctionPass {
|
|
public:
|
|
static char ID;
|
|
|
|
public:
|
|
SILowerI1Copies() : MachineFunctionPass(ID) {
|
|
initializeSILowerI1CopiesPass(*PassRegistry::getPassRegistry());
|
|
}
|
|
|
|
bool runOnMachineFunction(MachineFunction &MF) override;
|
|
|
|
StringRef getPassName() const override { return "SI Lower i1 Copies"; }
|
|
|
|
void getAnalysisUsage(AnalysisUsage &AU) const override {
|
|
AU.setPreservesCFG();
|
|
MachineFunctionPass::getAnalysisUsage(AU);
|
|
}
|
|
};
|
|
|
|
} // End anonymous namespace.
|
|
|
|
INITIALIZE_PASS(SILowerI1Copies, DEBUG_TYPE,
|
|
"SI Lower i1 Copies", false, false)
|
|
|
|
char SILowerI1Copies::ID = 0;
|
|
|
|
char &llvm::SILowerI1CopiesID = SILowerI1Copies::ID;
|
|
|
|
FunctionPass *llvm::createSILowerI1CopiesPass() {
|
|
return new SILowerI1Copies();
|
|
}
|
|
|
|
bool SILowerI1Copies::runOnMachineFunction(MachineFunction &MF) {
|
|
MachineRegisterInfo &MRI = MF.getRegInfo();
|
|
const SISubtarget &ST = MF.getSubtarget<SISubtarget>();
|
|
const SIInstrInfo *TII = ST.getInstrInfo();
|
|
const TargetRegisterInfo *TRI = &TII->getRegisterInfo();
|
|
|
|
std::vector<unsigned> I1Defs;
|
|
|
|
for (MachineFunction::iterator BI = MF.begin(), BE = MF.end();
|
|
BI != BE; ++BI) {
|
|
|
|
MachineBasicBlock &MBB = *BI;
|
|
MachineBasicBlock::iterator I, Next;
|
|
for (I = MBB.begin(); I != MBB.end(); I = Next) {
|
|
Next = std::next(I);
|
|
MachineInstr &MI = *I;
|
|
|
|
if (MI.getOpcode() == AMDGPU::IMPLICIT_DEF) {
|
|
unsigned Reg = MI.getOperand(0).getReg();
|
|
const TargetRegisterClass *RC = MRI.getRegClass(Reg);
|
|
if (RC == &AMDGPU::VReg_1RegClass)
|
|
MRI.setRegClass(Reg, &AMDGPU::SReg_64RegClass);
|
|
continue;
|
|
}
|
|
|
|
if (MI.getOpcode() != AMDGPU::COPY)
|
|
continue;
|
|
|
|
const MachineOperand &Dst = MI.getOperand(0);
|
|
const MachineOperand &Src = MI.getOperand(1);
|
|
|
|
if (!TargetRegisterInfo::isVirtualRegister(Src.getReg()) ||
|
|
!TargetRegisterInfo::isVirtualRegister(Dst.getReg()))
|
|
continue;
|
|
|
|
const TargetRegisterClass *DstRC = MRI.getRegClass(Dst.getReg());
|
|
const TargetRegisterClass *SrcRC = MRI.getRegClass(Src.getReg());
|
|
|
|
DebugLoc DL = MI.getDebugLoc();
|
|
MachineInstr *DefInst = MRI.getUniqueVRegDef(Src.getReg());
|
|
if (DstRC == &AMDGPU::VReg_1RegClass &&
|
|
TRI->getCommonSubClass(SrcRC, &AMDGPU::SGPR_64RegClass)) {
|
|
I1Defs.push_back(Dst.getReg());
|
|
|
|
if (DefInst->getOpcode() == AMDGPU::S_MOV_B64) {
|
|
if (DefInst->getOperand(1).isImm()) {
|
|
I1Defs.push_back(Dst.getReg());
|
|
|
|
int64_t Val = DefInst->getOperand(1).getImm();
|
|
assert(Val == 0 || Val == -1);
|
|
|
|
BuildMI(MBB, &MI, DL, TII->get(AMDGPU::V_MOV_B32_e32))
|
|
.add(Dst)
|
|
.addImm(Val);
|
|
MI.eraseFromParent();
|
|
continue;
|
|
}
|
|
}
|
|
|
|
BuildMI(MBB, &MI, DL, TII->get(AMDGPU::V_CNDMASK_B32_e64))
|
|
.add(Dst)
|
|
.addImm(0)
|
|
.addImm(-1)
|
|
.add(Src);
|
|
MI.eraseFromParent();
|
|
} else if (TRI->getCommonSubClass(DstRC, &AMDGPU::SGPR_64RegClass) &&
|
|
SrcRC == &AMDGPU::VReg_1RegClass) {
|
|
if (DefInst->getOpcode() == AMDGPU::V_CNDMASK_B32_e64 &&
|
|
DefInst->getOperand(1).isImm() && DefInst->getOperand(2).isImm() &&
|
|
DefInst->getOperand(1).getImm() == 0 &&
|
|
DefInst->getOperand(2).getImm() != 0 &&
|
|
DefInst->getOperand(3).isReg() &&
|
|
TargetRegisterInfo::isVirtualRegister(
|
|
DefInst->getOperand(3).getReg()) &&
|
|
TRI->getCommonSubClass(
|
|
MRI.getRegClass(DefInst->getOperand(3).getReg()),
|
|
&AMDGPU::SGPR_64RegClass)) {
|
|
BuildMI(MBB, &MI, DL, TII->get(AMDGPU::S_AND_B64))
|
|
.add(Dst)
|
|
.addReg(AMDGPU::EXEC)
|
|
.add(DefInst->getOperand(3));
|
|
} else {
|
|
BuildMI(MBB, &MI, DL, TII->get(AMDGPU::V_CMP_NE_U32_e64))
|
|
.add(Dst)
|
|
.add(Src)
|
|
.addImm(0);
|
|
}
|
|
MI.eraseFromParent();
|
|
}
|
|
}
|
|
}
|
|
|
|
for (unsigned Reg : I1Defs)
|
|
MRI.setRegClass(Reg, &AMDGPU::VGPR_32RegClass);
|
|
|
|
return false;
|
|
}
|