2005-10-16 05:39:50 +00:00
|
|
|
//===-- PPCISelLowering.h - PPC32 DAG Lowering Interface --------*- C++ -*-===//
|
2005-08-16 17:14:42 +00:00
|
|
|
//
|
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
2007-12-29 20:36:04 +00:00
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
2005-08-16 17:14:42 +00:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file defines the interfaces that PPC uses to lower LLVM code into a
|
|
|
|
// selection DAG.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#ifndef LLVM_TARGET_POWERPC_PPC32ISELLOWERING_H
|
|
|
|
#define LLVM_TARGET_POWERPC_PPC32ISELLOWERING_H
|
|
|
|
|
2005-10-14 23:51:18 +00:00
|
|
|
#include "PPC.h"
|
2006-11-02 01:44:04 +00:00
|
|
|
#include "PPCSubtarget.h"
|
2012-03-17 18:46:09 +00:00
|
|
|
#include "llvm/CodeGen/SelectionDAG.h"
|
2012-12-04 07:12:27 +00:00
|
|
|
#include "llvm/Target/TargetLowering.h"
|
2005-08-16 17:14:42 +00:00
|
|
|
|
|
|
|
namespace llvm {
|
2005-08-26 20:25:03 +00:00
|
|
|
namespace PPCISD {
|
|
|
|
enum NodeType {
|
2007-01-26 22:40:50 +00:00
|
|
|
// Start the numbering where the builtin ops and target ops leave off.
|
2008-09-23 18:42:32 +00:00
|
|
|
FIRST_NUMBER = ISD::BUILTIN_OP_END,
|
2005-08-26 20:25:03 +00:00
|
|
|
|
|
|
|
/// FSEL - Traditional three-operand fsel node.
|
|
|
|
///
|
|
|
|
FSEL,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-09-06 22:03:27 +00:00
|
|
|
/// FCFID - The FCFID instruction, taking an f64 operand and producing
|
|
|
|
/// and f64 value containing the FP representation of the integer that
|
|
|
|
/// was temporarily in the f64 operand.
|
|
|
|
FCFID,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
|
|
|
/// FCTI[D,W]Z - The FCTIDZ and FCTIWZ instructions, taking an f32 or f64
|
2005-09-06 22:03:27 +00:00
|
|
|
/// operand, producing an f64 value containing the integer representation
|
|
|
|
/// of that FP value.
|
|
|
|
FCTIDZ, FCTIWZ,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-03-01 05:50:56 +00:00
|
|
|
/// STFIWX - The STFIWX instruction. The first operand is an input token
|
2009-09-25 20:36:54 +00:00
|
|
|
/// chain, then an f64 value to store, then an address to store it to.
|
2006-03-01 05:50:56 +00:00
|
|
|
STFIWX,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-12-13 22:55:22 +00:00
|
|
|
// VMADDFP, VNMSUBFP - The VMADDFP and VNMSUBFP instructions, taking
|
|
|
|
// three v4f32 operands and producing a v4f32 result.
|
|
|
|
VMADDFP, VNMSUBFP,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-03-20 01:53:53 +00:00
|
|
|
/// VPERM - The PPC VPERM Instruction.
|
|
|
|
///
|
|
|
|
VPERM,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-11-17 07:30:41 +00:00
|
|
|
/// Hi/Lo - These represent the high and low 16-bit parts of a global
|
|
|
|
/// address respectively. These nodes have two operands, the first of
|
|
|
|
/// which must be a TargetGlobalAddress, and the second of which must be a
|
|
|
|
/// Constant. Selected naively, these turn into 'lis G+C' and 'li G+C',
|
|
|
|
/// though these are usually folded into other nodes.
|
|
|
|
Hi, Lo,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2009-08-15 11:54:46 +00:00
|
|
|
TOC_ENTRY,
|
|
|
|
|
2009-12-18 13:00:15 +00:00
|
|
|
/// The following three target-specific nodes are used for calls through
|
|
|
|
/// function pointers in the 64-bit SVR4 ABI.
|
|
|
|
|
|
|
|
/// Restore the TOC from the TOC save area of the current stack frame.
|
|
|
|
/// This is basically a hard coded load instruction which additionally
|
|
|
|
/// takes/produces a flag.
|
|
|
|
TOC_RESTORE,
|
|
|
|
|
|
|
|
/// Like a regular LOAD but additionally taking/producing a flag.
|
|
|
|
LOAD,
|
|
|
|
|
|
|
|
/// LOAD into r2 (also taking/producing a flag). Like TOC_RESTORE, this is
|
|
|
|
/// a hard coded load instruction.
|
|
|
|
LOAD_TOC,
|
|
|
|
|
2006-11-16 22:43:37 +00:00
|
|
|
/// OPRC, CHAIN = DYNALLOC(CHAIN, NEGSIZE, FRAME_INDEX)
|
|
|
|
/// This instruction is lowered in PPCRegisterInfo::eliminateFrameIndex to
|
|
|
|
/// compute an allocation on the stack.
|
|
|
|
DYNALLOC,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-11-17 07:30:41 +00:00
|
|
|
/// GlobalBaseReg - On Darwin, this node represents the result of the mflr
|
|
|
|
/// at function entry, used for PIC code.
|
|
|
|
GlobalBaseReg,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-12-06 02:10:38 +00:00
|
|
|
/// These nodes represent the 32-bit PPC shifts that operate on 6-bit
|
|
|
|
/// shift amounts. These nodes are generated by the multi-precision shift
|
|
|
|
/// code.
|
|
|
|
SRL, SRA, SHL,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
When possible, custom lower 32-bit SINT_TO_FP to this:
_foo2:
extsw r2, r3
std r2, -8(r1)
lfd f0, -8(r1)
fcfid f0, f0
frsp f1, f0
blr
instead of this:
_foo2:
lis r2, ha16(LCPI2_0)
lis r4, 17200
xoris r3, r3, 32768
stw r3, -4(r1)
stw r4, -8(r1)
lfs f0, lo16(LCPI2_0)(r2)
lfd f1, -8(r1)
fsub f0, f1, f0
frsp f1, f0
blr
This speeds up Misc/pi from 2.44s->2.09s with LLC and from 3.01->2.18s
with llcbeta (16.7% and 38.1% respectively).
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@26943 91177308-0d34-0410-b5e6-96231b3b80d8
2006-03-22 05:30:33 +00:00
|
|
|
/// EXTSW_32 - This is the EXTSW instruction for use with "32-bit"
|
|
|
|
/// registers.
|
|
|
|
EXTSW_32,
|
2005-12-20 00:26:01 +00:00
|
|
|
|
2006-05-17 19:00:46 +00:00
|
|
|
/// CALL - A direct function call.
|
2012-03-31 14:45:15 +00:00
|
|
|
/// CALL_NOP_SVR4 is a call with the special NOP which follows 64-bit
|
|
|
|
/// SVR4 calls.
|
|
|
|
CALL_Darwin, CALL_SVR4, CALL_NOP_SVR4,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2009-08-15 11:54:46 +00:00
|
|
|
/// NOP - Special NOP which follows 64-bit SVR4 calls.
|
|
|
|
NOP,
|
|
|
|
|
2006-05-17 19:00:46 +00:00
|
|
|
/// CHAIN,FLAG = MTCTR(VAL, CHAIN[, INFLAG]) - Directly corresponds to a
|
|
|
|
/// MTCTR instruction.
|
|
|
|
MTCTR,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-05-17 19:00:46 +00:00
|
|
|
/// CHAIN,FLAG = BCTRL(CHAIN, INFLAG) - Directly corresponds to a
|
|
|
|
/// BCTRL instruction.
|
2009-07-03 06:47:08 +00:00
|
|
|
BCTRL_Darwin, BCTRL_SVR4,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-12-20 00:26:01 +00:00
|
|
|
/// Return with a flag operand, matched by 'blr'
|
|
|
|
RET_FLAG,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2010-05-20 17:48:26 +00:00
|
|
|
/// R32 = MFCR(CRREG, INFLAG) - Represents the MFCRpseud/MFOCRF
|
|
|
|
/// instructions. This copies the bits corresponding to the specified
|
|
|
|
/// CRREG into the resultant GPR. Bits corresponding to other CR regs
|
|
|
|
/// are undefined.
|
2006-03-26 10:06:40 +00:00
|
|
|
MFCR,
|
2006-03-31 05:13:27 +00:00
|
|
|
|
|
|
|
/// RESVEC = VCMP(LHS, RHS, OPC) - Represents one of the altivec VCMP*
|
|
|
|
/// instructions. For lack of better number, we use the opcode number
|
|
|
|
/// encoding for the OPC field to identify the compare. For example, 838
|
|
|
|
/// is VCMPGTSH.
|
|
|
|
VCMP,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-03-26 10:06:40 +00:00
|
|
|
/// RESVEC, OUTFLAG = VCMPo(LHS, RHS, OPC) - Represents one of the
|
2011-02-25 21:41:48 +00:00
|
|
|
/// altivec VCMP*o instructions. For lack of better number, we use the
|
2006-03-26 10:06:40 +00:00
|
|
|
/// opcode number encoding for the OPC field to identify the compare. For
|
|
|
|
/// example, 838 is VCMPGTSH.
|
Implement an important entry from README_ALTIVEC:
If an altivec predicate compare is used immediately by a branch, don't
use a (serializing) MFCR instruction to read the CR6 register, which requires
a compare to get it back to CR's. Instead, just branch on CR6 directly. :)
For example, for:
void foo2(vector float *A, vector float *B) {
if (!vec_any_eq(*A, *B))
*B = (vector float){0,0,0,0};
}
We now generate:
_foo2:
mfspr r2, 256
oris r5, r2, 12288
mtspr 256, r5
lvx v2, 0, r4
lvx v3, 0, r3
vcmpeqfp. v2, v3, v2
bne cr6, LBB1_2 ; UnifiedReturnBlock
LBB1_1: ; cond_true
vxor v2, v2, v2
stvx v2, 0, r4
mtspr 256, r2
blr
LBB1_2: ; UnifiedReturnBlock
mtspr 256, r2
blr
instead of:
_foo2:
mfspr r2, 256
oris r5, r2, 12288
mtspr 256, r5
lvx v2, 0, r4
lvx v3, 0, r3
vcmpeqfp. v2, v3, v2
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
cmpwi cr0, r3, 0
beq cr0, LBB1_2 ; UnifiedReturnBlock
LBB1_1: ; cond_true
vxor v2, v2, v2
stvx v2, 0, r4
mtspr 256, r2
blr
LBB1_2: ; UnifiedReturnBlock
mtspr 256, r2
blr
This implements CodeGen/PowerPC/vec_br_cmp.ll.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27804 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-18 17:59:36 +00:00
|
|
|
VCMPo,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
Implement an important entry from README_ALTIVEC:
If an altivec predicate compare is used immediately by a branch, don't
use a (serializing) MFCR instruction to read the CR6 register, which requires
a compare to get it back to CR's. Instead, just branch on CR6 directly. :)
For example, for:
void foo2(vector float *A, vector float *B) {
if (!vec_any_eq(*A, *B))
*B = (vector float){0,0,0,0};
}
We now generate:
_foo2:
mfspr r2, 256
oris r5, r2, 12288
mtspr 256, r5
lvx v2, 0, r4
lvx v3, 0, r3
vcmpeqfp. v2, v3, v2
bne cr6, LBB1_2 ; UnifiedReturnBlock
LBB1_1: ; cond_true
vxor v2, v2, v2
stvx v2, 0, r4
mtspr 256, r2
blr
LBB1_2: ; UnifiedReturnBlock
mtspr 256, r2
blr
instead of:
_foo2:
mfspr r2, 256
oris r5, r2, 12288
mtspr 256, r5
lvx v2, 0, r4
lvx v3, 0, r3
vcmpeqfp. v2, v3, v2
mfcr r3, 2
rlwinm r3, r3, 27, 31, 31
cmpwi cr0, r3, 0
beq cr0, LBB1_2 ; UnifiedReturnBlock
LBB1_1: ; cond_true
vxor v2, v2, v2
stvx v2, 0, r4
mtspr 256, r2
blr
LBB1_2: ; UnifiedReturnBlock
mtspr 256, r2
blr
This implements CodeGen/PowerPC/vec_br_cmp.ll.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@27804 91177308-0d34-0410-b5e6-96231b3b80d8
2006-04-18 17:59:36 +00:00
|
|
|
/// CHAIN = COND_BRANCH CHAIN, CRRC, OPC, DESTBB [, INFLAG] - This
|
|
|
|
/// corresponds to the COND_BRANCH pseudo instruction. CRRC is the
|
|
|
|
/// condition register to branch on, OPC is the branch opcode to use (e.g.
|
|
|
|
/// PPC::BLE), DESTBB is the destination block to branch to, and INFLAG is
|
|
|
|
/// an optional input flag argument.
|
2006-07-10 20:56:58 +00:00
|
|
|
COND_BRANCH,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2007-10-10 01:01:31 +00:00
|
|
|
// The following 5 instructions are used only as part of the
|
|
|
|
// long double-to-int conversion sequence.
|
|
|
|
|
|
|
|
/// OUTFLAG = MFFS F8RC - This moves the FPSCR (not modelled) into the
|
|
|
|
/// register.
|
|
|
|
MFFS,
|
|
|
|
|
|
|
|
/// OUTFLAG = MTFSB0 INFLAG - This clears a bit in the FPSCR.
|
|
|
|
MTFSB0,
|
|
|
|
|
|
|
|
/// OUTFLAG = MTFSB1 INFLAG - This sets a bit in the FPSCR.
|
|
|
|
MTFSB1,
|
|
|
|
|
|
|
|
/// F8RC, OUTFLAG = FADDRTZ F8RC, F8RC, INFLAG - This is an FADD done with
|
2011-02-25 21:41:48 +00:00
|
|
|
/// rounding towards zero. It has flags added so it won't move past the
|
2007-10-10 01:01:31 +00:00
|
|
|
/// FPSCR-setting instructions.
|
|
|
|
FADDRTZ,
|
|
|
|
|
|
|
|
/// MTFSF = F8RC, INFLAG - This moves the register into the FPSCR.
|
2008-04-19 01:30:48 +00:00
|
|
|
MTFSF,
|
|
|
|
|
2008-04-19 02:30:38 +00:00
|
|
|
/// LARX = This corresponds to PPC l{w|d}arx instrcution: load and
|
2008-04-19 01:30:48 +00:00
|
|
|
/// reserve indexed. This is used to implement atomic operations.
|
2008-04-19 02:30:38 +00:00
|
|
|
LARX,
|
2008-04-19 01:30:48 +00:00
|
|
|
|
2008-04-19 02:30:38 +00:00
|
|
|
/// STCX = This corresponds to PPC stcx. instrcution: store conditional
|
|
|
|
/// indexed. This is used to implement atomic operations.
|
|
|
|
STCX,
|
2008-04-19 01:30:48 +00:00
|
|
|
|
2008-04-30 09:16:33 +00:00
|
|
|
/// TC_RETURN - A tail call return.
|
|
|
|
/// operand #0 chain
|
|
|
|
/// operand #1 callee (register or absolute)
|
|
|
|
/// operand #2 stack adjustment
|
|
|
|
/// operand #3 optional in flag
|
2009-09-25 20:36:54 +00:00
|
|
|
TC_RETURN,
|
|
|
|
|
2012-08-28 02:10:27 +00:00
|
|
|
/// ch, gl = CR6[UN]SET ch, inglue - Toggle CR bit 6 for SVR4 vararg calls
|
|
|
|
CR6SET,
|
|
|
|
CR6UNSET,
|
|
|
|
|
This patch improves the 64-bit PowerPC InitialExec TLS support by providing
for a wider range of GOT entries that can hold thread-relative offsets.
This matches the behavior of GCC, which was not documented in the PPC64 TLS
ABI. The ABI will be updated with the new code sequence.
Former sequence:
ld 9,x@got@tprel(2)
add 9,9,x@tls
New sequence:
addis 9,2,x@got@tprel@ha
ld 9,x@got@tprel@l(9)
add 9,9,x@tls
Note that a linker optimization exists to transform the new sequence into
the shorter sequence when appropriate, by replacing the addis with a nop
and modifying the base register and relocation type of the ld.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@170209 91177308-0d34-0410-b5e6-96231b3b80d8
2012-12-14 17:02:38 +00:00
|
|
|
/// G8RC = ADDIS_GOT_TPREL_HA %X2, Symbol - Used by the initial-exec
|
|
|
|
/// TLS model, produces an ADDIS8 instruction that adds the GOT
|
|
|
|
/// base to sym@got@tprel@ha.
|
|
|
|
ADDIS_GOT_TPREL_HA,
|
|
|
|
|
|
|
|
/// G8RC = LD_GOT_TPREL_L Symbol, G8RReg - Used by the initial-exec
|
2012-12-04 16:18:08 +00:00
|
|
|
/// TLS model, produces a LD instruction with base register G8RReg
|
This patch improves the 64-bit PowerPC InitialExec TLS support by providing
for a wider range of GOT entries that can hold thread-relative offsets.
This matches the behavior of GCC, which was not documented in the PPC64 TLS
ABI. The ABI will be updated with the new code sequence.
Former sequence:
ld 9,x@got@tprel(2)
add 9,9,x@tls
New sequence:
addis 9,2,x@got@tprel@ha
ld 9,x@got@tprel@l(9)
add 9,9,x@tls
Note that a linker optimization exists to transform the new sequence into
the shorter sequence when appropriate, by replacing the addis with a nop
and modifying the base register and relocation type of the ld.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@170209 91177308-0d34-0410-b5e6-96231b3b80d8
2012-12-14 17:02:38 +00:00
|
|
|
/// and offset sym@got@tprel@l. This completes the addition that
|
|
|
|
/// finds the offset of "sym" relative to the thread pointer.
|
|
|
|
LD_GOT_TPREL_L,
|
2012-12-04 16:18:08 +00:00
|
|
|
|
|
|
|
/// G8RC = ADD_TLS G8RReg, Symbol - Used by the initial-exec TLS
|
|
|
|
/// model, produces an ADD instruction that adds the contents of
|
|
|
|
/// G8RReg to the thread pointer. Symbol contains a relocation
|
|
|
|
/// sym@tls which is to be replaced by the thread pointer and
|
|
|
|
/// identifies to the linker that the instruction is part of a
|
|
|
|
/// TLS sequence.
|
|
|
|
ADD_TLS,
|
|
|
|
|
This patch implements the general dynamic TLS model for 64-bit PowerPC.
Given a thread-local symbol x with global-dynamic access, the generated
code to obtain x's address is:
Instruction Relocation Symbol
addis ra,r2,x@got@tlsgd@ha R_PPC64_GOT_TLSGD16_HA x
addi r3,ra,x@got@tlsgd@l R_PPC64_GOT_TLSGD16_L x
bl __tls_get_addr(x@tlsgd) R_PPC64_TLSGD x
R_PPC64_REL24 __tls_get_addr
nop
<use address in r3>
The implementation borrows from the medium code model work for introducing
special forms of ADDIS and ADDI into the DAG representation. This is made
slightly more complicated by having to introduce a call to the external
function __tls_get_addr. Using the full call machinery is overkill and,
more importantly, makes it difficult to add a special relocation. So I've
introduced another opcode GET_TLS_ADDR to represent the function call, and
surrounded it with register copies to set up the parameter and return value.
Most of the code is pretty straightforward. I ran into one peculiarity
when I introduced a new PPC opcode BL8_NOP_ELF_TLSGD, which is just like
BL8_NOP_ELF except that it takes another parameter to represent the symbol
("x" above) that requires a relocation on the call. Something in the
TblGen machinery causes BL8_NOP_ELF and BL8_NOP_ELF_TLSGD to be treated
identically during the emit phase, so this second operand was never
visited to generate relocations. This is the reason for the slightly
messy workaround in PPCMCCodeEmitter.cpp:getDirectBrEncoding().
Two new tests are included to demonstrate correct external assembly and
correct generation of relocations using the integrated assembler.
Comments welcome!
Thanks,
Bill
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@169910 91177308-0d34-0410-b5e6-96231b3b80d8
2012-12-11 20:30:11 +00:00
|
|
|
/// G8RC = ADDIS_TLSGD_HA %X2, Symbol - For the general-dynamic TLS
|
|
|
|
/// model, produces an ADDIS8 instruction that adds the GOT base
|
|
|
|
/// register to sym@got@tlsgd@ha.
|
|
|
|
ADDIS_TLSGD_HA,
|
|
|
|
|
|
|
|
/// G8RC = ADDI_TLSGD_L G8RReg, Symbol - For the general-dynamic TLS
|
|
|
|
/// model, produces an ADDI8 instruction that adds G8RReg to
|
|
|
|
/// sym@got@tlsgd@l.
|
|
|
|
ADDI_TLSGD_L,
|
|
|
|
|
|
|
|
/// G8RC = GET_TLS_ADDR %X3, Symbol - For the general-dynamic TLS
|
|
|
|
/// model, produces a call to __tls_get_addr(sym@tlsgd).
|
|
|
|
GET_TLS_ADDR,
|
|
|
|
|
2012-12-12 19:29:35 +00:00
|
|
|
/// G8RC = ADDIS_TLSLD_HA %X2, Symbol - For the local-dynamic TLS
|
|
|
|
/// model, produces an ADDIS8 instruction that adds the GOT base
|
|
|
|
/// register to sym@got@tlsld@ha.
|
|
|
|
ADDIS_TLSLD_HA,
|
|
|
|
|
|
|
|
/// G8RC = ADDI_TLSLD_L G8RReg, Symbol - For the local-dynamic TLS
|
|
|
|
/// model, produces an ADDI8 instruction that adds G8RReg to
|
|
|
|
/// sym@got@tlsld@l.
|
|
|
|
ADDI_TLSLD_L,
|
|
|
|
|
|
|
|
/// G8RC = GET_TLSLD_ADDR %X3, Symbol - For the local-dynamic TLS
|
|
|
|
/// model, produces a call to __tls_get_addr(sym@tlsld).
|
|
|
|
GET_TLSLD_ADDR,
|
|
|
|
|
|
|
|
/// G8RC = ADDIS_DTPREL_HA %X3, Symbol, Chain - For the
|
|
|
|
/// local-dynamic TLS model, produces an ADDIS8 instruction
|
|
|
|
/// that adds X3 to sym@dtprel@ha. The Chain operand is needed
|
|
|
|
/// to tie this in place following a copy to %X3 from the result
|
|
|
|
/// of a GET_TLSLD_ADDR.
|
|
|
|
ADDIS_DTPREL_HA,
|
|
|
|
|
|
|
|
/// G8RC = ADDI_DTPREL_L G8RReg, Symbol - For the local-dynamic TLS
|
|
|
|
/// model, produces an ADDI8 instruction that adds G8RReg to
|
|
|
|
/// sym@got@dtprel@l.
|
|
|
|
ADDI_DTPREL_L,
|
|
|
|
|
2013-02-20 15:50:31 +00:00
|
|
|
/// VRRC = VADD_SPLAT Elt, EltSize - Temporary node to be expanded
|
2013-02-20 20:41:42 +00:00
|
|
|
/// during instruction selection to optimize a BUILD_VECTOR into
|
|
|
|
/// operations on splats. This is necessary to avoid losing these
|
|
|
|
/// optimizations due to constant folding.
|
2013-02-20 15:50:31 +00:00
|
|
|
VADD_SPLAT,
|
|
|
|
|
2009-09-25 20:36:54 +00:00
|
|
|
/// STD_32 - This is the STD instruction for use with "32-bit" registers.
|
|
|
|
STD_32 = ISD::FIRST_TARGET_MEMORY_OPCODE,
|
2011-02-25 21:41:48 +00:00
|
|
|
|
|
|
|
/// CHAIN = STBRX CHAIN, GPRC, Ptr, Type - This is a
|
2009-09-25 20:36:54 +00:00
|
|
|
/// byte-swapping store instruction. It byte-swaps the low "Type" bits of
|
|
|
|
/// the GPRC input, then stores it through Ptr. Type can be either i16 or
|
|
|
|
/// i32.
|
2011-02-25 21:41:48 +00:00
|
|
|
STBRX,
|
|
|
|
|
|
|
|
/// GPRC, CHAIN = LBRX CHAIN, Ptr, Type - This is a
|
2009-09-25 20:36:54 +00:00
|
|
|
/// byte-swapping load instruction. It loads "Type" bits, byte swaps it,
|
|
|
|
/// then puts it in the bottom bits of the GPRC. TYPE can be either i16
|
|
|
|
/// or i32.
|
This patch implements medium code model support for 64-bit PowerPC.
The default for 64-bit PowerPC is small code model, in which TOC entries
must be addressable using a 16-bit offset from the TOC pointer. Additionally,
only TOC entries are addressed via the TOC pointer.
With medium code model, TOC entries and data sections can all be addressed
via the TOC pointer using a 32-bit offset. Cooperation with the linker
allows 16-bit offsets to be used when these are sufficient, reducing the
number of extra instructions that need to be executed. Medium code model
also does not generate explicit TOC entries in ".section toc" for variables
that are wholly internal to the compilation unit.
Consider a load of an external 4-byte integer. With small code model, the
compiler generates:
ld 3, .LC1@toc(2)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc ei[TC],ei
With medium model, it instead generates:
addis 3, 2, .LC1@toc@ha
ld 3, .LC1@toc@l(3)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc ei[TC],ei
Here .LC1@toc@ha is a relocation requesting the upper 16 bits of the
32-bit offset of ei's TOC entry from the TOC base pointer. Similarly,
.LC1@toc@l is a relocation requesting the lower 16 bits. Note that if
the linker determines that ei's TOC entry is within a 16-bit offset of
the TOC base pointer, it will replace the "addis" with a "nop", and
replace the "ld" with the identical "ld" instruction from the small
code model example.
Consider next a load of a function-scope static integer. For small code
model, the compiler generates:
ld 3, .LC1@toc(2)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc test_fn_static.si[TC],test_fn_static.si
.type test_fn_static.si,@object
.local test_fn_static.si
.comm test_fn_static.si,4,4
For medium code model, the compiler generates:
addis 3, 2, test_fn_static.si@toc@ha
addi 3, 3, test_fn_static.si@toc@l
lwz 4, 0(3)
.type test_fn_static.si,@object
.local test_fn_static.si
.comm test_fn_static.si,4,4
Again, the linker may replace the "addis" with a "nop", calculating only
a 16-bit offset when this is sufficient.
Note that it would be more efficient for the compiler to generate:
addis 3, 2, test_fn_static.si@toc@ha
lwz 4, test_fn_static.si@toc@l(3)
The current patch does not perform this optimization yet. This will be
addressed as a peephole optimization in a later patch.
For the moment, the default code model for 64-bit PowerPC will remain the
small code model. We plan to eventually change the default to medium code
model, which matches current upstream GCC behavior. Note that the different
code models are ABI-compatible, so code compiled with different models will
be linked and execute correctly.
I've tested the regression suite and the application/benchmark test suite in
two ways: Once with the patch as submitted here, and once with additional
logic to force medium code model as the default. The tests all compile
cleanly, with one exception. The mandel-2 application test fails due to an
unrelated ABI compatibility with passing complex numbers. It just so happens
that small code model was incredibly lucky, in that temporary values in
floating-point registers held the expected values needed by the external
library routine that was called incorrectly. My current thought is to correct
the ABI problems with _Complex before making medium code model the default,
to avoid introducing this "regression."
Here are a few comments on how the patch works, since the selection code
can be difficult to follow:
The existing logic for small code model defines three pseudo-instructions:
LDtoc for most uses, LDtocJTI for jump table addresses, and LDtocCPT for
constant pool addresses. These are expanded by SelectCodeCommon(). The
pseudo-instruction approach doesn't work for medium code model, because
we need to generate two instructions when we match the same pattern.
Instead, new logic in PPCDAGToDAGISel::Select() intercepts the TOC_ENTRY
node for medium code model, and generates an ADDIStocHA followed by either
a LDtocL or an ADDItocL. These new node types correspond naturally to
the sequences described above.
The addis/ld sequence is generated for the following cases:
* Jump table addresses
* Function addresses
* External global variables
* Tentative definitions of global variables (common linkage)
The addis/addi sequence is generated for the following cases:
* Constant pool entries
* File-scope static global variables
* Function-scope static variables
Expanding to the two-instruction sequences at select time exposes the
instructions to subsequent optimization, particularly scheduling.
The rest of the processing occurs at assembly time, in
PPCAsmPrinter::EmitInstruction. Each of the instructions is converted to
a "real" PowerPC instruction. When a TOC entry needs to be created, this
is done here in the same manner as for the existing LDtoc, LDtocJTI, and
LDtocCPT pseudo-instructions (I factored out a new routine to handle this).
I had originally thought that if a TOC entry was needed for LDtocL or
ADDItocL, it would already have been generated for the previous ADDIStocHA.
However, at higher optimization levels, the ADDIStocHA may appear in a
different block, which may be assembled textually following the block
containing the LDtocL or ADDItocL. So it is necessary to include the
possibility of creating a new TOC entry for those two instructions.
Note that for LDtocL, we generate a new form of LD called LDrs. This
allows specifying the @toc@l relocation for the offset field of the LD
instruction (i.e., the offset is replaced by a SymbolLo relocation).
When the peephole optimization described above is added, we will need
to do similar things for all immediate-form load and store operations.
The seven "mcm-n.ll" test cases are kept separate because otherwise the
intermingling of various TOC entries and so forth makes the tests fragile
and hard to understand.
The above assumes use of an external assembler. For use of the
integrated assembler, new relocations are added and used by
PPCELFObjectWriter. Testing is done with "mcm-obj.ll", which tests for
proper generation of the various relocations for the same sequences
tested with the external assembler.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@168708 91177308-0d34-0410-b5e6-96231b3b80d8
2012-11-27 17:35:46 +00:00
|
|
|
LBRX,
|
|
|
|
|
2013-02-21 17:12:27 +00:00
|
|
|
/// G8RC = ADDIS_TOC_HA %X2, Symbol - For medium and large code model,
|
|
|
|
/// produces an ADDIS8 instruction that adds the TOC base register to
|
|
|
|
/// sym@toc@ha.
|
This patch implements medium code model support for 64-bit PowerPC.
The default for 64-bit PowerPC is small code model, in which TOC entries
must be addressable using a 16-bit offset from the TOC pointer. Additionally,
only TOC entries are addressed via the TOC pointer.
With medium code model, TOC entries and data sections can all be addressed
via the TOC pointer using a 32-bit offset. Cooperation with the linker
allows 16-bit offsets to be used when these are sufficient, reducing the
number of extra instructions that need to be executed. Medium code model
also does not generate explicit TOC entries in ".section toc" for variables
that are wholly internal to the compilation unit.
Consider a load of an external 4-byte integer. With small code model, the
compiler generates:
ld 3, .LC1@toc(2)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc ei[TC],ei
With medium model, it instead generates:
addis 3, 2, .LC1@toc@ha
ld 3, .LC1@toc@l(3)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc ei[TC],ei
Here .LC1@toc@ha is a relocation requesting the upper 16 bits of the
32-bit offset of ei's TOC entry from the TOC base pointer. Similarly,
.LC1@toc@l is a relocation requesting the lower 16 bits. Note that if
the linker determines that ei's TOC entry is within a 16-bit offset of
the TOC base pointer, it will replace the "addis" with a "nop", and
replace the "ld" with the identical "ld" instruction from the small
code model example.
Consider next a load of a function-scope static integer. For small code
model, the compiler generates:
ld 3, .LC1@toc(2)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc test_fn_static.si[TC],test_fn_static.si
.type test_fn_static.si,@object
.local test_fn_static.si
.comm test_fn_static.si,4,4
For medium code model, the compiler generates:
addis 3, 2, test_fn_static.si@toc@ha
addi 3, 3, test_fn_static.si@toc@l
lwz 4, 0(3)
.type test_fn_static.si,@object
.local test_fn_static.si
.comm test_fn_static.si,4,4
Again, the linker may replace the "addis" with a "nop", calculating only
a 16-bit offset when this is sufficient.
Note that it would be more efficient for the compiler to generate:
addis 3, 2, test_fn_static.si@toc@ha
lwz 4, test_fn_static.si@toc@l(3)
The current patch does not perform this optimization yet. This will be
addressed as a peephole optimization in a later patch.
For the moment, the default code model for 64-bit PowerPC will remain the
small code model. We plan to eventually change the default to medium code
model, which matches current upstream GCC behavior. Note that the different
code models are ABI-compatible, so code compiled with different models will
be linked and execute correctly.
I've tested the regression suite and the application/benchmark test suite in
two ways: Once with the patch as submitted here, and once with additional
logic to force medium code model as the default. The tests all compile
cleanly, with one exception. The mandel-2 application test fails due to an
unrelated ABI compatibility with passing complex numbers. It just so happens
that small code model was incredibly lucky, in that temporary values in
floating-point registers held the expected values needed by the external
library routine that was called incorrectly. My current thought is to correct
the ABI problems with _Complex before making medium code model the default,
to avoid introducing this "regression."
Here are a few comments on how the patch works, since the selection code
can be difficult to follow:
The existing logic for small code model defines three pseudo-instructions:
LDtoc for most uses, LDtocJTI for jump table addresses, and LDtocCPT for
constant pool addresses. These are expanded by SelectCodeCommon(). The
pseudo-instruction approach doesn't work for medium code model, because
we need to generate two instructions when we match the same pattern.
Instead, new logic in PPCDAGToDAGISel::Select() intercepts the TOC_ENTRY
node for medium code model, and generates an ADDIStocHA followed by either
a LDtocL or an ADDItocL. These new node types correspond naturally to
the sequences described above.
The addis/ld sequence is generated for the following cases:
* Jump table addresses
* Function addresses
* External global variables
* Tentative definitions of global variables (common linkage)
The addis/addi sequence is generated for the following cases:
* Constant pool entries
* File-scope static global variables
* Function-scope static variables
Expanding to the two-instruction sequences at select time exposes the
instructions to subsequent optimization, particularly scheduling.
The rest of the processing occurs at assembly time, in
PPCAsmPrinter::EmitInstruction. Each of the instructions is converted to
a "real" PowerPC instruction. When a TOC entry needs to be created, this
is done here in the same manner as for the existing LDtoc, LDtocJTI, and
LDtocCPT pseudo-instructions (I factored out a new routine to handle this).
I had originally thought that if a TOC entry was needed for LDtocL or
ADDItocL, it would already have been generated for the previous ADDIStocHA.
However, at higher optimization levels, the ADDIStocHA may appear in a
different block, which may be assembled textually following the block
containing the LDtocL or ADDItocL. So it is necessary to include the
possibility of creating a new TOC entry for those two instructions.
Note that for LDtocL, we generate a new form of LD called LDrs. This
allows specifying the @toc@l relocation for the offset field of the LD
instruction (i.e., the offset is replaced by a SymbolLo relocation).
When the peephole optimization described above is added, we will need
to do similar things for all immediate-form load and store operations.
The seven "mcm-n.ll" test cases are kept separate because otherwise the
intermingling of various TOC entries and so forth makes the tests fragile
and hard to understand.
The above assumes use of an external assembler. For use of the
integrated assembler, new relocations are added and used by
PPCELFObjectWriter. Testing is done with "mcm-obj.ll", which tests for
proper generation of the various relocations for the same sequences
tested with the external assembler.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@168708 91177308-0d34-0410-b5e6-96231b3b80d8
2012-11-27 17:35:46 +00:00
|
|
|
ADDIS_TOC_HA,
|
|
|
|
|
2013-02-21 17:12:27 +00:00
|
|
|
/// G8RC = LD_TOC_L Symbol, G8RReg - For medium and large code model,
|
|
|
|
/// produces a LD instruction with base register G8RReg and offset
|
|
|
|
/// sym@toc@l. Preceded by an ADDIS_TOC_HA to form a full 32-bit offset.
|
This patch implements medium code model support for 64-bit PowerPC.
The default for 64-bit PowerPC is small code model, in which TOC entries
must be addressable using a 16-bit offset from the TOC pointer. Additionally,
only TOC entries are addressed via the TOC pointer.
With medium code model, TOC entries and data sections can all be addressed
via the TOC pointer using a 32-bit offset. Cooperation with the linker
allows 16-bit offsets to be used when these are sufficient, reducing the
number of extra instructions that need to be executed. Medium code model
also does not generate explicit TOC entries in ".section toc" for variables
that are wholly internal to the compilation unit.
Consider a load of an external 4-byte integer. With small code model, the
compiler generates:
ld 3, .LC1@toc(2)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc ei[TC],ei
With medium model, it instead generates:
addis 3, 2, .LC1@toc@ha
ld 3, .LC1@toc@l(3)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc ei[TC],ei
Here .LC1@toc@ha is a relocation requesting the upper 16 bits of the
32-bit offset of ei's TOC entry from the TOC base pointer. Similarly,
.LC1@toc@l is a relocation requesting the lower 16 bits. Note that if
the linker determines that ei's TOC entry is within a 16-bit offset of
the TOC base pointer, it will replace the "addis" with a "nop", and
replace the "ld" with the identical "ld" instruction from the small
code model example.
Consider next a load of a function-scope static integer. For small code
model, the compiler generates:
ld 3, .LC1@toc(2)
lwz 4, 0(3)
.section .toc,"aw",@progbits
.LC1:
.tc test_fn_static.si[TC],test_fn_static.si
.type test_fn_static.si,@object
.local test_fn_static.si
.comm test_fn_static.si,4,4
For medium code model, the compiler generates:
addis 3, 2, test_fn_static.si@toc@ha
addi 3, 3, test_fn_static.si@toc@l
lwz 4, 0(3)
.type test_fn_static.si,@object
.local test_fn_static.si
.comm test_fn_static.si,4,4
Again, the linker may replace the "addis" with a "nop", calculating only
a 16-bit offset when this is sufficient.
Note that it would be more efficient for the compiler to generate:
addis 3, 2, test_fn_static.si@toc@ha
lwz 4, test_fn_static.si@toc@l(3)
The current patch does not perform this optimization yet. This will be
addressed as a peephole optimization in a later patch.
For the moment, the default code model for 64-bit PowerPC will remain the
small code model. We plan to eventually change the default to medium code
model, which matches current upstream GCC behavior. Note that the different
code models are ABI-compatible, so code compiled with different models will
be linked and execute correctly.
I've tested the regression suite and the application/benchmark test suite in
two ways: Once with the patch as submitted here, and once with additional
logic to force medium code model as the default. The tests all compile
cleanly, with one exception. The mandel-2 application test fails due to an
unrelated ABI compatibility with passing complex numbers. It just so happens
that small code model was incredibly lucky, in that temporary values in
floating-point registers held the expected values needed by the external
library routine that was called incorrectly. My current thought is to correct
the ABI problems with _Complex before making medium code model the default,
to avoid introducing this "regression."
Here are a few comments on how the patch works, since the selection code
can be difficult to follow:
The existing logic for small code model defines three pseudo-instructions:
LDtoc for most uses, LDtocJTI for jump table addresses, and LDtocCPT for
constant pool addresses. These are expanded by SelectCodeCommon(). The
pseudo-instruction approach doesn't work for medium code model, because
we need to generate two instructions when we match the same pattern.
Instead, new logic in PPCDAGToDAGISel::Select() intercepts the TOC_ENTRY
node for medium code model, and generates an ADDIStocHA followed by either
a LDtocL or an ADDItocL. These new node types correspond naturally to
the sequences described above.
The addis/ld sequence is generated for the following cases:
* Jump table addresses
* Function addresses
* External global variables
* Tentative definitions of global variables (common linkage)
The addis/addi sequence is generated for the following cases:
* Constant pool entries
* File-scope static global variables
* Function-scope static variables
Expanding to the two-instruction sequences at select time exposes the
instructions to subsequent optimization, particularly scheduling.
The rest of the processing occurs at assembly time, in
PPCAsmPrinter::EmitInstruction. Each of the instructions is converted to
a "real" PowerPC instruction. When a TOC entry needs to be created, this
is done here in the same manner as for the existing LDtoc, LDtocJTI, and
LDtocCPT pseudo-instructions (I factored out a new routine to handle this).
I had originally thought that if a TOC entry was needed for LDtocL or
ADDItocL, it would already have been generated for the previous ADDIStocHA.
However, at higher optimization levels, the ADDIStocHA may appear in a
different block, which may be assembled textually following the block
containing the LDtocL or ADDItocL. So it is necessary to include the
possibility of creating a new TOC entry for those two instructions.
Note that for LDtocL, we generate a new form of LD called LDrs. This
allows specifying the @toc@l relocation for the offset field of the LD
instruction (i.e., the offset is replaced by a SymbolLo relocation).
When the peephole optimization described above is added, we will need
to do similar things for all immediate-form load and store operations.
The seven "mcm-n.ll" test cases are kept separate because otherwise the
intermingling of various TOC entries and so forth makes the tests fragile
and hard to understand.
The above assumes use of an external assembler. For use of the
integrated assembler, new relocations are added and used by
PPCELFObjectWriter. Testing is done with "mcm-obj.ll", which tests for
proper generation of the various relocations for the same sequences
tested with the external assembler.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@168708 91177308-0d34-0410-b5e6-96231b3b80d8
2012-11-27 17:35:46 +00:00
|
|
|
LD_TOC_L,
|
|
|
|
|
|
|
|
/// G8RC = ADDI_TOC_L G8RReg, Symbol - For medium code model, produces
|
|
|
|
/// an ADDI8 instruction that adds G8RReg to sym@toc@l.
|
|
|
|
/// Preceded by an ADDIS_TOC_HA to form a full 32-bit offset.
|
|
|
|
ADDI_TOC_L
|
2006-01-27 23:34:02 +00:00
|
|
|
};
|
2006-03-20 06:15:45 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Define some predicates that are used for node matching.
|
|
|
|
namespace PPC {
|
2006-04-06 17:23:16 +00:00
|
|
|
/// isVPKUHUMShuffleMask - Return true if this is the shuffle mask for a
|
|
|
|
/// VPKUHUM instruction.
|
2009-04-27 18:41:29 +00:00
|
|
|
bool isVPKUHUMShuffleMask(ShuffleVectorSDNode *N, bool isUnary);
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-04-06 17:23:16 +00:00
|
|
|
/// isVPKUWUMShuffleMask - Return true if this is the shuffle mask for a
|
|
|
|
/// VPKUWUM instruction.
|
2009-04-27 18:41:29 +00:00
|
|
|
bool isVPKUWUMShuffleMask(ShuffleVectorSDNode *N, bool isUnary);
|
2006-04-06 21:11:54 +00:00
|
|
|
|
|
|
|
/// isVMRGLShuffleMask - Return true if this is a shuffle mask suitable for
|
|
|
|
/// a VRGL* instruction with the specified unit size (1,2 or 4 bytes).
|
2009-04-27 18:41:29 +00:00
|
|
|
bool isVMRGLShuffleMask(ShuffleVectorSDNode *N, unsigned UnitSize,
|
|
|
|
bool isUnary);
|
2006-04-06 21:11:54 +00:00
|
|
|
|
|
|
|
/// isVMRGHShuffleMask - Return true if this is a shuffle mask suitable for
|
|
|
|
/// a VRGH* instruction with the specified unit size (1,2 or 4 bytes).
|
2009-04-27 18:41:29 +00:00
|
|
|
bool isVMRGHShuffleMask(ShuffleVectorSDNode *N, unsigned UnitSize,
|
|
|
|
bool isUnary);
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-04-06 18:26:28 +00:00
|
|
|
/// isVSLDOIShuffleMask - If this is a vsldoi shuffle mask, return the shift
|
|
|
|
/// amount, otherwise return -1.
|
2006-04-06 22:28:36 +00:00
|
|
|
int isVSLDOIShuffleMask(SDNode *N, bool isUnary);
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-03-20 06:15:45 +00:00
|
|
|
/// isSplatShuffleMask - Return true if the specified VECTOR_SHUFFLE operand
|
|
|
|
/// specifies a splat of a single element that is suitable for input to
|
|
|
|
/// VSPLTB/VSPLTH/VSPLTW.
|
2009-04-27 18:41:29 +00:00
|
|
|
bool isSplatShuffleMask(ShuffleVectorSDNode *N, unsigned EltSize);
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2007-07-30 07:51:22 +00:00
|
|
|
/// isAllNegativeZeroVector - Returns true if all elements of build_vector
|
|
|
|
/// are -0.0.
|
|
|
|
bool isAllNegativeZeroVector(SDNode *N);
|
|
|
|
|
2006-03-20 06:15:45 +00:00
|
|
|
/// getVSPLTImmediate - Return the appropriate VSPLT* immediate to splat the
|
|
|
|
/// specified isSplatShuffleMask VECTOR_SHUFFLE mask.
|
2006-04-04 17:25:31 +00:00
|
|
|
unsigned getVSPLTImmediate(SDNode *N, unsigned EltSize);
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-04-12 17:37:20 +00:00
|
|
|
/// get_VSPLTI_elt - If this is a build_vector of constants which can be
|
2006-04-08 06:46:53 +00:00
|
|
|
/// formed by using a vspltis[bhw] instruction of the specified element
|
|
|
|
/// size, return the constant being splatted. The ByteSize field indicates
|
|
|
|
/// the number of bytes of each element [124] -> [bhw].
|
2008-07-27 21:46:04 +00:00
|
|
|
SDValue get_VSPLTI_elt(SDNode *N, unsigned ByteSize, SelectionDAG &DAG);
|
2006-03-20 06:15:45 +00:00
|
|
|
}
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-10-16 05:39:50 +00:00
|
|
|
class PPCTargetLowering : public TargetLowering {
|
2006-11-02 01:44:04 +00:00
|
|
|
const PPCSubtarget &PPCSubTarget;
|
2010-04-17 14:41:14 +00:00
|
|
|
|
2005-08-16 17:14:42 +00:00
|
|
|
public:
|
2007-08-02 21:21:54 +00:00
|
|
|
explicit PPCTargetLowering(PPCTargetMachine &TM);
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-01-09 23:52:17 +00:00
|
|
|
/// getTargetNodeName() - This method returns the name of a target specific
|
|
|
|
/// DAG node.
|
|
|
|
virtual const char *getTargetNodeName(unsigned Opcode) const;
|
2006-11-08 02:15:41 +00:00
|
|
|
|
2013-03-01 18:40:30 +00:00
|
|
|
virtual MVT getScalarShiftAmountTy(EVT LHSTy) const { return MVT::i32; }
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2008-03-10 15:42:14 +00:00
|
|
|
/// getSetCCResultType - Return the ISD::SETCC ValueType
|
2011-09-06 19:07:46 +00:00
|
|
|
virtual EVT getSetCCResultType(EVT VT) const;
|
2008-03-10 15:42:14 +00:00
|
|
|
|
2006-11-08 02:15:41 +00:00
|
|
|
/// getPreIndexedAddressParts - returns true by value, base pointer and
|
|
|
|
/// offset pointer and addressing mode by reference if the node's address
|
|
|
|
/// can be legally represented as pre-indexed load / store address.
|
2008-07-27 21:46:04 +00:00
|
|
|
virtual bool getPreIndexedAddressParts(SDNode *N, SDValue &Base,
|
|
|
|
SDValue &Offset,
|
2006-11-09 17:55:04 +00:00
|
|
|
ISD::MemIndexedMode &AM,
|
2009-01-15 16:29:45 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-11-08 02:15:41 +00:00
|
|
|
/// SelectAddressRegReg - Given the specified addressed, check to see if it
|
|
|
|
/// can be represented as an indexed [r+r] operation. Returns false if it
|
|
|
|
/// can be more efficiently represented with [r+imm].
|
2008-07-27 21:46:04 +00:00
|
|
|
bool SelectAddressRegReg(SDValue N, SDValue &Base, SDValue &Index,
|
2009-01-15 16:29:45 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-11-08 02:15:41 +00:00
|
|
|
/// SelectAddressRegImm - Returns true if the address N can be represented
|
|
|
|
/// by a base register plus a signed 16-bit displacement [r+imm], and if it
|
|
|
|
/// is not better represented as reg+reg.
|
2008-07-27 21:46:04 +00:00
|
|
|
bool SelectAddressRegImm(SDValue N, SDValue &Disp, SDValue &Base,
|
2009-01-15 16:29:45 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-11-08 02:15:41 +00:00
|
|
|
/// SelectAddressRegRegOnly - Given the specified addressed, force it to be
|
|
|
|
/// represented as an indexed [r+r] operation.
|
2008-07-27 21:46:04 +00:00
|
|
|
bool SelectAddressRegRegOnly(SDValue N, SDValue &Base, SDValue &Index,
|
2009-01-15 16:29:45 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2006-11-08 02:15:41 +00:00
|
|
|
|
|
|
|
/// SelectAddressRegImmShift - Returns true if the address N can be
|
|
|
|
/// represented by a base register plus a signed 14-bit displacement
|
|
|
|
/// [r+imm*4]. Suitable for use by STD and friends.
|
2008-07-27 21:46:04 +00:00
|
|
|
bool SelectAddressRegImmShift(SDValue N, SDValue &Disp, SDValue &Base,
|
2009-01-15 16:29:45 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2006-11-08 02:15:41 +00:00
|
|
|
|
2012-04-01 19:23:08 +00:00
|
|
|
Sched::Preference getSchedulingPreference(SDNode *N) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2005-08-26 00:52:45 +00:00
|
|
|
/// LowerOperation - Provide custom lowering hooks for some operations.
|
|
|
|
///
|
2010-04-17 15:26:15 +00:00
|
|
|
virtual SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const;
|
2007-11-28 18:44:47 +00:00
|
|
|
|
2008-12-01 11:39:25 +00:00
|
|
|
/// ReplaceNodeResults - Replace the results of node with an illegal result
|
|
|
|
/// type with new values built out of custom code.
|
|
|
|
///
|
|
|
|
virtual void ReplaceNodeResults(SDNode *N, SmallVectorImpl<SDValue>&Results,
|
2010-04-17 15:26:15 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2008-12-01 11:39:25 +00:00
|
|
|
|
2008-07-27 21:46:04 +00:00
|
|
|
virtual SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2008-07-27 21:46:04 +00:00
|
|
|
virtual void computeMaskedBitsForTargetNode(const SDValue Op,
|
2011-02-25 21:41:48 +00:00
|
|
|
APInt &KnownZero,
|
2008-02-13 00:35:47 +00:00
|
|
|
APInt &KnownOne,
|
2007-06-22 14:59:07 +00:00
|
|
|
const SelectionDAG &DAG,
|
2006-04-02 06:26:07 +00:00
|
|
|
unsigned Depth = 0) const;
|
2005-10-18 23:23:37 +00:00
|
|
|
|
2010-05-01 00:01:06 +00:00
|
|
|
virtual MachineBasicBlock *
|
|
|
|
EmitInstrWithCustomInserter(MachineInstr *MI,
|
|
|
|
MachineBasicBlock *MBB) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
MachineBasicBlock *EmitAtomicBinary(MachineInstr *MI,
|
2008-08-25 22:34:37 +00:00
|
|
|
MachineBasicBlock *MBB, bool is64Bit,
|
2009-02-07 16:15:20 +00:00
|
|
|
unsigned BinOpcode) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
MachineBasicBlock *EmitPartwordAtomicBinary(MachineInstr *MI,
|
|
|
|
MachineBasicBlock *MBB,
|
2009-02-07 16:15:20 +00:00
|
|
|
bool is8bit, unsigned Opcode) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2007-03-25 02:14:49 +00:00
|
|
|
ConstraintType getConstraintType(const std::string &Constraint) const;
|
2010-10-29 17:29:13 +00:00
|
|
|
|
|
|
|
/// Examine constraint string and operand type and determine a weight value.
|
|
|
|
/// The operand object must already have been set up with the operand type.
|
|
|
|
ConstraintWeight getSingleConstraintMatchWeight(
|
|
|
|
AsmOperandInfo &info, const char *constraint) const;
|
|
|
|
|
2011-02-25 21:41:48 +00:00
|
|
|
std::pair<unsigned, const TargetRegisterClass*>
|
2006-11-02 01:44:04 +00:00
|
|
|
getRegForInlineAsmConstraint(const std::string &Constraint,
|
2009-08-10 22:56:29 +00:00
|
|
|
EVT VT) const;
|
2006-03-13 23:20:37 +00:00
|
|
|
|
2008-02-28 22:31:51 +00:00
|
|
|
/// getByValTypeAlignment - Return the desired alignment for ByVal aggregate
|
|
|
|
/// function arguments in the caller parameter area. This is the actual
|
|
|
|
/// alignment, not its logarithm.
|
2011-07-18 04:54:35 +00:00
|
|
|
unsigned getByValTypeAlignment(Type *Ty) const;
|
2008-02-28 22:31:51 +00:00
|
|
|
|
2007-08-25 00:47:38 +00:00
|
|
|
/// LowerAsmOperandForConstraint - Lower the specified operand into the Ops
|
2010-06-25 21:55:36 +00:00
|
|
|
/// vector. If it is invalid, don't add anything to Ops.
|
2008-07-27 21:46:04 +00:00
|
|
|
virtual void LowerAsmOperandForConstraint(SDValue Op,
|
2011-06-02 23:16:42 +00:00
|
|
|
std::string &Constraint,
|
2008-07-27 21:46:04 +00:00
|
|
|
std::vector<SDValue> &Ops,
|
2008-04-26 23:02:14 +00:00
|
|
|
SelectionDAG &DAG) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2007-03-30 23:15:24 +00:00
|
|
|
/// isLegalAddressingMode - Return true if the addressing mode represented
|
|
|
|
/// by AM is legal for this target, for a load/store of the specified type.
|
2011-07-18 04:54:35 +00:00
|
|
|
virtual bool isLegalAddressingMode(const AddrMode &AM, Type *Ty)const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2006-03-13 23:20:37 +00:00
|
|
|
/// isLegalAddressImmediate - Return true if the integer value can be used
|
2007-03-12 23:29:01 +00:00
|
|
|
/// as the offset of the target addressing mode for load / store of the
|
|
|
|
/// given type.
|
2011-07-18 04:54:35 +00:00
|
|
|
virtual bool isLegalAddressImmediate(int64_t V, Type *Ty) const;
|
2007-03-12 23:29:01 +00:00
|
|
|
|
|
|
|
/// isLegalAddressImmediate - Return true if the GlobalValue can be used as
|
|
|
|
/// the offset of the target addressing mode.
|
|
|
|
virtual bool isLegalAddressImmediate(GlobalValue *GV) const;
|
2007-03-01 13:11:38 +00:00
|
|
|
|
2008-10-21 03:41:46 +00:00
|
|
|
virtual bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const;
|
2011-02-25 21:41:48 +00:00
|
|
|
|
2010-04-01 20:10:42 +00:00
|
|
|
/// getOptimalMemOpType - Returns the target specific optimal type for load
|
2010-04-02 19:36:14 +00:00
|
|
|
/// and store operations as a result of memset, memcpy, and memmove
|
|
|
|
/// lowering. If DstAlign is zero that means it's safe to destination
|
|
|
|
/// alignment can satisfy any constraint. Similarly if SrcAlign is zero it
|
|
|
|
/// means there isn't a need to check it against alignment requirement,
|
2012-12-12 02:34:41 +00:00
|
|
|
/// probably because the source does not need to be loaded. If 'IsMemset' is
|
|
|
|
/// true, that means it's expanding a memset. If 'ZeroMemset' is true, that
|
|
|
|
/// means it's a memset of zero. 'MemcpyStrSrc' indicates whether the memcpy
|
|
|
|
/// source is constant so it does not need to be loaded.
|
2010-04-16 20:11:05 +00:00
|
|
|
/// It returns EVT::Other if the type should be determined using generic
|
|
|
|
/// target-independent logic.
|
2010-04-02 19:36:14 +00:00
|
|
|
virtual EVT
|
2012-12-12 02:34:41 +00:00
|
|
|
getOptimalMemOpType(uint64_t Size, unsigned DstAlign, unsigned SrcAlign,
|
|
|
|
bool IsMemset, bool ZeroMemset, bool MemcpyStrSrc,
|
2010-04-16 20:11:05 +00:00
|
|
|
MachineFunction &MF) const;
|
2008-10-21 03:41:46 +00:00
|
|
|
|
2013-03-15 15:27:13 +00:00
|
|
|
/// Is unaligned memory access allowed for the given type, and is it fast
|
|
|
|
/// relative to software emulation.
|
|
|
|
virtual bool allowsUnalignedMemoryAccesses(EVT VT, bool *Fast = 0) const;
|
|
|
|
|
2012-06-22 00:49:52 +00:00
|
|
|
/// isFMAFasterThanMulAndAdd - Return true if an FMA operation is faster than
|
|
|
|
/// a pair of mul and add instructions. fmuladd intrinsics will be expanded to
|
|
|
|
/// FMAs when this method returns true (and FMAs are legal), otherwise fmuladd
|
|
|
|
/// is expanded to mul + add.
|
|
|
|
virtual bool isFMAFasterThanMulAndAdd(EVT VT) const;
|
|
|
|
|
2008-04-19 01:30:48 +00:00
|
|
|
private:
|
2008-07-27 21:46:04 +00:00
|
|
|
SDValue getFramePointerFrameIndex(SelectionDAG & DAG) const;
|
|
|
|
SDValue getReturnAddrFrameIndex(SelectionDAG & DAG) const;
|
2008-04-30 09:16:33 +00:00
|
|
|
|
2010-01-27 00:07:07 +00:00
|
|
|
bool
|
|
|
|
IsEligibleForTailCallOptimization(SDValue Callee,
|
|
|
|
CallingConv::ID CalleeCC,
|
|
|
|
bool isVarArg,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
SelectionDAG& DAG) const;
|
|
|
|
|
2008-07-27 21:46:04 +00:00
|
|
|
SDValue EmitTailCallLoadFPAndRetAddr(SelectionDAG & DAG,
|
2009-02-04 20:06:27 +00:00
|
|
|
int SPDiff,
|
|
|
|
SDValue Chain,
|
|
|
|
SDValue &LROpOut,
|
|
|
|
SDValue &FPOpOut,
|
2009-07-03 06:47:08 +00:00
|
|
|
bool isDarwinABI,
|
2010-04-17 15:26:15 +00:00
|
|
|
DebugLoc dl) const;
|
|
|
|
|
|
|
|
SDValue LowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerFRAMEADDR(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerConstantPool(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerBlockAddress(SDValue Op, SelectionDAG &DAG) const;
|
2012-06-04 17:36:38 +00:00
|
|
|
SDValue LowerGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const;
|
2010-04-17 15:26:15 +00:00
|
|
|
SDValue LowerGlobalAddress(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerJumpTable(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerSETCC(SDValue Op, SelectionDAG &DAG) const;
|
2011-09-06 13:37:06 +00:00
|
|
|
SDValue LowerINIT_TRAMPOLINE(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerADJUST_TRAMPOLINE(SDValue Op, SelectionDAG &DAG) const;
|
2008-07-27 21:46:04 +00:00
|
|
|
SDValue LowerVASTART(SDValue Op, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
const PPCSubtarget &Subtarget) const;
|
2010-04-17 14:41:14 +00:00
|
|
|
SDValue LowerVAARG(SDValue Op, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
const PPCSubtarget &Subtarget) const;
|
2008-07-27 21:46:04 +00:00
|
|
|
SDValue LowerSTACKRESTORE(SDValue Op, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
const PPCSubtarget &Subtarget) const;
|
2008-07-27 21:46:04 +00:00
|
|
|
SDValue LowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
const PPCSubtarget &Subtarget) const;
|
|
|
|
SDValue LowerSELECT_CC(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerFP_TO_INT(SDValue Op, SelectionDAG &DAG, DebugLoc dl) const;
|
|
|
|
SDValue LowerSINT_TO_FP(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerFLT_ROUNDS_(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerSHL_PARTS(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerSRL_PARTS(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerSRA_PARTS(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerSCALAR_TO_VECTOR(SDValue Op, SelectionDAG &DAG) const;
|
|
|
|
SDValue LowerMUL(SDValue Op, SelectionDAG &DAG) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
|
|
|
|
SDValue LowerCallResult(SDValue Chain, SDValue InFlag,
|
2009-09-02 08:44:58 +00:00
|
|
|
CallingConv::ID CallConv, bool isVarArg,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
2009-09-02 08:44:58 +00:00
|
|
|
SDValue FinishCall(CallingConv::ID CallConv, DebugLoc dl, bool isTailCall,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
bool isVarArg,
|
|
|
|
SelectionDAG &DAG,
|
|
|
|
SmallVector<std::pair<unsigned, SDValue>, 8>
|
|
|
|
&RegsToPass,
|
|
|
|
SDValue InFlag, SDValue Chain,
|
|
|
|
SDValue &Callee,
|
|
|
|
int SPDiff, unsigned NumBytes,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
2010-04-17 15:26:15 +00:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
|
|
|
|
virtual SDValue
|
|
|
|
LowerFormalArguments(SDValue Chain,
|
2009-09-02 08:44:58 +00:00
|
|
|
CallingConv::ID CallConv, bool isVarArg,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
|
|
|
|
virtual SDValue
|
2012-05-25 16:35:28 +00:00
|
|
|
LowerCall(TargetLowering::CallLoweringInfo &CLI,
|
2010-04-17 15:26:15 +00:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
|
2011-10-14 19:51:36 +00:00
|
|
|
virtual bool
|
|
|
|
CanLowerReturn(CallingConv::ID CallConv, MachineFunction &MF,
|
|
|
|
bool isVarArg,
|
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
|
|
|
LLVMContext &Context) const;
|
|
|
|
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
virtual SDValue
|
|
|
|
LowerReturn(SDValue Chain,
|
2009-09-02 08:44:58 +00:00
|
|
|
CallingConv::ID CallConv, bool isVarArg,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
2010-07-07 15:54:55 +00:00
|
|
|
const SmallVectorImpl<SDValue> &OutVals,
|
2010-04-17 15:26:15 +00:00
|
|
|
DebugLoc dl, SelectionDAG &DAG) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
|
2012-10-23 15:51:16 +00:00
|
|
|
SDValue
|
|
|
|
extendArgForPPC64(ISD::ArgFlagsTy Flags, EVT ObjectVT, SelectionDAG &DAG,
|
|
|
|
SDValue ArgVal, DebugLoc dl) const;
|
|
|
|
|
|
|
|
void
|
|
|
|
setMinReservedArea(MachineFunction &MF, SelectionDAG &DAG,
|
|
|
|
unsigned nAltivecParamsAtEnd,
|
|
|
|
unsigned MinReservedArea, bool isPPC64) const;
|
|
|
|
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
SDValue
|
2012-10-05 21:27:08 +00:00
|
|
|
LowerFormalArguments_Darwin(SDValue Chain,
|
|
|
|
CallingConv::ID CallConv, bool isVarArg,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
|
|
|
SDValue
|
|
|
|
LowerFormalArguments_64SVR4(SDValue Chain,
|
2009-09-02 08:44:58 +00:00
|
|
|
CallingConv::ID CallConv, bool isVarArg,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
SDValue
|
2012-09-19 15:42:13 +00:00
|
|
|
LowerFormalArguments_32SVR4(SDValue Chain,
|
|
|
|
CallingConv::ID CallConv, bool isVarArg,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
|
|
|
|
SDValue
|
2012-10-23 15:51:16 +00:00
|
|
|
createMemcpyOutsideCallSeq(SDValue Arg, SDValue PtrOff,
|
|
|
|
SDValue CallSeqStart, ISD::ArgFlagsTy Flags,
|
|
|
|
SelectionDAG &DAG, DebugLoc dl) const;
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
LowerCall_Darwin(SDValue Chain, SDValue Callee,
|
|
|
|
CallingConv::ID CallConv,
|
|
|
|
bool isVarArg, bool isTailCall,
|
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
|
|
|
const SmallVectorImpl<SDValue> &OutVals,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
|
|
|
SDValue
|
|
|
|
LowerCall_64SVR4(SDValue Chain, SDValue Callee,
|
2012-09-19 15:42:13 +00:00
|
|
|
CallingConv::ID CallConv,
|
2012-02-28 18:51:51 +00:00
|
|
|
bool isVarArg, bool isTailCall,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
2010-07-07 15:54:55 +00:00
|
|
|
const SmallVectorImpl<SDValue> &OutVals,
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
2010-04-17 15:26:15 +00:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
Major calling convention code refactoring.
Instead of awkwardly encoding calling-convention information with ISD::CALL,
ISD::FORMAL_ARGUMENTS, ISD::RET, and ISD::ARG_FLAGS nodes, TargetLowering
provides three virtual functions for targets to override:
LowerFormalArguments, LowerCall, and LowerRet, which replace the custom
lowering done on the special nodes. They provide the same information, but
in a more immediately usable format.
This also reworks much of the target-independent tail call logic. The
decision of whether or not to perform a tail call is now cleanly split
between target-independent portions, and the target dependent portion
in IsEligibleForTailCallOptimization.
This also synchronizes all in-tree targets, to help enable future
refactoring and feature work.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@78142 91177308-0d34-0410-b5e6-96231b3b80d8
2009-08-05 01:29:28 +00:00
|
|
|
SDValue
|
2012-09-19 15:42:13 +00:00
|
|
|
LowerCall_32SVR4(SDValue Chain, SDValue Callee, CallingConv::ID CallConv,
|
|
|
|
bool isVarArg, bool isTailCall,
|
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
|
|
|
const SmallVectorImpl<SDValue> &OutVals,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
DebugLoc dl, SelectionDAG &DAG,
|
|
|
|
SmallVectorImpl<SDValue> &InVals) const;
|
2005-08-16 17:14:42 +00:00
|
|
|
};
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif // LLVM_TARGET_POWERPC_PPC32ISELLOWERING_H
|