mirror of
https://github.com/RPCSX/llvm.git
synced 2024-12-11 21:57:55 +00:00
Update README file.
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@30244 91177308-0d34-0410-b5e6-96231b3b80d8
This commit is contained in:
parent
23f8497724
commit
0f4aa6ee20
@ -80,15 +80,6 @@ allocator. Delay codegen until post register allocation.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
Model X86 EFLAGS as a real register to avoid redudant cmp / test. e.g.
|
||||
|
||||
cmpl $1, %eax
|
||||
setg %al
|
||||
testb %al, %al # unnecessary
|
||||
jne .BB7
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
Count leading zeros and count trailing zeros:
|
||||
|
||||
int clz(int X) { return __builtin_clz(X); }
|
||||
@ -126,6 +117,8 @@ commutative, it is not matched with the load on both sides. The dag combiner
|
||||
should be made smart enough to cannonicalize the load into the RHS of a compare
|
||||
when it can invert the result of the compare for free.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
How about intrinsics? An example is:
|
||||
*res = _mm_mulhi_epu16(*A, _mm_mul_epu32(*B, *C));
|
||||
|
||||
@ -140,51 +133,6 @@ target specific hook.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
The DAG Isel doesn't fold the loads into the adds in this testcase. The
|
||||
pattern selector does. This is because the chain value of the load gets
|
||||
selected first, and the loads aren't checking to see if they are only used by
|
||||
and add.
|
||||
|
||||
.ll:
|
||||
|
||||
int %test(int* %x, int* %y, int* %z) {
|
||||
%X = load int* %x
|
||||
%Y = load int* %y
|
||||
%Z = load int* %z
|
||||
%a = add int %X, %Y
|
||||
%b = add int %a, %Z
|
||||
ret int %b
|
||||
}
|
||||
|
||||
dag isel:
|
||||
|
||||
_test:
|
||||
movl 4(%esp), %eax
|
||||
movl (%eax), %eax
|
||||
movl 8(%esp), %ecx
|
||||
movl (%ecx), %ecx
|
||||
addl %ecx, %eax
|
||||
movl 12(%esp), %ecx
|
||||
movl (%ecx), %ecx
|
||||
addl %ecx, %eax
|
||||
ret
|
||||
|
||||
pattern isel:
|
||||
|
||||
_test:
|
||||
movl 12(%esp), %ecx
|
||||
movl 4(%esp), %edx
|
||||
movl 8(%esp), %eax
|
||||
movl (%eax), %eax
|
||||
addl (%edx), %eax
|
||||
addl (%ecx), %eax
|
||||
ret
|
||||
|
||||
This is bad for register pressure, though the dag isel is producing a
|
||||
better schedule. :)
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
In many cases, LLVM generates code like this:
|
||||
|
||||
_test:
|
||||
@ -198,7 +146,7 @@ on some processors (which ones?), it is more efficient to do this:
|
||||
|
||||
_test:
|
||||
movl 8(%esp), %ebx
|
||||
xor %eax, %eax
|
||||
xor %eax, %eax
|
||||
cmpl %ebx, 4(%esp)
|
||||
setl %al
|
||||
ret
|
||||
@ -207,38 +155,6 @@ Doing this correctly is tricky though, as the xor clobbers the flags.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
We should generate 'test' instead of 'cmp' in various cases, e.g.:
|
||||
|
||||
bool %test(int %X) {
|
||||
%Y = shl int %X, ubyte 1
|
||||
%C = seteq int %Y, 0
|
||||
ret bool %C
|
||||
}
|
||||
bool %test(int %X) {
|
||||
%Y = and int %X, 8
|
||||
%C = seteq int %Y, 0
|
||||
ret bool %C
|
||||
}
|
||||
|
||||
This may just be a matter of using 'test' to write bigger patterns for X86cmp.
|
||||
|
||||
An important case is comparison against zero:
|
||||
|
||||
if (X == 0) ...
|
||||
|
||||
instead of:
|
||||
|
||||
cmpl $0, %eax
|
||||
je LBB4_2 #cond_next
|
||||
|
||||
use:
|
||||
test %eax, %eax
|
||||
jz LBB4_2
|
||||
|
||||
which is smaller.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
We should generate bts/btr/etc instructions on targets where they are cheap or
|
||||
when codesize is important. e.g., for:
|
||||
|
||||
@ -564,17 +480,6 @@ nicer, while still handling the hard cases.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
Some ideas for instruction selection code simplification: 1. A pre-pass to
|
||||
determine which chain producing node can or cannot be folded. The generated
|
||||
isel code would then use the information. 2. The same pre-pass can force
|
||||
ordering of TokenFactor operands to allow load / store folding. 3. During isel,
|
||||
instead of recursively going up the chain operand chain, mark the chain operand
|
||||
as available and put it in some work list. Select other nodes in the normal
|
||||
manner. The chain operands are selected after all other nodes are selected. Uses
|
||||
of chain nodes are modified after instruction selection is completed.
|
||||
|
||||
//===---------------------------------------------------------------------===//
|
||||
|
||||
Another instruction selector deficiency:
|
||||
|
||||
void %bar() {
|
||||
|
Loading…
Reference in New Issue
Block a user