llvm-mirror/test/CodeGen/X86/test-pic-jtbl.ll
Dale Johannesen fd7f1ad2f7 Implement smarter algorithm for choosing which blocks to tail-merge.
See test/CodeGen/X86/test-pic-jtbl.ll for a case where it works well;
shaves another 10K off our favorite benchmark.  I was hesitant about
this because of compile speed, but seems to do OK on a bootstrap.

llvm-svn: 37392
2007-06-01 23:02:45 +00:00

59 lines
1.4 KiB
LLVM

; RUN: llvm-as < %s | llc -mtriple=i686-pc-linux-gnu -relocation-model=pic \
; RUN: -o %t -f
; RUN: grep _GLOBAL_OFFSET_TABLE_ %t
; RUN: grep piclabel %t | wc -l | grep 3
; RUN: grep PLT %t | wc -l | grep 6
; RUN: grep GOTOFF %t | wc -l | grep 1
; RUN: grep JTI %t | wc -l | grep 8
define void @bar(i32 %n.u) {
entry:
switch i32 %n.u, label %bb12 [i32 1, label %bb i32 2, label %bb6 i32 4, label %bb7 i32 5, label %bb8 i32 6, label %bb10 i32 7, label %bb1 i32 8, label %bb3 i32 9, label %bb4 i32 10, label %bb9 i32 11, label %bb2 i32 12, label %bb5 i32 13, label %bb11 ]
bb:
tail call void(...)* @foo1()
ret void
bb1:
tail call void(...)* @foo2()
ret void
bb2:
tail call void(...)* @foo6()
ret void
bb3:
tail call void(...)* @foo3()
ret void
bb4:
tail call void(...)* @foo4()
ret void
bb5:
tail call void(...)* @foo5()
ret void
bb6:
tail call void(...)* @foo1()
ret void
bb7:
tail call void(...)* @foo2()
ret void
bb8:
tail call void(...)* @foo6()
ret void
bb9:
tail call void(...)* @foo3()
ret void
bb10:
tail call void(...)* @foo4()
ret void
bb11:
tail call void(...)* @foo5()
ret void
bb12:
tail call void(...)* @foo6()
ret void
}
declare void @foo1(...)
declare void @foo2(...)
declare void @foo6(...)
declare void @foo3(...)
declare void @foo4(...)
declare void @foo5(...)