2010-08-03 18:20:32 +00:00
|
|
|
; RUN: opt < %s -instcombine -S | FileCheck %s
|
2013-09-03 21:05:48 +00:00
|
|
|
target datalayout = "E-p:64:64:64-p1:32:32:32-a0:0:8-f32:32:32-f64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-v64:64:64-v128:128:128"
|
2008-04-10 18:43:06 +00:00
|
|
|
|
|
|
|
; Instcombine should be able to prove vector alignment in the
|
|
|
|
; presence of a few mild address computation tricks.
|
|
|
|
|
2013-07-14 01:42:54 +00:00
|
|
|
; CHECK-LABEL: @test0(
|
2010-08-03 18:20:32 +00:00
|
|
|
; CHECK: align 16
|
|
|
|
|
|
|
|
define void @test0(i8* %b, i64 %n, i64 %u, i64 %y) nounwind {
|
2008-04-10 18:43:06 +00:00
|
|
|
entry:
|
|
|
|
%c = ptrtoint i8* %b to i64
|
|
|
|
%d = and i64 %c, -16
|
|
|
|
%e = inttoptr i64 %d to double*
|
|
|
|
%v = mul i64 %u, 2
|
|
|
|
%z = and i64 %y, -2
|
|
|
|
%t1421 = icmp eq i64 %n, 0
|
|
|
|
br i1 %t1421, label %return, label %bb
|
|
|
|
|
|
|
|
bb:
|
|
|
|
%i = phi i64 [ %indvar.next, %bb ], [ 20, %entry ]
|
|
|
|
%j = mul i64 %i, %v
|
|
|
|
%h = add i64 %j, %z
|
|
|
|
%t8 = getelementptr double* %e, i64 %h
|
|
|
|
%p = bitcast double* %t8 to <2 x double>*
|
|
|
|
store <2 x double><double 0.0, double 0.0>, <2 x double>* %p, align 8
|
|
|
|
%indvar.next = add i64 %i, 1
|
|
|
|
%exitcond = icmp eq i64 %indvar.next, %n
|
|
|
|
br i1 %exitcond, label %return, label %bb
|
|
|
|
|
|
|
|
return:
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2010-08-03 18:20:32 +00:00
|
|
|
; When we see a unaligned load from an insufficiently aligned global or
|
|
|
|
; alloca, increase the alignment of the load, turning it into an aligned load.
|
|
|
|
|
2013-07-14 01:42:54 +00:00
|
|
|
; CHECK-LABEL: @test1(
|
2010-08-03 18:20:32 +00:00
|
|
|
; CHECK: tmp = load
|
|
|
|
; CHECK: GLOBAL{{.*}}align 16
|
|
|
|
|
|
|
|
@GLOBAL = internal global [4 x i32] zeroinitializer
|
|
|
|
|
|
|
|
define <16 x i8> @test1(<2 x i64> %x) {
|
|
|
|
entry:
|
|
|
|
%tmp = load <16 x i8>* bitcast ([4 x i32]* @GLOBAL to <16 x i8>*), align 1
|
|
|
|
ret <16 x i8> %tmp
|
|
|
|
}
|
|
|
|
|
2013-09-03 21:05:48 +00:00
|
|
|
@GLOBAL_as1 = internal addrspace(1) global [4 x i32] zeroinitializer
|
|
|
|
|
|
|
|
define <16 x i8> @test1_as1(<2 x i64> %x) {
|
|
|
|
; CHECK-LABEL: @test1_as1(
|
|
|
|
; CHECK: tmp = load
|
|
|
|
; CHECK: GLOBAL_as1{{.*}}align 16
|
|
|
|
%tmp = load <16 x i8> addrspace(1)* bitcast ([4 x i32] addrspace(1)* @GLOBAL_as1 to <16 x i8> addrspace(1)*), align 1
|
|
|
|
ret <16 x i8> %tmp
|
|
|
|
}
|
|
|
|
|
|
|
|
@GLOBAL_as1_gep = internal addrspace(1) global [8 x i32] zeroinitializer
|
|
|
|
|
|
|
|
define <16 x i8> @test1_as1_gep(<2 x i64> %x) {
|
|
|
|
; CHECK-LABEL: @test1_as1_gep(
|
|
|
|
; CHECK: tmp = load
|
|
|
|
; CHECK: GLOBAL_as1_gep{{.*}}align 16
|
|
|
|
%tmp = load <16 x i8> addrspace(1)* bitcast (i32 addrspace(1)* getelementptr ([8 x i32] addrspace(1)* @GLOBAL_as1_gep, i16 0, i16 4) to <16 x i8> addrspace(1)*), align 1
|
|
|
|
ret <16 x i8> %tmp
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2010-08-03 18:20:32 +00:00
|
|
|
; When a load or store lacks an explicit alignment, add one.
|
|
|
|
|
2013-07-14 01:42:54 +00:00
|
|
|
; CHECK-LABEL: @test2(
|
2010-08-03 18:20:32 +00:00
|
|
|
; CHECK: load double* %p, align 8
|
|
|
|
; CHECK: store double %n, double* %p, align 8
|
|
|
|
|
|
|
|
define double @test2(double* %p, double %n) nounwind {
|
|
|
|
%t = load double* %p
|
|
|
|
store double %n, double* %p
|
|
|
|
ret double %t
|
|
|
|
}
|
2012-10-04 13:36:31 +00:00
|
|
|
|
|
|
|
declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind
|
|
|
|
|
|
|
|
declare void @use(i8*)
|
|
|
|
|
|
|
|
%struct.s = type { i32, i32, i32, i32 }
|
|
|
|
|
|
|
|
define void @test3(%struct.s* sret %a4) {
|
|
|
|
; Check that the alignment is bumped up the alignment of the sret type.
|
2013-07-14 01:42:54 +00:00
|
|
|
; CHECK-LABEL: @test3(
|
2012-10-04 13:36:31 +00:00
|
|
|
%a4.cast = bitcast %struct.s* %a4 to i8*
|
|
|
|
call void @llvm.memset.p0i8.i64(i8* %a4.cast, i8 0, i64 16, i32 1, i1 false)
|
|
|
|
; CHECK: call void @llvm.memset.p0i8.i64(i8* %a4.cast, i8 0, i64 16, i32 4, i1 false)
|
|
|
|
call void @use(i8* %a4.cast)
|
|
|
|
ret void
|
|
|
|
}
|