; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt < %s -basic-aa -dse -S | FileCheck %s ; RUN: opt < %s -aa-pipeline=basic-aa -passes=dse -S | FileCheck %s target datalayout = "E-p:64:64:64-a0:0:8-f32:32:32-f64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-v64:64:64-v128:128:128" declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i1) nounwind declare void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* nocapture, i8, i64, i32) nounwind declare void @llvm.memcpy.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i1) nounwind declare void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* nocapture, i8* nocapture, i64, i32) nounwind declare void @llvm.init.trampoline(i8*, i8*, i8*) define void @test1(i32* %Q, i32* %P) { ; CHECK-LABEL: @test1( ; CHECK-NEXT: store i32 0, i32* [[P:%.*]] ; CHECK-NEXT: ret void ; %DEAD = load i32, i32* %Q store i32 %DEAD, i32* %P store i32 0, i32* %P ret void } ; PR8576 - Should delete store of 10 even though p/q are may aliases. define void @test2(i32 *%p, i32 *%q) { ; CHECK-LABEL: @test2( ; CHECK-NEXT: store i32 20, i32* [[Q:%.*]], align 4 ; CHECK-NEXT: store i32 30, i32* [[P:%.*]], align 4 ; CHECK-NEXT: ret void ; store i32 10, i32* %p, align 4 store i32 20, i32* %q, align 4 store i32 30, i32* %p, align 4 ret void } ; PR8677 @g = global i32 1 define i32 @test3(i32* %g_addr) nounwind { ; CHECK-LABEL: @test3( ; CHECK-NEXT: [[G_VALUE:%.*]] = load i32, i32* [[G_ADDR:%.*]], align 4 ; CHECK-NEXT: store i32 -1, i32* @g, align 4 ; CHECK-NEXT: store i32 [[G_VALUE]], i32* [[G_ADDR]], align 4 ; CHECK-NEXT: [[TMP3:%.*]] = load i32, i32* @g, align 4 ; CHECK-NEXT: ret i32 [[TMP3]] ; %g_value = load i32, i32* %g_addr, align 4 store i32 -1, i32* @g, align 4 store i32 %g_value, i32* %g_addr, align 4 %tmp3 = load i32, i32* @g, align 4 ret i32 %tmp3 } define void @test4(i32* %Q) { ; CHECK-LABEL: @test4( ; CHECK-NEXT: [[A:%.*]] = load i32, i32* [[Q:%.*]], align 4 ; CHECK-NEXT: store volatile i32 [[A]], i32* [[Q]] ; CHECK-NEXT: ret void ; %a = load i32, i32* %Q store volatile i32 %a, i32* %Q ret void } define void @test5(i32* %Q) { ; CHECK-LABEL: @test5( ; CHECK-NEXT: [[A:%.*]] = load volatile i32, i32* [[Q:%.*]], align 4 ; CHECK-NEXT: ret void ; %a = load volatile i32, i32* %Q store i32 %a, i32* %Q ret void } ; Should delete store of 10 even though memset is a may-store to P (P and Q may ; alias). define void @test6(i32 *%p, i8 *%q) { ; CHECK-LABEL: @test6( ; CHECK-NEXT: call void @llvm.memset.p0i8.i64(i8* [[Q:%.*]], i8 42, i64 900, i1 false) ; CHECK-NEXT: store i32 30, i32* [[P:%.*]], align 4 ; CHECK-NEXT: ret void ; store i32 10, i32* %p, align 4 ;; dead. call void @llvm.memset.p0i8.i64(i8* %q, i8 42, i64 900, i1 false) store i32 30, i32* %p, align 4 ret void } ; Should delete store of 10 even though memset is a may-store to P (P and Q may ; alias). define void @test6_atomic(i32* align 4 %p, i8* align 4 %q) { ; CHECK-LABEL: @test6_atomic( ; CHECK-NEXT: call void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* align 4 [[Q:%.*]], i8 42, i64 900, i32 4) ; CHECK-NEXT: store atomic i32 30, i32* [[P:%.*]] unordered, align 4 ; CHECK-NEXT: ret void ; store atomic i32 10, i32* %p unordered, align 4 ;; dead. call void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* align 4 %q, i8 42, i64 900, i32 4) store atomic i32 30, i32* %p unordered, align 4 ret void } ; Should delete store of 10 even though memcpy is a may-store to P (P and Q may ; alias). define void @test7(i32 *%p, i8 *%q, i8* noalias %r) { ; CHECK-LABEL: @test7( ; CHECK-NEXT: call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[Q:%.*]], i8* [[R:%.*]], i64 900, i1 false) ; CHECK-NEXT: store i32 30, i32* [[P:%.*]], align 4 ; CHECK-NEXT: ret void ; store i32 10, i32* %p, align 4 ;; dead. call void @llvm.memcpy.p0i8.p0i8.i64(i8* %q, i8* %r, i64 900, i1 false) store i32 30, i32* %p, align 4 ret void } ; Should delete store of 10 even though memcpy is a may-store to P (P and Q may ; alias). define void @test7_atomic(i32* align 4 %p, i8* align 4 %q, i8* noalias align 4 %r) { ; CHECK-LABEL: @test7_atomic( ; CHECK-NEXT: call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 4 [[Q:%.*]], i8* align 4 [[R:%.*]], i64 900, i32 4) ; CHECK-NEXT: store atomic i32 30, i32* [[P:%.*]] unordered, align 4 ; CHECK-NEXT: ret void ; store atomic i32 10, i32* %p unordered, align 4 ;; dead. call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 4 %q, i8* align 4 %r, i64 900, i32 4) store atomic i32 30, i32* %p unordered, align 4 ret void } ; Do not delete stores that are only partially killed. define i32 @test8() { ; CHECK-LABEL: @test8( ; CHECK-NEXT: [[V:%.*]] = alloca i32 ; CHECK-NEXT: store i32 1234567, i32* [[V]] ; CHECK-NEXT: [[X:%.*]] = load i32, i32* [[V]] ; CHECK-NEXT: ret i32 [[X]] ; %V = alloca i32 store i32 1234567, i32* %V %V2 = bitcast i32* %V to i8* store i8 0, i8* %V2 %X = load i32, i32* %V ret i32 %X } ; Test for byval handling. %struct.x = type { i32, i32, i32, i32 } define void @test9(%struct.x* byval %a) nounwind { ; CHECK-LABEL: @test9( ; CHECK-NEXT: ret void ; %tmp2 = getelementptr %struct.x, %struct.x* %a, i32 0, i32 0 store i32 1, i32* %tmp2, align 4 ret void } ; Test for inalloca handling. define void @test9_2(%struct.x* inalloca %a) nounwind { ; CHECK-LABEL: @test9_2( ; CHECK-NEXT: ret void ; %tmp2 = getelementptr %struct.x, %struct.x* %a, i32 0, i32 0 store i32 1, i32* %tmp2, align 4 ret void } ; Test for preallocated handling. define void @test9_3(%struct.x* preallocated(%struct.x) %a) nounwind { ; CHECK-LABEL: @test9_3( ; CHECK-NEXT: ret void ; %tmp2 = getelementptr %struct.x, %struct.x* %a, i32 0, i32 0 store i32 1, i32* %tmp2, align 4 ret void } ; va_arg has fuzzy dependence, the store shouldn't be zapped. define double @test10(i8* %X) { ; CHECK-LABEL: @test10( ; CHECK-NEXT: [[X_ADDR:%.*]] = alloca i8* ; CHECK-NEXT: store i8* [[X:%.*]], i8** [[X_ADDR]] ; CHECK-NEXT: [[TMP_0:%.*]] = va_arg i8** [[X_ADDR]], double ; CHECK-NEXT: ret double [[TMP_0]] ; %X_addr = alloca i8* store i8* %X, i8** %X_addr %tmp.0 = va_arg i8** %X_addr, double ret double %tmp.0 } ; DSE should delete the dead trampoline. declare void @test11f() define void @test11() { ; CHECK-LABEL: @test11( ; CHECK-NEXT: ret void ; %storage = alloca [10 x i8], align 16 ; <[10 x i8]*> [#uses=1] %cast = getelementptr [10 x i8], [10 x i8]* %storage, i32 0, i32 0 ; [#uses=1] call void @llvm.init.trampoline( i8* %cast, i8* bitcast (void ()* @test11f to i8*), i8* null ) ; [#uses=1] ret void } ; PR2599 - load -> store to same address. define void @test12({ i32, i32 }* %x) nounwind { ; CHECK-LABEL: @test12( ; CHECK-NEXT: [[TMP7:%.*]] = getelementptr { i32, i32 }, { i32, i32 }* [[X:%.*]], i32 0, i32 1 ; CHECK-NEXT: [[TMP8:%.*]] = load i32, i32* [[TMP7]], align 4 ; CHECK-NEXT: [[TMP17:%.*]] = sub i32 0, [[TMP8]] ; CHECK-NEXT: store i32 [[TMP17]], i32* [[TMP7]], align 4 ; CHECK-NEXT: ret void ; %tmp4 = getelementptr { i32, i32 }, { i32, i32 }* %x, i32 0, i32 0 %tmp5 = load i32, i32* %tmp4, align 4 %tmp7 = getelementptr { i32, i32 }, { i32, i32 }* %x, i32 0, i32 1 %tmp8 = load i32, i32* %tmp7, align 4 %tmp17 = sub i32 0, %tmp8 store i32 %tmp5, i32* %tmp4, align 4 store i32 %tmp17, i32* %tmp7, align 4 ret void } ; %P doesn't escape, the DEAD instructions should be removed. declare void @test13f() define i32* @test13() { ; CHECK-LABEL: @test13( ; CHECK-NEXT: [[PTR:%.*]] = tail call i8* @malloc(i32 4) ; CHECK-NEXT: [[P:%.*]] = bitcast i8* [[PTR]] to i32* ; CHECK-NEXT: call void @test13f() ; CHECK-NEXT: store i32 0, i32* [[P]] ; CHECK-NEXT: ret i32* [[P]] ; %ptr = tail call i8* @malloc(i32 4) %P = bitcast i8* %ptr to i32* %DEAD = load i32, i32* %P %DEAD2 = add i32 %DEAD, 1 store i32 %DEAD2, i32* %P call void @test13f( ) store i32 0, i32* %P ret i32* %P } define i32 addrspace(1)* @test13_addrspacecast() { ; CHECK-LABEL: @test13_addrspacecast( ; CHECK-NEXT: [[P:%.*]] = tail call i8* @malloc(i32 4) ; CHECK-NEXT: [[P_BC:%.*]] = bitcast i8* [[P]] to i32* ; CHECK-NEXT: [[P:%.*]] = addrspacecast i32* [[P_BC]] to i32 addrspace(1)* ; CHECK-NEXT: call void @test13f() ; CHECK-NEXT: store i32 0, i32 addrspace(1)* [[P]] ; CHECK-NEXT: ret i32 addrspace(1)* [[P]] ; %p = tail call i8* @malloc(i32 4) %p.bc = bitcast i8* %p to i32* %P = addrspacecast i32* %p.bc to i32 addrspace(1)* %DEAD = load i32, i32 addrspace(1)* %P %DEAD2 = add i32 %DEAD, 1 store i32 %DEAD2, i32 addrspace(1)* %P call void @test13f( ) store i32 0, i32 addrspace(1)* %P ret i32 addrspace(1)* %P } declare noalias i8* @malloc(i32) declare noalias i8* @calloc(i32, i32) declare noalias i8* @aligned_alloc(i32, i32) declare void @free(i8*) define void @test14(i32* %Q) { ; CHECK-LABEL: @test14( ; CHECK-NEXT: ret void ; %P = alloca i32 %DEAD = load i32, i32* %Q store i32 %DEAD, i32* %P ret void } ; Dead store on an aligned_alloc: should know that %M doesn't alias with %A. define i32 @test14a(i8* %M, i8 %value) { ; CHECK-LABEL: @test14a( ; CHECK-NOT: store ; CHECK: ret i32 0 ; %A = tail call i8* @aligned_alloc(i32 32, i32 1024) store i8 %value, i8* %A tail call void @free(i8* %A) ret i32 0 } ; PR8701 ;; Fully dead overwrite of memcpy. define void @test15(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test15( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) ret void } ;; Fully dead overwrite of memcpy. define void @test15_atomic(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test15_atomic( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } ;; Fully dead overwrite of memcpy. define void @test15_atomic_weaker(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test15_atomic_weaker( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i1 false) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } ;; Fully dead overwrite of memcpy. define void @test15_atomic_weaker_2(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test15_atomic_weaker_2( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i1 false) ret void } ;; Full overwrite of smaller memcpy. define void @test16(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test16( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 8, i1 false) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) ret void } ;; Full overwrite of smaller memcpy. define void @test16_atomic(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test16_atomic( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 8, i32 1) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } ;; Full overwrite of smaller memory where overwrite has stronger atomicity define void @test16_atomic_weaker(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test16_atomic_weaker( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 8, i1 false) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } ;; Full overwrite of smaller memory where overwrite has weaker atomicity. define void @test16_atomic_weaker_2(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test16_atomic_weaker_2( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 8, i32 1) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i1 false) ret void } ;; Overwrite of memset by memcpy. define void @test17(i8* %P, i8* noalias %Q) nounwind ssp { ; CHECK-LABEL: @test17( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memset.p0i8.i64(i8* %P, i8 42, i64 8, i1 false) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) ret void } ;; Overwrite of memset by memcpy. define void @test17_atomic(i8* %P, i8* noalias %Q) nounwind ssp { ; CHECK-LABEL: @test17_atomic( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* align 1 %P, i8 42, i64 8, i32 1) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } ;; Overwrite of memset by memcpy. Overwrite is stronger atomicity. We can ;; remove the memset. define void @test17_atomic_weaker(i8* %P, i8* noalias %Q) nounwind ssp { ; CHECK-LABEL: @test17_atomic_weaker( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memset.p0i8.i64(i8* align 1 %P, i8 42, i64 8, i1 false) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } ;; Overwrite of memset by memcpy. Overwrite is weaker atomicity. We can remove ;; the memset. define void @test17_atomic_weaker_2(i8* %P, i8* noalias %Q) nounwind ssp { ; CHECK-LABEL: @test17_atomic_weaker_2( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memset.element.unordered.atomic.p0i8.i64(i8* align 1 %P, i8 42, i64 8, i32 1) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i1 false) ret void } ; Should not delete the volatile memset. define void @test17v(i8* %P, i8* %Q) nounwind ssp { ; CHECK-LABEL: @test17v( ; CHECK-NEXT: tail call void @llvm.memset.p0i8.i64(i8* [[P:%.*]], i8 42, i64 8, i1 true) ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memset.p0i8.i64(i8* %P, i8 42, i64 8, i1 true) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) ret void } ; PR8728 ; Do not delete instruction where possible situation is: ; A = B ; A = A ; ; NB! See PR11763 - currently LLVM allows memcpy's source and destination to be ; equal (but not inequal and overlapping). define void @test18(i8* %P, i8* %Q, i8* %R) nounwind ssp { ; CHECK-LABEL: @test18( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P]], i8* [[R:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %R, i64 12, i1 false) ret void } define void @test18_atomic(i8* %P, i8* %Q, i8* %R) nounwind ssp { ; CHECK-LABEL: @test18_atomic( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P]], i8* align 1 [[R:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %R, i64 12, i32 1) ret void } ; The store here is not dead because the byval call reads it. declare void @test19f({i32}* byval align 4 %P) define void @test19({i32} * nocapture byval align 4 %arg5) nounwind ssp { ; CHECK-LABEL: @test19( ; CHECK-NEXT: bb: ; CHECK-NEXT: [[TMP7:%.*]] = getelementptr inbounds { i32 }, { i32 }* [[ARG5:%.*]], i32 0, i32 0 ; CHECK-NEXT: store i32 912, i32* [[TMP7]] ; CHECK-NEXT: call void @test19f({ i32 }* byval align 4 [[ARG5]]) ; CHECK-NEXT: ret void ; bb: %tmp7 = getelementptr inbounds {i32}, {i32}* %arg5, i32 0, i32 0 store i32 912, i32* %tmp7 call void @test19f({i32}* byval align 4 %arg5) ret void } define void @test20() { ; CHECK-LABEL: @test20( ; CHECK-NEXT: ret void ; %m = call i8* @malloc(i32 24) store i8 0, i8* %m ret void } define void @test21() { ; CHECK-LABEL: @test21( ; CHECK-NEXT: ret void ; %m = call i8* @calloc(i32 9, i32 7) store i8 0, i8* %m ret void } define void @test22(i1 %i, i32 %k, i32 %m) nounwind { ; CHECK-LABEL: @test22( ; CHECK-NEXT: ret void ; %k.addr = alloca i32 %m.addr = alloca i32 %k.addr.m.addr = select i1 %i, i32* %k.addr, i32* %m.addr store i32 0, i32* %k.addr.m.addr, align 4 ret void } ; PR13547 declare noalias i8* @strdup(i8* nocapture) nounwind define noalias i8* @test23() nounwind uwtable ssp { ; CHECK-LABEL: @test23( ; CHECK-NEXT: [[X:%.*]] = alloca [2 x i8], align 1 ; CHECK-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [2 x i8], [2 x i8]* [[X]], i64 0, i64 0 ; CHECK-NEXT: store i8 97, i8* [[ARRAYIDX]], align 1 ; CHECK-NEXT: [[ARRAYIDX1:%.*]] = getelementptr inbounds [2 x i8], [2 x i8]* [[X]], i64 0, i64 1 ; CHECK-NEXT: store i8 0, i8* [[ARRAYIDX1]], align 1 ; CHECK-NEXT: [[CALL:%.*]] = call i8* @strdup(i8* [[ARRAYIDX]]) #3 ; CHECK-NEXT: ret i8* [[CALL]] ; %x = alloca [2 x i8], align 1 %arrayidx = getelementptr inbounds [2 x i8], [2 x i8]* %x, i64 0, i64 0 store i8 97, i8* %arrayidx, align 1 %arrayidx1 = getelementptr inbounds [2 x i8], [2 x i8]* %x, i64 0, i64 1 store i8 0, i8* %arrayidx1, align 1 %call = call i8* @strdup(i8* %arrayidx) nounwind ret i8* %call } ; Make sure same sized store to later element is deleted define void @test24([2 x i32]* %a, i32 %b, i32 %c) nounwind { ; CHECK-LABEL: @test24( ; CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds [2 x i32], [2 x i32]* [[A:%.*]], i64 0, i64 0 ; CHECK-NEXT: store i32 [[B:%.*]], i32* [[TMP1]], align 4 ; CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds [2 x i32], [2 x i32]* [[A]], i64 0, i64 1 ; CHECK-NEXT: store i32 [[C:%.*]], i32* [[TMP2]], align 4 ; CHECK-NEXT: ret void ; %1 = getelementptr inbounds [2 x i32], [2 x i32]* %a, i64 0, i64 0 store i32 0, i32* %1, align 4 %2 = getelementptr inbounds [2 x i32], [2 x i32]* %a, i64 0, i64 1 store i32 0, i32* %2, align 4 %3 = getelementptr inbounds [2 x i32], [2 x i32]* %a, i64 0, i64 0 store i32 %b, i32* %3, align 4 %4 = getelementptr inbounds [2 x i32], [2 x i32]* %a, i64 0, i64 1 store i32 %c, i32* %4, align 4 ret void } ; Check another case like PR13547 where strdup is not like malloc. define i8* @test25(i8* %p) nounwind { ; CHECK-LABEL: @test25( ; CHECK-NEXT: [[P_4:%.*]] = getelementptr i8, i8* [[P:%.*]], i64 4 ; CHECK-NEXT: [[TMP:%.*]] = load i8, i8* [[P_4]], align 1 ; CHECK-NEXT: store i8 0, i8* [[P_4]], align 1 ; CHECK-NEXT: [[Q:%.*]] = call i8* @strdup(i8* [[P]]) #6 ; CHECK-NEXT: store i8 [[TMP]], i8* [[P_4]], align 1 ; CHECK-NEXT: ret i8* [[Q]] ; %p.4 = getelementptr i8, i8* %p, i64 4 %tmp = load i8, i8* %p.4, align 1 store i8 0, i8* %p.4, align 1 %q = call i8* @strdup(i8* %p) nounwind optsize store i8 %tmp, i8* %p.4, align 1 ret i8* %q } ; Remove redundant store if loaded value is in another block. define i32 @test26(i1 %c, i32* %p) { ; CHECK-LABEL: @test26( ; CHECK-NEXT: entry: ; CHECK-NEXT: br i1 [[C:%.*]], label [[BB1:%.*]], label [[BB2:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br label [[BB3:%.*]] ; CHECK: bb2: ; CHECK-NEXT: br label [[BB3]] ; CHECK: bb3: ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 br i1 %c, label %bb1, label %bb2 bb1: br label %bb3 bb2: store i32 %v, i32* %p, align 4 br label %bb3 bb3: ret i32 0 } ; Remove redundant store if loaded value is in another block. define i32 @test27(i1 %c, i32* %p) { ; CHECK-LABEL: @test27( ; CHECK-NEXT: entry: ; CHECK-NEXT: br i1 [[C:%.*]], label [[BB1:%.*]], label [[BB2:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br label [[BB3:%.*]] ; CHECK: bb2: ; CHECK-NEXT: br label [[BB3]] ; CHECK: bb3: ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 br i1 %c, label %bb1, label %bb2 bb1: br label %bb3 bb2: br label %bb3 bb3: store i32 %v, i32* %p, align 4 ret i32 0 } ; Don't remove redundant store because of may-aliased store. define i32 @test28(i1 %c, i32* %p, i32* %p2, i32 %i) { ; CHECK-LABEL: @test28( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P:%.*]], align 4 ; CHECK-NEXT: store i32 [[I:%.*]], i32* [[P2:%.*]], align 4 ; CHECK-NEXT: br i1 [[C:%.*]], label [[BB1:%.*]], label [[BB2:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br label [[BB3:%.*]] ; CHECK: bb2: ; CHECK-NEXT: br label [[BB3]] ; CHECK: bb3: ; CHECK-NEXT: store i32 [[V]], i32* [[P]], align 4 ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 ; Might overwrite value at %p store i32 %i, i32* %p2, align 4 br i1 %c, label %bb1, label %bb2 bb1: br label %bb3 bb2: br label %bb3 bb3: store i32 %v, i32* %p, align 4 ret i32 0 } ; Don't remove redundant store because of may-aliased store. define i32 @test29(i1 %c, i32* %p, i32* %p2, i32 %i) { ; CHECK-LABEL: @test29( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P:%.*]], align 4 ; CHECK-NEXT: br i1 [[C:%.*]], label [[BB1:%.*]], label [[BB2:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br label [[BB3:%.*]] ; CHECK: bb2: ; CHECK-NEXT: store i32 [[I:%.*]], i32* [[P2:%.*]], align 4 ; CHECK-NEXT: br label [[BB3]] ; CHECK: bb3: ; CHECK-NEXT: store i32 [[V]], i32* [[P]], align 4 ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 br i1 %c, label %bb1, label %bb2 bb1: br label %bb3 bb2: ; Might overwrite value at %p store i32 %i, i32* %p2, align 4 br label %bb3 bb3: store i32 %v, i32* %p, align 4 ret i32 0 } declare void @unknown_func() ; Don't remove redundant store because of unknown call. define i32 @test30(i1 %c, i32* %p, i32 %i) { ; CHECK-LABEL: @test30( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P:%.*]], align 4 ; CHECK-NEXT: br i1 [[C:%.*]], label [[BB1:%.*]], label [[BB2:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br label [[BB3:%.*]] ; CHECK: bb2: ; CHECK-NEXT: call void @unknown_func() ; CHECK-NEXT: br label [[BB3]] ; CHECK: bb3: ; CHECK-NEXT: store i32 [[V]], i32* [[P]], align 4 ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 br i1 %c, label %bb1, label %bb2 bb1: br label %bb3 bb2: ; Might overwrite value at %p call void @unknown_func() br label %bb3 bb3: store i32 %v, i32* %p, align 4 ret i32 0 } ; Remove redundant store if loaded value is in another block inside a loop. define i32 @test31(i1 %c, i32* %p, i32 %i) { ; CHECK-LABEL: @test31( ; CHECK-NEXT: entry: ; CHECK-NEXT: br label [[BB1:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br i1 undef, label [[BB1]], label [[BB2:%.*]] ; CHECK: bb2: ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 br label %bb1 bb1: store i32 %v, i32* %p, align 4 br i1 undef, label %bb1, label %bb2 bb2: ret i32 0 } ; Don't remove redundant store in a loop with a may-alias store. define i32 @test32(i1 %c, i32* %p, i32 %i) { ; CHECK-LABEL: @test32( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[V:%.*]] = load i32, i32* [[P:%.*]], align 4 ; CHECK-NEXT: br label [[BB1:%.*]] ; CHECK: bb1: ; CHECK-NEXT: store i32 [[V]], i32* [[P]], align 4 ; CHECK-NEXT: call void @unknown_func() ; CHECK-NEXT: br i1 undef, label [[BB1]], label [[BB2:%.*]] ; CHECK: bb2: ; CHECK-NEXT: ret i32 0 ; entry: %v = load i32, i32* %p, align 4 br label %bb1 bb1: store i32 %v, i32* %p, align 4 ; Might read and overwrite value at %p call void @unknown_func() br i1 undef, label %bb1, label %bb2 bb2: ret i32 0 } ; Remove redundant store, which is in the lame loop as the load. define i32 @test33(i1 %c, i32* %p, i32 %i) { ; CHECK-LABEL: @test33( ; CHECK-NEXT: entry: ; CHECK-NEXT: br label [[BB1:%.*]] ; CHECK: bb1: ; CHECK-NEXT: br label [[BB2:%.*]] ; CHECK: bb2: ; CHECK-NEXT: call void @unknown_func() ; CHECK-NEXT: br i1 undef, label [[BB1]], label [[BB3:%.*]] ; CHECK: bb3: ; CHECK-NEXT: ret i32 0 ; entry: br label %bb1 bb1: %v = load i32, i32* %p, align 4 br label %bb2 bb2: store i32 %v, i32* %p, align 4 ; Might read and overwrite value at %p, but doesn't matter. call void @unknown_func() br i1 undef, label %bb1, label %bb3 bb3: ret i32 0 } ; Don't remove redundant store: unknown_func could unwind define void @test34(i32* noalias %p) { ; CHECK-LABEL: @test34( ; CHECK-NEXT: store i32 1, i32* [[P:%.*]] ; CHECK-NEXT: call void @unknown_func() ; CHECK-NEXT: store i32 0, i32* [[P]] ; CHECK-NEXT: ret void ; store i32 1, i32* %p call void @unknown_func() store i32 0, i32* %p ret void } ; Remove redundant store even with an unwinding function in the same block define void @test35(i32* noalias %p) { ; CHECK-LABEL: @test35( ; CHECK-NEXT: call void @unknown_func() ; CHECK-NEXT: store i32 0, i32* [[P:%.*]] ; CHECK-NEXT: ret void ; call void @unknown_func() store i32 1, i32* %p store i32 0, i32* %p ret void } ; We cannot optimize away the first memmove since %P could overlap with %Q. define void @test36(i8* %P, i8* %Q) { ; CHECK-LABEL: @test36( ; CHECK-NEXT: tail call void @llvm.memmove.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: tail call void @llvm.memmove.p0i8.p0i8.i64(i8* [[P]], i8* [[Q]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memmove.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) tail call void @llvm.memmove.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) ret void } define void @test36_atomic(i8* %P, i8* %Q) { ; CHECK-LABEL: @test36_atomic( ; CHECK-NEXT: tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P]], i8* align 1 [[Q]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) ret void } define void @test37(i8* %P, i8* %Q, i8* %R) { ; CHECK-LABEL: @test37( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: tail call void @llvm.memmove.p0i8.p0i8.i64(i8* [[P]], i8* [[R:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) tail call void @llvm.memmove.p0i8.p0i8.i64(i8* %P, i8* %R, i64 12, i1 false) ret void } define void @test37_atomic(i8* %P, i8* %Q, i8* %R) { ; CHECK-LABEL: @test37_atomic( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P]], i8* align 1 [[R:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %R, i64 12, i32 1) ret void } ; Same caveat about memcpy as in @test18 applies here. define void @test38(i8* %P, i8* %Q, i8* %R) { ; CHECK-LABEL: @test38( ; CHECK-NEXT: tail call void @llvm.memmove.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P]], i8* [[R:%.*]], i64 12, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memmove.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %R, i64 12, i1 false) ret void } define void @test38_atomic(i8* %P, i8* %Q, i8* %R) { ; CHECK-LABEL: @test38_atomic( ; CHECK-NEXT: tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P]], i8* align 1 [[R:%.*]], i64 12, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %R, i64 12, i32 1) ret void } define void @test39(i8* %P, i8* %Q, i8* %R) { ; CHECK-LABEL: @test39( ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P:%.*]], i8* [[Q:%.*]], i64 12, i1 false) ; CHECK-NEXT: tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* [[P]], i8* [[R:%.*]], i64 8, i1 false) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %Q, i64 12, i1 false) tail call void @llvm.memcpy.p0i8.p0i8.i64(i8* %P, i8* %R, i64 8, i1 false) ret void } define void @test39_atomic(i8* %P, i8* %Q, i8* %R) { ; CHECK-LABEL: @test39_atomic( ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P:%.*]], i8* align 1 [[Q:%.*]], i64 12, i32 1) ; CHECK-NEXT: tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 [[P]], i8* align 1 [[R:%.*]], i64 8, i32 1) ; CHECK-NEXT: ret void ; tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %Q, i64 12, i32 1) tail call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i64(i8* align 1 %P, i8* align 1 %R, i64 8, i32 1) ret void } define i32 @test40() { ; CHECK-LABEL: @test40( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: br label [[LOOP:%.*]] ; CHECK: loop: ; CHECK-NEXT: [[INDVARS_IV:%.*]] = phi i64 [ 0, [[ENTRY:%.*]] ], [ [[INDVARS_IV_NEXT:%.*]], [[LOOP]] ] ; CHECK-NEXT: [[INDVARS_IV_NEXT]] = add nuw nsw i64 [[INDVARS_IV]], 1 ; CHECK-NEXT: [[P_NEXT:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV_NEXT]] ; CHECK-NEXT: store i8 1, i8* [[P_NEXT]] ; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV]] ; CHECK-NEXT: store i8 0, i8* [[P]] ; CHECK-NEXT: [[CONTINUE:%.*]] = icmp ugt i64 [[INDVARS_IV]], 15 ; CHECK-NEXT: br i1 [[CONTINUE]], label [[LOOP]], label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) br label %loop loop: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %loop ] %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %p.next = getelementptr inbounds i8, i8* %m, i64 %indvars.iv.next store i8 1, i8* %p.next %p = getelementptr inbounds i8, i8* %m, i64 %indvars.iv store i8 0, i8* %p %continue = icmp ugt i64 %indvars.iv, 15 br i1 %continue, label %loop, label %return return: ret i32 0 } define i32 @test41() { ; CHECK-LABEL: @test41( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: br label [[LOOP:%.*]] ; CHECK: loop: ; CHECK-NEXT: [[INDVARS_IV:%.*]] = phi i64 [ 0, [[ENTRY:%.*]] ], [ [[INDVARS_IV_NEXT:%.*]], [[CONT:%.*]] ] ; CHECK-NEXT: [[INDVARS_IV_NEXT]] = add nuw nsw i64 [[INDVARS_IV]], 1 ; CHECK-NEXT: [[P_NEXT:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV_NEXT]] ; CHECK-NEXT: store i8 1, i8* [[P_NEXT]] ; CHECK-NEXT: br label [[CONT]] ; CHECK: cont: ; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV]] ; CHECK-NEXT: store i8 0, i8* [[P]] ; CHECK-NEXT: [[CONTINUE:%.*]] = icmp ugt i64 [[INDVARS_IV]], 15 ; CHECK-NEXT: br i1 [[CONTINUE]], label [[LOOP]], label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) br label %loop loop: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %cont ] %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %p.next = getelementptr inbounds i8, i8* %m, i64 %indvars.iv.next store i8 1, i8* %p.next br label %cont cont: %p = getelementptr inbounds i8, i8* %m, i64 %indvars.iv store i8 0, i8* %p %continue = icmp ugt i64 %indvars.iv, 15 br i1 %continue, label %loop, label %return return: ret i32 0 } ; The store is redundant here, but currently we fail to eliminate it. ; We are walking from the store up to the calloc and translate phis as ; needed. In this case we fail to translate %p while going over the ; backedge. Because of that we conservatively assume that zero initialized ; memory is clobbered. define i32 @test42() { ; CHECK-LABEL: @test42( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: br label [[LOOP:%.*]] ; CHECK: loop: ; CHECK-NEXT: [[INDVARS_IV:%.*]] = phi i64 [ 0, [[ENTRY:%.*]] ], [ [[INDVARS_IV_NEXT:%.*]], [[CONT:%.*]] ] ; CHECK-NEXT: [[INDVARS_IV_NEXT]] = add nuw nsw i64 [[INDVARS_IV]], 1 ; CHECK-NEXT: br label [[CONT]] ; CHECK: cont: ; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV]] ; CHECK-NEXT: store i8 0, i8* [[P]] ; CHECK-NEXT: [[CONTINUE:%.*]] = icmp ugt i64 [[INDVARS_IV]], 15 ; CHECK-NEXT: br i1 [[CONTINUE]], label [[LOOP]], label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) br label %loop loop: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %cont ] %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 br label %cont cont: %p = getelementptr inbounds i8, i8* %m, i64 %indvars.iv store i8 0, i8* %p %continue = icmp ugt i64 %indvars.iv, 15 br i1 %continue, label %loop, label %return return: ret i32 0 } define i32 @test43() { ; CHECK-LABEL: @test43( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: br label [[LOOP:%.*]] ; CHECK: loop: ; CHECK-NEXT: [[INDVARS_IV:%.*]] = phi i64 [ 0, [[ENTRY:%.*]] ], [ [[INDVARS_IV_NEXT:%.*]], [[CONT_2:%.*]] ] ; CHECK-NEXT: [[INDVARS_IV_NEXT]] = add nuw nsw i64 [[INDVARS_IV]], 1 ; CHECK-NEXT: [[P_NEXT:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV_NEXT]] ; CHECK-NEXT: store i8 1, i8* [[P_NEXT]] ; CHECK-NEXT: br label [[CONT:%.*]] ; CHECK: cont: ; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV]] ; CHECK-NEXT: store i8 0, i8* [[P]] ; CHECK-NEXT: br label [[CONT_2]] ; CHECK: cont.2: ; CHECK-NEXT: [[CONTINUE:%.*]] = icmp ugt i64 [[INDVARS_IV]], 15 ; CHECK-NEXT: br i1 [[CONTINUE]], label [[LOOP]], label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) br label %loop loop: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %cont.2 ] %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %p.next = getelementptr inbounds i8, i8* %m, i64 %indvars.iv.next store i8 1, i8* %p.next br label %cont cont: %p = getelementptr inbounds i8, i8* %m, i64 %indvars.iv store i8 0, i8* %p br label %cont.2 cont.2: %continue = icmp ugt i64 %indvars.iv, 15 br i1 %continue, label %loop, label %return return: ret i32 0 } define i32 @test44() { ; CHECK-LABEL: @test44( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: br label [[LOOP:%.*]] ; CHECK: loop: ; CHECK-NEXT: [[INDVARS_IV:%.*]] = phi i64 [ 0, [[ENTRY:%.*]] ], [ [[INDVARS_IV_NEXT:%.*]], [[CONT_2:%.*]] ] ; CHECK-NEXT: [[INDVARS_IV_NEXT]] = add nuw nsw i64 [[INDVARS_IV]], 1 ; CHECK-NEXT: [[P_NEXT:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV_NEXT]] ; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[INDVARS_IV]] ; CHECK-NEXT: store i8 0, i8* [[P]] ; CHECK-NEXT: br label [[CONT:%.*]] ; CHECK: cont: ; CHECK-NEXT: store i8 1, i8* [[P_NEXT]] ; CHECK-NEXT: br label [[CONT_2]] ; CHECK: cont.2: ; CHECK-NEXT: [[CONTINUE:%.*]] = icmp ugt i64 [[INDVARS_IV]], 15 ; CHECK-NEXT: br i1 [[CONTINUE]], label [[LOOP]], label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) br label %loop loop: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %cont.2 ] %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %p.next = getelementptr inbounds i8, i8* %m, i64 %indvars.iv.next %p = getelementptr inbounds i8, i8* %m, i64 %indvars.iv store i8 0, i8* %p br label %cont cont: store i8 1, i8* %p.next br label %cont.2 cont.2: %continue = icmp ugt i64 %indvars.iv, 15 br i1 %continue, label %loop, label %return return: ret i32 0 } ; This is an example which can potentially benefit from PHI translation. ; Current implementation doesn't handle this case though. This is because ; we don't visit the same block with different addresses while looking for ; clobbering instructions. define i32 @test45(i1 %c) { ; CHECK-LABEL: @test45( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE:%.*]], label [[FALSE:%.*]] ; CHECK: true: ; CHECK-NEXT: [[P_1:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 1 ; CHECK-NEXT: store i8 1, i8* [[P_1]] ; CHECK-NEXT: br label [[CONT:%.*]] ; CHECK: false: ; CHECK-NEXT: [[P_2:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 2 ; CHECK-NEXT: store i8 1, i8* [[P_2]] ; CHECK-NEXT: br label [[CONT]] ; CHECK: cont: ; CHECK-NEXT: [[OFFSET:%.*]] = phi i64 [ 2, [[TRUE]] ], [ 1, [[FALSE]] ] ; CHECK-NEXT: [[P:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 [[OFFSET]] ; CHECK-NEXT: store i8 0, i8* [[P]] ; CHECK-NEXT: br label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) br i1 %c, label %true, label %false true: %p.1 = getelementptr inbounds i8, i8* %m, i64 1 store i8 1, i8* %p.1 br label %cont false: %p.2 = getelementptr inbounds i8, i8* %m, i64 2 store i8 1, i8* %p.2 br label %cont cont: %offset = phi i64 [ 2, %true ], [ 1, %false ] %p = getelementptr inbounds i8, i8* %m, i64 %offset store i8 0, i8* %p br label %return return: ret i32 0 } ; This is test45 modified in a way to demonstrate PHI translation ; improving the accuracy of the analysis (on a slightly convoluted ; case though). define i32 @test46(i1 %c) { ; CHECK-LABEL: @test46( ; CHECK-NEXT: entry: ; CHECK-NEXT: [[M:%.*]] = call i8* @calloc(i32 9, i32 20) ; CHECK-NEXT: [[P_1:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 1 ; CHECK-NEXT: [[P_2:%.*]] = getelementptr inbounds i8, i8* [[M]], i64 2 ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE:%.*]], label [[FALSE:%.*]] ; CHECK: true: ; CHECK-NEXT: store i8 1, i8* [[P_1]] ; CHECK-NEXT: br label [[CONT:%.*]] ; CHECK: false: ; CHECK-NEXT: store i8 1, i8* [[P_1]] ; CHECK-NEXT: br label [[CONT]] ; CHECK: cont: ; CHECK-NEXT: br label [[RETURN:%.*]] ; CHECK: return: ; CHECK-NEXT: ret i32 0 ; entry: %m = call i8* @calloc(i32 9, i32 20) %p.1 = getelementptr inbounds i8, i8* %m, i64 1 %p.2 = getelementptr inbounds i8, i8* %m, i64 2 br i1 %c, label %true, label %false true: store i8 1, i8* %p.1 br label %cont false: store i8 1, i8* %p.1 br label %cont cont: %offset = phi i64 [ 2, %true ], [ 2, %false ] %p = getelementptr inbounds i8, i8* %m, i64 %offset store i8 0, i8* %p br label %return return: ret i32 0 } declare void @llvm.memmove.p0i8.p0i8.i64(i8* nocapture, i8* nocapture readonly, i64, i1) declare void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i64(i8* nocapture, i8* nocapture readonly, i64, i32)