mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-12-09 12:34:25 +00:00
f2f1248de9
Summary: Some vector loads and stores generated from AArch64 intrinsics alias each other unnecessarily, preventing better scheduling. We just need to transfer memory operands during lowering. Reviewers: mcrosier, t.p.northover, jmolloy Subscribers: aemerson, rengolin, llvm-commits Differential Revision: https://reviews.llvm.org/D26313 llvm-svn: 286168
61 lines
3.0 KiB
LLVM
61 lines
3.0 KiB
LLVM
; RUN: llc < %s -mcpu=cortex-a53 -enable-post-misched=false -enable-aa-sched-mi | FileCheck %s
|
|
|
|
; Check that the vector store intrinsic does not prevent fmla instructions from
|
|
; being scheduled together. Since the vector loads and stores generated from
|
|
; the intrinsics do not alias each other, the store can be pushed past the load.
|
|
; This allows fmla instructions to be scheduled together.
|
|
|
|
|
|
; CHECK: fmla
|
|
; CHECK-NEXT: fmla
|
|
; CHECK-NEXT: fmla
|
|
; CHECK-NEXT: fmla
|
|
target datalayout = "e-m:e-i64:64-i128:128-n8:16:32:64-S128"
|
|
target triple = "aarch64--linux-gnu"
|
|
|
|
%Struct = type { i64*, [9 x double], [16 x {float, float}], [16 x {float, float}], i32, i32 }
|
|
|
|
; Function Attrs: nounwind
|
|
define linkonce_odr void @func(%Struct* nocapture %this) unnamed_addr #0 align 2 {
|
|
entry:
|
|
%0 = insertelement <4 x float> undef, float undef, i32 0
|
|
%1 = insertelement <4 x float> %0, float undef, i32 1
|
|
%2 = insertelement <4 x float> %1, float undef, i32 2
|
|
%3 = insertelement <4 x float> %2, float undef, i32 3
|
|
%scevgep = getelementptr %Struct, %Struct* %this, i64 0, i32 2, i64 8, i32 0
|
|
%struct_ptr = bitcast float* %scevgep to i8*
|
|
%vec1 = tail call { <4 x float>, <4 x float> } @llvm.aarch64.neon.ld2.v4f32.p0i8(i8* %struct_ptr)
|
|
%ev1 = extractvalue { <4 x float>, <4 x float> } %vec1, 1
|
|
%fm1 = fmul <4 x float> %0, %ev1
|
|
%av1 = fadd <4 x float> %1, %fm1
|
|
%ev2 = extractvalue { <4 x float>, <4 x float> } %vec1, 0
|
|
%fm2 = fmul <4 x float> %2, %ev2
|
|
%av2 = fadd <4 x float> %3, %fm2
|
|
%scevgep2 = getelementptr %Struct, %Struct* %this, i64 0, i32 3, i64 8, i32 0
|
|
%struct_ptr2 = bitcast float* %scevgep2 to i8*
|
|
tail call void @llvm.aarch64.neon.st2.v4f32.p0i8(<4 x float> %av2, <4 x float> %av1, i8* %struct_ptr2)
|
|
%scevgep3 = getelementptr %Struct, %Struct* %this, i64 0, i32 2, i64 12, i32 0
|
|
%struct_ptr3 = bitcast float* %scevgep3 to i8*
|
|
%vec2 = tail call { <4 x float>, <4 x float> } @llvm.aarch64.neon.ld2.v4f32.p0i8(i8* %struct_ptr3)
|
|
%ev3 = extractvalue { <4 x float>, <4 x float> } %vec2, 1
|
|
%fm3 = fmul <4 x float> %0, %ev3
|
|
%av3 = fadd <4 x float> %1, %fm3
|
|
%ev4 = extractvalue { <4 x float>, <4 x float> } %vec2, 0
|
|
%fm4 = fmul <4 x float> %2, %ev4
|
|
%av4 = fadd <4 x float> %3, %fm4
|
|
%scevgep4 = getelementptr %Struct, %Struct* %this, i64 0, i32 3, i64 12, i32 0
|
|
%struct_ptr4 = bitcast float* %scevgep4 to i8*
|
|
tail call void @llvm.aarch64.neon.st2.v4f32.p0i8(<4 x float> %av4, <4 x float> %av3, i8* %struct_ptr4)
|
|
ret void
|
|
}
|
|
|
|
; Function Attrs: nounwind readonly
|
|
declare { <4 x float>, <4 x float> } @llvm.aarch64.neon.ld2.v4f32.p0i8(i8*) #2
|
|
|
|
; Function Attrs: nounwind
|
|
declare void @llvm.aarch64.neon.st2.v4f32.p0i8(<4 x float>, <4 x float>, i8* nocapture) #1
|
|
|
|
attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "stack-protector-buffer-size"="8" "unsafe-fp-math"="true" "use-soft-float"="false" }
|
|
attributes #1 = { nounwind }
|
|
attributes #2 = { nounwind readonly }
|