mirror of
https://github.com/RPCSX/llvm.git
synced 2024-12-03 17:31:50 +00:00
cf0fa9b9dd
The z13 vector facility includes some instructions that operate only on the high f64 in a v2f64, effectively extending the FP register set from 16 to 32 registers. It's still better to use the old instructions if the operands happen to fit though, since the older instructions have a shorter encoding. Based on a patch by Richard Sandiford. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@236524 91177308-0d34-0410-b5e6-96231b3b80d8
122 lines
3.6 KiB
LLVM
122 lines
3.6 KiB
LLVM
; Test multiplication of two f64s, producing an f64 result.
|
|
;
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z10 \
|
|
; RUN: | FileCheck -check-prefix=CHECK -check-prefix=CHECK-SCALAR %s
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z13 | FileCheck %s
|
|
|
|
declare double @foo()
|
|
|
|
; Check register multiplication.
|
|
define double @f1(double %f1, double %f2) {
|
|
; CHECK-LABEL: f1:
|
|
; CHECK: mdbr %f0, %f2
|
|
; CHECK: br %r14
|
|
%res = fmul double %f1, %f2
|
|
ret double %res
|
|
}
|
|
|
|
; Check the low end of the MDB range.
|
|
define double @f2(double %f1, double *%ptr) {
|
|
; CHECK-LABEL: f2:
|
|
; CHECK: mdb %f0, 0(%r2)
|
|
; CHECK: br %r14
|
|
%f2 = load double , double *%ptr
|
|
%res = fmul double %f1, %f2
|
|
ret double %res
|
|
}
|
|
|
|
; Check the high end of the aligned MDB range.
|
|
define double @f3(double %f1, double *%base) {
|
|
; CHECK-LABEL: f3:
|
|
; CHECK: mdb %f0, 4088(%r2)
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr double, double *%base, i64 511
|
|
%f2 = load double , double *%ptr
|
|
%res = fmul double %f1, %f2
|
|
ret double %res
|
|
}
|
|
|
|
; Check the next doubleword up, which needs separate address logic.
|
|
; Other sequences besides this one would be OK.
|
|
define double @f4(double %f1, double *%base) {
|
|
; CHECK-LABEL: f4:
|
|
; CHECK: aghi %r2, 4096
|
|
; CHECK: mdb %f0, 0(%r2)
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr double, double *%base, i64 512
|
|
%f2 = load double , double *%ptr
|
|
%res = fmul double %f1, %f2
|
|
ret double %res
|
|
}
|
|
|
|
; Check negative displacements, which also need separate address logic.
|
|
define double @f5(double %f1, double *%base) {
|
|
; CHECK-LABEL: f5:
|
|
; CHECK: aghi %r2, -8
|
|
; CHECK: mdb %f0, 0(%r2)
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr double, double *%base, i64 -1
|
|
%f2 = load double , double *%ptr
|
|
%res = fmul double %f1, %f2
|
|
ret double %res
|
|
}
|
|
|
|
; Check that MDB allows indices.
|
|
define double @f6(double %f1, double *%base, i64 %index) {
|
|
; CHECK-LABEL: f6:
|
|
; CHECK: sllg %r1, %r3, 3
|
|
; CHECK: mdb %f0, 800(%r1,%r2)
|
|
; CHECK: br %r14
|
|
%ptr1 = getelementptr double, double *%base, i64 %index
|
|
%ptr2 = getelementptr double, double *%ptr1, i64 100
|
|
%f2 = load double , double *%ptr2
|
|
%res = fmul double %f1, %f2
|
|
ret double %res
|
|
}
|
|
|
|
; Check that multiplications of spilled values can use MDB rather than MDBR.
|
|
define double @f7(double *%ptr0) {
|
|
; CHECK-LABEL: f7:
|
|
; CHECK: brasl %r14, foo@PLT
|
|
; CHECK-SCALAR: mdb %f0, 160(%r15)
|
|
; CHECK: br %r14
|
|
%ptr1 = getelementptr double, double *%ptr0, i64 2
|
|
%ptr2 = getelementptr double, double *%ptr0, i64 4
|
|
%ptr3 = getelementptr double, double *%ptr0, i64 6
|
|
%ptr4 = getelementptr double, double *%ptr0, i64 8
|
|
%ptr5 = getelementptr double, double *%ptr0, i64 10
|
|
%ptr6 = getelementptr double, double *%ptr0, i64 12
|
|
%ptr7 = getelementptr double, double *%ptr0, i64 14
|
|
%ptr8 = getelementptr double, double *%ptr0, i64 16
|
|
%ptr9 = getelementptr double, double *%ptr0, i64 18
|
|
%ptr10 = getelementptr double, double *%ptr0, i64 20
|
|
|
|
%val0 = load double , double *%ptr0
|
|
%val1 = load double , double *%ptr1
|
|
%val2 = load double , double *%ptr2
|
|
%val3 = load double , double *%ptr3
|
|
%val4 = load double , double *%ptr4
|
|
%val5 = load double , double *%ptr5
|
|
%val6 = load double , double *%ptr6
|
|
%val7 = load double , double *%ptr7
|
|
%val8 = load double , double *%ptr8
|
|
%val9 = load double , double *%ptr9
|
|
%val10 = load double , double *%ptr10
|
|
|
|
%ret = call double @foo()
|
|
|
|
%mul0 = fmul double %ret, %val0
|
|
%mul1 = fmul double %mul0, %val1
|
|
%mul2 = fmul double %mul1, %val2
|
|
%mul3 = fmul double %mul2, %val3
|
|
%mul4 = fmul double %mul3, %val4
|
|
%mul5 = fmul double %mul4, %val5
|
|
%mul6 = fmul double %mul5, %val6
|
|
%mul7 = fmul double %mul6, %val7
|
|
%mul8 = fmul double %mul7, %val8
|
|
%mul9 = fmul double %mul8, %val9
|
|
%mul10 = fmul double %mul9, %val10
|
|
|
|
ret double %mul10
|
|
}
|