2019-05-23 18:08:13 +00:00
|
|
|
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
2019-07-09 15:48:05 +00:00
|
|
|
; RUN: llc -march=mips -mattr=+msa,+fp64,+mips32r2 < %s | FileCheck %s --check-prefixes=ALL,MIPS
|
|
|
|
; RUN: llc -march=mipsel -mattr=+msa,+fp64,+mips32r2 < %s | FileCheck %s --check-prefixes=ALL,MIPSEL
|
2013-09-11 12:39:25 +00:00
|
|
|
|
|
|
|
define void @add_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: addv.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = add <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @add_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: addv.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = add <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @add_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: addv.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = add <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @add_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: addv.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = add <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
2013-09-23 14:29:55 +00:00
|
|
|
|
|
|
|
define void @add_v16i8_i(<16 x i8>* %c, <16 x i8>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v16i8_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($5)
|
|
|
|
; ALL-NEXT: addvi.b $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
2013-09-24 12:18:31 +00:00
|
|
|
%2 = add <16 x i8> %1, <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1,
|
2019-05-23 18:08:13 +00:00
|
|
|
i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>
|
2013-09-23 14:29:55 +00:00
|
|
|
store <16 x i8> %2, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @add_v8i16_i(<8 x i16>* %c, <8 x i16>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v8i16_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($5)
|
|
|
|
; ALL-NEXT: addvi.h $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
2013-09-24 12:18:31 +00:00
|
|
|
%2 = add <8 x i16> %1, <i16 1, i16 1, i16 1, i16 1,
|
2019-05-23 18:08:13 +00:00
|
|
|
i16 1, i16 1, i16 1, i16 1>
|
2013-09-23 14:29:55 +00:00
|
|
|
store <8 x i16> %2, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @add_v4i32_i(<4 x i32>* %c, <4 x i32>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v4i32_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($5)
|
|
|
|
; ALL-NEXT: addvi.w $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
2013-09-23 14:29:55 +00:00
|
|
|
%2 = add <4 x i32> %1, <i32 1, i32 1, i32 1, i32 1>
|
|
|
|
store <4 x i32> %2, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @add_v2i64_i(<2 x i64>* %c, <2 x i64>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: add_v2i64_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($5)
|
|
|
|
; ALL-NEXT: addvi.d $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
2013-09-23 14:29:55 +00:00
|
|
|
%2 = add <2 x i64> %1, <i64 1, i64 1>
|
|
|
|
store <2 x i64> %2, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-11 12:39:25 +00:00
|
|
|
define void @sub_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: subv.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sub <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @sub_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: subv.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sub <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @sub_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: subv.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sub <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @sub_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: subv.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sub <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-23 14:29:55 +00:00
|
|
|
define void @sub_v16i8_i(<16 x i8>* %c, <16 x i8>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v16i8_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($5)
|
|
|
|
; ALL-NEXT: subvi.b $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
2013-09-24 12:18:31 +00:00
|
|
|
%2 = sub <16 x i8> %1, <i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1,
|
2019-05-23 18:08:13 +00:00
|
|
|
i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1, i8 1>
|
2013-09-23 14:29:55 +00:00
|
|
|
store <16 x i8> %2, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2019-09-18 19:34:24 +00:00
|
|
|
define void @sub_v16i8_i_negated(<16 x i8>* %c, <16 x i8>* %a) nounwind {
|
|
|
|
; ALL-LABEL: sub_v16i8_i_negated:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($5)
|
[MIPS] For vectors, select `add %x, C` as `sub %x, -C` if it results in inline immediate
Summary:
As discussed in https://reviews.llvm.org/D62341#1515637,
for MIPS `add %x, -1` isn't optimal. Unlike X86 there
are no fastpaths to matearialize such `-1`/`1` vector constants,
and `sub %x, 1` results in better codegen,
so undo canonicalization
Reviewers: atanasyan, Petar.Avramovic, RKSimon
Reviewed By: atanasyan
Subscribers: sdardis, arichardson, hiraditya, jrtc27, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D66805
llvm-svn: 372254
2019-09-18 19:34:41 +00:00
|
|
|
; ALL-NEXT: subvi.b $w0, $w0, 1
|
2019-09-18 19:34:24 +00:00
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = add <16 x i8> %1, <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1,
|
|
|
|
i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>
|
|
|
|
store <16 x i8> %2, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-23 14:29:55 +00:00
|
|
|
define void @sub_v8i16_i(<8 x i16>* %c, <8 x i16>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v8i16_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($5)
|
|
|
|
; ALL-NEXT: subvi.h $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
2013-09-24 12:18:31 +00:00
|
|
|
%2 = sub <8 x i16> %1, <i16 1, i16 1, i16 1, i16 1,
|
2019-05-23 18:08:13 +00:00
|
|
|
i16 1, i16 1, i16 1, i16 1>
|
2013-09-23 14:29:55 +00:00
|
|
|
store <8 x i16> %2, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2019-09-18 19:34:24 +00:00
|
|
|
define void @sub_v8i16_i_negated(<8 x i16>* %c, <8 x i16>* %a) nounwind {
|
|
|
|
; ALL-LABEL: sub_v8i16_i_negated:
|
|
|
|
; ALL: # %bb.0:
|
[MIPS] For vectors, select `add %x, C` as `sub %x, -C` if it results in inline immediate
Summary:
As discussed in https://reviews.llvm.org/D62341#1515637,
for MIPS `add %x, -1` isn't optimal. Unlike X86 there
are no fastpaths to matearialize such `-1`/`1` vector constants,
and `sub %x, 1` results in better codegen,
so undo canonicalization
Reviewers: atanasyan, Petar.Avramovic, RKSimon
Reviewed By: atanasyan
Subscribers: sdardis, arichardson, hiraditya, jrtc27, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D66805
llvm-svn: 372254
2019-09-18 19:34:41 +00:00
|
|
|
; ALL-NEXT: ld.h $w0, 0($5)
|
|
|
|
; ALL-NEXT: subvi.h $w0, $w0, 1
|
2019-09-18 19:34:24 +00:00
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = add <8 x i16> %1, <i16 -1, i16 -1, i16 -1, i16 -1,
|
|
|
|
i16 -1, i16 -1, i16 -1, i16 -1>
|
|
|
|
store <8 x i16> %2, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-23 14:29:55 +00:00
|
|
|
define void @sub_v4i32_i(<4 x i32>* %c, <4 x i32>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v4i32_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($5)
|
|
|
|
; ALL-NEXT: subvi.w $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
2013-09-23 14:29:55 +00:00
|
|
|
%2 = sub <4 x i32> %1, <i32 1, i32 1, i32 1, i32 1>
|
|
|
|
store <4 x i32> %2, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2019-09-18 19:34:24 +00:00
|
|
|
define void @sub_v4i32_i_negated(<4 x i32>* %c, <4 x i32>* %a) nounwind {
|
|
|
|
; ALL-LABEL: sub_v4i32_i_negated:
|
|
|
|
; ALL: # %bb.0:
|
[MIPS] For vectors, select `add %x, C` as `sub %x, -C` if it results in inline immediate
Summary:
As discussed in https://reviews.llvm.org/D62341#1515637,
for MIPS `add %x, -1` isn't optimal. Unlike X86 there
are no fastpaths to matearialize such `-1`/`1` vector constants,
and `sub %x, 1` results in better codegen,
so undo canonicalization
Reviewers: atanasyan, Petar.Avramovic, RKSimon
Reviewed By: atanasyan
Subscribers: sdardis, arichardson, hiraditya, jrtc27, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D66805
llvm-svn: 372254
2019-09-18 19:34:41 +00:00
|
|
|
; ALL-NEXT: ld.w $w0, 0($5)
|
|
|
|
; ALL-NEXT: subvi.w $w0, $w0, 1
|
2019-09-18 19:34:24 +00:00
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = add <4 x i32> %1, <i32 -1, i32 -1, i32 -1, i32 -1>
|
|
|
|
store <4 x i32> %2, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-23 14:29:55 +00:00
|
|
|
define void @sub_v2i64_i(<2 x i64>* %c, <2 x i64>* %a) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: sub_v2i64_i:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($5)
|
|
|
|
; ALL-NEXT: subvi.d $w0, $w0, 1
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
2013-09-23 14:29:55 +00:00
|
|
|
%2 = sub <2 x i64> %1, <i64 1, i64 1>
|
|
|
|
store <2 x i64> %2, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2019-09-18 19:34:24 +00:00
|
|
|
define void @sub_v2i64_i_negated(<2 x i64>* %c, <2 x i64>* %a) nounwind {
|
|
|
|
; MIPS-LABEL: sub_v2i64_i_negated:
|
|
|
|
; MIPS: # %bb.0:
|
|
|
|
; MIPS-NEXT: ldi.b $w0, -1
|
|
|
|
; MIPS-NEXT: shf.w $w0, $w0, 177
|
|
|
|
; MIPS-NEXT: ld.d $w1, 0($5)
|
|
|
|
; MIPS-NEXT: addv.d $w0, $w1, $w0
|
|
|
|
; MIPS-NEXT: jr $ra
|
|
|
|
; MIPS-NEXT: st.d $w0, 0($4)
|
|
|
|
;
|
|
|
|
; MIPSEL-LABEL: sub_v2i64_i_negated:
|
|
|
|
; MIPSEL: # %bb.0:
|
|
|
|
; MIPSEL-NEXT: ldi.b $w0, -1
|
|
|
|
; MIPSEL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; MIPSEL-NEXT: addv.d $w0, $w1, $w0
|
|
|
|
; MIPSEL-NEXT: jr $ra
|
|
|
|
; MIPSEL-NEXT: st.d $w0, 0($4)
|
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = add <2 x i64> %1, <i64 -1, i64 -1>
|
|
|
|
store <2 x i64> %2, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-11 12:39:25 +00:00
|
|
|
define void @mul_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mul_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: mulv.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = mul <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mul_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mul_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: mulv.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = mul <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mul_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mul_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: mulv.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = mul <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mul_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mul_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: mulv.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = mul <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-10-11 10:50:42 +00:00
|
|
|
define void @maddv_v16i8(<16 x i8>* %d, <16 x i8>* %a, <16 x i8>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: maddv_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w2, 0($5)
|
|
|
|
; ALL-NEXT: maddv.b $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w2, 0($4)
|
|
|
|
<16 x i8>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
|
|
|
%3 = load <16 x i8>, <16 x i8>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <16 x i8> %2, %3
|
|
|
|
%5 = add <16 x i8> %4, %1
|
|
|
|
store <16 x i8> %5, <16 x i8>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @maddv_v8i16(<8 x i16>* %d, <8 x i16>* %a, <8 x i16>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: maddv_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w2, 0($5)
|
|
|
|
; ALL-NEXT: maddv.h $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w2, 0($4)
|
|
|
|
<8 x i16>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
|
|
|
%3 = load <8 x i16>, <8 x i16>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <8 x i16> %2, %3
|
|
|
|
%5 = add <8 x i16> %4, %1
|
|
|
|
store <8 x i16> %5, <8 x i16>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @maddv_v4i32(<4 x i32>* %d, <4 x i32>* %a, <4 x i32>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: maddv_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w2, 0($5)
|
|
|
|
; ALL-NEXT: maddv.w $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w2, 0($4)
|
|
|
|
<4 x i32>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
|
|
|
%3 = load <4 x i32>, <4 x i32>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <4 x i32> %2, %3
|
|
|
|
%5 = add <4 x i32> %4, %1
|
|
|
|
store <4 x i32> %5, <4 x i32>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @maddv_v2i64(<2 x i64>* %d, <2 x i64>* %a, <2 x i64>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: maddv_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w2, 0($5)
|
|
|
|
; ALL-NEXT: maddv.d $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w2, 0($4)
|
|
|
|
<2 x i64>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
|
|
|
%3 = load <2 x i64>, <2 x i64>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <2 x i64> %2, %3
|
|
|
|
%5 = add <2 x i64> %4, %1
|
|
|
|
store <2 x i64> %5, <2 x i64>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @msubv_v16i8(<16 x i8>* %d, <16 x i8>* %a, <16 x i8>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: msubv_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w2, 0($5)
|
|
|
|
; ALL-NEXT: msubv.b $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w2, 0($4)
|
|
|
|
<16 x i8>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
|
|
|
%3 = load <16 x i8>, <16 x i8>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <16 x i8> %2, %3
|
|
|
|
%5 = sub <16 x i8> %1, %4
|
|
|
|
store <16 x i8> %5, <16 x i8>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @msubv_v8i16(<8 x i16>* %d, <8 x i16>* %a, <8 x i16>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: msubv_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w2, 0($5)
|
|
|
|
; ALL-NEXT: msubv.h $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w2, 0($4)
|
|
|
|
<8 x i16>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
|
|
|
%3 = load <8 x i16>, <8 x i16>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <8 x i16> %2, %3
|
|
|
|
%5 = sub <8 x i16> %1, %4
|
|
|
|
store <8 x i16> %5, <8 x i16>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @msubv_v4i32(<4 x i32>* %d, <4 x i32>* %a, <4 x i32>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: msubv_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w2, 0($5)
|
|
|
|
; ALL-NEXT: msubv.w $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w2, 0($4)
|
|
|
|
<4 x i32>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
|
|
|
%3 = load <4 x i32>, <4 x i32>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <4 x i32> %2, %3
|
|
|
|
%5 = sub <4 x i32> %1, %4
|
|
|
|
store <4 x i32> %5, <4 x i32>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @msubv_v2i64(<2 x i64>* %d, <2 x i64>* %a, <2 x i64>* %b,
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: msubv_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($7)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w2, 0($5)
|
|
|
|
; ALL-NEXT: msubv.d $w2, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w2, 0($4)
|
|
|
|
<2 x i64>* %c) nounwind {
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
|
|
|
%3 = load <2 x i64>, <2 x i64>* %c
|
2013-10-11 10:50:42 +00:00
|
|
|
%4 = mul <2 x i64> %2, %3
|
|
|
|
%5 = sub <2 x i64> %1, %4
|
|
|
|
store <2 x i64> %5, <2 x i64>* %d
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
2013-09-11 12:39:25 +00:00
|
|
|
define void @div_s_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_s_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_s.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sdiv <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_s_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_s_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_s.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sdiv <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_s_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_s_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_s.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sdiv <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_s_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_s_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_s.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = sdiv <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_u_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_u_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_u.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = udiv <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_u_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_u_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_u.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = udiv <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_u_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_u_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_u.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = udiv <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @div_u_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: div_u_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: div_u.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-09-11 12:39:25 +00:00
|
|
|
%3 = udiv <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
2013-10-01 10:22:35 +00:00
|
|
|
|
|
|
|
define void @mod_s_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_s_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_s.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = srem <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_s_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_s_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_s.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = srem <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_s_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_s_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_s.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = srem <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_s_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_s_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_s.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = srem <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_u_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_u_v16i8:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.b $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.b $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_u.b $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.b $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <16 x i8>, <16 x i8>* %a
|
|
|
|
%2 = load <16 x i8>, <16 x i8>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = urem <16 x i8> %1, %2
|
|
|
|
store <16 x i8> %3, <16 x i8>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_u_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_u_v8i16:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.h $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.h $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_u.h $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.h $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <8 x i16>, <8 x i16>* %a
|
|
|
|
%2 = load <8 x i16>, <8 x i16>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = urem <8 x i16> %1, %2
|
|
|
|
store <8 x i16> %3, <8 x i16>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_u_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_u_v4i32:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.w $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.w $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_u.w $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.w $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <4 x i32>, <4 x i32>* %a
|
|
|
|
%2 = load <4 x i32>, <4 x i32>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = urem <4 x i32> %1, %2
|
|
|
|
store <4 x i32> %3, <4 x i32>* %c
|
|
|
|
ret void
|
|
|
|
}
|
|
|
|
|
|
|
|
define void @mod_u_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind {
|
2019-05-23 18:08:13 +00:00
|
|
|
; ALL-LABEL: mod_u_v2i64:
|
|
|
|
; ALL: # %bb.0:
|
|
|
|
; ALL-NEXT: ld.d $w0, 0($6)
|
|
|
|
; ALL-NEXT: ld.d $w1, 0($5)
|
|
|
|
; ALL-NEXT: mod_u.d $w0, $w1, $w0
|
|
|
|
; ALL-NEXT: jr $ra
|
|
|
|
; ALL-NEXT: st.d $w0, 0($4)
|
2015-02-27 21:17:42 +00:00
|
|
|
%1 = load <2 x i64>, <2 x i64>* %a
|
|
|
|
%2 = load <2 x i64>, <2 x i64>* %b
|
2013-10-01 10:22:35 +00:00
|
|
|
%3 = urem <2 x i64> %1, %2
|
|
|
|
store <2 x i64> %3, <2 x i64>* %c
|
|
|
|
ret void
|
|
|
|
}
|