mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-28 13:51:09 +00:00
[PowerPC] Set setMaxAtomicSizeInBitsSupported appropriately for 32-bit PowerPC in PPCTargetLowering
Reviewed By: nemanjai Differential Revision: https://reviews.llvm.org/D86165
This commit is contained in:
parent
621df3de9a
commit
99286ab2f0
@ -1199,6 +1199,9 @@ PPCTargetLowering::PPCTargetLowering(const PPCTargetMachine &TM,
|
||||
setLibcallName(RTLIB::SRA_I128, nullptr);
|
||||
}
|
||||
|
||||
if (!isPPC64)
|
||||
setMaxAtomicSizeInBitsSupported(32);
|
||||
|
||||
setStackPointerRegisterToSaveRestore(isPPC64 ? PPC::X1 : PPC::R1);
|
||||
|
||||
// We have target-specific dag combine patterns for the following nodes:
|
||||
|
@ -1,3 +1,4 @@
|
||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||
; RUN: llc < %s -mtriple=powerpc-unknown-linux-gnu -verify-machineinstrs -ppc-asm-full-reg-names | FileCheck %s --check-prefix=CHECK --check-prefix=PPC32
|
||||
; FIXME: -verify-machineinstrs currently fail on ppc64 (mismatched register/instruction).
|
||||
; This is already checked for in Atomics-64.ll
|
||||
@ -8,9 +9,25 @@
|
||||
|
||||
; Indexed version of loads
|
||||
define i8 @load_x_i8_seq_cst([100000 x i8]* %mem) {
|
||||
; CHECK-LABEL: load_x_i8_seq_cst
|
||||
; CHECK: sync
|
||||
; CHECK: lbzx [[VAL:r[0-9]+]]
|
||||
; PPC32-LABEL: load_x_i8_seq_cst:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: lis r4, 1
|
||||
; PPC32-NEXT: sync
|
||||
; PPC32-NEXT: ori r4, r4, 24464
|
||||
; PPC32-NEXT: lbzx r3, r3, r4
|
||||
; PPC32-NEXT: lwsync
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: load_x_i8_seq_cst:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: lis r4, 1
|
||||
; PPC64-NEXT: sync
|
||||
; PPC64-NEXT: ori r4, r4, 24464
|
||||
; PPC64-NEXT: lbzx r3, r3, r4
|
||||
; PPC64-NEXT: cmpd cr7, r3, r3
|
||||
; PPC64-NEXT: bne- cr7, .+4
|
||||
; PPC64-NEXT: isync
|
||||
; PPC64-NEXT: blr
|
||||
; CHECK-PPC32: lwsync
|
||||
; CHECK-PPC64: cmpw [[CR:cr[0-9]+]], [[VAL]], [[VAL]]
|
||||
; CHECK-PPC64: bne- [[CR]], .+4
|
||||
@ -20,8 +37,23 @@ define i8 @load_x_i8_seq_cst([100000 x i8]* %mem) {
|
||||
ret i8 %val
|
||||
}
|
||||
define i16 @load_x_i16_acquire([100000 x i16]* %mem) {
|
||||
; CHECK-LABEL: load_x_i16_acquire
|
||||
; CHECK: lhzx [[VAL:r[0-9]+]]
|
||||
; PPC32-LABEL: load_x_i16_acquire:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: lis r4, 2
|
||||
; PPC32-NEXT: ori r4, r4, 48928
|
||||
; PPC32-NEXT: lhzx r3, r3, r4
|
||||
; PPC32-NEXT: lwsync
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: load_x_i16_acquire:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: lis r4, 2
|
||||
; PPC64-NEXT: ori r4, r4, 48928
|
||||
; PPC64-NEXT: lhzx r3, r3, r4
|
||||
; PPC64-NEXT: cmpd cr7, r3, r3
|
||||
; PPC64-NEXT: bne- cr7, .+4
|
||||
; PPC64-NEXT: isync
|
||||
; PPC64-NEXT: blr
|
||||
; CHECK-PPC32: lwsync
|
||||
; CHECK-PPC64: cmpw [[CR:cr[0-9]+]], [[VAL]], [[VAL]]
|
||||
; CHECK-PPC64: bne- [[CR]], .+4
|
||||
@ -31,19 +63,39 @@ define i16 @load_x_i16_acquire([100000 x i16]* %mem) {
|
||||
ret i16 %val
|
||||
}
|
||||
define i32 @load_x_i32_monotonic([100000 x i32]* %mem) {
|
||||
; CHECK-LABEL: load_x_i32_monotonic
|
||||
; CHECK: lwzx
|
||||
; CHECK-NOT: sync
|
||||
; CHECK-LABEL: load_x_i32_monotonic:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lis r4, 5
|
||||
; CHECK-NEXT: ori r4, r4, 32320
|
||||
; CHECK-NEXT: lwzx r3, r3, r4
|
||||
; CHECK-NEXT: blr
|
||||
%ptr = getelementptr inbounds [100000 x i32], [100000 x i32]* %mem, i64 0, i64 90000
|
||||
%val = load atomic i32, i32* %ptr monotonic, align 4
|
||||
ret i32 %val
|
||||
}
|
||||
define i64 @load_x_i64_unordered([100000 x i64]* %mem) {
|
||||
; CHECK-LABEL: load_x_i64_unordered
|
||||
; PPC32: __sync_
|
||||
; PPC64-NOT: __sync_
|
||||
; PPC64: ldx
|
||||
; CHECK-NOT: sync
|
||||
; PPC32-LABEL: load_x_i64_unordered:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: mflr r0
|
||||
; PPC32-NEXT: stw r0, 4(r1)
|
||||
; PPC32-NEXT: stwu r1, -16(r1)
|
||||
; PPC32-NEXT: .cfi_def_cfa_offset 16
|
||||
; PPC32-NEXT: .cfi_offset lr, 4
|
||||
; PPC32-NEXT: addi r3, r3, -896
|
||||
; PPC32-NEXT: addis r3, r3, 11
|
||||
; PPC32-NEXT: li r4, 0
|
||||
; PPC32-NEXT: bl __atomic_load_8
|
||||
; PPC32-NEXT: lwz r0, 20(r1)
|
||||
; PPC32-NEXT: addi r1, r1, 16
|
||||
; PPC32-NEXT: mtlr r0
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: load_x_i64_unordered:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: lis r4, 10
|
||||
; PPC64-NEXT: ori r4, r4, 64640
|
||||
; PPC64-NEXT: ldx r3, r3, r4
|
||||
; PPC64-NEXT: blr
|
||||
%ptr = getelementptr inbounds [100000 x i64], [100000 x i64]* %mem, i64 0, i64 90000
|
||||
%val = load atomic i64, i64* %ptr unordered, align 8
|
||||
ret i64 %val
|
||||
@ -51,35 +103,69 @@ define i64 @load_x_i64_unordered([100000 x i64]* %mem) {
|
||||
|
||||
; Indexed version of stores
|
||||
define void @store_x_i8_seq_cst([100000 x i8]* %mem) {
|
||||
; CHECK-LABEL: store_x_i8_seq_cst
|
||||
; CHECK: sync
|
||||
; CHECK: stbx
|
||||
; CHECK-LABEL: store_x_i8_seq_cst:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lis r4, 1
|
||||
; CHECK-NEXT: ori r4, r4, 24464
|
||||
; CHECK-NEXT: li r5, 42
|
||||
; CHECK-NEXT: sync
|
||||
; CHECK-NEXT: stbx r5, r3, r4
|
||||
; CHECK-NEXT: blr
|
||||
%ptr = getelementptr inbounds [100000 x i8], [100000 x i8]* %mem, i64 0, i64 90000
|
||||
store atomic i8 42, i8* %ptr seq_cst, align 1
|
||||
ret void
|
||||
}
|
||||
define void @store_x_i16_release([100000 x i16]* %mem) {
|
||||
; CHECK-LABEL: store_x_i16_release
|
||||
; CHECK: lwsync
|
||||
; CHECK: sthx
|
||||
; CHECK-LABEL: store_x_i16_release:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lis r4, 2
|
||||
; CHECK-NEXT: ori r4, r4, 48928
|
||||
; CHECK-NEXT: li r5, 42
|
||||
; CHECK-NEXT: lwsync
|
||||
; CHECK-NEXT: sthx r5, r3, r4
|
||||
; CHECK-NEXT: blr
|
||||
%ptr = getelementptr inbounds [100000 x i16], [100000 x i16]* %mem, i64 0, i64 90000
|
||||
store atomic i16 42, i16* %ptr release, align 2
|
||||
ret void
|
||||
}
|
||||
define void @store_x_i32_monotonic([100000 x i32]* %mem) {
|
||||
; CHECK-LABEL: store_x_i32_monotonic
|
||||
; CHECK-NOT: sync
|
||||
; CHECK: stwx
|
||||
; CHECK-LABEL: store_x_i32_monotonic:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lis r4, 5
|
||||
; CHECK-NEXT: ori r4, r4, 32320
|
||||
; CHECK-NEXT: li r5, 42
|
||||
; CHECK-NEXT: stwx r5, r3, r4
|
||||
; CHECK-NEXT: blr
|
||||
%ptr = getelementptr inbounds [100000 x i32], [100000 x i32]* %mem, i64 0, i64 90000
|
||||
store atomic i32 42, i32* %ptr monotonic, align 4
|
||||
ret void
|
||||
}
|
||||
define void @store_x_i64_unordered([100000 x i64]* %mem) {
|
||||
; CHECK-LABEL: store_x_i64_unordered
|
||||
; CHECK-NOT: sync
|
||||
; PPC32: __sync_
|
||||
; PPC64-NOT: __sync_
|
||||
; PPC64: stdx
|
||||
; PPC32-LABEL: store_x_i64_unordered:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: mflr r0
|
||||
; PPC32-NEXT: stw r0, 4(r1)
|
||||
; PPC32-NEXT: stwu r1, -16(r1)
|
||||
; PPC32-NEXT: .cfi_def_cfa_offset 16
|
||||
; PPC32-NEXT: .cfi_offset lr, 4
|
||||
; PPC32-NEXT: addi r3, r3, -896
|
||||
; PPC32-NEXT: addis r3, r3, 11
|
||||
; PPC32-NEXT: li r5, 0
|
||||
; PPC32-NEXT: li r6, 42
|
||||
; PPC32-NEXT: li r7, 0
|
||||
; PPC32-NEXT: bl __atomic_store_8
|
||||
; PPC32-NEXT: lwz r0, 20(r1)
|
||||
; PPC32-NEXT: addi r1, r1, 16
|
||||
; PPC32-NEXT: mtlr r0
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: store_x_i64_unordered:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: lis r4, 10
|
||||
; PPC64-NEXT: ori r4, r4, 64640
|
||||
; PPC64-NEXT: li r5, 42
|
||||
; PPC64-NEXT: stdx r5, r3, r4
|
||||
; PPC64-NEXT: blr
|
||||
%ptr = getelementptr inbounds [100000 x i64], [100000 x i64]* %mem, i64 0, i64 90000
|
||||
store atomic i64 42, i64* %ptr unordered, align 8
|
||||
ret void
|
||||
|
@ -1,3 +1,4 @@
|
||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --force-update
|
||||
; RUN: llc -verify-machineinstrs < %s -mtriple=powerpc-unknown-linux-gnu -verify-machineinstrs -ppc-asm-full-reg-names | FileCheck %s --check-prefix=CHECK --check-prefix=PPC32
|
||||
; This is already checked for in Atomics-64.ll
|
||||
; RUN: llc -verify-machineinstrs < %s -mtriple=powerpc64-unknown-linux-gnu -ppc-asm-full-reg-names | FileCheck %s --check-prefix=CHECK --check-prefix=PPC64
|
||||
@ -9,22 +10,35 @@
|
||||
; We first check loads, for all sizes from i8 to i64.
|
||||
; We also vary orderings to check for barriers.
|
||||
define i8 @load_i8_unordered(i8* %mem) {
|
||||
; CHECK-LABEL: load_i8_unordered
|
||||
; CHECK: lbz
|
||||
; CHECK-NOT: sync
|
||||
; CHECK-LABEL: load_i8_unordered:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lbz r3, 0(r3)
|
||||
; CHECK-NEXT: blr
|
||||
%val = load atomic i8, i8* %mem unordered, align 1
|
||||
ret i8 %val
|
||||
}
|
||||
define i16 @load_i16_monotonic(i16* %mem) {
|
||||
; CHECK-LABEL: load_i16_monotonic
|
||||
; CHECK: lhz
|
||||
; CHECK-NOT: sync
|
||||
; CHECK-LABEL: load_i16_monotonic:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lhz r3, 0(r3)
|
||||
; CHECK-NEXT: blr
|
||||
%val = load atomic i16, i16* %mem monotonic, align 2
|
||||
ret i16 %val
|
||||
}
|
||||
define i32 @load_i32_acquire(i32* %mem) {
|
||||
; CHECK-LABEL: load_i32_acquire
|
||||
; CHECK: lwz [[VAL:r[0-9]+]]
|
||||
; PPC32-LABEL: load_i32_acquire:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: lwz r3, 0(r3)
|
||||
; PPC32-NEXT: lwsync
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: load_i32_acquire:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: lwz r3, 0(r3)
|
||||
; PPC64-NEXT: cmpd cr7, r3, r3
|
||||
; PPC64-NEXT: bne- cr7, .+4
|
||||
; PPC64-NEXT: isync
|
||||
; PPC64-NEXT: blr
|
||||
%val = load atomic i32, i32* %mem acquire, align 4
|
||||
; CHECK-PPC32: lwsync
|
||||
; CHECK-PPC64: cmpw [[CR:cr[0-9]+]], [[VAL]], [[VAL]]
|
||||
@ -33,11 +47,28 @@ define i32 @load_i32_acquire(i32* %mem) {
|
||||
ret i32 %val
|
||||
}
|
||||
define i64 @load_i64_seq_cst(i64* %mem) {
|
||||
; CHECK-LABEL: load_i64_seq_cst
|
||||
; CHECK: sync
|
||||
; PPC32: __sync_
|
||||
; PPC64-NOT: __sync_
|
||||
; PPC64: ld [[VAL:r[0-9]+]]
|
||||
; PPC32-LABEL: load_i64_seq_cst:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: mflr r0
|
||||
; PPC32-NEXT: stw r0, 4(r1)
|
||||
; PPC32-NEXT: stwu r1, -16(r1)
|
||||
; PPC32-NEXT: .cfi_def_cfa_offset 16
|
||||
; PPC32-NEXT: .cfi_offset lr, 4
|
||||
; PPC32-NEXT: li r4, 5
|
||||
; PPC32-NEXT: bl __atomic_load_8
|
||||
; PPC32-NEXT: lwz r0, 20(r1)
|
||||
; PPC32-NEXT: addi r1, r1, 16
|
||||
; PPC32-NEXT: mtlr r0
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: load_i64_seq_cst:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: sync
|
||||
; PPC64-NEXT: ld r3, 0(r3)
|
||||
; PPC64-NEXT: cmpd cr7, r3, r3
|
||||
; PPC64-NEXT: bne- cr7, .+4
|
||||
; PPC64-NEXT: isync
|
||||
; PPC64-NEXT: blr
|
||||
%val = load atomic i64, i64* %mem seq_cst, align 8
|
||||
; CHECK-PPC32: lwsync
|
||||
; CHECK-PPC64: cmpw [[CR:cr[0-9]+]], [[VAL]], [[VAL]]
|
||||
@ -48,95 +79,401 @@ define i64 @load_i64_seq_cst(i64* %mem) {
|
||||
|
||||
; Stores
|
||||
define void @store_i8_unordered(i8* %mem) {
|
||||
; CHECK-LABEL: store_i8_unordered
|
||||
; CHECK-NOT: sync
|
||||
; CHECK: stb
|
||||
; CHECK-LABEL: store_i8_unordered:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: li r4, 42
|
||||
; CHECK-NEXT: stb r4, 0(r3)
|
||||
; CHECK-NEXT: blr
|
||||
store atomic i8 42, i8* %mem unordered, align 1
|
||||
ret void
|
||||
}
|
||||
define void @store_i16_monotonic(i16* %mem) {
|
||||
; CHECK-LABEL: store_i16_monotonic
|
||||
; CHECK-NOT: sync
|
||||
; CHECK: sth
|
||||
; CHECK-LABEL: store_i16_monotonic:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: li r4, 42
|
||||
; CHECK-NEXT: sth r4, 0(r3)
|
||||
; CHECK-NEXT: blr
|
||||
store atomic i16 42, i16* %mem monotonic, align 2
|
||||
ret void
|
||||
}
|
||||
define void @store_i32_release(i32* %mem) {
|
||||
; CHECK-LABEL: store_i32_release
|
||||
; CHECK: lwsync
|
||||
; CHECK: stw
|
||||
; CHECK-LABEL: store_i32_release:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: li r4, 42
|
||||
; CHECK-NEXT: lwsync
|
||||
; CHECK-NEXT: stw r4, 0(r3)
|
||||
; CHECK-NEXT: blr
|
||||
store atomic i32 42, i32* %mem release, align 4
|
||||
ret void
|
||||
}
|
||||
define void @store_i64_seq_cst(i64* %mem) {
|
||||
; CHECK-LABEL: store_i64_seq_cst
|
||||
; CHECK: sync
|
||||
; PPC32: __sync_
|
||||
; PPC64-NOT: __sync_
|
||||
; PPC64: std
|
||||
; PPC32-LABEL: store_i64_seq_cst:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: mflr r0
|
||||
; PPC32-NEXT: stw r0, 4(r1)
|
||||
; PPC32-NEXT: stwu r1, -16(r1)
|
||||
; PPC32-NEXT: .cfi_def_cfa_offset 16
|
||||
; PPC32-NEXT: .cfi_offset lr, 4
|
||||
; PPC32-NEXT: li r5, 0
|
||||
; PPC32-NEXT: li r6, 42
|
||||
; PPC32-NEXT: li r7, 5
|
||||
; PPC32-NEXT: bl __atomic_store_8
|
||||
; PPC32-NEXT: lwz r0, 20(r1)
|
||||
; PPC32-NEXT: addi r1, r1, 16
|
||||
; PPC32-NEXT: mtlr r0
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: store_i64_seq_cst:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: li r4, 42
|
||||
; PPC64-NEXT: sync
|
||||
; PPC64-NEXT: std r4, 0(r3)
|
||||
; PPC64-NEXT: blr
|
||||
store atomic i64 42, i64* %mem seq_cst, align 8
|
||||
ret void
|
||||
}
|
||||
|
||||
; Atomic CmpXchg
|
||||
define i8 @cas_strong_i8_sc_sc(i8* %mem) {
|
||||
; CHECK-LABEL: cas_strong_i8_sc_sc
|
||||
; CHECK: sync
|
||||
; PPC32-LABEL: cas_strong_i8_sc_sc:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: rlwinm r8, r3, 3, 27, 28
|
||||
; PPC32-NEXT: li r5, 1
|
||||
; PPC32-NEXT: li r6, 0
|
||||
; PPC32-NEXT: li r7, 255
|
||||
; PPC32-NEXT: rlwinm r4, r3, 0, 0, 29
|
||||
; PPC32-NEXT: xori r3, r8, 24
|
||||
; PPC32-NEXT: slw r5, r5, r3
|
||||
; PPC32-NEXT: slw r8, r6, r3
|
||||
; PPC32-NEXT: slw r6, r7, r3
|
||||
; PPC32-NEXT: and r7, r5, r6
|
||||
; PPC32-NEXT: and r8, r8, r6
|
||||
; PPC32-NEXT: sync
|
||||
; PPC32-NEXT: .LBB8_1:
|
||||
; PPC32-NEXT: lwarx r9, 0, r4
|
||||
; PPC32-NEXT: and r5, r9, r6
|
||||
; PPC32-NEXT: cmpw r5, r8
|
||||
; PPC32-NEXT: bne cr0, .LBB8_3
|
||||
; PPC32-NEXT: # %bb.2:
|
||||
; PPC32-NEXT: andc r9, r9, r6
|
||||
; PPC32-NEXT: or r9, r9, r7
|
||||
; PPC32-NEXT: stwcx. r9, 0, r4
|
||||
; PPC32-NEXT: bne cr0, .LBB8_1
|
||||
; PPC32-NEXT: b .LBB8_4
|
||||
; PPC32-NEXT: .LBB8_3:
|
||||
; PPC32-NEXT: stwcx. r9, 0, r4
|
||||
; PPC32-NEXT: .LBB8_4:
|
||||
; PPC32-NEXT: srw r3, r5, r3
|
||||
; PPC32-NEXT: lwsync
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: cas_strong_i8_sc_sc:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: rlwinm r8, r3, 3, 27, 28
|
||||
; PPC64-NEXT: li r5, 1
|
||||
; PPC64-NEXT: li r6, 0
|
||||
; PPC64-NEXT: li r7, 255
|
||||
; PPC64-NEXT: rldicr r4, r3, 0, 61
|
||||
; PPC64-NEXT: xori r3, r8, 24
|
||||
; PPC64-NEXT: slw r5, r5, r3
|
||||
; PPC64-NEXT: slw r8, r6, r3
|
||||
; PPC64-NEXT: slw r6, r7, r3
|
||||
; PPC64-NEXT: and r7, r5, r6
|
||||
; PPC64-NEXT: and r8, r8, r6
|
||||
; PPC64-NEXT: sync
|
||||
; PPC64-NEXT: .LBB8_1:
|
||||
; PPC64-NEXT: lwarx r9, 0, r4
|
||||
; PPC64-NEXT: and r5, r9, r6
|
||||
; PPC64-NEXT: cmpw r5, r8
|
||||
; PPC64-NEXT: bne cr0, .LBB8_3
|
||||
; PPC64-NEXT: # %bb.2:
|
||||
; PPC64-NEXT: andc r9, r9, r6
|
||||
; PPC64-NEXT: or r9, r9, r7
|
||||
; PPC64-NEXT: stwcx. r9, 0, r4
|
||||
; PPC64-NEXT: bne cr0, .LBB8_1
|
||||
; PPC64-NEXT: b .LBB8_4
|
||||
; PPC64-NEXT: .LBB8_3:
|
||||
; PPC64-NEXT: stwcx. r9, 0, r4
|
||||
; PPC64-NEXT: .LBB8_4:
|
||||
; PPC64-NEXT: srw r3, r5, r3
|
||||
; PPC64-NEXT: lwsync
|
||||
; PPC64-NEXT: blr
|
||||
%val = cmpxchg i8* %mem, i8 0, i8 1 seq_cst seq_cst
|
||||
; CHECK: lwsync
|
||||
%loaded = extractvalue { i8, i1} %val, 0
|
||||
ret i8 %loaded
|
||||
}
|
||||
define i16 @cas_weak_i16_acquire_acquire(i16* %mem) {
|
||||
; CHECK-LABEL: cas_weak_i16_acquire_acquire
|
||||
;CHECK-NOT: sync
|
||||
; PPC32-LABEL: cas_weak_i16_acquire_acquire:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: li r6, 0
|
||||
; PPC32-NEXT: rlwinm r4, r3, 3, 27, 27
|
||||
; PPC32-NEXT: li r5, 1
|
||||
; PPC32-NEXT: ori r7, r6, 65535
|
||||
; PPC32-NEXT: xori r4, r4, 16
|
||||
; PPC32-NEXT: slw r8, r5, r4
|
||||
; PPC32-NEXT: slw r9, r6, r4
|
||||
; PPC32-NEXT: slw r5, r7, r4
|
||||
; PPC32-NEXT: rlwinm r3, r3, 0, 0, 29
|
||||
; PPC32-NEXT: and r6, r8, r5
|
||||
; PPC32-NEXT: and r8, r9, r5
|
||||
; PPC32-NEXT: .LBB9_1:
|
||||
; PPC32-NEXT: lwarx r9, 0, r3
|
||||
; PPC32-NEXT: and r7, r9, r5
|
||||
; PPC32-NEXT: cmpw r7, r8
|
||||
; PPC32-NEXT: bne cr0, .LBB9_3
|
||||
; PPC32-NEXT: # %bb.2:
|
||||
; PPC32-NEXT: andc r9, r9, r5
|
||||
; PPC32-NEXT: or r9, r9, r6
|
||||
; PPC32-NEXT: stwcx. r9, 0, r3
|
||||
; PPC32-NEXT: bne cr0, .LBB9_1
|
||||
; PPC32-NEXT: b .LBB9_4
|
||||
; PPC32-NEXT: .LBB9_3:
|
||||
; PPC32-NEXT: stwcx. r9, 0, r3
|
||||
; PPC32-NEXT: .LBB9_4:
|
||||
; PPC32-NEXT: srw r3, r7, r4
|
||||
; PPC32-NEXT: lwsync
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: cas_weak_i16_acquire_acquire:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: li r6, 0
|
||||
; PPC64-NEXT: rlwinm r4, r3, 3, 27, 27
|
||||
; PPC64-NEXT: li r5, 1
|
||||
; PPC64-NEXT: ori r7, r6, 65535
|
||||
; PPC64-NEXT: xori r4, r4, 16
|
||||
; PPC64-NEXT: slw r8, r5, r4
|
||||
; PPC64-NEXT: slw r9, r6, r4
|
||||
; PPC64-NEXT: slw r5, r7, r4
|
||||
; PPC64-NEXT: rldicr r3, r3, 0, 61
|
||||
; PPC64-NEXT: and r6, r8, r5
|
||||
; PPC64-NEXT: and r8, r9, r5
|
||||
; PPC64-NEXT: .LBB9_1:
|
||||
; PPC64-NEXT: lwarx r9, 0, r3
|
||||
; PPC64-NEXT: and r7, r9, r5
|
||||
; PPC64-NEXT: cmpw r7, r8
|
||||
; PPC64-NEXT: bne cr0, .LBB9_3
|
||||
; PPC64-NEXT: # %bb.2:
|
||||
; PPC64-NEXT: andc r9, r9, r5
|
||||
; PPC64-NEXT: or r9, r9, r6
|
||||
; PPC64-NEXT: stwcx. r9, 0, r3
|
||||
; PPC64-NEXT: bne cr0, .LBB9_1
|
||||
; PPC64-NEXT: b .LBB9_4
|
||||
; PPC64-NEXT: .LBB9_3:
|
||||
; PPC64-NEXT: stwcx. r9, 0, r3
|
||||
; PPC64-NEXT: .LBB9_4:
|
||||
; PPC64-NEXT: srw r3, r7, r4
|
||||
; PPC64-NEXT: lwsync
|
||||
; PPC64-NEXT: blr
|
||||
%val = cmpxchg weak i16* %mem, i16 0, i16 1 acquire acquire
|
||||
; CHECK: lwsync
|
||||
%loaded = extractvalue { i16, i1} %val, 0
|
||||
ret i16 %loaded
|
||||
}
|
||||
define i32 @cas_strong_i32_acqrel_acquire(i32* %mem) {
|
||||
; CHECK-LABEL: cas_strong_i32_acqrel_acquire
|
||||
; CHECK: lwsync
|
||||
; CHECK-LABEL: cas_strong_i32_acqrel_acquire:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: li r5, 1
|
||||
; CHECK-NEXT: li r6, 0
|
||||
; CHECK-NEXT: lwsync
|
||||
; CHECK-NEXT: .LBB10_1:
|
||||
; CHECK-NEXT: lwarx r4, 0, r3
|
||||
; CHECK-NEXT: cmpw r6, r4
|
||||
; CHECK-NEXT: bne cr0, .LBB10_3
|
||||
; CHECK-NEXT: # %bb.2:
|
||||
; CHECK-NEXT: stwcx. r5, 0, r3
|
||||
; CHECK-NEXT: bne cr0, .LBB10_1
|
||||
; CHECK-NEXT: b .LBB10_4
|
||||
; CHECK-NEXT: .LBB10_3:
|
||||
; CHECK-NEXT: stwcx. r4, 0, r3
|
||||
; CHECK-NEXT: .LBB10_4:
|
||||
; CHECK-NEXT: mr r3, r4
|
||||
; CHECK-NEXT: lwsync
|
||||
; CHECK-NEXT: blr
|
||||
%val = cmpxchg i32* %mem, i32 0, i32 1 acq_rel acquire
|
||||
; CHECK: lwsync
|
||||
%loaded = extractvalue { i32, i1} %val, 0
|
||||
ret i32 %loaded
|
||||
}
|
||||
define i64 @cas_weak_i64_release_monotonic(i64* %mem) {
|
||||
; CHECK-LABEL: cas_weak_i64_release_monotonic
|
||||
; CHECK: lwsync
|
||||
; PPC32-LABEL: cas_weak_i64_release_monotonic:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: mflr r0
|
||||
; PPC32-NEXT: stw r0, 4(r1)
|
||||
; PPC32-NEXT: stwu r1, -16(r1)
|
||||
; PPC32-NEXT: .cfi_def_cfa_offset 16
|
||||
; PPC32-NEXT: .cfi_offset lr, 4
|
||||
; PPC32-NEXT: li r4, 0
|
||||
; PPC32-NEXT: stw r4, 12(r1)
|
||||
; PPC32-NEXT: li r5, 0
|
||||
; PPC32-NEXT: stw r4, 8(r1)
|
||||
; PPC32-NEXT: addi r4, r1, 8
|
||||
; PPC32-NEXT: li r6, 1
|
||||
; PPC32-NEXT: li r7, 3
|
||||
; PPC32-NEXT: li r8, 0
|
||||
; PPC32-NEXT: bl __atomic_compare_exchange_8
|
||||
; PPC32-NEXT: lwz r4, 12(r1)
|
||||
; PPC32-NEXT: lwz r3, 8(r1)
|
||||
; PPC32-NEXT: lwz r0, 20(r1)
|
||||
; PPC32-NEXT: addi r1, r1, 16
|
||||
; PPC32-NEXT: mtlr r0
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: cas_weak_i64_release_monotonic:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: li r5, 1
|
||||
; PPC64-NEXT: li r6, 0
|
||||
; PPC64-NEXT: lwsync
|
||||
; PPC64-NEXT: .LBB11_1:
|
||||
; PPC64-NEXT: ldarx r4, 0, r3
|
||||
; PPC64-NEXT: cmpd r6, r4
|
||||
; PPC64-NEXT: bne cr0, .LBB11_4
|
||||
; PPC64-NEXT: # %bb.2:
|
||||
; PPC64-NEXT: stdcx. r5, 0, r3
|
||||
; PPC64-NEXT: bne cr0, .LBB11_1
|
||||
; PPC64-NEXT: # %bb.3:
|
||||
; PPC64-NEXT: mr r3, r4
|
||||
; PPC64-NEXT: blr
|
||||
; PPC64-NEXT: .LBB11_4:
|
||||
; PPC64-NEXT: stdcx. r4, 0, r3
|
||||
; PPC64-NEXT: mr r3, r4
|
||||
; PPC64-NEXT: blr
|
||||
%val = cmpxchg weak i64* %mem, i64 0, i64 1 release monotonic
|
||||
; CHECK-NOT: [sync ]
|
||||
%loaded = extractvalue { i64, i1} %val, 0
|
||||
ret i64 %loaded
|
||||
}
|
||||
|
||||
; AtomicRMW
|
||||
define i8 @add_i8_monotonic(i8* %mem, i8 %operand) {
|
||||
; CHECK-LABEL: add_i8_monotonic
|
||||
; CHECK-NOT: sync
|
||||
; PPC32-LABEL: add_i8_monotonic:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: rlwinm r7, r3, 3, 27, 28
|
||||
; PPC32-NEXT: li r6, 255
|
||||
; PPC32-NEXT: rlwinm r5, r3, 0, 0, 29
|
||||
; PPC32-NEXT: xori r3, r7, 24
|
||||
; PPC32-NEXT: slw r4, r4, r3
|
||||
; PPC32-NEXT: slw r6, r6, r3
|
||||
; PPC32-NEXT: .LBB12_1:
|
||||
; PPC32-NEXT: lwarx r7, 0, r5
|
||||
; PPC32-NEXT: add r8, r4, r7
|
||||
; PPC32-NEXT: andc r9, r7, r6
|
||||
; PPC32-NEXT: and r8, r8, r6
|
||||
; PPC32-NEXT: or r8, r8, r9
|
||||
; PPC32-NEXT: stwcx. r8, 0, r5
|
||||
; PPC32-NEXT: bne cr0, .LBB12_1
|
||||
; PPC32-NEXT: # %bb.2:
|
||||
; PPC32-NEXT: srw r3, r7, r3
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: add_i8_monotonic:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: rlwinm r7, r3, 3, 27, 28
|
||||
; PPC64-NEXT: li r6, 255
|
||||
; PPC64-NEXT: rldicr r5, r3, 0, 61
|
||||
; PPC64-NEXT: xori r3, r7, 24
|
||||
; PPC64-NEXT: slw r4, r4, r3
|
||||
; PPC64-NEXT: slw r6, r6, r3
|
||||
; PPC64-NEXT: .LBB12_1:
|
||||
; PPC64-NEXT: lwarx r7, 0, r5
|
||||
; PPC64-NEXT: add r8, r4, r7
|
||||
; PPC64-NEXT: andc r9, r7, r6
|
||||
; PPC64-NEXT: and r8, r8, r6
|
||||
; PPC64-NEXT: or r8, r8, r9
|
||||
; PPC64-NEXT: stwcx. r8, 0, r5
|
||||
; PPC64-NEXT: bne cr0, .LBB12_1
|
||||
; PPC64-NEXT: # %bb.2:
|
||||
; PPC64-NEXT: srw r3, r7, r3
|
||||
; PPC64-NEXT: blr
|
||||
%val = atomicrmw add i8* %mem, i8 %operand monotonic
|
||||
ret i8 %val
|
||||
}
|
||||
define i16 @xor_i16_seq_cst(i16* %mem, i16 %operand) {
|
||||
; CHECK-LABEL: xor_i16_seq_cst
|
||||
; CHECK: sync
|
||||
; PPC32-LABEL: xor_i16_seq_cst:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: li r6, 0
|
||||
; PPC32-NEXT: rlwinm r7, r3, 3, 27, 27
|
||||
; PPC32-NEXT: rlwinm r5, r3, 0, 0, 29
|
||||
; PPC32-NEXT: ori r6, r6, 65535
|
||||
; PPC32-NEXT: xori r3, r7, 16
|
||||
; PPC32-NEXT: slw r4, r4, r3
|
||||
; PPC32-NEXT: slw r6, r6, r3
|
||||
; PPC32-NEXT: sync
|
||||
; PPC32-NEXT: .LBB13_1:
|
||||
; PPC32-NEXT: lwarx r7, 0, r5
|
||||
; PPC32-NEXT: xor r8, r4, r7
|
||||
; PPC32-NEXT: andc r9, r7, r6
|
||||
; PPC32-NEXT: and r8, r8, r6
|
||||
; PPC32-NEXT: or r8, r8, r9
|
||||
; PPC32-NEXT: stwcx. r8, 0, r5
|
||||
; PPC32-NEXT: bne cr0, .LBB13_1
|
||||
; PPC32-NEXT: # %bb.2:
|
||||
; PPC32-NEXT: srw r3, r7, r3
|
||||
; PPC32-NEXT: lwsync
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: xor_i16_seq_cst:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: li r6, 0
|
||||
; PPC64-NEXT: rlwinm r7, r3, 3, 27, 27
|
||||
; PPC64-NEXT: rldicr r5, r3, 0, 61
|
||||
; PPC64-NEXT: ori r6, r6, 65535
|
||||
; PPC64-NEXT: xori r3, r7, 16
|
||||
; PPC64-NEXT: slw r4, r4, r3
|
||||
; PPC64-NEXT: slw r6, r6, r3
|
||||
; PPC64-NEXT: sync
|
||||
; PPC64-NEXT: .LBB13_1:
|
||||
; PPC64-NEXT: lwarx r7, 0, r5
|
||||
; PPC64-NEXT: xor r8, r4, r7
|
||||
; PPC64-NEXT: andc r9, r7, r6
|
||||
; PPC64-NEXT: and r8, r8, r6
|
||||
; PPC64-NEXT: or r8, r8, r9
|
||||
; PPC64-NEXT: stwcx. r8, 0, r5
|
||||
; PPC64-NEXT: bne cr0, .LBB13_1
|
||||
; PPC64-NEXT: # %bb.2:
|
||||
; PPC64-NEXT: srw r3, r7, r3
|
||||
; PPC64-NEXT: lwsync
|
||||
; PPC64-NEXT: blr
|
||||
%val = atomicrmw xor i16* %mem, i16 %operand seq_cst
|
||||
; CHECK: lwsync
|
||||
ret i16 %val
|
||||
}
|
||||
define i32 @xchg_i32_acq_rel(i32* %mem, i32 %operand) {
|
||||
; CHECK-LABEL: xchg_i32_acq_rel
|
||||
; CHECK: lwsync
|
||||
; CHECK-LABEL: xchg_i32_acq_rel:
|
||||
; CHECK: # %bb.0:
|
||||
; CHECK-NEXT: lwsync
|
||||
; CHECK-NEXT: .LBB14_1:
|
||||
; CHECK-NEXT: lwarx r5, 0, r3
|
||||
; CHECK-NEXT: stwcx. r4, 0, r3
|
||||
; CHECK-NEXT: bne cr0, .LBB14_1
|
||||
; CHECK-NEXT: # %bb.2:
|
||||
; CHECK-NEXT: mr r3, r5
|
||||
; CHECK-NEXT: lwsync
|
||||
; CHECK-NEXT: blr
|
||||
%val = atomicrmw xchg i32* %mem, i32 %operand acq_rel
|
||||
; CHECK: lwsync
|
||||
ret i32 %val
|
||||
}
|
||||
define i64 @and_i64_release(i64* %mem, i64 %operand) {
|
||||
; CHECK-LABEL: and_i64_release
|
||||
; CHECK: lwsync
|
||||
; PPC32-LABEL: and_i64_release:
|
||||
; PPC32: # %bb.0:
|
||||
; PPC32-NEXT: mflr r0
|
||||
; PPC32-NEXT: stw r0, 4(r1)
|
||||
; PPC32-NEXT: stwu r1, -16(r1)
|
||||
; PPC32-NEXT: .cfi_def_cfa_offset 16
|
||||
; PPC32-NEXT: .cfi_offset lr, 4
|
||||
; PPC32-NEXT: li r7, 3
|
||||
; PPC32-NEXT: bl __atomic_fetch_and_8
|
||||
; PPC32-NEXT: lwz r0, 20(r1)
|
||||
; PPC32-NEXT: addi r1, r1, 16
|
||||
; PPC32-NEXT: mtlr r0
|
||||
; PPC32-NEXT: blr
|
||||
;
|
||||
; PPC64-LABEL: and_i64_release:
|
||||
; PPC64: # %bb.0:
|
||||
; PPC64-NEXT: lwsync
|
||||
; PPC64-NEXT: .LBB15_1:
|
||||
; PPC64-NEXT: ldarx r5, 0, r3
|
||||
; PPC64-NEXT: and r6, r4, r5
|
||||
; PPC64-NEXT: stdcx. r6, 0, r3
|
||||
; PPC64-NEXT: bne cr0, .LBB15_1
|
||||
; PPC64-NEXT: # %bb.2:
|
||||
; PPC64-NEXT: mr r3, r5
|
||||
; PPC64-NEXT: blr
|
||||
%val = atomicrmw and i64* %mem, i64 %operand release
|
||||
; CHECK-NOT: [sync ]
|
||||
ret i64 %val
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user