mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2025-03-04 10:30:30 +00:00
[X86][BMI] Added fast-isel tests for BMI1 intrinsics
A lot of the codegen is pretty awful for these as they are mostly implemented as generic bit twiddling ops llvm-svn: 272508
This commit is contained in:
parent
ed661e0e70
commit
abf516f3bc
165
test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll
Normal file
165
test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll
Normal file
@ -0,0 +1,165 @@
|
||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||
; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X64
|
||||
|
||||
; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/bmi-builtins.c
|
||||
|
||||
;
|
||||
; AMD Intrinsics
|
||||
;
|
||||
|
||||
define i64 @test__andn_u64(i64 %a0, i64 %a1) {
|
||||
; X64-LABEL: test__andn_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorq $-1, %rdi
|
||||
; X64-NEXT: andq %rsi, %rdi
|
||||
; X64-NEXT: movq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%xor = xor i64 %a0, -1
|
||||
%res = and i64 %xor, %a1
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test__bextr_u64(i64 %a0, i64 %a1) {
|
||||
; X64-LABEL: test__bextr_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: bextrq %rsi, %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%res = call i64 @llvm.x86.bmi.bextr.64(i64 %a0, i64 %a1)
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test__blsi_u64(i64 %a0) {
|
||||
; X64-LABEL: test__blsi_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorl %eax, %eax
|
||||
; X64-NEXT: subq %rdi, %rax
|
||||
; X64-NEXT: andq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%neg = sub i64 0, %a0
|
||||
%res = and i64 %a0, %neg
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test__blsmsk_u64(i64 %a0) {
|
||||
; X64-LABEL: test__blsmsk_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movq %rdi, %rax
|
||||
; X64-NEXT: subq $1, %rax
|
||||
; X64-NEXT: xorq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i64 %a0, 1
|
||||
%res = xor i64 %a0, %dec
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test__blsr_u64(i64 %a0) {
|
||||
; X64-LABEL: test__blsr_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movq %rdi, %rax
|
||||
; X64-NEXT: subq $1, %rax
|
||||
; X64-NEXT: andq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i64 %a0, 1
|
||||
%res = and i64 %a0, %dec
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test__tzcnt_u64(i64 %a0) {
|
||||
; X64-LABEL: test__tzcnt_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl $64, %ecx
|
||||
; X64-NEXT: tzcntq %rdi, %rax
|
||||
; X64-NEXT: cmovbq %rcx, %rax
|
||||
; X64-NEXT: retq
|
||||
%cmp = icmp ne i64 %a0, 0
|
||||
%cttz = call i64 @llvm.cttz.i64(i64 %a0, i1 true)
|
||||
%res = select i1 %cmp, i64 %cttz, i64 64
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
;
|
||||
; Intel intrinsics
|
||||
;
|
||||
|
||||
define i64 @test_andn_u64(i64 %a0, i64 %a1) {
|
||||
; X64-LABEL: test_andn_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorq $-1, %rdi
|
||||
; X64-NEXT: andq %rsi, %rdi
|
||||
; X64-NEXT: movq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%xor = xor i64 %a0, -1
|
||||
%res = and i64 %xor, %a1
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test_bextr_u64(i64 %a0, i32 %a1, i32 %a2) {
|
||||
; X64-LABEL: test_bextr_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: andl $255, %esi
|
||||
; X64-NEXT: andl $255, %edx
|
||||
; X64-NEXT: shll $8, %edx
|
||||
; X64-NEXT: orl %esi, %edx
|
||||
; X64-NEXT: movl %edx, %eax
|
||||
; X64-NEXT: bextrq %rax, %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%and1 = and i32 %a1, 255
|
||||
%and2 = and i32 %a2, 255
|
||||
%shl = shl i32 %and2, 8
|
||||
%or = or i32 %and1, %shl
|
||||
%zext = zext i32 %or to i64
|
||||
%res = call i64 @llvm.x86.bmi.bextr.64(i64 %a0, i64 %zext)
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test_blsi_u64(i64 %a0) {
|
||||
; X64-LABEL: test_blsi_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorl %eax, %eax
|
||||
; X64-NEXT: subq %rdi, %rax
|
||||
; X64-NEXT: andq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%neg = sub i64 0, %a0
|
||||
%res = and i64 %a0, %neg
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test_blsmsk_u64(i64 %a0) {
|
||||
; X64-LABEL: test_blsmsk_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movq %rdi, %rax
|
||||
; X64-NEXT: subq $1, %rax
|
||||
; X64-NEXT: xorq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i64 %a0, 1
|
||||
%res = xor i64 %a0, %dec
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test_blsr_u64(i64 %a0) {
|
||||
; X64-LABEL: test_blsr_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movq %rdi, %rax
|
||||
; X64-NEXT: subq $1, %rax
|
||||
; X64-NEXT: andq %rdi, %rax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i64 %a0, 1
|
||||
%res = and i64 %a0, %dec
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
define i64 @test_tzcnt_u64(i64 %a0) {
|
||||
; X64-LABEL: test_tzcnt_u64:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl $64, %ecx
|
||||
; X64-NEXT: tzcntq %rdi, %rax
|
||||
; X64-NEXT: cmovbq %rcx, %rax
|
||||
; X64-NEXT: retq
|
||||
%cmp = icmp ne i64 %a0, 0
|
||||
%cttz = call i64 @llvm.cttz.i64(i64 %a0, i1 true)
|
||||
%res = select i1 %cmp, i64 %cttz, i64 64
|
||||
ret i64 %res
|
||||
}
|
||||
|
||||
declare i64 @llvm.cttz.i64(i64, i1)
|
||||
declare i64 @llvm.x86.bmi.bextr.64(i64, i64)
|
326
test/CodeGen/X86/bmi-intrinsics-fast-isel.ll
Normal file
326
test/CodeGen/X86/bmi-intrinsics-fast-isel.ll
Normal file
@ -0,0 +1,326 @@
|
||||
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||
; RUN: llc < %s -fast-isel -mtriple=i686-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X32
|
||||
; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X64
|
||||
|
||||
; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/bmi-builtins.c
|
||||
|
||||
;
|
||||
; AMD Intrinsics
|
||||
;
|
||||
|
||||
define i16 @test__tzcnt_u16(i16 %a0) {
|
||||
; X32-LABEL: test__tzcnt_u16:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movzwl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: movzwl %ax, %ecx
|
||||
; X32-NEXT: cmpl $0, %ecx
|
||||
; X32-NEXT: jne .LBB0_1
|
||||
; X32-NEXT: # BB#2:
|
||||
; X32-NEXT: movw $16, %ax
|
||||
; X32-NEXT: retl
|
||||
; X32-NEXT: .LBB0_1:
|
||||
; X32-NEXT: tzcntw %ax, %ax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__tzcnt_u16:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movw $16, %cx
|
||||
; X64-NEXT: movzwl %di, %edx
|
||||
; X64-NEXT: tzcntw %dx, %ax
|
||||
; X64-NEXT: cmpl $0, %edx
|
||||
; X64-NEXT: cmovew %cx, %ax
|
||||
; X64-NEXT: retq
|
||||
%zext = zext i16 %a0 to i32
|
||||
%cmp = icmp ne i32 %zext, 0
|
||||
%cttz = call i16 @llvm.cttz.i16(i16 %a0, i1 true)
|
||||
%res = select i1 %cmp, i16 %cttz, i16 16
|
||||
ret i16 %res
|
||||
}
|
||||
|
||||
define i32 @test__andn_u32(i32 %a0, i32 %a1) {
|
||||
; X32-LABEL: test__andn_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: xorl $-1, %eax
|
||||
; X32-NEXT: andl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__andn_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorl $-1, %edi
|
||||
; X64-NEXT: andl %esi, %edi
|
||||
; X64-NEXT: movl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%xor = xor i32 %a0, -1
|
||||
%res = and i32 %xor, %a1
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test__bextr_u32(i32 %a0, i32 %a1) {
|
||||
; X32-LABEL: test__bextr_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: bextrl %eax, {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__bextr_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: bextrl %esi, %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%res = call i32 @llvm.x86.bmi.bextr.32(i32 %a0, i32 %a1)
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test__blsi_u32(i32 %a0) {
|
||||
; X32-LABEL: test__blsi_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: xorl %eax, %eax
|
||||
; X32-NEXT: subl %ecx, %eax
|
||||
; X32-NEXT: andl %ecx, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__blsi_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorl %eax, %eax
|
||||
; X64-NEXT: subl %edi, %eax
|
||||
; X64-NEXT: andl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%neg = sub i32 0, %a0
|
||||
%res = and i32 %a0, %neg
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test__blsmsk_u32(i32 %a0) {
|
||||
; X32-LABEL: test__blsmsk_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: movl %ecx, %eax
|
||||
; X32-NEXT: subl $1, %eax
|
||||
; X32-NEXT: xorl %ecx, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__blsmsk_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl %edi, %eax
|
||||
; X64-NEXT: subl $1, %eax
|
||||
; X64-NEXT: xorl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i32 %a0, 1
|
||||
%res = xor i32 %a0, %dec
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test__blsr_u32(i32 %a0) {
|
||||
; X32-LABEL: test__blsr_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: movl %ecx, %eax
|
||||
; X32-NEXT: subl $1, %eax
|
||||
; X32-NEXT: andl %ecx, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__blsr_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl %edi, %eax
|
||||
; X64-NEXT: subl $1, %eax
|
||||
; X64-NEXT: andl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i32 %a0, 1
|
||||
%res = and i32 %a0, %dec
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test__tzcnt_u32(i32 %a0) {
|
||||
; X32-LABEL: test__tzcnt_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: cmpl $0, %eax
|
||||
; X32-NEXT: jne .LBB6_1
|
||||
; X32-NEXT: # BB#2:
|
||||
; X32-NEXT: movl $32, %eax
|
||||
; X32-NEXT: retl
|
||||
; X32-NEXT: .LBB6_1:
|
||||
; X32-NEXT: tzcntl %eax, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test__tzcnt_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl $32, %ecx
|
||||
; X64-NEXT: tzcntl %edi, %eax
|
||||
; X64-NEXT: cmovbl %ecx, %eax
|
||||
; X64-NEXT: retq
|
||||
%cmp = icmp ne i32 %a0, 0
|
||||
%cttz = call i32 @llvm.cttz.i32(i32 %a0, i1 true)
|
||||
%res = select i1 %cmp, i32 %cttz, i32 32
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
;
|
||||
; Intel intrinsics
|
||||
;
|
||||
|
||||
define i16 @test_tzcnt_u16(i16 %a0) {
|
||||
; X32-LABEL: test_tzcnt_u16:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movzwl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: movzwl %ax, %ecx
|
||||
; X32-NEXT: cmpl $0, %ecx
|
||||
; X32-NEXT: jne .LBB7_1
|
||||
; X32-NEXT: # BB#2:
|
||||
; X32-NEXT: movw $16, %ax
|
||||
; X32-NEXT: retl
|
||||
; X32-NEXT: .LBB7_1:
|
||||
; X32-NEXT: tzcntw %ax, %ax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_tzcnt_u16:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movw $16, %cx
|
||||
; X64-NEXT: movzwl %di, %edx
|
||||
; X64-NEXT: tzcntw %dx, %ax
|
||||
; X64-NEXT: cmpl $0, %edx
|
||||
; X64-NEXT: cmovew %cx, %ax
|
||||
; X64-NEXT: retq
|
||||
%zext = zext i16 %a0 to i32
|
||||
%cmp = icmp ne i32 %zext, 0
|
||||
%cttz = call i16 @llvm.cttz.i16(i16 %a0, i1 true)
|
||||
%res = select i1 %cmp, i16 %cttz, i16 16
|
||||
ret i16 %res
|
||||
}
|
||||
|
||||
define i32 @test_andn_u32(i32 %a0, i32 %a1) {
|
||||
; X32-LABEL: test_andn_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: xorl $-1, %eax
|
||||
; X32-NEXT: andl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_andn_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorl $-1, %edi
|
||||
; X64-NEXT: andl %esi, %edi
|
||||
; X64-NEXT: movl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%xor = xor i32 %a0, -1
|
||||
%res = and i32 %xor, %a1
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test_bextr_u32(i32 %a0, i32 %a1, i32 %a2) {
|
||||
; X32-LABEL: test_bextr_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: andl $255, %ecx
|
||||
; X32-NEXT: andl $255, %eax
|
||||
; X32-NEXT: shll $8, %eax
|
||||
; X32-NEXT: orl %ecx, %eax
|
||||
; X32-NEXT: bextrl %eax, {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_bextr_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: andl $255, %esi
|
||||
; X64-NEXT: andl $255, %edx
|
||||
; X64-NEXT: shll $8, %edx
|
||||
; X64-NEXT: orl %esi, %edx
|
||||
; X64-NEXT: bextrl %edx, %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%and1 = and i32 %a1, 255
|
||||
%and2 = and i32 %a2, 255
|
||||
%shl = shl i32 %and2, 8
|
||||
%or = or i32 %and1, %shl
|
||||
%res = call i32 @llvm.x86.bmi.bextr.32(i32 %a0, i32 %or)
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test_blsi_u32(i32 %a0) {
|
||||
; X32-LABEL: test_blsi_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: xorl %eax, %eax
|
||||
; X32-NEXT: subl %ecx, %eax
|
||||
; X32-NEXT: andl %ecx, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_blsi_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: xorl %eax, %eax
|
||||
; X64-NEXT: subl %edi, %eax
|
||||
; X64-NEXT: andl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%neg = sub i32 0, %a0
|
||||
%res = and i32 %a0, %neg
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test_blsmsk_u32(i32 %a0) {
|
||||
; X32-LABEL: test_blsmsk_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: movl %ecx, %eax
|
||||
; X32-NEXT: subl $1, %eax
|
||||
; X32-NEXT: xorl %ecx, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_blsmsk_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl %edi, %eax
|
||||
; X64-NEXT: subl $1, %eax
|
||||
; X64-NEXT: xorl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i32 %a0, 1
|
||||
%res = xor i32 %a0, %dec
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test_blsr_u32(i32 %a0) {
|
||||
; X32-LABEL: test_blsr_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
||||
; X32-NEXT: movl %ecx, %eax
|
||||
; X32-NEXT: subl $1, %eax
|
||||
; X32-NEXT: andl %ecx, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_blsr_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl %edi, %eax
|
||||
; X64-NEXT: subl $1, %eax
|
||||
; X64-NEXT: andl %edi, %eax
|
||||
; X64-NEXT: retq
|
||||
%dec = sub i32 %a0, 1
|
||||
%res = and i32 %a0, %dec
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
define i32 @test_tzcnt_u32(i32 %a0) {
|
||||
; X32-LABEL: test_tzcnt_u32:
|
||||
; X32: # BB#0:
|
||||
; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
|
||||
; X32-NEXT: cmpl $0, %eax
|
||||
; X32-NEXT: jne .LBB13_1
|
||||
; X32-NEXT: # BB#2:
|
||||
; X32-NEXT: movl $32, %eax
|
||||
; X32-NEXT: retl
|
||||
; X32-NEXT: .LBB13_1:
|
||||
; X32-NEXT: tzcntl %eax, %eax
|
||||
; X32-NEXT: retl
|
||||
;
|
||||
; X64-LABEL: test_tzcnt_u32:
|
||||
; X64: # BB#0:
|
||||
; X64-NEXT: movl $32, %ecx
|
||||
; X64-NEXT: tzcntl %edi, %eax
|
||||
; X64-NEXT: cmovbl %ecx, %eax
|
||||
; X64-NEXT: retq
|
||||
%cmp = icmp ne i32 %a0, 0
|
||||
%cttz = call i32 @llvm.cttz.i32(i32 %a0, i1 true)
|
||||
%res = select i1 %cmp, i32 %cttz, i32 32
|
||||
ret i32 %res
|
||||
}
|
||||
|
||||
declare i16 @llvm.cttz.i16(i16, i1)
|
||||
declare i32 @llvm.cttz.i32(i32, i1)
|
||||
declare i32 @llvm.x86.bmi.bextr.32(i32, i32)
|
Loading…
x
Reference in New Issue
Block a user