mirror of
https://github.com/capstone-engine/llvm-capstone.git
synced 2025-03-04 16:41:43 +00:00

Pursuant to discussions at https://discourse.llvm.org/t/rfc-c-23-p1467r9-extended-floating-point-types-and-standard-names/70033/22, this commit enhances the handling of the __bf16 type in Clang. - Firstly, it upgrades __bf16 from a storage-only type to an arithmetic type. - Secondly, it changes the mangling of __bf16 to DF16b on all architectures except ARM. This change has been made in accordance with the finalization of the mangling for the std::bfloat16_t type, as discussed at https://github.com/itanium-cxx-abi/cxx-abi/pull/147. - Finally, this commit extends the existing excess precision support to the __bf16 type. This applies to hardware architectures that do not natively support bfloat16 arithmetic. Appropriate tests have been added to verify the effects of these changes and ensure no regressions in other areas of the compiler. Reviewed By: rjmccall, pengfei, zahiraam Differential Revision: https://reviews.llvm.org/D150913
58 lines
2.2 KiB
Plaintext
58 lines
2.2 KiB
Plaintext
// REQUIRES: amdgpu-registered-target
|
|
// REQUIRES: x86-registered-target
|
|
|
|
// RUN: %clang_cc1 "-aux-triple" "x86_64-unknown-linux-gnu" "-triple" "r600-unknown-unknown"\
|
|
// RUN: -fcuda-is-device "-aux-target-cpu" "x86-64" -fsyntax-only -verify=r600 %s
|
|
|
|
// AMDGCN has storage-only support for bf16. R600 does not support it should error out when
|
|
// it's the main target.
|
|
|
|
#include "Inputs/cuda.h"
|
|
|
|
// There should be no errors on using the type itself, or when loading/storing values for amdgcn.
|
|
// r600 should error on all uses of the type.
|
|
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(2))) __bf16 bf16_x2;
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(4))) __bf16 bf16_x4;
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(8))) __bf16 bf16_x8;
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(16))) __bf16 bf16_x16;
|
|
|
|
// r600-error@+1 2 {{__bf16 is not supported on this target}}
|
|
__device__ void test(bool b, __bf16 *out, __bf16 in) {
|
|
__bf16 bf16 = in; // r600-error {{__bf16 is not supported on this target}}
|
|
*out = bf16;
|
|
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(2))) __bf16 bf16_x2;
|
|
bf16_x2 vec2_a, vec2_b;
|
|
vec2_a = vec2_b;
|
|
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(4))) __bf16 bf16_x4;
|
|
bf16_x4 vec4_a, vec4_b;
|
|
vec4_a = vec4_b;
|
|
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(8))) __bf16 bf16_x8;
|
|
bf16_x8 vec8_a, vec8_b;
|
|
vec8_a = vec8_b;
|
|
|
|
// r600-error@+1 {{__bf16 is not supported on this target}}
|
|
typedef __attribute__((ext_vector_type(16))) __bf16 bf16_x16;
|
|
bf16_x16 vec16_a, vec16_b;
|
|
vec16_a = vec16_b;
|
|
}
|
|
|
|
// r600-error@+1 2 {{__bf16 is not supported on this target}}
|
|
__bf16 hostfn(__bf16 a) {
|
|
return a;
|
|
}
|
|
|
|
// r600-error@+2 {{__bf16 is not supported on this target}}
|
|
// r600-error@+1 {{vector size not an integral multiple of component size}}
|
|
typedef __bf16 foo __attribute__((__vector_size__(16), __aligned__(16)));
|