// RUN: %clang_cc1 -x c++ -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s -msve-vector-bits=128 | FileCheck %s -D#VBITS=128 --check-prefixes=CHECK,CHECK128 // RUN: %clang_cc1 -x c++ -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s -msve-vector-bits=256 | FileCheck %s -D#VBITS=256 --check-prefixes=CHECK,CHECKWIDE // RUN: %clang_cc1 -x c++ -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s -msve-vector-bits=512 | FileCheck %s -D#VBITS=512 --check-prefixes=CHECK,CHECKWIDE // RUN: %clang_cc1 -x c++ -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s -msve-vector-bits=1024 | FileCheck %s -D#VBITS=1024 --check-prefixes=CHECK,CHECKWIDE // RUN: %clang_cc1 -x c++ -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s -msve-vector-bits=2048 | FileCheck %s -D#VBITS=2048 --check-prefixes=CHECK,CHECKWIDE // REQUIRES: aarch64-registered-target // Examples taken from section "3.7.3.3 Behavior specific to SVE // vectors" of the SVE ACLE (Version 00bet6) that can be found at // https://developer.arm.com/documentation/100987/latest // // Example has been expanded to work with mutiple values of // -msve-vector-bits. #include // Page 26, first paragraph of 3.7.3.3: sizeof and alignof #if __ARM_FEATURE_SVE_BITS #define N __ARM_FEATURE_SVE_BITS typedef svfloat32_t fixed_svfloat __attribute__((arm_sve_vector_bits(N))); void test01() { static_assert(alignof(fixed_svfloat) == 16, "Invalid align of Vector Length Specific Type."); static_assert(sizeof(fixed_svfloat) == N / 8, "Invalid size of Vector Length Specific Type."); } #endif // Page 26, items 1 and 2 of 3.7.3.3: how VLST and GNUT are related. #if __ARM_FEATURE_SVE_BITS && __ARM_FEATURE_SVE_VECTOR_OPERATORS #define N __ARM_FEATURE_SVE_BITS typedef svfloat64_t fixed_svfloat64 __attribute__((arm_sve_vector_bits(N))); typedef float64_t gnufloat64 __attribute__((vector_size(N / 8))); void test02() { static_assert(alignof(fixed_svfloat64) == alignof(gnufloat64), "Align of Vector Length Specific Type and GNU Vector Types " "should be the same."); static_assert(sizeof(fixed_svfloat64) == sizeof(gnufloat64), "Size of Vector Length Specific Type and GNU Vector Types " "should be the same."); } #endif // Page 27, item 1. #if __ARM_FEATURE_SVE_BITS && __ARM_FEATURE_SVE_VECTOR_OPERATORS #define N __ARM_FEATURE_SVE_BITS // CHECK-LABEL: define @_Z1f9__SVE_VLSIu11__SVInt32_tLj // CHECK-SAME: [[#VBITS]] // CHECK-SAME: EES_( %x.coerce, %y.coerce) // CHECK-NEXT: entry: // CHECK-NEXT: %x = alloca <[[#div(VBITS,32)]] x i32>, align 16 // CHECK-NEXT: %y = alloca <[[#div(VBITS,32)]] x i32>, align 16 // CHECK-NEXT: %retval.coerce = alloca , align 16 // CHECK-NEXT: %0 = bitcast <[[#div(VBITS,32)]] x i32>* %x to * // CHECK-NEXT: store %x.coerce, * %0, align 16 // CHECK-NEXT: %x1 = load <[[#div(VBITS,32)]] x i32>, <[[#div(VBITS,32)]] x i32>* %x, align 16 // CHECK-NEXT: %1 = bitcast <[[#div(VBITS,32)]] x i32>* %y to * // CHECK-NEXT: store %y.coerce, * %1, align 16 // CHECK-NEXT: %y2 = load <[[#div(VBITS,32)]] x i32>, <[[#div(VBITS,32)]] x i32>* %y, align 16 // CHECK-NEXT: %add = add <[[#div(VBITS,32)]] x i32> %y2, %x1 // CHECK-NEXT: %retval.0..sroa_cast = bitcast * %retval.coerce to <[[#div(VBITS,32)]] x i32>* // CHECK-NEXT: store <[[#div(VBITS,32)]] x i32> %add, <[[#div(VBITS,32)]] x i32>* %retval.0..sroa_cast, align 16 // CHECK-NEXT: %2 = load , * %retval.coerce, align 16 // CHECK-NEXT: ret %2 typedef svint32_t vec __attribute__((arm_sve_vector_bits(N))); auto f(vec x, vec y) { return x + y; } // Returns a vec. #endif // Page 27, item 3, adapted for a generic value of __ARM_FEATURE_SVE_BITS #if __ARM_FEATURE_SVE_BITS && __ARM_FEATURE_SVE_VECTOR_OPERATORS #define N __ARM_FEATURE_SVE_BITS typedef int16_t vec1 __attribute__((vector_size(N / 8))); void f(vec1); typedef svint16_t vec2 __attribute__((arm_sve_vector_bits(N))); // CHECK-LABEL: define void @_Z1g9__SVE_VLSIu11__SVInt16_tLj // CHECK-SAME: [[#VBITS]] // CHECK-SAME: EE( %x.coerce) // CHECK-NEXT: entry: // CHECK128-NEXT: %x = alloca <[[#div(VBITS,16)]] x i16>, align 16 // CHECK128-NEXT: %0 = bitcast <[[#div(VBITS,16)]] x i16>* %x to * // CHECK128-NEXT: store %x.coerce, * %0, align 16 // CHECK128-NEXT: %x1 = load <[[#div(VBITS,16)]] x i16>, <[[#div(VBITS,16)]] x i16>* %x, align 16 // CHECK128-NEXT: call void @_Z1fDv[[#div(VBITS,16)]]_s(<[[#div(VBITS,16)]] x i16> %x1) // CHECK128-NEXT: ret void // CHECKWIDE-NEXT: %x = alloca <[[#div(VBITS,16)]] x i16>, align 16 // CHECKWIDE-NEXT: %indirect-arg-temp = alloca <[[#div(VBITS,16)]] x i16>, align 16 // CHECKWIDE-NEXT: %0 = bitcast <[[#div(VBITS,16)]] x i16>* %x to * // CHECKWIDE-NEXT: store %x.coerce, * %0, align 16 // CHECKWIDE-NEXT: %x1 = load <[[#div(VBITS,16)]] x i16>, <[[#div(VBITS,16)]] x i16>* %x, align 16 // CHECKWIDE-NEXT: store <[[#div(VBITS,16)]] x i16> %x1, <[[#div(VBITS,16)]] x i16>* %indirect-arg-temp, align 16 // CHECKWIDE-NEXT: call void @_Z1fDv[[#div(VBITS,16)]]_s(<[[#div(VBITS,16)]] x i16>* nonnull %indirect-arg-temp) // CHECKWIDE-NEXT: ret void void g(vec2 x) { f(x); } // OK #endif