/*===----------- avx10_2satcvtintrin.h - AVX10_2SATCVT intrinsics ----------=== * * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. * See https://llvm.org/LICENSE.txt for license information. * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception * *===-----------------------------------------------------------------------=== */ #ifndef __IMMINTRIN_H #error \ "Never use directly; include instead." #endif // __IMMINTRIN_H #ifndef __AVX10_2SATCVTINTRIN_H #define __AVX10_2SATCVTINTRIN_H #define _mm_ipcvtnebf16_epi8(A) \ ((__m128i)__builtin_ia32_vcvtnebf162ibs128((__v8bf)(__m128bh)(A))) #define _mm_mask_ipcvtnebf16_epi8(W, U, A) \ ((__m128i)__builtin_ia32_selectw_128( \ (__mmask8)(U), (__v8hi)_mm_ipcvtnebf16_epi8(A), (__v8hi)(__m128i)(W))) #define _mm_maskz_ipcvtnebf16_epi8(U, A) \ ((__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \ (__v8hi)_mm_ipcvtnebf16_epi8(A), \ (__v8hi)_mm_setzero_si128())) #define _mm256_ipcvtnebf16_epi8(A) \ ((__m256i)__builtin_ia32_vcvtnebf162ibs256((__v16bf)(__m256bh)(A))) #define _mm256_mask_ipcvtnebf16_epi8(W, U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvtnebf16_epi8(A), \ (__v16hi)(__m256i)(W))) #define _mm256_maskz_ipcvtnebf16_epi8(U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvtnebf16_epi8(A), \ (__v16hi)_mm256_setzero_si256())) #define _mm_ipcvtnebf16_epu8(A) \ ((__m128i)__builtin_ia32_vcvtnebf162iubs128((__v8bf)(__m128bh)(A))) #define _mm_mask_ipcvtnebf16_epu8(W, U, A) \ ((__m128i)__builtin_ia32_selectw_128( \ (__mmask8)(U), (__v8hi)_mm_ipcvtnebf16_epu8(A), (__v8hi)(__m128i)(W))) #define _mm_maskz_ipcvtnebf16_epu8(U, A) \ ((__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \ (__v8hi)_mm_ipcvtnebf16_epu8(A), \ (__v8hi)_mm_setzero_si128())) #define _mm256_ipcvtnebf16_epu8(A) \ ((__m256i)__builtin_ia32_vcvtnebf162iubs256((__v16bf)(__m256bh)(A))) #define _mm256_mask_ipcvtnebf16_epu8(W, U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvtnebf16_epu8(A), \ (__v16hi)(__m256i)(W))) #define _mm256_maskz_ipcvtnebf16_epu8(U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvtnebf16_epu8(A), \ (__v16hi)_mm256_setzero_si256())) #define _mm_ipcvtph_epi8(A) \ ((__m128i)__builtin_ia32_vcvtph2ibs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvtph_epi8(W, U, A) \ ((__m128i)__builtin_ia32_vcvtph2ibs128_mask((__v8hf)(__m128h)(A), \ (__v8hu)(W), (__mmask8)(U))) #define _mm_maskz_ipcvtph_epi8(U, A) \ ((__m128i)__builtin_ia32_vcvtph2ibs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvtph_epi8(A) \ ((__m256i)__builtin_ia32_vcvtph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvtph_epi8(W, U, A) \ ((__m256i)__builtin_ia32_vcvtph2ibs256_mask((__v16hf)(__m256h)(A), \ (__v16hu)(W), (__mmask16)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvtph_epi8(U, A) \ ((__m256i)__builtin_ia32_vcvtph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(_mm256_setzero_si256()), \ (__mmask16)(U), _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvt_roundph_epi8(A, R) \ ((__m256i)__builtin_ia32_vcvtph2ibs256_mask((__v16hf)(__m256h)(A), \ (__v16hu)_mm256_setzero_si256(), \ (__mmask16)-1, (const int)R)) #define _mm256_mask_ipcvt_roundph_epi8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvtph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(W), (__mmask16)(U), (const int)R)) #define _mm256_maskz_ipcvt_roundph_epi8(U, A, R) \ ((__m256i)__builtin_ia32_vcvtph2ibs256_mask((__v16hf)(__m256h)(A), \ (__v16hu)_mm256_setzero_si256(), \ (__mmask16)(U), (const int)R)) #define _mm_ipcvtph_epu8(A) \ ((__m128i)__builtin_ia32_vcvtph2iubs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvtph_epu8(W, U, A) \ ((__m128i)__builtin_ia32_vcvtph2iubs128_mask((__v8hf)(__m128h)(A), \ (__v8hu)(W), (__mmask8)(U))) #define _mm_maskz_ipcvtph_epu8(U, A) \ ((__m128i)__builtin_ia32_vcvtph2iubs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvtph_epu8(A) \ ((__m256i)__builtin_ia32_vcvtph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvtph_epu8(W, U, A) \ ((__m256i)__builtin_ia32_vcvtph2iubs256_mask((__v16hf)(__m256h)(A), \ (__v16hu)(W), (__mmask16)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvtph_epu8(U, A) \ ((__m256i)__builtin_ia32_vcvtph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(_mm256_setzero_si256()), \ (__mmask16)(U), _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvt_roundph_epu8(A, R) \ ((__m256i)__builtin_ia32_vcvtph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ (const int)R)) #define _mm256_mask_ipcvt_roundph_epu8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvtph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(W), (__mmask16)(U), (const int)R)) #define _mm256_maskz_ipcvt_roundph_epu8(U, A, R) \ ((__m256i)__builtin_ia32_vcvtph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)(U), \ (const int)R)) #define _mm_ipcvtps_epi8(A) \ ((__m128i)__builtin_ia32_vcvtps2ibs128_mask( \ (__v4sf)(__m128)(A), (__v4su)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvtps_epi8(W, U, A) \ ((__m128i)__builtin_ia32_vcvtps2ibs128_mask((__v4sf)(__m128)(A), \ (__v4su)(W), (__mmask8)(U))) #define _mm_maskz_ipcvtps_epi8(U, A) \ ((__m128i)__builtin_ia32_vcvtps2ibs128_mask( \ (__v4sf)(__m128)(A), (__v4su)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvtps_epi8(A) \ ((__m256i)__builtin_ia32_vcvtps2ibs256_mask( \ (__v8sf)(__m256)(A), (__v8su)_mm256_setzero_si256(), (__mmask8)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvtps_epi8(W, U, A) \ ((__m256i)__builtin_ia32_vcvtps2ibs256_mask((__v8sf)(__m256)(A), \ (__v8su)(W), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvtps_epi8(U, A) \ ((__m256i)__builtin_ia32_vcvtps2ibs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(_mm256_setzero_si256()), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvt_roundps_epi8(A, R) \ ((__m256i)__builtin_ia32_vcvtps2ibs256_mask((__v8sf)(__m256)(A), \ (__v8su)_mm256_setzero_si256(), \ (__mmask8)-1, (const int)R)) #define _mm256_mask_ipcvt_roundps_epi8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvtps2ibs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(W), (__mmask8)(U), (const int)R)) #define _mm256_maskz_ipcvt_roundps_epi8(U, A, R) \ ((__m256i)__builtin_ia32_vcvtps2ibs256_mask((__v8sf)(__m256)(A), \ (__v8su)_mm256_setzero_si256(), \ (__mmask8)(U), (const int)R)) #define _mm_ipcvtps_epu8(A) \ ((__m128i)__builtin_ia32_vcvtps2iubs128_mask( \ (__v4sf)(__m128)(A), (__v4su)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvtps_epu8(W, U, A) \ ((__m128i)__builtin_ia32_vcvtps2iubs128_mask((__v4sf)(__m128)(A), \ (__v4su)(W), (__mmask8)(U))) #define _mm_maskz_ipcvtps_epu8(U, A) \ ((__m128i)__builtin_ia32_vcvtps2iubs128_mask( \ (__v4sf)(__m128)(A), (__v4su)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvtps_epu8(A) \ ((__m256i)__builtin_ia32_vcvtps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)_mm256_setzero_si256(), (__mmask8)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvtps_epu8(W, U, A) \ ((__m256i)__builtin_ia32_vcvtps2iubs256_mask((__v8sf)(__m256)(A), \ (__v8su)(W), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvtps_epu8(U, A) \ ((__m256i)__builtin_ia32_vcvtps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(_mm256_setzero_si256()), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvt_roundps_epu8(A, R) \ ((__m256i)__builtin_ia32_vcvtps2iubs256_mask((__v8sf)(__m256)(A), \ (__v8su)_mm256_setzero_si256(), \ (__mmask8)-1, (const int)R)) #define _mm256_mask_ipcvt_roundps_epu8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvtps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(W), (__mmask8)(U), (const int)R)) #define _mm256_maskz_ipcvt_roundps_epu8(U, A, R) \ ((__m256i)__builtin_ia32_vcvtps2iubs256_mask((__v8sf)(__m256)(A), \ (__v8su)_mm256_setzero_si256(), \ (__mmask8)(U), (const int)R)) #define _mm_ipcvttnebf16_epi8(A) \ ((__m128i)__builtin_ia32_vcvttnebf162ibs128((__v8bf)(__m128bh)(A))) #define _mm_mask_ipcvttnebf16_epi8(W, U, A) \ ((__m128i)__builtin_ia32_selectw_128( \ (__mmask8)(U), (__v8hi)_mm_ipcvttnebf16_epi8(A), (__v8hi)(__m128i)(W))) #define _mm_maskz_ipcvttnebf16_epi8(U, A) \ ((__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \ (__v8hi)_mm_ipcvttnebf16_epi8(A), \ (__v8hi)_mm_setzero_si128())) #define _mm256_ipcvttnebf16_epi8(A) \ ((__m256i)__builtin_ia32_vcvttnebf162ibs256((__v16bf)(__m256bh)(A))) #define _mm256_mask_ipcvttnebf16_epi8(W, U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvttnebf16_epi8(A), \ (__v16hi)(__m256i)(W))) #define _mm256_maskz_ipcvttnebf16_epi8(U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvttnebf16_epi8(A), \ (__v16hi)_mm256_setzero_si256())) #define _mm_ipcvttnebf16_epu8(A) \ ((__m128i)__builtin_ia32_vcvttnebf162iubs128((__v8bf)(__m128bh)(A))) #define _mm_mask_ipcvttnebf16_epu8(W, U, A) \ ((__m128i)__builtin_ia32_selectw_128( \ (__mmask8)(U), (__v8hi)_mm_ipcvttnebf16_epu8(A), (__v8hi)(__m128i)(W))) #define _mm_maskz_ipcvttnebf16_epu8(U, A) \ ((__m128i)__builtin_ia32_selectw_128((__mmask8)(U), \ (__v8hi)_mm_ipcvttnebf16_epu8(A), \ (__v8hi)_mm_setzero_si128())) #define _mm256_ipcvttnebf16_epu8(A) \ ((__m256i)__builtin_ia32_vcvttnebf162iubs256((__v16bf)(__m256bh)(A))) #define _mm256_mask_ipcvttnebf16_epu8(W, U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvttnebf16_epu8(A), \ (__v16hi)(__m256i)(W))) #define _mm256_maskz_ipcvttnebf16_epu8(U, A) \ ((__m256i)__builtin_ia32_selectw_256((__mmask16)(U), \ (__v16hi)_mm256_ipcvttnebf16_epu8(A), \ (__v16hi)_mm256_setzero_si256())) #define _mm_ipcvttph_epi8(A) \ ((__m128i)__builtin_ia32_vcvttph2ibs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvttph_epi8(W, U, A) \ ((__m128i)__builtin_ia32_vcvttph2ibs128_mask((__v8hf)(__m128h)(A), \ (__v8hu)(W), (__mmask8)(U))) #define _mm_maskz_ipcvttph_epi8(U, A) \ ((__m128i)__builtin_ia32_vcvttph2ibs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvttph_epi8(A) \ ((__m256i)__builtin_ia32_vcvttph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvttph_epi8(W, U, A) \ ((__m256i)__builtin_ia32_vcvttph2ibs256_mask((__v16hf)(__m256h)(A), \ (__v16hu)(W), (__mmask16)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvttph_epi8(U, A) \ ((__m256i)__builtin_ia32_vcvttph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(_mm256_setzero_si256()), \ (__mmask16)(U), _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvtt_roundph_epi8(A, R) \ ((__m256i)__builtin_ia32_vcvttph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ (const int)R)) #define _mm256_mask_ipcvtt_roundph_epi8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvttph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(W), (__mmask16)(U), (const int)R)) #define _mm256_maskz_ipcvtt_roundph_epi8(U, A, R) \ ((__m256i)__builtin_ia32_vcvttph2ibs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)(U), \ (const int)R)) #define _mm_ipcvttph_epu8(A) \ ((__m128i)__builtin_ia32_vcvttph2iubs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvttph_epu8(W, U, A) \ ((__m128i)__builtin_ia32_vcvttph2iubs128_mask((__v8hf)(__m128h)(A), \ (__v8hu)(W), (__mmask8)(U))) #define _mm_maskz_ipcvttph_epu8(U, A) \ ((__m128i)__builtin_ia32_vcvttph2iubs128_mask( \ (__v8hf)(__m128h)(A), (__v8hu)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvttph_epu8(A) \ ((__m256i)__builtin_ia32_vcvttph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvttph_epu8(W, U, A) \ ((__m256i)__builtin_ia32_vcvttph2iubs256_mask((__v16hf)(__m256h)(A), \ (__v16hu)(W), (__mmask16)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvttph_epu8(U, A) \ ((__m256i)__builtin_ia32_vcvttph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(_mm256_setzero_si256()), \ (__mmask16)(U), _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvtt_roundph_epu8(A, R) \ ((__m256i)__builtin_ia32_vcvttph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)-1, \ (const int)R)) #define _mm256_mask_ipcvtt_roundph_epu8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvttph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)(W), (__mmask16)(U), (const int)R)) #define _mm256_maskz_ipcvtt_roundph_epu8(U, A, R) \ ((__m256i)__builtin_ia32_vcvttph2iubs256_mask( \ (__v16hf)(__m256h)(A), (__v16hu)_mm256_setzero_si256(), (__mmask16)(U), \ (const int)R)) #define _mm_ipcvttps_epi8(A) \ ((__m128i)__builtin_ia32_vcvttps2ibs128_mask( \ (__v4sf)(__m128)(A), (__v4su)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvttps_epi8(W, U, A) \ ((__m128i)__builtin_ia32_vcvttps2ibs128_mask((__v4sf)(__m128)(A), \ (__v4su)(W), (__mmask8)(U))) #define _mm_maskz_ipcvttps_epi8(U, A) \ ((__m128i)__builtin_ia32_vcvttps2ibs128_mask( \ (__v4sf)(__m128)(A), (__v4su)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvttps_epi8(A) \ ((__m256i)__builtin_ia32_vcvttps2ibs256_mask( \ (__v8sf)(__m256)(A), (__v8su)_mm256_setzero_si256(), (__mmask8)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvttps_epi8(W, U, A) \ ((__m256i)__builtin_ia32_vcvttps2ibs256_mask((__v8sf)(__m256)(A), \ (__v8su)(W), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvttps_epi8(U, A) \ ((__m256i)__builtin_ia32_vcvttps2ibs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(_mm256_setzero_si256()), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvtt_roundps_epi8(A, R) \ ((__m256i)__builtin_ia32_vcvttps2ibs256_mask((__v8sf)(__m256)(A), \ (__v8su)_mm256_setzero_si256(), \ (__mmask8)-1, (const int)R)) #define _mm256_mask_ipcvtt_roundps_epi8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvttps2ibs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(W), (__mmask8)(U), (const int)R)) #define _mm256_maskz_ipcvtt_roundps_epi8(U, A, R) \ ((__m256i)__builtin_ia32_vcvttps2ibs256_mask((__v8sf)(__m256)(A), \ (__v8su)_mm256_setzero_si256(), \ (__mmask8)(U), (const int)R)) #define _mm_ipcvttps_epu8(A) \ ((__m128i)__builtin_ia32_vcvttps2iubs128_mask( \ (__v4sf)(__m128)(A), (__v4su)_mm_setzero_si128(), (__mmask8)-1)) #define _mm_mask_ipcvttps_epu8(W, U, A) \ ((__m128i)__builtin_ia32_vcvttps2iubs128_mask((__v4sf)(__m128)(A), \ (__v4su)(W), (__mmask8)(U))) #define _mm_maskz_ipcvttps_epu8(U, A) \ ((__m128i)__builtin_ia32_vcvttps2iubs128_mask( \ (__v4sf)(__m128)(A), (__v4su)(_mm_setzero_si128()), (__mmask8)(U))) #define _mm256_ipcvttps_epu8(A) \ ((__m256i)__builtin_ia32_vcvttps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)_mm256_setzero_si256(), (__mmask8)-1, \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_mask_ipcvttps_epu8(W, U, A) \ ((__m256i)__builtin_ia32_vcvttps2iubs256_mask((__v8sf)(__m256)(A), \ (__v8su)(W), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_maskz_ipcvttps_epu8(U, A) \ ((__m256i)__builtin_ia32_vcvttps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(_mm256_setzero_si256()), (__mmask8)(U), \ _MM_FROUND_CUR_DIRECTION)) #define _mm256_ipcvtt_roundps_epu8(A, R) \ ((__m256i)__builtin_ia32_vcvttps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)_mm256_setzero_si256(), (__mmask8)-1, \ (const int)R)) #define _mm256_mask_ipcvtt_roundps_epu8(W, U, A, R) \ ((__m256i)__builtin_ia32_vcvttps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)(W), (__mmask8)(U), (const int)R)) #define _mm256_maskz_ipcvtt_roundps_epu8(U, A, R) \ ((__m256i)__builtin_ia32_vcvttps2iubs256_mask( \ (__v8sf)(__m256)(A), (__v8su)_mm256_setzero_si256(), (__mmask8)(U), \ (const int)R)) #endif // __AVX10_2SATCVTINTRIN_H