/external/libaom/libaom/aom_dsp/simd/ |
D | v256_intrinsics.h | 27 SIMD_INLINE uint32_t v256_low_u32(v256 a) { return c_v256_low_u32(a); } in v256_low_u32() 28 SIMD_INLINE v64 v256_low_v64(v256 a) { return c_v256_low_v64(a); } in v256_low_v64() 29 SIMD_INLINE uint64_t v256_low_u64(v256 a) { return c_v256_low_u64(a); } in v256_low_u64() 30 SIMD_INLINE v128 v256_low_v128(v256 a) { return c_v256_low_v128(a); } in v256_low_v128() 31 SIMD_INLINE v128 v256_high_v128(v256 a) { return c_v256_high_v128(a); } in v256_high_v128() 32 SIMD_INLINE v256 v256_from_v128(v128 hi, v128 lo) { in v256_from_v128() 35 SIMD_INLINE v256 v256_from_64(uint64_t a, uint64_t b, uint64_t c, uint64_t d) { in v256_from_64() 38 SIMD_INLINE v256 v256_from_v64(v64 a, v64 b, v64 c, v64 d) { in v256_from_v64() 42 SIMD_INLINE v256 v256_load_unaligned(const void *p) { in v256_load_unaligned() 45 SIMD_INLINE v256 v256_load_aligned(const void *p) { in v256_load_aligned() [all …]
|
D | v64_intrinsics.h | 24 SIMD_INLINE uint32_t v64_low_u32(v64 a) { return c_v64_low_u32(a); } in v64_low_u32() 25 SIMD_INLINE uint32_t v64_high_u32(v64 a) { return c_v64_high_u32(a); } in v64_high_u32() 26 SIMD_INLINE int32_t v64_low_s32(v64 a) { return c_v64_low_s32(a); } in v64_low_s32() 27 SIMD_INLINE int32_t v64_high_s32(v64 a) { return c_v64_high_s32(a); } in v64_high_s32() 28 SIMD_INLINE v64 v64_from_32(uint32_t x, uint32_t y) { in v64_from_32() 31 SIMD_INLINE v64 v64_from_64(uint64_t x) { return c_v64_from_64(x); } in v64_from_64() 32 SIMD_INLINE uint64_t v64_u64(v64 x) { return c_v64_u64(x); } in v64_u64() 33 SIMD_INLINE v64 v64_from_16(uint16_t a, uint16_t b, uint16_t c, uint16_t d) { in v64_from_16() 37 SIMD_INLINE uint32_t u32_load_unaligned(const void *p) { in u32_load_unaligned() 40 SIMD_INLINE uint32_t u32_load_aligned(const void *p) { in u32_load_aligned() [all …]
|
D | v128_intrinsics.h | 26 SIMD_INLINE uint32_t v128_low_u32(v128 a) { return c_v128_low_u32(a); } in v128_low_u32() 27 SIMD_INLINE v64 v128_low_v64(v128 a) { return c_v128_low_v64(a); } in v128_low_v64() 28 SIMD_INLINE v64 v128_high_v64(v128 a) { return c_v128_high_v64(a); } in v128_high_v64() 29 SIMD_INLINE v128 v128_from_64(uint64_t hi, uint64_t lo) { in v128_from_64() 32 SIMD_INLINE v128 v128_from_v64(v64 hi, v64 lo) { in v128_from_v64() 35 SIMD_INLINE v128 v128_from_32(uint32_t a, uint32_t b, uint32_t c, uint32_t d) { in v128_from_32() 39 SIMD_INLINE v128 v128_load_unaligned(const void *p) { in v128_load_unaligned() 42 SIMD_INLINE v128 v128_load_aligned(const void *p) { in v128_load_aligned() 46 SIMD_INLINE void v128_store_unaligned(void *p, v128 a) { in v128_store_unaligned() 49 SIMD_INLINE void v128_store_aligned(void *p, v128 a) { in v128_store_aligned() [all …]
|
D | v64_intrinsics_arm.h | 26 SIMD_INLINE uint32_t v64_low_u32(v64 a) { in v64_low_u32() 30 SIMD_INLINE uint32_t v64_high_u32(v64 a) { in v64_high_u32() 34 SIMD_INLINE int32_t v64_low_s32(v64 a) { in v64_low_s32() 38 SIMD_INLINE int32_t v64_high_s32(v64 a) { in v64_high_s32() 42 SIMD_INLINE v64 v64_from_16(uint16_t a, uint16_t b, uint16_t c, uint16_t d) { in v64_from_16() 47 SIMD_INLINE v64 v64_from_32(uint32_t x, uint32_t y) { in v64_from_32() 51 SIMD_INLINE v64 v64_from_64(uint64_t x) { return vcreate_s64(x); } in v64_from_64() 53 SIMD_INLINE uint64_t v64_u64(v64 x) { return (uint64_t)x; } in v64_u64() 55 SIMD_INLINE uint32_t u32_load_aligned(const void *p) { in u32_load_aligned() 59 SIMD_INLINE uint32_t u32_load_unaligned(const void *p) { in u32_load_unaligned() [all …]
|
D | v64_intrinsics_x86.h | 25 SIMD_INLINE uint32_t v64_low_u32(v64 a) { in v64_low_u32() 29 SIMD_INLINE uint32_t v64_high_u32(v64 a) { in v64_high_u32() 33 SIMD_INLINE int32_t v64_low_s32(v64 a) { return (int32_t)_mm_cvtsi128_si32(a); } in v64_low_s32() 35 SIMD_INLINE int32_t v64_high_s32(v64 a) { in v64_high_s32() 39 SIMD_INLINE v64 v64_from_16(uint16_t a, uint16_t b, uint16_t c, uint16_t d) { in v64_from_16() 45 SIMD_INLINE v64 v64_from_32(uint32_t x, uint32_t y) { in v64_from_32() 49 SIMD_INLINE v64 v64_from_64(uint64_t x) { in v64_from_64() 57 SIMD_INLINE uint64_t v64_u64(v64 x) { in v64_u64() 61 SIMD_INLINE uint32_t u32_load_aligned(const void *p) { in u32_load_aligned() 65 SIMD_INLINE uint32_t u32_load_unaligned(const void *p) { in u32_load_unaligned() [all …]
|
D | v128_intrinsics_x86.h | 20 SIMD_INLINE uint32_t v128_low_u32(v128 a) { in v128_low_u32() 24 SIMD_INLINE v64 v128_low_v64(v128 a) { in v128_low_v64() 28 SIMD_INLINE v64 v128_high_v64(v128 a) { return _mm_srli_si128(a, 8); } in v128_high_v64() 30 SIMD_INLINE v128 v128_from_v64(v64 a, v64 b) { in v128_from_v64() 34 SIMD_INLINE v128 v128_from_64(uint64_t a, uint64_t b) { in v128_from_64() 38 SIMD_INLINE v128 v128_from_32(uint32_t a, uint32_t b, uint32_t c, uint32_t d) { in v128_from_32() 42 SIMD_INLINE v128 v128_load_aligned(const void *p) { in v128_load_aligned() 46 SIMD_INLINE v128 v128_load_unaligned(const void *p) { in v128_load_unaligned() 54 SIMD_INLINE void v128_store_aligned(void *p, v128 a) { in v128_store_aligned() 58 SIMD_INLINE void v128_store_unaligned(void *p, v128 a) { in v128_store_unaligned() [all …]
|
D | v256_intrinsics_x86.h | 34 SIMD_INLINE uint32_t v256_low_u32(v256 a) { in v256_low_u32() 38 SIMD_INLINE v64 v256_low_v64(v256 a) { in v256_low_v64() 42 SIMD_INLINE uint64_t v256_low_u64(v256 a) { return v64_u64(v256_low_v64(a)); } in v256_low_u64() 44 SIMD_INLINE v128 v256_low_v128(v256 a) { return _mm256_castsi256_si128(a); } in v256_low_v128() 46 SIMD_INLINE v128 v256_high_v128(v256 a) { in v256_high_v128() 50 SIMD_INLINE v256 v256_from_v128(v128 a, v128 b) { in v256_from_v128() 55 SIMD_INLINE v256 v256_from_v64(v64 a, v64 b, v64 c, v64 d) { in v256_from_v64() 59 SIMD_INLINE v256 v256_from_64(uint64_t a, uint64_t b, uint64_t c, uint64_t d) { in v256_from_64() 63 SIMD_INLINE v256 v256_load_aligned(const void *p) { in v256_load_aligned() 67 SIMD_INLINE v256 v256_load_unaligned(const void *p) { in v256_load_unaligned() [all …]
|
D | v128_intrinsics_arm.h | 21 SIMD_INLINE uint32_t v128_low_u32(v128 a) { in v128_low_u32() 25 SIMD_INLINE v64 v128_low_v64(v128 a) { return vget_low_s64(a); } in v128_low_v64() 27 SIMD_INLINE v64 v128_high_v64(v128 a) { return vget_high_s64(a); } in v128_high_v64() 29 SIMD_INLINE v128 v128_from_v64(v64 a, v64 b) { return vcombine_s64(b, a); } in v128_from_v64() 31 SIMD_INLINE v128 v128_from_64(uint64_t a, uint64_t b) { in v128_from_64() 35 SIMD_INLINE v128 v128_from_32(uint32_t a, uint32_t b, uint32_t c, uint32_t d) { in v128_from_32() 39 SIMD_INLINE v128 v128_load_aligned(const void *p) { in v128_load_aligned() 43 SIMD_INLINE v128 v128_load_unaligned(const void *p) { in v128_load_unaligned() 47 SIMD_INLINE void v128_store_aligned(void *p, v128 r) { in v128_store_aligned() 51 SIMD_INLINE void v128_store_unaligned(void *p, v128 r) { in v128_store_unaligned() [all …]
|
D | v128_intrinsics_c.h | 34 SIMD_INLINE uint32_t c_v128_low_u32(c_v128 a) { return a.u32[0]; } in c_v128_low_u32() 36 SIMD_INLINE c_v64 c_v128_low_v64(c_v128 a) { return a.v64[0]; } in c_v128_low_v64() 38 SIMD_INLINE c_v64 c_v128_high_v64(c_v128 a) { return a.v64[1]; } in c_v128_high_v64() 40 SIMD_INLINE c_v128 c_v128_from_64(uint64_t hi, uint64_t lo) { in c_v128_from_64() 47 SIMD_INLINE c_v128 c_v128_from_v64(c_v64 hi, c_v64 lo) { in c_v128_from_v64() 54 SIMD_INLINE c_v128 c_v128_from_32(uint32_t a, uint32_t b, uint32_t c, in c_v128_from_32() 64 SIMD_INLINE c_v128 c_v128_load_unaligned(const void *p) { in c_v128_load_unaligned() 73 SIMD_INLINE c_v128 c_v128_load_aligned(const void *p) { in c_v128_load_aligned() 81 SIMD_INLINE void c_v128_store_unaligned(void *p, c_v128 a) { in c_v128_store_unaligned() 88 SIMD_INLINE void c_v128_store_aligned(void *p, c_v128 a) { in c_v128_store_aligned() [all …]
|
D | v256_intrinsics_c.h | 35 SIMD_INLINE uint32_t c_v256_low_u32(c_v256 a) { return a.u32[0]; } in c_v256_low_u32() 37 SIMD_INLINE c_v64 c_v256_low_v64(c_v256 a) { return a.v64[0]; } in c_v256_low_v64() 39 SIMD_INLINE uint64_t c_v256_low_u64(c_v256 a) { return a.u64[0]; } in c_v256_low_u64() 41 SIMD_INLINE c_v128 c_v256_low_v128(c_v256 a) { return a.v128[0]; } in c_v256_low_v128() 43 SIMD_INLINE c_v128 c_v256_high_v128(c_v256 a) { return a.v128[1]; } in c_v256_high_v128() 45 SIMD_INLINE c_v256 c_v256_from_v128(c_v128 hi, c_v128 lo) { in c_v256_from_v128() 52 SIMD_INLINE c_v256 c_v256_from_64(uint64_t a, uint64_t b, uint64_t c, in c_v256_from_64() 62 SIMD_INLINE c_v256 c_v256_from_v64(c_v64 a, c_v64 b, c_v64 c, c_v64 d) { in c_v256_from_v64() 71 SIMD_INLINE c_v256 c_v256_load_unaligned(const void *p) { in c_v256_load_unaligned() 80 SIMD_INLINE c_v256 c_v256_load_aligned(const void *p) { in c_v256_load_aligned() [all …]
|
D | v256_intrinsics_v128.h | 31 SIMD_INLINE uint32_t v256_low_u32(v256 a) { return v128_low_u32(a.val[0]); } in v256_low_u32() 33 SIMD_INLINE v64 v256_low_v64(v256 a) { return v128_low_v64(a.val[0]); } in v256_low_v64() 35 SIMD_INLINE uint64_t v256_low_u64(v256 a) { return v64_u64(v256_low_v64(a)); } in v256_low_u64() 37 SIMD_INLINE v128 v256_low_v128(v256 a) { return a.val[0]; } in v256_low_v128() 39 SIMD_INLINE v128 v256_high_v128(v256 a) { return a.val[1]; } in v256_high_v128() 41 SIMD_INLINE v256 v256_from_v128(v128 hi, v128 lo) { in v256_from_v128() 48 SIMD_INLINE v256 v256_from_64(uint64_t a, uint64_t b, uint64_t c, uint64_t d) { in v256_from_64() 52 SIMD_INLINE v256 v256_from_v64(v64 a, v64 b, v64 c, v64 d) { in v256_from_v64() 56 SIMD_INLINE v256 v256_load_unaligned(const void *p) { in v256_load_unaligned() 61 SIMD_INLINE v256 v256_load_aligned(const void *p) { in v256_load_aligned() [all …]
|
D | v64_intrinsics_c.h | 34 SIMD_INLINE uint32_t c_v64_low_u32(c_v64 a) { in c_v64_low_u32() 38 SIMD_INLINE uint32_t c_v64_high_u32(c_v64 a) { in c_v64_high_u32() 42 SIMD_INLINE int32_t c_v64_low_s32(c_v64 a) { in c_v64_low_s32() 46 SIMD_INLINE int32_t c_v64_high_s32(c_v64 a) { in c_v64_high_s32() 50 SIMD_INLINE c_v64 c_v64_from_32(uint32_t x, uint32_t y) { in c_v64_from_32() 57 SIMD_INLINE c_v64 c_v64_from_64(uint64_t x) { in c_v64_from_64() 63 SIMD_INLINE uint64_t c_v64_u64(c_v64 x) { return x.u64; } in c_v64_u64() 65 SIMD_INLINE c_v64 c_v64_from_16(uint16_t a, uint16_t b, uint16_t c, in c_v64_from_16() 82 SIMD_INLINE uint32_t c_u32_load_unaligned(const void *p) { in c_u32_load_unaligned() 91 SIMD_INLINE void c_u32_store_unaligned(void *p, uint32_t a) { in c_u32_store_unaligned() [all …]
|
/external/oboe/samples/RhythmGame/third_party/glm/ |
D | CMakeLists.txt | 21 file(GLOB_RECURSE SIMD_INLINE ./simd/*.inl) 35 source_group("SIMD Files" FILES ${SIMD_INLINE}) 47 ${SIMD_SOURCE} ${SIMD_INLINE} ${SIMD_HEADER}) 56 ${SIMD_SOURCE} ${SIMD_INLINE} ${SIMD_HEADER}) 65 ${SIMD_SOURCE} ${SIMD_INLINE} ${SIMD_HEADER})
|
/external/libaom/libaom/aom_dsp/ |
D | aom_simd_inline.h | 17 #ifndef SIMD_INLINE 18 #define SIMD_INLINE static AOM_FORCE_INLINE macro
|
/external/libaom/libaom/av1/common/ |
D | cdef_block_simd.h | 201 SIMD_INLINE v256 constrain16(v256 a, v256 b, unsigned int threshold, in constrain16() 212 SIMD_INLINE v128 constrain(v256 a, v256 b, unsigned int strength, in constrain()
|
/external/libaom/libaom/test/ |
D | simd_cmp_impl.h | 19 #undef SIMD_INLINE 20 #define SIMD_INLINE static // Don't enforce inlining macro
|