Home
last modified time | relevance | path

Searched refs:SIMD_INLINE (Results 1 – 16 of 16) sorted by relevance

/external/libaom/libaom/aom_dsp/simd/
Dv256_intrinsics.h27 SIMD_INLINE uint32_t v256_low_u32(v256 a) { return c_v256_low_u32(a); } in v256_low_u32()
28 SIMD_INLINE v64 v256_low_v64(v256 a) { return c_v256_low_v64(a); } in v256_low_v64()
29 SIMD_INLINE uint64_t v256_low_u64(v256 a) { return c_v256_low_u64(a); } in v256_low_u64()
30 SIMD_INLINE v128 v256_low_v128(v256 a) { return c_v256_low_v128(a); } in v256_low_v128()
31 SIMD_INLINE v128 v256_high_v128(v256 a) { return c_v256_high_v128(a); } in v256_high_v128()
32 SIMD_INLINE v256 v256_from_v128(v128 hi, v128 lo) { in v256_from_v128()
35 SIMD_INLINE v256 v256_from_64(uint64_t a, uint64_t b, uint64_t c, uint64_t d) { in v256_from_64()
38 SIMD_INLINE v256 v256_from_v64(v64 a, v64 b, v64 c, v64 d) { in v256_from_v64()
42 SIMD_INLINE v256 v256_load_unaligned(const void *p) { in v256_load_unaligned()
45 SIMD_INLINE v256 v256_load_aligned(const void *p) { in v256_load_aligned()
[all …]
Dv64_intrinsics.h24 SIMD_INLINE uint32_t v64_low_u32(v64 a) { return c_v64_low_u32(a); } in v64_low_u32()
25 SIMD_INLINE uint32_t v64_high_u32(v64 a) { return c_v64_high_u32(a); } in v64_high_u32()
26 SIMD_INLINE int32_t v64_low_s32(v64 a) { return c_v64_low_s32(a); } in v64_low_s32()
27 SIMD_INLINE int32_t v64_high_s32(v64 a) { return c_v64_high_s32(a); } in v64_high_s32()
28 SIMD_INLINE v64 v64_from_32(uint32_t x, uint32_t y) { in v64_from_32()
31 SIMD_INLINE v64 v64_from_64(uint64_t x) { return c_v64_from_64(x); } in v64_from_64()
32 SIMD_INLINE uint64_t v64_u64(v64 x) { return c_v64_u64(x); } in v64_u64()
33 SIMD_INLINE v64 v64_from_16(uint16_t a, uint16_t b, uint16_t c, uint16_t d) { in v64_from_16()
37 SIMD_INLINE uint32_t u32_load_unaligned(const void *p) { in u32_load_unaligned()
40 SIMD_INLINE uint32_t u32_load_aligned(const void *p) { in u32_load_aligned()
[all …]
Dv128_intrinsics.h26 SIMD_INLINE uint32_t v128_low_u32(v128 a) { return c_v128_low_u32(a); } in v128_low_u32()
27 SIMD_INLINE v64 v128_low_v64(v128 a) { return c_v128_low_v64(a); } in v128_low_v64()
28 SIMD_INLINE v64 v128_high_v64(v128 a) { return c_v128_high_v64(a); } in v128_high_v64()
29 SIMD_INLINE v128 v128_from_64(uint64_t hi, uint64_t lo) { in v128_from_64()
32 SIMD_INLINE v128 v128_from_v64(v64 hi, v64 lo) { in v128_from_v64()
35 SIMD_INLINE v128 v128_from_32(uint32_t a, uint32_t b, uint32_t c, uint32_t d) { in v128_from_32()
39 SIMD_INLINE v128 v128_load_unaligned(const void *p) { in v128_load_unaligned()
42 SIMD_INLINE v128 v128_load_aligned(const void *p) { in v128_load_aligned()
46 SIMD_INLINE void v128_store_unaligned(void *p, v128 a) { in v128_store_unaligned()
49 SIMD_INLINE void v128_store_aligned(void *p, v128 a) { in v128_store_aligned()
[all …]
Dv64_intrinsics_arm.h26 SIMD_INLINE uint32_t v64_low_u32(v64 a) { in v64_low_u32()
30 SIMD_INLINE uint32_t v64_high_u32(v64 a) { in v64_high_u32()
34 SIMD_INLINE int32_t v64_low_s32(v64 a) { in v64_low_s32()
38 SIMD_INLINE int32_t v64_high_s32(v64 a) { in v64_high_s32()
42 SIMD_INLINE v64 v64_from_16(uint16_t a, uint16_t b, uint16_t c, uint16_t d) { in v64_from_16()
47 SIMD_INLINE v64 v64_from_32(uint32_t x, uint32_t y) { in v64_from_32()
51 SIMD_INLINE v64 v64_from_64(uint64_t x) { return vcreate_s64(x); } in v64_from_64()
53 SIMD_INLINE uint64_t v64_u64(v64 x) { return (uint64_t)x; } in v64_u64()
55 SIMD_INLINE uint32_t u32_load_aligned(const void *p) { in u32_load_aligned()
59 SIMD_INLINE uint32_t u32_load_unaligned(const void *p) { in u32_load_unaligned()
[all …]
Dv64_intrinsics_x86.h25 SIMD_INLINE uint32_t v64_low_u32(v64 a) { in v64_low_u32()
29 SIMD_INLINE uint32_t v64_high_u32(v64 a) { in v64_high_u32()
33 SIMD_INLINE int32_t v64_low_s32(v64 a) { return (int32_t)_mm_cvtsi128_si32(a); } in v64_low_s32()
35 SIMD_INLINE int32_t v64_high_s32(v64 a) { in v64_high_s32()
39 SIMD_INLINE v64 v64_from_16(uint16_t a, uint16_t b, uint16_t c, uint16_t d) { in v64_from_16()
45 SIMD_INLINE v64 v64_from_32(uint32_t x, uint32_t y) { in v64_from_32()
49 SIMD_INLINE v64 v64_from_64(uint64_t x) { in v64_from_64()
57 SIMD_INLINE uint64_t v64_u64(v64 x) { in v64_u64()
61 SIMD_INLINE uint32_t u32_load_aligned(const void *p) { in u32_load_aligned()
65 SIMD_INLINE uint32_t u32_load_unaligned(const void *p) { in u32_load_unaligned()
[all …]
Dv128_intrinsics_x86.h20 SIMD_INLINE uint32_t v128_low_u32(v128 a) { in v128_low_u32()
24 SIMD_INLINE v64 v128_low_v64(v128 a) { in v128_low_v64()
28 SIMD_INLINE v64 v128_high_v64(v128 a) { return _mm_srli_si128(a, 8); } in v128_high_v64()
30 SIMD_INLINE v128 v128_from_v64(v64 a, v64 b) { in v128_from_v64()
34 SIMD_INLINE v128 v128_from_64(uint64_t a, uint64_t b) { in v128_from_64()
38 SIMD_INLINE v128 v128_from_32(uint32_t a, uint32_t b, uint32_t c, uint32_t d) { in v128_from_32()
42 SIMD_INLINE v128 v128_load_aligned(const void *p) { in v128_load_aligned()
46 SIMD_INLINE v128 v128_load_unaligned(const void *p) { in v128_load_unaligned()
54 SIMD_INLINE void v128_store_aligned(void *p, v128 a) { in v128_store_aligned()
58 SIMD_INLINE void v128_store_unaligned(void *p, v128 a) { in v128_store_unaligned()
[all …]
Dv256_intrinsics_x86.h34 SIMD_INLINE uint32_t v256_low_u32(v256 a) { in v256_low_u32()
38 SIMD_INLINE v64 v256_low_v64(v256 a) { in v256_low_v64()
42 SIMD_INLINE uint64_t v256_low_u64(v256 a) { return v64_u64(v256_low_v64(a)); } in v256_low_u64()
44 SIMD_INLINE v128 v256_low_v128(v256 a) { return _mm256_castsi256_si128(a); } in v256_low_v128()
46 SIMD_INLINE v128 v256_high_v128(v256 a) { in v256_high_v128()
50 SIMD_INLINE v256 v256_from_v128(v128 a, v128 b) { in v256_from_v128()
55 SIMD_INLINE v256 v256_from_v64(v64 a, v64 b, v64 c, v64 d) { in v256_from_v64()
59 SIMD_INLINE v256 v256_from_64(uint64_t a, uint64_t b, uint64_t c, uint64_t d) { in v256_from_64()
63 SIMD_INLINE v256 v256_load_aligned(const void *p) { in v256_load_aligned()
67 SIMD_INLINE v256 v256_load_unaligned(const void *p) { in v256_load_unaligned()
[all …]
Dv128_intrinsics_arm.h21 SIMD_INLINE uint32_t v128_low_u32(v128 a) { in v128_low_u32()
25 SIMD_INLINE v64 v128_low_v64(v128 a) { return vget_low_s64(a); } in v128_low_v64()
27 SIMD_INLINE v64 v128_high_v64(v128 a) { return vget_high_s64(a); } in v128_high_v64()
29 SIMD_INLINE v128 v128_from_v64(v64 a, v64 b) { return vcombine_s64(b, a); } in v128_from_v64()
31 SIMD_INLINE v128 v128_from_64(uint64_t a, uint64_t b) { in v128_from_64()
35 SIMD_INLINE v128 v128_from_32(uint32_t a, uint32_t b, uint32_t c, uint32_t d) { in v128_from_32()
39 SIMD_INLINE v128 v128_load_aligned(const void *p) { in v128_load_aligned()
43 SIMD_INLINE v128 v128_load_unaligned(const void *p) { in v128_load_unaligned()
47 SIMD_INLINE void v128_store_aligned(void *p, v128 r) { in v128_store_aligned()
51 SIMD_INLINE void v128_store_unaligned(void *p, v128 r) { in v128_store_unaligned()
[all …]
Dv128_intrinsics_c.h34 SIMD_INLINE uint32_t c_v128_low_u32(c_v128 a) { return a.u32[0]; } in c_v128_low_u32()
36 SIMD_INLINE c_v64 c_v128_low_v64(c_v128 a) { return a.v64[0]; } in c_v128_low_v64()
38 SIMD_INLINE c_v64 c_v128_high_v64(c_v128 a) { return a.v64[1]; } in c_v128_high_v64()
40 SIMD_INLINE c_v128 c_v128_from_64(uint64_t hi, uint64_t lo) { in c_v128_from_64()
47 SIMD_INLINE c_v128 c_v128_from_v64(c_v64 hi, c_v64 lo) { in c_v128_from_v64()
54 SIMD_INLINE c_v128 c_v128_from_32(uint32_t a, uint32_t b, uint32_t c, in c_v128_from_32()
64 SIMD_INLINE c_v128 c_v128_load_unaligned(const void *p) { in c_v128_load_unaligned()
73 SIMD_INLINE c_v128 c_v128_load_aligned(const void *p) { in c_v128_load_aligned()
81 SIMD_INLINE void c_v128_store_unaligned(void *p, c_v128 a) { in c_v128_store_unaligned()
88 SIMD_INLINE void c_v128_store_aligned(void *p, c_v128 a) { in c_v128_store_aligned()
[all …]
Dv256_intrinsics_c.h35 SIMD_INLINE uint32_t c_v256_low_u32(c_v256 a) { return a.u32[0]; } in c_v256_low_u32()
37 SIMD_INLINE c_v64 c_v256_low_v64(c_v256 a) { return a.v64[0]; } in c_v256_low_v64()
39 SIMD_INLINE uint64_t c_v256_low_u64(c_v256 a) { return a.u64[0]; } in c_v256_low_u64()
41 SIMD_INLINE c_v128 c_v256_low_v128(c_v256 a) { return a.v128[0]; } in c_v256_low_v128()
43 SIMD_INLINE c_v128 c_v256_high_v128(c_v256 a) { return a.v128[1]; } in c_v256_high_v128()
45 SIMD_INLINE c_v256 c_v256_from_v128(c_v128 hi, c_v128 lo) { in c_v256_from_v128()
52 SIMD_INLINE c_v256 c_v256_from_64(uint64_t a, uint64_t b, uint64_t c, in c_v256_from_64()
62 SIMD_INLINE c_v256 c_v256_from_v64(c_v64 a, c_v64 b, c_v64 c, c_v64 d) { in c_v256_from_v64()
71 SIMD_INLINE c_v256 c_v256_load_unaligned(const void *p) { in c_v256_load_unaligned()
80 SIMD_INLINE c_v256 c_v256_load_aligned(const void *p) { in c_v256_load_aligned()
[all …]
Dv256_intrinsics_v128.h31 SIMD_INLINE uint32_t v256_low_u32(v256 a) { return v128_low_u32(a.val[0]); } in v256_low_u32()
33 SIMD_INLINE v64 v256_low_v64(v256 a) { return v128_low_v64(a.val[0]); } in v256_low_v64()
35 SIMD_INLINE uint64_t v256_low_u64(v256 a) { return v64_u64(v256_low_v64(a)); } in v256_low_u64()
37 SIMD_INLINE v128 v256_low_v128(v256 a) { return a.val[0]; } in v256_low_v128()
39 SIMD_INLINE v128 v256_high_v128(v256 a) { return a.val[1]; } in v256_high_v128()
41 SIMD_INLINE v256 v256_from_v128(v128 hi, v128 lo) { in v256_from_v128()
48 SIMD_INLINE v256 v256_from_64(uint64_t a, uint64_t b, uint64_t c, uint64_t d) { in v256_from_64()
52 SIMD_INLINE v256 v256_from_v64(v64 a, v64 b, v64 c, v64 d) { in v256_from_v64()
56 SIMD_INLINE v256 v256_load_unaligned(const void *p) { in v256_load_unaligned()
61 SIMD_INLINE v256 v256_load_aligned(const void *p) { in v256_load_aligned()
[all …]
Dv64_intrinsics_c.h34 SIMD_INLINE uint32_t c_v64_low_u32(c_v64 a) { in c_v64_low_u32()
38 SIMD_INLINE uint32_t c_v64_high_u32(c_v64 a) { in c_v64_high_u32()
42 SIMD_INLINE int32_t c_v64_low_s32(c_v64 a) { in c_v64_low_s32()
46 SIMD_INLINE int32_t c_v64_high_s32(c_v64 a) { in c_v64_high_s32()
50 SIMD_INLINE c_v64 c_v64_from_32(uint32_t x, uint32_t y) { in c_v64_from_32()
57 SIMD_INLINE c_v64 c_v64_from_64(uint64_t x) { in c_v64_from_64()
63 SIMD_INLINE uint64_t c_v64_u64(c_v64 x) { return x.u64; } in c_v64_u64()
65 SIMD_INLINE c_v64 c_v64_from_16(uint16_t a, uint16_t b, uint16_t c, in c_v64_from_16()
82 SIMD_INLINE uint32_t c_u32_load_unaligned(const void *p) { in c_u32_load_unaligned()
91 SIMD_INLINE void c_u32_store_unaligned(void *p, uint32_t a) { in c_u32_store_unaligned()
[all …]
/external/oboe/samples/RhythmGame/third_party/glm/
DCMakeLists.txt21 file(GLOB_RECURSE SIMD_INLINE ./simd/*.inl)
35 source_group("SIMD Files" FILES ${SIMD_INLINE})
47 ${SIMD_SOURCE} ${SIMD_INLINE} ${SIMD_HEADER})
56 ${SIMD_SOURCE} ${SIMD_INLINE} ${SIMD_HEADER})
65 ${SIMD_SOURCE} ${SIMD_INLINE} ${SIMD_HEADER})
/external/libaom/libaom/aom_dsp/
Daom_simd_inline.h17 #ifndef SIMD_INLINE
18 #define SIMD_INLINE static AOM_FORCE_INLINE macro
/external/libaom/libaom/av1/common/
Dcdef_block_simd.h201 SIMD_INLINE v256 constrain16(v256 a, v256 b, unsigned int threshold, in constrain16()
212 SIMD_INLINE v128 constrain(v256 a, v256 b, unsigned int strength, in constrain()
/external/libaom/libaom/test/
Dsimd_cmp_impl.h19 #undef SIMD_INLINE
20 #define SIMD_INLINE static // Don't enforce inlining macro