Home
last modified time | relevance | path

Searched refs:va12 (Results 1 – 25 of 34) sorted by relevance

12

/external/llvm/test/CodeGen/X86/
Dsoft-fp.ll18 %va12 = bitcast [1 x %struct.__va_list_tag]* %va to i8* ; <i8*> [#uses=2]
19 call void @llvm.va_start(i8* %va12)
22 call void @llvm.va_end(i8* %va12)
/external/XNNPACK/src/qs8-igemm/gen/
D2x8c2-minmax-rndnu-neon-mull-ld1r.c77 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r() local
112 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r()
D3x8c2-minmax-rndnu-neon-mull-ld1r.c87 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local
136 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r()
D2x8c2-minmax-fp32-neonv8-mlal-ld1r.c184 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local
219 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
D2x8c2-minmax-rndnu-neon-mlal-ld1r.c183 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r() local
218 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r()
D2x16c2-minmax-rndnu-neon-mull-ld1r.c81 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r() local
140 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r()
D2x8c2-minmax-fp32-neon-mlal-ld1r.c183 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local
218 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
D4x8c2-minmax-rndnu-neon-mull-ld1r.c97 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r() local
160 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r()
D3x8c2-minmax-rndnu-neon-mlal-ld1r.c233 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r() local
282 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r()
D3x16c2-minmax-rndnu-neon-mull-ld1r.c93 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r() local
174 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r()
D2x16c2-minmax-rndnu-neon-mlal-ld1r.c251 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mlal_ld1r() local
310 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mlal_ld1r()
/external/XNNPACK/src/qs8-gemm/gen/
D2x8c2-minmax-rndnu-neon-mull-ld1r.c64 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r() local
99 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r()
D3x8c2-minmax-rndnu-neon-mull-ld1r.c72 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local
121 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r()
D2x8c2-minmax-fp32-neon-mlal-ld1r.c169 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local
204 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
D2x8c2-minmax-rndnu-neon-mlal-ld1r.c169 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r() local
204 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r()
D2x8c2-minmax-fp32-neonv8-mlal-ld1r.c170 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local
205 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
D2x16c2-minmax-rndnu-neon-mull-ld1r.c68 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r() local
127 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r()
D4x8c2-minmax-rndnu-neon-mull-ld1r.c80 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r() local
143 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r()
D3x8c2-minmax-rndnu-neon-mlal-ld1r.c217 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r() local
266 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r()
D3x16c2-minmax-rndnu-neon-mull-ld1r.c78 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r() local
159 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r()
D4x16c2-minmax-rndnu-neon-mull-ld1r.c88 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r() local
191 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r()
/external/XNNPACK/src/qc8-gemm/gen/
D2x8c2-minmax-fp32-neon-mlal-ld1r.c169 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local
204 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
D2x8c2-minmax-fp32-neonv8-mlal-ld1r.c170 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local
205 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
/external/XNNPACK/src/qc8-igemm/gen/
D2x8c2-minmax-fp32-neon-mlal-ld1r.c183 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local
218 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
D2x8c2-minmax-fp32-neonv8-mlal-ld1r.c184 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local
219 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()

12