/external/llvm/test/CodeGen/X86/ |
D | soft-fp.ll | 18 %va12 = bitcast [1 x %struct.__va_list_tag]* %va to i8* ; <i8*> [#uses=2] 19 call void @llvm.va_start(i8* %va12) 22 call void @llvm.va_end(i8* %va12)
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 2x8c2-minmax-rndnu-neon-mull-ld1r.c | 77 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r() local 112 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r()
|
D | 3x8c2-minmax-rndnu-neon-mull-ld1r.c | 87 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local 136 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 184 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 219 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
D | 2x8c2-minmax-rndnu-neon-mlal-ld1r.c | 183 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r() local 218 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x16c2-minmax-rndnu-neon-mull-ld1r.c | 81 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r() local 140 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r()
|
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 183 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 218 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 4x8c2-minmax-rndnu-neon-mull-ld1r.c | 97 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r() local 160 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r()
|
D | 3x8c2-minmax-rndnu-neon-mlal-ld1r.c | 233 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r() local 282 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r()
|
D | 3x16c2-minmax-rndnu-neon-mull-ld1r.c | 93 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r() local 174 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r()
|
D | 2x16c2-minmax-rndnu-neon-mlal-ld1r.c | 251 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mlal_ld1r() local 310 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_igemm_minmax_rndnu_ukernel_2x16c2__neon_mlal_ld1r()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 2x8c2-minmax-rndnu-neon-mull-ld1r.c | 64 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r() local 99 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mull_ld1r()
|
D | 3x8c2-minmax-rndnu-neon-mull-ld1r.c | 72 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r() local 121 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mull_ld1r()
|
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 169 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 204 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-rndnu-neon-mlal-ld1r.c | 169 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r() local 204 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 170 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 205 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
D | 2x16c2-minmax-rndnu-neon-mull-ld1r.c | 68 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r() local 127 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_2x16c2__neon_mull_ld1r()
|
D | 4x8c2-minmax-rndnu-neon-mull-ld1r.c | 80 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r() local 143 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x8c2__neon_mull_ld1r()
|
D | 3x8c2-minmax-rndnu-neon-mlal-ld1r.c | 217 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r() local 266 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x8c2__neon_mlal_ld1r()
|
D | 3x16c2-minmax-rndnu-neon-mull-ld1r.c | 78 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r() local 159 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_3x16c2__neon_mull_ld1r()
|
D | 4x16c2-minmax-rndnu-neon-mull-ld1r.c | 88 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r() local 191 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qs8_gemm_minmax_rndnu_ukernel_4x16c2__neon_mull_ld1r()
|
/external/XNNPACK/src/qc8-gemm/gen/ |
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 169 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 204 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 170 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 205 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_gemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|
/external/XNNPACK/src/qc8-igemm/gen/ |
D | 2x8c2-minmax-fp32-neon-mlal-ld1r.c | 183 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r() local 218 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neon_mlal_ld1r()
|
D | 2x8c2-minmax-fp32-neonv8-mlal-ld1r.c | 184 const int16x4_t va12 = vld1_dup_s16((const void*)(a1 + 4)); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r() local 219 const int8x8_t va1c2 = vreinterpret_s8_s16(va12); in xnn_qc8_igemm_minmax_fp32_ukernel_2x8c2__neonv8_mlal_ld1r()
|