| /external/v8/src/codegen/arm64/ |
| D | macro-assembler-arm64.h | 1560 void Ld1(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld1() 1564 void Ld1(const VRegister& vt, const VRegister& vt2, const VRegister& vt3, in Ld1() 1569 void Ld1(const VRegister& vt, const VRegister& vt2, const VRegister& vt3, in Ld1() 1582 void Ld2(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld2() 1586 void Ld2(const VRegister& vt, const VRegister& vt2, int lane, in Ld2() 1591 void Ld2r(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld2r() 1595 void Ld3(const VRegister& vt, const VRegister& vt2, const VRegister& vt3, in Ld3() 1600 void Ld3(const VRegister& vt, const VRegister& vt2, const VRegister& vt3, in Ld3() 1605 void Ld3r(const VRegister& vt, const VRegister& vt2, const VRegister& vt3, in Ld3r() 1610 void Ld4(const VRegister& vt, const VRegister& vt2, const VRegister& vt3, in Ld4() [all …]
|
| /external/XNNPACK/src/u8-clamp/ |
| D | scalar.c | 25 uint8_t vt2 = x[2]; in xnn_u8_clamp_ukernel__scalar() local
|
| /external/XNNPACK/src/x8-lut/ |
| D | scalar.c | 31 const uint8_t vt2 = t[vx2]; in xnn_x8_lut_ukernel__scalar() local
|
| /external/XNNPACK/src/x32-packx/ |
| D | x4-psimd.c | 49 const psimd_u32 vt2 = psimd_interleave_lo_u32(vx2, vx3); in xnn_x32_packx_ukernel_4x__psimd() local
|
| D | x4-sse.c | 51 const __m128 vt2 = _mm_unpacklo_ps(vx2, vx3); in xnn_x32_packx_ukernel_4x__sse() local
|
| /external/vixl/src/aarch64/ |
| D | macro-assembler-aarch64.h | 3064 void Ld1(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld1() 3070 const VRegister& vt2, in Ld1() 3078 const VRegister& vt2, in Ld1() 3096 void Ld2(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld2() 3102 const VRegister& vt2, in Ld2() 3109 void Ld2r(const VRegister& vt, const VRegister& vt2, const MemOperand& src) { in Ld2r() 3115 const VRegister& vt2, in Ld3() 3123 const VRegister& vt2, in Ld3() 3132 const VRegister& vt2, in Ld3r() 3140 const VRegister& vt2, in Ld4() [all …]
|
| /external/XNNPACK/src/f32-vscaleextexp/gen/ |
| D | avx512f-p5-scalef-x48.c | 58 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() local
|
| D | avx512f-p5-scalef-x64.c | 60 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x64() local
|
| D | avx2-p5-x24.c | 64 __m256 vt2 = _mm256_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x24() local
|
| D | avx2-p5-x32.c | 66 __m256 vt2 = _mm256_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x32() local
|
| D | avx512f-p5-scalef-x80.c | 62 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x80() local
|
| /external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
| D | avx512f-p5-scalef-x48.c | 62 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() local
|
| D | avx2-p5-x24.c | 77 __m256 vt2 = _mm256_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x24() local
|
| D | avx512f-p5-scalef-x64.c | 65 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x64() local
|
| D | avx512f-p5-scalef-x80.c | 68 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x80() local
|
| D | avx2-p5-x32.c | 82 __m256 vt2 = _mm256_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x32() local
|
| /external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
| D | scalar-p5-x4.c | 83 float vt2 = vn2 * vminus_ln2_hi + vx2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4() local
|
| D | scalar-p5-x4-acc2.c | 84 float vt2 = vn2 * vminus_ln2_hi + vx2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc2() local
|
| D | scalar-p5-x4-acc4.c | 86 float vt2 = vn2 * vminus_ln2_hi + vx2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc4() local
|
| D | scalar-lut64-p2-x4-acc2.c | 106 float vt2 = vn2 * vminus_ln2_o64_hi + vx2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_lut64_p2_x4_acc2() local
|
| D | scalar-lut64-p2-x4.c | 105 float vt2 = vn2 * vminus_ln2_o64_hi + vx2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_lut64_p2_x4() local
|
| D | scalar-lut64-p2-x4-acc4.c | 108 float vt2 = vn2 * vminus_ln2_o64_hi + vx2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_lut64_p2_x4_acc4() local
|
| /external/XNNPACK/src/f32-sigmoid/gen/ |
| D | scalar-p5-div-x4.c | 90 float vt2 = vn2 * vln2_hi + vz2; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x4() local
|
| D | scalar-lut2048-p1-div-x4.c | 111 float vt2 = vn2 * vln2_o2048_hi + vz2; in xnn_f32_sigmoid_ukernel__scalar_lut2048_p1_div_x4() local
|
| D | scalar-lut64-p2-div-x4.c | 111 float vt2 = vn2 * vln2_o64_hi + vz2; in xnn_f32_sigmoid_ukernel__scalar_lut64_p2_div_x4() local
|