Searched refs:vld1q_dup_s32 (Results 1 – 17 of 17) sorted by relevance
25 const int32x4_t va_multiplier = vld1q_dup_s32(¶ms->neon.a_multiplier); in xnn_q8_vadd_ukernel__neon()26 const int32x4_t vb_multiplier = vld1q_dup_s32(¶ms->neon.b_multiplier); in xnn_q8_vadd_ukernel__neon()27 const int32x4_t vright_shift = vld1q_dup_s32(¶ms->neon.right_shift); in xnn_q8_vadd_ukernel__neon()
33 const int32x4_t vbias = vld1q_dup_s32(¶ms->neon.bias); in xnn_q8_avgpool_ukernel_up9__neon()35 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_avgpool_ukernel_up9__neon()
33 const int32x4_t vbias = vld1q_dup_s32(¶ms->neon.bias); in xnn_q8_avgpool_ukernel_mp9p8q__neon()35 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
56 const int32x4_t vbias = vld1q_dup_s32(¶ms->neon.bias); in xnn_q8_gavgpool_ukernel_up7__neon()58 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_gavgpool_ukernel_up7__neon()
40 const int32x4_t vbias = vld1q_dup_s32(¶ms->neon.bias); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()104 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_gavgpool_ukernel_mp7p7q__neon()
40 statev = vld1q_dup_s32(filter_state_ch1); in WebRtcIsacfix_AllpassFilter2FixDec16Neon()
289 vst1q_s32( psDelDec->LF_AR_Q14, vld1q_dup_s32( &NSQ->sLF_AR_shp_Q14 ) ); in silk_NSQ_del_dec_neon()290 vst1q_s32( psDelDec->Diff_Q14, vld1q_dup_s32( &NSQ->sDiff_shp_Q14 ) ); in silk_NSQ_del_dec_neon()291 …vst1q_s32( psDelDec->Shape_Q14[ 0 ], vld1q_dup_s32( &NSQ->sLTP_shp_Q14[ psEncC->ltp_mem_length - 1… in silk_NSQ_del_dec_neon()293 vst1q_s32( psDelDec->sLPC_Q14[ i ], vld1q_dup_s32( &NSQ->sLPC_Q14[ i ] ) ); in silk_NSQ_del_dec_neon()296 vst1q_s32( psDelDec->sAR2_Q14[ i ], vld1q_dup_s32( &NSQ->sAR2_Q14[ i ] ) ); in silk_NSQ_del_dec_neon()
320 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_igemm_ukernel_4x8__neon()330 const int32x4_t vright_shift = vld1q_dup_s32(¶ms->neon.right_shift); in xnn_q8_igemm_ukernel_4x8__neon()
496 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_igemm_ukernel_8x8__neon()514 const int32x4_t vright_shift = vld1q_dup_s32(¶ms->neon.right_shift); in xnn_q8_igemm_ukernel_8x8__neon()
283 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_gemm_ukernel_4x8__neon()293 const int32x4_t vright_shift = vld1q_dup_s32(¶ms->neon.right_shift); in xnn_q8_gemm_ukernel_4x8__neon()
451 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_gemm_ukernel_8x8__neon()469 const int32x4_t vright_shift = vld1q_dup_s32(¶ms->neon.right_shift); in xnn_q8_gemm_ukernel_8x8__neon()
26 const int32x4_t vmultiplier = vld1q_dup_s32(¶ms->neon.multiplier); in xnn_q8_dwconv_ukernel_up8x9__neon()27 const int32x4_t vright_shift = vld1q_dup_s32(¶ms->neon.right_shift); in xnn_q8_dwconv_ukernel_up8x9__neon()
77 return vld1q_dup_s32(a); in test_vld1q_dup_s32()
4381 return vld1q_dup_s32(a); in test_vld1q_dup_s32()
61 #define vld1q_dup_s8x4(src) vld1q_dup_s32(reinterpret_cast<const int32*>(src))
79 #define vld1q_dup_s8x4(src) vld1q_dup_s32(reinterpret_cast<const int32*>(src))
1281 _NEON2SSESTORAGE int32x4_t vld1q_dup_s32(__transfersize(1) int32_t const * ptr); // VLD1.32 {d0[]},…9471 _NEON2SSESTORAGE int32x4_t vld1q_dup_s32(__transfersize(1) int32_t const * ptr); // VLD1.32 {d0[]},…9472 #define vld1q_dup_s32(ptr) _mm_set1_epi32 (*(ptr)) macro