/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | mbloopfilter_neon.c | 348 q2tmp0 = vtrnq_u32(vreinterpretq_u32_u8(q3), vreinterpretq_u32_u8(q7)); in vp8_mbloop_filter_vertical_edge_y_neon() 349 q2tmp1 = vtrnq_u32(vreinterpretq_u32_u8(q4), vreinterpretq_u32_u8(q8)); in vp8_mbloop_filter_vertical_edge_y_neon() 350 q2tmp2 = vtrnq_u32(vreinterpretq_u32_u8(q5), vreinterpretq_u32_u8(q9)); in vp8_mbloop_filter_vertical_edge_y_neon() 351 q2tmp3 = vtrnq_u32(vreinterpretq_u32_u8(q6), vreinterpretq_u32_u8(q10)); in vp8_mbloop_filter_vertical_edge_y_neon() 383 q2tmp0 = vtrnq_u32(vreinterpretq_u32_u8(q3), vreinterpretq_u32_u8(q7)); in vp8_mbloop_filter_vertical_edge_y_neon() 384 q2tmp1 = vtrnq_u32(vreinterpretq_u32_u8(q4), vreinterpretq_u32_u8(q8)); in vp8_mbloop_filter_vertical_edge_y_neon() 385 q2tmp2 = vtrnq_u32(vreinterpretq_u32_u8(q5), vreinterpretq_u32_u8(q9)); in vp8_mbloop_filter_vertical_edge_y_neon() 386 q2tmp3 = vtrnq_u32(vreinterpretq_u32_u8(q6), vreinterpretq_u32_u8(q10)); in vp8_mbloop_filter_vertical_edge_y_neon() 512 q2tmp0 = vtrnq_u32(vreinterpretq_u32_u8(q3), vreinterpretq_u32_u8(q7)); in vp8_mbloop_filter_vertical_edge_uv_neon() 513 q2tmp1 = vtrnq_u32(vreinterpretq_u32_u8(q4), vreinterpretq_u32_u8(q8)); in vp8_mbloop_filter_vertical_edge_uv_neon() [all …]
|
D | vp8_loopfilter_neon.c | 377 q2tmp0 = vtrnq_u32(vreinterpretq_u32_u8(q3), vreinterpretq_u32_u8(q7)); in vp8_loop_filter_vertical_edge_y_neon() 378 q2tmp1 = vtrnq_u32(vreinterpretq_u32_u8(q4), vreinterpretq_u32_u8(q8)); in vp8_loop_filter_vertical_edge_y_neon() 379 q2tmp2 = vtrnq_u32(vreinterpretq_u32_u8(q5), vreinterpretq_u32_u8(q9)); in vp8_loop_filter_vertical_edge_y_neon() 380 q2tmp3 = vtrnq_u32(vreinterpretq_u32_u8(q6), vreinterpretq_u32_u8(q10)); in vp8_loop_filter_vertical_edge_y_neon() 490 q2tmp0 = vtrnq_u32(vreinterpretq_u32_u8(q3), vreinterpretq_u32_u8(q7)); in vp8_loop_filter_vertical_edge_uv_neon() 491 q2tmp1 = vtrnq_u32(vreinterpretq_u32_u8(q4), vreinterpretq_u32_u8(q8)); in vp8_loop_filter_vertical_edge_uv_neon() 492 q2tmp2 = vtrnq_u32(vreinterpretq_u32_u8(q5), vreinterpretq_u32_u8(q9)); in vp8_loop_filter_vertical_edge_uv_neon() 493 q2tmp3 = vtrnq_u32(vreinterpretq_u32_u8(q6), vreinterpretq_u32_u8(q10)); in vp8_loop_filter_vertical_edge_uv_neon()
|
/external/webp/src/dsp/ |
D | lossless_neon.c | 150 #define GETQ_U8_AS_U32(IN) vgetq_lane_u32(vreinterpretq_u32_u8((IN)), 0); 151 #define STOREQ_U8_AS_U32P(OUT, IN) vst1q_u32((OUT), vreinterpretq_u32_u8((IN))); 270 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 294 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 357 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 388 vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE)); \ 575 const uint32x4_t a0g0 = vandq_u32(vreinterpretq_u32_u8(in), mask_ag); in TransformColorInverse_NEON()
|
/external/XNNPACK/src/q8-igemm/ |
D | 8x8-neon.c | 615 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 2… in xnn_q8_igemm_ukernel_8x8__neon() 616 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 0… in xnn_q8_igemm_ukernel_8x8__neon() 617 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 2… in xnn_q8_igemm_ukernel_8x8__neon() 618 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 0… in xnn_q8_igemm_ukernel_8x8__neon() 619 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_q8_igemm_ukernel_8x8__neon() 620 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_q8_igemm_ukernel_8x8__neon() 621 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_q8_igemm_ukernel_8x8__neon() 622 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_q8_igemm_ukernel_8x8__neon()
|
D | 4x8-neon.c | 386 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_q8_igemm_ukernel_4x8__neon() 387 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_q8_igemm_ukernel_4x8__neon() 388 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_q8_igemm_ukernel_4x8__neon() 389 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_q8_igemm_ukernel_4x8__neon()
|
/external/XNNPACK/src/q8-gemm/ |
D | 8x8-neon.c | 577 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_q8_gemm_ukernel_8x8__neon() 578 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_q8_gemm_ukernel_8x8__neon() 579 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_q8_gemm_ukernel_8x8__neon() 580 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_q8_gemm_ukernel_8x8__neon() 581 …vst1q_lane_u32(__builtin_assume_aligned(c4, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 0… in xnn_q8_gemm_ukernel_8x8__neon() 582 …vst1q_lane_u32(__builtin_assume_aligned(c5, 1), vreinterpretq_u32_u8(vout4x01234567_5x01234567), 2… in xnn_q8_gemm_ukernel_8x8__neon() 583 …vst1q_lane_u32(__builtin_assume_aligned(c6, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 0… in xnn_q8_gemm_ukernel_8x8__neon() 584 …vst1q_lane_u32(__builtin_assume_aligned(c7, 1), vreinterpretq_u32_u8(vout6x01234567_7x01234567), 2… in xnn_q8_gemm_ukernel_8x8__neon()
|
D | 4x8-neon.c | 353 …vst1q_lane_u32(__builtin_assume_aligned(c0, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 0… in xnn_q8_gemm_ukernel_4x8__neon() 354 …vst1q_lane_u32(__builtin_assume_aligned(c1, 1), vreinterpretq_u32_u8(vout0x01234567_1x01234567), 2… in xnn_q8_gemm_ukernel_4x8__neon() 355 …vst1q_lane_u32(__builtin_assume_aligned(c2, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 0… in xnn_q8_gemm_ukernel_4x8__neon() 356 …vst1q_lane_u32(__builtin_assume_aligned(c3, 1), vreinterpretq_u32_u8(vout2x01234567_3x01234567), 2… in xnn_q8_gemm_ukernel_4x8__neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | mem_neon.h | 146 const uint32x4_t a_u32 = vreinterpretq_u32_u8(a); in store_unaligned_u8q()
|
D | vpx_convolve8_neon.c | 387 d0123 = vreinterpretq_u32_u8( in vpx_convolve8_avg_horiz_neon() 472 d0415 = vreinterpretq_u32_u8( in vpx_convolve8_avg_horiz_neon() 474 d2637 = vreinterpretq_u32_u8( in vpx_convolve8_avg_horiz_neon() 843 d0123 = vreinterpretq_u32_u8( in vpx_convolve8_avg_vert_neon()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 21102 return vreinterpretq_u32_u8(a); in test_vreinterpretq_u32_u8()
|
D | arm_neon_intrinsics.c | 15061 return vreinterpretq_u32_u8(a); in test_vreinterpretq_u32_u8()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 16646 _NEON2SSESTORAGE uint32x4_t vreinterpretq_u32_u8 (uint8x16_t t); 16647 #define vreinterpretq_u32_u8 macro
|