/external/scrypt/lib/crypto/ |
D | crypto_scrypt-neon-salsa208.h | 75 diag3 ^= vsriq_n_u32(vshlq_n_u32(a0,7),a0,25); in salsa20_8_intrinsic() 77 diag2 ^= vsriq_n_u32(vshlq_n_u32(a1,9),a1,23); in salsa20_8_intrinsic() 79 diag1 ^= vsriq_n_u32(vshlq_n_u32(a2,13),a2,19); in salsa20_8_intrinsic() 81 diag0 ^= vsriq_n_u32(vshlq_n_u32(a3,18),a3,14); in salsa20_8_intrinsic() 88 diag1 ^= vsriq_n_u32(vshlq_n_u32(a0,7),a0,25); in salsa20_8_intrinsic() 90 diag2 ^= vsriq_n_u32(vshlq_n_u32(a1,9),a1,23); in salsa20_8_intrinsic() 92 diag3 ^= vsriq_n_u32(vshlq_n_u32(a2,13),a2,19); in salsa20_8_intrinsic() 94 diag0 ^= vsriq_n_u32(vshlq_n_u32(a3,18),a3,14); in salsa20_8_intrinsic()
|
/external/scrypt/patches/ |
D | arm_neon.patch | 81 + diag3 ^= vsriq_n_u32(vshlq_n_u32(a0,7),a0,25); 83 + diag2 ^= vsriq_n_u32(vshlq_n_u32(a1,9),a1,23); 85 + diag1 ^= vsriq_n_u32(vshlq_n_u32(a2,13),a2,19); 87 + diag0 ^= vsriq_n_u32(vshlq_n_u32(a3,18),a3,14); 94 + diag1 ^= vsriq_n_u32(vshlq_n_u32(a0,7),a0,25); 96 + diag2 ^= vsriq_n_u32(vshlq_n_u32(a1,9),a1,23); 98 + diag3 ^= vsriq_n_u32(vshlq_n_u32(a2,13),a2,19); 100 + diag0 ^= vsriq_n_u32(vshlq_n_u32(a3,18),a3,14);
|
/external/rust/crates/libz-sys/src/zlib-ng/arch/arm/ |
D | adler32_neon.c | 33 s2acc = vaddq_u32(s2acc, vshlq_n_u32(adacc, 5)); in NEON_accum32() 49 s2acc = vaddq_u32(s2acc, vshlq_n_u32(adacc, 4)); in NEON_accum32()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | depthwiseconv_uint8_transitional.h | 3336 filter_reg_0_a_shifted = vshlq_n_u32(filter_reg_0_a, 8); 3337 filter_reg_1_a_shifted = vshlq_n_u32(filter_reg_1_a, 8); 3338 filter_reg_2_a_shifted = vshlq_n_u32(filter_reg_2_a, 8); 3757 filter_reg_0_a_shifted = vshlq_n_u32(filter_reg_0_a, 8); 3758 filter_reg_1_a_shifted = vshlq_n_u32(filter_reg_1_a, 8); 3759 filter_reg_2_a_shifted = vshlq_n_u32(filter_reg_2_a, 8); 4435 filter_reg_0_a_shifted = vshlq_n_u32(filter_reg_0_a, 8); 4436 filter_reg_1_a_shifted = vshlq_n_u32(filter_reg_1_a, 8); 4437 filter_reg_2_a_shifted = vshlq_n_u32(filter_reg_2_a, 8); 4903 filter_reg_0_a_shifted = vshlq_n_u32(filter_reg_0_a, 8); [all …]
|
/external/zlib/ |
D | adler32_simd.c | 284 v_s2 = vshlq_n_u32(v_s2, 5); in adler32_simd_()
|
/external/angle/third_party/zlib/ |
D | adler32_simd.c | 284 v_s2 = vshlq_n_u32(v_s2, 5); in adler32_simd_()
|
/external/llvm-project/clang/test/CodeGen/arm-mve-intrinsics/ |
D | vector-shift-imm.c | 129 return vshlq_n_u32(a, 7); in test_vshlq_n_u32() 171 return vshlq_n_u32(a, 0); in test_vshlq_n_u32_trivial()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | film_grain_neon.cc | 584 const uint32x4_t line_increment8 = vshlq_n_u32(line_increment4, 1); in InitializeScalingLookupTable_NEON()
|
D | loop_restoration_neon.cc | 892 const uint32x4_t sum4 = vshlq_n_u32(sum, 2); in Sum565W() 1108 sum_b444->val[0] = vshlq_n_u32(sum_b111.val[0], 2); in Store343_444() 1109 sum_b444->val[1] = vshlq_n_u32(sum_b111.val[1], 2); in Store343_444()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v128_intrinsics_arm.h | 836 return c ? vreinterpretq_s64_u32(vshlq_n_u32(vreinterpretq_u32_s64(a), c)) in v128_shl_n_32()
|
/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 4708 return vshlq_n_u32(a, 3); in test_vshlq_n_u32()
|
D | arm_neon_intrinsics.c | 14694 return vshlq_n_u32(a, 1); in test_vshlq_n_u32()
|
/external/clang/test/CodeGen/ |
D | arm_neon_intrinsics.c | 17167 return vshlq_n_u32(a, 1); in test_vshlq_n_u32()
|
D | aarch64-neon-intrinsics.c | 5363 return vshlq_n_u32(a, 3); in test_vshlq_n_u32()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 1055 _NEON2SSESTORAGE uint32x4_t vshlq_n_u32(uint32x4_t a, __constrange(0,31) int b); // VSHL.I32 q0,q0,… 7757 _NEON2SSESTORAGE uint32x4_t vshlq_n_u32(uint32x4_t a, __constrange(0,31) int b); // VSHL.I32 q0,q0,… 7758 #define vshlq_n_u32 vshlq_n_s32 macro
|