Home
last modified time | relevance | path

Searched refs:in3 (Results 1 – 25 of 102) sorted by relevance

12345

/third_party/openh264/codec/common/inc/
Dmsa_macros.h231 #define MSA_ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
234 MSA_ST_V2(RTYPE, in2, in3, (pdst) + 2 * (stride), stride); \
237 #define MSA_ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
239 MSA_ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
359 #define MSA_ST_D8(in0, in1, in2, in3, idx0, idx1, idx2, idx3, \ argument
363 MSA_ST_D4(in2, in3, idx4, idx5, idx6, idx7, pdst + 4 * stride, stride) \
379 #define MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument
382 MSA_VSHF_B(RTYPE, in2, in3, mask1, out1) \
385 #define MSA_VSHF_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
388 MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1); \
[all …]
/third_party/ffmpeg/libavutil/mips/
Dgeneric_macros_msa.h241 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument
246 SW(in3, (pdst) + 3 * stride); \
256 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument
261 SD(in3, (pdst) + 3 * stride); \
369 #define ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument
372 ST_V2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \
379 #define ST_V6(RTYPE, in0, in1, in2, in3, in4, in5, pdst, stride) \ argument
381 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
386 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument
388 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \
[all …]
/third_party/ffmpeg/libavcodec/mips/
Didctdsp_msa.c28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_pixels_clamped_msa() local
30 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa()
31 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa()
32 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_pixels_clamped_msa()
38 in3_d = __msa_copy_u_d((v2i64) in3, 0); in put_pixels_clamped_msa()
52 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_signed_pixels_clamped_msa() local
54 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa()
59 in3 += 128; in put_signed_pixels_clamped_msa()
65 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa()
66 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_signed_pixels_clamped_msa()
[all …]
Dh263dsp_msa.c32 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in h263_h_loop_filter_msa() local
38 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in h263_h_loop_filter_msa()
39 TRANSPOSE8x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in h263_h_loop_filter_msa()
40 in0, in3, in2, in1); in h263_h_loop_filter_msa()
44 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_h_loop_filter_msa()
82 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa()
83 in3 = (v16u8) __msa_adds_s_b((v16i8) in3, (v16i8) d0); in h263_h_loop_filter_msa()
84 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa()
86 ILVR_B2_SH(in3, in0, in1, in2, temp0, temp1); in h263_h_loop_filter_msa()
88 in3 = (v16u8) __msa_ilvl_h(temp1, temp0); in h263_h_loop_filter_msa()
[all …]
Dhevc_idct_msa.c94 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument
104 ILVR_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \
106 ILVL_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \
146 PCKEV_H2_SH(sum1_l, sum1_r, sum2_l, sum2_r, in3, in4); \
334 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_idct_8x8_msa() local
336 LD_SH8(coeffs, 8, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa()
337 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in hevc_idct_8x8_msa()
338 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in hevc_idct_8x8_msa()
339 in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa()
340 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 12); in hevc_idct_8x8_msa()
[all …]
Dvp9_idct_msa.c86 #define VP9_DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
94 DOTP_SH4_SW(in2, in3, in2, in3, in6, in6, in7, in7, \
116 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument
134 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \
137 in4, in3); \
163 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
229 #define VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3) \ argument
240 ADD4(res0_m, in0, res1_m, in1, res2_m, in2, res3_m, in3, \
247 #define VP9_IDCT4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
261 step1_m = __msa_ilvr_h(in3, in1); \
[all …]
Dvc1dsp_msa.c30 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x8_msa() local
44 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_msa()
48 UNPCK_SH_SW(in3, in_r3, in_l3); in ff_vc1_inv_trans_8x8_msa()
134 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_msa()
137 ST_SH8(in0, in1, in2, in3, in4, in5, in6, in7, block, 8); in ff_vc1_inv_trans_8x8_msa()
142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_4x8_msa() local
159 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_4x8_msa()
163 UNPCK_R_SH_SW(in3, in_r3); in ff_vc1_inv_trans_4x8_msa()
191 in0, in1, in2, in3); in ff_vc1_inv_trans_4x8_msa()
192 ST_D8(in0, in1, in2, in3, 0, 1, 0, 1, 0, 1, 0, 1, block, 8); in ff_vc1_inv_trans_4x8_msa()
[all …]
Dsimple_idct_msa.c28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in simple_idct_msa() local
41 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
46 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa()
47 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
48 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7; in simple_idct_msa()
79 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa()
111 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp, in simple_idct_msa()
121 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa()
122 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa()
152 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa()
[all …]
Dvp8_idct_msa.c29 #define VP8_IDCT_1D_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument
40 c_tmp2_m = in3 + (((in3) * const_cospi8sqrt2minus1_m) >> 16); \
43 d_tmp2_m = ((in3) * sinpi8_sqrt2_m) >> 16; \
51 v4i32 in0, in1, in2, in3, hz0, hz1, hz2, hz3, vt0, vt1, vt2, vt3; in ff_vp8_idct_add_msa() local
60 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_idct_add_msa()
61 VP8_IDCT_1D_W(in0, in1, in2, in3, hz0, hz1, hz2, hz3); in ff_vp8_idct_add_msa()
108 v4i32 in0, in1, in2, in3, a1, b1, c1, d1; in ff_vp8_luma_dc_wht_msa() local
114 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_luma_dc_wht_msa()
115 BUTTERFLY_4(in0, in1, in2, in3, a1, b1, c1, d1); in ff_vp8_luma_dc_wht_msa()
Dmpegvideoencdsp_msa.c27 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in sum_u8src_16width_msa() local
30 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in sum_u8src_16width_msa()
34 HADD_UB4_UB(in0, in1, in2, in3, in0, in1, in2, in3); in sum_u8src_16width_msa()
42 sum += HADD_UH_U32(in3); in sum_u8src_16width_msa()
Dhevc_mc_bi_msa.c38 #define HEVC_BI_RND_CLIP4(in0, in1, in2, in3, \ argument
43 HEVC_BI_RND_CLIP2(in2, in3, vec2, vec3, rnd_val, out2, out3); \
54 #define HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, vec0, vec1, vec2, \ argument
58 HEVC_BI_RND_CLIP2_MAX_SATU(in2, in3, vec2, vec3, rnd_val, out2, out3); \
73 v8i16 in0 = { 0 }, in1 = { 0 }, in2 = { 0 }, in3 = { 0 }; in hevc_bi_copy_4w_msa() local
116 INSERT_D2_SH(tpd2, tpd3, in3); in hevc_bi_copy_4w_msa()
120 HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, dst0, dst1, dst2, in hevc_bi_copy_4w_msa()
142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_6w_msa() local
154 LD_SH8(src1_ptr, src2_stride, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_bi_copy_6w_msa()
162 HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, dst0, dst1, dst2, dst3, in hevc_bi_copy_6w_msa()
[all …]
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/spirv_assembly/
DvktSpvAsmTrinaryMinMaxTests.cpp196 static inline void runOpFunc (F f, void* out, const void* in1, const void* in2, const void* in3) in runOpFunc() argument
198 …interpret_cast<const T*>(in1), *reinterpret_cast<const T*>(in2), *reinterpret_cast<const T*>(in3)); in runOpFunc()
202 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (min3<deInt8>, out, in1,… in minInt8() argument
203 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (max3<deInt8>, out, in1,… in maxInt8() argument
204 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (mid3<deInt8>, out, in1,… in midInt8() argument
205 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (min3<deUint8>, out, in1… in minUint8() argument
206 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (max3<deUint8>, out, in1… in maxUint8() argument
207 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (mid3<deUint8>, out, in1… in midUint8() argument
208 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (min3<deInt16>, out, in1… in minInt16() argument
209 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (max3<deInt16>, out, in1… in maxInt16() argument
[all …]
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/spirv_assembly/
DvktSpvAsmTrinaryMinMaxTests.cpp196 static inline void runOpFunc (F f, void* out, const void* in1, const void* in2, const void* in3) in runOpFunc() argument
198 …interpret_cast<const T*>(in1), *reinterpret_cast<const T*>(in2), *reinterpret_cast<const T*>(in3)); in runOpFunc()
202 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (min3<deInt8>, out, in1,… in minInt8() argument
203 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (max3<deInt8>, out, in1,… in maxInt8() argument
204 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (mid3<deInt8>, out, in1,… in midInt8() argument
205 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (min3<deUint8>, out, in1… in minUint8() argument
206 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (max3<deUint8>, out, in1… in maxUint8() argument
207 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (mid3<deUint8>, out, in1… in midUint8() argument
208 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (min3<deInt16>, out, in1… in minInt16() argument
209 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (max3<deInt16>, out, in1… in maxInt16() argument
[all …]
/third_party/skia/third_party/externals/libwebp/src/dsp/
Dmsa_macro.h168 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument
176 SW(in3, ptmp); \
202 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument
210 SD(in3, ptmp); \
303 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
305 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument
312 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \
338 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument
340 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \
[all …]
Dcommon_sse2.h70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument
78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b()
133 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE2() argument
165 __m128i* const in3) { in VP8L32bToPlanar_SSE2() argument
169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2()
170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2()
185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar_SSE2()
Dcommon_sse41.h40 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE41() argument
66 WEBP_SSE41_SHUFF(G, in2, in3) in VP8PlanarTo24b_SSE41()
91 *in3 = _mm_or_si128(RG3, B3); in VP8PlanarTo24b_SSE41()
104 __m128i* const in3) { in VP8L32bToPlanar_SSE41() argument
111 const __m128i A3 = _mm_shuffle_epi8(*in3, shuff0); in VP8L32bToPlanar_SSE41()
120 *in3 = _mm_unpacklo_epi64(B0, B2); in VP8L32bToPlanar_SSE41()
/third_party/ffmpeg/libavcodec/loongarch/
Dsimple_idct_lasx.c41 in0, in1, in2, in3); \
42 LASX_TRANSPOSE4x16(in0, in1, in2, in3, in0, in1, in2, in3); \
56 b2 = __lasx_xvpermi_d(in3, 0xD8); \
58 b3 = __lasx_xvpermi_d(in3, 0x8D); \
89 DUP2_ARG2(__lasx_xvilvl_h, in3, in2, w6, w4, temp0, temp1); \
100 DUP2_ARG2(__lasx_xvilvh_h, in3, in2, w7, w5, temp0, temp1); \
118 in0, in1, in2, in3); \
122 DUP4_ARG2(__lasx_xvpickev_h, in1, in0, in3, in2, a2, a3, a0, a1, \
123 in0, in1, in2, in3); \
124 DUP4_ARG2(__lasx_xvpermi_d, in0, 0xD8, in1, 0xD8, in2, 0xD8, in3, 0xD8, \
[all …]
Dhevc_idct_lsx.c101 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument
111 DUP4_ARG2(__lsx_vilvl_h, in4, in0, in6, in2, in5, in1, in3, in7, \
113 DUP4_ARG2(__lsx_vilvh_h, in4, in0, in6, in2, in5, in1, in3, in7, \
152 in3 = __lsx_vssrarni_h_w(sum1_l, sum1_r, shift); \
363 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in ff_hevc_idct_8x8_lsx() local
366 coeffs, 48, in0, in1, in2, in3); in ff_hevc_idct_8x8_lsx()
369 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in ff_hevc_idct_8x8_lsx()
370 LSX_TRANSPOSE8x8_H(in0, in1, in2, in3, in4, in5, in6, in7, in ff_hevc_idct_8x8_lsx()
371 in0, in1, in2, in3, in4, in5, in6, in7); in ff_hevc_idct_8x8_lsx()
372 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 12); in ff_hevc_idct_8x8_lsx()
[all …]
Dvc1dsp_lasx.c29 __m256i in0, in1, in2, in3; in ff_vc1_inv_trans_8x8_lasx() local
57 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx()
58 DUP4_ARG2(__lasx_xvpermi_d, in0, 0xD8, in1, 0xD8, in2, 0xD8, in3, 0xD8, in ff_vc1_inv_trans_8x8_lasx()
59 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx()
61 DUP2_ARG2(__lasx_xvilvl_h, in2, in0, in3, in1, temp0, temp1); in ff_vc1_inv_trans_8x8_lasx()
72 DUP2_ARG2(__lasx_xvilvh_h, in1, in0, in3, in2, temp0, temp1); in ff_vc1_inv_trans_8x8_lasx()
85 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx()
88 DUP4_ARG2(__lasx_xvsrai_w, in0, 3, in1, 3, in2, 3, in3, 3, in ff_vc1_inv_trans_8x8_lasx()
89 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx()
93 in3, in2, temp0, temp1, temp2, temp3); in ff_vc1_inv_trans_8x8_lasx()
[all …]
/third_party/openssl/crypto/des/asm/
Ddes_enc.m477 #define in3 %i3
123 ! parameter 7 1 for move in1 to in3
124 ! parameter 8 1 for move in3 to in4, 2 for move in4 to in3
125 ! parameter 9 1 for load ks3 and ks2 to in4 and in3
136 ifelse($7,1,{mov in1, in3},{nop})
140 ifelse($8,1,{mov in3, in4},{})
141 ifelse($8,2,{mov in4, in3},{})
155 ifelse($9,1,{LDPTR KS2, in3},{})
618 ! parameter 5 1: load ks1/ks2 to in3/in4, add 120 to in4
619 ! 2: mov in4 to in3
[all …]
/third_party/node/deps/openssl/openssl/crypto/des/asm/
Ddes_enc.m477 #define in3 %i3
123 ! parameter 7 1 for move in1 to in3
124 ! parameter 8 1 for move in3 to in4, 2 for move in4 to in3
125 ! parameter 9 1 for load ks3 and ks2 to in4 and in3
136 ifelse($7,1,{mov in1, in3},{nop})
140 ifelse($8,1,{mov in3, in4},{})
141 ifelse($8,2,{mov in4, in3},{})
155 ifelse($9,1,{LDPTR KS2, in3},{})
618 ! parameter 5 1: load ks1/ks2 to in3/in4, add 120 to in4
619 ! 2: mov in4 to in3
[all …]
/third_party/openssl/crypto/poly1305/asm/
Dpoly1305-ppcfp.pl69 my ($in0,$in1,$in2,$in3,$i1,$i2,$i3) = map("r$_",(7..12,6));
122 li $in3,12
126 $LWXLE $in3,$in3,$inp
133 andc $in3,$in3,$i2
138 stw $in3,`8*7+(4^$LITTLE_ENDIAN)`($ctx)
263 li $in3,1
267 stw $in3,`$LOCALS+8*4+(4^$LITTLE_ENDIAN)`($sp)
282 oris $in3,$padbit,`(1023+52+96)<<4`
285 stw $in3,`$LOCALS+8*3+(0^$LITTLE_ENDIAN)`($sp)
293 $LWXLE $in3,$i3,$inp
[all …]
/third_party/node/deps/openssl/openssl/crypto/poly1305/asm/
Dpoly1305-ppcfp.pl69 my ($in0,$in1,$in2,$in3,$i1,$i2,$i3) = map("r$_",(7..12,6));
122 li $in3,12
126 $LWXLE $in3,$in3,$inp
133 andc $in3,$in3,$i2
138 stw $in3,`8*7+(4^$LITTLE_ENDIAN)`($ctx)
263 li $in3,1
267 stw $in3,`$LOCALS+8*4+(4^$LITTLE_ENDIAN)`($sp)
282 oris $in3,$padbit,`(1023+52+96)<<4`
285 stw $in3,`$LOCALS+8*3+(0^$LITTLE_ENDIAN)`($sp)
293 $LWXLE $in3,$i3,$inp
[all …]
/third_party/openssl/crypto/aes/asm/
Daesp8-ppc.pl678 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
682 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment
784 lvx_u $in3,$x30,$inp
790 le?vperm $in3,$in3,$in3,$inpperm
798 vxor $out3,$in3,$rndkey0
905 vxor $in3,$in3,v31
921 vncipherlast $out4,$out4,$in3
923 lvx_u $in3,$x30,$inp
928 le?vperm $in3,$in3,$in3,$inpperm
952 vxor $out3,$in3,$rndkey0
[all …]
/third_party/node/deps/openssl/openssl/crypto/aes/asm/
Daesp8-ppc.pl678 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13));
682 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment
784 lvx_u $in3,$x30,$inp
790 le?vperm $in3,$in3,$in3,$inpperm
798 vxor $out3,$in3,$rndkey0
905 vxor $in3,$in3,v31
921 vncipherlast $out4,$out4,$in3
923 lvx_u $in3,$x30,$inp
928 le?vperm $in3,$in3,$in3,$inpperm
952 vxor $out3,$in3,$rndkey0
[all …]

12345