/third_party/openh264/codec/common/inc/ |
D | msa_macros.h | 231 #define MSA_ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 234 MSA_ST_V2(RTYPE, in2, in3, (pdst) + 2 * (stride), stride); \ 237 #define MSA_ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument 239 MSA_ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \ 359 #define MSA_ST_D8(in0, in1, in2, in3, idx0, idx1, idx2, idx3, \ argument 363 MSA_ST_D4(in2, in3, idx4, idx5, idx6, idx7, pdst + 4 * stride, stride) \ 379 #define MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ argument 382 MSA_VSHF_B(RTYPE, in2, in3, mask1, out1) \ 385 #define MSA_VSHF_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 388 MSA_VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1); \ [all …]
|
/third_party/ffmpeg/libavutil/mips/ |
D | generic_macros_msa.h | 241 #define SW4(in0, in1, in2, in3, pdst, stride) \ argument 246 SW(in3, (pdst) + 3 * stride); \ 256 #define SD4(in0, in1, in2, in3, pdst, stride) \ argument 261 SD(in3, (pdst) + 3 * stride); \ 369 #define ST_V4(RTYPE, in0, in1, in2, in3, pdst, stride) \ argument 372 ST_V2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \ 379 #define ST_V6(RTYPE, in0, in1, in2, in3, in4, in5, pdst, stride) \ argument 381 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \ 386 #define ST_V8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) \ argument 388 ST_V4(RTYPE, in0, in1, in2, in3, (pdst), stride); \ [all …]
|
/third_party/ffmpeg/libavcodec/mips/ |
D | idctdsp_msa.c | 28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_pixels_clamped_msa() local 30 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa() 31 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_pixels_clamped_msa() 32 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_pixels_clamped_msa() 38 in3_d = __msa_copy_u_d((v2i64) in3, 0); in put_pixels_clamped_msa() 52 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in put_signed_pixels_clamped_msa() local 54 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa() 59 in3 += 128; in put_signed_pixels_clamped_msa() 65 CLIP_SH8_0_255(in0, in1, in2, in3, in4, in5, in6, in7); in put_signed_pixels_clamped_msa() 66 PCKEV_B4_SH(in0, in0, in1, in1, in2, in2, in3, in3, in0, in1, in2, in3); in put_signed_pixels_clamped_msa() [all …]
|
D | h263dsp_msa.c | 32 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in h263_h_loop_filter_msa() local 38 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in h263_h_loop_filter_msa() 39 TRANSPOSE8x4_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, in h263_h_loop_filter_msa() 40 in0, in3, in2, in1); in h263_h_loop_filter_msa() 44 temp2 = (v8i16) __msa_ilvr_b((v16i8) in2, (v16i8) in3); in h263_h_loop_filter_msa() 82 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa() 83 in3 = (v16u8) __msa_adds_s_b((v16i8) in3, (v16i8) d0); in h263_h_loop_filter_msa() 84 in3 = __msa_xori_b(in3, 128); in h263_h_loop_filter_msa() 86 ILVR_B2_SH(in3, in0, in1, in2, temp0, temp1); in h263_h_loop_filter_msa() 88 in3 = (v16u8) __msa_ilvl_h(temp1, temp0); in h263_h_loop_filter_msa() [all …]
|
D | hevc_idct_msa.c | 94 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument 104 ILVR_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \ 106 ILVL_H4_SH(in4, in0, in6, in2, in5, in1, in3, in7, \ 146 PCKEV_H2_SH(sum1_l, sum1_r, sum2_l, sum2_r, in3, in4); \ 334 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_idct_8x8_msa() local 336 LD_SH8(coeffs, 8, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa() 337 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in hevc_idct_8x8_msa() 338 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in hevc_idct_8x8_msa() 339 in0, in1, in2, in3, in4, in5, in6, in7); in hevc_idct_8x8_msa() 340 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 12); in hevc_idct_8x8_msa() [all …]
|
D | vp9_idct_msa.c | 86 #define VP9_DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 94 DOTP_SH4_SW(in2, in3, in2, in3, in6, in6, in7, in7, \ 116 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 134 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \ 137 in4, in3); \ 163 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \ 229 #define VP9_ADDBLK_ST8x4_UB(dst, dst_stride, in0, in1, in2, in3) \ argument 240 ADD4(res0_m, in0, res1_m, in1, res2_m, in2, res3_m, in3, \ 247 #define VP9_IDCT4x4(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 261 step1_m = __msa_ilvr_h(in3, in1); \ [all …]
|
D | vc1dsp_msa.c | 30 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_8x8_msa() local 44 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_8x8_msa() 48 UNPCK_SH_SW(in3, in_r3, in_l3); in ff_vc1_inv_trans_8x8_msa() 134 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_msa() 137 ST_SH8(in0, in1, in2, in3, in4, in5, in6, in7, block, 8); in ff_vc1_inv_trans_8x8_msa() 142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in ff_vc1_inv_trans_4x8_msa() local 159 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in ff_vc1_inv_trans_4x8_msa() 163 UNPCK_R_SH_SW(in3, in_r3); in ff_vc1_inv_trans_4x8_msa() 191 in0, in1, in2, in3); in ff_vc1_inv_trans_4x8_msa() 192 ST_D8(in0, in1, in2, in3, 0, 1, 0, 1, 0, 1, 0, 1, block, 8); in ff_vc1_inv_trans_4x8_msa() [all …]
|
D | simple_idct_msa.c | 28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in simple_idct_msa() local 41 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa() 46 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa() 47 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa() 48 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7; in simple_idct_msa() 79 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa() 111 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp, in simple_idct_msa() 121 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in simple_idct_msa() 122 in0, in1, in2, in3, in4, in5, in6, in7); in simple_idct_msa() 152 ILVRL_H2_SW(in1, in3, b3_r, b3_l); in simple_idct_msa() [all …]
|
D | vp8_idct_msa.c | 29 #define VP8_IDCT_1D_W(in0, in1, in2, in3, out0, out1, out2, out3) \ argument 40 c_tmp2_m = in3 + (((in3) * const_cospi8sqrt2minus1_m) >> 16); \ 43 d_tmp2_m = ((in3) * sinpi8_sqrt2_m) >> 16; \ 51 v4i32 in0, in1, in2, in3, hz0, hz1, hz2, hz3, vt0, vt1, vt2, vt3; in ff_vp8_idct_add_msa() local 60 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_idct_add_msa() 61 VP8_IDCT_1D_W(in0, in1, in2, in3, hz0, hz1, hz2, hz3); in ff_vp8_idct_add_msa() 108 v4i32 in0, in1, in2, in3, a1, b1, c1, d1; in ff_vp8_luma_dc_wht_msa() local 114 UNPCK_SH_SW(input1, in2, in3); in ff_vp8_luma_dc_wht_msa() 115 BUTTERFLY_4(in0, in1, in2, in3, a1, b1, c1, d1); in ff_vp8_luma_dc_wht_msa()
|
D | mpegvideoencdsp_msa.c | 27 v16u8 in0, in1, in2, in3, in4, in5, in6, in7; in sum_u8src_16width_msa() local 30 LD_UB8(src, stride, in0, in1, in2, in3, in4, in5, in6, in7); in sum_u8src_16width_msa() 34 HADD_UB4_UB(in0, in1, in2, in3, in0, in1, in2, in3); in sum_u8src_16width_msa() 42 sum += HADD_UH_U32(in3); in sum_u8src_16width_msa()
|
D | hevc_mc_bi_msa.c | 38 #define HEVC_BI_RND_CLIP4(in0, in1, in2, in3, \ argument 43 HEVC_BI_RND_CLIP2(in2, in3, vec2, vec3, rnd_val, out2, out3); \ 54 #define HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, vec0, vec1, vec2, \ argument 58 HEVC_BI_RND_CLIP2_MAX_SATU(in2, in3, vec2, vec3, rnd_val, out2, out3); \ 73 v8i16 in0 = { 0 }, in1 = { 0 }, in2 = { 0 }, in3 = { 0 }; in hevc_bi_copy_4w_msa() local 116 INSERT_D2_SH(tpd2, tpd3, in3); in hevc_bi_copy_4w_msa() 120 HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, dst0, dst1, dst2, in hevc_bi_copy_4w_msa() 142 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in hevc_bi_copy_6w_msa() local 154 LD_SH8(src1_ptr, src2_stride, in0, in1, in2, in3, in4, in5, in6, in7); in hevc_bi_copy_6w_msa() 162 HEVC_BI_RND_CLIP4_MAX_SATU(in0, in1, in2, in3, dst0, dst1, dst2, dst3, in hevc_bi_copy_6w_msa() [all …]
|
/third_party/vk-gl-cts/external/vulkancts/modules/vulkan/spirv_assembly/ |
D | vktSpvAsmTrinaryMinMaxTests.cpp | 196 static inline void runOpFunc (F f, void* out, const void* in1, const void* in2, const void* in3) in runOpFunc() argument 198 …interpret_cast<const T*>(in1), *reinterpret_cast<const T*>(in2), *reinterpret_cast<const T*>(in3)); in runOpFunc() 202 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (min3<deInt8>, out, in1,… in minInt8() argument 203 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (max3<deInt8>, out, in1,… in maxInt8() argument 204 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (mid3<deInt8>, out, in1,… in midInt8() argument 205 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (min3<deUint8>, out, in1… in minUint8() argument 206 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (max3<deUint8>, out, in1… in maxUint8() argument 207 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (mid3<deUint8>, out, in1… in midUint8() argument 208 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (min3<deInt16>, out, in1… in minInt16() argument 209 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (max3<deInt16>, out, in1… in maxInt16() argument [all …]
|
/third_party/vk-gl-cts/external/vulkancts/modules_no_buildgn/vulkan/spirv_assembly/ |
D | vktSpvAsmTrinaryMinMaxTests.cpp | 196 static inline void runOpFunc (F f, void* out, const void* in1, const void* in2, const void* in3) in runOpFunc() argument 198 …interpret_cast<const T*>(in1), *reinterpret_cast<const T*>(in2), *reinterpret_cast<const T*>(in3)); in runOpFunc() 202 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (min3<deInt8>, out, in1,… in minInt8() argument 203 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (max3<deInt8>, out, in1,… in maxInt8() argument 204 …const void* in1, const void* in2, const void* in3) { runOpFunc<deInt8> (mid3<deInt8>, out, in1,… in midInt8() argument 205 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (min3<deUint8>, out, in1… in minUint8() argument 206 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (max3<deUint8>, out, in1… in maxUint8() argument 207 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deUint8> (mid3<deUint8>, out, in1… in midUint8() argument 208 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (min3<deInt16>, out, in1… in minInt16() argument 209 …onst void* in1, const void* in2, const void* in3) { runOpFunc<deInt16> (max3<deInt16>, out, in1… in maxInt16() argument [all …]
|
/third_party/skia/third_party/externals/libwebp/src/dsp/ |
D | msa_macro.h | 168 #define SW4(in0, in1, in2, in3, pdst, stride) do { \ argument 176 SW(in3, ptmp); \ 202 #define SD4(in0, in1, in2, in3, pdst, stride) do { \ argument 210 SD(in3, ptmp); \ 303 #define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 305 ST_B2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ 310 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 312 ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride); \ 338 #define ST_W4(RTYPE, in0, in1, in2, in3, pdst, stride) do { \ argument 340 ST_W2(RTYPE, in2, in3, pdst + 2 * stride, stride); \ [all …]
|
D | common_sse2.h | 70 const __m128i* const in2, const __m128i* const in3, __m128i* const out0, in VP8Transpose_2_4x4_16b() argument 78 const __m128i transpose0_1 = _mm_unpacklo_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 80 const __m128i transpose0_3 = _mm_unpackhi_epi16(*in2, *in3); in VP8Transpose_2_4x4_16b() 133 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE2() argument 165 __m128i* const in3) { in VP8L32bToPlanar_SSE2() argument 169 const __m128i A2 = _mm_unpacklo_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2() 170 const __m128i A3 = _mm_unpackhi_epi8(*in2, *in3); in VP8L32bToPlanar_SSE2() 185 *in3 = _mm_unpacklo_epi64(C0, C2); in VP8L32bToPlanar_SSE2()
|
D | common_sse41.h | 40 __m128i* const in3, __m128i* const in4, __m128i* const in5) { in VP8PlanarTo24b_SSE41() argument 66 WEBP_SSE41_SHUFF(G, in2, in3) in VP8PlanarTo24b_SSE41() 91 *in3 = _mm_or_si128(RG3, B3); in VP8PlanarTo24b_SSE41() 104 __m128i* const in3) { in VP8L32bToPlanar_SSE41() argument 111 const __m128i A3 = _mm_shuffle_epi8(*in3, shuff0); in VP8L32bToPlanar_SSE41() 120 *in3 = _mm_unpacklo_epi64(B0, B2); in VP8L32bToPlanar_SSE41()
|
/third_party/ffmpeg/libavcodec/loongarch/ |
D | simple_idct_lasx.c | 41 in0, in1, in2, in3); \ 42 LASX_TRANSPOSE4x16(in0, in1, in2, in3, in0, in1, in2, in3); \ 56 b2 = __lasx_xvpermi_d(in3, 0xD8); \ 58 b3 = __lasx_xvpermi_d(in3, 0x8D); \ 89 DUP2_ARG2(__lasx_xvilvl_h, in3, in2, w6, w4, temp0, temp1); \ 100 DUP2_ARG2(__lasx_xvilvh_h, in3, in2, w7, w5, temp0, temp1); \ 118 in0, in1, in2, in3); \ 122 DUP4_ARG2(__lasx_xvpickev_h, in1, in0, in3, in2, a2, a3, a0, a1, \ 123 in0, in1, in2, in3); \ 124 DUP4_ARG2(__lasx_xvpermi_d, in0, 0xD8, in1, 0xD8, in2, 0xD8, in3, 0xD8, \ [all …]
|
D | hevc_idct_lsx.c | 101 #define HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, shift) \ argument 111 DUP4_ARG2(__lsx_vilvl_h, in4, in0, in6, in2, in5, in1, in3, in7, \ 113 DUP4_ARG2(__lsx_vilvh_h, in4, in0, in6, in2, in5, in1, in3, in7, \ 152 in3 = __lsx_vssrarni_h_w(sum1_l, sum1_r, shift); \ 363 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in ff_hevc_idct_8x8_lsx() local 366 coeffs, 48, in0, in1, in2, in3); in ff_hevc_idct_8x8_lsx() 369 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 7); in ff_hevc_idct_8x8_lsx() 370 LSX_TRANSPOSE8x8_H(in0, in1, in2, in3, in4, in5, in6, in7, in ff_hevc_idct_8x8_lsx() 371 in0, in1, in2, in3, in4, in5, in6, in7); in ff_hevc_idct_8x8_lsx() 372 HEVC_IDCT8x8_COL(in0, in1, in2, in3, in4, in5, in6, in7, 12); in ff_hevc_idct_8x8_lsx() [all …]
|
D | vc1dsp_lasx.c | 29 __m256i in0, in1, in2, in3; in ff_vc1_inv_trans_8x8_lasx() local 57 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx() 58 DUP4_ARG2(__lasx_xvpermi_d, in0, 0xD8, in1, 0xD8, in2, 0xD8, in3, 0xD8, in ff_vc1_inv_trans_8x8_lasx() 59 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx() 61 DUP2_ARG2(__lasx_xvilvl_h, in2, in0, in3, in1, temp0, temp1); in ff_vc1_inv_trans_8x8_lasx() 72 DUP2_ARG2(__lasx_xvilvh_h, in1, in0, in3, in2, temp0, temp1); in ff_vc1_inv_trans_8x8_lasx() 85 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx() 88 DUP4_ARG2(__lasx_xvsrai_w, in0, 3, in1, 3, in2, 3, in3, 3, in ff_vc1_inv_trans_8x8_lasx() 89 in0, in1, in2, in3); in ff_vc1_inv_trans_8x8_lasx() 93 in3, in2, temp0, temp1, temp2, temp3); in ff_vc1_inv_trans_8x8_lasx() [all …]
|
/third_party/openssl/crypto/des/asm/ |
D | des_enc.m4 | 77 #define in3 %i3 123 ! parameter 7 1 for move in1 to in3 124 ! parameter 8 1 for move in3 to in4, 2 for move in4 to in3 125 ! parameter 9 1 for load ks3 and ks2 to in4 and in3 136 ifelse($7,1,{mov in1, in3},{nop}) 140 ifelse($8,1,{mov in3, in4},{}) 141 ifelse($8,2,{mov in4, in3},{}) 155 ifelse($9,1,{LDPTR KS2, in3},{}) 618 ! parameter 5 1: load ks1/ks2 to in3/in4, add 120 to in4 619 ! 2: mov in4 to in3 [all …]
|
/third_party/node/deps/openssl/openssl/crypto/des/asm/ |
D | des_enc.m4 | 77 #define in3 %i3 123 ! parameter 7 1 for move in1 to in3 124 ! parameter 8 1 for move in3 to in4, 2 for move in4 to in3 125 ! parameter 9 1 for load ks3 and ks2 to in4 and in3 136 ifelse($7,1,{mov in1, in3},{nop}) 140 ifelse($8,1,{mov in3, in4},{}) 141 ifelse($8,2,{mov in4, in3},{}) 155 ifelse($9,1,{LDPTR KS2, in3},{}) 618 ! parameter 5 1: load ks1/ks2 to in3/in4, add 120 to in4 619 ! 2: mov in4 to in3 [all …]
|
/third_party/openssl/crypto/poly1305/asm/ |
D | poly1305-ppcfp.pl | 69 my ($in0,$in1,$in2,$in3,$i1,$i2,$i3) = map("r$_",(7..12,6)); 122 li $in3,12 126 $LWXLE $in3,$in3,$inp 133 andc $in3,$in3,$i2 138 stw $in3,`8*7+(4^$LITTLE_ENDIAN)`($ctx) 263 li $in3,1 267 stw $in3,`$LOCALS+8*4+(4^$LITTLE_ENDIAN)`($sp) 282 oris $in3,$padbit,`(1023+52+96)<<4` 285 stw $in3,`$LOCALS+8*3+(0^$LITTLE_ENDIAN)`($sp) 293 $LWXLE $in3,$i3,$inp [all …]
|
/third_party/node/deps/openssl/openssl/crypto/poly1305/asm/ |
D | poly1305-ppcfp.pl | 69 my ($in0,$in1,$in2,$in3,$i1,$i2,$i3) = map("r$_",(7..12,6)); 122 li $in3,12 126 $LWXLE $in3,$in3,$inp 133 andc $in3,$in3,$i2 138 stw $in3,`8*7+(4^$LITTLE_ENDIAN)`($ctx) 263 li $in3,1 267 stw $in3,`$LOCALS+8*4+(4^$LITTLE_ENDIAN)`($sp) 282 oris $in3,$padbit,`(1023+52+96)<<4` 285 stw $in3,`$LOCALS+8*3+(0^$LITTLE_ENDIAN)`($sp) 293 $LWXLE $in3,$i3,$inp [all …]
|
/third_party/openssl/crypto/aes/asm/ |
D | aesp8-ppc.pl | 678 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13)); 682 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment 784 lvx_u $in3,$x30,$inp 790 le?vperm $in3,$in3,$in3,$inpperm 798 vxor $out3,$in3,$rndkey0 905 vxor $in3,$in3,v31 921 vncipherlast $out4,$out4,$in3 923 lvx_u $in3,$x30,$inp 928 le?vperm $in3,$in3,$in3,$inpperm 952 vxor $out3,$in3,$rndkey0 [all …]
|
/third_party/node/deps/openssl/openssl/crypto/aes/asm/ |
D | aesp8-ppc.pl | 678 my ($in0, $in1, $in2, $in3, $in4, $in5, $in6, $in7 )=map("v$_",(0..3,10..13)); 682 my ($tmp,$keyperm)=($in3,$in4); # aliases with "caller", redundant assignment 784 lvx_u $in3,$x30,$inp 790 le?vperm $in3,$in3,$in3,$inpperm 798 vxor $out3,$in3,$rndkey0 905 vxor $in3,$in3,v31 921 vncipherlast $out4,$out4,$in3 923 lvx_u $in3,$x30,$inp 928 le?vperm $in3,$in3,$in3,$inpperm 952 vxor $out3,$in3,$rndkey0 [all …]
|