/external/libavc/common/armv8/ |
D | ih264_padding_neon_av8.s | 98 mov x4, x5 103 st1 {v0.8b, v1.8b}, [x4], x6 180 sub x4, x0, x3 192 st1 {v0.16b}, [x4], x1 // 16 bytes store 194 st1 {v2.16b}, [x4], x1 // 16 bytes store 199 st1 {v4.16b}, [x4], x1 // 16 bytes store 202 st1 {v6.16b}, [x4], x1 // 16 bytes store 208 st1 {v0.16b}, [x4], x1 // 16 bytes store 212 st1 {v2.16b}, [x4], x1 // 16 bytes store 216 st1 {v4.16b}, [x4], x1 // 16 bytes store [all …]
|
D | ih264_default_weighted_pred_av8.s | 117 sxtw x4, w4 128 ld1 {v2.s}[0], [x1], x4 //load row 1 in source 2 129 ld1 {v2.s}[1], [x1], x4 //load row 2 in source 2 133 ld1 {v3.s}[0], [x1], x4 //load row 3 in source 2 134 ld1 {v3.s}[1], [x1], x4 //load row 4 in source 2 147 ld1 {v4.8b}, [x1], x4 //load row 1 in source 2 149 ld1 {v5.8b}, [x1], x4 //load row 2 in source 2 153 ld1 {v6.8b}, [x1], x4 //load row 3 in source 2 156 ld1 {v7.8b}, [x1], x4 //load row 4 in source 2 169 ld1 {v16.8b, v17.8b}, [x1], x4 //load row 1 in source 2 [all …]
|
D | ih264_iquant_itrans_recon_dc_av8.s | 123 sxtw x4, w4 164 st1 {v1.s}[0], [x2], x4 165 st1 {v1.s}[1], [x2], x4 166 st1 {v2.s}[0], [x2], x4 229 sxtw x4, w4 248 ld1 {v11.d}[0], [x2], x4 //load pu1_out for interleaving 249 ld1 {v11.d}[1], [x2], x4 250 ld1 {v12.d}[0], [x2], x4 268 st1 {v11.d}[0], [x0], x4 269 st1 {v11.d}[1], [x0], x4 [all …]
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_fwd_txfm_sse2.c | 132 __m128i x4[8]; in fdct4x8_new_sse2() local 133 x4[0] = x3[0]; in fdct4x8_new_sse2() 134 x4[1] = x3[1]; in fdct4x8_new_sse2() 135 x4[2] = x3[2]; in fdct4x8_new_sse2() 136 x4[3] = x3[3]; in fdct4x8_new_sse2() 138 &x3[7], &x4[4], &x4[7]); in fdct4x8_new_sse2() 140 &x3[6], &x4[5], &x4[6]); in fdct4x8_new_sse2() 143 output[0] = x4[0]; in fdct4x8_new_sse2() 144 output[1] = x4[4]; in fdct4x8_new_sse2() 145 output[2] = x4[2]; in fdct4x8_new_sse2() [all …]
|
D | av1_fwd_txfm1d_sse4.c | 812 __m128i x4[64]; in av1_fdct64_new_sse4_1() local 813 x4[0] = _mm_add_epi32(x3[0], x3[7]); in av1_fdct64_new_sse4_1() 814 x4[7] = _mm_sub_epi32(x3[0], x3[7]); in av1_fdct64_new_sse4_1() 815 x4[1] = _mm_add_epi32(x3[1], x3[6]); in av1_fdct64_new_sse4_1() 816 x4[6] = _mm_sub_epi32(x3[1], x3[6]); in av1_fdct64_new_sse4_1() 817 x4[2] = _mm_add_epi32(x3[2], x3[5]); in av1_fdct64_new_sse4_1() 818 x4[5] = _mm_sub_epi32(x3[2], x3[5]); in av1_fdct64_new_sse4_1() 819 x4[3] = _mm_add_epi32(x3[3], x3[4]); in av1_fdct64_new_sse4_1() 820 x4[4] = _mm_sub_epi32(x3[3], x3[4]); in av1_fdct64_new_sse4_1() 821 x4[8] = x3[8]; in av1_fdct64_new_sse4_1() [all …]
|
/external/libhevc/common/arm64/ |
D | ihevc_padding.s | 99 sub x4,x0,x3 115 add x5,x4,x1 117 st1 {v0.16b},[x4],#16 //128/8 = 16 bytes store 118 st1 {v0.16b},[x4],#16 // 16 bytes store 119 st1 {v0.16b},[x4],#16 // 16 bytes store 120 st1 {v0.16b},[x4],#16 // 16 bytes store 121 st1 {v0.16b},[x4] // 16 bytes store 217 sub x4,x0,x3 233 add x5,x4,x1 235 st1 {v0.16b},[x4],#16 //128/8 = 16 bytes store [all …]
|
D | ihevc_inter_pred_filters_luma_horz.s | 123 mov x15,x4 // pi1_coeff 131 mov x4,x15 //loads pi1_coeff 135 ld1 {v0.8b},[x4] //coeff = vld1_s8(pi1_coeff) 147 add x4,x12,x2 //pu1_src_tmp2_8 = pu1_src + src_strd 198 add x4,x12,x2 //pu1_src + src_strd 236 ld1 {v12.2s},[x4],x11 //vector load pu1_src + src_strd 238 ld1 {v13.2s},[x4],x11 240 ld1 {v14.2s},[x4],x11 242 ld1 {v15.2s},[x4],x11 244 ld1 {v16.2s},[x4],x11 //vector load pu1_src + src_strd [all …]
|
D | ihevc_inter_pred_chroma_horz_w16out.s | 113 mov x15,x4 // pi1_coeff 117 mov x4,x15 //loads pi1_coeff 121 ld1 {v0.8b},[x4] //coeff = vld1_s8(pi1_coeff) 134 add x4,x12,x2 //pu1_src_tmp2_8 = pu1_src + src_strd 174 add x4,x12,x2 193 add x20,x4, x2 , lsl #1 205 add x19,x4,#8 207 ld1 { v29.2s},[x4],x11 //vector load pu1_src 212 ld1 { v10.2s},[x4],x11 //vector load pu1_src 217 ld1 { v12.2s},[x4],x11 //vector load pu1_src [all …]
|
D | ihevc_inter_pred_chroma_horz.s | 116 mov x15,x4 // pi1_coeff 121 mov x4,x15 //loads pi1_coeff 125 ld1 {v0.8b},[x4] //coeff = vld1_s8(pi1_coeff) 134 add x4,x12,x2 //pu1_src_tmp2_8 = pu1_src + src_strd 164 add x4,x12,x2 177 add x20,x4, x2 , lsl #1 190 add x19,x4,#8 192 ld1 { v29.2s},[x4],x11 //vector load pu1_src 197 ld1 { v10.2s},[x4],x11 //vector load pu1_src 202 ld1 { v12.2s},[x4],x11 //vector load pu1_src [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | selects.ll | 30 %base.x4 = shl i32 %base, 2 31 %base.x4.p1 = add i32 %base.x4, 1 32 %base.x4.p2 = add i32 %base.x4, 2 33 %base.x4.p3 = add i32 %base.x4, 3 34 %zext.x4 = zext i32 %base.x4 to i64 35 %zext.x4.p1 = zext i32 %base.x4.p1 to i64 36 %zext.x4.p2 = zext i32 %base.x4.p2 to i64 37 %zext.x4.p3 = zext i32 %base.x4.p3 to i64 38 %base.x16 = mul i64 %zext.x4, 4 39 %base.x16.p4 = shl i64 %zext.x4.p1, 2 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/AArch64/ |
D | arm64-memory.s | 8 ldr w5, [x4, #20] 9 ldr x4, [x3] 17 ldrb w5, [x4, #20] 41 prfm pstl3strm, [x4, x5, lsl #3] 43 ; CHECK: ldr w5, [x4, #20] ; encoding: [0x85,0x14,0x40,0xb9] 44 ; CHECK: ldr x4, [x3] ; encoding: [0x64,0x00,0x40,0xf9] 52 ; CHECK: ldrb w5, [x4, #20] ; encoding: [0x85,0x50,0x40,0x39] 75 ; CHECK: prfm pstl3strm, [x4, x5, lsl #3] ; encoding: [0x95,0x78,0xa5,0xf8] 81 str x4, [x3] 83 str w5, [x4, #20] [all …]
|
/external/llvm/test/MC/AArch64/ |
D | arm64-memory.s | 8 ldr w5, [x4, #20] 9 ldr x4, [x3] 17 ldrb w5, [x4, #20] 41 prfm pstl3strm, [x4, x5, lsl #3] 43 ; CHECK: ldr w5, [x4, #20] ; encoding: [0x85,0x14,0x40,0xb9] 44 ; CHECK: ldr x4, [x3] ; encoding: [0x64,0x00,0x40,0xf9] 52 ; CHECK: ldrb w5, [x4, #20] ; encoding: [0x85,0x50,0x40,0x39] 75 ; CHECK: prfm pstl3strm, [x4, x5, lsl #3] ; encoding: [0x95,0x78,0xa5,0xf8] 81 str x4, [x3] 83 str w5, [x4, #20] [all …]
|
/external/mesa3d/src/amd/addrlib/r800/ |
D | siaddrlib.cpp | 232 ADDR_CHANNEL_SETTING x4 = InitChannel(1, 0, log2BytesPP + bankXStart + 1); in ComputeBankEquation() local 241 x4.value = (threshX > bankXStart + 1) ? x4.value : 0; in ComputeBankEquation() 258 pEquation->xor2[1] = x4; in ComputeBankEquation() 270 pEquation->xor2[1] = x4; in ComputeBankEquation() 280 pEquation->addr[1] = x4; in ComputeBankEquation() 292 pEquation->addr[1] = x4; in ComputeBankEquation() 313 pEquation->xor2[1] = x4; in ComputeBankEquation() 323 pEquation->xor2[1] = x4; in ComputeBankEquation() 331 pEquation->addr[1] = x4; in ComputeBankEquation() 349 pEquation->xor1[1] = x4; in ComputeBankEquation() [all …]
|
/external/libmpeg2/common/armv8/ |
D | impeg2_inter_pred.s | 109 ldr x4, [x0] //src->y 113 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 117 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 119 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 121 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 123 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 125 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 127 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 129 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src 131 ld1 {v0.8b, v1.8b}, [x4], x2 //Load and increment src [all …]
|
/external/llvm/test/Bitcode/ |
D | binaryFloatInstructions.3.2.ll | 8 define void @fadd(float %x1, double %x2 ,half %x3, fp128 %x4, x86_fp80 %x5, ppc_fp128 %x6){ 19 ; CHECK-NEXT: %res4 = fadd fp128 %x4, %x4 20 %res4 = fadd fp128 %x4, %x4 31 define void @faddFloatVec(<2 x float> %x1, <3 x float> %x2 ,<4 x float> %x3, <8 x float> %x4, <16 x… 42 ; CHECK-NEXT: %res4 = fadd <8 x float> %x4, %x4 43 %res4 = fadd <8 x float> %x4, %x4 51 define void @faddDoubleVec(<2 x double> %x1, <3 x double> %x2 ,<4 x double> %x3, <8 x double> %x4, … 62 ; CHECK-NEXT: %res4 = fadd <8 x double> %x4, %x4 63 %res4 = fadd <8 x double> %x4, %x4 71 define void @faddHalfVec(<2 x half> %x1, <3 x half> %x2 ,<4 x half> %x3, <8 x half> %x4, <16 x half… [all …]
|
D | binaryIntInstructions.3.2.ll | 8 define void @add(i1 %x1, i8 %x2 ,i16 %x3, i32 %x4, i64 %x5){ 19 ; CHECK-NEXT: %res4 = add i32 %x4, %x4 20 %res4 = add i32 %x4, %x4 37 define void @addvec8NuwNsw(<2 x i8> %x1, <3 x i8> %x2 ,<4 x i8> %x3, <8 x i8> %x4, <16 x i8> %x5){ 48 ; CHECK-NEXT: %res4 = add nuw nsw <8 x i8> %x4, %x4 49 %res4 = add nuw nsw <8 x i8> %x4, %x4 57 define void @addvec16NuwNsw(<2 x i16> %x1, <3 x i16> %x2 ,<4 x i16> %x3, <8 x i16> %x4, <16 x i16> … 68 ; CHECK-NEXT: %res4 = add nuw nsw <8 x i16> %x4, %x4 69 %res4 = add nuw nsw <8 x i16> %x4, %x4 77 define void @addvec32NuwNsw(<2 x i32> %x1, <3 x i32> %x2 ,<4 x i32> %x3, <8 x i32> %x4, <16 x i32> … [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Bitcode/ |
D | binaryFloatInstructions.3.2.ll | 8 define void @fadd(float %x1, double %x2 ,half %x3, fp128 %x4, x86_fp80 %x5, ppc_fp128 %x6){ 19 ; CHECK-NEXT: %res4 = fadd fp128 %x4, %x4 20 %res4 = fadd fp128 %x4, %x4 31 define void @faddFloatVec(<2 x float> %x1, <3 x float> %x2 ,<4 x float> %x3, <8 x float> %x4, <16 x… 42 ; CHECK-NEXT: %res4 = fadd <8 x float> %x4, %x4 43 %res4 = fadd <8 x float> %x4, %x4 51 define void @faddDoubleVec(<2 x double> %x1, <3 x double> %x2 ,<4 x double> %x3, <8 x double> %x4, … 62 ; CHECK-NEXT: %res4 = fadd <8 x double> %x4, %x4 63 %res4 = fadd <8 x double> %x4, %x4 71 define void @faddHalfVec(<2 x half> %x1, <3 x half> %x2 ,<4 x half> %x3, <8 x half> %x4, <16 x half… [all …]
|
D | binaryIntInstructions.3.2.ll | 8 define void @add(i1 %x1, i8 %x2 ,i16 %x3, i32 %x4, i64 %x5){ 19 ; CHECK-NEXT: %res4 = add i32 %x4, %x4 20 %res4 = add i32 %x4, %x4 37 define void @addvec8NuwNsw(<2 x i8> %x1, <3 x i8> %x2 ,<4 x i8> %x3, <8 x i8> %x4, <16 x i8> %x5){ 48 ; CHECK-NEXT: %res4 = add nuw nsw <8 x i8> %x4, %x4 49 %res4 = add nuw nsw <8 x i8> %x4, %x4 57 define void @addvec16NuwNsw(<2 x i16> %x1, <3 x i16> %x2 ,<4 x i16> %x3, <8 x i16> %x4, <16 x i16> … 68 ; CHECK-NEXT: %res4 = add nuw nsw <8 x i16> %x4, %x4 69 %res4 = add nuw nsw <8 x i16> %x4, %x4 77 define void @addvec32NuwNsw(<2 x i32> %x1, <3 x i32> %x2 ,<4 x i32> %x3, <8 x i32> %x4, <16 x i32> … [all …]
|
/external/pdfium/third_party/agg23/ |
D | agg_curves.cpp | 32 float x4, float y4) in init() argument 37 bezier(x1, y1, x2, y2, x3, y3, x4, y4); in init() 43 float x4, float y4, in recursive_bezier() argument 53 float x34 = (x3 + x4) / 2; in recursive_bezier() 61 float dx = x4 - x1; in recursive_bezier() 63 float d2 = fabs(((x2 - x4) * dy) - ((y2 - y4) * dx)); in recursive_bezier() 64 float d3 = fabs(((x3 - x4) * dy) - ((y3 - y4) * dx)); in recursive_bezier() 69 fabs(x2 + x4 - x3 - x3) + fabs(y2 + y4 - y3 - y3) <= in recursive_bezier() 98 recursive_bezier(x1234, y1234, x234, y234, x34, y34, x4, y4, level + 1); in recursive_bezier() 103 float x4, float y4) in bezier() argument [all …]
|
D | agg_curves.h | 28 float x4, float y4) in curve4_points() 36 cp[6] = x4; in curve4_points() 42 float x4, float y4) in init() 50 cp[6] = x4; in init() 71 float x4, float y4) : in curve4_div() argument 74 init(x1, y1, x2, y2, x3, y3, x4, y4); in curve4_div() 89 float x4, float y4); 127 float x4, float y4); 131 float x4, float y4, 145 float x4, float y4) in curve4() argument [all …]
|
/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_highbd_iht8x8_add_sse4.c | 58 __m128i x0[2], x1[2], x2[2], x3[2], x4[2], x5[2], x6[2], x7[2]; in highbd_iadst8_sse4_1() local 69 x4[0] = _mm_sub_epi64(s0[0], s4[0]); in highbd_iadst8_sse4_1() 70 x4[1] = _mm_sub_epi64(s0[1], s4[1]); in highbd_iadst8_sse4_1() 93 x4[0] = dct_const_round_shift_64bit(x4[0]); in highbd_iadst8_sse4_1() 94 x4[1] = dct_const_round_shift_64bit(x4[1]); in highbd_iadst8_sse4_1() 105 x4[0] = pack_4(x4[0], x4[1]); in highbd_iadst8_sse4_1() 116 highbd_iadst_butterfly_sse4_1(x4[0], x5[0], cospi_8_64, cospi_24_64, s4, s5); in highbd_iadst8_sse4_1() 119 x4[0] = _mm_add_epi64(s4[0], s6[0]); in highbd_iadst8_sse4_1() 120 x4[1] = _mm_add_epi64(s4[1], s6[1]); in highbd_iadst8_sse4_1() 127 x4[0] = dct_const_round_shift_64bit(x4[0]); in highbd_iadst8_sse4_1() [all …]
|
/external/libopus/celt/x86/ |
D | vq_sse2.c | 70 __m128 x4, s4; in op_pvq_search_sse2() local 71 x4 = _mm_loadu_ps(&X[j]); in op_pvq_search_sse2() 72 s4 = _mm_cmplt_ps(x4, _mm_setzero_ps()); in op_pvq_search_sse2() 74 x4 = _mm_andnot_ps(signmask, x4); in op_pvq_search_sse2() 75 sums = _mm_add_ps(sums, x4); in op_pvq_search_sse2() 79 _mm_storeu_ps(&X[j], x4); in op_pvq_search_sse2() 113 __m128 rx4, x4, y4; in op_pvq_search_sse2() local 115 x4 = _mm_loadu_ps(&X[j]); in op_pvq_search_sse2() 116 rx4 = _mm_mul_ps(x4, rcp4); in op_pvq_search_sse2() 121 xy4 = _mm_add_ps(xy4, _mm_mul_ps(x4, y4)); in op_pvq_search_sse2() [all …]
|
/external/libhevc/decoder/arm64/ |
D | ihevcd_fmt_conv_420sp_to_420p.s | 91 mov x15, x4 109 MOV x4,x9 //// Copying height 136 SUBS x4, x4, #1 158 sxtw x4,w4 159 CMP x4,#0 ////Swap U and V dest if is_u_first_flag is zero 160 csel x4, x5, x4,EQ 162 csel x3, x4, x3,EQ 164 MOV x4,x9 //// Copying height 196 SUBS x4, x4, #1
|
/external/libxaac/decoder/armv8/ |
D | ixheaacd_cos_sin_mod_loop1.s | 42 ADD x4, x0, x1, lsl #3 //psubband1 43 SUB x4, x4, #4 61 ld1 {v3.s}[0], [x4] 62 ADD x7, x4, #256 64 SUB x4, x4, #4 98 ld1 {v3.s}[0], [x4] 99 ADD x7, x4, #256 101 SUB x4, x4, #4 134 ld1 {v3.s}[0], [x4] 135 ADD x7, x4, #256 [all …]
|
/external/eigen/bench/ |
D | README.txt | 18 3d-3x3 / 4d-4x4 / Xd-4x4 / Xd-20x20 / 23 3d-3x3 / 4d-4x4 / Xd-4x4 / Xd-20x20 / 28 3d-3x3 / 4d-4x4 / Xd-4x4 / Xd-20x20 / 33 3d-3x3 / 4d-4x4 / Xd-4x4 / Xd-20x20 / 41 double, fixed-size 4x4: 0.00165105s 0.0778739s 48 double, fixed-size 4x4: 0.000589848s 0.019949s
|