/external/libhevc/common/arm64/ |
D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 211 shrn v19.4h, v19.4s, #6 230 shrn v20.4h, v20.4s, #6 257 shrn v21.4h, v21.4s, #6 275 shrn v31.4h, v31.4s, #6 302 shrn v19.4h, v19.4s, #6 323 shrn v20.4h, v20.4s, #6 345 shrn v21.4h, v21.4s, #6 360 shrn v31.4h, v31.4s, #6 374 shrn v19.4h, v19.4s, #6 387 shrn v20.4h, v20.4s, #6 [all …]
|
D | ihevc_intra_pred_filters_chroma_mode_19_to_25.s | 256 shrn v5.8b, v2.8h,#5 //idx = pos >> 5 387 shrn v3.8b, v2.8h,#5 //idx = pos >> 5 488 shrn v3.8b, v2.8h,#5 //idx = pos >> 5
|
D | ihevc_intra_pred_luma_mode_27_to_33.s | 148 shrn v5.8b, v2.8h,#5 //idx = pos >> 5 281 shrn v3.8b, v2.8h,#5 //idx = pos >> 5 379 shrn v3.8b, v2.8h,#5 //idx = pos >> 5
|
D | ihevc_intra_pred_chroma_mode_27_to_33.s | 143 shrn v5.8b, v2.8h,#5 //idx = pos >> 5 276 shrn v3.8b, v2.8h,#5 //idx = pos >> 5 373 shrn v3.8b, v2.8h,#5 //idx = pos >> 5
|
D | ihevc_intra_pred_filters_luma_mode_19_to_25.s | 260 shrn v5.8b, v2.8h,#5 //idx = pos >> 5 387 shrn v3.8b, v2.8h,#5 //idx = pos >> 5 487 shrn v3.8b, v2.8h,#5 //idx = pos >> 5
|
/external/libavc/encoder/armv8/ |
D | ih264e_half_pel_av8.s | 374 …shrn v21.4h, v20.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 376 …shrn v20.4h, v26.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 392 …shrn v28.4h, v2.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 399 …shrn v29.4h, v26.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 425 …shrn v28.4h, v22.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 490 …shrn v21.4h, v20.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 492 …shrn v20.4h, v26.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 508 …shrn v28.4h, v6.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 515 …shrn v29.4h, v26.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… 540 …shrn v28.4h, v22.4s, #8 //// shift by 8 and later we will shift by 2 more with roundin… [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-vecFold.ll | 9 ; CHECK: shrn.8b v0, v0, #5 25 ; CHECK: shrn.4h v0, v0, #5 41 ; CHECK: shrn.2s v0, v0, #5 138 declare <8 x i8> @llvm.aarch64.neon.shrn.v8i8(<8 x i16>, i32) nounwind readnone 139 declare <4 x i16> @llvm.aarch64.neon.shrn.v4i16(<4 x i32>, i32) nounwind readnone 140 declare <2 x i32> @llvm.aarch64.neon.shrn.v2i32(<2 x i64>, i32) nounwind readnone
|
D | arm64-neon-simd-shift.ll | 215 ; CHECK: shrn {{v[0-9]+}}.8b, {{v[0-9]+}}.8h, #3 223 ; CHECK: shrn {{v[0-9]+}}.4h, {{v[0-9]+}}.4s, #9 231 ; CHECK: shrn {{v[0-9]+}}.2s, {{v[0-9]+}}.2d, #19 239 ; CHECK: shrn {{v[0-9]+}}.8b, {{v[0-9]+}}.8h, #3 247 ; CHECK: shrn {{v[0-9]+}}.4h, {{v[0-9]+}}.4s, #9 255 ; CHECK: shrn {{v[0-9]+}}.2s, {{v[0-9]+}}.2d, #19
|
D | arm64-vshift.ll | 679 ;CHECK: shrn.8b v0, {{v[0-9]+}}, #1 688 ;CHECK: shrn.4h v0, {{v[0-9]+}}, #1 697 ;CHECK: shrn.2s v0, {{v[0-9]+}}, #1 737 declare <8 x i8> @llvm.aarch64.neon.shrn.v8i8(<8 x i16>, i32) nounwind readnone 738 declare <4 x i16> @llvm.aarch64.neon.shrn.v4i16(<4 x i32>, i32) nounwind readnone 739 declare <2 x i32> @llvm.aarch64.neon.shrn.v2i32(<2 x i64>, i32) nounwind readnone
|
/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | arm64-vecFold.ll | 9 ; CHECK: shrn.8b v0, v0, #5 25 ; CHECK: shrn.4h v0, v0, #5 41 ; CHECK: shrn.2s v0, v0, #5 138 declare <8 x i8> @llvm.aarch64.neon.shrn.v8i8(<8 x i16>, i32) nounwind readnone 139 declare <4 x i16> @llvm.aarch64.neon.shrn.v4i16(<4 x i32>, i32) nounwind readnone 140 declare <2 x i32> @llvm.aarch64.neon.shrn.v2i32(<2 x i64>, i32) nounwind readnone
|
D | arm64-neon-simd-shift.ll | 215 ; CHECK: shrn {{v[0-9]+}}.8b, {{v[0-9]+}}.8h, #3 223 ; CHECK: shrn {{v[0-9]+}}.4h, {{v[0-9]+}}.4s, #9 231 ; CHECK: shrn {{v[0-9]+}}.2s, {{v[0-9]+}}.2d, #19 239 ; CHECK: shrn {{v[0-9]+}}.8b, {{v[0-9]+}}.8h, #3 247 ; CHECK: shrn {{v[0-9]+}}.4h, {{v[0-9]+}}.4s, #9 255 ; CHECK: shrn {{v[0-9]+}}.2s, {{v[0-9]+}}.2d, #19
|
D | vec_umulo.ll | 23 ; CHECK-NEXT: shrn v0.2s, v1.2d, #32 41 ; CHECK-NEXT: shrn v0.2s, v1.2d, #32 309 ; CHECK-NEXT: shrn v1.4h, v0.4s, #16
|
D | arm64-vshift.ll | 679 ;CHECK: shrn.8b v0, {{v[0-9]+}}, #1 688 ;CHECK: shrn.4h v0, {{v[0-9]+}}, #1 697 ;CHECK: shrn.2s v0, {{v[0-9]+}}, #1 737 declare <8 x i8> @llvm.aarch64.neon.shrn.v8i8(<8 x i16>, i32) nounwind readnone 738 declare <4 x i16> @llvm.aarch64.neon.shrn.v4i16(<4 x i32>, i32) nounwind readnone 739 declare <2 x i32> @llvm.aarch64.neon.shrn.v2i32(<2 x i64>, i32) nounwind readnone
|
/external/capstone/suite/MC/AArch64/ |
D | neon-simd-shift.s.cs | 92 0x20,0x84,0x0d,0x0f = shrn v0.8b, v1.8h, #3 93 0x20,0x84,0x1d,0x0f = shrn v0.4h, v1.4s, #3 94 0x20,0x84,0x3d,0x0f = shrn v0.2s, v1.2d, #3
|
/external/llvm/test/MC/AArch64/ |
D | neon-simd-shift.s | 262 shrn v0.8b, v1.8h, #3 263 shrn v0.4h, v1.4s, #3 264 shrn v0.2s, v1.2d, #3
|
D | arm64-advsimd.s | 1473 shrn.8b v0, v0, #1 1475 shrn.4h v0, v0, #3 1477 shrn.2s v0, v0, #5 1645 ; CHECK: shrn.8b v0, v0, #1 ; encoding: [0x00,0x84,0x0f,0x0f] 1647 ; CHECK: shrn.4h v0, v0, #3 ; encoding: [0x00,0x84,0x1d,0x0f] 1649 ; CHECK: shrn.2s v0, v0, #5 ; encoding: [0x00,0x84,0x3b,0x0f] 1805 shrn v9.8b, v11.8h, #1 1807 shrn v7.4h, v8.4s, #3 1809 shrn v5.2s, v6.2d, #5 1875 ; CHECK: shrn.8b v9, v11, #1 ; encoding: [0x69,0x85,0x0f,0x0f] [all …]
|
D | neon-diagnostics.s | 1807 shrn v0.8b, v1.8b, #3 1808 shrn v0.4h, v1.4h, #3 1809 shrn v0.2s, v1.2s, #3
|
/external/llvm-project/llvm/test/MC/AArch64/ |
D | neon-simd-shift.s | 262 shrn v0.8b, v1.8h, #3 263 shrn v0.4h, v1.4s, #3 264 shrn v0.2s, v1.2d, #3
|
D | arm64-advsimd.s | 1473 shrn.8b v0, v0, #1 1475 shrn.4h v0, v0, #3 1477 shrn.2s v0, v0, #5 1645 ; CHECK: shrn.8b v0, v0, #1 ; encoding: [0x00,0x84,0x0f,0x0f] 1647 ; CHECK: shrn.4h v0, v0, #3 ; encoding: [0x00,0x84,0x1d,0x0f] 1649 ; CHECK: shrn.2s v0, v0, #5 ; encoding: [0x00,0x84,0x3b,0x0f] 1805 shrn v9.8b, v11.8h, #1 1807 shrn v7.4h, v8.4s, #3 1809 shrn v5.2s, v6.2d, #5 1875 ; CHECK: shrn.8b v9, v11, #1 ; encoding: [0x69,0x85,0x0f,0x0f] [all …]
|
D | neon-diagnostics.s | 1812 shrn v0.8b, v1.8b, #3 1813 shrn v0.4h, v1.4h, #3 1814 shrn v0.2s, v1.2s, #3
|
/external/llvm-project/llvm/test/CodeGen/Thumb2/LowOverheadLoops/ |
D | safe-retaining.mir | 25 …%shrn = call <8 x i16> @llvm.arm.mve.vshrn.v8i16.v4i32(<8 x i16> %bitcast.a, <4 x i32> %masked.loa… 26 %bitcast = bitcast <8 x i16> %shrn to <4 x i32> 60 …%shrn = call <16 x i8> @llvm.arm.mve.vshrn.v16i8.v8i16(<16 x i8> %bitcast.a, <8 x i16> %masked.loa… 61 %bitcast = bitcast <16 x i8> %shrn to <8 x i16>
|
D | ctlz-non-zeros.mir | 28 …%shrn = call <16 x i8> @llvm.arm.mve.vshrn.v16i8.v8i16(<16 x i8> %ctlz, <8 x i16> %masked.load.b, … 29 %bitcast = bitcast <16 x i8> %shrn to <8 x i16> 64 …%shrn = call <8 x i16> @llvm.arm.mve.vshrn.v8i16.v4i32(<8 x i16> %ctlz, <4 x i32> %masked.load.b, … 65 %bitcast = bitcast <8 x i16> %shrn to <4 x i32> 100 …%shrn = call <8 x i16> @llvm.arm.mve.vshrn.v8i16.v4i32(<8 x i16> %bitcast.a, <4 x i32> %ctlz, i32 … 101 %bitcast = bitcast <8 x i16> %shrn to <4 x i32>
|
D | unsafe-retaining.mir | 26 …%shrn = call <8 x i16> @llvm.arm.mve.vshrn.v8i16.v4i32(<8 x i16> %bitcast.a, <4 x i32> %not, i32 1… 27 %bitcast = bitcast <8 x i16> %shrn to <4 x i32> 63 …%shrn = call <8 x i16> @llvm.arm.mve.vshrn.v8i16.v4i32(<8 x i16> %bitcast.a, <4 x i32> %or, i32 3,… 64 %bitcast = bitcast <8 x i16> %shrn to <4 x i32>
|
/external/libavc/common/armv8/ |
D | ih264_resi_trans_quant_av8.s | 524 shrn v0.4h, v22.4s, #1 //i4_value = (x0 + x1) >> 1; 526 shrn v1.4h, v24.4s, #1 //i4_value = (x0 - x1) >> 1;
|
/external/llvm-project/llvm/test/MC/Disassembler/AArch64/ |
D | arm64-advsimd.txt | 2048 # CHECK: shrn.8b v0, v0, #0x7 2050 # CHECK: shrn.4h v0, v0, #0xd 2052 # CHECK: shrn.2s v0, v0, #0x1b
|