/external/llvm/unittests/ADT/ |
D | MapVectorTest.cpp | 43 MapVector<int, int> MV; in TEST() local 46 R = MV.insert(std::make_pair(1, 2)); in TEST() 47 ASSERT_EQ(R.first, MV.begin()); in TEST() 52 R = MV.insert(std::make_pair(1, 3)); in TEST() 53 ASSERT_EQ(R.first, MV.begin()); in TEST() 58 R = MV.insert(std::make_pair(4, 5)); in TEST() 59 ASSERT_NE(R.first, MV.end()); in TEST() 64 EXPECT_EQ(MV.size(), 2u); in TEST() 65 EXPECT_EQ(MV[1], 2); in TEST() 66 EXPECT_EQ(MV[4], 5); in TEST() [all …]
|
/external/llvm-project/llvm/unittests/ADT/ |
D | MapVectorTest.cpp | 42 MapVector<int, int> MV; in TEST() local 45 R = MV.insert(std::make_pair(1, 2)); in TEST() 46 ASSERT_EQ(R.first, MV.begin()); in TEST() 51 R = MV.insert(std::make_pair(1, 3)); in TEST() 52 ASSERT_EQ(R.first, MV.begin()); in TEST() 57 R = MV.insert(std::make_pair(4, 5)); in TEST() 58 ASSERT_NE(R.first, MV.end()); in TEST() 63 EXPECT_EQ(MV.size(), 2u); in TEST() 64 EXPECT_EQ(MV[1], 2); in TEST() 65 EXPECT_EQ(MV[4], 5); in TEST() [all …]
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_mcomp.h | 38 MV ss_mv[8 * MAX_MVSEARCH_STEPS]; // Motion vector 45 const MV *mv) { in get_buf_from_mv() 52 void vp9_set_mv_search_range(MvLimits *mv_limits, const MV *mv); 53 int vp9_mv_bit_cost(const MV *mv, const MV *ref, const int *mvjcost, 57 int vp9_get_mvpred_var(const MACROBLOCK *x, const MV *best_mv, 58 const MV *center_mv, const vp9_variance_fn_ptr_t *vfp, 60 int vp9_get_mvpred_av_var(const MACROBLOCK *x, const MV *best_mv, 61 const MV *center_mv, const uint8_t *second_pred, 78 const MV *ref_mv); 81 const MACROBLOCK *x, MV *bestmv, const MV *ref_mv, int allow_hp, [all …]
|
D | vp9_mcomp.c | 32 void vp9_set_mv_search_range(MvLimits *mv_limits, const MV *mv) { in vp9_set_mv_search_range() 53 const MV *ref_mv) { in vp9_set_subpel_mv_search_range() 80 static INLINE int mv_cost(const MV *mv, const int *joint_cost, in mv_cost() 88 int vp9_mv_bit_cost(const MV *mv, const MV *ref, const int *mvjcost, in vp9_mv_bit_cost() 90 const MV diff = { mv->row - ref->row, mv->col - ref->col }; in vp9_mv_bit_cost() 95 static int mv_err_cost(const MV *mv, const MV *ref, const int *mvjcost, in mv_err_cost() 98 const MV diff = { mv->row - ref->row, mv->col - ref->col }; in mv_err_cost() 107 static int mvsad_err_cost(const MACROBLOCK *x, const MV *mv, const MV *ref, in mvsad_err_cost() 109 const MV diff = { mv->row - ref->row, mv->col - ref->col }; in mvsad_err_cost() 121 const MV ss_mvs[] = { { -len, 0 }, { len, 0 }, { 0, -len }, { 0, len } }; in vp9_init_dsmotion_compensation() [all …]
|
/external/llvm-project/llvm/test/Transforms/MemCpyOpt/ |
D | form-memset.ll | 83 %struct.MV = type { i16, i16 } 90 ; CHECK-NEXT: [[LEFT_MVD:%.*]] = alloca [8 x %struct.MV], align 8 91 ; CHECK-NEXT: [[UP_MVD:%.*]] = alloca [8 x %struct.MV], align 8 100 ; CHECK-NEXT: [[TMP43:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … 102 ; CHECK-NEXT: [[TMP46:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … 103 ; CHECK-NEXT: [[TMP57:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … 104 ; CHECK-NEXT: [[TMP60:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … 105 ; CHECK-NEXT: [[TMP71:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … 106 ; CHECK-NEXT: [[TMP74:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … 107 ; CHECK-NEXT: [[TMP85:%.*]] = getelementptr [8 x %struct.MV], [8 x %struct.MV]* [[UP_MVD]], i32 … [all …]
|
/external/llvm/test/Transforms/MemCpyOpt/ |
D | form-memset.ll | 60 %struct.MV = type { i16, i16 } 66 %left_mvd = alloca [8 x %struct.MV] ; <[8 x %struct.MV]*> [#uses=17] 67 %up_mvd = alloca [8 x %struct.MV] ; <[8 x %struct.MV]*> [#uses=17] 84 …%tmp43 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 7, i32 0 ; <i16*> … 86 …%tmp46 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 7, i32 1 ; <i16*> … 88 …%tmp57 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 6, i32 0 ; <i16*> … 90 …%tmp60 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 6, i32 1 ; <i16*> … 92 …%tmp71 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 5, i32 0 ; <i16*> … 94 …%tmp74 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 5, i32 1 ; <i16*> … 96 …%tmp85 = getelementptr [8 x %struct.MV], [8 x %struct.MV]* %up_mvd, i32 0, i32 4, i32 0 ; <i16*> … [all …]
|
/external/libaom/libaom/av1/encoder/ |
D | encodemv.h | 21 void av1_encode_mv(AV1_COMP *cpi, aom_writer *w, const MV *mv, const MV *ref, 24 void av1_update_mv_stats(const MV *mv, const MV *ref, nmv_context *mvctx, 33 void av1_encode_dv(aom_writer *w, const MV *mv, const MV *ref, 46 static INLINE MV_JOINT_TYPE av1_get_mv_joint(const MV *mv) { in av1_get_mv_joint()
|
D | mcomp.c | 36 const MACROBLOCK *x, const MV *ref_mv) { in init_mv_cost_params() 59 const MACROBLOCK *x, BLOCK_SIZE bsize, const MV *ref_mv, in av1_make_default_fullpel_ms_params() 90 const MV *ref_mv, const int *cost_list) { in av1_make_default_subpel_ms_params() 124 void av1_set_mv_search_range(FullMvLimits *mv_limits, const MV *mv) { in av1_set_mv_search_range() 178 static INLINE int mv_cost(const MV *mv, const int *joint_cost, in mv_cost() 189 int av1_mv_bit_cost(const MV *mv, const MV *ref_mv, const int *mvjcost, in av1_mv_bit_cost() 191 const MV diff = { mv->row - ref_mv->row, mv->col - ref_mv->col }; in av1_mv_bit_cost() 199 static INLINE int mv_err_cost(const MV *mv, const MV *ref_mv, in mv_err_cost() 202 const MV diff = { mv->row - ref_mv->row, mv->col - ref_mv->col }; in mv_err_cost() 203 const MV abs_diff = { abs(diff.row), abs(diff.col) }; in mv_err_cost() [all …]
|
D | mcomp.h | 64 const MV *ref_mv; 73 int av1_mv_bit_cost(const MV *mv, const MV *ref_mv, const int *mvjcost, 77 const MV *ref_mv, const aom_variance_fn_ptr_t *vfp); 159 const MV *ref_mv, 204 void av1_set_mv_search_range(FullMvLimits *mv_limits, const MV *mv); 211 const MV *ref_mv); 273 const MV *ref_mv, const int *cost_list); 277 MV start_mv, MV *bestmv, int *distortion, 302 const MV *ref_mv) { in av1_set_subpel_mv_search_range() 320 MV mv) { in av1_is_subpelmv_in_range()
|
/external/libvpx/libvpx/test/ |
D | non_greedy_mv_test.cc | 19 MV **buffer_ptr) { in read_in_mf() 28 *buffer_ptr = (MV *)malloc((*rows_ptr) * (*cols_ptr) * sizeof(MV)); in read_in_mf() 63 static void compare_mf(const MV *mf1, const MV *mf2, int rows, int cols, in compare_mf() 70 MV mv1 = mf1[idx]; in compare_mf() 71 MV mv2 = mf2[idx]; in compare_mf() 132 MV *search_mf = NULL; in TEST() 133 MV *smooth_mf = NULL; in TEST() 134 MV *estimation = NULL; in TEST() 135 MV *ground_truth = NULL; in TEST() 149 smooth_mf = (MV *)malloc(rows * cols * sizeof(MV)); in TEST() [all …]
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_mv.h | 27 } MV; typedef 31 MV as_mv; 39 static INLINE int is_zero_mv(const MV *mv) { in is_zero_mv() 43 static INLINE int is_equal_mv(const MV *a, const MV *b) { in is_equal_mv() 47 static INLINE void clamp_mv(MV *mv, int min_col, int max_col, int min_row, in clamp_mv()
|
D | vp9_reconinter.c | 25 const MV *src_mv, const struct scale_factors *sf, int w, int h, int ref, in vp9_highbd_build_inter_predictor() 29 const MV mv_q4 = { is_q4 ? src_mv->row : src_mv->row * 2, in vp9_highbd_build_inter_predictor() 44 int dst_stride, const MV *src_mv, in vp9_build_inter_predictor() 49 const MV mv_q4 = { is_q4 ? src_mv->row : src_mv->row * 2, in vp9_build_inter_predictor() 65 static MV mi_mv_pred_q4(const MODE_INFO *mi, int idx) { in mi_mv_pred_q4() 66 MV res = { round_mv_comp_q4(mi->bmi[0].as_mv[idx].as_mv.row + in mi_mv_pred_q4() 81 static MV mi_mv_pred_q2(const MODE_INFO *mi, int idx, int block0, int block1) { in mi_mv_pred_q2() 82 MV res = { round_mv_comp_q2(mi->bmi[block0].as_mv[idx].as_mv.row + in mi_mv_pred_q2() 90 MV clamp_mv_to_umv_border_sb(const MACROBLOCKD *xd, const MV *src_mv, int bw, in clamp_mv_to_umv_border_sb() 99 MV clamped_mv = { (short)(src_mv->row * (1 << (1 - ss_y))), in clamp_mv_to_umv_border_sb() [all …]
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/GlobalISel/ |
D | legalize-sextload-constant-32bit.mir | 13 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 14 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 4, addrspace 6) 31 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 32 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 2, addrspace 6) 34 ; CI: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C1]](s64) 60 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 61 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 1, addrspace 6) 63 ; CI: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C1]](s64) 66 ; CI: [[PTR_ADD1:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C2]](s64) 69 ; CI: [[PTR_ADD2:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C3]](s64) [all …]
|
D | regbankselect-merge-values.mir | 16 ; CHECK: [[MV:%[0-9]+]]:sgpr(s64) = G_MERGE_VALUES [[EXTRACT]](s32), [[EXTRACT1]](s32) 17 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 36 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[EXTRACT]](s32), [[EXTRACT1]](s32) 37 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 55 ; CHECK: [[MV:%[0-9]+]]:agpr(s64) = G_MERGE_VALUES [[COPY]](s32), [[COPY1]](s32) 56 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 75 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[COPY2]](s32), [[COPY3]](s32) 76 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 95 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[COPY2]](s32), [[COPY3]](s32) 96 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64)
|
D | legalize-zextload-constant-32bit.mir | 13 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 14 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 4, addrspace 6) 31 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 32 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 2, addrspace 6) 34 ; CI: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C1]](s64) 60 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 61 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 1, addrspace 6) 63 ; CI: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C1]](s64) 66 ; CI: [[PTR_ADD1:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C2]](s64) 69 ; CI: [[PTR_ADD2:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C3]](s64) [all …]
|
D | legalize-anyext.mir | 286 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[DEF]](s32) 288 ; CHECK: [[MV1:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[MV]](s64), [[DEF1]](s64), [[DEF1]](s64) 305 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[DEF]](s32) 307 ; CHECK: [[MV1:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[MV]](s64), [[DEF1]](s64) 323 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[DEF]](s32) 325 …; CHECK: [[MV1:%[0-9]+]]:_(s320) = G_MERGE_VALUES [[MV]](s64), [[DEF1]](s64), [[DEF1]](s64), [[DEF… 342 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[DEF]](s32) 344 ; CHECK: [[MV1:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[MV]](s64), [[DEF1]](s64), [[DEF1]](s64) 360 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[DEF]](s32) 362 …; CHECK: [[MV1:%[0-9]+]]:_(s448) = G_MERGE_VALUES [[MV]](s64), [[DEF1]](s64), [[DEF1]](s64), [[DEF… [all …]
|
D | regbankselect-sext-inreg.mir | 127 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[SEXT_INREG]](s32), [[ASHR]](s32) 128 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 149 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[SEXT_INREG]](s32), [[ASHR]](s32) 150 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 171 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[COPY1]](s32), [[ASHR]](s32) 172 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 192 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[COPY1]](s32), [[SEXT_INREG]](s32) 193 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) 213 ; CHECK: [[MV:%[0-9]+]]:vgpr(s64) = G_MERGE_VALUES [[COPY1]](s32), [[SEXT_INREG]](s32) 214 ; CHECK: S_ENDPGM 0, implicit [[MV]](s64) [all …]
|
D | legalize-load-constant-32bit.mir | 13 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 14 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 1, addrspace 6) 16 ; CI: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C1]](s64) 19 ; CI: [[PTR_ADD1:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C2]](s64) 22 ; CI: [[PTR_ADD2:%[0-9]+]]:_(p4) = G_PTR_ADD [[MV]], [[C3]](s64) 57 ; CI: [[MV:%[0-9]+]]:_(p4) = G_MERGE_VALUES [[COPY]](p6), [[C]](p6) 58 ; CI: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[MV]](p4) :: (load 4, addrspace 6)
|
D | legalize-zext.mir | 82 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[C]](s32) 84 ; CHECK: [[MV1:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[MV]](s64), [[C1]](s64), [[C1]](s64) 345 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[C]](s32) 347 ; CHECK: [[MV1:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[MV]](s64), [[C1]](s64) 363 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[C]](s32) 365 …; CHECK: [[MV1:%[0-9]+]]:_(s320) = G_MERGE_VALUES [[MV]](s64), [[C1]](s64), [[C1]](s64), [[C1]](s6… 383 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[C]](s32) 385 ; CHECK: [[MV1:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[MV]](s64), [[C1]](s64), [[C1]](s64) 401 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[C]](s32) 403 …; CHECK: [[MV1:%[0-9]+]]:_(s448) = G_MERGE_VALUES [[MV]](s64), [[C1]](s64), [[C1]](s64), [[C1]](s6… [all …]
|
D | irtranslator-memory-intrinsics.ll | 14 ; CHECK: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY]](s32), [[COPY1]](s32) 17 …; CHECK: G_MEMCPY [[MV]](p1), [[COPY2]](p3), [[TRUNC]](s32), 0 :: (store 1 into %ir.dst, addrspa… 32 ; CHECK: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY]](s32), [[COPY1]](s32) 34 …; CHECK: G_MEMCPY [[MV]](p1), [[COPY2]](p3), [[C]](s32), 0 :: (store 1 into %ir.dst, addrspace 1… 49 ; CHECK: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY]](s32), [[COPY1]](s32) 52 …; CHECK: G_MEMCPY [[MV]](p1), [[COPY2]](p3), [[ZEXT]](s32), 0 :: (store 1 into %ir.dst, addrspac… 67 ; CHECK: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY1]](s32), [[COPY2]](s32) 70 …; CHECK: G_MEMCPY [[COPY]](p3), [[MV]](p1), [[TRUNC]](s32), 0 :: (store 1 into %ir.dst, addrspac… 85 ; CHECK: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY1]](s32), [[COPY2]](s32) 87 …; CHECK: G_MEMCPY [[COPY]](p3), [[MV]](p1), [[C]](s32), 0 :: (store 1 into %ir.dst, addrspace 3)… [all …]
|
D | legalize-sext.mir | 308 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[ASHR]](s32) 310 ; CHECK: [[MV2:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[MV]](s64), [[MV1]](s64), [[MV1]](s64) 328 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[ASHR]](s32) 330 ; CHECK: [[MV2:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[MV]](s64), [[MV1]](s64) 347 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[ASHR]](s32) 349 …; CHECK: [[MV2:%[0-9]+]]:_(s320) = G_MERGE_VALUES [[MV]](s64), [[MV1]](s64), [[MV1]](s64), [[MV1]]… 367 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[ASHR]](s32) 369 ; CHECK: [[MV2:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[MV]](s64), [[MV1]](s64), [[MV1]](s64) 386 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[ASHR]](s32) 388 …; CHECK: [[MV2:%[0-9]+]]:_(s448) = G_MERGE_VALUES [[MV]](s64), [[MV1]](s64), [[MV1]](s64), [[MV1]]… [all …]
|
D | artifact-combiner-trunc.mir | 50 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[C]](s32), [[C1]](s32) 51 ; CHECK: $vgpr0_vgpr1 = COPY [[MV]](s64) 68 ; CHECK: [[MV:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[C]](p0), [[C1]](p0) 69 ; CHECK: [[TRUNC:%[0-9]+]]:_(s32) = G_TRUNC [[MV]](s128) 87 ; CHECK: [[MV:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[C]](p0), [[C1]](p0) 88 ; CHECK: [[TRUNC:%[0-9]+]]:_(s64) = G_TRUNC [[MV]](s128) 106 ; CHECK: [[MV:%[0-9]+]]:_(s192) = G_MERGE_VALUES [[C]](p0), [[C1]](p0), [[C]](p0) 107 ; CHECK: [[TRUNC:%[0-9]+]]:_(s128) = G_TRUNC [[MV]](s192)
|
/external/libaom/libaom/av1/common/ |
D | mv.h | 43 } MV; typedef 45 static const MV kZeroMv = { 0, 0 }; 50 MV as_mv; 75 static AOM_INLINE FULLPEL_MV get_fullmv_from_mv(const MV *subpel_mv) { in get_fullmv_from_mv() 81 static AOM_INLINE MV get_mv_from_fullmv(const FULLPEL_MV *full_mv) { in get_mv_from_fullmv() 82 const MV subpel_mv = { (int16_t)GET_MV_SUBPEL(full_mv->row), in get_mv_from_fullmv() 227 static INLINE void integer_mv_precision(MV *mv) { in integer_mv_precision() 332 static INLINE int is_zero_mv(const MV *mv) { in is_zero_mv() 336 static INLINE int is_equal_mv(const MV *a, const MV *b) { in is_equal_mv() 340 static INLINE void clamp_mv(MV *mv, const SubpelMvLimits *mv_limits) { in clamp_mv()
|
/external/llvm-project/llvm/test/CodeGen/Mips/GlobalISel/irtranslator/ |
D | split_args.ll | 9 ; MIPS32: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[COPY1]](s32) 10 ; MIPS32: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[MV]](s64) 33 ; MIPS32: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[LOAD]](s32), [[LOAD1]](s32) 34 ; MIPS32: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[MV]](s64) 49 ; MIPS32: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY1]](s32), [[COPY2]](s32) 50 ; MIPS32: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[MV]](s64) 76 ; MIPS32: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[LOAD1]](s32), [[LOAD2]](s32) 77 ; MIPS32: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[MV]](s64) 99 ; MIPS32: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[LOAD]](s32), [[LOAD1]](s32) 100 ; MIPS32: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[MV]](s64)
|
/external/llvm-project/llvm/test/CodeGen/AArch64/GlobalISel/ |
D | legalize-sext-zext-128.mir | 16 ; CHECK: [[MV:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[COPY]](s64), [[ASHR]](s64) 17 ; CHECK: G_STORE [[MV]](s128), [[COPY1]](p0) :: (store 16) 38 ; CHECK: [[MV:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[COPY]](s64), [[C]](s64) 39 ; CHECK: G_STORE [[MV]](s128), [[COPY1]](p0) :: (store 16) 60 ; CHECK: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY]](s32), [[C]](s32) 62 ; CHECK: [[MV1:%[0-9]+]]:_(s128) = G_MERGE_VALUES [[MV]](s64), [[C1]](s64)
|