/external/libvpx/libvpx/vp8/encoder/ |
D | mcomp.c | 33 clamp((mv->as_mv.row - ref->as_mv.row) >> 1, 0, MVvals); in vp8_mv_bit_cost() 35 clamp((mv->as_mv.col - ref->as_mv.col) >> 1, 0, MVvals); in vp8_mv_bit_cost() 44 clamp((mv->as_mv.row - ref->as_mv.row) >> 1, 0, MVvals); in mv_err_cost() 46 clamp((mv->as_mv.col - ref->as_mv.col) >> 1, 0, MVvals); in mv_err_cost() 59 return ((mvsadcost[0][(mv->as_mv.row - ref->as_mv.row)] + in mvsad_err_cost() 60 mvsadcost[1][(mv->as_mv.col - ref->as_mv.col)]) * in mvsad_err_cost() 229 int rr = ref_mv->as_mv.row >> 1, rc = ref_mv->as_mv.col >> 1; in vp8_find_best_sub_pixel_step_iteratively() 230 int br = bestmv->as_mv.row * 4, bc = bestmv->as_mv.col * 4; in vp8_find_best_sub_pixel_step_iteratively() 241 (ref_mv->as_mv.col >> 1) - ((1 << mvlong_width) - 1)); in vp8_find_best_sub_pixel_step_iteratively() 243 (ref_mv->as_mv.col >> 1) + ((1 << mvlong_width) - 1)); in vp8_find_best_sub_pixel_step_iteratively() [all …]
|
D | pickinter.c | 138 bestmv->as_mv.row *= 8; in vp8_skip_fractional_mv_step() 139 bestmv->as_mv.col *= 8; in vp8_skip_fractional_mv_step() 152 int xoffset = this_mv.as_mv.col & 7; in vp8_get_inter_mbpred_error() 153 int yoffset = this_mv.as_mv.row & 7; in vp8_get_inter_mbpred_error() 155 in_what += (this_mv.as_mv.row >> 3) * pre_stride + (this_mv.as_mv.col >> 3); in vp8_get_inter_mbpred_error() 383 x->MVcount[0][mv_max + ((xd->mode_info_context->mbmi.mv.as_mv.row - in update_mvcount() 384 best_ref_mv->as_mv.row) >> in update_mvcount() 386 x->MVcount[1][mv_max + ((xd->mode_info_context->mbmi.mv.as_mv.col - in update_mvcount() 387 best_ref_mv->as_mv.col) >> in update_mvcount() 431 (*parent_ref_mv).as_mv.row = store_mode_info[parent_mb_index].mv.as_mv.row * in get_lower_res_motion_info() [all …]
|
D | mr_dissim.c | 33 mvx[cnt] = x->mbmi.mv.as_mv.row; \ 34 mvy[cnt] = x->mbmi.mv.as_mv.col; \ 40 mvx[cnt] = x->mbmi.mv.as_mv.row; \ 41 mvy[cnt] = x->mbmi.mv.as_mv.col; \ 175 mmvx = VPXMAX(abs(min_mvx - here->mbmi.mv.as_mv.row), in vp8_cal_dissimilarity() 176 abs(max_mvx - here->mbmi.mv.as_mv.row)); in vp8_cal_dissimilarity() 177 mmvy = VPXMAX(abs(min_mvy - here->mbmi.mv.as_mv.col), in vp8_cal_dissimilarity() 178 abs(max_mvy - here->mbmi.mv.as_mv.col)); in vp8_cal_dissimilarity()
|
D | rdopt.c | 379 int mv_row = x->e_mbd.mode_info_context->mbmi.mv.as_mv.row; in VP8_UVSSE() 380 int mv_col = x->e_mbd.mode_info_context->mbmi.mv.as_mv.col; in VP8_UVSSE() 1057 mvp_full.as_mv.row = bsi->mvp.as_mv.row >> 3; in rd_check_segment() 1058 mvp_full.as_mv.col = bsi->mvp.as_mv.col >> 3; in rd_check_segment() 1128 if (((mode_mv[this_mode].as_mv.row >> 3) < x->mv_row_min) || in rd_check_segment() 1129 ((mode_mv[this_mode].as_mv.row >> 3) > x->mv_row_max) || in rd_check_segment() 1130 ((mode_mv[this_mode].as_mv.col >> 3) < x->mv_col_min) || in rd_check_segment() 1131 ((mode_mv[this_mode].as_mv.col >> 3) > x->mv_col_max)) { in rd_check_segment() 1178 bsi->mvs[i].as_mv = x->partition_info->bmi[i].mv.as_mv; in rd_check_segment() 1233 int col_min = ((best_ref_mv->as_mv.col + 7) >> 3) - MAX_FULL_PEL_VAL; in vp8_rd_pick_best_mbsegmentation() [all …]
|
D | temporal_filter.c | 138 best_ref_mv1_full.as_mv.col = best_ref_mv1.as_mv.col >> 3; in vp8_temporal_filter_find_matching_mb_c() 139 best_ref_mv1_full.as_mv.row = best_ref_mv1.as_mv.row >> 3; in vp8_temporal_filter_find_matching_mb_c() 242 mbd->block[0].bmi.mv.as_mv.row = 0; in vp8_temporal_filter_iterate_c() 243 mbd->block[0].bmi.mv.as_mv.col = 0; in vp8_temporal_filter_iterate_c() 270 cpi->frames[frame]->y_stride, mbd->block[0].bmi.mv.as_mv.row, in vp8_temporal_filter_iterate_c() 271 mbd->block[0].bmi.mv.as_mv.col, predictor); in vp8_temporal_filter_iterate_c()
|
D | firstpass.c | 433 ref_mv_full.as_mv.col = ref_mv->as_mv.col >> 3; in first_pass_motion_search() 434 ref_mv_full.as_mv.row = ref_mv->as_mv.row >> 3; in first_pass_motion_search() 442 best_mv->row = tmp_mv.as_mv.row; in first_pass_motion_search() 443 best_mv->col = tmp_mv.as_mv.col; in first_pass_motion_search() 465 best_mv->row = tmp_mv.as_mv.row; in first_pass_motion_search() 466 best_mv->col = tmp_mv.as_mv.col; in first_pass_motion_search() 609 d->bmi.mv.as_mv.row = 0; in vp8_first_pass() 610 d->bmi.mv.as_mv.col = 0; in vp8_first_pass() 619 first_pass_motion_search(cpi, x, &best_ref_mv, &d->bmi.mv.as_mv, in vp8_first_pass() 632 d->bmi.mv.as_mv.row = tmp_mv.row; in vp8_first_pass() [all …]
|
/external/libvpx/libvpx/vp8/common/ |
D | reconinter.c | 64 ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + in vp8_build_inter_predictors_b() 65 (d->bmi.mv.as_mv.col >> 3); in vp8_build_inter_predictors_b() 67 if (d->bmi.mv.as_mv.row & 7 || d->bmi.mv.as_mv.col & 7) { in vp8_build_inter_predictors_b() 68 sppf(ptr, pre_stride, d->bmi.mv.as_mv.col & 7, d->bmi.mv.as_mv.row & 7, in vp8_build_inter_predictors_b() 86 ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + in build_inter_predictors4b() 87 (d->bmi.mv.as_mv.col >> 3); in build_inter_predictors4b() 89 if (d->bmi.mv.as_mv.row & 7 || d->bmi.mv.as_mv.col & 7) { in build_inter_predictors4b() 90 x->subpixel_predict8x8(ptr, pre_stride, d->bmi.mv.as_mv.col & 7, in build_inter_predictors4b() 91 d->bmi.mv.as_mv.row & 7, dst, dst_stride); in build_inter_predictors4b() 101 ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + in build_inter_predictors2b() [all …]
|
D | findnearmv.h | 27 mvp->as_mv.row *= -1; in mv_bias() 28 mvp->as_mv.col *= -1; in mv_bias() 35 if (mv->as_mv.col < (xd->mb_to_left_edge - LEFT_TOP_MARGIN)) { in vp8_clamp_mv2() 36 mv->as_mv.col = xd->mb_to_left_edge - LEFT_TOP_MARGIN; in vp8_clamp_mv2() 37 } else if (mv->as_mv.col > xd->mb_to_right_edge + RIGHT_BOTTOM_MARGIN) { in vp8_clamp_mv2() 38 mv->as_mv.col = xd->mb_to_right_edge + RIGHT_BOTTOM_MARGIN; in vp8_clamp_mv2() 41 if (mv->as_mv.row < (xd->mb_to_top_edge - LEFT_TOP_MARGIN)) { in vp8_clamp_mv2() 42 mv->as_mv.row = xd->mb_to_top_edge - LEFT_TOP_MARGIN; in vp8_clamp_mv2() 43 } else if (mv->as_mv.row > xd->mb_to_bottom_edge + RIGHT_BOTTOM_MARGIN) { in vp8_clamp_mv2() 44 mv->as_mv.row = xd->mb_to_bottom_edge + RIGHT_BOTTOM_MARGIN; in vp8_clamp_mv2() [all …]
|
D | debugmodes.c | 101 fprintf(mvs, "%5d:%-5d", mi[mb_index].mbmi.mv.as_mv.row / 2, in vp8_print_modes_and_motion_vectors() 102 mi[mb_index].mbmi.mv.as_mv.col / 2); in vp8_print_modes_and_motion_vectors() 125 fprintf(mvs, "%3d:%-3d ", mi[mb_index].bmi[bindex].mv.as_mv.row, in vp8_print_modes_and_motion_vectors() 126 mi[mb_index].bmi[bindex].mv.as_mv.col); in vp8_print_modes_and_motion_vectors()
|
D | findnearmv.c | 125 inv->as_mv.row = src->as_mv.row * -1; in invert_and_clamp_mvs() 126 inv->as_mv.col = src->as_mv.col * -1; in invert_and_clamp_mvs()
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_reconinter.c | 68 mi->bmi[0].as_mv[idx].as_mv.row + mi->bmi[1].as_mv[idx].as_mv.row + in mi_mv_pred_q4() 69 mi->bmi[2].as_mv[idx].as_mv.row + mi->bmi[3].as_mv[idx].as_mv.row), in mi_mv_pred_q4() 71 mi->bmi[0].as_mv[idx].as_mv.col + mi->bmi[1].as_mv[idx].as_mv.col + in mi_mv_pred_q4() 72 mi->bmi[2].as_mv[idx].as_mv.col + mi->bmi[3].as_mv[idx].as_mv.col) in mi_mv_pred_q4() 82 MV res = { round_mv_comp_q2(mi->bmi[block0].as_mv[idx].as_mv.row + in mi_mv_pred_q2() 83 mi->bmi[block1].as_mv[idx].as_mv.row), in mi_mv_pred_q2() 84 round_mv_comp_q2(mi->bmi[block0].as_mv[idx].as_mv.col + in mi_mv_pred_q2() 85 mi->bmi[block1].as_mv[idx].as_mv.col) }; in mi_mv_pred_q2() 117 case 0: res = mi->bmi[block].as_mv[ref].as_mv; break; in average_split_mvs() 142 : mi->mv[ref].as_mv; in build_inter_predictors()
|
D | vp9_mvref_common.c | 105 mv.as_mv.row *= -1; in find_mv_refs_idx() 106 mv.as_mv.col *= -1; in find_mv_refs_idx() 117 mv.as_mv.row *= -1; in find_mv_refs_idx() 118 mv.as_mv.col *= -1; in find_mv_refs_idx() 130 clamp_mv_ref(&mv_ref_list[i].as_mv, xd); in find_mv_refs_idx() 146 lower_mv_precision(&mvlist[i].as_mv, allow_hp); in vp9_find_best_ref_mvs() 147 clamp_mv2(&mvlist[i].as_mv, xd); in vp9_find_best_ref_mvs() 175 nearest_mv->as_int = bmi[0].as_mv[ref].as_int; in vp9_append_sub8x8_mvs_for_idx() 184 candidates[0] = bmi[1].as_mv[ref]; in vp9_append_sub8x8_mvs_for_idx() 185 candidates[1] = bmi[0].as_mv[ref]; in vp9_append_sub8x8_mvs_for_idx() [all …]
|
D | vp9_mvref_common.h | 229 .as_mv[which_mv] in get_sub_block_mv() 239 mv.as_mv.row *= -1; in scale_mv() 240 mv.as_mv.col *= -1; in scale_mv()
|
/external/libaom/libaom/av1/common/ |
D | mv.h | 34 MV as_mv; member 233 res.as_mv.row = gm->wmmat[0] >> GM_TRANS_ONLY_PREC_DIFF; in gm_get_motion_vector() 234 res.as_mv.col = gm->wmmat[1] >> GM_TRANS_ONLY_PREC_DIFF; in gm_get_motion_vector() 235 assert(IMPLIES(1 & (res.as_mv.row | res.as_mv.col), allow_hp)); in gm_get_motion_vector() 237 integer_mv_precision(&res.as_mv); in gm_get_motion_vector() 257 res.as_mv.row = ty; in gm_get_motion_vector() 258 res.as_mv.col = tx; in gm_get_motion_vector() 261 integer_mv_precision(&res.as_mv); in gm_get_motion_vector()
|
D | mvref_common.c | 61 if ((abs(mi->mv[idx].as_mv.row) > REFMVS_LIMIT) || in av1_copy_frame_mvs() 62 (abs(mi->mv[idx].as_mv.col) > REFMVS_LIMIT)) in av1_copy_frame_mvs() 359 get_mv_projection(&this_refmv.as_mv, prev_frame_mvs->mfmv0.as_mv, in add_tpl_ref_mv() 361 lower_mv_precision(&this_refmv.as_mv, cm->allow_high_precision_mv, in add_tpl_ref_mv() 365 if (abs(this_refmv.as_mv.row - gm_mv_candidates[0].as_mv.row) >= 16 || in add_tpl_ref_mv() 366 abs(this_refmv.as_mv.col - gm_mv_candidates[0].as_mv.col) >= 16) in add_tpl_ref_mv() 399 get_mv_projection(&this_refmv.as_mv, prev_frame_mvs->mfmv0.as_mv, in add_tpl_ref_mv() 401 get_mv_projection(&comp_refmv.as_mv, prev_frame_mvs->mfmv0.as_mv, in add_tpl_ref_mv() 404 lower_mv_precision(&this_refmv.as_mv, cm->allow_high_precision_mv, in add_tpl_ref_mv() 406 lower_mv_precision(&comp_refmv.as_mv, cm->allow_high_precision_mv, in add_tpl_ref_mv() [all …]
|
D | mvref_common.h | 275 ref_dv->as_mv.row = 0; in av1_find_ref_dv() 276 ref_dv->as_mv.col = -MI_SIZE * mib_size - INTRABC_DELAY_PIXELS; in av1_find_ref_dv() 278 ref_dv->as_mv.row = -MI_SIZE * mib_size; in av1_find_ref_dv() 279 ref_dv->as_mv.col = 0; in av1_find_ref_dv() 281 ref_dv->as_mv.row *= 8; in av1_find_ref_dv() 282 ref_dv->as_mv.col *= 8; in av1_find_ref_dv()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_diamond_search_sad_avx.c | 31 result.as_mv.row = row; in pack_int_mv() 32 result.as_mv.col = col; in pack_int_mv() 45 return joint_cost[get_mv_joint(mv)] + comp_cost[0][mv.as_mv.row] + in mv_cost() 46 comp_cost[1][mv.as_mv.col]; in mv_cost() 52 pack_int_mv(mv.as_mv.row - ref->row, mv.as_mv.col - ref->col); in mvsad_err_cost() 101 const int ref_row = clamp(ref_mv->row, minmv.as_mv.row, maxmv.as_mv.row); in vp9_diamond_search_sad_avx() 102 const int ref_col = clamp(ref_mv->col, minmv.as_mv.col, maxmv.as_mv.col); in vp9_diamond_search_sad_avx() 134 best_sad += mvsad_err_cost(x, bmv, &fcenter_mv.as_mv, sad_per_bit); in vp9_diamond_search_sad_avx() 308 *best_mv = bmv.as_mv; in vp9_diamond_search_sad_avx()
|
/external/libvpx/libvpx/vp8/decoder/ |
D | error_concealment.c | 164 new_row = row - bmi->mv.as_mv.row; in calculate_overlaps() 165 new_col = col - bmi->mv.as_mv.col; in calculate_overlaps() 223 col_acc += overlaps[i].overlap * overlaps[i].bmi->mv.as_mv.col; in estimate_mv() 224 row_acc += overlaps[i].overlap * overlaps[i].bmi->mv.as_mv.row; in estimate_mv() 229 bmi->mv.as_mv.col = col_acc / overlap_sum; in estimate_mv() 230 bmi->mv.as_mv.row = row_acc / overlap_sum; in estimate_mv() 232 bmi->mv.as_mv.col = 0; in estimate_mv() 233 bmi->mv.as_mv.row = 0; in estimate_mv() 245 MV *const filtered_mv = &(mi->mbmi.mv.as_mv); in estimate_mb_mvs() 265 filtered_mv->col += bmi[i].mv.as_mv.col; in estimate_mb_mvs() [all …]
|
D | decodemv.c | 246 blockmv.as_mv.row = read_mvcomponent(bc, &mvc[0]) * 2; in decode_split_mv() 247 blockmv.as_mv.row += best_mv.as_mv.row; in decode_split_mv() 248 blockmv.as_mv.col = read_mvcomponent(bc, &mvc[1]) * 2; in decode_split_mv() 249 blockmv.as_mv.col += best_mv.as_mv.col; in decode_split_mv() 416 read_mv(bc, &mbmi_mv->as_mv, (const MV_CONTEXT *)mvc); in read_mb_modes_mv() 417 mbmi_mv->as_mv.row += near_mvs[near_index].as_mv.row; in read_mb_modes_mv() 418 mbmi_mv->as_mv.col += near_mvs[near_index].as_mv.col; in read_mb_modes_mv()
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_pickmode.c | 135 base_mv->as_mv.row = (candidate->mv[0].as_mv.row * 2); in mv_refs_rt() 136 base_mv->as_mv.col = (candidate->mv[0].as_mv.col * 2); in mv_refs_rt() 137 clamp_mv_ref(&base_mv->as_mv, xd); in mv_refs_rt() 149 clamp_mv_ref(&mv_ref_list[i].as_mv, xd); in mv_refs_rt() 165 const MV ref_mv = x->mbmi_ext->ref_mvs[ref][0].as_mv; in combined_motion_search() 195 mvp_full = x->mbmi_ext->ref_mvs[ref][x->mv_best_ref_index[ref]].as_mv; in combined_motion_search() 205 center_mv = tmp_mv->as_mv; in combined_motion_search() 208 tmp_mv->as_mv.row = x->sb_mvrow_part >> 3; in combined_motion_search() 209 tmp_mv->as_mv.col = x->sb_mvcol_part >> 3; in combined_motion_search() 213 cond_cost_list(cpi, cost_list), ¢er_mv, &tmp_mv->as_mv, INT_MAX, 0); in combined_motion_search() [all …]
|
D | vp9_rdopt.c | 1446 thismvcost += vp9_mv_bit_cost(&this_mv[0].as_mv, &best_ref_mv[0]->as_mv, in set_and_cost_bmi_mvs() 1450 thismvcost += vp9_mv_bit_cost(&this_mv[1].as_mv, &best_ref_mv[1]->as_mv, in set_and_cost_bmi_mvs() 1467 mi->bmi[i].as_mv[0].as_int = this_mv[0].as_int; in set_and_cost_bmi_mvs() 1468 if (is_compound) mi->bmi[i].as_mv[1].as_int = this_mv[1].as_int; in set_and_cost_bmi_mvs() 1525 pd->dst.stride, &mi->bmi[i].as_mv[ref].as_mv, in encode_inter_mb_segment() 1531 pre, y_stride, dst, pd->dst.stride, &mi->bmi[i].as_mv[ref].as_mv, in encode_inter_mb_segment() 1537 pre, y_stride, dst, pd->dst.stride, &mi->bmi[i].as_mv[ref].as_mv, in encode_inter_mb_segment() 1781 second_pred_alloc_16, pw, &frame_mv[refs[!id]].as_mv, &sf, pw, ph, 0, in joint_motion_search() 1786 second_pred, pw, &frame_mv[refs[!id]].as_mv, in joint_motion_search() 1792 second_pred, pw, &frame_mv[refs[!id]].as_mv, &sf, in joint_motion_search() [all …]
|
/external/libaom/libaom/av1/decoder/ |
D | decodemv.c | 655 read_mv(r, &mv->as_mv, &ref_mv->as_mv, &ec_ctx->ndvc, MV_SUBPEL_NONE); in assign_dv() 657 assert((mv->as_mv.col & 7) == 0); in assign_dv() 658 assert((mv->as_mv.row & 7) == 0); in assign_dv() 659 mv->as_mv.col = (mv->as_mv.col >> 3) * 8; in assign_dv() 660 mv->as_mv.row = (mv->as_mv.row >> 3) * 8; in assign_dv() 661 int valid = is_mv_valid(&mv->as_mv) && in assign_dv() 662 av1_is_dv_valid(mv->as_mv, cm, xd, mi_row, mi_col, bsize, in assign_dv() 694 int valid_dv = (dv_ref.as_mv.col & 7) == 0 && (dv_ref.as_mv.row & 7) == 0; in read_intrabc_info() 695 dv_ref.as_mv.col = (dv_ref.as_mv.col >> 3) * 8; in read_intrabc_info() 696 dv_ref.as_mv.row = (dv_ref.as_mv.row >> 3) * 8; in read_intrabc_info() [all …]
|
D | inspection.c | 76 mi->mv[0].row = mbmi->mv[0].as_mv.row; in ifd_inspect() 77 mi->mv[0].col = mbmi->mv[0].as_mv.col; in ifd_inspect() 78 mi->mv[1].row = mbmi->mv[1].as_mv.row; in ifd_inspect() 79 mi->mv[1].col = mbmi->mv[1].as_mv.col; in ifd_inspect()
|
/external/libvpx/libvpx/vp9/decoder/ |
D | vp9_decodemv.c | 414 read_mv(r, &mv[i].as_mv, &ref_mv[i].as_mv, &cm->fc->nmvc, mv_counts, in assign_mv() 416 ret = ret && is_mv_valid(&mv[i].as_mv); in assign_mv() 453 lower_mv_precision(&mvlist[i].as_mv, allow_hp); in dec_find_best_ref_mvs() 586 mv.as_mv.row *= -1; in dec_find_mv_refs() 587 mv.as_mv.col *= -1; in dec_find_mv_refs() 598 mv.as_mv.row *= -1; in dec_find_mv_refs() 599 mv.as_mv.col *= -1; in dec_find_mv_refs() 613 for (i = 0; i < refmv_count; ++i) clamp_mv_ref(&mv_ref_list[i].as_mv, xd); in dec_find_mv_refs() 640 best_sub8x8->as_int = bmi[0].as_mv[ref].as_int; in append_sub8x8_mvs_for_idx() 644 if (bmi[0].as_mv[ref].as_int != mv_list[n].as_int) { in append_sub8x8_mvs_for_idx() [all …]
|
/external/libavc/encoder/ |
D | ih264e.h | 671 ih264e_mv_t as_mv[1]; member 689 ih264e_mv_t as_mv[1]; member 705 ih264e_mv_t as_mv[4]; member 726 ih264e_mv_t as_mv[16]; member
|