/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/ |
D | mcomp.c | 35 …[0][(mv->as_mv.row - ref->as_mv.row) >> 1] + mvcost[1][(mv->as_mv.col - ref->as_mv.col) >> 1]) * W… in vp8_mv_bit_cost() 43 mvcost[1][(mv->as_mv.col - ref->as_mv.col) >> 1]) in mv_err_cost() 54 mvsadcost[1][(mv->as_mv.col - ref->as_mv.col)]) in mvsad_err_cost() 67 x->ss[search_site_count].mv.col = 0; in vp8_init_dsmotion_compensation() 76 x->ss[search_site_count].mv.col = 0; in vp8_init_dsmotion_compensation() 82 x->ss[search_site_count].mv.col = 0; in vp8_init_dsmotion_compensation() 88 x->ss[search_site_count].mv.col = -Len; in vp8_init_dsmotion_compensation() 94 x->ss[search_site_count].mv.col = Len; in vp8_init_dsmotion_compensation() 114 x->ss[search_site_count].mv.col = 0; in vp8_init3smotion_compensation() 123 x->ss[search_site_count].mv.col = 0; in vp8_init3smotion_compensation() [all …]
|
D | lookahead.c | 112 int row, col, active_end; in vp8_lookahead_push() local 130 col = 0; in vp8_lookahead_push() 135 for (; col < mb_cols; ++col) in vp8_lookahead_push() 137 if (active_map[col]) in vp8_lookahead_push() 142 if (col == mb_cols) in vp8_lookahead_push() 146 active_end = col; in vp8_lookahead_push() 157 col << 4, 16, in vp8_lookahead_push() 158 (active_end - col) << 4); in vp8_lookahead_push() 161 col = active_end; in vp8_lookahead_push()
|
D | pickinter.c | 60 bestmv->as_mv.col <<= 3; in vp8_skip_fractional_mv_step() 78 int xoffset = this_mv.as_mv.col & 7; in vp8_get_inter_mbpred_error() 81 in_what += (this_mv.as_mv.row >> 3) * pre_stride + (this_mv.as_mv.col >> 3); in vp8_get_inter_mbpred_error() 403 x->MVcount[1][mv_max+((xd->mode_info_context->mbmi.mv.as_mv.col - in update_mvcount() 404 best_ref_mv->as_mv.col) >> 1)]++; in update_mvcount() 453 (*parent_ref_mv).as_mv.col = store_mode_info[parent_mb_index].mv.as_mv.col in get_lower_res_motion_info() 541 if( abs(mv_l.as_mv.row) < 8 && abs(mv_l.as_mv.col) < 8) in calculate_zeromv_rd_adjustment() 549 if( abs(mv_al.as_mv.row) < 8 && abs(mv_al.as_mv.col) < 8) in calculate_zeromv_rd_adjustment() 557 if( abs(mv_a.as_mv.row) < 8 && abs(mv_a.as_mv.col) < 8) in calculate_zeromv_rd_adjustment() 866 int col_min = ((best_ref_mv.as_mv.col+7)>>3) - MAX_FULL_PEL_VAL; in vp8_pick_inter_mode() [all …]
|
D | mr_dissim.c | 35 mvy[cnt] = x->mbmi.mv.as_mv.col; \ 43 mvy[cnt] = x->mbmi.mv.as_mv.col; \ 196 mmvy = MAX(abs(min_mvy - here->mbmi.mv.as_mv.col), in vp8_cal_dissimilarity() 197 abs(max_mvy - here->mbmi.mv.as_mv.col)); in vp8_cal_dissimilarity()
|
D | rdopt.c | 480 int mv_col = x->e_mbd.mode_info_context->mbmi.mv.as_mv.col; in VP8_UVSSE() 973 const int row = i >> 2, col = i & 3; in labels2mode() local 980 if (col && labelings[i] == labelings[i-1]) in labels2mode() 995 this_mv->as_int = col ? d[-1].bmi.mv.as_int : (uint32_t)left_block_mv(mic, i); in labels2mode() 1011 left_mv.as_int = col ? d[-1].bmi.mv.as_int : in labels2mode() 1230 mvp_full.as_mv.col = bsi->mvp.as_mv.col >>3; in rd_check_segment() 1312 …((mode_mv[this_mode].as_mv.col >> 3) < x->mv_col_min) || ((mode_mv[this_mode].as_mv.col >> 3) > x-… in rd_check_segment() 1425 int col_min = ((best_ref_mv->as_mv.col+7)>>3) - MAX_FULL_PEL_VAL; in vp8_rd_pick_best_mbsegmentation() 1427 int col_max = (best_ref_mv->as_mv.col>>3) + MAX_FULL_PEL_VAL; in vp8_rd_pick_best_mbsegmentation() 1454 …v.row - bsi.sv_mvp[2].as_mv.row))>>3, (abs(bsi.sv_mvp[0].as_mv.col - bsi.sv_mvp[2].as_mv.col))>>3); in vp8_rd_pick_best_mbsegmentation() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/decoder/ |
D | error_concealment.c | 29 int col; member 139 int row, col; in calculate_overlaps_mb() local 159 for (col = 0; col < end_col; ++col) in calculate_overlaps_mb() 165 (((first_blk_col + col) * in calculate_overlaps_mb() 167 assign_overlap(b_ol_ul[row * 4 + col].overlaps, bmi, overlap); in calculate_overlaps_mb() 178 int row, col, rel_row, rel_col; in vp8_calculate_overlaps() local 186 col = (4 * b_col) << 3; /* Q3 */ in vp8_calculate_overlaps() 190 new_col = col - bmi->mv.as_mv.col; in vp8_calculate_overlaps() 261 col_acc += overlaps[i].overlap * overlaps[i].bmi->mv.as_mv.col; in estimate_mv() 268 bmi->mv.as_mv.col = col_acc / overlap_sum; in estimate_mv() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/ |
D | vp9_mcomp.c | 28 return &buf->buf[mv->row * buf->stride + mv->col]; in get_buf_from_mv() 32 int col_min = (mv->col >> 3) - MAX_FULL_PEL_VAL + (mv->col & 7 ? 1 : 0); in vp9_set_mv_search_range() 34 int col_max = (mv->col >> 3) + MAX_FULL_PEL_VAL; in vp9_set_mv_search_range() 71 comp_cost[0][mv->row] + comp_cost[1][mv->col]; in mv_cost() 77 mv->col - ref->col }; in vp9_mv_bit_cost() 86 mv->col - ref->col }; in mv_err_cost() 98 mv->col - ref->col }; in mvsad_err_cost() 108 x->ss[0].mv.col = x->ss[0].mv.row = 0; in vp9_init_dsmotion_compensation() 118 ss->offset = ss->mv.row * stride + ss->mv.col; in vp9_init_dsmotion_compensation() 129 x->ss[0].mv.col = x->ss[0].mv.row = 0; in vp9_init3smotion_compensation() [all …]
|
D | vp9_lookahead.c | 97 int row, col, active_end; in vp9_lookahead_push() local 117 col = 0; in vp9_lookahead_push() 121 for (; col < mb_cols; ++col) { in vp9_lookahead_push() 122 if (active_map[col]) in vp9_lookahead_push() 127 if (col == mb_cols) in vp9_lookahead_push() 131 active_end = col; in vp9_lookahead_push() 141 col << 4, 16, in vp9_lookahead_push() 142 (active_end - col) << 4); in vp9_lookahead_push() 145 col = active_end; in vp9_lookahead_push()
|
D | vp9_encodemv.c | 206 mv->col - ref->col}; in vp9_encode_mv() 215 encode_mv_component(w, diff.col, &mvctx->comps[1], usehp); in vp9_encode_mv() 220 unsigned int maxv = MAX(abs(mv->row), abs(mv->col)) >> 3; in vp9_encode_mv() 239 mvs[i].as_mv.col - ref->col}; in inc_mvs()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ |
D | findnearmv.h | 31 mvp->as_mv.col *= -1; in mv_bias() 39 if (mv->as_mv.col < (xd->mb_to_left_edge - LEFT_TOP_MARGIN)) in vp8_clamp_mv2() 40 mv->as_mv.col = xd->mb_to_left_edge - LEFT_TOP_MARGIN; in vp8_clamp_mv2() 41 else if (mv->as_mv.col > xd->mb_to_right_edge + RIGHT_BOTTOM_MARGIN) in vp8_clamp_mv2() 42 mv->as_mv.col = xd->mb_to_right_edge + RIGHT_BOTTOM_MARGIN; in vp8_clamp_mv2() 53 mv->as_mv.col = (mv->as_mv.col < mb_to_left_edge) ? in vp8_clamp_mv() 54 mb_to_left_edge : mv->as_mv.col; in vp8_clamp_mv() 55 mv->as_mv.col = (mv->as_mv.col > mb_to_right_edge) ? in vp8_clamp_mv() 56 mb_to_right_edge : mv->as_mv.col; in vp8_clamp_mv() 67 need_to_clamp = (mv->as_mv.col < mb_to_left_edge); in vp8_check_mv_bounds() [all …]
|
D | reconinter.c | 131 … ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + (d->bmi.mv.as_mv.col >> 3); in vp8_build_inter_predictors_b() 133 if (d->bmi.mv.as_mv.row & 7 || d->bmi.mv.as_mv.col & 7) in vp8_build_inter_predictors_b() 135 sppf(ptr, pre_stride, d->bmi.mv.as_mv.col & 7, d->bmi.mv.as_mv.row & 7, pred_ptr, pitch); in vp8_build_inter_predictors_b() 154 … ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + (d->bmi.mv.as_mv.col >> 3); in build_inter_predictors4b() 156 if (d->bmi.mv.as_mv.row & 7 || d->bmi.mv.as_mv.col & 7) in build_inter_predictors4b() 158 …x->subpixel_predict8x8(ptr, pre_stride, d->bmi.mv.as_mv.col & 7, d->bmi.mv.as_mv.row & 7, dst, dst… in build_inter_predictors4b() 169 … ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + (d->bmi.mv.as_mv.col >> 3); in build_inter_predictors2b() 171 if (d->bmi.mv.as_mv.row & 7 || d->bmi.mv.as_mv.col & 7) in build_inter_predictors2b() 173 …x->subpixel_predict8x4(ptr, pre_stride, d->bmi.mv.as_mv.col & 7, d->bmi.mv.as_mv.row & 7, dst, dst… in build_inter_predictors2b() 185 … ptr = base_pre + d->offset + (d->bmi.mv.as_mv.row >> 3) * pre_stride + (d->bmi.mv.as_mv.col >> 3); in build_inter_predictors_b() [all …]
|
D | postproc.c | 139 int col; in vp8_post_proc_down_and_across_mb_row_c() local 149 for (col = 0; col < cols; col++) in vp8_post_proc_down_and_across_mb_row_c() 151 unsigned char p_above2 = p_src[col - 2 * src_pixels_per_line]; in vp8_post_proc_down_and_across_mb_row_c() 152 unsigned char p_above1 = p_src[col - src_pixels_per_line]; in vp8_post_proc_down_and_across_mb_row_c() 153 unsigned char p_below1 = p_src[col + src_pixels_per_line]; in vp8_post_proc_down_and_across_mb_row_c() 154 unsigned char p_below2 = p_src[col + 2 * src_pixels_per_line]; in vp8_post_proc_down_and_across_mb_row_c() 156 v = p_src[col]; in vp8_post_proc_down_and_across_mb_row_c() 158 if ((abs(v - p_above2) < f[col]) && (abs(v - p_above1) < f[col]) in vp8_post_proc_down_and_across_mb_row_c() 159 && (abs(v - p_below1) < f[col]) && (abs(v - p_below2) < f[col])) in vp8_post_proc_down_and_across_mb_row_c() 168 p_dst[col] = v; in vp8_post_proc_down_and_across_mb_row_c() [all …]
|
/hardware/intel/common/libmix/mix_vbp/viddec_fw/fw/codecs/vc1/parser/ |
D | vc1parse_bitplane.c | 157 int32_t row[2], col[2]; in vc1_Norm2ModeDecode() local 169 col[0] = 0; /* i%width; */ in vc1_Norm2ModeDecode() 171 col[1] = 1; /* (i+1)%width; */ in vc1_Norm2ModeDecode() 183 col[0] = 1; /* i%width; */ in vc1_Norm2ModeDecode() 184 col[1] = 2; /* (i+1)%width; */ in vc1_Norm2ModeDecode() 189 col[0] = 0; /* i%width; */ in vc1_Norm2ModeDecode() 191 col[1] = 0; /* (i+1)%width; */ in vc1_Norm2ModeDecode() 209 put_bit(0, col[0],row[0], width, height, pBitplane->invert, in vc1_Norm2ModeDecode() 211 put_bit(0, col[1],row[1], width, height, pBitplane->invert, in vc1_Norm2ModeDecode() 219 put_bit(1, col[0],row[0], width, height, pBitplane->invert, in vc1_Norm2ModeDecode() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/ |
D | vp9_mv.h | 24 int16_t col; member 34 int32_t col; member 39 mv->col = clamp(mv->col, min_col, max_col); in clamp_mv()
|
D | vp9_mvref_common.c | 18 int col; member 143 mv.as_mv.col *= -1; in scale_mv() 184 mi_col + mi_pos->col < tile->mi_col_start || in is_inside() 186 mi_col + mi_pos->col >= tile->mi_col_end); in is_inside() 218 const MODE_INFO *const candidate_mi = xd->mi[mv_ref->col + mv_ref->row * in find_mv_refs_idx() 226 ADD_MV_REF_LIST(get_sub_block_mv(candidate_mi, 0, mv_ref->col, block)); in find_mv_refs_idx() 228 ADD_MV_REF_LIST(get_sub_block_mv(candidate_mi, 1, mv_ref->col, block)); in find_mv_refs_idx() 238 const MB_MODE_INFO *const candidate = &xd->mi[mv_ref->col + mv_ref->row * in find_mv_refs_idx() 264 const MB_MODE_INFO *const candidate = &xd->mi[mv_ref->col + mv_ref->row in find_mv_refs_idx() 300 if (mv->col & 1) in lower_mv_precision() [all …]
|
D | vp9_postproc.c | 143 int col; in vp9_post_proc_down_and_across_c() local 155 for (col = 0; col < cols; col++) { in vp9_post_proc_down_and_across_c() 157 int v = p_src[col]; in vp9_post_proc_down_and_across_c() 160 if (abs(v - p_src[col + i * pitch]) > flimit) in vp9_post_proc_down_and_across_c() 163 kernel += kernel5[2 + i] * p_src[col + i * pitch]; in vp9_post_proc_down_and_across_c() 168 p_dst[col] = v; in vp9_post_proc_down_and_across_c() 178 for (col = 0; col < cols; col++) { in vp9_post_proc_down_and_across_c() 180 v = p_src[col]; in vp9_post_proc_down_and_across_c() 182 d[col & 7] = v; in vp9_post_proc_down_and_across_c() 185 if (abs(v - p_src[col + i]) > flimit) in vp9_post_proc_down_and_across_c() [all …]
|
D | vp9_reconinter.c | 89 is_q4 ? src_mv->col : src_mv->col * 2 }; in vp9_build_inter_predictor() 91 const int subpel_x = mv.col & SUBPEL_MASK; in vp9_build_inter_predictor() 94 src += (mv.row >> SUBPEL_BITS) * src_stride + (mv.col >> SUBPEL_BITS); in vp9_build_inter_predictor() 109 round_mv_comp_q4(mi->bmi[0].as_mv[idx].as_mv.col + in mi_mv_pred_q4() 110 mi->bmi[1].as_mv[idx].as_mv.col + in mi_mv_pred_q4() 111 mi->bmi[2].as_mv[idx].as_mv.col + in mi_mv_pred_q4() 112 mi->bmi[3].as_mv[idx].as_mv.col) }; in mi_mv_pred_q4() 128 src_mv->col * (1 << (1 - ss_x)) in clamp_mv_to_umv_border_sb() 188 scaled_mv.col = mv_q4.col; in build_inter_predictors() 191 subpel_x = scaled_mv.col & SUBPEL_MASK; in build_inter_predictors() [all …]
|
D | vp9_tile_common.c | 24 void vp9_tile_init(TileInfo *tile, const VP9_COMMON *cm, int row, int col) { in vp9_tile_init() argument 27 tile->mi_col_start = get_tile_offset(col, cm->mi_cols, cm->log2_tile_cols); in vp9_tile_init() 28 tile->mi_col_end = get_tile_offset(col + 1, cm->mi_cols, cm->log2_tile_cols); in vp9_tile_init()
|
/hardware/ti/omap4xxx/camera/ |
D | NV12_resize.c | 36 mmUint16 row,col; in VT_resizeFrame_Video_opt2_lp() local 122 for (col=0; col < codx; col++) in VT_resizeFrame_Video_opt2_lp() 133 x = (mmUint16) ((mmUint32) (col*resizeFactorX) >> 9); in VT_resizeFrame_Video_opt2_lp() 134 xf = (mmUchar) ((mmUint32) ((col*resizeFactorX) >> 6) & 0x7); in VT_resizeFrame_Video_opt2_lp() 207 for (col=0; col < (((codx)>>1)); col++) in VT_resizeFrame_Video_opt2_lp() 220 x = (mmUint16) ((mmUint32) (col*resizeFactorX) >> 9); in VT_resizeFrame_Video_opt2_lp() 221 xf = (mmUchar) ((mmUint32) ((col*resizeFactorX) >> 6) & 0x7); in VT_resizeFrame_Video_opt2_lp()
|
/hardware/ti/omap4-aah/camera/ |
D | NV12_resize.cpp | 49 mmUint16 row,col; in VT_resizeFrame_Video_opt2_lp() local 124 for ( col = 0; col < codx; col++ ) { in VT_resizeFrame_Video_opt2_lp() 132 x = (mmUint16) ((mmUint32) (col*resizeFactorX) >> 9); in VT_resizeFrame_Video_opt2_lp() 133 xf = (mmUchar) ((mmUint32) ((col*resizeFactorX) >> 6) & 0x7); in VT_resizeFrame_Video_opt2_lp() 203 for ( col = 0; col < (((codx)>>1)); col++ ) { in VT_resizeFrame_Video_opt2_lp() 214 x = (mmUint16) ((mmUint32) (col*resizeFactorX) >> 9); in VT_resizeFrame_Video_opt2_lp() 215 xf = (mmUchar) ((mmUint32) ((col*resizeFactorX) >> 6) & 0x7); in VT_resizeFrame_Video_opt2_lp()
|
/hardware/ti/omap4-aah/ion/ |
D | ion_test.c | 67 size_t row, col; in _ion_tiler_map_test() local 70 for (col = 0; col < width; col++) { in _ion_tiler_map_test() 71 int i = (row * stride) + col; in _ion_tiler_map_test() 75 for (col = 0; col < width; col++) { in _ion_tiler_map_test() 76 int i = (row * stride) + col; in _ion_tiler_map_test()
|
D | ion_test_2.c | 121 size_t row, col; in _ion_tiler_map_test() local 124 for (col = 0; col < width; col++) { in _ion_tiler_map_test() 125 int i = (row * stride) + col; in _ion_tiler_map_test() 129 for (col = 0; col < width; col++) { in _ion_tiler_map_test() 130 int i = (row * stride) + col; in _ion_tiler_map_test()
|
/hardware/intel/common/libva/test/ |
D | loadsurface.h | 29 int row=0, col=0; in scale_2dimage() local 32 for (col=0; col<dst_imgw; col++) { in scale_2dimage() 33 …*(dst_img + row * dst_imgw + col) = *(src_img + (row * src_imgh/dst_imgh) * src_imgw + col * src_i… in scale_2dimage() 59 int row, col; in YUV_blend_with_pic() local 98 for (col=0; col<width; col++, q++) { in YUV_blend_with_pic() 127 for (col=0; col<width/2; col++, qU++, qV++) { in YUV_blend_with_pic()
|
/hardware/intel/common/libva/test/videoprocess/ |
D | videoprocess.cpp | 215 uint32_t row, col; in construct_nv12_mask_surface() local 241 for (col = 0; col < surface_image.width / 2; col++) { in construct_nv12_mask_surface() 242 u_dst[col * 2] = 128; in construct_nv12_mask_surface() 243 u_dst[col * 2 + 1] = 128; in construct_nv12_mask_surface() 264 uint32_t frame_size, i, row, col; in upload_yv12_frame_to_yuv_surface() local 329 for (col = 0; col < surface_image.width / 2; col++) { in upload_yv12_frame_to_yuv_surface() 330 u_dst[col * 2] = u_src[col]; in upload_yv12_frame_to_yuv_surface() 331 u_dst[col * 2 + 1] = v_src[col]; in upload_yv12_frame_to_yuv_surface() 366 uint32_t frame_size, row, col; in store_yuv_surface_to_yv12_frame() local 430 for (col = 0; col < surface_image.width /2; col++) { in store_yuv_surface_to_yv12_frame() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
D | vp9_idct_intrin_sse2.c | 3504 __m128i in[32], col[32]; in vp9_idct32x32_34_add_sse2() local 3560 col[0] = _mm_add_epi16(stp1_0, stp1_31); in vp9_idct32x32_34_add_sse2() local 3561 col[1] = _mm_add_epi16(stp1_1, stp1_30); in vp9_idct32x32_34_add_sse2() 3562 col[2] = _mm_add_epi16(stp1_2, stp1_29); in vp9_idct32x32_34_add_sse2() 3563 col[3] = _mm_add_epi16(stp1_3, stp1_28); in vp9_idct32x32_34_add_sse2() 3564 col[4] = _mm_add_epi16(stp1_4, stp1_27); in vp9_idct32x32_34_add_sse2() 3565 col[5] = _mm_add_epi16(stp1_5, stp1_26); in vp9_idct32x32_34_add_sse2() 3566 col[6] = _mm_add_epi16(stp1_6, stp1_25); in vp9_idct32x32_34_add_sse2() 3567 col[7] = _mm_add_epi16(stp1_7, stp1_24); in vp9_idct32x32_34_add_sse2() 3568 col[8] = _mm_add_epi16(stp1_8, stp1_23); in vp9_idct32x32_34_add_sse2() [all …]
|