/external/libvpx/libvpx/vp8/decoder/ |
D | error_concealment.c | 53 pbi->overlaps = vpx_calloc(pbi->common.mb_rows * pbi->common.mb_cols, in vp8_alloc_overlap_lists() 171 int mb_rows, int mb_cols, in vp8_calculate_overlaps() argument 190 if (new_row >= ((16*mb_rows) << 3) || new_col >= ((16*mb_cols) << 3)) in vp8_calculate_overlaps() 212 end_col = MIN(mb_cols - overlap_mb_col, 2); in vp8_calculate_overlaps() 230 mb_overlap = overlap_ul + (overlap_mb_row + rel_row) * mb_cols + in vp8_calculate_overlaps() 328 int mb_rows, int mb_cols) in calc_prev_mb_overlaps() argument 337 overlaps, mb_rows, mb_cols, in calc_prev_mb_overlaps() 349 int mb_rows, int mb_cols, in estimate_missing_mvs() argument 353 memset(overlaps, 0, sizeof(MB_OVERLAP) * mb_rows * mb_cols); in estimate_missing_mvs() 357 for (mb_col = 0; mb_col < mb_cols; ++mb_col) in estimate_missing_mvs() [all …]
|
D | error_concealment.h | 37 int mb_rows, int mb_cols,
|
D | decodeframe.c | 598 for (mb_col = 0; mb_col < pc->mb_cols; mb_col++) in decode_mb_rows() 605 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; in decode_mb_rows() 627 pc->mb_rows, pc->mb_cols, in decode_mb_rows() 709 lf_mic += pc->mb_cols; in decode_mb_rows() 1336 pbi->mvs_corrupt_from_mb < (unsigned int)pc->mb_cols * pc->mb_rows) in vp8_decode_frame() 1344 memset(pc->above_context, 0, sizeof(ENTROPY_CONTEXT_PLANES) * pc->mb_cols); in vp8_decode_frame()
|
D | threading.c | 305 const int first_row_no_sync_above = pc->mb_cols + nsync; in mt_decode_mb_rows() 407 for (mb_col = 0; mb_col < pc->mb_cols; mb_col++) in mt_decode_mb_rows() 425 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; in mt_decode_mb_rows() 451 pc->mb_rows, pc->mb_cols, in mt_decode_mb_rows() 508 if(mb_col != pc->mb_cols-1) in mt_decode_mb_rows()
|
D | decodemv.c | 630 mb_to_right_edge_start = ((pbi->common.mb_cols - 1) * 16) << 3; in vp8_decode_mode_mvs() 639 while (++mb_col < pbi->common.mb_cols) in vp8_decode_mode_mvs() 642 int mb_num = mb_row * pbi->common.mb_cols + mb_col; in vp8_decode_mode_mvs()
|
D | onyxd_if.c | 377 for (col = 0; col < pbi->common.mb_cols; ++col) in vp8dx_receive_compressed_data() 445 for (mb_col = 0; mb_col < oci->mb_cols; mb_col++,mi++) in vp8dx_references_buffer()
|
/external/libvpx/libvpx/test/ |
D | set_roi.cc | 54 cpi.common.mb_cols = 320 >> 4; in TEST() 55 const int mbs = (cpi.common.mb_rows * cpi.common.mb_cols); in TEST() 70 cpi.common.mb_cols, delta_q, delta_lf, in TEST() 138 cpi.common.mb_cols, rand_deltas, in TEST() 148 cpi.common.mb_cols, delta_q, in TEST() 160 cpi.common.mb_cols, delta_q, in TEST() 167 cpi.common.mb_cols, delta_q, in TEST() 172 cpi.common.mb_cols - 1, delta_q, in TEST()
|
/external/libvpx/libvpx/vp8/common/ |
D | alloccommon.c | 84 oci->mb_cols = width >> 4; in vp8_alloc_frame_buffers() 85 oci->MBs = oci->mb_rows * oci->mb_cols; in vp8_alloc_frame_buffers() 86 oci->mode_info_stride = oci->mb_cols + 1; in vp8_alloc_frame_buffers() 87 oci->mip = vpx_calloc((oci->mb_cols + 1) * (oci->mb_rows + 1), sizeof(MODE_INFO)); in vp8_alloc_frame_buffers() 97 oci->above_context = vpx_calloc(sizeof(ENTROPY_CONTEXT_PLANES) * oci->mb_cols, 1); in vp8_alloc_frame_buffers() 115 oci->pp_limits_buffer = vpx_memalign(16, 24 * ((oci->mb_cols + 1) & ~1)); in vp8_alloc_frame_buffers()
|
D | vp8_loopfilter.c | 208 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_loop_filter_row_normal() 265 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_loop_filter_row_simple() 316 int mb_cols = cm->mb_cols; in vp8_loop_filter_frame() local 340 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_loop_filter_frame() 396 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_loop_filter_frame() 482 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_loop_filter_frame_yonly() 565 int mb_cols = post->y_width >> 4; in vp8_loop_filter_partial_frame() local 592 mode_info_context = cm->mi + (post->y_height >> 5) * (mb_cols + 1); in vp8_loop_filter_partial_frame() 597 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_loop_filter_partial_frame()
|
D | mfqe.c | 305 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_multiframe_quality_enhance() 377 y_ptr += show->y_stride * 16 - 16 * cm->mb_cols; in vp8_multiframe_quality_enhance() 378 u_ptr += show->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance() 379 v_ptr += show->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance() 380 yd_ptr += dest->y_stride * 16 - 16 * cm->mb_cols; in vp8_multiframe_quality_enhance() 381 ud_ptr += dest->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance() 382 vd_ptr += dest->uv_stride * 8 - 8 * cm->mb_cols; in vp8_multiframe_quality_enhance()
|
D | postproc.c | 339 unsigned char *uvlimits = cm->pp_limits_buffer + 16 * cm->mb_cols; in vp8_deblock() 349 for (mbc = 0; mbc < cm->mb_cols; mbc++) in vp8_deblock() 400 int mb_cols = cm->mb_cols; in vp8_de_noise() local 406 memset(limits, (unsigned char)ppl, 16 * mb_cols); in vp8_de_noise() 849 oci->mb_cols, oci->mb_rows); in vp8_post_proc_frame() 859 int mb_cols = post->y_width >> 4; in vp8_post_proc_frame() local 868 for (j = 0; j < mb_cols; j++) in vp8_post_proc_frame() 891 int mb_cols = post->y_width >> 4; in vp8_post_proc_frame() local 900 for (j = 0; j < mb_cols; j++) in vp8_post_proc_frame()
|
D | onyxc_int.h | 104 int mb_cols; member
|
/external/libvpx/libvpx/vp8/encoder/ |
D | lookahead.c | 114 int mb_cols = (src->y_width + 15) >> 4; in vp8_lookahead_push() local 135 for (; col < mb_cols; ++col) in vp8_lookahead_push() 142 if (col == mb_cols) in vp8_lookahead_push() 148 for (; active_end < mb_cols; ++active_end) in vp8_lookahead_push() 164 active_map += mb_cols; in vp8_lookahead_push()
|
D | segmentation.c | 26 memset(cpi->gf_active_flags, 1, (cm->mb_rows * cm->mb_cols)); in vp8_update_gf_useage_maps() 27 cpi->gf_active_count = cm->mb_rows * cm->mb_cols; in vp8_update_gf_useage_maps() 35 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in vp8_update_gf_useage_maps()
|
D | mr_dissim.c | 93 for (mb_col = 0; mb_col < cm->mb_cols; mb_col ++) in vp8_cal_dissimilarity() 123 if(mb_col < (cm->mb_cols-1)) in vp8_cal_dissimilarity() 139 if(mb_col < (cm->mb_cols-1) in vp8_cal_dissimilarity() 152 if(mb_col < (cm->mb_cols-1)) in vp8_cal_dissimilarity() 168 if(mb_col < (cm->mb_cols-1) in vp8_cal_dissimilarity()
|
D | ethreading.c | 87 int map_index = (mb_row * cm->mb_cols); in thread_encoding_proc() 94 tp = cpi->tok + (mb_row * (cm->mb_cols * 16 * 24)); in thread_encoding_proc() 114 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in thread_encoding_proc() 136 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 16) << 3; in thread_encoding_proc() 144 x->mv_col_max = ((cm->mb_cols - 1 - mb_col) * 16) + (VP8BORDERINPIXELS - 16); in thread_encoding_proc() 305 … x->src.y_buffer += 16 * x->src.y_stride * (cpi->encoding_thread_count + 1) - 16 * cm->mb_cols; in thread_encoding_proc() 306 … x->src.u_buffer += 8 * x->src.uv_stride * (cpi->encoding_thread_count + 1) - 8 * cm->mb_cols; in thread_encoding_proc() 307 … x->src.v_buffer += 8 * x->src.uv_stride * (cpi->encoding_thread_count + 1) - 8 * cm->mb_cols; in thread_encoding_proc() 311 x->gf_active_ptr += cm->mb_cols * cpi->encoding_thread_count; in thread_encoding_proc() 532 if(th_count > ((cm->mb_cols / cpi->mt_sync_range) - 1)) in vp8cx_create_encoder_threads() [all …]
|
D | encodeframe.c | 227 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in calc_activity_index() 289 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in build_activity_map() 317 x->src.y_buffer += 16 * x->src.y_stride - 16 * cm->mb_cols; in build_activity_map() 378 int map_index = (mb_row * cpi->common.mb_cols); in encode_mb_row() 388 const int rightmost_col = cm->mb_cols + nsync; in encode_mb_row() 432 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) in encode_mb_row() 443 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 16) << 3; in encode_mb_row() 449 x->mv_col_max = ((cm->mb_cols - 1 - mb_col) * 16) in encode_mb_row() 669 memset(cm->above_context, 0, sizeof(ENTROPY_CONTEXT_PLANES) * cm->mb_cols); in init_encode_frame_mb_context() 846 tp = cpi->tok + mb_row * (cm->mb_cols * 16 * 24); in vp8_encode_frame() [all …]
|
D | temporal_filter.c | 237 int mb_cols = cpi->common.mb_cols; in vp8_temporal_filter_iterate_c() local 272 for (mb_col = 0; mb_col < mb_cols; mb_col++) in vp8_temporal_filter_iterate_c() 282 cpi->mb.mv_col_max = ((cpi->common.mb_cols - 1 - mb_col) * 16) in vp8_temporal_filter_iterate_c() 420 mb_y_offset += 16*(f->y_stride-mb_cols); in vp8_temporal_filter_iterate_c() 421 mb_uv_offset += 8*(f->uv_stride-mb_cols); in vp8_temporal_filter_iterate_c()
|
D | onyx_if.c | 493 memcpy(cpi->segmentation_map, segmentation_map, (cpi->common.mb_rows * cpi->common.mb_cols)); in set_segmentation_map() 525 CHECK_MEM_ERROR(seg_map, vpx_calloc(cpi->common.mb_rows * cpi->common.mb_cols, 1)); in segmentation_test_function() 561 int mbs_in_frame = cpi->common.mb_rows * cpi->common.mb_cols; in cyclic_background_refresh() 574 (cpi->common.mb_rows * cpi->common.mb_cols) / 10; in cyclic_background_refresh() 581 (cpi->common.mb_rows * cpi->common.mb_cols) / 20; in cyclic_background_refresh() 1190 cpi->mb.pip = vpx_calloc((cpi->common.mb_cols + 1) * in vp8_alloc_partition_data() 1240 unsigned int tokens = cm->mb_rows * cm->mb_cols * 24 * 16; in vp8_alloc_compressor_data() 1253 cm->mb_rows * cm->mb_cols)); in vp8_alloc_compressor_data() 1254 cpi->gf_active_count = cm->mb_rows * cm->mb_cols; in vp8_alloc_compressor_data() 1259 cm->mb_rows * cm->mb_cols)); in vp8_alloc_compressor_data() [all …]
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_lookahead.c | 102 int mb_cols = (src->y_width + 15) >> 4; in vp9_lookahead_push() local 141 for (; col < mb_cols; ++col) { in vp9_lookahead_push() 147 if (col == mb_cols) in vp9_lookahead_push() 153 for (; active_end < mb_cols; ++active_end) { in vp9_lookahead_push() 168 active_map += mb_cols; in vp9_lookahead_push()
|
D | vp9_mbgraph.c | 267 x->mv_col_max = (cm->mb_cols - 1) * 8 + BORDER_MV_PIXELS_B16; in update_mbgraph_frame_stats() 270 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { in update_mbgraph_frame_stats() 294 offset += cm->mb_cols; in update_mbgraph_frame_stats() 309 vpx_calloc(cm->mb_rows * cm->mb_cols * sizeof(*arf_not_zz), in separate_arf_mbs() 321 offset += cm->mb_cols, mb_row++) { in separate_arf_mbs() 322 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { in separate_arf_mbs() 345 if (arf_not_zz[mi_row / 2 * cm->mb_cols + mi_col / 2]) { in separate_arf_mbs() 396 cm->mb_rows * cm->mb_cols * sizeof(*cpi->mbgraph_stats[i].mb_stats)); in vp9_update_mbgraph_stats()
|
D | vp9_temporal_filter.c | 280 int mb_cols = (frames[alt_ref_index]->y_crop_width + 15) >> 4; in temporal_filter_iterate_c() local 329 for (mb_col = 0; mb_col < mb_cols; mb_col++) { in temporal_filter_iterate_c() 337 cpi->td.mb.mv_col_max = ((mb_cols - 1 - mb_col) * 16) in temporal_filter_iterate_c() 581 mb_y_offset += 16 * (f->y_stride - mb_cols); in temporal_filter_iterate_c() 582 mb_uv_offset += mb_uv_height * f->uv_stride - mb_uv_width * mb_cols; in temporal_filter_iterate_c()
|
D | vp9_encoder.h | 584 static INLINE int get_token_alloc(int mb_rows, int mb_cols) { in get_token_alloc() argument 590 return mb_rows * mb_cols * (16 * 16 * 3 + 4); in get_token_alloc()
|
D | vp9_firstpass.c | 648 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { in vp9_first_pass() 656 const int mb_index = mb_row * cm->mb_cols + mb_col; in vp9_first_pass() 752 x->mv_col_max = ((cm->mb_cols - 1 - mb_col) * 16) + BORDER_MV_PIXELS_B16; in vp9_first_pass() 981 if (mb_col < cm->mb_cols / 2) { in vp9_first_pass() 986 } else if (mb_col > cm->mb_cols / 2) { in vp9_first_pass() 1009 x->plane[0].src.buf += 16 * x->plane[0].src.stride - 16 * cm->mb_cols; in vp9_first_pass() 1011 uv_mb_height * cm->mb_cols; in vp9_first_pass() 1013 uv_mb_height * cm->mb_cols; in vp9_first_pass() 1027 MAX(0, intra_skip_count - (image_data_start_row * cm->mb_cols * 2)); in vp9_first_pass()
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_alloccommon.c | 46 cm->mb_cols = (cm->mi_cols + 1) >> 1; in vp9_set_mb_mi() 48 cm->MBs = cm->mb_rows * cm->mb_cols; in vp9_set_mb_mi()
|