/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/ |
D | vp9_speed_features.c | 45 SPEED_FEATURES *sf, in set_good_speed_feature() argument 48 sf->adaptive_rd_thresh = 1; in set_good_speed_feature() 49 sf->recode_loop = ((speed < 1) ? ALLOW_RECODE : ALLOW_RECODE_KFMAXBW); in set_good_speed_feature() 50 sf->allow_skip_recode = 1; in set_good_speed_feature() 53 sf->use_square_partition_only = !frame_is_intra_only(cm); in set_good_speed_feature() 54 sf->less_rectangular_check = 1; in set_good_speed_feature() 55 sf->tx_size_search_method = vp9_frame_is_boosted(cpi) in set_good_speed_feature() 59 sf->disable_split_mask = cm->show_frame ? in set_good_speed_feature() 62 sf->disable_split_mask = DISABLE_COMPOUND_SPLIT; in set_good_speed_feature() 63 sf->use_rd_breakout = 1; in set_good_speed_feature() [all …]
|
D | vp9_pickmode.c | 88 if (cpi->sf.search_method == FAST_DIAMOND) { in full_pixel_motion_search() 93 } else if (cpi->sf.search_method == FAST_HEX) { in full_pixel_motion_search() 98 } else if (cpi->sf.search_method == HEX) { in full_pixel_motion_search() 103 } else if (cpi->sf.search_method == SQUARE) { in full_pixel_motion_search() 108 } else if (cpi->sf.search_method == BIGDIA) { in full_pixel_motion_search() 114 int further_steps = (cpi->sf.max_step_search_steps - 1) - step_param; in full_pixel_motion_search() 161 cpi->sf.subpel_force_stop, in sub_pixel_motion_search() 162 cpi->sf.subpel_iters_per_step, in sub_pixel_motion_search() 238 x->skip_encode = cpi->sf.skip_encode_frame && x->q_index < QIDX_SKIP_THRESH; in vp9_pick_inter_mode() 291 if (cpi->sf.disable_inter_mode_mask[bsize] & in vp9_pick_inter_mode()
|
D | vp9_encodeframe.c | 487 const struct scale_factors *const sf = &cm->frame_refs[LAST_FRAME - 1].sf; in choose_partitioning() local 501 vp9_setup_pre_planes(xd, 0, yv12, mi_row, mi_col, sf); in choose_partitioning() 1627 cpi->sf.partition_search_type == SEARCH_PARTITION && in rd_use_partition() 1628 cpi->sf.adjust_partitioning_from_last_frame) { in rd_use_partition() 1758 && cpi->sf.adjust_partitioning_from_last_frame in rd_use_partition() 1759 && cpi->sf.partition_search_type == SEARCH_PARTITION in rd_use_partition() 1964 if (cpi->sf.auto_min_max_partition_size == RELAXED_NEIGHBORING_MIN_MAX) { in rd_auto_partition_range() 1979 if (cpi->sf.use_square_partition_only && in rd_auto_partition_range() 2051 if (cpi->sf.auto_min_max_partition_size) { in rd_pick_partition() 2052 partition_none_allowed &= (bsize <= cpi->sf.max_partition_size && in rd_pick_partition() [all …]
|
D | vp9_rdopt.c | 289 x->select_txfm_size = (cpi->sf.tx_size_search_method == USE_LARGESTALL && in vp9_initialize_rd_consts() 294 if (!cpi->sf.use_nonrd_pick_mode || cm->frame_type == KEY_FRAME) { in vp9_initialize_rd_consts() 302 if (!cpi->sf.use_nonrd_pick_mode || (cm->current_video_frame & 0x07) == 1 || in vp9_initialize_rd_consts() 788 mbmi->tx_size, cpi->sf.use_fast_coef_costing); in choose_largest_txfm_size() 933 cpi->sf.use_fast_coef_costing); in choose_txfm_size_from_modelrd() 962 if (cpi->sf.tx_size_search_method == USE_LARGESTALL) { in inter_super_block_yrd() 971 if (cpi->sf.tx_size_search_method == USE_LARGESTINTRA_MODELINTER) { in inter_super_block_yrd() 982 cpi->sf.use_fast_coef_costing); in inter_super_block_yrd() 1000 if (cpi->sf.tx_size_search_method != USE_FULL_RD) { in intra_super_block_yrd() 1012 cpi->sf.use_fast_coef_costing); in intra_super_block_yrd() [all …]
|
D | vp9_temporal_filter.c | 156 step_param = cpi->sf.reduce_first_step_size + ((cpi->speed > 5) ? 1 : 0); in temporal_filter_find_matching_mb_c() 158 step_param = cpi->sf.reduce_first_step_size + 2; in temporal_filter_find_matching_mb_c() 159 step_param = MIN(step_param, (cpi->sf.max_step_search_steps - 2)); in temporal_filter_find_matching_mb_c() 177 0, cpi->sf.subpel_iters_per_step, in temporal_filter_find_matching_mb_c() 388 struct scale_factors sf; in vp9_temporal_filter_prepare() local 447 vp9_setup_scale_factors_for_frame(&sf, in vp9_temporal_filter_prepare() 462 strength, &sf); in vp9_temporal_filter_prepare()
|
D | vp9_mbgraph.c | 40 int step_param = cpi->sf.reduce_first_step_size + in do_16x16_motion_iteration() 42 step_param = MIN(step_param, (cpi->sf.max_step_search_steps - 2)); in do_16x16_motion_iteration() 60 &v_fn_ptr, 0, cpi->sf.subpel_iters_per_step, NULL, NULL, &distortion, in do_16x16_motion_iteration()
|
D | vp9_onyx_if.c | 507 const SPEED_FEATURES *const sf = &cpi->sf; in set_rd_speed_thresholds_sub8x8() local 522 if (sf->disable_split_mask & (1 << i)) in set_rd_speed_thresholds_sub8x8() 657 if (cpi->sf.search_method == NSTEP) { in update_frame_size() 659 } else if (cpi->sf.search_method == DIAMOND) { in update_frame_size() 2021 } else if ((cpi->sf.recode_loop == ALLOW_RECODE) || in recode_loop_test() 2022 ((cpi->sf.recode_loop == ALLOW_RECODE_KFARFGF) && in recode_loop_test() 2112 vp9_pick_filter_level(cpi->Source, cpi, cpi->sf.lpf_pick); 2338 if (cpi->sf.recode_loop >= ALLOW_RECODE_KFARFGF) { 2341 if (!cpi->sf.use_nonrd_pick_mode) 2547 const SPEED_FEATURES *const sf = &cpi->sf; local [all …]
|
D | vp9_ratectrl.c | 982 if (cpi->sf.use_nonrd_pick_mode) { in vp9_rc_pick_q_and_bounds() 985 if (cpi->sf.force_frame_boost == 1) in vp9_rc_pick_q_and_bounds() 986 q -= cpi->sf.max_delta_qindex; in vp9_rc_pick_q_and_bounds() 1006 (cpi->sf.recode_tolerance * this_frame_target) / 100; in vp9_rc_compute_frame_size_bounds() 1083 cpi, (cpi->sf.recode_loop >= ALLOW_RECODE_KFARFGF || in vp9_rc_postencode_update()
|
D | vp9_aq_cyclicrefresh.c | 214 if (cpi->sf.use_nonrd_pick_mode) { in vp9_cyclic_refresh_setup()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/ |
D | vp9_scale.c | 15 static INLINE int scaled_x(int val, const struct scale_factors *sf) { in scaled_x() argument 16 return (int)((int64_t)val * sf->x_scale_fp >> REF_SCALE_SHIFT); in scaled_x() 19 static INLINE int scaled_y(int val, const struct scale_factors *sf) { in scaled_y() argument 20 return (int)((int64_t)val * sf->y_scale_fp >> REF_SCALE_SHIFT); in scaled_y() 23 static int unscaled_value(int val, const struct scale_factors *sf) { in unscaled_value() argument 24 (void) sf; in unscaled_value() 44 MV32 vp9_scale_mv(const MV *mv, int x, int y, const struct scale_factors *sf) { in vp9_scale_mv() argument 45 const int x_off_q4 = scaled_x(x << SUBPEL_BITS, sf) & SUBPEL_MASK; in vp9_scale_mv() 46 const int y_off_q4 = scaled_y(y << SUBPEL_BITS, sf) & SUBPEL_MASK; in vp9_scale_mv() 48 scaled_y(mv->row, sf) + y_off_q4, in vp9_scale_mv() [all …]
|
D | vp9_scale.h | 31 int (*scale_value_x)(int val, const struct scale_factors *sf); 32 int (*scale_value_y)(int val, const struct scale_factors *sf); 37 MV32 vp9_scale_mv(const MV *mv, int x, int y, const struct scale_factors *sf); 39 void vp9_setup_scale_factors_for_frame(struct scale_factors *sf, 43 static INLINE int vp9_is_valid_scale(const struct scale_factors *sf) { in vp9_is_valid_scale() argument 44 return sf->x_scale_fp != REF_INVALID_SCALE && in vp9_is_valid_scale() 45 sf->y_scale_fp != REF_INVALID_SCALE; in vp9_is_valid_scale() 48 static INLINE int vp9_is_scaled(const struct scale_factors *sf) { in vp9_is_scaled() argument 49 return sf->x_scale_fp != REF_NO_SCALE || in vp9_is_scaled() 50 sf->y_scale_fp != REF_NO_SCALE; in vp9_is_scaled()
|
D | vp9_reconinter.h | 36 const struct scale_factors *sf, 43 const struct scale_factors *sf) { in scaled_buffer_offset() argument 44 const int x = sf ? sf->scale_value_x(x_offset, sf) : x_offset; in scaled_buffer_offset() 45 const int y = sf ? sf->scale_value_y(y_offset, sf) : y_offset; in scaled_buffer_offset() 65 const struct scale_factors *sf);
|
D | vp9_reconinter.c | 70 const struct scale_factors *sf, in inter_predictor() argument 74 sf->predict[subpel_x != 0][subpel_y != 0][ref]( in inter_predictor() 82 const struct scale_factors *sf, in vp9_build_inter_predictor() argument 90 MV32 mv = vp9_scale_mv(&mv_q4, x, y, sf); in vp9_build_inter_predictor() 97 sf, w, h, ref, kernel, sf->x_step_q4, sf->y_step_q4); in vp9_build_inter_predictor() 153 const struct scale_factors *const sf = &xd->block_refs[ref]->sf; in build_inter_predictors() local 180 if (vp9_is_scaled(sf)) { in build_inter_predictors() 181 pre = pre_buf->buf + scaled_buffer_offset(x, y, pre_buf->stride, sf); in build_inter_predictors() 182 scaled_mv = vp9_scale_mv(&mv_q4, mi_x + x, mi_y + y, sf); in build_inter_predictors() 183 xs = sf->x_step_q4; in build_inter_predictors() [all …]
|
D | vp9_blockd.h | 201 struct scale_factors sf; member
|
/hardware/invensense/60xx/mlsdk/mllite/ |
D | mlSetGyroBias.c | 56 int sf; in inv_convert_bias() local 60 sf = 2000 * 131 / mldl_cfg->gyro_sens_trim; in inv_convert_bias() 62 sf = 2000; in inv_convert_bias() 75 biasTmp[i] = (long)(biasTmp[i] * 1.39882274201861f / sf); in inv_convert_bias() 110 int sf, i; in inv_set_gyro_bias_in_dps() local 116 sf = 2000 * 131 / mldl_cfg->gyro_sens_trim; in inv_set_gyro_bias_in_dps() 118 sf = 2000; in inv_set_gyro_bias_in_dps() 122 biasTmp = -bias[i] / sf; in inv_set_gyro_bias_in_dps()
|
D | mlarray.c | 2133 long sf = 0; in inv_set_gyro_bias() local 2139 sf = 2000 * 131 / mldl_cfg->gyro_sens_trim; in inv_set_gyro_bias() 2141 sf = 2000; in inv_set_gyro_bias() 2145 biasTmp = -inv_obj.gyro_bias[i] / sf; in inv_set_gyro_bias() 2284 long sf; in inv_set_gyro_temp_slope() local 2292 sf = -inv_obj.temp_slope[i] / 1118; in inv_set_gyro_temp_slope() 2293 if (sf > 127) { in inv_set_gyro_temp_slope() 2294 sf -= 256; in inv_set_gyro_temp_slope() 2296 regs[i] = (unsigned char)sf; in inv_set_gyro_temp_slope()
|
D | ml.c | 777 unsigned long sf; in inv_set_accel_calibration() local 871 sf = (1073741824L / inv_obj.accel_sens); in inv_set_accel_calibration() 873 sf = 0; in inv_set_accel_calibration() 875 regs[0] = (unsigned char)((sf >> 8) & 0xff); in inv_set_accel_calibration() 876 regs[1] = (unsigned char)(sf & 0xff); in inv_set_accel_calibration() 931 int sf = 0; in inv_set_gyro_calibration() local 1020 sf = (long)((long long)23832619764371LL / inv_obj.gyro_sens); in inv_set_gyro_calibration() 1022 sf = 0; in inv_set_gyro_calibration() 1025 result = inv_set_mpu_memory(KEY_D_0_24, 4, inv_int32_to_big8(sf, regs)); in inv_set_gyro_calibration()
|
D | mldl.c | 687 long sf; in inv_set_offset() local 689 sf = (2000L * 131) / mldlCfg.gyro_sens_trim; in inv_set_offset() 694 inv_obj.gyro_bias[ii] = -offset[ii] * sf; in inv_set_offset()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/ |
D | onyx_if.c | 751 SPEED_FEATURES *sf = &cpi->sf; in vp8_set_speed_features() local 756 int last_improved_quant = sf->improved_quant; in vp8_set_speed_features() 768 sf->RD = 1; in vp8_set_speed_features() 769 sf->search_method = NSTEP; in vp8_set_speed_features() 770 sf->improved_quant = 1; in vp8_set_speed_features() 771 sf->improved_dct = 1; in vp8_set_speed_features() 772 sf->auto_filter = 1; in vp8_set_speed_features() 773 sf->recode_loop = 1; in vp8_set_speed_features() 774 sf->quarter_pixel_search = 1; in vp8_set_speed_features() 775 sf->half_pixel_search = 1; in vp8_set_speed_features() [all …]
|
D | quantize.c | 448 invert_quant(cpi->sf.improved_quant, cpi->Y1quant[Q] + 0, in vp8cx_init_quantizer() 457 invert_quant(cpi->sf.improved_quant, cpi->Y2quant[Q] + 0, in vp8cx_init_quantizer() 466 invert_quant(cpi->sf.improved_quant, cpi->UVquant[Q] + 0, in vp8cx_init_quantizer() 476 invert_quant(cpi->sf.improved_quant, cpi->Y1quant[Q] + 1, in vp8cx_init_quantizer() 485 invert_quant(cpi->sf.improved_quant, cpi->Y2quant[Q] + 1, in vp8cx_init_quantizer() 494 invert_quant(cpi->sf.improved_quant, cpi->UVquant[Q] + 1, in vp8cx_init_quantizer()
|
D | temporal_filter.c | 181 step_param = cpi->sf.first_step + (cpi->Speed > 5); in vp8_temporal_filter_find_matching_mb_c() 185 step_param = cpi->sf.first_step + 2; in vp8_temporal_filter_find_matching_mb_c()
|
D | encodeframe.c | 1170 if (cpi->sf.RD && cpi->compressor_speed != 2) in vp8cx_encode_intra_macroblock() 1235 if (cpi->sf.RD) in vp8cx_encode_inter_macroblock() 1240 if(cpi->sf.use_fastquant_for_pick) in vp8cx_encode_inter_macroblock() 1253 if (cpi->sf.improved_quant) in vp8cx_encode_inter_macroblock() 1329 if (cpi->sf.improved_quant) in vp8cx_encode_inter_macroblock()
|
D | pickinter.c | 503 !(cpi->sf.half_pixel_search) || cpi->common.full_pixel==1) in evaluate_inter_mode() 599 int sf_improved_mv_pred = cpi->sf.improved_mv_pred; in vp8_pick_inter_mode() 881 step_param = cpi->sf.first_step + speed_adjust; in vp8_pick_inter_mode() 962 0: (cpi->sf.max_step_search_steps - 1 - step_param); in vp8_pick_inter_mode() 964 if (cpi->sf.search_method == HEX) in vp8_pick_inter_mode()
|
D | rdopt.c | 285 if (cpi->sf.thresh_mult[i] < INT_MAX) in vp8_initialize_rd_consts() 287 x->rd_threshes[i] = cpi->sf.thresh_mult[i] * q / 100; in vp8_initialize_rd_consts() 303 if (cpi->sf.thresh_mult[i] < (INT_MAX / q)) in vp8_initialize_rd_consts() 305 x->rd_threshes[i] = cpi->sf.thresh_mult[i] * q; in vp8_initialize_rd_consts() 1476 …if (cpi->sf.no_skip_block4x4_search || bsi.segment_num == BLOCK_8X8) /* || (sv_segment_rd8x8-bsi.… in vp8_rd_pick_best_mbsegmentation() 2239 int step_param = cpi->sf.first_step; in vp8_rd_pick_inter_mode() 2296 further_steps = (cpi->sf.max_step_search_steps - 1) - step_param; in vp8_rd_pick_inter_mode()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vpx/src/ |
D | svc_encodeframe.c | 179 static void sf_create_index(struct Superframe *sf) { in sf_create_index() argument 185 if (sf->count == 0 || sf->count >= 8) return; in sf_create_index() 188 marker |= sf->count - 1; in sf_create_index() 192 if (sf->magnitude < mask) break; in sf_create_index() 199 sf->index_size = 2 + (mag + 1) * sf->count; in sf_create_index() 200 bufp = sf->buffer; in sf_create_index() 203 for (i = 0; i < sf->count; ++i) { in sf_create_index() 204 int this_sz = sf->sizes[i]; in sf_create_index()
|