/external/libvpx/libvpx/vpx_util/ |
D | vpx_debug_util.c | 88 static int frame_stride = MAX_FRAME_STRIDE; variable 123 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_record_block_pre() 131 [(r + pixel_r) * frame_stride + c + pixel_c] = in mismatch_record_block_pre() 157 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_record_block_tx() 165 [(r + pixel_r) * frame_stride + c + pixel_c] = in mismatch_record_block_tx() 192 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_check_block_pre() 200 [(r + pixel_r) * frame_stride + c + pixel_c] != in mismatch_check_block_pre() 218 [(rr + pixel_r) * frame_stride + cc + pixel_c]); in mismatch_check_block_pre() 240 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_check_block_tx() 248 [(r + pixel_r) * frame_stride + c + pixel_c] != in mismatch_check_block_tx() [all …]
|
/external/libaom/libaom/aom_util/ |
D | debug_util.c | 94 static int frame_stride = MAX_FRAME_STRIDE; variable 125 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_record_block_pre() 134 [(r + pixel_r) * frame_stride + c + pixel_c] = in mismatch_record_block_pre() 158 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_record_block_tx() 167 [(r + pixel_r) * frame_stride + c + pixel_c] = in mismatch_record_block_tx() 190 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_check_block_pre() 200 [(r + pixel_r) * frame_stride + c + pixel_c] != in mismatch_check_block_pre() 217 [(rr + pixel_r) * frame_stride + cc + pixel_c]); in mismatch_check_block_pre() 236 if (pixel_c + blk_w >= frame_stride || pixel_r + blk_h >= frame_height) { in mismatch_check_block_tx() 246 [(r + pixel_r) * frame_stride + c + pixel_c] != in mismatch_check_block_tx() [all …]
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/python/kernel_tests/ |
D | audio_microfrontend_op_test.py | 96 frame_stride=2) 138 frame_stride=2) 158 frame_stride=3,
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/python/ops/ |
D | audio_microfrontend_op.py | 53 frame_stride=1, argument 113 right_context, frame_stride, zero_padding, out_scale, out_type)
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/ |
D | audio_microfrontend.cc | 35 int frame_stride; member 72 data->frame_stride = m["frame_stride"].AsInt32(); in Init() 112 data->state->window.step / data->frame_stride + in Prepare() 160 for (anchor = 0; anchor < frame_buffer.size(); anchor += data->frame_stride) { in GenerateFeatures()
|
D | README.md | 66 **Striding.** An integer `frame_stride` that indicates the striding step used to 68 example above, with a `frame_stride=3`, the output tensor would have a shape of
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/ops/ |
D | audio_microfrontend_op.cc | 84 int frame_stride; in __anon74d2595d0102() local 85 TF_RETURN_IF_ERROR(ctx->GetAttr("frame_stride", &frame_stride)); in __anon74d2595d0102() 95 ctx->Divide(num_frames, frame_stride, false, &num_frames)); in __anon74d2595d0102()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | temporal_filter_sse2.c | 245 const uint32_t frame_stride = ref_frame->strides[plane == 0 ? 0 : 1]; in av1_apply_temporal_filter_planewise_sse2() local 246 const int frame_offset = mb_row * plane_h * frame_stride + mb_col * plane_w; in av1_apply_temporal_filter_planewise_sse2() 255 ref, frame_stride, pred + mb_pels * plane, plane_w, plane_w, plane_h, in av1_apply_temporal_filter_planewise_sse2()
|
D | temporal_filter_avx2.c | 267 const uint32_t frame_stride = ref_frame->strides[plane == 0 ? 0 : 1]; in av1_apply_temporal_filter_planewise_avx2() local 268 const int frame_offset = mb_row * plane_h * frame_stride + mb_col * plane_w; in av1_apply_temporal_filter_planewise_avx2() 277 ref, frame_stride, pred + mb_pels * plane, plane_w, plane_w, plane_h, in av1_apply_temporal_filter_planewise_avx2()
|
/external/libaom/libaom/av1/encoder/ |
D | temporal_filter.c | 543 const int frame_stride = frame_to_filter->strides[plane == 0 ? 0 : 1]; in av1_apply_temporal_filter_yuv_c() local 544 const int frame_offset = mb_row * plane_h * frame_stride + mb_col * plane_w; in av1_apply_temporal_filter_yuv_c() 546 compute_square_diff(ref, frame_offset, frame_stride, pred, plane_offset, in av1_apply_temporal_filter_yuv_c() 679 const int frame_stride = frame_to_filter->strides[plane == 0 ? 0 : 1]; in av1_apply_temporal_filter_planewise_c() local 680 const int frame_offset = mb_row * plane_h * frame_stride + mb_col * plane_w; in av1_apply_temporal_filter_planewise_c() 682 compute_square_diff(ref, frame_offset, frame_stride, pred, plane_offset, in av1_apply_temporal_filter_planewise_c() 877 const int frame_stride = result_buffer->strides[plane == 0 ? 0 : 1]; in tf_normalize_filtered_frame() local 878 const int frame_offset = mb_row * plane_h * frame_stride + mb_col * plane_w; in tf_normalize_filtered_frame() 897 frame_idx += (frame_stride - plane_w); in tf_normalize_filtered_frame()
|