Home
last modified time | relevance | path

Searched refs:input_frame (Results 1 – 11 of 11) sorted by relevance

/external/webrtc/webrtc/common_video/libyuv/
Dwebrtc_libyuv.cc126 int ExtractBuffer(const VideoFrame& input_frame, size_t size, uint8_t* buffer) { in ExtractBuffer() argument
128 if (input_frame.IsZeroSize()) in ExtractBuffer()
131 CalcBufferSize(kI420, input_frame.width(), input_frame.height()); in ExtractBuffer()
140 int width = (plane ? (input_frame.width() + 1) / 2 : in ExtractBuffer()
141 input_frame.width()); in ExtractBuffer()
142 int height = (plane ? (input_frame.height() + 1) / 2 : in ExtractBuffer()
143 input_frame.height()); in ExtractBuffer()
144 const uint8_t* plane_ptr = input_frame.buffer( in ExtractBuffer()
149 plane_ptr += input_frame.stride(static_cast<PlaneType>(plane)); in ExtractBuffer()
Dscaler_unittest.cc354 VideoFrame input_frame; in ScaleSequence() local
370 input_frame.CreateFrame(frame_buffer.get(), in ScaleSequence()
378 EXPECT_EQ(0, test_scaler_.Scale(input_frame, &output_frame)); in ScaleSequence()
/external/webrtc/talk/media/base/
Dvideoframefactory.cc36 const CapturedFrame* input_frame, in CreateAliasedFrame() argument
42 input_frame, cropped_input_width, cropped_input_height)); in CreateAliasedFrame()
54 (input_frame->rotation == webrtc::kVideoRotation_90 || in CreateAliasedFrame()
55 input_frame->rotation == webrtc::kVideoRotation_270)) { in CreateAliasedFrame()
Dvideoframefactory.h52 virtual VideoFrame* CreateAliasedFrame(const CapturedFrame* input_frame,
59 virtual VideoFrame* CreateAliasedFrame(const CapturedFrame* input_frame,
/external/tensorflow/tensorflow/core/common_runtime/
Dexecutor.cc1189 FrameState* input_frame = nullptr; member
1196 input_frame = in_frame; in TaggedNode()
1372 Entry* GetInputTensors(FrameState* input_frame, in GetInputTensors() argument
1374 return input_frame->GetIteration(input_iter)->input_tensors; in GetInputTensors()
1666 FrameState* input_frame = tagged_node.input_frame; in Process() local
1674 mutex_lock l(input_frame->mu); in Process()
1675 input_frame->GetIteration(input_iter)->mark_started(item.pending_id); in Process()
1700 Entry* input_tensors = GetInputTensors(input_frame, input_iter); in Process()
1723 MaybeMarkCompleted(input_frame, input_iter, id); in Process()
1732 params.frame_iter = FrameAndIter(input_frame->frame_id, input_iter); in Process()
[all …]
/external/webrtc/talk/app/webrtc/
Dandroidvideocapturer.cc99 const cricket::CapturedFrame* input_frame, in CreateAliasedFrame() argument
111 apply_rotation_ ? input_frame->rotation : in CreateAliasedFrame()
114 scaled_buffer, input_frame->time_stamp, in CreateAliasedFrame()
115 apply_rotation_ ? webrtc::kVideoRotation_0 : input_frame->rotation); in CreateAliasedFrame()
117 return VideoFrameFactory::CreateAliasedFrame(input_frame, in CreateAliasedFrame()
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/
Dvp8_sequence_coder.cc148 webrtc::VideoFrame input_frame; in SequenceCoder() local
163 input_frame.CreateEmptyFrame(width, height, width, half_width, half_width); in SequenceCoder()
170 height, 0, webrtc::kVideoRotation_0, &input_frame); in SequenceCoder()
171 encoder->Encode(input_frame, NULL, NULL); in SequenceCoder()
/external/webrtc/webrtc/modules/audio_processing/
Daudio_processing_performance_unittest.cc58 input_frame.resize(2); in AudioFrameData()
59 input_frame[0] = &input_framechannels[0]; in AudioFrameData()
60 input_frame[1] = &input_framechannels[max_frame_size]; in AudioFrameData()
71 std::vector<float*> input_frame; member
335 &frame_data_.input_frame[0], frame_data_.input_stream_config, in ProcessCapture()
363 &frame_data_.input_frame[0], frame_data_.input_stream_config, in ProcessRender()
412 rand_gen_, &frame_data_.input_frame[0]); in PrepareFrame()
Daudio_processing_impl_locking_unittest.cc95 input_frame.resize(2); in AudioFrameData()
96 input_frame[0] = &input_framechannels[0]; in AudioFrameData()
97 input_frame[1] = &input_framechannels[max_frame_size]; in AudioFrameData()
112 std::vector<float*> input_frame; member
685 PopulateAudioFrame(&frame_data_.input_frame[0], kCaptureInputFloatLevel, in PrepareFrame()
723 &frame_data_.input_frame[0], frame_data_.input_samples_per_channel, in CallApmCaptureSide()
730 &frame_data_.input_frame[0], frame_data_.input_stream_config, in CallApmCaptureSide()
958 PopulateAudioFrame(&frame_data_.input_frame[0], kRenderInputFloatLevel, in PrepareFrame()
993 &frame_data_.input_frame[0], frame_data_.input_stream_config, in CallApmRenderSide()
1001 &frame_data_.input_frame[0], frame_data_.input_samples_per_channel, in CallApmRenderSide()
/external/webrtc/talk/app/webrtc/java/jni/
Dandroidmediaencoder_jni.cc581 VideoFrame input_frame = frame; in EncodeOnCodecThread() local
596 input_frame.set_video_frame_buffer(scaled_buffer); in EncodeOnCodecThread()
598 input_frame = quality_scaler_.GetScaledFrame(frame); in EncodeOnCodecThread()
603 if (!MaybeReconfigureEncoderOnCodecThread(input_frame)) { in EncodeOnCodecThread()
625 if (!input_frame.native_handle()) { in EncodeOnCodecThread()
641 encode_status = EncodeByteBufferOnCodecThread(jni, key_frame, input_frame, in EncodeOnCodecThread()
644 encode_status = EncodeTextureOnCodecThread(jni, key_frame, input_frame); in EncodeOnCodecThread()
648 ALOGE << "Failed encode frame with timestamp: " << input_frame.timestamp(); in EncodeOnCodecThread()
658 timestamps_.push_back(input_frame.timestamp()); in EncodeOnCodecThread()
659 render_times_ms_.push_back(input_frame.render_time_ms()); in EncodeOnCodecThread()
/external/webrtc/webrtc/common_video/libyuv/include/
Dwebrtc_libyuv.h95 int ExtractBuffer(const VideoFrame& input_frame, size_t size, uint8_t* buffer);