/external/webrtc/webrtc/common_video/libyuv/ |
D | webrtc_libyuv.cc | 253 dst_frame->buffer(kVPlane), in ConvertToI420() 254 dst_frame->stride(kVPlane), in ConvertToI420() 270 src_frame.buffer(kVPlane), in ConvertFromI420() 271 src_frame.stride(kVPlane), in ConvertFromI420() 285 src_frame.buffer(kVPlane), in ConvertFromYV12() 286 src_frame.stride(kVPlane), in ConvertFromYV12() 308 ref_frame->buffer(kVPlane), in I420PSNR() 309 ref_frame->stride(kVPlane), in I420PSNR() 314 test_frame->buffer(kVPlane), in I420PSNR() 315 test_frame->stride(kVPlane), in I420PSNR() [all …]
|
D | scaler.cc | 78 const uint8_t* v_ptr = src_frame.buffer(kVPlane) + in Scale() 79 src_offset_y / 2 * src_frame.stride(kVPlane) + in Scale() 87 src_frame.stride(kVPlane), in Scale() 93 dst_frame->buffer(kVPlane), in Scale() 94 dst_frame->stride(kVPlane), in Scale()
|
/external/webrtc/webrtc/common_video/ |
D | i420_video_frame_unittest.cc | 61 int stride_v = frame.stride(kVPlane); in TEST() 67 EXPECT_EQ(ExpectedSize(stride_v, height, kVPlane), in TEST() 68 frame.allocated_size(kVPlane)); in TEST() 112 memset(small_frame.buffer(kVPlane), 3, small_frame.allocated_size(kVPlane)); in TEST() 154 EXPECT_TRUE(const_frame1_ptr->buffer(kVPlane) == in TEST() 155 const_frame2_ptr->buffer(kVPlane)); in TEST() 209 EXPECT_TRUE(EqualPlane(buffer_v, frame2.buffer(kVPlane), stride_uv, 8, 8)); in TEST() 214 EXPECT_LE(kSizeUv, frame2.allocated_size(kVPlane)); in TEST() 222 const uint8_t* v = frame.buffer(kVPlane); in TEST() 226 EXPECT_EQ(v, frame.buffer(kVPlane)); in TEST() [all …]
|
D | video_frame.cc | 82 stride_u == stride(kUPlane) && stride_v == stride(kVPlane)) { in CreateEmptyFrame() 120 memcpy(buffer(kVPlane), buffer_v, expected_size_v); in CreateFrame() 146 videoFrame.buffer(kVPlane), videoFrame.width(), in CopyFrame() 148 videoFrame.stride(kUPlane), videoFrame.stride(kVPlane)); in CopyFrame() 229 stride(kVPlane) != frame.stride(kVPlane) || in EqualsFrame() 241 EqualPlane(buffer(kVPlane), frame.buffer(kVPlane), in EqualsFrame() 242 stride(kVPlane), half_width, half_height); in EqualsFrame()
|
D | video_frame_buffer.cc | 69 case kVPlane: in data() 90 case kVPlane: in stride() 176 case kVPlane: in data() 190 case kVPlane: in stride() 228 const uint8_t* v_plane = buffer->data(kVPlane) + in ShallowCenterCrop() 229 buffer->stride(kVPlane) * uv_offset_y + uv_offset_x; in ShallowCenterCrop() 234 v_plane, buffer->stride(kVPlane), in ShallowCenterCrop()
|
D | i420_buffer_pool_unittest.cc | 26 const uint8_t* v_ptr = buffer->data(kVPlane); in TEST() 33 EXPECT_EQ(v_ptr, buffer->data(kVPlane)); in TEST() 43 const uint8_t* v_ptr = buffer->data(kVPlane); in TEST() 51 EXPECT_NE(v_ptr, buffer->data(kVPlane)); in TEST()
|
/external/webrtc/talk/media/webrtc/ |
D | webrtcvideoframe.cc | 38 using webrtc::kVPlane; 108 return video_frame_buffer_ ? video_frame_buffer_->data(kVPlane) : nullptr; in GetVPlane() 122 return video_frame_buffer_ ? video_frame_buffer_->MutableData(kVPlane) in GetVPlane() 135 return video_frame_buffer_ ? video_frame_buffer_->stride(kVPlane) : 0; in GetVPitch() 170 video_frame_buffer_->stride(kVPlane)); in MakeExclusive() 174 new_buffer->MutableData(kVPlane), new_buffer->stride(kYPlane), in MakeExclusive() 175 new_buffer->stride(kUPlane), new_buffer->stride(kVPlane))) { in MakeExclusive()
|
/external/webrtc/webrtc/video/ |
D | video_capture_input_unittest.cc | 268 (frame1.stride(kVPlane) == frame2.stride(kVPlane)) && in EqualBufferFrames() 272 (frame1.allocated_size(kVPlane) == frame2.allocated_size(kVPlane)) && in EqualBufferFrames() 277 (memcmp(frame1.buffer(kVPlane), frame2.buffer(kVPlane), in EqualBufferFrames() 278 frame1.allocated_size(kVPlane)) == 0)); in EqualBufferFrames()
|
D | video_encoder_unittest.cc | 114 memset(frame_.buffer(webrtc::kVPlane), 128, in EncodeFrame() 115 frame_.allocated_size(webrtc::kVPlane)); in EncodeFrame()
|
/external/webrtc/webrtc/modules/video_processing/ |
D | video_denoiser.cc | 65 int stride_v = frame.stride(kVPlane); in DenoiseFrame() 72 frame.buffer(kVPlane), width_, height_, in DenoiseFrame() 87 uint8_t* v_dst = denoised_frame->buffer(kVPlane); in DenoiseFrame() 90 const uint8_t* v_src = frame.buffer(kVPlane); in DenoiseFrame()
|
/external/webrtc/webrtc/test/ |
D | frame_generator.cc | 44 memset(frame_.buffer(kVPlane), v, frame_.allocated_size(kVPlane)); in NextFrame() 209 int offset_v = (current_source_frame_->stride(PlaneType::kVPlane) * in CropSourceToScrolledImage() 216 ¤t_source_frame_->buffer(PlaneType::kVPlane)[offset_v], in CropSourceToScrolledImage() 220 current_source_frame_->stride(PlaneType::kVPlane)); in CropSourceToScrolledImage()
|
D | frame_generator_unittest.cc | 70 ASSERT_EQ(uv_size, frame->allocated_size(PlaneType::kVPlane)); in CheckFrameAndMutate() 71 buffer = frame->buffer(PlaneType::kVPlane); in CheckFrameAndMutate()
|
D | fake_texture_frame.h | 47 memset(buffer->MutableData(kVPlane), 0, half_height * half_width); in NativeToI420Buffer()
|
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/ |
D | simulcast_unittest.h | 130 EXPECT_NEAR(kColorV, decoded_image.buffer(kVPlane)[i], 4); in Decoded() 317 memset(input_frame_.buffer(kVPlane), 0, in SetUpCodec() 318 input_frame_.allocated_size(kVPlane)); in SetUpCodec() 567 memset(input_frame_.buffer(kVPlane), 0, in SwitchingToOneStream() 568 input_frame_.allocated_size(kVPlane)); in SwitchingToOneStream() 608 memset(input_frame_.buffer(kVPlane), 0, in SwitchingToOneStream() 609 input_frame_.allocated_size(kVPlane)); in SwitchingToOneStream() 696 plane_offset[kVPlane] = kColorV; in TestRPSIEncodeDecode() 712 plane_offset[kVPlane] += 1; in TestRPSIEncodeDecode() 720 plane_offset[kVPlane] += 1; in TestRPSIEncodeDecode() [all …]
|
D | simulcast_encoder_adapter.cc | 296 input_image.buffer(kVPlane), input_image.stride(kVPlane), src_width, in Encode() 299 dst_frame.buffer(kVPlane), dst_frame.stride(kVPlane), dst_width, in Encode()
|
/external/webrtc/webrtc/modules/video_coding/codecs/h264/ |
D | h264_video_toolbox_decoder.cc | 69 buffer->MutableData(webrtc::kVPlane), buffer->stride(webrtc::kVPlane), in VideoFrameBufferForPixelBuffer()
|
D | h264_video_toolbox_encoder.cc | 145 frame.buffer(webrtc::kVPlane), frame.stride(webrtc::kVPlane), dst_y, in CopyVideoFrameToPixelBuffer()
|
/external/webrtc/talk/app/webrtc/java/jni/ |
D | androidvideocapturer_jni.cc | 168 buffer->MutableData(webrtc::kVPlane), buffer->stride(webrtc::kVPlane), in OnMemoryBufferFrame()
|
D | androidmediadecoder_jni.cc | 705 frame_buffer->MutableData(webrtc::kVPlane), in DeliverPendingOutputs() 706 frame_buffer->stride(webrtc::kVPlane), in DeliverPendingOutputs() 719 frame_buffer->MutableData(webrtc::kVPlane), in DeliverPendingOutputs() 720 frame_buffer->stride(webrtc::kVPlane), in DeliverPendingOutputs()
|
/external/webrtc/webrtc/modules/video_render/android/ |
D | video_render_opengles20.cc | 391 GlTexSubImage2D(width / 2, height / 2, frameToRender.stride(kVPlane), in UpdateTextures() 392 frameToRender.buffer(kVPlane)); in UpdateTextures()
|
/external/webrtc/webrtc/modules/video_capture/test/ |
D | video_capture_unittest.cc | 70 (frame1.stride(webrtc::kVPlane) == frame2.stride(webrtc::kVPlane)) && in CompareFrames() 460 memset(test_frame_.buffer(webrtc::kVPlane), 127, in SetUp()
|
/external/webrtc/webrtc/common_video/include/ |
D | video_frame_buffer.h | 25 kVPlane = 2, enumerator
|
/external/webrtc/webrtc/modules/video_processing/test/ |
D | video_processing_unittest.cc | 77 memset(video_frame_.buffer(kVPlane), 0, video_frame_.allocated_size(kVPlane)); in SetUp()
|
/external/webrtc/webrtc/modules/video_render/test/testAPI/ |
D | testAPI.cc | 281 memset(frame->buffer(kVPlane), color, frame->allocated_size(kVPlane)); in GetTestVideoFrame()
|
/external/webrtc/webrtc/modules/video_render/ios/ |
D | open_gles20.mm | 329 width / 2, height / 2, frame.stride(kVPlane), frame.buffer(kVPlane));
|