• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // Copyright 2013 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
4 
5 #include "content/renderer/media/rtc_video_decoder.h"
6 
7 #include "base/bind.h"
8 #include "base/logging.h"
9 #include "base/memory/ref_counted.h"
10 #include "base/message_loop/message_loop_proxy.h"
11 #include "base/metrics/histogram.h"
12 #include "base/safe_numerics.h"
13 #include "base/stl_util.h"
14 #include "base/task_runner_util.h"
15 #include "content/child/child_thread.h"
16 #include "content/renderer/media/native_handle_impl.h"
17 #include "media/base/bind_to_loop.h"
18 #include "media/filters/gpu_video_accelerator_factories.h"
19 #include "third_party/webrtc/common_video/interface/texture_video_frame.h"
20 #include "third_party/webrtc/system_wrappers/interface/ref_count.h"
21 
22 namespace content {
23 
24 const int32 RTCVideoDecoder::ID_LAST = 0x3FFFFFFF;
25 const int32 RTCVideoDecoder::ID_HALF = 0x20000000;
26 const int32 RTCVideoDecoder::ID_INVALID = -1;
27 
28 // Maximum number of concurrent VDA::Decode() operations RVD will maintain.
29 // Higher values allow better pipelining in the GPU, but also require more
30 // resources.
31 static const size_t kMaxInFlightDecodes = 8;
32 
33 // Size of shared-memory segments we allocate.  Since we reuse them we let them
34 // be on the beefy side.
35 static const size_t kSharedMemorySegmentBytes = 100 << 10;
36 
37 // Maximum number of allocated shared-memory segments.
38 static const int kMaxNumSharedMemorySegments = 16;
39 
40 // Maximum number of pending WebRTC buffers that are waiting for the shared
41 // memory. 10 seconds for 30 fps.
42 static const size_t kMaxNumOfPendingBuffers = 300;
43 
44 // A shared memory segment and its allocated size. This class has the ownership
45 // of |shm|.
46 class RTCVideoDecoder::SHMBuffer {
47  public:
48   SHMBuffer(base::SharedMemory* shm, size_t size);
49   ~SHMBuffer();
50   base::SharedMemory* const shm;
51   const size_t size;
52 };
53 
SHMBuffer(base::SharedMemory * shm,size_t size)54 RTCVideoDecoder::SHMBuffer::SHMBuffer(base::SharedMemory* shm, size_t size)
55     : shm(shm), size(size) {}
56 
~SHMBuffer()57 RTCVideoDecoder::SHMBuffer::~SHMBuffer() { shm->Close(); }
58 
BufferData(int32 bitstream_buffer_id,uint32_t timestamp,int width,int height,size_t size)59 RTCVideoDecoder::BufferData::BufferData(int32 bitstream_buffer_id,
60                                         uint32_t timestamp,
61                                         int width,
62                                         int height,
63                                         size_t size)
64     : bitstream_buffer_id(bitstream_buffer_id),
65       timestamp(timestamp),
66       width(width),
67       height(height),
68       size(size) {}
69 
BufferData()70 RTCVideoDecoder::BufferData::BufferData() {}
71 
~BufferData()72 RTCVideoDecoder::BufferData::~BufferData() {}
73 
RTCVideoDecoder(const scoped_refptr<media::GpuVideoAcceleratorFactories> & factories)74 RTCVideoDecoder::RTCVideoDecoder(
75     const scoped_refptr<media::GpuVideoAcceleratorFactories>& factories)
76     : factories_(factories),
77       vda_loop_proxy_(factories->GetMessageLoop()),
78       decoder_texture_target_(0),
79       next_picture_buffer_id_(0),
80       state_(UNINITIALIZED),
81       decode_complete_callback_(NULL),
82       num_shm_buffers_(0),
83       next_bitstream_buffer_id_(0),
84       reset_bitstream_buffer_id_(ID_INVALID),
85       weak_factory_(this) {
86   DCHECK(!vda_loop_proxy_->BelongsToCurrentThread());
87 
88   weak_this_ = weak_factory_.GetWeakPtr();
89 
90   base::WaitableEvent message_loop_async_waiter(false, false);
91   // Waiting here is safe. The media thread is stopped in the child thread and
92   // the child thread is blocked when VideoDecoderFactory::CreateVideoDecoder
93   // runs.
94   vda_loop_proxy_->PostTask(FROM_HERE,
95                             base::Bind(&RTCVideoDecoder::Initialize,
96                                        base::Unretained(this),
97                                        &message_loop_async_waiter));
98   message_loop_async_waiter.Wait();
99 }
100 
~RTCVideoDecoder()101 RTCVideoDecoder::~RTCVideoDecoder() {
102   DVLOG(2) << "~RTCVideoDecoder";
103   // Destroy VDA and remove |this| from the observer if this is vda thread.
104   if (vda_loop_proxy_->BelongsToCurrentThread()) {
105     base::MessageLoop::current()->RemoveDestructionObserver(this);
106     DestroyVDA();
107   } else {
108     // VDA should have been destroyed in WillDestroyCurrentMessageLoop.
109     DCHECK(!vda_);
110   }
111 
112   // Delete all shared memories.
113   STLDeleteElements(&available_shm_segments_);
114   STLDeleteValues(&bitstream_buffers_in_decoder_);
115   STLDeleteContainerPairFirstPointers(decode_buffers_.begin(),
116                                       decode_buffers_.end());
117   decode_buffers_.clear();
118 
119   // Delete WebRTC input buffers.
120   for (std::deque<std::pair<webrtc::EncodedImage, BufferData> >::iterator it =
121            pending_buffers_.begin();
122        it != pending_buffers_.end();
123        ++it) {
124     delete[] it->first._buffer;
125   }
126 }
127 
Create(webrtc::VideoCodecType type,const scoped_refptr<media::GpuVideoAcceleratorFactories> & factories)128 scoped_ptr<RTCVideoDecoder> RTCVideoDecoder::Create(
129     webrtc::VideoCodecType type,
130     const scoped_refptr<media::GpuVideoAcceleratorFactories>& factories) {
131   scoped_ptr<RTCVideoDecoder> decoder;
132   // Convert WebRTC codec type to media codec profile.
133   media::VideoCodecProfile profile;
134   switch (type) {
135     case webrtc::kVideoCodecVP8:
136       profile = media::VP8PROFILE_MAIN;
137       break;
138     default:
139       DVLOG(2) << "Video codec not supported:" << type;
140       return decoder.Pass();
141   }
142 
143   decoder.reset(new RTCVideoDecoder(factories));
144   decoder->vda_ =
145       factories->CreateVideoDecodeAccelerator(profile, decoder.get()).Pass();
146   // vda can be NULL if VP8 is not supported.
147   if (decoder->vda_ != NULL) {
148     decoder->state_ = INITIALIZED;
149   } else {
150     factories->GetMessageLoop()->DeleteSoon(FROM_HERE, decoder.release());
151   }
152   return decoder.Pass();
153 }
154 
InitDecode(const webrtc::VideoCodec * codecSettings,int32_t)155 int32_t RTCVideoDecoder::InitDecode(const webrtc::VideoCodec* codecSettings,
156                                     int32_t /*numberOfCores*/) {
157   DVLOG(2) << "InitDecode";
158   DCHECK_EQ(codecSettings->codecType, webrtc::kVideoCodecVP8);
159   if (codecSettings->codecSpecific.VP8.feedbackModeOn) {
160     LOG(ERROR) << "Feedback mode not supported";
161     return RecordInitDecodeUMA(WEBRTC_VIDEO_CODEC_ERROR);
162   }
163 
164   base::AutoLock auto_lock(lock_);
165   if (state_ == UNINITIALIZED || state_ == DECODE_ERROR) {
166     LOG(ERROR) << "VDA is not initialized. state=" << state_;
167     return RecordInitDecodeUMA(WEBRTC_VIDEO_CODEC_UNINITIALIZED);
168   }
169   // Create some shared memory if the queue is empty.
170   if (available_shm_segments_.size() == 0) {
171     vda_loop_proxy_->PostTask(FROM_HERE,
172                               base::Bind(&RTCVideoDecoder::CreateSHM,
173                                          weak_this_,
174                                          kMaxInFlightDecodes,
175                                          kSharedMemorySegmentBytes));
176   }
177   return RecordInitDecodeUMA(WEBRTC_VIDEO_CODEC_OK);
178 }
179 
Decode(const webrtc::EncodedImage & inputImage,bool missingFrames,const webrtc::RTPFragmentationHeader *,const webrtc::CodecSpecificInfo *,int64_t)180 int32_t RTCVideoDecoder::Decode(
181     const webrtc::EncodedImage& inputImage,
182     bool missingFrames,
183     const webrtc::RTPFragmentationHeader* /*fragmentation*/,
184     const webrtc::CodecSpecificInfo* /*codecSpecificInfo*/,
185     int64_t /*renderTimeMs*/) {
186   DVLOG(3) << "Decode";
187 
188   base::AutoLock auto_lock(lock_);
189 
190   if (state_ == UNINITIALIZED || decode_complete_callback_ == NULL) {
191     LOG(ERROR) << "The decoder has not initialized.";
192     return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
193   }
194 
195   if (state_ == DECODE_ERROR) {
196     LOG(ERROR) << "Decoding error occurred.";
197     return WEBRTC_VIDEO_CODEC_ERROR;
198   }
199 
200   if (missingFrames || !inputImage._completeFrame) {
201     DLOG(ERROR) << "Missing or incomplete frames.";
202     // Unlike the SW decoder in libvpx, hw decoder cannot handle broken frames.
203     // Return an error to request a key frame.
204     return WEBRTC_VIDEO_CODEC_ERROR;
205   }
206 
207   // Most platforms' VDA implementations support mid-stream resolution change
208   // internally.  Platforms whose VDAs fail to support mid-stream resolution
209   // change gracefully need to have their clients cover for them, and we do that
210   // here.
211 #ifdef ANDROID
212   const bool kVDACanHandleMidstreamResize = false;
213 #else
214   const bool kVDACanHandleMidstreamResize = true;
215 #endif
216 
217   bool need_to_reset_for_midstream_resize = false;
218   if (inputImage._frameType == webrtc::kKeyFrame) {
219     DVLOG(2) << "Got key frame. size=" << inputImage._encodedWidth << "x"
220              << inputImage._encodedHeight;
221     gfx::Size prev_frame_size = frame_size_;
222     frame_size_.SetSize(inputImage._encodedWidth, inputImage._encodedHeight);
223     if (!kVDACanHandleMidstreamResize && !prev_frame_size.IsEmpty() &&
224         prev_frame_size != frame_size_) {
225       need_to_reset_for_midstream_resize = true;
226     }
227   } else if (IsFirstBufferAfterReset(next_bitstream_buffer_id_,
228                                      reset_bitstream_buffer_id_)) {
229     // TODO(wuchengli): VDA should handle it. Remove this when
230     // http://crosbug.com/p/21913 is fixed.
231     DVLOG(1) << "The first frame should be a key frame. Drop this.";
232     return WEBRTC_VIDEO_CODEC_ERROR;
233   }
234 
235   // Create buffer metadata.
236   BufferData buffer_data(next_bitstream_buffer_id_,
237                          inputImage._timeStamp,
238                          frame_size_.width(),
239                          frame_size_.height(),
240                          inputImage._length);
241   // Mask against 30 bits, to avoid (undefined) wraparound on signed integer.
242   next_bitstream_buffer_id_ = (next_bitstream_buffer_id_ + 1) & ID_LAST;
243 
244   // If a shared memory segment is available, there are no pending buffers, and
245   // this isn't a mid-stream resolution change, then send the buffer for decode
246   // immediately. Otherwise, save the buffer in the queue for later decode.
247   scoped_ptr<SHMBuffer> shm_buffer;
248   if (!need_to_reset_for_midstream_resize && pending_buffers_.size() == 0)
249     shm_buffer = GetSHM_Locked(inputImage._length);
250   if (!shm_buffer) {
251     if (!SaveToPendingBuffers_Locked(inputImage, buffer_data))
252       return WEBRTC_VIDEO_CODEC_ERROR;
253     if (need_to_reset_for_midstream_resize) {
254       base::AutoUnlock auto_unlock(lock_);
255       Reset();
256     }
257     return WEBRTC_VIDEO_CODEC_OK;
258   }
259 
260   SaveToDecodeBuffers_Locked(inputImage, shm_buffer.Pass(), buffer_data);
261   vda_loop_proxy_->PostTask(
262       FROM_HERE, base::Bind(&RTCVideoDecoder::RequestBufferDecode, weak_this_));
263   return WEBRTC_VIDEO_CODEC_OK;
264 }
265 
RegisterDecodeCompleteCallback(webrtc::DecodedImageCallback * callback)266 int32_t RTCVideoDecoder::RegisterDecodeCompleteCallback(
267     webrtc::DecodedImageCallback* callback) {
268   DVLOG(2) << "RegisterDecodeCompleteCallback";
269   base::AutoLock auto_lock(lock_);
270   decode_complete_callback_ = callback;
271   return WEBRTC_VIDEO_CODEC_OK;
272 }
273 
Release()274 int32_t RTCVideoDecoder::Release() {
275   DVLOG(2) << "Release";
276   // Do not destroy VDA because WebRTC can call InitDecode and start decoding
277   // again.
278   return Reset();
279 }
280 
Reset()281 int32_t RTCVideoDecoder::Reset() {
282   DVLOG(2) << "Reset";
283   base::AutoLock auto_lock(lock_);
284   if (state_ == UNINITIALIZED) {
285     LOG(ERROR) << "Decoder not initialized.";
286     return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
287   }
288   if (next_bitstream_buffer_id_ != 0)
289     reset_bitstream_buffer_id_ = next_bitstream_buffer_id_ - 1;
290   else
291     reset_bitstream_buffer_id_ = ID_LAST;
292   // If VDA is already resetting, no need to request the reset again.
293   if (state_ != RESETTING) {
294     state_ = RESETTING;
295     vda_loop_proxy_->PostTask(
296         FROM_HERE, base::Bind(&RTCVideoDecoder::ResetInternal, weak_this_));
297   }
298   return WEBRTC_VIDEO_CODEC_OK;
299 }
300 
NotifyInitializeDone()301 void RTCVideoDecoder::NotifyInitializeDone() {
302   DVLOG(2) << "NotifyInitializeDone";
303   NOTREACHED();
304 }
305 
ProvidePictureBuffers(uint32 count,const gfx::Size & size,uint32 texture_target)306 void RTCVideoDecoder::ProvidePictureBuffers(uint32 count,
307                                             const gfx::Size& size,
308                                             uint32 texture_target) {
309   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
310   DVLOG(3) << "ProvidePictureBuffers. texture_target=" << texture_target;
311 
312   if (!vda_)
313     return;
314 
315   std::vector<uint32> texture_ids;
316   std::vector<gpu::Mailbox> texture_mailboxes;
317   decoder_texture_target_ = texture_target;
318   // Discards the sync point returned here since PictureReady will imply that
319   // the produce has already happened, and the texture is ready for use.
320   if (!factories_->CreateTextures(count,
321                                   size,
322                                   &texture_ids,
323                                   &texture_mailboxes,
324                                   decoder_texture_target_)) {
325     NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
326     return;
327   }
328   DCHECK_EQ(count, texture_ids.size());
329   DCHECK_EQ(count, texture_mailboxes.size());
330 
331   std::vector<media::PictureBuffer> picture_buffers;
332   for (size_t i = 0; i < texture_ids.size(); ++i) {
333     picture_buffers.push_back(media::PictureBuffer(
334         next_picture_buffer_id_++, size, texture_ids[i], texture_mailboxes[i]));
335     bool inserted = assigned_picture_buffers_.insert(std::make_pair(
336         picture_buffers.back().id(), picture_buffers.back())).second;
337     DCHECK(inserted);
338   }
339   vda_->AssignPictureBuffers(picture_buffers);
340 }
341 
DismissPictureBuffer(int32 id)342 void RTCVideoDecoder::DismissPictureBuffer(int32 id) {
343   DVLOG(3) << "DismissPictureBuffer. id=" << id;
344   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
345 
346   std::map<int32, media::PictureBuffer>::iterator it =
347       assigned_picture_buffers_.find(id);
348   if (it == assigned_picture_buffers_.end()) {
349     NOTREACHED() << "Missing picture buffer: " << id;
350     return;
351   }
352 
353   media::PictureBuffer buffer_to_dismiss = it->second;
354   assigned_picture_buffers_.erase(it);
355 
356   std::set<int32>::iterator at_display_it =
357       picture_buffers_at_display_.find(id);
358 
359   if (at_display_it == picture_buffers_at_display_.end()) {
360     // We can delete the texture immediately as it's not being displayed.
361     factories_->DeleteTexture(buffer_to_dismiss.texture_id());
362   } else {
363     // Texture in display. Postpone deletion until after it's returned to us.
364     bool inserted = dismissed_picture_buffers_
365         .insert(std::make_pair(id, buffer_to_dismiss)).second;
366     DCHECK(inserted);
367   }
368 }
369 
PictureReady(const media::Picture & picture)370 void RTCVideoDecoder::PictureReady(const media::Picture& picture) {
371   DVLOG(3) << "PictureReady";
372   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
373 
374   std::map<int32, media::PictureBuffer>::iterator it =
375       assigned_picture_buffers_.find(picture.picture_buffer_id());
376   if (it == assigned_picture_buffers_.end()) {
377     NOTREACHED() << "Missing picture buffer: " << picture.picture_buffer_id();
378     NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
379     return;
380   }
381   const media::PictureBuffer& pb = it->second;
382 
383   // Create a media::VideoFrame.
384   uint32_t timestamp = 0, width = 0, height = 0;
385   size_t size = 0;
386   GetBufferData(
387       picture.bitstream_buffer_id(), &timestamp, &width, &height, &size);
388   scoped_refptr<media::VideoFrame> frame =
389       CreateVideoFrame(picture, pb, timestamp, width, height, size);
390   bool inserted =
391       picture_buffers_at_display_.insert(picture.picture_buffer_id()).second;
392   DCHECK(inserted);
393 
394   // Create a WebRTC video frame.
395   webrtc::RefCountImpl<NativeHandleImpl>* handle =
396       new webrtc::RefCountImpl<NativeHandleImpl>(frame);
397   webrtc::TextureVideoFrame decoded_image(handle, width, height, timestamp, 0);
398 
399   // Invoke decode callback. WebRTC expects no callback after Reset or Release.
400   {
401     base::AutoLock auto_lock(lock_);
402     DCHECK(decode_complete_callback_ != NULL);
403     if (IsBufferAfterReset(picture.bitstream_buffer_id(),
404                            reset_bitstream_buffer_id_)) {
405       decode_complete_callback_->Decoded(decoded_image);
406     }
407   }
408 }
409 
CreateVideoFrame(const media::Picture & picture,const media::PictureBuffer & pb,uint32_t timestamp,uint32_t width,uint32_t height,size_t size)410 scoped_refptr<media::VideoFrame> RTCVideoDecoder::CreateVideoFrame(
411     const media::Picture& picture,
412     const media::PictureBuffer& pb,
413     uint32_t timestamp,
414     uint32_t width,
415     uint32_t height,
416     size_t size) {
417   gfx::Rect visible_rect(width, height);
418   gfx::Size natural_size(width, height);
419   DCHECK(decoder_texture_target_);
420   // Convert timestamp from 90KHz to ms.
421   base::TimeDelta timestamp_ms = base::TimeDelta::FromInternalValue(
422       base::checked_numeric_cast<uint64_t>(timestamp) * 1000 / 90);
423   return media::VideoFrame::WrapNativeTexture(
424       make_scoped_ptr(new media::VideoFrame::MailboxHolder(
425           pb.texture_mailbox(),
426           0,  // sync_point
427           media::BindToCurrentLoop(
428               base::Bind(&RTCVideoDecoder::ReusePictureBuffer,
429                          weak_this_,
430                          picture.picture_buffer_id())))),
431       decoder_texture_target_,
432       pb.size(),
433       visible_rect,
434       natural_size,
435       timestamp_ms,
436       base::Bind(&media::GpuVideoAcceleratorFactories::ReadPixels,
437                  factories_,
438                  pb.texture_id(),
439                  natural_size),
440       base::Closure());
441 }
442 
NotifyEndOfBitstreamBuffer(int32 id)443 void RTCVideoDecoder::NotifyEndOfBitstreamBuffer(int32 id) {
444   DVLOG(3) << "NotifyEndOfBitstreamBuffer. id=" << id;
445   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
446 
447   std::map<int32, SHMBuffer*>::iterator it =
448       bitstream_buffers_in_decoder_.find(id);
449   if (it == bitstream_buffers_in_decoder_.end()) {
450     NotifyError(media::VideoDecodeAccelerator::PLATFORM_FAILURE);
451     NOTREACHED() << "Missing bitstream buffer: " << id;
452     return;
453   }
454 
455   {
456     base::AutoLock auto_lock(lock_);
457     PutSHM_Locked(scoped_ptr<SHMBuffer>(it->second));
458   }
459   bitstream_buffers_in_decoder_.erase(it);
460 
461   RequestBufferDecode();
462 }
463 
NotifyFlushDone()464 void RTCVideoDecoder::NotifyFlushDone() {
465   DVLOG(3) << "NotifyFlushDone";
466   NOTREACHED() << "Unexpected flush done notification.";
467 }
468 
NotifyResetDone()469 void RTCVideoDecoder::NotifyResetDone() {
470   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
471   DVLOG(3) << "NotifyResetDone";
472 
473   if (!vda_)
474     return;
475 
476   input_buffer_data_.clear();
477   {
478     base::AutoLock auto_lock(lock_);
479     state_ = INITIALIZED;
480   }
481   // Send the pending buffers for decoding.
482   RequestBufferDecode();
483 }
484 
NotifyError(media::VideoDecodeAccelerator::Error error)485 void RTCVideoDecoder::NotifyError(media::VideoDecodeAccelerator::Error error) {
486   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
487   if (!vda_)
488     return;
489 
490   LOG(ERROR) << "VDA Error:" << error;
491   UMA_HISTOGRAM_ENUMERATION("Media.RTCVideoDecoderError",
492                             error,
493                             media::VideoDecodeAccelerator::LARGEST_ERROR_ENUM);
494   DestroyVDA();
495 
496   base::AutoLock auto_lock(lock_);
497   state_ = DECODE_ERROR;
498 }
499 
WillDestroyCurrentMessageLoop()500 void RTCVideoDecoder::WillDestroyCurrentMessageLoop() {
501   DVLOG(2) << "WillDestroyCurrentMessageLoop";
502   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
503   factories_->Abort();
504   weak_factory_.InvalidateWeakPtrs();
505   DestroyVDA();
506 }
507 
Initialize(base::WaitableEvent * waiter)508 void RTCVideoDecoder::Initialize(base::WaitableEvent* waiter) {
509   DVLOG(2) << "Initialize";
510   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
511   base::MessageLoop::current()->AddDestructionObserver(this);
512   waiter->Signal();
513 }
514 
RequestBufferDecode()515 void RTCVideoDecoder::RequestBufferDecode() {
516   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
517   if (!vda_)
518     return;
519 
520   MovePendingBuffersToDecodeBuffers();
521 
522   while (CanMoreDecodeWorkBeDone()) {
523     // Get a buffer and data from the queue.
524     SHMBuffer* shm_buffer = NULL;
525     BufferData buffer_data;
526     {
527       base::AutoLock auto_lock(lock_);
528       // Do not request decode if VDA is resetting.
529       if (decode_buffers_.size() == 0 || state_ == RESETTING)
530         return;
531       shm_buffer = decode_buffers_.front().first;
532       buffer_data = decode_buffers_.front().second;
533       decode_buffers_.pop_front();
534       // Drop the buffers before Reset or Release is called.
535       if (!IsBufferAfterReset(buffer_data.bitstream_buffer_id,
536                               reset_bitstream_buffer_id_)) {
537         PutSHM_Locked(scoped_ptr<SHMBuffer>(shm_buffer));
538         continue;
539       }
540     }
541 
542     // Create a BitstreamBuffer and send to VDA to decode.
543     media::BitstreamBuffer bitstream_buffer(buffer_data.bitstream_buffer_id,
544                                             shm_buffer->shm->handle(),
545                                             buffer_data.size);
546     bool inserted = bitstream_buffers_in_decoder_
547         .insert(std::make_pair(bitstream_buffer.id(), shm_buffer)).second;
548     DCHECK(inserted);
549     RecordBufferData(buffer_data);
550     vda_->Decode(bitstream_buffer);
551   }
552 }
553 
CanMoreDecodeWorkBeDone()554 bool RTCVideoDecoder::CanMoreDecodeWorkBeDone() {
555   return bitstream_buffers_in_decoder_.size() < kMaxInFlightDecodes;
556 }
557 
IsBufferAfterReset(int32 id_buffer,int32 id_reset)558 bool RTCVideoDecoder::IsBufferAfterReset(int32 id_buffer, int32 id_reset) {
559   if (id_reset == ID_INVALID)
560     return true;
561   int32 diff = id_buffer - id_reset;
562   if (diff <= 0)
563     diff += ID_LAST + 1;
564   return diff < ID_HALF;
565 }
566 
IsFirstBufferAfterReset(int32 id_buffer,int32 id_reset)567 bool RTCVideoDecoder::IsFirstBufferAfterReset(int32 id_buffer, int32 id_reset) {
568   if (id_reset == ID_INVALID)
569     return id_buffer == 0;
570   return id_buffer == ((id_reset + 1) & ID_LAST);
571 }
572 
SaveToDecodeBuffers_Locked(const webrtc::EncodedImage & input_image,scoped_ptr<SHMBuffer> shm_buffer,const BufferData & buffer_data)573 void RTCVideoDecoder::SaveToDecodeBuffers_Locked(
574     const webrtc::EncodedImage& input_image,
575     scoped_ptr<SHMBuffer> shm_buffer,
576     const BufferData& buffer_data) {
577   memcpy(shm_buffer->shm->memory(), input_image._buffer, input_image._length);
578   std::pair<SHMBuffer*, BufferData> buffer_pair =
579       std::make_pair(shm_buffer.release(), buffer_data);
580 
581   // Store the buffer and the metadata to the queue.
582   decode_buffers_.push_back(buffer_pair);
583 }
584 
SaveToPendingBuffers_Locked(const webrtc::EncodedImage & input_image,const BufferData & buffer_data)585 bool RTCVideoDecoder::SaveToPendingBuffers_Locked(
586     const webrtc::EncodedImage& input_image,
587     const BufferData& buffer_data) {
588   DVLOG(2) << "SaveToPendingBuffers_Locked"
589            << ". pending_buffers size=" << pending_buffers_.size()
590            << ". decode_buffers_ size=" << decode_buffers_.size()
591            << ". available_shm size=" << available_shm_segments_.size();
592   // Queued too many buffers. Something goes wrong.
593   if (pending_buffers_.size() >= kMaxNumOfPendingBuffers) {
594     LOG(WARNING) << "Too many pending buffers!";
595     return false;
596   }
597 
598   // Clone the input image and save it to the queue.
599   uint8_t* buffer = new uint8_t[input_image._length];
600   // TODO(wuchengli): avoid memcpy. Extend webrtc::VideoDecoder::Decode()
601   // interface to take a non-const ptr to the frame and add a method to the
602   // frame that will swap buffers with another.
603   memcpy(buffer, input_image._buffer, input_image._length);
604   webrtc::EncodedImage encoded_image(
605       buffer, input_image._length, input_image._length);
606   std::pair<webrtc::EncodedImage, BufferData> buffer_pair =
607       std::make_pair(encoded_image, buffer_data);
608 
609   pending_buffers_.push_back(buffer_pair);
610   return true;
611 }
612 
MovePendingBuffersToDecodeBuffers()613 void RTCVideoDecoder::MovePendingBuffersToDecodeBuffers() {
614   base::AutoLock auto_lock(lock_);
615   while (pending_buffers_.size() > 0) {
616     // Get a pending buffer from the queue.
617     const webrtc::EncodedImage& input_image = pending_buffers_.front().first;
618     const BufferData& buffer_data = pending_buffers_.front().second;
619 
620     // Drop the frame if it comes before Reset or Release.
621     if (!IsBufferAfterReset(buffer_data.bitstream_buffer_id,
622                             reset_bitstream_buffer_id_)) {
623       delete[] input_image._buffer;
624       pending_buffers_.pop_front();
625       continue;
626     }
627     // Get shared memory and save it to decode buffers.
628     scoped_ptr<SHMBuffer> shm_buffer = GetSHM_Locked(input_image._length);
629     if (!shm_buffer)
630       return;
631     SaveToDecodeBuffers_Locked(input_image, shm_buffer.Pass(), buffer_data);
632     delete[] input_image._buffer;
633     pending_buffers_.pop_front();
634   }
635 }
636 
ResetInternal()637 void RTCVideoDecoder::ResetInternal() {
638   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
639   DVLOG(2) << "ResetInternal";
640   if (vda_)
641     vda_->Reset();
642 }
643 
ReusePictureBuffer(int64 picture_buffer_id,uint32 sync_point)644 void RTCVideoDecoder::ReusePictureBuffer(int64 picture_buffer_id,
645                                          uint32 sync_point) {
646   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
647   DVLOG(3) << "ReusePictureBuffer. id=" << picture_buffer_id;
648 
649   if (!vda_)
650     return;
651 
652   CHECK(!picture_buffers_at_display_.empty());
653 
654   size_t num_erased = picture_buffers_at_display_.erase(picture_buffer_id);
655   DCHECK(num_erased);
656 
657   std::map<int32, media::PictureBuffer>::iterator it =
658       assigned_picture_buffers_.find(picture_buffer_id);
659 
660   if (it == assigned_picture_buffers_.end()) {
661     // This picture was dismissed while in display, so we postponed deletion.
662     it = dismissed_picture_buffers_.find(picture_buffer_id);
663     DCHECK(it != dismissed_picture_buffers_.end());
664     factories_->DeleteTexture(it->second.texture_id());
665     dismissed_picture_buffers_.erase(it);
666     return;
667   }
668 
669   factories_->WaitSyncPoint(sync_point);
670 
671   vda_->ReusePictureBuffer(picture_buffer_id);
672 }
673 
DestroyTextures()674 void RTCVideoDecoder::DestroyTextures() {
675   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
676   std::map<int32, media::PictureBuffer>::iterator it;
677 
678   for (it = assigned_picture_buffers_.begin();
679        it != assigned_picture_buffers_.end();
680        ++it) {
681     factories_->DeleteTexture(it->second.texture_id());
682   }
683   assigned_picture_buffers_.clear();
684 
685   for (it = dismissed_picture_buffers_.begin();
686        it != dismissed_picture_buffers_.end();
687        ++it) {
688     factories_->DeleteTexture(it->second.texture_id());
689   }
690   dismissed_picture_buffers_.clear();
691 }
692 
DestroyVDA()693 void RTCVideoDecoder::DestroyVDA() {
694   DVLOG(2) << "DestroyVDA";
695   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
696   if (vda_)
697     vda_.release()->Destroy();
698   DestroyTextures();
699   base::AutoLock auto_lock(lock_);
700   state_ = UNINITIALIZED;
701 }
702 
GetSHM_Locked(size_t min_size)703 scoped_ptr<RTCVideoDecoder::SHMBuffer> RTCVideoDecoder::GetSHM_Locked(
704     size_t min_size) {
705   // Reuse a SHM if possible.
706   SHMBuffer* ret = NULL;
707   if (!available_shm_segments_.empty() &&
708       available_shm_segments_.back()->size >= min_size) {
709     ret = available_shm_segments_.back();
710     available_shm_segments_.pop_back();
711   }
712   // Post to vda thread to create shared memory if SHM cannot be reused or the
713   // queue is almost empty.
714   if (num_shm_buffers_ < kMaxNumSharedMemorySegments &&
715       (ret == NULL || available_shm_segments_.size() <= 1)) {
716     vda_loop_proxy_->PostTask(
717         FROM_HERE,
718         base::Bind(&RTCVideoDecoder::CreateSHM, weak_this_, 1, min_size));
719   }
720   return scoped_ptr<SHMBuffer>(ret);
721 }
722 
PutSHM_Locked(scoped_ptr<SHMBuffer> shm_buffer)723 void RTCVideoDecoder::PutSHM_Locked(scoped_ptr<SHMBuffer> shm_buffer) {
724   available_shm_segments_.push_back(shm_buffer.release());
725 }
726 
CreateSHM(int number,size_t min_size)727 void RTCVideoDecoder::CreateSHM(int number, size_t min_size) {
728   DCHECK(vda_loop_proxy_->BelongsToCurrentThread());
729   DVLOG(2) << "CreateSHM. size=" << min_size;
730   int number_to_allocate;
731   {
732     base::AutoLock auto_lock(lock_);
733     number_to_allocate =
734         std::min(kMaxNumSharedMemorySegments - num_shm_buffers_, number);
735   }
736   size_t size_to_allocate = std::max(min_size, kSharedMemorySegmentBytes);
737   for (int i = 0; i < number_to_allocate; i++) {
738     base::SharedMemory* shm = factories_->CreateSharedMemory(size_to_allocate);
739     if (shm != NULL) {
740       base::AutoLock auto_lock(lock_);
741       num_shm_buffers_++;
742       PutSHM_Locked(
743           scoped_ptr<SHMBuffer>(new SHMBuffer(shm, size_to_allocate)));
744     }
745   }
746   // Kick off the decoding.
747   RequestBufferDecode();
748 }
749 
RecordBufferData(const BufferData & buffer_data)750 void RTCVideoDecoder::RecordBufferData(const BufferData& buffer_data) {
751   input_buffer_data_.push_front(buffer_data);
752   // Why this value?  Because why not.  avformat.h:MAX_REORDER_DELAY is 16, but
753   // that's too small for some pathological B-frame test videos.  The cost of
754   // using too-high a value is low (192 bits per extra slot).
755   static const size_t kMaxInputBufferDataSize = 128;
756   // Pop from the back of the list, because that's the oldest and least likely
757   // to be useful in the future data.
758   if (input_buffer_data_.size() > kMaxInputBufferDataSize)
759     input_buffer_data_.pop_back();
760 }
761 
GetBufferData(int32 bitstream_buffer_id,uint32_t * timestamp,uint32_t * width,uint32_t * height,size_t * size)762 void RTCVideoDecoder::GetBufferData(int32 bitstream_buffer_id,
763                                     uint32_t* timestamp,
764                                     uint32_t* width,
765                                     uint32_t* height,
766                                     size_t* size) {
767   for (std::list<BufferData>::iterator it = input_buffer_data_.begin();
768        it != input_buffer_data_.end();
769        ++it) {
770     if (it->bitstream_buffer_id != bitstream_buffer_id)
771       continue;
772     *timestamp = it->timestamp;
773     *width = it->width;
774     *height = it->height;
775     return;
776   }
777   NOTREACHED() << "Missing bitstream buffer id: " << bitstream_buffer_id;
778 }
779 
RecordInitDecodeUMA(int32_t status)780 int32_t RTCVideoDecoder::RecordInitDecodeUMA(int32_t status) {
781   // Logging boolean is enough to know if HW decoding has been used. Also,
782   // InitDecode is less likely to return an error so enum is not used here.
783   bool sample = (status == WEBRTC_VIDEO_CODEC_OK) ? true : false;
784   UMA_HISTOGRAM_BOOLEAN("Media.RTCVideoDecoderInitDecodeSuccess", sample);
785   return status;
786 }
787 
788 }  // namespace content
789