• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  *  Copyright (c) 2016 The WebRTC project authors. All Rights Reserved.
3  *
4  *  Use of this source code is governed by a BSD-style license
5  *  that can be found in the LICENSE file in the root of the source
6  *  tree. An additional intellectual property rights grant can be found
7  *  in the file PATENTS.  All contributing project authors may
8  *  be found in the AUTHORS file in the root of the source tree.
9  */
10 
11 #include "modules/video_coding/frame_buffer2.h"
12 
13 #include <algorithm>
14 #include <cstdlib>
15 #include <iterator>
16 #include <queue>
17 #include <utility>
18 #include <vector>
19 
20 #include "absl/memory/memory.h"
21 #include "api/video/encoded_image.h"
22 #include "api/video/video_timing.h"
23 #include "modules/video_coding/include/video_coding_defines.h"
24 #include "modules/video_coding/jitter_estimator.h"
25 #include "modules/video_coding/timing.h"
26 #include "rtc_base/checks.h"
27 #include "rtc_base/experiments/rtt_mult_experiment.h"
28 #include "rtc_base/logging.h"
29 #include "rtc_base/numerics/sequence_number_util.h"
30 #include "rtc_base/trace_event.h"
31 #include "system_wrappers/include/clock.h"
32 #include "system_wrappers/include/field_trial.h"
33 
34 namespace webrtc {
35 namespace video_coding {
36 
37 namespace {
38 // Max number of frames the buffer will hold.
39 constexpr size_t kMaxFramesBuffered = 800;
40 
41 // Max number of decoded frame info that will be saved.
42 constexpr int kMaxFramesHistory = 1 << 13;
43 
44 // The time it's allowed for a frame to be late to its rendering prediction and
45 // still be rendered.
46 constexpr int kMaxAllowedFrameDelayMs = 5;
47 
48 constexpr int64_t kLogNonDecodedIntervalMs = 5000;
49 }  // namespace
50 
FrameBuffer(Clock * clock,VCMTiming * timing,VCMReceiveStatisticsCallback * stats_callback)51 FrameBuffer::FrameBuffer(Clock* clock,
52                          VCMTiming* timing,
53                          VCMReceiveStatisticsCallback* stats_callback)
54     : decoded_frames_history_(kMaxFramesHistory),
55       clock_(clock),
56       callback_queue_(nullptr),
57       jitter_estimator_(clock),
58       timing_(timing),
59       inter_frame_delay_(clock_->TimeInMilliseconds()),
60       stopped_(false),
61       protection_mode_(kProtectionNack),
62       stats_callback_(stats_callback),
63       last_log_non_decoded_ms_(-kLogNonDecodedIntervalMs),
64       add_rtt_to_playout_delay_(
65           webrtc::field_trial::IsEnabled("WebRTC-AddRttToPlayoutDelay")),
66       rtt_mult_settings_(RttMultExperiment::GetRttMultValue()) {
67   callback_checker_.Detach();
68 }
69 
~FrameBuffer()70 FrameBuffer::~FrameBuffer() {
71   RTC_DCHECK_RUN_ON(&construction_checker_);
72 }
73 
NextFrame(int64_t max_wait_time_ms,bool keyframe_required,rtc::TaskQueue * callback_queue,std::function<void (std::unique_ptr<EncodedFrame>,ReturnReason)> handler)74 void FrameBuffer::NextFrame(
75     int64_t max_wait_time_ms,
76     bool keyframe_required,
77     rtc::TaskQueue* callback_queue,
78     std::function<void(std::unique_ptr<EncodedFrame>, ReturnReason)> handler) {
79   RTC_DCHECK_RUN_ON(&callback_checker_);
80   RTC_DCHECK(callback_queue->IsCurrent());
81   TRACE_EVENT0("webrtc", "FrameBuffer::NextFrame");
82   int64_t latest_return_time_ms =
83       clock_->TimeInMilliseconds() + max_wait_time_ms;
84 
85   MutexLock lock(&mutex_);
86   if (stopped_) {
87     return;
88   }
89   latest_return_time_ms_ = latest_return_time_ms;
90   keyframe_required_ = keyframe_required;
91   frame_handler_ = handler;
92   callback_queue_ = callback_queue;
93   StartWaitForNextFrameOnQueue();
94 }
95 
StartWaitForNextFrameOnQueue()96 void FrameBuffer::StartWaitForNextFrameOnQueue() {
97   RTC_DCHECK(callback_queue_);
98   RTC_DCHECK(!callback_task_.Running());
99   int64_t wait_ms = FindNextFrame(clock_->TimeInMilliseconds());
100   callback_task_ = RepeatingTaskHandle::DelayedStart(
101       callback_queue_->Get(), TimeDelta::Millis(wait_ms), [this] {
102         RTC_DCHECK_RUN_ON(&callback_checker_);
103         // If this task has not been cancelled, we did not get any new frames
104         // while waiting. Continue with frame delivery.
105         MutexLock lock(&mutex_);
106         if (!frames_to_decode_.empty()) {
107           // We have frames, deliver!
108           frame_handler_(absl::WrapUnique(GetNextFrame()), kFrameFound);
109           CancelCallback();
110           return TimeDelta::Zero();  // Ignored.
111         } else if (clock_->TimeInMilliseconds() >= latest_return_time_ms_) {
112           // We have timed out, signal this and stop repeating.
113           frame_handler_(nullptr, kTimeout);
114           CancelCallback();
115           return TimeDelta::Zero();  // Ignored.
116         } else {
117           // If there's no frames to decode and there is still time left, it
118           // means that the frame buffer was cleared between creation and
119           // execution of this task. Continue waiting for the remaining time.
120           int64_t wait_ms = FindNextFrame(clock_->TimeInMilliseconds());
121           return TimeDelta::Millis(wait_ms);
122         }
123       });
124 }
125 
FindNextFrame(int64_t now_ms)126 int64_t FrameBuffer::FindNextFrame(int64_t now_ms) {
127   int64_t wait_ms = latest_return_time_ms_ - now_ms;
128   frames_to_decode_.clear();
129 
130   // |last_continuous_frame_| may be empty below, but nullopt is smaller
131   // than everything else and loop will immediately terminate as expected.
132   for (auto frame_it = frames_.begin();
133        frame_it != frames_.end() && frame_it->first <= last_continuous_frame_;
134        ++frame_it) {
135     if (!frame_it->second.continuous ||
136         frame_it->second.num_missing_decodable > 0) {
137       continue;
138     }
139 
140     EncodedFrame* frame = frame_it->second.frame.get();
141 
142     if (keyframe_required_ && !frame->is_keyframe())
143       continue;
144 
145     auto last_decoded_frame_timestamp =
146         decoded_frames_history_.GetLastDecodedFrameTimestamp();
147 
148     // TODO(https://bugs.webrtc.org/9974): consider removing this check
149     // as it may make a stream undecodable after a very long delay between
150     // frames.
151     if (last_decoded_frame_timestamp &&
152         AheadOf(*last_decoded_frame_timestamp, frame->Timestamp())) {
153       continue;
154     }
155 
156     // Only ever return all parts of a superframe. Therefore skip this
157     // frame if it's not a beginning of a superframe.
158     if (frame->inter_layer_predicted) {
159       continue;
160     }
161 
162     // Gather all remaining frames for the same superframe.
163     std::vector<FrameMap::iterator> current_superframe;
164     current_superframe.push_back(frame_it);
165     bool last_layer_completed = frame_it->second.frame->is_last_spatial_layer;
166     FrameMap::iterator next_frame_it = frame_it;
167     while (true) {
168       ++next_frame_it;
169       if (next_frame_it == frames_.end() ||
170           next_frame_it->first.picture_id != frame->id.picture_id ||
171           !next_frame_it->second.continuous) {
172         break;
173       }
174       // Check if the next frame has some undecoded references other than
175       // the previous frame in the same superframe.
176       size_t num_allowed_undecoded_refs =
177           (next_frame_it->second.frame->inter_layer_predicted) ? 1 : 0;
178       if (next_frame_it->second.num_missing_decodable >
179           num_allowed_undecoded_refs) {
180         break;
181       }
182       // All frames in the superframe should have the same timestamp.
183       if (frame->Timestamp() != next_frame_it->second.frame->Timestamp()) {
184         RTC_LOG(LS_WARNING) << "Frames in a single superframe have different"
185                                " timestamps. Skipping undecodable superframe.";
186         break;
187       }
188       current_superframe.push_back(next_frame_it);
189       last_layer_completed = next_frame_it->second.frame->is_last_spatial_layer;
190     }
191     // Check if the current superframe is complete.
192     // TODO(bugs.webrtc.org/10064): consider returning all available to
193     // decode frames even if the superframe is not complete yet.
194     if (!last_layer_completed) {
195       continue;
196     }
197 
198     frames_to_decode_ = std::move(current_superframe);
199 
200     if (frame->RenderTime() == -1) {
201       frame->SetRenderTime(timing_->RenderTimeMs(frame->Timestamp(), now_ms));
202     }
203     wait_ms = timing_->MaxWaitingTime(frame->RenderTime(), now_ms);
204 
205     // This will cause the frame buffer to prefer high framerate rather
206     // than high resolution in the case of the decoder not decoding fast
207     // enough and the stream has multiple spatial and temporal layers.
208     // For multiple temporal layers it may cause non-base layer frames to be
209     // skipped if they are late.
210     if (wait_ms < -kMaxAllowedFrameDelayMs)
211       continue;
212 
213     break;
214   }
215   wait_ms = std::min<int64_t>(wait_ms, latest_return_time_ms_ - now_ms);
216   wait_ms = std::max<int64_t>(wait_ms, 0);
217   return wait_ms;
218 }
219 
GetNextFrame()220 EncodedFrame* FrameBuffer::GetNextFrame() {
221   RTC_DCHECK_RUN_ON(&callback_checker_);
222   int64_t now_ms = clock_->TimeInMilliseconds();
223   // TODO(ilnik): remove |frames_out| use frames_to_decode_ directly.
224   std::vector<EncodedFrame*> frames_out;
225 
226   RTC_DCHECK(!frames_to_decode_.empty());
227   bool superframe_delayed_by_retransmission = false;
228   size_t superframe_size = 0;
229   EncodedFrame* first_frame = frames_to_decode_[0]->second.frame.get();
230   int64_t render_time_ms = first_frame->RenderTime();
231   int64_t receive_time_ms = first_frame->ReceivedTime();
232   // Gracefully handle bad RTP timestamps and render time issues.
233   if (HasBadRenderTiming(*first_frame, now_ms)) {
234     jitter_estimator_.Reset();
235     timing_->Reset();
236     render_time_ms = timing_->RenderTimeMs(first_frame->Timestamp(), now_ms);
237   }
238 
239   for (FrameMap::iterator& frame_it : frames_to_decode_) {
240     RTC_DCHECK(frame_it != frames_.end());
241     EncodedFrame* frame = frame_it->second.frame.release();
242 
243     frame->SetRenderTime(render_time_ms);
244 
245     superframe_delayed_by_retransmission |= frame->delayed_by_retransmission();
246     receive_time_ms = std::max(receive_time_ms, frame->ReceivedTime());
247     superframe_size += frame->size();
248 
249     PropagateDecodability(frame_it->second);
250     decoded_frames_history_.InsertDecoded(frame_it->first, frame->Timestamp());
251 
252     // Remove decoded frame and all undecoded frames before it.
253     if (stats_callback_) {
254       unsigned int dropped_frames = std::count_if(
255           frames_.begin(), frame_it,
256           [](const std::pair<const VideoLayerFrameId, FrameInfo>& frame) {
257             return frame.second.frame != nullptr;
258           });
259       if (dropped_frames > 0) {
260         stats_callback_->OnDroppedFrames(dropped_frames);
261       }
262     }
263 
264     frames_.erase(frames_.begin(), ++frame_it);
265 
266     frames_out.push_back(frame);
267   }
268 
269   if (!superframe_delayed_by_retransmission) {
270     int64_t frame_delay;
271 
272     if (inter_frame_delay_.CalculateDelay(first_frame->Timestamp(),
273                                           &frame_delay, receive_time_ms)) {
274       jitter_estimator_.UpdateEstimate(frame_delay, superframe_size);
275     }
276 
277     float rtt_mult = protection_mode_ == kProtectionNackFEC ? 0.0 : 1.0;
278     absl::optional<float> rtt_mult_add_cap_ms = absl::nullopt;
279     if (rtt_mult_settings_.has_value()) {
280       rtt_mult = rtt_mult_settings_->rtt_mult_setting;
281       rtt_mult_add_cap_ms = rtt_mult_settings_->rtt_mult_add_cap_ms;
282     }
283     timing_->SetJitterDelay(
284         jitter_estimator_.GetJitterEstimate(rtt_mult, rtt_mult_add_cap_ms));
285     timing_->UpdateCurrentDelay(render_time_ms, now_ms);
286   } else {
287     if (RttMultExperiment::RttMultEnabled() || add_rtt_to_playout_delay_)
288       jitter_estimator_.FrameNacked();
289   }
290 
291   UpdateJitterDelay();
292   UpdateTimingFrameInfo();
293 
294   if (frames_out.size() == 1) {
295     return frames_out[0];
296   } else {
297     return CombineAndDeleteFrames(frames_out);
298   }
299 }
300 
HasBadRenderTiming(const EncodedFrame & frame,int64_t now_ms)301 bool FrameBuffer::HasBadRenderTiming(const EncodedFrame& frame,
302                                      int64_t now_ms) {
303   // Assume that render timing errors are due to changes in the video stream.
304   int64_t render_time_ms = frame.RenderTimeMs();
305   // Zero render time means render immediately.
306   if (render_time_ms == 0) {
307     return false;
308   }
309   if (render_time_ms < 0) {
310     return true;
311   }
312   const int64_t kMaxVideoDelayMs = 10000;
313   if (std::abs(render_time_ms - now_ms) > kMaxVideoDelayMs) {
314     int frame_delay = static_cast<int>(std::abs(render_time_ms - now_ms));
315     RTC_LOG(LS_WARNING)
316         << "A frame about to be decoded is out of the configured "
317            "delay bounds ("
318         << frame_delay << " > " << kMaxVideoDelayMs
319         << "). Resetting the video jitter buffer.";
320     return true;
321   }
322   if (static_cast<int>(timing_->TargetVideoDelay()) > kMaxVideoDelayMs) {
323     RTC_LOG(LS_WARNING) << "The video target delay has grown larger than "
324                         << kMaxVideoDelayMs << " ms.";
325     return true;
326   }
327   return false;
328 }
329 
SetProtectionMode(VCMVideoProtection mode)330 void FrameBuffer::SetProtectionMode(VCMVideoProtection mode) {
331   TRACE_EVENT0("webrtc", "FrameBuffer::SetProtectionMode");
332   MutexLock lock(&mutex_);
333   protection_mode_ = mode;
334 }
335 
Start()336 void FrameBuffer::Start() {
337   TRACE_EVENT0("webrtc", "FrameBuffer::Start");
338   MutexLock lock(&mutex_);
339   stopped_ = false;
340 }
341 
Stop()342 void FrameBuffer::Stop() {
343   TRACE_EVENT0("webrtc", "FrameBuffer::Stop");
344   MutexLock lock(&mutex_);
345   if (stopped_)
346     return;
347   stopped_ = true;
348 
349   CancelCallback();
350 }
351 
Clear()352 void FrameBuffer::Clear() {
353   MutexLock lock(&mutex_);
354   ClearFramesAndHistory();
355 }
356 
UpdateRtt(int64_t rtt_ms)357 void FrameBuffer::UpdateRtt(int64_t rtt_ms) {
358   MutexLock lock(&mutex_);
359   jitter_estimator_.UpdateRtt(rtt_ms);
360 }
361 
ValidReferences(const EncodedFrame & frame) const362 bool FrameBuffer::ValidReferences(const EncodedFrame& frame) const {
363   for (size_t i = 0; i < frame.num_references; ++i) {
364     if (frame.references[i] >= frame.id.picture_id)
365       return false;
366 
367     for (size_t j = i + 1; j < frame.num_references; ++j) {
368       if (frame.references[i] == frame.references[j])
369         return false;
370     }
371   }
372 
373   if (frame.inter_layer_predicted && frame.id.spatial_layer == 0)
374     return false;
375 
376   return true;
377 }
378 
CancelCallback()379 void FrameBuffer::CancelCallback() {
380   // Called from the callback queue or from within Stop().
381   frame_handler_ = {};
382   callback_task_.Stop();
383   callback_queue_ = nullptr;
384   callback_checker_.Detach();
385 }
386 
IsCompleteSuperFrame(const EncodedFrame & frame)387 bool FrameBuffer::IsCompleteSuperFrame(const EncodedFrame& frame) {
388   if (frame.inter_layer_predicted) {
389     // Check that all previous spatial layers are already inserted.
390     VideoLayerFrameId id = frame.id;
391     RTC_DCHECK_GT(id.spatial_layer, 0);
392     --id.spatial_layer;
393     FrameMap::iterator prev_frame = frames_.find(id);
394     if (prev_frame == frames_.end() || !prev_frame->second.frame)
395       return false;
396     while (prev_frame->second.frame->inter_layer_predicted) {
397       if (prev_frame == frames_.begin())
398         return false;
399       --prev_frame;
400       --id.spatial_layer;
401       if (!prev_frame->second.frame ||
402           prev_frame->first.picture_id != id.picture_id ||
403           prev_frame->first.spatial_layer != id.spatial_layer) {
404         return false;
405       }
406     }
407   }
408 
409   if (!frame.is_last_spatial_layer) {
410     // Check that all following spatial layers are already inserted.
411     VideoLayerFrameId id = frame.id;
412     ++id.spatial_layer;
413     FrameMap::iterator next_frame = frames_.find(id);
414     if (next_frame == frames_.end() || !next_frame->second.frame)
415       return false;
416     while (!next_frame->second.frame->is_last_spatial_layer) {
417       ++next_frame;
418       ++id.spatial_layer;
419       if (next_frame == frames_.end() || !next_frame->second.frame ||
420           next_frame->first.picture_id != id.picture_id ||
421           next_frame->first.spatial_layer != id.spatial_layer) {
422         return false;
423       }
424     }
425   }
426 
427   return true;
428 }
429 
InsertFrame(std::unique_ptr<EncodedFrame> frame)430 int64_t FrameBuffer::InsertFrame(std::unique_ptr<EncodedFrame> frame) {
431   TRACE_EVENT0("webrtc", "FrameBuffer::InsertFrame");
432   RTC_DCHECK(frame);
433 
434   MutexLock lock(&mutex_);
435 
436   const VideoLayerFrameId& id = frame->id;
437   int64_t last_continuous_picture_id =
438       !last_continuous_frame_ ? -1 : last_continuous_frame_->picture_id;
439 
440   if (!ValidReferences(*frame)) {
441     RTC_LOG(LS_WARNING) << "Frame with (picture_id:spatial_id) ("
442                         << id.picture_id << ":"
443                         << static_cast<int>(id.spatial_layer)
444                         << ") has invalid frame references, dropping frame.";
445     return last_continuous_picture_id;
446   }
447 
448   if (frames_.size() >= kMaxFramesBuffered) {
449     if (frame->is_keyframe()) {
450       RTC_LOG(LS_WARNING) << "Inserting keyframe (picture_id:spatial_id) ("
451                           << id.picture_id << ":"
452                           << static_cast<int>(id.spatial_layer)
453                           << ") but buffer is full, clearing"
454                              " buffer and inserting the frame.";
455       ClearFramesAndHistory();
456     } else {
457       RTC_LOG(LS_WARNING) << "Frame with (picture_id:spatial_id) ("
458                           << id.picture_id << ":"
459                           << static_cast<int>(id.spatial_layer)
460                           << ") could not be inserted due to the frame "
461                              "buffer being full, dropping frame.";
462       return last_continuous_picture_id;
463     }
464   }
465 
466   auto last_decoded_frame = decoded_frames_history_.GetLastDecodedFrameId();
467   auto last_decoded_frame_timestamp =
468       decoded_frames_history_.GetLastDecodedFrameTimestamp();
469   if (last_decoded_frame && id <= *last_decoded_frame) {
470     if (AheadOf(frame->Timestamp(), *last_decoded_frame_timestamp) &&
471         frame->is_keyframe()) {
472       // If this frame has a newer timestamp but an earlier picture id then we
473       // assume there has been a jump in the picture id due to some encoder
474       // reconfiguration or some other reason. Even though this is not according
475       // to spec we can still continue to decode from this frame if it is a
476       // keyframe.
477       RTC_LOG(LS_WARNING)
478           << "A jump in picture id was detected, clearing buffer.";
479       ClearFramesAndHistory();
480       last_continuous_picture_id = -1;
481     } else {
482       RTC_LOG(LS_WARNING) << "Frame with (picture_id:spatial_id) ("
483                           << id.picture_id << ":"
484                           << static_cast<int>(id.spatial_layer)
485                           << ") inserted after frame ("
486                           << last_decoded_frame->picture_id << ":"
487                           << static_cast<int>(last_decoded_frame->spatial_layer)
488                           << ") was handed off for decoding, dropping frame.";
489       return last_continuous_picture_id;
490     }
491   }
492 
493   // Test if inserting this frame would cause the order of the frames to become
494   // ambiguous (covering more than half the interval of 2^16). This can happen
495   // when the picture id make large jumps mid stream.
496   if (!frames_.empty() && id < frames_.begin()->first &&
497       frames_.rbegin()->first < id) {
498     RTC_LOG(LS_WARNING)
499         << "A jump in picture id was detected, clearing buffer.";
500     ClearFramesAndHistory();
501     last_continuous_picture_id = -1;
502   }
503 
504   auto info = frames_.emplace(id, FrameInfo()).first;
505 
506   if (info->second.frame) {
507     return last_continuous_picture_id;
508   }
509 
510   if (!UpdateFrameInfoWithIncomingFrame(*frame, info))
511     return last_continuous_picture_id;
512 
513   if (!frame->delayed_by_retransmission())
514     timing_->IncomingTimestamp(frame->Timestamp(), frame->ReceivedTime());
515 
516   if (stats_callback_ && IsCompleteSuperFrame(*frame)) {
517     stats_callback_->OnCompleteFrame(frame->is_keyframe(), frame->size(),
518                                      frame->contentType());
519   }
520 
521   info->second.frame = std::move(frame);
522 
523   if (info->second.num_missing_continuous == 0) {
524     info->second.continuous = true;
525     PropagateContinuity(info);
526     last_continuous_picture_id = last_continuous_frame_->picture_id;
527 
528     // Since we now have new continuous frames there might be a better frame
529     // to return from NextFrame.
530     if (callback_queue_) {
531       callback_queue_->PostTask([this] {
532         MutexLock lock(&mutex_);
533         if (!callback_task_.Running())
534           return;
535         RTC_CHECK(frame_handler_);
536         callback_task_.Stop();
537         StartWaitForNextFrameOnQueue();
538       });
539     }
540   }
541 
542   return last_continuous_picture_id;
543 }
544 
PropagateContinuity(FrameMap::iterator start)545 void FrameBuffer::PropagateContinuity(FrameMap::iterator start) {
546   TRACE_EVENT0("webrtc", "FrameBuffer::PropagateContinuity");
547   RTC_DCHECK(start->second.continuous);
548 
549   std::queue<FrameMap::iterator> continuous_frames;
550   continuous_frames.push(start);
551 
552   // A simple BFS to traverse continuous frames.
553   while (!continuous_frames.empty()) {
554     auto frame = continuous_frames.front();
555     continuous_frames.pop();
556 
557     if (!last_continuous_frame_ || *last_continuous_frame_ < frame->first) {
558       last_continuous_frame_ = frame->first;
559     }
560 
561     // Loop through all dependent frames, and if that frame no longer has
562     // any unfulfilled dependencies then that frame is continuous as well.
563     for (size_t d = 0; d < frame->second.dependent_frames.size(); ++d) {
564       auto frame_ref = frames_.find(frame->second.dependent_frames[d]);
565       RTC_DCHECK(frame_ref != frames_.end());
566 
567       // TODO(philipel): Look into why we've seen this happen.
568       if (frame_ref != frames_.end()) {
569         --frame_ref->second.num_missing_continuous;
570         if (frame_ref->second.num_missing_continuous == 0) {
571           frame_ref->second.continuous = true;
572           continuous_frames.push(frame_ref);
573         }
574       }
575     }
576   }
577 }
578 
PropagateDecodability(const FrameInfo & info)579 void FrameBuffer::PropagateDecodability(const FrameInfo& info) {
580   TRACE_EVENT0("webrtc", "FrameBuffer::PropagateDecodability");
581   for (size_t d = 0; d < info.dependent_frames.size(); ++d) {
582     auto ref_info = frames_.find(info.dependent_frames[d]);
583     RTC_DCHECK(ref_info != frames_.end());
584     // TODO(philipel): Look into why we've seen this happen.
585     if (ref_info != frames_.end()) {
586       RTC_DCHECK_GT(ref_info->second.num_missing_decodable, 0U);
587       --ref_info->second.num_missing_decodable;
588     }
589   }
590 }
591 
UpdateFrameInfoWithIncomingFrame(const EncodedFrame & frame,FrameMap::iterator info)592 bool FrameBuffer::UpdateFrameInfoWithIncomingFrame(const EncodedFrame& frame,
593                                                    FrameMap::iterator info) {
594   TRACE_EVENT0("webrtc", "FrameBuffer::UpdateFrameInfoWithIncomingFrame");
595   const VideoLayerFrameId& id = frame.id;
596 
597   auto last_decoded_frame = decoded_frames_history_.GetLastDecodedFrameId();
598   RTC_DCHECK(!last_decoded_frame || *last_decoded_frame < info->first);
599 
600   // In this function we determine how many missing dependencies this |frame|
601   // has to become continuous/decodable. If a frame that this |frame| depend
602   // on has already been decoded then we can ignore that dependency since it has
603   // already been fulfilled.
604   //
605   // For all other frames we will register a backwards reference to this |frame|
606   // so that |num_missing_continuous| and |num_missing_decodable| can be
607   // decremented as frames become continuous/are decoded.
608   struct Dependency {
609     VideoLayerFrameId id;
610     bool continuous;
611   };
612   std::vector<Dependency> not_yet_fulfilled_dependencies;
613 
614   // Find all dependencies that have not yet been fulfilled.
615   for (size_t i = 0; i < frame.num_references; ++i) {
616     VideoLayerFrameId ref_key(frame.references[i], frame.id.spatial_layer);
617     // Does |frame| depend on a frame earlier than the last decoded one?
618     if (last_decoded_frame && ref_key <= *last_decoded_frame) {
619       // Was that frame decoded? If not, this |frame| will never become
620       // decodable.
621       if (!decoded_frames_history_.WasDecoded(ref_key)) {
622         int64_t now_ms = clock_->TimeInMilliseconds();
623         if (last_log_non_decoded_ms_ + kLogNonDecodedIntervalMs < now_ms) {
624           RTC_LOG(LS_WARNING)
625               << "Frame with (picture_id:spatial_id) (" << id.picture_id << ":"
626               << static_cast<int>(id.spatial_layer)
627               << ") depends on a non-decoded frame more previous than"
628                  " the last decoded frame, dropping frame.";
629           last_log_non_decoded_ms_ = now_ms;
630         }
631         return false;
632       }
633     } else {
634       auto ref_info = frames_.find(ref_key);
635       bool ref_continuous =
636           ref_info != frames_.end() && ref_info->second.continuous;
637       not_yet_fulfilled_dependencies.push_back({ref_key, ref_continuous});
638     }
639   }
640 
641   // Does |frame| depend on the lower spatial layer?
642   if (frame.inter_layer_predicted) {
643     VideoLayerFrameId ref_key(frame.id.picture_id, frame.id.spatial_layer - 1);
644     auto ref_info = frames_.find(ref_key);
645 
646     bool lower_layer_decoded =
647         last_decoded_frame && *last_decoded_frame == ref_key;
648     bool lower_layer_continuous =
649         lower_layer_decoded ||
650         (ref_info != frames_.end() && ref_info->second.continuous);
651 
652     if (!lower_layer_continuous || !lower_layer_decoded) {
653       not_yet_fulfilled_dependencies.push_back(
654           {ref_key, lower_layer_continuous});
655     }
656   }
657 
658   info->second.num_missing_continuous = not_yet_fulfilled_dependencies.size();
659   info->second.num_missing_decodable = not_yet_fulfilled_dependencies.size();
660 
661   for (const Dependency& dep : not_yet_fulfilled_dependencies) {
662     if (dep.continuous)
663       --info->second.num_missing_continuous;
664 
665     frames_[dep.id].dependent_frames.push_back(id);
666   }
667 
668   return true;
669 }
670 
UpdateJitterDelay()671 void FrameBuffer::UpdateJitterDelay() {
672   TRACE_EVENT0("webrtc", "FrameBuffer::UpdateJitterDelay");
673   if (!stats_callback_)
674     return;
675 
676   int max_decode_ms;
677   int current_delay_ms;
678   int target_delay_ms;
679   int jitter_buffer_ms;
680   int min_playout_delay_ms;
681   int render_delay_ms;
682   if (timing_->GetTimings(&max_decode_ms, &current_delay_ms, &target_delay_ms,
683                           &jitter_buffer_ms, &min_playout_delay_ms,
684                           &render_delay_ms)) {
685     stats_callback_->OnFrameBufferTimingsUpdated(
686         max_decode_ms, current_delay_ms, target_delay_ms, jitter_buffer_ms,
687         min_playout_delay_ms, render_delay_ms);
688   }
689 }
690 
UpdateTimingFrameInfo()691 void FrameBuffer::UpdateTimingFrameInfo() {
692   TRACE_EVENT0("webrtc", "FrameBuffer::UpdateTimingFrameInfo");
693   absl::optional<TimingFrameInfo> info = timing_->GetTimingFrameInfo();
694   if (info && stats_callback_)
695     stats_callback_->OnTimingFrameInfoUpdated(*info);
696 }
697 
ClearFramesAndHistory()698 void FrameBuffer::ClearFramesAndHistory() {
699   TRACE_EVENT0("webrtc", "FrameBuffer::ClearFramesAndHistory");
700   if (stats_callback_) {
701     unsigned int dropped_frames = std::count_if(
702         frames_.begin(), frames_.end(),
703         [](const std::pair<const VideoLayerFrameId, FrameInfo>& frame) {
704           return frame.second.frame != nullptr;
705         });
706     if (dropped_frames > 0) {
707       stats_callback_->OnDroppedFrames(dropped_frames);
708     }
709   }
710   frames_.clear();
711   last_continuous_frame_.reset();
712   frames_to_decode_.clear();
713   decoded_frames_history_.Clear();
714 }
715 
716 // TODO(philipel): Avoid the concatenation of frames here, by replacing
717 // NextFrame and GetNextFrame with methods returning multiple frames.
CombineAndDeleteFrames(const std::vector<EncodedFrame * > & frames) const718 EncodedFrame* FrameBuffer::CombineAndDeleteFrames(
719     const std::vector<EncodedFrame*>& frames) const {
720   RTC_DCHECK(!frames.empty());
721   EncodedFrame* first_frame = frames[0];
722   EncodedFrame* last_frame = frames.back();
723   size_t total_length = 0;
724   for (size_t i = 0; i < frames.size(); ++i) {
725     total_length += frames[i]->size();
726   }
727   auto encoded_image_buffer = EncodedImageBuffer::Create(total_length);
728   uint8_t* buffer = encoded_image_buffer->data();
729   first_frame->SetSpatialLayerFrameSize(first_frame->id.spatial_layer,
730                                         first_frame->size());
731   memcpy(buffer, first_frame->data(), first_frame->size());
732   buffer += first_frame->size();
733 
734   // Spatial index of combined frame is set equal to spatial index of its top
735   // spatial layer.
736   first_frame->SetSpatialIndex(last_frame->id.spatial_layer);
737   first_frame->id.spatial_layer = last_frame->id.spatial_layer;
738 
739   first_frame->video_timing_mutable()->network2_timestamp_ms =
740       last_frame->video_timing().network2_timestamp_ms;
741   first_frame->video_timing_mutable()->receive_finish_ms =
742       last_frame->video_timing().receive_finish_ms;
743 
744   // Append all remaining frames to the first one.
745   for (size_t i = 1; i < frames.size(); ++i) {
746     EncodedFrame* next_frame = frames[i];
747     first_frame->SetSpatialLayerFrameSize(next_frame->id.spatial_layer,
748                                           next_frame->size());
749     memcpy(buffer, next_frame->data(), next_frame->size());
750     buffer += next_frame->size();
751     delete next_frame;
752   }
753   first_frame->SetEncodedData(encoded_image_buffer);
754   return first_frame;
755 }
756 
757 FrameBuffer::FrameInfo::FrameInfo() = default;
758 FrameBuffer::FrameInfo::FrameInfo(FrameInfo&&) = default;
759 FrameBuffer::FrameInfo::~FrameInfo() = default;
760 
761 }  // namespace video_coding
762 }  // namespace webrtc
763