X-Git-Url: https://main.carlh.net/gitweb/?a=blobdiff_plain;f=src%2Flib%2Fvideo_decoder.cc;h=f68f0815c1f156c75421d853049e712b21c337be;hb=6f344b876689a1234a5eb75041882f06f5d9fe5c;hp=88f88c1296a16c35c833c36172cc438fdaa61804;hpb=aeb835a18c8df347e0ed68fb24631b320abeb611;p=dcpomatic.git diff --git a/src/lib/video_decoder.cc b/src/lib/video_decoder.cc index 88f88c129..f68f0815c 100644 --- a/src/lib/video_decoder.cc +++ b/src/lib/video_decoder.cc @@ -1,5 +1,5 @@ /* - Copyright (C) 2012-2015 Carl Hetherington + Copyright (C) 2012-2016 Carl Hetherington This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by @@ -34,17 +34,18 @@ using std::back_inserter; using boost::shared_ptr; using boost::optional; -VideoDecoder::VideoDecoder (shared_ptr c) +VideoDecoder::VideoDecoder (shared_ptr c, shared_ptr log) #ifdef DCPOMATIC_DEBUG : test_gaps (0) , _video_content (c) #else : _video_content (c) #endif + , _log (log) , _last_seek_accurate (true) , _ignore_video (false) { - _black_image.reset (new Image (PIX_FMT_RGB24, _video_content->video_size(), true)); + _black_image.reset (new Image (AV_PIX_FMT_RGB24, _video_content->video_size(), true)); _black_image->make_black (); } @@ -70,12 +71,16 @@ VideoDecoder::decoded_video (Frame frame) list VideoDecoder::get_video (Frame frame, bool accurate) { + if (_no_data_frame && frame >= _no_data_frame.get()) { + return list (); + } + /* At this stage, if we have get_video()ed before, _decoded_video will contain the last frame that this method returned (and possibly a few more). If the requested frame is not in _decoded_video and it is not the next one after the end of _decoded_video we need to seek. */ - _video_content->film()->log()->log (String::compose ("VD has request for %1", frame), LogEntry::TYPE_DEBUG_DECODE); + _log->log (String::compose ("VD has request for %1", frame), LogEntry::TYPE_DEBUG_DECODE); if (_decoded_video.empty() || frame < _decoded_video.front().frame || frame > (_decoded_video.back().frame + 1)) { seek (ContentTime::from_frames (frame, _video_content->video_frame_rate()), accurate); @@ -92,14 +97,17 @@ VideoDecoder::get_video (Frame frame, bool accurate) /* We are being accurate, so we want the right frame. * This could all be one statement but it's split up for clarity. */ + bool no_data = false; + while (true) { if (!decoded_video(frame).empty ()) { /* We got what we want */ break; } - if (pass ()) { + if (pass (PASS_REASON_VIDEO, accurate)) { /* The decoder has nothing more for us */ + no_data = true; break; } @@ -113,17 +121,22 @@ VideoDecoder::get_video (Frame frame, bool accurate) } dec = decoded_video (frame); + + if (no_data && dec.empty()) { + _no_data_frame = frame; + } + } else { /* Any frame will do: use the first one that comes out of pass() */ - while (_decoded_video.empty() && !pass ()) {} + while (_decoded_video.empty() && !pass (PASS_REASON_VIDEO, accurate)) {} if (!_decoded_video.empty ()) { dec.push_back (_decoded_video.front ()); } } - /* Clean up _decoded_video; keep the frame we are returning (which may have two images + /* Clean up _decoded_video; keep the frame we are returning, if any (which may have two images for 3D), but nothing before that */ - while (!_decoded_video.empty() && _decoded_video.front().frame < dec.front().frame) { + while (!_decoded_video.empty() && !dec.empty() && _decoded_video.front().frame < dec.front().frame) { _decoded_video.pop_front (); } @@ -240,12 +253,7 @@ VideoDecoder::video (shared_ptr image, Frame frame) return; } - _video_content->film()->log()->log (String::compose ("VD receives %1", frame), LogEntry::TYPE_DEBUG_DECODE); - - /* We may receive the same frame index twice for 3D, and we need to know - when that happens. - */ - bool const same = (!_decoded_video.empty() && frame == _decoded_video.back().frame); + _log->log (String::compose ("VD receives %1", frame), LogEntry::TYPE_DEBUG_DECODE); /* Work out what we are going to push into _decoded_video next */ list to_push; @@ -254,8 +262,14 @@ VideoDecoder::video (shared_ptr image, Frame frame) to_push.push_back (ContentVideo (image, EYES_BOTH, PART_WHOLE, frame)); break; case VIDEO_FRAME_TYPE_3D_ALTERNATE: + { + /* We receive the same frame index twice for 3D-alternate; hence we know which + frame this one is. + */ + bool const same = (!_decoded_video.empty() && frame == _decoded_video.back().frame); to_push.push_back (ContentVideo (image, same ? EYES_RIGHT : EYES_LEFT, PART_WHOLE, frame)); break; + } case VIDEO_FRAME_TYPE_3D_LEFT_RIGHT: to_push.push_back (ContentVideo (image, EYES_LEFT, PART_LEFT_HALF, frame)); to_push.push_back (ContentVideo (image, EYES_RIGHT, PART_RIGHT_HALF, frame)); @@ -313,15 +327,20 @@ VideoDecoder::video (shared_ptr image, Frame frame) break; case VIDEO_FRAME_TYPE_3D_RIGHT: fill_one_eye (from.get(), to.get (), EYES_RIGHT); + break; } } copy (to_push.begin(), to_push.end(), back_inserter (_decoded_video)); - /* We can't let this build up too much or we will run out of memory. We need to allow - the most frames that can exist between blocks of sound in a multiplexed file. + /* We can't let this build up too much or we will run out of memory. There is a + `best' value for the allowed size of _decoded_video which balances memory use + with decoding efficiency (lack of seeks). Throwing away video frames here + is not a problem for correctness, so do it. */ - DCPOMATIC_ASSERT (_decoded_video.size() <= 96); + while (_decoded_video.size() > 96) { + _decoded_video.pop_back (); + } } void