X-Git-Url: https://main.carlh.net/gitweb/?a=blobdiff_plain;ds=sidebyside;f=src%2Flib%2Fvideo_decoder.cc;h=38b9e8ced058c77d23dd0396a1621837181a809f;hb=aebfa24afe42d80693df66318e5d2818ebf5989b;hp=b861f49e5657ac90ad43be6be6b53454e10a72da;hpb=4dbc6ef917aeceb906b1ef1caf6911033e7e2c54;p=dcpomatic.git diff --git a/src/lib/video_decoder.cc b/src/lib/video_decoder.cc index b861f49e5..38b9e8ced 100644 --- a/src/lib/video_decoder.cc +++ b/src/lib/video_decoder.cc @@ -1,5 +1,5 @@ /* - Copyright (C) 2012-2014 Carl Hetherington + Copyright (C) 2012-2016 Carl Hetherington This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by @@ -19,9 +19,11 @@ #include "video_decoder.h" #include "image.h" -#include "image_proxy.h" #include "raw_image_proxy.h" -#include "content_video.h" +#include "film.h" +#include "log.h" +#include "compose.hpp" +#include #include "i18n.h" @@ -39,18 +41,18 @@ VideoDecoder::VideoDecoder (shared_ptr c) #else : _video_content (c) #endif - , _same (false) , _last_seek_accurate (true) + , _ignore_video (false) { - _black_image.reset (new Image (PIX_FMT_RGB24, _video_content->video_size(), true)); + _black_image.reset (new Image (AV_PIX_FMT_RGB24, _video_content->video_size(), true)); _black_image->make_black (); } list -VideoDecoder::decoded_video (VideoFrame frame) +VideoDecoder::decoded_video (Frame frame) { list output; - + for (list::const_iterator i = _decoded_video.begin(); i != _decoded_video.end(); ++i) { if (i->frame == frame) { output.push_back (*i); @@ -66,13 +68,19 @@ VideoDecoder::decoded_video (VideoFrame frame) * @return Frames; there may be none (if there is no video there), 1 for 2D or 2 for 3D. */ list -VideoDecoder::get_video (VideoFrame frame, bool accurate) +VideoDecoder::get_video (Frame frame, bool accurate) { + if (_no_data_frame && frame >= _no_data_frame.get()) { + return list (); + } + /* At this stage, if we have get_video()ed before, _decoded_video will contain the last frame that this method returned (and possibly a few more). If the requested frame is not in _decoded_video and it is not the next one after the end of _decoded_video we need to seek. */ + _video_content->film()->log()->log (String::compose ("VD has request for %1", frame), LogEntry::TYPE_DEBUG_DECODE); + if (_decoded_video.empty() || frame < _decoded_video.front().frame || frame > (_decoded_video.back().frame + 1)) { seek (ContentTime::from_frames (frame, _video_content->video_frame_rate()), accurate); } @@ -94,8 +102,9 @@ VideoDecoder::get_video (VideoFrame frame, bool accurate) break; } - if (pass ()) { + if (pass (PASS_REASON_VIDEO, accurate)) { /* The decoder has nothing more for us */ + _no_data_frame = frame; break; } @@ -111,24 +120,27 @@ VideoDecoder::get_video (VideoFrame frame, bool accurate) dec = decoded_video (frame); } else { /* Any frame will do: use the first one that comes out of pass() */ - while (_decoded_video.empty() && !pass ()) {} + while (_decoded_video.empty() && !pass (PASS_REASON_VIDEO, accurate)) {} if (!_decoded_video.empty ()) { dec.push_back (_decoded_video.front ()); } } - /* Clean up _decoded_video; keep the frame we are returning (which may have two images + /* Clean up _decoded_video; keep the frame we are returning, if any (which may have two images for 3D), but nothing before that */ - while (!_decoded_video.empty() && _decoded_video.front().frame < dec.front().frame) { + while (!_decoded_video.empty() && !dec.empty() && _decoded_video.front().frame < dec.front().frame) { _decoded_video.pop_front (); } return dec; } -/** Fill _decoded_video from `from' up to, but not including, `to' */ +/** Fill _decoded_video from `from' up to, but not including, `to' with + * a frame for one particular Eyes value (which could be EYES_BOTH, + * EYES_LEFT or EYES_RIGHT) + */ void -VideoDecoder::fill_2d (VideoFrame from, VideoFrame to) +VideoDecoder::fill_one_eye (Frame from, Frame to, Eyes eye) { if (to == 0) { /* Already OK */ @@ -136,7 +148,7 @@ VideoDecoder::fill_2d (VideoFrame from, VideoFrame to) } /* Fill with black... */ - boost::shared_ptr filler_image (new RawImageProxy (_black_image)); + shared_ptr filler_image (new RawImageProxy (_black_image)); Part filler_part = PART_WHOLE; /* ...unless there's some video we can fill with */ @@ -145,24 +157,21 @@ VideoDecoder::fill_2d (VideoFrame from, VideoFrame to) filler_part = _decoded_video.back().part; } - VideoFrame filler_frame = from; - - while (filler_frame < to) { - + for (Frame i = from; i < to; ++i) { #ifdef DCPOMATIC_DEBUG test_gaps++; #endif _decoded_video.push_back ( - ContentVideo (filler_image, EYES_BOTH, filler_part, filler_frame) + ContentVideo (filler_image, eye, filler_part, i) ); - - ++filler_frame; } } -/** Fill _decoded_video from `from' up to, but not including, `to' */ +/** Fill _decoded_video from `from' up to, but not including, `to' + * adding both left and right eye frames. + */ void -VideoDecoder::fill_3d (VideoFrame from, VideoFrame to, Eyes eye) +VideoDecoder::fill_both_eyes (Frame from, Frame to, Eyes eye) { if (to == 0 && eye == EYES_LEFT) { /* Already OK */ @@ -170,8 +179,8 @@ VideoDecoder::fill_3d (VideoFrame from, VideoFrame to, Eyes eye) } /* Fill with black... */ - boost::shared_ptr filler_left_image (new RawImageProxy (_black_image)); - boost::shared_ptr filler_right_image (new RawImageProxy (_black_image)); + shared_ptr filler_left_image (new RawImageProxy (_black_image)); + shared_ptr filler_right_image (new RawImageProxy (_black_image)); Part filler_left_part = PART_WHOLE; Part filler_right_part = PART_WHOLE; @@ -190,7 +199,7 @@ VideoDecoder::fill_3d (VideoFrame from, VideoFrame to, Eyes eye) } } - VideoFrame filler_frame = from; + Frame filler_frame = from; Eyes filler_eye = _decoded_video.empty() ? EYES_LEFT : _decoded_video.back().eyes; if (_decoded_video.empty ()) { @@ -227,15 +236,16 @@ VideoDecoder::fill_3d (VideoFrame from, VideoFrame to, Eyes eye) } } } - + /** Called by subclasses when they have a video frame ready */ void -VideoDecoder::video (shared_ptr image, VideoFrame frame) +VideoDecoder::video (shared_ptr image, Frame frame) { - /* We may receive the same frame index twice for 3D, and we need to know - when that happens. - */ - _same = (!_decoded_video.empty() && frame == _decoded_video.back().frame); + if (_ignore_video) { + return; + } + + _video_content->film()->log()->log (String::compose ("VD receives %1", frame), LogEntry::TYPE_DEBUG_DECODE); /* Work out what we are going to push into _decoded_video next */ list to_push; @@ -244,8 +254,14 @@ VideoDecoder::video (shared_ptr image, VideoFrame frame) to_push.push_back (ContentVideo (image, EYES_BOTH, PART_WHOLE, frame)); break; case VIDEO_FRAME_TYPE_3D_ALTERNATE: - to_push.push_back (ContentVideo (image, _same ? EYES_RIGHT : EYES_LEFT, PART_WHOLE, frame)); + { + /* We receive the same frame index twice for 3D-alternate; hence we know which + frame this one is. + */ + bool const same = (!_decoded_video.empty() && frame == _decoded_video.back().frame); + to_push.push_back (ContentVideo (image, same ? EYES_RIGHT : EYES_LEFT, PART_WHOLE, frame)); break; + } case VIDEO_FRAME_TYPE_3D_LEFT_RIGHT: to_push.push_back (ContentVideo (image, EYES_LEFT, PART_LEFT_HALF, frame)); to_push.push_back (ContentVideo (image, EYES_RIGHT, PART_RIGHT_HALF, frame)); @@ -269,26 +285,54 @@ VideoDecoder::video (shared_ptr image, VideoFrame frame) and the things we are about to push. */ - boost::optional from; - boost::optional to; - + optional from; + optional to; + if (_decoded_video.empty() && _last_seek_time && _last_seek_accurate) { - from = _last_seek_time->frames (_video_content->video_frame_rate ()); + from = _last_seek_time->frames_round (_video_content->video_frame_rate ()); to = to_push.front().frame; } else if (!_decoded_video.empty ()) { from = _decoded_video.back().frame + 1; to = to_push.front().frame; } + /* If we've pre-rolled on a seek we may now receive out-of-order frames + (frames before the last seek time) which we can just ignore. + */ + + if (from && to && from.get() > to.get()) { + return; + } + if (from) { - if (_video_content->video_frame_type() == VIDEO_FRAME_TYPE_2D) { - fill_2d (from.get(), to.get ()); - } else { - fill_3d (from.get(), to.get(), to_push.front().eyes); + switch (_video_content->video_frame_type ()) { + case VIDEO_FRAME_TYPE_2D: + fill_one_eye (from.get(), to.get (), EYES_BOTH); + break; + case VIDEO_FRAME_TYPE_3D_LEFT_RIGHT: + case VIDEO_FRAME_TYPE_3D_TOP_BOTTOM: + case VIDEO_FRAME_TYPE_3D_ALTERNATE: + fill_both_eyes (from.get(), to.get(), to_push.front().eyes); + break; + case VIDEO_FRAME_TYPE_3D_LEFT: + fill_one_eye (from.get(), to.get (), EYES_LEFT); + break; + case VIDEO_FRAME_TYPE_3D_RIGHT: + fill_one_eye (from.get(), to.get (), EYES_RIGHT); + break; } } copy (to_push.begin(), to_push.end(), back_inserter (_decoded_video)); + + /* We can't let this build up too much or we will run out of memory. There is a + `best' value for the allowed size of _decoded_video which balances memory use + with decoding efficiency (lack of seeks). Throwing away video frames here + is not a problem for correctness, so do it. + */ + while (_decoded_video.size() > 96) { + _decoded_video.pop_back (); + } } void @@ -298,3 +342,10 @@ VideoDecoder::seek (ContentTime s, bool accurate) _last_seek_time = s; _last_seek_accurate = accurate; } + +/** Set this player never to produce any video data */ +void +VideoDecoder::set_ignore_video () +{ + _ignore_video = true; +}