/*
- Copyright (C) 2012-2014 Carl Hetherington <cth@carlh.net>
+ Copyright (C) 2012-2021 Carl Hetherington <cth@carlh.net>
- This program is free software; you can redistribute it and/or modify
+ This file is part of DCP-o-matic.
+
+ DCP-o-matic is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
- This program is distributed in the hope that it will be useful,
+ DCP-o-matic is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
- along with this program; if not, write to the Free Software
- Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
+ along with DCP-o-matic. If not, see <http://www.gnu.org/licenses/>.
*/
-#include "video_decoder.h"
+
+#include "compose.hpp"
+#include "frame_interval_checker.h"
#include "image.h"
-#include "image_proxy.h"
-#include "content_video.h"
+#include "j2k_image_proxy.h"
+#include "log.h"
+#include "raw_image_proxy.h"
+#include "video_decoder.h"
+#include <iostream>
#include "i18n.h"
+
using std::cout;
-using std::list;
-using std::max;
-using boost::shared_ptr;
-using boost::optional;
-
-VideoDecoder::VideoDecoder (shared_ptr<const VideoContent> c)
-#ifdef DCPOMATIC_DEBUG
- : test_gaps (0)
- , _video_content (c)
-#else
- : _video_content (c)
-#endif
- , _same (false)
-{
+using std::dynamic_pointer_cast;
+using std::shared_ptr;
+using namespace dcpomatic;
-}
-list<ContentVideo>
-VideoDecoder::decoded_video (VideoFrame frame)
+VideoDecoder::VideoDecoder (Decoder* parent, shared_ptr<const Content> c)
+ : DecoderPart (parent)
+ , _content (c)
+ , _frame_interval_checker (new FrameIntervalChecker())
{
- list<ContentVideo> output;
-
- for (list<ContentVideo>::const_iterator i = _decoded_video.begin(); i != _decoded_video.end(); ++i) {
- if (i->frame == frame) {
- output.push_back (*i);
- }
- }
- return output;
}
-/** Get all frames which exist in the content at a given frame index.
- * @param frame Frame index.
- * @param accurate true to try hard to return frames at the precise time that was requested, otherwise frames nearby may be returned.
- * @return Frames; there may be none (if there is no video there), 1 for 2D or 2 for 3D.
- */
-list<ContentVideo>
-VideoDecoder::get_video (VideoFrame frame, bool accurate)
+
+/** Called by decoder classes when they have a video frame ready */
+void
+VideoDecoder::emit(shared_ptr<const Film> film, shared_ptr<const ImageProxy> image, ContentTime time)
{
- /* At this stage, if we have get_video()ed before, _decoded_video will contain the last frame that this
- method returned (and possibly a few more). If the requested frame is not in _decoded_video and it is not the next
- one after the end of _decoded_video we need to seek.
- */
-
- if (_decoded_video.empty() || frame < _decoded_video.front().frame || frame > (_decoded_video.back().frame + 1)) {
- seek (ContentTime::from_frames (frame, _video_content->video_frame_rate()), accurate);
+ if (ignore ()) {
+ return;
}
- list<ContentVideo> dec;
+ auto const afr = _content->active_video_frame_rate(film);
+ auto const vft = _content->video->frame_type();
- /* Now enough pass() calls should either:
- * (a) give us what we want, or
- * (b) give us something after what we want, indicating that we will never get what we want, or
- * (c) hit the end of the decoder.
+ /* Do some heuristics to try and spot the case where the user sets content to 3D
+ * when it is not. We try to tell this by looking at the differences in time between
+ * the first few frames. Real 3D content should have two frames for each timestamp.
*/
- if (accurate) {
- /* We are being accurate, so we want the right frame.
- * This could all be one statement but it's split up for clarity.
- */
- while (true) {
- if (!decoded_video(frame).empty ()) {
- /* We got what we want */
- break;
- }
-
- if (pass ()) {
- /* The decoder has nothing more for us */
- break;
- }
-
- if (!_decoded_video.empty() && _decoded_video.front().frame > frame) {
- /* We're never going to get the frame we want. Perhaps the caller is asking
- * for a video frame before the content's video starts (if its audio
- * begins before its video, for example).
- */
- break;
- }
+ if (_frame_interval_checker) {
+ _frame_interval_checker->feed(time, afr);
+ if (_frame_interval_checker->guess() == FrameIntervalChecker::PROBABLY_NOT_3D && vft == VideoFrameType::THREE_D) {
+ boost::throw_exception (
+ DecodeError(
+ String::compose(
+ _("The content file %1 is set as 3D but does not appear to contain 3D images. Please set it to 2D. "
+ "You can still make a 3D DCP from this content by ticking the 3D option in the DCP video tab."),
+ _content->path(0)
+ )
+ )
+ );
}
- dec = decoded_video (frame);
- } else {
- /* Any frame will do: use the first one that comes out of pass() */
- while (_decoded_video.empty() && !pass ()) {}
- if (!_decoded_video.empty ()) {
- dec.push_back (_decoded_video.front ());
+ if (_frame_interval_checker->guess() != FrameIntervalChecker::AGAIN) {
+ _frame_interval_checker.reset ();
}
}
- /* Clean up _decoded_video; keep the frame we are returning, but nothing before that */
- while (!_decoded_video.empty() && _decoded_video.front().frame < dec.front().frame) {
- _decoded_video.pop_front ();
- }
-
- return dec;
-}
-
+ switch (vft) {
+ case VideoFrameType::TWO_D:
+ Data(ContentVideo(image, time, Eyes::BOTH, Part::WHOLE));
+ break;
+ case VideoFrameType::THREE_D:
+ {
+ auto eyes = Eyes::LEFT;
+ auto j2k = dynamic_pointer_cast<const J2KImageProxy>(image);
+ if (j2k && j2k->eye()) {
+ eyes = *j2k->eye() == dcp::Eye::LEFT ? Eyes::LEFT : Eyes::RIGHT;
+ }
-/** Called by subclasses when they have a video frame ready */
-void
-VideoDecoder::video (shared_ptr<const ImageProxy> image, VideoFrame frame)
-{
- /* We may receive the same frame index twice for 3D, and we need to know
- when that happens.
- */
- _same = (!_decoded_video.empty() && frame == _decoded_video.back().frame);
-
- /* Fill in gaps */
- /* XXX: 3D */
-
- while (!_decoded_video.empty () && (_decoded_video.back().frame + 1) < frame) {
-#ifdef DCPOMATIC_DEBUG
- test_gaps++;
-#endif
- _decoded_video.push_back (
- ContentVideo (
- _decoded_video.back().image,
- _decoded_video.back().eyes,
- _decoded_video.back().part,
- _decoded_video.back().frame + 1
- )
- );
- }
-
- switch (_video_content->video_frame_type ()) {
- case VIDEO_FRAME_TYPE_2D:
- _decoded_video.push_back (ContentVideo (image, EYES_BOTH, PART_WHOLE, frame));
+ Data(ContentVideo(image, time, eyes, Part::WHOLE));
break;
- case VIDEO_FRAME_TYPE_3D_ALTERNATE:
- _decoded_video.push_back (ContentVideo (image, _same ? EYES_RIGHT : EYES_LEFT, PART_WHOLE, frame));
+ }
+ case VideoFrameType::THREE_D_ALTERNATE:
+ {
+ Eyes eyes;
+ if (_last_emitted_eyes) {
+ eyes = _last_emitted_eyes.get() == Eyes::LEFT ? Eyes::RIGHT : Eyes::LEFT;
+ } else {
+ /* We don't know what eye this frame is, so just guess */
+ auto frame = time.frames_round(_content->video_frame_rate().get_value_or(24));
+ eyes = (frame % 2) ? Eyes::RIGHT : Eyes::LEFT;
+ }
+ Data(ContentVideo(image, time, eyes, Part::WHOLE));
+ _last_emitted_eyes = eyes;
break;
- case VIDEO_FRAME_TYPE_3D_LEFT_RIGHT:
- _decoded_video.push_back (ContentVideo (image, EYES_LEFT, PART_LEFT_HALF, frame));
- _decoded_video.push_back (ContentVideo (image, EYES_RIGHT, PART_RIGHT_HALF, frame));
+ }
+ case VideoFrameType::THREE_D_LEFT_RIGHT:
+ Data(ContentVideo(image, time, Eyes::LEFT, Part::LEFT_HALF));
+ Data(ContentVideo(image, time, Eyes::RIGHT, Part::RIGHT_HALF));
break;
- case VIDEO_FRAME_TYPE_3D_TOP_BOTTOM:
- _decoded_video.push_back (ContentVideo (image, EYES_LEFT, PART_TOP_HALF, frame));
- _decoded_video.push_back (ContentVideo (image, EYES_RIGHT, PART_BOTTOM_HALF, frame));
+ case VideoFrameType::THREE_D_TOP_BOTTOM:
+ Data(ContentVideo(image, time, Eyes::LEFT, Part::TOP_HALF));
+ Data(ContentVideo(image, time, Eyes::RIGHT, Part::BOTTOM_HALF));
break;
- case VIDEO_FRAME_TYPE_3D_LEFT:
- _decoded_video.push_back (ContentVideo (image, EYES_LEFT, PART_WHOLE, frame));
+ case VideoFrameType::THREE_D_LEFT:
+ Data(ContentVideo(image, time, Eyes::LEFT, Part::WHOLE));
break;
- case VIDEO_FRAME_TYPE_3D_RIGHT:
- _decoded_video.push_back (ContentVideo (image, EYES_RIGHT, PART_WHOLE, frame));
+ case VideoFrameType::THREE_D_RIGHT:
+ Data(ContentVideo(image, time, Eyes::RIGHT, Part::WHOLE));
break;
default:
- assert (false);
+ DCPOMATIC_ASSERT (false);
}
+
+ _position = time;
}
+
void
-VideoDecoder::seek (ContentTime, bool)
+VideoDecoder::seek ()
{
- _decoded_video.clear ();
+ _position = boost::none;
+ _last_emitted_eyes.reset ();
+ _frame_interval_checker.reset (new FrameIntervalChecker());
}
-