X-Git-Url: https://main.carlh.net/gitweb/?p=dcpomatic.git;a=blobdiff_plain;f=src%2Flib%2Fdcp_decoder.cc;h=59b44ae975ee5080c48ec6599535bf21d390c915;hp=2433ad0fb8bf1a4ca6ee6b82b6526ebfff56cbaa;hb=037b18b34e3e3fd9b9e692d85f35de17ab67aff8;hpb=e35d2120a831fe6ccd993a4f86eac084e057e78d diff --git a/src/lib/dcp_decoder.cc b/src/lib/dcp_decoder.cc index 2433ad0fb..59b44ae97 100644 --- a/src/lib/dcp_decoder.cc +++ b/src/lib/dcp_decoder.cc @@ -1,5 +1,5 @@ /* - Copyright (C) 2014-2016 Carl Hetherington + Copyright (C) 2014-2020 Carl Hetherington This file is part of DCP-o-matic. @@ -18,15 +18,19 @@ */ +#include "atmos_decoder.h" #include "dcp_decoder.h" #include "dcp_content.h" #include "audio_content.h" #include "video_decoder.h" #include "audio_decoder.h" #include "j2k_image_proxy.h" -#include "subtitle_decoder.h" +#include "text_decoder.h" +#include "ffmpeg_image_proxy.h" #include "image.h" #include "config.h" +#include "digester.h" +#include "frame_interval_checker.h" #include #include #include @@ -37,62 +41,89 @@ #include #include #include +#include #include #include #include #include -#include +#include +#include +#include #include #include "i18n.h" using std::list; using std::cout; -using boost::shared_ptr; -using boost::dynamic_pointer_cast; +using std::map; +using std::string; +using std::vector; +using std::shared_ptr; +using std::dynamic_pointer_cast; +using std::make_shared; using boost::optional; +using namespace dcpomatic; -DCPDecoder::DCPDecoder (shared_ptr c, shared_ptr log, bool fast) - : DCP (c) - , _decode_referenced (false) +DCPDecoder::DCPDecoder (shared_ptr film, shared_ptr c, bool fast, bool tolerant, shared_ptr old) + : DCP (c, tolerant) + , Decoder (film) { - if (c->video) { - video.reset (new VideoDecoder (this, c, log)); - } - if (c->audio) { - audio.reset (new AudioDecoder (this, c->audio, log, fast)); - } - if (c->subtitle) { - /* XXX: this time here should be the time of the first subtitle, not 0 */ - subtitle.reset (new SubtitleDecoder (this, c->subtitle, log, ContentTime())); + if (c->can_be_played()) { + if (c->video) { + video = make_shared(this, c); + } + if (c->audio) { + audio = make_shared(this, c->audio, fast); + } + for (auto i: c->text) { + /* XXX: this time here should be the time of the first subtitle, not 0 */ + text.push_back (make_shared(this, i, ContentTime())); + } + if (c->atmos) { + atmos = make_shared(this, c); + } } - list > cpl_list = cpls (); + /* We try to avoid re-scanning the DCP's files every time we make a new DCPDecoder; we do this + by re-using the _reels list. Before we do this we need to check that nothing too serious + has changed in the DCPContent. - if (cpl_list.empty()) { - throw DCPError (_("No CPLs found in DCP.")); - } + We do this by storing a digest of the important bits of the DCPContent and then checking that's + the same before we re-use _reels. + */ + + _lazy_digest = calculate_lazy_digest (c); - shared_ptr cpl; - BOOST_FOREACH (shared_ptr i, cpl_list) { - if (_dcp_content->cpl() && i->id() == _dcp_content->cpl().get()) { - cpl = i; + if (old && old->lazy_digest() == _lazy_digest) { + _reels = old->_reels; + } else { + + auto cpl_list = cpls (); + + if (cpl_list.empty()) { + throw DCPError (_("No CPLs found in DCP.")); } - } - if (!cpl) { - /* No CPL found; probably an old file that doesn't specify it; - just use the first one. - */ - cpl = cpls().front (); + shared_ptr cpl; + for (auto i: cpl_list) { + if (_dcp_content->cpl() && i->id() == _dcp_content->cpl().get()) { + cpl = i; + } + } + + if (!cpl) { + /* No CPL found; probably an old file that doesn't specify it; + just use the first one. + */ + cpl = cpls().front (); + } + + _reels = cpl->reels (); } set_decode_referenced (false); - _reels = cpl->reels (); - _reel = _reels.begin (); - _offset = 0; get_readers (); } @@ -100,58 +131,64 @@ DCPDecoder::DCPDecoder (shared_ptr c, shared_ptr log, boo bool DCPDecoder::pass () { - if (_reel == _reels.end () || !_dcp_content->can_be_played ()) { + if (!_dcp_content->can_be_played()) { return true; } - double const vfr = _dcp_content->active_video_frame_rate (); + if (_reel == _reels.end()) { + if (audio) { + audio->flush (); + } + return true; + } + + double const vfr = _dcp_content->active_video_frame_rate (film()); /* Frame within the (played part of the) reel that is coming up next */ int64_t const frame = _next.frames_round (vfr); - /* We must emit subtitles first as when we emit the video for this frame - it will expect already to have the subs. + shared_ptr picture_asset = (*_reel)->main_picture()->asset(); + DCPOMATIC_ASSERT (picture_asset); + + /* We must emit texts first as when we emit the video for this frame + it will expect already to have the texts. */ - pass_subtitles (_next); + pass_texts (_next, picture_asset->size()); if ((_mono_reader || _stereo_reader) && (_decode_referenced || !_dcp_content->reference_video())) { - shared_ptr asset = (*_reel)->main_picture()->asset (); - int64_t const entry_point = (*_reel)->main_picture()->entry_point (); + int64_t const entry_point = (*_reel)->main_picture()->entry_point().get_value_or(0); if (_mono_reader) { video->emit ( - shared_ptr ( - new J2KImageProxy ( - _mono_reader->get_frame (entry_point + frame), - asset->size(), - AV_PIX_FMT_XYZ12LE, - _forced_reduction - ) + film(), + std::make_shared( + _mono_reader->get_frame (entry_point + frame), + picture_asset->size(), + AV_PIX_FMT_XYZ12LE, + _forced_reduction ), _offset + frame ); } else { video->emit ( - shared_ptr ( - new J2KImageProxy ( - _stereo_reader->get_frame (entry_point + frame), - asset->size(), - dcp::EYE_LEFT, - AV_PIX_FMT_XYZ12LE, - _forced_reduction - ) + film(), + std::make_shared( + _stereo_reader->get_frame (entry_point + frame), + picture_asset->size(), + dcp::Eye::LEFT, + AV_PIX_FMT_XYZ12LE, + _forced_reduction ), _offset + frame ); video->emit ( - shared_ptr ( - new J2KImageProxy ( - _stereo_reader->get_frame (entry_point + frame), - asset->size(), - dcp::EYE_RIGHT, - AV_PIX_FMT_XYZ12LE, - _forced_reduction - ) + film(), + std::make_shared( + _stereo_reader->get_frame (entry_point + frame), + picture_asset->size(), + dcp::Eye::RIGHT, + AV_PIX_FMT_XYZ12LE, + _forced_reduction ), _offset + frame ); @@ -159,7 +196,7 @@ DCPDecoder::pass () } if (_sound_reader && (_decode_referenced || !_dcp_content->reference_audio())) { - int64_t const entry_point = (*_reel)->main_sound()->entry_point (); + int64_t const entry_point = (*_reel)->main_sound()->entry_point().get_value_or(0); shared_ptr sf = _sound_reader->get_frame (entry_point + frame); uint8_t const * from = sf->data (); @@ -174,7 +211,13 @@ DCPDecoder::pass () } } - audio->emit (_dcp_content->audio->stream(), data, ContentTime::from_frames (_offset, vfr) + _next); + audio->emit (film(), _dcp_content->audio->stream(), data, ContentTime::from_frames (_offset, vfr) + _next); + } + + if (_atmos_reader) { + DCPOMATIC_ASSERT (_atmos_metadata); + int64_t const entry_point = (*_reel)->atmos()->entry_point().get_value_or(0); + atmos->emit (film(), _atmos_reader->get_frame(entry_point + frame), _offset + frame, *_atmos_metadata); } _next += ContentTime::from_frames (1, vfr); @@ -190,30 +233,100 @@ DCPDecoder::pass () } void -DCPDecoder::pass_subtitles (ContentTime next) +DCPDecoder::pass_texts (ContentTime next, dcp::Size size) { - double const vfr = _dcp_content->active_video_frame_rate (); + auto decoder = text.begin (); + if (decoder == text.end()) { + /* It's possible that there is now a main subtitle but no TextDecoders, for example if + the CPL has just changed but the TextContent's texts have not been recreated yet. + */ + return; + } + + if ((*_reel)->main_subtitle()) { + pass_texts ( + next, + (*_reel)->main_subtitle()->asset(), + _dcp_content->reference_text(TextType::OPEN_SUBTITLE), + (*_reel)->main_subtitle()->entry_point().get_value_or(0), + *decoder, + size + ); + ++decoder; + } + + for (auto i: (*_reel)->closed_captions()) { + pass_texts ( + next, i->asset(), _dcp_content->reference_text(TextType::CLOSED_CAPTION), i->entry_point().get_value_or(0), *decoder, size + ); + ++decoder; + } +} + +void +DCPDecoder::pass_texts ( + ContentTime next, shared_ptr asset, bool reference, int64_t entry_point, shared_ptr decoder, dcp::Size size + ) +{ + double const vfr = _dcp_content->active_video_frame_rate (film()); /* Frame within the (played part of the) reel that is coming up next */ int64_t const frame = next.frames_round (vfr); - if ((*_reel)->main_subtitle() && (_decode_referenced || !_dcp_content->reference_subtitle())) { - int64_t const entry_point = (*_reel)->main_subtitle()->entry_point (); - list subs = (*_reel)->main_subtitle()->asset()->subtitles_during ( + if (_decode_referenced || !reference) { + auto subs = asset->subtitles_during ( dcp::Time (entry_point + frame, vfr, vfr), dcp::Time (entry_point + frame + 1, vfr, vfr), true ); - BOOST_FOREACH (dcp::SubtitleString i, subs) { - list s; - s.push_back (i); - subtitle->emit_text ( + list strings; + + for (auto i: subs) { + auto is = dynamic_pointer_cast(i); + if (is) { + if (!strings.empty() && (strings.back().in() != is->in() || strings.back().out() != is->out())) { + auto b = strings.back(); + decoder->emit_plain ( + ContentTimePeriod ( + ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.in().as_seconds()), + ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.out().as_seconds()) + ), + strings + ); + strings.clear (); + } + + strings.push_back (*is); + } + + /* XXX: perhaps these image subs should also be collected together like the string ones are; + this would need to be done both here and in DCPSubtitleDecoder. + */ + + auto ii = dynamic_pointer_cast(i); + if (ii) { + emit_subtitle_image ( + ContentTimePeriod ( + ContentTime::from_frames (_offset - entry_point, vfr) + ContentTime::from_seconds (i->in().as_seconds ()), + ContentTime::from_frames (_offset - entry_point, vfr) + ContentTime::from_seconds (i->out().as_seconds ()) + ), + *ii, + size, + decoder + ); + } + } + + if (!strings.empty()) { + auto b = strings.back(); + decoder->emit_plain ( ContentTimePeriod ( - ContentTime::from_frames (_offset - entry_point, vfr) + ContentTime::from_seconds (i.in().as_seconds ()), - ContentTime::from_frames (_offset - entry_point, vfr) + ContentTime::from_seconds (i.out().as_seconds ()) + ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.in().as_seconds()), + ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.out().as_seconds()) ), - s + strings ); + strings.clear (); } } } @@ -221,7 +334,7 @@ DCPDecoder::pass_subtitles (ContentTime next) void DCPDecoder::next_reel () { - _offset += (*_reel)->main_picture()->duration(); + _offset += (*_reel)->main_picture()->actual_duration(); ++_reel; get_readers (); } @@ -233,6 +346,7 @@ DCPDecoder::get_readers () _mono_reader.reset (); _stereo_reader.reset (); _sound_reader.reset (); + _atmos_reader.reset (); return; } @@ -243,9 +357,11 @@ DCPDecoder::get_readers () DCPOMATIC_ASSERT (mono || stereo); if (mono) { _mono_reader = mono->start_read (); + _mono_reader->set_check_hmac (false); _stereo_reader.reset (); } else { _stereo_reader = stereo->start_read (); + _stereo_reader->set_check_hmac (false); _mono_reader.reset (); } } else { @@ -255,9 +371,20 @@ DCPDecoder::get_readers () if ((*_reel)->main_sound()) { _sound_reader = (*_reel)->main_sound()->asset()->start_read (); + _sound_reader->set_check_hmac (false); } else { _sound_reader.reset (); } + + if ((*_reel)->atmos()) { + shared_ptr asset = (*_reel)->atmos()->asset(); + _atmos_reader = asset->start_read(); + _atmos_reader->set_check_hmac (false); + _atmos_metadata = AtmosMetadata (asset); + } else { + _atmos_reader.reset (); + _atmos_metadata = boost::none; + } } void @@ -284,25 +411,33 @@ DCPDecoder::seek (ContentTime t, bool accurate) /* Seek to pre-roll position */ - while (_reel != _reels.end() && pre >= ContentTime::from_frames ((*_reel)->main_picture()->duration(), _dcp_content->active_video_frame_rate ())) { - ContentTime rd = ContentTime::from_frames ((*_reel)->main_picture()->duration(), _dcp_content->active_video_frame_rate ()); + while ( + _reel != _reels.end() && + pre >= ContentTime::from_frames ((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())) + ) { + + ContentTime rd = ContentTime::from_frames ((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())); pre -= rd; t -= rd; next_reel (); } - /* Pass subtitles in the pre-roll */ + /* Pass texts in the pre-roll */ - double const vfr = _dcp_content->active_video_frame_rate (); + double const vfr = _dcp_content->active_video_frame_rate (film()); for (int i = 0; i < pre_roll_seconds * vfr; ++i) { - pass_subtitles (pre); + pass_texts (pre, (*_reel)->main_picture()->asset()->size()); pre += ContentTime::from_frames (1, vfr); } /* Seek to correct position */ - while (_reel != _reels.end() && t >= ContentTime::from_frames ((*_reel)->main_picture()->duration(), _dcp_content->active_video_frame_rate ())) { - t -= ContentTime::from_frames ((*_reel)->main_picture()->duration(), _dcp_content->active_video_frame_rate ()); + while ( + _reel != _reels.end() && + t >= ContentTime::from_frames ((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())) + ) { + + t -= ContentTime::from_frames ((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())); next_reel (); } @@ -327,3 +462,43 @@ DCPDecoder::set_forced_reduction (optional reduction) { _forced_reduction = reduction; } + +string +DCPDecoder::calculate_lazy_digest (shared_ptr c) const +{ + Digester d; + for (auto i: c->paths()) { + d.add (i.string()); + } + if (_dcp_content->kdm()) { + d.add(_dcp_content->kdm()->id()); + } + d.add (static_cast(c->cpl())); + if (c->cpl()) { + d.add (c->cpl().get()); + } + return d.get (); +} + +ContentTime +DCPDecoder::position () const +{ + return ContentTime::from_frames(_offset, _dcp_content->active_video_frame_rate(film())) + _next; +} + + +vector +DCPDecoder::fonts () const +{ + vector data; + for (auto i: _reels) { + if (i->main_subtitle() && i->main_subtitle()->asset()) { + map fm = i->main_subtitle()->asset()->font_data(); + for (map::const_iterator j = fm.begin(); j != fm.end(); ++j) { + data.push_back (FontData(j->first, j->second)); + } + } + } + return data; +} +