X-Git-Url: https://main.carlh.net/gitweb/?p=dcpomatic.git;a=blobdiff_plain;f=src%2Flib%2Fffmpeg.cc;h=3bd08e84a6ea7c1be2f27426e1755171dce484d4;hp=bbf9bbed9405fb602b1fd2d69f69d408daa5fc00;hb=254b3044d72de6b033d7c584f5abd2b9aa70aad5;hpb=6ec390c3a6259848cb6a85ee5dfda79f76161942 diff --git a/src/lib/ffmpeg.cc b/src/lib/ffmpeg.cc index bbf9bbed9..3bd08e84a 100644 --- a/src/lib/ffmpeg.cc +++ b/src/lib/ffmpeg.cc @@ -23,12 +23,12 @@ #include "film.h" #include "exceptions.h" #include "util.h" -#include "raw_convert.h" #include "log.h" #include "ffmpeg_subtitle_stream.h" #include "ffmpeg_audio_stream.h" #include "digester.h" #include "compose.hpp" +#include extern "C" { #include #include @@ -46,6 +46,7 @@ using std::cerr; using std::vector; using boost::shared_ptr; using boost::optional; +using dcp::raw_convert; boost::mutex FFmpeg::_mutex; boost::weak_ptr FFmpeg::_ffmpeg_log; @@ -109,12 +110,9 @@ FFmpeg::ffmpeg_log_callback (void* ptr, int level, const char* fmt, va_list vl) void FFmpeg::setup_general () { - av_register_all (); - /* This might not work too well in some cases of multiple FFmpeg decoders, but it's probably good enough. */ - _ffmpeg_log = _ffmpeg_content->film()->log (); av_log_set_callback (FFmpeg::ffmpeg_log_callback); _file_group.set_paths (_ffmpeg_content->paths ()); @@ -129,9 +127,13 @@ FFmpeg::setup_general () */ av_dict_set (&options, "analyzeduration", raw_convert (5 * 60 * 1000000).c_str(), 0); av_dict_set (&options, "probesize", raw_convert (5 * 60 * 1000000).c_str(), 0); + if (_ffmpeg_content->decryption_key()) { + av_dict_set (&options, "decryption_key", _ffmpeg_content->decryption_key()->c_str(), 0); + } - if (avformat_open_input (&_format_context, 0, 0, &options) < 0) { - throw OpenFileError (_ffmpeg_content->path(0).string ()); + int e = avformat_open_input (&_format_context, 0, 0, &options); + if (e < 0) { + throw OpenFileError (_ffmpeg_content->path(0).string(), e, true); } if (avformat_find_stream_info (_format_context, 0) < 0) { @@ -144,7 +146,7 @@ FFmpeg::setup_general () for (uint32_t i = 0; i < _format_context->nb_streams; ++i) { AVStream* s = _format_context->streams[i]; - if (s->codec->codec_type == AVMEDIA_TYPE_VIDEO) { + if (s->codec->codec_type == AVMEDIA_TYPE_VIDEO && avcodec_find_decoder(s->codec->codec_id)) { if (s->avg_frame_rate.num > 0 && s->avg_frame_rate.den > 0) { /* This is definitely our video stream */ _video_stream = i; @@ -215,9 +217,9 @@ FFmpeg::setup_decoders () if (avcodec_open2 (context, codec, &options) < 0) { throw DecodeError (N_("could not open decoder")); } + } else { + dcpomatic_log->log (String::compose ("No codec found for stream %1", i), LogEntry::TYPE_WARNING); } - - /* We are silently ignoring any failures to find suitable decoders here */ } } @@ -273,59 +275,6 @@ FFmpeg::subtitle_period (AVSubtitle const & sub) ); } -string -FFmpeg::subtitle_id (AVSubtitle const & sub) -{ - Digester digester; - digester.add (sub.pts); - for (unsigned int i = 0; i < sub.num_rects; ++i) { - AVSubtitleRect* rect = sub.rects[i]; - digester.add (rect->x); - digester.add (rect->y); - digester.add (rect->w); - digester.add (rect->h); -#ifdef DCPOMATIC_HAVE_AVSUBTITLERECT_PICT - int const line = rect->pict.linesize[0]; - for (int j = 0; j < rect->h; ++j) { - digester.add (rect->pict.data[0] + j * line, line); - } -#else - int const line = rect->linesize[0]; - for (int j = 0; j < rect->h; ++j) { - digester.add (rect->data[0] + j * line, line); - } -#endif - } - return digester.get (); -} - -/** @return true if sub starts a new image subtitle */ -bool -FFmpeg::subtitle_starts_image (AVSubtitle const & sub) -{ - bool image = false; - bool text = false; - - for (unsigned int i = 0; i < sub.num_rects; ++i) { - switch (sub.rects[i]->type) { - case SUBTITLE_BITMAP: - image = true; - break; - case SUBTITLE_TEXT: - case SUBTITLE_ASS: - text = true; - break; - default: - break; - } - } - - /* We can't cope with mixed image/text in one AVSubtitle */ - DCPOMATIC_ASSERT (!image || !text); - - return image; -} - /** Compute the pts offset to use given a set of audio streams and some video details. * Sometimes these parameters will have just been determined by an Examiner, sometimes * they will have been retrieved from a piece of Content, hence the need for this method @@ -373,7 +322,7 @@ FFmpeg::pts_offset (vector > audio_streams, option /* Now adjust so that the video pts starts on a frame */ if (first_video) { ContentTime const fvc = first_video.get() + po; - po += fvc.round_up (video_frame_rate) - fvc; + po += fvc.ceil (video_frame_rate) - fvc; } return po;