X-Git-Url: https://main.carlh.net/gitweb/?p=dcpomatic.git;a=blobdiff_plain;f=src%2Flib%2Fffmpeg.cc;h=53829c5f2391919056b4248a368693b7fef1d640;hp=ac179e1d62df8ffa0830056b36a39e09447d697e;hb=a5be11a965c2c38442e4e069874e7e21b5b43a5c;hpb=ecddd0e7b0cd4645fdc0fa565928b547b925e840 diff --git a/src/lib/ffmpeg.cc b/src/lib/ffmpeg.cc index ac179e1d6..53829c5f2 100644 --- a/src/lib/ffmpeg.cc +++ b/src/lib/ffmpeg.cc @@ -24,6 +24,7 @@ #include "exceptions.h" #include "util.h" #include "log.h" +#include "dcpomatic_log.h" #include "ffmpeg_subtitle_stream.h" #include "ffmpeg_audio_stream.h" #include "digester.h" @@ -49,7 +50,6 @@ using boost::optional; using dcp::raw_convert; boost::mutex FFmpeg::_mutex; -boost::weak_ptr FFmpeg::_ffmpeg_log; FFmpeg::FFmpeg (boost::shared_ptr c) : _ffmpeg_content (c) @@ -97,14 +97,9 @@ FFmpeg::ffmpeg_log_callback (void* ptr, int level, const char* fmt, va_list vl) char line[1024]; static int prefix = 0; av_log_format_line (ptr, level, fmt, vl, line, sizeof (line), &prefix); - shared_ptr log = _ffmpeg_log.lock (); - if (log) { - string str (line); - boost::algorithm::trim (str); - log->log (String::compose ("FFmpeg: %1", str), LogEntry::TYPE_GENERAL); - } else { - cerr << line; - } + string str (line); + boost::algorithm::trim (str); + dcpomatic_log->log (String::compose ("FFmpeg: %1", str), LogEntry::TYPE_GENERAL); } void @@ -113,7 +108,6 @@ FFmpeg::setup_general () /* This might not work too well in some cases of multiple FFmpeg decoders, but it's probably good enough. */ - _ffmpeg_log = _ffmpeg_content->film()->log (); av_log_set_callback (FFmpeg::ffmpeg_log_callback); _file_group.set_paths (_ffmpeg_content->paths ()); @@ -128,6 +122,9 @@ FFmpeg::setup_general () */ av_dict_set (&options, "analyzeduration", raw_convert (5 * 60 * 1000000).c_str(), 0); av_dict_set (&options, "probesize", raw_convert (5 * 60 * 1000000).c_str(), 0); + if (_ffmpeg_content->decryption_key()) { + av_dict_set (&options, "decryption_key", _ffmpeg_content->decryption_key()->c_str(), 0); + } int e = avformat_open_input (&_format_context, 0, 0, &options); if (e < 0) { @@ -216,10 +213,7 @@ FFmpeg::setup_decoders () throw DecodeError (N_("could not open decoder")); } } else { - shared_ptr log = _ffmpeg_log.lock (); - if (log) { - log->log (String::compose ("No codec found for stream %1", i), LogEntry::TYPE_WARNING); - } + dcpomatic_log->log (String::compose ("No codec found for stream %1", i), LogEntry::TYPE_WARNING); } } } @@ -276,65 +270,6 @@ FFmpeg::subtitle_period (AVSubtitle const & sub) ); } -string -FFmpeg::subtitle_id (AVSubtitle const & sub) -{ - Digester digester; - digester.add (sub.pts); - for (unsigned int i = 0; i < sub.num_rects; ++i) { - AVSubtitleRect* rect = sub.rects[i]; - if (rect->type == SUBTITLE_BITMAP) { - digester.add (rect->x); - digester.add (rect->y); - digester.add (rect->w); - digester.add (rect->h); -#ifdef DCPOMATIC_HAVE_AVSUBTITLERECT_PICT - int const line = rect->pict.linesize[0]; - for (int j = 0; j < rect->h; ++j) { - digester.add (rect->pict.data[0] + j * line, line); - } -#else - int const line = rect->linesize[0]; - for (int j = 0; j < rect->h; ++j) { - digester.add (rect->data[0] + j * line, line); - } -#endif - } else if (rect->type == SUBTITLE_TEXT) { - digester.add (string (rect->text)); - } else if (rect->type == SUBTITLE_ASS) { - digester.add (string (rect->ass)); - } - } - return digester.get (); -} - -/** @return true if sub starts a new image subtitle */ -bool -FFmpeg::subtitle_starts_image (AVSubtitle const & sub) -{ - bool image = false; - bool text = false; - - for (unsigned int i = 0; i < sub.num_rects; ++i) { - switch (sub.rects[i]->type) { - case SUBTITLE_BITMAP: - image = true; - break; - case SUBTITLE_TEXT: - case SUBTITLE_ASS: - text = true; - break; - default: - break; - } - } - - /* We can't cope with mixed image/text in one AVSubtitle */ - DCPOMATIC_ASSERT (!image || !text); - - return image; -} - /** Compute the pts offset to use given a set of audio streams and some video details. * Sometimes these parameters will have just been determined by an Examiner, sometimes * they will have been retrieved from a piece of Content, hence the need for this method