#include <iomanip>
#include <sstream>
#include <iostream>
+#include <fstream>
#include <unistd.h>
#include <errno.h>
#include <boost/array.hpp>
#include "dcp_video_frame.h"
#include "lut.h"
#include "config.h"
-#include "film_state.h"
#include "options.h"
#include "exceptions.h"
#include "server.h"
#include "scaler.h"
#include "image.h"
#include "log.h"
+#include "subtitle.h"
-#ifdef DEBUG_HASH
-#include <mhash.h>
-#endif
+#include "i18n.h"
-using namespace std;
-using namespace boost;
+using std::string;
+using std::stringstream;
+using std::ofstream;
+using std::cout;
+using boost::shared_ptr;
+using libdcp::Size;
/** Construct a DCP video frame.
* @param input Input image.
* @param out Required size of output, in pixels (including any padding).
* @param s Scaler to use.
* @param p Number of pixels of padding either side of the image.
- * @param f Index of the frame within the Film.
- * @param fps Frames per second of the Film.
+ * @param f Index of the frame within the DCP.
+ * @param fps Frames per second of the Film's source.
* @param pp FFmpeg post-processing string to use.
* @param clut Colour look-up table to use (see Config::colour_lut_index ())
* @param bw J2K bandwidth to use (see Config::j2k_bandwidth ())
* @param l Log to write to.
*/
DCPVideoFrame::DCPVideoFrame (
- shared_ptr<Image> yuv, Size out, int p, Scaler const * s, int f, float fps, string pp, int clut, int bw, Log* l)
+ shared_ptr<const Image> yuv, shared_ptr<Subtitle> sub,
+ Size out, int p, int subtitle_offset, float subtitle_scale,
+ Scaler const * s, int f, int dcp_fps, string pp, int clut, int bw, shared_ptr<Log> l
+ )
: _input (yuv)
+ , _subtitle (sub)
, _out_size (out)
, _padding (p)
+ , _subtitle_offset (subtitle_offset)
+ , _subtitle_scale (subtitle_scale)
, _scaler (s)
, _frame (f)
- /* we round here; not sure if this is right */
- , _frames_per_second (rint (fps))
+ , _frames_per_second (dcp_fps)
, _post_process (pp)
- , _colour_lut_index (clut)
+ , _colour_lut (clut)
, _j2k_bandwidth (bw)
, _log (l)
, _image (0)
_image = opj_image_create (3, &_cmptparm[0], CLRSPC_SRGB);
if (_image == 0) {
- throw EncodeError ("could not create libopenjpeg image");
+ throw EncodeError (N_("could not create libopenjpeg image"));
}
_image->x0 = 0;
if (_parameters) {
free (_parameters->cp_comment);
- free (_parameters->cp_matrice);
}
delete _parameters;
shared_ptr<EncodedData>
DCPVideoFrame::encode_locally ()
{
- shared_ptr<Image> prepared = _input;
-
if (!_post_process.empty ()) {
- prepared = prepared->post_process (_post_process);
+ _input = _input->post_process (_post_process, true);
}
- prepared = prepared->scale_and_convert_to_rgb (_out_size, _padding, _scaler);
+ shared_ptr<Image> prepared = _input->scale_and_convert_to_rgb (_out_size, _padding, _scaler, true);
- create_openjpeg_container ();
+ if (_subtitle) {
+ Rect tx = subtitle_transformed_area (
+ float (_out_size.width) / _input->size().width,
+ float (_out_size.height) / _input->size().height,
+ _subtitle->area(), _subtitle_offset, _subtitle_scale
+ );
+
+ shared_ptr<Image> im = _subtitle->image()->scale (tx.size(), _scaler, true);
+ prepared->alpha_blend (im, tx.position());
+ }
- int const size = _out_size.width * _out_size.height;
+ create_openjpeg_container ();
struct {
double r, g, b;
/* Copy our RGB into the openjpeg container, converting to XYZ in the process */
- uint8_t* p = prepared->data()[0];
- for (int i = 0; i < size; ++i) {
- /* In gamma LUT (converting 8-bit input to 12-bit) */
- s.r = lut_in[_colour_lut_index][*p++ << 4];
- s.g = lut_in[_colour_lut_index][*p++ << 4];
- s.b = lut_in[_colour_lut_index][*p++ << 4];
-
- /* RGB to XYZ Matrix */
- d.x = ((s.r * color_matrix[_colour_lut_index][0][0]) + (s.g * color_matrix[_colour_lut_index][0][1]) + (s.b * color_matrix[_colour_lut_index][0][2]));
- d.y = ((s.r * color_matrix[_colour_lut_index][1][0]) + (s.g * color_matrix[_colour_lut_index][1][1]) + (s.b * color_matrix[_colour_lut_index][1][2]));
- d.z = ((s.r * color_matrix[_colour_lut_index][2][0]) + (s.g * color_matrix[_colour_lut_index][2][1]) + (s.b * color_matrix[_colour_lut_index][2][2]));
-
- /* DCI companding */
- d.x = d.x * DCI_COEFFICENT * (DCI_LUT_SIZE - 1);
- d.y = d.y * DCI_COEFFICENT * (DCI_LUT_SIZE - 1);
- d.z = d.z * DCI_COEFFICENT * (DCI_LUT_SIZE - 1);
-
- /* Out gamma LUT */
- _image->comps[0].data[i] = lut_out[LO_DCI][(int) d.x];
- _image->comps[1].data[i] = lut_out[LO_DCI][(int) d.y];
- _image->comps[2].data[i] = lut_out[LO_DCI][(int) d.z];
+ int jn = 0;
+ for (int y = 0; y < _out_size.height; ++y) {
+ uint8_t* p = prepared->data()[0] + y * prepared->stride()[0];
+ for (int x = 0; x < _out_size.width; ++x) {
+
+ /* In gamma LUT (converting 8-bit input to 12-bit) */
+ s.r = lut_in[_colour_lut][*p++ << 4];
+ s.g = lut_in[_colour_lut][*p++ << 4];
+ s.b = lut_in[_colour_lut][*p++ << 4];
+
+ /* RGB to XYZ Matrix */
+ d.x = ((s.r * color_matrix[_colour_lut][0][0]) +
+ (s.g * color_matrix[_colour_lut][0][1]) +
+ (s.b * color_matrix[_colour_lut][0][2]));
+
+ d.y = ((s.r * color_matrix[_colour_lut][1][0]) +
+ (s.g * color_matrix[_colour_lut][1][1]) +
+ (s.b * color_matrix[_colour_lut][1][2]));
+
+ d.z = ((s.r * color_matrix[_colour_lut][2][0]) +
+ (s.g * color_matrix[_colour_lut][2][1]) +
+ (s.b * color_matrix[_colour_lut][2][2]));
+
+ /* DCI companding */
+ d.x = d.x * DCI_COEFFICENT * (DCI_LUT_SIZE - 1);
+ d.y = d.y * DCI_COEFFICENT * (DCI_LUT_SIZE - 1);
+ d.z = d.z * DCI_COEFFICENT * (DCI_LUT_SIZE - 1);
+
+ /* Out gamma LUT */
+ _image->comps[0].data[jn] = lut_out[LO_DCI][(int) d.x];
+ _image->comps[1].data[jn] = lut_out[LO_DCI][(int) d.y];
+ _image->comps[2].data[jn] = lut_out[LO_DCI][(int) d.z];
+
+ ++jn;
+ }
}
/* Set the max image and component sizes based on frame_rate */
_parameters->tcp_numlayers++;
_parameters->cp_disto_alloc = 1;
_parameters->cp_rsiz = CINEMA2K;
- _parameters->cp_comment = strdup ("DVD-o-matic");
+ _parameters->cp_comment = strdup (N_("DVD-o-matic"));
_parameters->cp_cinema = CINEMA2K_24;
/* 3 components, so use MCT */
/* get a J2K compressor handle */
_cinfo = opj_create_compress (CODEC_J2K);
+ if (_cinfo == 0) {
+ throw EncodeError (N_("could not create JPEG2000 encoder"));
+ }
/* Set event manager to null (openjpeg 1.3 bug) */
_cinfo->event_mgr = 0;
-#ifdef DEBUG_HASH
- md5_data ("J2K in X frame " + lexical_cast<string> (_frame), _image->comps[0].data, size * sizeof (int));
- md5_data ("J2K in Y frame " + lexical_cast<string> (_frame), _image->comps[1].data, size * sizeof (int));
- md5_data ("J2K in Z frame " + lexical_cast<string> (_frame), _image->comps[2].data, size * sizeof (int));
-#endif
-
/* Setup the encoder parameters using the current image and user parameters */
opj_setup_encoder (_cinfo, _parameters, _image);
_cio = opj_cio_open ((opj_common_ptr) _cinfo, 0, 0);
+ if (_cio == 0) {
+ throw EncodeError (N_("could not open JPEG2000 stream"));
+ }
int const r = opj_encode (_cinfo, _cio, _image, 0);
if (r == 0) {
- throw EncodeError ("jpeg2000 encoding failed");
+ throw EncodeError (N_("JPEG2000 encoding failed"));
}
-#ifdef DEBUG_HASH
- md5_data ("J2K out frame " + lexical_cast<string> (_frame), _cio->buffer, cio_tell (_cio));
-#endif
-
- {
- stringstream s;
- s << "Finished locally-encoded frame " << _frame;
- _log->log (s.str ());
- }
+ _log->log (String::compose (N_("Finished locally-encoded frame %1"), _frame));
return shared_ptr<EncodedData> (new LocallyEncodedData (_cio->buffer, cio_tell (_cio)));
}
shared_ptr<EncodedData>
DCPVideoFrame::encode_remotely (ServerDescription const * serv)
{
- asio::io_service io_service;
- asio::ip::tcp::resolver resolver (io_service);
+ boost::asio::io_service io_service;
+ boost::asio::ip::tcp::resolver resolver (io_service);
+ boost::asio::ip::tcp::resolver::query query (serv->host_name(), boost::lexical_cast<string> (Config::instance()->server_port ()));
+ boost::asio::ip::tcp::resolver::iterator endpoint_iterator = resolver.resolve (query);
- asio::ip::tcp::resolver::query query (serv->host_name(), boost::lexical_cast<string> (Config::instance()->server_port ()));
- asio::ip::tcp::resolver::iterator endpoint_iterator = resolver.resolve (query);
+ shared_ptr<Socket> socket (new Socket);
- shared_ptr<asio::ip::tcp::socket> socket (new asio::ip::tcp::socket (io_service));
+ socket->connect (*endpoint_iterator);
- DeadlineWrapper wrapper (io_service);
- wrapper.set_socket (socket);
+ stringstream s;
+ s << N_("encode please\n")
+ << N_("input_width ") << _input->size().width << N_("\n")
+ << N_("input_height ") << _input->size().height << N_("\n")
+ << N_("input_pixel_format ") << _input->pixel_format() << N_("\n")
+ << N_("output_width ") << _out_size.width << N_("\n")
+ << N_("output_height ") << _out_size.height << N_("\n")
+ << N_("padding ") << _padding << N_("\n")
+ << N_("subtitle_offset ") << _subtitle_offset << N_("\n")
+ << N_("subtitle_scale ") << _subtitle_scale << N_("\n")
+ << N_("scaler ") << _scaler->id () << N_("\n")
+ << N_("frame ") << _frame << N_("\n")
+ << N_("frames_per_second ") << _frames_per_second << N_("\n");
+
+ if (!_post_process.empty()) {
+ s << N_("post_process ") << _post_process << N_("\n");
+ }
+
+ s << N_("colour_lut ") << _colour_lut << N_("\n")
+ << N_("j2k_bandwidth ") << _j2k_bandwidth << N_("\n");
+
+ if (_subtitle) {
+ s << N_("subtitle_x ") << _subtitle->position().x << N_("\n")
+ << N_("subtitle_y ") << _subtitle->position().y << N_("\n")
+ << N_("subtitle_width ") << _subtitle->image()->size().width << N_("\n")
+ << N_("subtitle_height ") << _subtitle->image()->size().height << N_("\n");
+ }
- wrapper.connect (*endpoint_iterator, 30);
+ _log->log (String::compose (
+ N_("Sending to remote; pixel format %1, components %2, lines (%3,%4,%5), line sizes (%6,%7,%8)"),
+ _input->pixel_format(), _input->components(),
+ _input->lines(0), _input->lines(1), _input->lines(2),
+ _input->line_size()[0], _input->line_size()[1], _input->line_size()[2]
+ ));
-#ifdef DEBUG_HASH
- _input->hash ("Input for remote encoding (before sending)");
-#endif
+ socket->write (s.str().length() + 1);
+ socket->write ((uint8_t *) s.str().c_str(), s.str().length() + 1);
- stringstream s;
- s << "encode "
- << _input->size().width << " " << _input->size().height << " "
- << _input->pixel_format() << " "
- << _out_size.width << " " << _out_size.height << " "
- << _padding << " "
- << _scaler->id () << " "
- << _frame << " "
- << _frames_per_second << " "
- << (_post_process.empty() ? "none" : _post_process) << " "
- << Config::instance()->colour_lut_index () << " "
- << Config::instance()->j2k_bandwidth () << " ";
-
- for (int i = 0; i < _input->components(); ++i) {
- s << _input->line_size()[i] << " ";
+ _input->write_to_socket (socket);
+ if (_subtitle) {
+ _subtitle->image()->write_to_socket (socket);
}
- wrapper.write ((uint8_t *) s.str().c_str(), s.str().length() + 1, 10);
+ shared_ptr<EncodedData> e (new RemotelyEncodedData (socket->read_uint32 ()));
+ socket->read (e->data(), e->size());
- for (int i = 0; i < _input->components(); ++i) {
- wrapper.write (_input->data()[i], _input->line_size()[i] * _input->lines(i), 10);
- }
+ _log->log (String::compose (N_("Finished remotely-encoded frame %1"), _frame));
+
+ return e;
+}
- char buffer[32];
- wrapper.read_indefinite ((uint8_t *) buffer, sizeof (buffer), 30);
- wrapper.consume (strlen (buffer) + 1);
- shared_ptr<EncodedData> e (new RemotelyEncodedData (atoi (buffer)));
+EncodedData::EncodedData (int s)
+ : _data (new uint8_t[s])
+ , _size (s)
+{
- /* now read the rest */
- wrapper.read_definite_and_consume (e->data(), e->size(), 30);
+}
-#ifdef DEBUG_HASH
- e->hash ("Encoded image (after receiving)");
-#endif
+EncodedData::EncodedData (string file)
+{
+ _size = boost::filesystem::file_size (file);
+ _data = new uint8_t[_size];
- {
- stringstream s;
- s << "Finished remotely-encoded frame " << _frame;
- _log->log (s.str ());
+ FILE* f = fopen (file.c_str(), N_("rb"));
+ if (!f) {
+ throw FileError (_("could not open file for reading"), file);
}
- return e;
+ fread (_data, 1, _size, f);
+ fclose (f);
+}
+
+
+EncodedData::~EncodedData ()
+{
+ delete[] _data;
}
/** Write this data to a J2K file.
- * @param opt Options.
- * @param frame Frame index.
+ * @param Film Film.
+ * @param frame DCP frame index.
*/
void
-EncodedData::write (shared_ptr<const Options> opt, int frame)
+EncodedData::write (shared_ptr<const Film> film, int frame) const
{
- string const tmp_j2k = opt->frame_out_path (frame, true);
+ string const tmp_j2c = film->j2c_path (frame, true);
- FILE* f = fopen (tmp_j2k.c_str (), "wb");
+ FILE* f = fopen (tmp_j2c.c_str (), N_("wb"));
if (!f) {
- throw WriteFileError (tmp_j2k, errno);
+ throw WriteFileError (tmp_j2c, errno);
}
fwrite (_data, 1, _size, f);
fclose (f);
+ string const real_j2c = film->j2c_path (frame, false);
+
/* Rename the file from foo.j2c.tmp to foo.j2c now that it is complete */
- filesystem::rename (tmp_j2k, opt->frame_out_path (frame, false));
+ boost::filesystem::rename (tmp_j2c, real_j2c);
+}
+
+void
+EncodedData::write_info (shared_ptr<const Film> film, int frame, libdcp::FrameInfo fin) const
+{
+ string const info = film->info_path (frame);
+ ofstream h (info.c_str());
+ fin.write (h);
}
/** Send this data to a socket.
* @param socket Socket
*/
void
-EncodedData::send (shared_ptr<asio::ip::tcp::socket> socket)
+EncodedData::send (shared_ptr<Socket> socket)
{
- stringstream s;
- s << _size;
- asio::write (*socket, asio::buffer (s.str().c_str(), s.str().length() + 1));
- asio::write (*socket, asio::buffer (_data, _size));
+ socket->write (_size);
+ socket->write (_data, _size);
}
-#ifdef DEBUG_HASH
-void
-EncodedData::hash (string n) const
+LocallyEncodedData::LocallyEncodedData (uint8_t* d, int s)
+ : EncodedData (s)
{
- md5_data (n, _data, _size);
+ memcpy (_data, d, s);
}
-#endif
/** @param s Size of data in bytes */
RemotelyEncodedData::RemotelyEncodedData (int s)
- : EncodedData (new uint8_t[s], s)
+ : EncodedData (s)
{
}
-
-RemotelyEncodedData::~RemotelyEncodedData ()
-{
- delete[] _data;
-}