X-Git-Url: https://main.carlh.net/gitweb/?a=blobdiff_plain;f=src%2Flib%2Fimage.cc;h=2d4bc0af05457f6c00eaa123f7514fdf252f8378;hb=8889cf7126810fb9b754643a45dcc94ad578125f;hp=620e71aa7a188d43aaaa560b2d48ca9a3c9c5e45;hpb=493fa08920fbed2d8cda546d86a2b5e68083ed58;p=dcpomatic.git diff --git a/src/lib/image.cc b/src/lib/image.cc index 620e71aa7..2d4bc0af0 100644 --- a/src/lib/image.cc +++ b/src/lib/image.cc @@ -26,6 +26,7 @@ #include #include #include +#include #include extern "C" { #include @@ -39,8 +40,17 @@ extern "C" { #include "exceptions.h" #include "scaler.h" +#include "i18n.h" + using namespace std; using namespace boost; +using libdcp::Size; + +void +Image::swap (Image& other) +{ + std::swap (_pixel_format, other._pixel_format); +} /** @param n Component index. * @return Number of lines in the image for the given component. @@ -57,9 +67,18 @@ Image::lines (int n) const } break; case PIX_FMT_RGB24: + case PIX_FMT_RGBA: + case PIX_FMT_YUV422P10LE: + case PIX_FMT_YUV422P: + case PIX_FMT_YUV444P: + case PIX_FMT_YUV444P9BE: + case PIX_FMT_YUV444P9LE: + case PIX_FMT_YUV444P10BE: + case PIX_FMT_YUV444P10LE: + case PIX_FMT_UYVY422: return size().height; default: - assert (false); + throw PixelFormatError (N_("lines()"), _pixel_format); } return 0; @@ -71,30 +90,72 @@ Image::components () const { switch (_pixel_format) { case PIX_FMT_YUV420P: + case PIX_FMT_YUV422P10LE: + case PIX_FMT_YUV422P: + case PIX_FMT_YUV444P: + case PIX_FMT_YUV444P9BE: + case PIX_FMT_YUV444P9LE: + case PIX_FMT_YUV444P10BE: + case PIX_FMT_YUV444P10LE: return 3; case PIX_FMT_RGB24: + case PIX_FMT_RGBA: + case PIX_FMT_UYVY422: return 1; default: - assert (false); + throw PixelFormatError (N_("components()"), _pixel_format); } return 0; } +shared_ptr +Image::scale (libdcp::Size out_size, Scaler const * scaler, bool result_aligned) const +{ + assert (scaler); + /* Empirical testing suggests that sws_scale() will crash if + the input image is not aligned. + */ + assert (aligned ()); + + shared_ptr scaled (new SimpleImage (pixel_format(), out_size, result_aligned)); + + struct SwsContext* scale_context = sws_getContext ( + size().width, size().height, pixel_format(), + out_size.width, out_size.height, pixel_format(), + scaler->ffmpeg_id (), 0, 0, 0 + ); + + sws_scale ( + scale_context, + data(), stride(), + 0, size().height, + scaled->data(), scaled->stride() + ); + + sws_freeContext (scale_context); + + return scaled; +} + /** Scale this image to a given size and convert it to RGB. * @param out_size Output image size in pixels. * @param scaler Scaler to use. */ -shared_ptr -Image::scale_and_convert_to_rgb (Size out_size, int padding, Scaler const * scaler) const +shared_ptr +Image::scale_and_convert_to_rgb (libdcp::Size out_size, int padding, Scaler const * scaler, bool result_aligned) const { assert (scaler); + /* Empirical testing suggests that sws_scale() will crash if + the input image is not aligned. + */ + assert (aligned ()); - Size content_size = out_size; + libdcp::Size content_size = out_size; content_size.width -= (padding * 2); - shared_ptr rgb (new RGBFrameImage (content_size)); - + shared_ptr rgb (new SimpleImage (PIX_FMT_RGB24, content_size, result_aligned)); + struct SwsContext* scale_context = sws_getContext ( size().width, size().height, pixel_format(), content_size.width, content_size.height, PIX_FMT_RGB24, @@ -104,9 +165,9 @@ Image::scale_and_convert_to_rgb (Size out_size, int padding, Scaler const * scal /* Scale and convert to RGB from whatever its currently in (which may be RGB) */ sws_scale ( scale_context, - data(), line_size(), + data(), stride(), 0, size().height, - rgb->data (), rgb->line_size () + rgb->data(), rgb->stride() ); /* Put the image in the right place in a black frame if are padding; this is @@ -114,7 +175,7 @@ Image::scale_and_convert_to_rgb (Size out_size, int padding, Scaler const * scal scheme of things. */ if (padding > 0) { - shared_ptr padded_rgb (new RGBFrameImage (out_size)); + shared_ptr padded_rgb (new SimpleImage (PIX_FMT_RGB24, out_size, result_aligned)); padded_rgb->make_black (); /* XXX: we are cheating a bit here; we know the frame is RGB so we can @@ -124,8 +185,8 @@ Image::scale_and_convert_to_rgb (Size out_size, int padding, Scaler const * scal uint8_t* q = rgb->data()[0]; for (int j = 0; j < rgb->lines(0); ++j) { memcpy (p, q, rgb->line_size()[0]); - p += padded_rgb->line_size()[0]; - q += rgb->line_size()[0]; + p += padded_rgb->stride()[0]; + q += rgb->stride()[0]; } rgb = padded_rgb; @@ -140,17 +201,37 @@ Image::scale_and_convert_to_rgb (Size out_size, int padding, Scaler const * scal * @param pp Flags for the required set of post processes. * @return Post-processed image. */ -shared_ptr -Image::post_process (string pp) const +shared_ptr +Image::post_process (string pp, bool aligned) const { - shared_ptr out (new PostProcessImage (PIX_FMT_YUV420P, size ())); - + shared_ptr out (new SimpleImage (pixel_format(), size (), aligned)); + + int pp_format = 0; + switch (pixel_format()) { + case PIX_FMT_YUV420P: + pp_format = PP_FORMAT_420; + break; + case PIX_FMT_YUV422P10LE: + case PIX_FMT_YUV422P: + case PIX_FMT_UYVY422: + pp_format = PP_FORMAT_422; + break; + case PIX_FMT_YUV444P: + case PIX_FMT_YUV444P9BE: + case PIX_FMT_YUV444P9LE: + case PIX_FMT_YUV444P10BE: + case PIX_FMT_YUV444P10LE: + pp_format = PP_FORMAT_444; + default: + throw PixelFormatError (N_("post_process"), pixel_format()); + } + pp_mode* mode = pp_get_mode_by_name_and_quality (pp.c_str (), PP_QUALITY_MAX); - pp_context* context = pp_get_context (size().width, size().height, PP_FORMAT_420 | PP_CPU_CAPS_MMX2); + pp_context* context = pp_get_context (size().width, size().height, pp_format | PP_CPU_CAPS_MMX2); pp_postprocess ( - (const uint8_t **) data(), line_size(), - out->data(), out->line_size(), + (const uint8_t **) data(), stride(), + out->data(), out->stride(), size().width, size().height, 0, 0, mode, context, 0 ); @@ -161,196 +242,447 @@ Image::post_process (string pp) const return out; } +shared_ptr +Image::crop (Crop crop, bool aligned) const +{ + libdcp::Size cropped_size = size (); + cropped_size.width -= crop.left + crop.right; + cropped_size.height -= crop.top + crop.bottom; + + shared_ptr out (new SimpleImage (pixel_format(), cropped_size, aligned)); + + for (int c = 0; c < components(); ++c) { + int const crop_left_in_bytes = bytes_per_pixel(c) * crop.left; + int const cropped_width_in_bytes = bytes_per_pixel(c) * cropped_size.width; + + /* Start of the source line, cropped from the top but not the left */ + uint8_t* in_p = data()[c] + crop.top * stride()[c]; + uint8_t* out_p = out->data()[c]; + + for (int y = 0; y < cropped_size.height; ++y) { + memcpy (out_p, in_p + crop_left_in_bytes, cropped_width_in_bytes); + in_p += stride()[c]; + out_p += out->stride()[c]; + } + } + + return out; +} + +/** Blacken a YUV image whose bits per pixel is rounded up to 16 */ +void +Image::yuv_16_black (uint16_t v) +{ + memset (data()[0], 0, lines(0) * stride()[0]); + for (int i = 1; i < 3; ++i) { + int16_t* p = reinterpret_cast (data()[i]); + for (int y = 0; y < size().height; ++y) { + for (int x = 0; x < line_size()[i] / 2; ++x) { + p[x] = v; + } + p += stride()[i] / 2; + } + } +} + +uint16_t +Image::swap_16 (uint16_t v) +{ + return ((v >> 8) & 0xff) | ((v & 0xff) << 8); +} + void Image::make_black () { + /* U/V black value for 8-bit colour */ + static uint8_t const eight_bit_uv = (1 << 7) - 1; + + /* U/V black value for 9-bit colour */ + static uint16_t const nine_bit_uv = (1 << 8) - 1; + + /* U/V black value for 10-bit colour */ + static uint16_t const ten_bit_uv = (1 << 9) - 1; + switch (_pixel_format) { case PIX_FMT_YUV420P: - memset (data()[0], 0, lines(0) * line_size()[0]); - memset (data()[1], 0x80, lines(1) * line_size()[1]); - memset (data()[2], 0x80, lines(2) * line_size()[2]); + case PIX_FMT_YUV422P: + case PIX_FMT_YUV444P: + memset (data()[0], 0, lines(0) * stride()[0]); + memset (data()[1], eight_bit_uv, lines(1) * stride()[1]); + memset (data()[2], eight_bit_uv, lines(2) * stride()[2]); break; + case PIX_FMT_YUV422P9LE: + case PIX_FMT_YUV444P9LE: + yuv_16_black (nine_bit_uv); + break; + + case PIX_FMT_YUV422P9BE: + case PIX_FMT_YUV444P9BE: + yuv_16_black (swap_16 (nine_bit_uv)); + break; + + case PIX_FMT_YUV422P10LE: + case PIX_FMT_YUV444P10LE: + yuv_16_black (ten_bit_uv); + break; + + case PIX_FMT_YUV444P10BE: + case PIX_FMT_YUV422P10BE: + yuv_16_black (swap_16 (ten_bit_uv)); + case PIX_FMT_RGB24: - memset (data()[0], 0, lines(0) * line_size()[0]); + memset (data()[0], 0, lines(0) * stride()[0]); + break; + + case PIX_FMT_UYVY422: + { + int const Y = lines(0); + int const X = line_size()[0]; + uint8_t* p = data()[0]; + for (int y = 0; y < Y; ++y) { + for (int x = 0; x < X / 4; ++x) { + *p++ = eight_bit_uv; // Cb + *p++ = 0; // Y0 + *p++ = eight_bit_uv; // Cr + *p++ = 0; // Y1 + } + } break; + } default: - assert (false); + throw PixelFormatError (N_("make_black()"), _pixel_format); + } +} + +void +Image::alpha_blend (shared_ptr other, Position position) +{ + /* Only implemented for RGBA onto RGB24 so far */ + assert (_pixel_format == PIX_FMT_RGB24 && other->pixel_format() == PIX_FMT_RGBA); + + int start_tx = position.x; + int start_ox = 0; + + if (start_tx < 0) { + start_ox = -start_tx; + start_tx = 0; + } + + int start_ty = position.y; + int start_oy = 0; + + if (start_ty < 0) { + start_oy = -start_ty; + start_ty = 0; + } + + for (int ty = start_ty, oy = start_oy; ty < size().height && oy < other->size().height; ++ty, ++oy) { + uint8_t* tp = data()[0] + ty * stride()[0] + position.x * 3; + uint8_t* op = other->data()[0] + oy * other->stride()[0]; + for (int tx = start_tx, ox = start_ox; tx < size().width && ox < other->size().width; ++tx, ++ox) { + float const alpha = float (op[3]) / 255; + tp[0] = (tp[0] * (1 - alpha)) + op[0] * alpha; + tp[1] = (tp[1] * (1 - alpha)) + op[1] * alpha; + tp[2] = (tp[2] * (1 - alpha)) + op[2] * alpha; + tp += 3; + op += 4; + } + } +} + +void +Image::read_from_socket (shared_ptr socket) +{ + for (int i = 0; i < components(); ++i) { + uint8_t* p = data()[i]; + for (int y = 0; y < lines(i); ++y) { + socket->read (p, line_size()[i]); + p += stride()[i]; + } + } +} + +void +Image::write_to_socket (shared_ptr socket) const +{ + for (int i = 0; i < components(); ++i) { + uint8_t* p = data()[i]; + for (int y = 0; y < lines(i); ++y) { + socket->write (p, line_size()[i]); + p += stride()[i]; + } } } + +float +Image::bytes_per_pixel (int c) const +{ + if (c == 3) { + return 0; + } + + switch (_pixel_format) { + case PIX_FMT_RGB24: + if (c == 0) { + return 3; + } else { + return 0; + } + case PIX_FMT_RGBA: + if (c == 0) { + return 4; + } else { + return 0; + } + case PIX_FMT_YUV420P: + case PIX_FMT_YUV422P: + if (c == 0) { + return 1; + } else { + return 0.5; + } + case PIX_FMT_YUV422P10LE: + if (c == 0) { + return 2; + } else { + return 1; + } + case PIX_FMT_UYVY422: + return 2; + case PIX_FMT_YUV444P: + return 3; + case PIX_FMT_YUV444P9BE: + case PIX_FMT_YUV444P9LE: + case PIX_FMT_YUV444P10LE: + case PIX_FMT_YUV444P10BE: + return 6; + default: + throw PixelFormatError (N_("bytes_per_pixel()"), _pixel_format); + } + + return 0; +} + + /** Construct a SimpleImage of a given size and format, allocating memory * as required. * * @param p Pixel format. * @param s Size in pixels. */ -SimpleImage::SimpleImage (PixelFormat p, Size s) +SimpleImage::SimpleImage (AVPixelFormat p, libdcp::Size s, bool aligned) : Image (p) , _size (s) + , _aligned (aligned) +{ + allocate (); +} + +void +SimpleImage::allocate () { - _data = (uint8_t **) av_malloc (components() * sizeof (uint8_t *)); - _line_size = (int *) av_malloc (components() * sizeof (int)); + _data = (uint8_t **) av_malloc (4 * sizeof (uint8_t *)); + _data[0] = _data[1] = _data[2] = _data[3] = 0; + + _line_size = (int *) av_malloc (4 * sizeof (int)); + _line_size[0] = _line_size[1] = _line_size[2] = _line_size[3] = 0; + _stride = (int *) av_malloc (4 * sizeof (int)); + _stride[0] = _stride[1] = _stride[2] = _stride[3] = 0; + for (int i = 0; i < components(); ++i) { - _data[i] = 0; - _line_size[i] = 0; + _line_size[i] = _size.width * bytes_per_pixel(i); + _stride[i] = stride_round_up (i, _line_size, _aligned ? 32 : 1); + _data[i] = (uint8_t *) av_malloc (_stride[i] * lines (i)); } } -/** Destroy a SimpleImage */ -SimpleImage::~SimpleImage () +SimpleImage::SimpleImage (SimpleImage const & other) + : Image (other) { + _size = other._size; + _aligned = other._aligned; + + allocate (); + for (int i = 0; i < components(); ++i) { - av_free (_data[i]); + uint8_t* p = _data[i]; + uint8_t* q = other._data[i]; + for (int j = 0; j < lines(i); ++j) { + memcpy (p, q, _line_size[i]); + p += stride()[i]; + q += other.stride()[i]; + } } - - av_free (_data); - av_free (_line_size); } -/** Set the size in bytes of each horizontal line of a given component. - * @param i Component index. - * @param s Size of line in bytes. - */ -void -SimpleImage::set_line_size (int i, int s) +SimpleImage::SimpleImage (shared_ptr other) + : Image (*other.get()) { - _line_size[i] = s; - _data[i] = (uint8_t *) av_malloc (s * lines (i)); -} + _size = other->size (); + _aligned = true; -uint8_t ** -SimpleImage::data () const -{ - return _data; + allocate (); + + for (int i = 0; i < components(); ++i) { + assert(line_size()[i] == other->line_size()[i]); + uint8_t* p = _data[i]; + uint8_t* q = other->data()[i]; + for (int j = 0; j < lines(i); ++j) { + memcpy (p, q, line_size()[i]); + p += stride()[i]; + q += other->stride()[i]; + } + } } -int * -SimpleImage::line_size () const +SimpleImage& +SimpleImage::operator= (SimpleImage const & other) { - return _line_size; + if (this == &other) { + return *this; + } + + SimpleImage tmp (other); + swap (tmp); + return *this; } -Size -SimpleImage::size () const +void +SimpleImage::swap (SimpleImage & other) { - return _size; -} + Image::swap (other); + + std::swap (_size, other._size); + for (int i = 0; i < 4; ++i) { + std::swap (_data[i], other._data[i]); + std::swap (_line_size[i], other._line_size[i]); + std::swap (_stride[i], other._stride[i]); + } -FilterBufferImage::FilterBufferImage (PixelFormat p, AVFilterBufferRef* b) - : Image (p) - , _buffer (b) + std::swap (_aligned, other._aligned); +} + +/** Destroy a SimpleImage */ +SimpleImage::~SimpleImage () { + for (int i = 0; i < components(); ++i) { + av_free (_data[i]); + } + av_free (_data); + av_free (_line_size); + av_free (_stride); } -FilterBufferImage::~FilterBufferImage () +uint8_t ** +SimpleImage::data () const { - avfilter_unref_buffer (_buffer); + return _data; } -uint8_t ** -FilterBufferImage::data () const +int * +SimpleImage::line_size () const { - return _buffer->data; + return _line_size; } int * -FilterBufferImage::line_size () const +SimpleImage::stride () const { - return _buffer->linesize; + return _stride; } -Size -FilterBufferImage::size () const +libdcp::Size +SimpleImage::size () const { - return Size (_buffer->video->w, _buffer->video->h); + return _size; } -/** XXX: this could be generalised to use any format, but I don't - * understand how avpicture_fill is supposed to be called with - * multi-planar images. - */ -RGBFrameImage::RGBFrameImage (Size s) - : Image (PIX_FMT_RGB24) - , _size (s) +bool +SimpleImage::aligned () const { - _frame = avcodec_alloc_frame (); - if (_frame == 0) { - throw EncodeError ("could not allocate frame"); - } + return _aligned; +} - _data = (uint8_t *) av_malloc (size().width * size().height * 3); - avpicture_fill ((AVPicture *) _frame, _data, PIX_FMT_RGB24, size().width, size().height); - _frame->width = size().width; - _frame->height = size().height; - _frame->format = PIX_FMT_RGB24; +FrameImage::FrameImage (AVFrame* frame) + : Image (static_cast (frame->format)) + , _frame (frame) +{ + _line_size = (int *) av_malloc (4 * sizeof (int)); + _line_size[0] = _line_size[1] = _line_size[2] = _line_size[3] = 0; + + for (int i = 0; i < components(); ++i) { + _line_size[i] = size().width * bytes_per_pixel(i); + } } -RGBFrameImage::~RGBFrameImage () +FrameImage::~FrameImage () { - av_free (_data); - av_free (_frame); + av_frame_free (&_frame); + av_free (_line_size); } uint8_t ** -RGBFrameImage::data () const +FrameImage::data () const { return _frame->data; } int * -RGBFrameImage::line_size () const +FrameImage::line_size () const { - return _frame->linesize; + return _line_size; } -Size -RGBFrameImage::size () const +int * +FrameImage::stride () const { - return _size; + /* AVFrame's `linesize' is what we call `stride' */ + return _frame->linesize; } -PostProcessImage::PostProcessImage (PixelFormat p, Size s) - : Image (p) - , _size (s) +libdcp::Size +FrameImage::size () const { - _data = new uint8_t*[4]; - _line_size = new int[4]; - - for (int i = 0; i < 4; ++i) { - _data[i] = (uint8_t *) av_malloc (s.width * s.height); - _line_size[i] = s.width; - } + return libdcp::Size (_frame->width, _frame->height); } -PostProcessImage::~PostProcessImage () +bool +FrameImage::aligned () const { - for (int i = 0; i < 4; ++i) { - av_free (_data[i]); - } - - delete[] _data; - delete[] _line_size; + return true; } -uint8_t ** -PostProcessImage::data () const +RGBPlusAlphaImage::RGBPlusAlphaImage (shared_ptr im) + : SimpleImage (im->pixel_format(), im->size(), false) { - return _data; -} + assert (im->pixel_format() == PIX_FMT_RGBA); + + _alpha = (uint8_t *) av_malloc (im->size().width * im->size().height); + + uint8_t* in = im->data()[0]; + uint8_t* out = data()[0]; + uint8_t* out_alpha = _alpha; + for (int y = 0; y < im->size().height; ++y) { + uint8_t* in_r = in; + for (int x = 0; x < im->size().width; ++x) { + *out++ = *in_r++; + *out++ = *in_r++; + *out++ = *in_r++; + *out_alpha++ = *in_r++; + } -int * -PostProcessImage::line_size () const -{ - return _line_size; + in += im->stride()[0]; + } } -Size -PostProcessImage::size () const +RGBPlusAlphaImage::~RGBPlusAlphaImage () { - return _size; + av_free (_alpha); } +