/* Copyright (C) 2012 Carl Hetherington This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ /** @file src/image.cc * @brief A set of classes to describe video images. */ #include #include #include #include #include #include #include extern "C" { #include #include #include #include #include #include } #include "image.h" #include "exceptions.h" #include "scaler.h" using namespace std; using namespace boost; /** @param n Component index. * @return Number of lines in the image for the given component. */ int Image::lines (int n) const { switch (_pixel_format) { case PIX_FMT_YUV420P: if (n == 0) { return size().height; } else { return size().height / 2; } break; case PIX_FMT_RGB24: case PIX_FMT_RGBA: case PIX_FMT_YUV422P10LE: case PIX_FMT_YUV422P: return size().height; default: assert (false); } return 0; } /** @return Number of components */ int Image::components () const { switch (_pixel_format) { case PIX_FMT_YUV420P: case PIX_FMT_YUV422P10LE: case PIX_FMT_YUV422P: return 3; case PIX_FMT_RGB24: case PIX_FMT_RGBA: return 1; default: assert (false); } return 0; } shared_ptr Image::scale (Size out_size, Scaler const * scaler) const { assert (scaler); shared_ptr scaled (new AlignedImage (pixel_format(), out_size)); struct SwsContext* scale_context = sws_getContext ( size().width, size().height, pixel_format(), out_size.width, out_size.height, pixel_format(), scaler->ffmpeg_id (), 0, 0, 0 ); sws_scale ( scale_context, data(), stride(), 0, size().height, scaled->data(), scaled->stride() ); sws_freeContext (scale_context); return scaled; } /** Scale this image to a given size and convert it to RGB. * @param out_size Output image size in pixels. * @param scaler Scaler to use. */ shared_ptr Image::scale_and_convert_to_rgb (Size out_size, int padding, Scaler const * scaler) const { assert (scaler); Size content_size = out_size; content_size.width -= (padding * 2); shared_ptr rgb (new AlignedImage (PIX_FMT_RGB24, content_size)); struct SwsContext* scale_context = sws_getContext ( size().width, size().height, pixel_format(), content_size.width, content_size.height, PIX_FMT_RGB24, scaler->ffmpeg_id (), 0, 0, 0 ); /* Scale and convert to RGB from whatever its currently in (which may be RGB) */ sws_scale ( scale_context, data(), stride(), 0, size().height, rgb->data(), rgb->stride() ); /* Put the image in the right place in a black frame if are padding; this is a bit grubby and expensive, but probably inconsequential in the great scheme of things. */ if (padding > 0) { shared_ptr padded_rgb (new AlignedImage (PIX_FMT_RGB24, out_size)); padded_rgb->make_black (); /* XXX: we are cheating a bit here; we know the frame is RGB so we can make assumptions about its composition. */ uint8_t* p = padded_rgb->data()[0] + padding * 3; uint8_t* q = rgb->data()[0]; for (int j = 0; j < rgb->lines(0); ++j) { memcpy (p, q, rgb->line_size()[0]); p += padded_rgb->stride()[0]; q += rgb->stride()[0]; } rgb = padded_rgb; } sws_freeContext (scale_context); return rgb; } /** Run a FFmpeg post-process on this image and return the processed version. * @param pp Flags for the required set of post processes. * @return Post-processed image. */ shared_ptr Image::post_process (string pp) const { shared_ptr out (new AlignedImage (pixel_format(), size ())); int pp_format = 0; switch (pixel_format()) { case PIX_FMT_YUV420P: pp_format = PP_FORMAT_420; break; case PIX_FMT_YUV422P10LE: case PIX_FMT_YUV422P: pp_format = PP_FORMAT_422; break; default: assert (false); } pp_mode* mode = pp_get_mode_by_name_and_quality (pp.c_str (), PP_QUALITY_MAX); pp_context* context = pp_get_context (size().width, size().height, pp_format | PP_CPU_CAPS_MMX2); pp_postprocess ( (const uint8_t **) data(), stride(), out->data(), out->stride(), size().width, size().height, 0, 0, mode, context, 0 ); pp_free_mode (mode); pp_free_context (context); return out; } void Image::make_black () { switch (_pixel_format) { case PIX_FMT_YUV420P: case PIX_FMT_YUV422P10LE: case PIX_FMT_YUV422P: memset (data()[0], 0, lines(0) * stride()[0]); memset (data()[1], 0x80, lines(1) * stride()[1]); memset (data()[2], 0x80, lines(2) * stride()[2]); break; case PIX_FMT_RGB24: memset (data()[0], 0, lines(0) * stride()[0]); break; default: assert (false); } } void Image::alpha_blend (shared_ptr other, Position position) { /* Only implemented for RGBA onto RGB24 so far */ assert (_pixel_format == PIX_FMT_RGB24 && other->pixel_format() == PIX_FMT_RGBA); int start_tx = position.x; int start_ox = 0; if (start_tx < 0) { start_ox = -start_tx; start_tx = 0; } int start_ty = position.y; int start_oy = 0; if (start_ty < 0) { start_oy = -start_ty; start_ty = 0; } for (int ty = start_ty, oy = start_oy; ty < size().height && oy < other->size().height; ++ty, ++oy) { uint8_t* tp = data()[0] + ty * stride()[0] + position.x * 3; uint8_t* op = other->data()[0] + oy * other->stride()[0]; for (int tx = start_tx, ox = start_ox; tx < size().width && ox < other->size().width; ++tx, ++ox) { float const alpha = float (op[3]) / 255; tp[0] = (tp[0] * (1 - alpha)) + op[0] * alpha; tp[1] = (tp[1] * (1 - alpha)) + op[1] * alpha; tp[2] = (tp[2] * (1 - alpha)) + op[2] * alpha; tp += 3; op += 4; } } } void Image::read_from_socket (shared_ptr socket) { for (int i = 0; i < components(); ++i) { uint8_t* p = data()[i]; for (int y = 0; y < lines(i); ++y) { socket->read_definite_and_consume (p, line_size()[i], 30); p += stride()[i]; } } } void Image::write_to_socket (shared_ptr socket) const { for (int i = 0; i < components(); ++i) { uint8_t* p = data()[i]; for (int y = 0; y < lines(i); ++y) { socket->write (p, line_size()[i], 30); p += stride()[i]; } } } /** Construct a SimpleImage of a given size and format, allocating memory * as required. * * @param p Pixel format. * @param s Size in pixels. */ SimpleImage::SimpleImage (AVPixelFormat p, Size s, function rounder) : Image (p) , _size (s) { _data = (uint8_t **) av_malloc (4 * sizeof (uint8_t *)); _data[0] = _data[1] = _data[2] = _data[3] = 0; _line_size = (int *) av_malloc (4 * sizeof (int)); _line_size[0] = _line_size[1] = _line_size[2] = _line_size[3] = 0; _stride = (int *) av_malloc (4 * sizeof (int)); _stride[0] = _stride[1] = _stride[2] = _stride[3] = 0; switch (p) { case PIX_FMT_RGB24: _line_size[0] = s.width * 3; break; case PIX_FMT_RGBA: _line_size[0] = s.width * 4; break; case PIX_FMT_YUV420P: case PIX_FMT_YUV422P: _line_size[0] = s.width; _line_size[1] = s.width / 2; _line_size[2] = s.width / 2; break; case PIX_FMT_YUV422P10LE: _line_size[0] = s.width * 2; _line_size[1] = s.width; _line_size[2] = s.width; break; default: assert (false); } for (int i = 0; i < components(); ++i) { _stride[i] = rounder (_line_size[i]); _data[i] = (uint8_t *) av_malloc (_stride[i] * lines (i)); } } /** Destroy a SimpleImage */ SimpleImage::~SimpleImage () { for (int i = 0; i < components(); ++i) { av_free (_data[i]); } av_free (_data); av_free (_line_size); av_free (_stride); } uint8_t ** SimpleImage::data () const { return _data; } int * SimpleImage::line_size () const { return _line_size; } int * SimpleImage::stride () const { return _stride; } Size SimpleImage::size () const { return _size; } AlignedImage::AlignedImage (AVPixelFormat f, Size s) : SimpleImage (f, s, boost::bind (round_up, _1, 32)) { } CompactImage::CompactImage (AVPixelFormat f, Size s) : SimpleImage (f, s, boost::bind (round_up, _1, 1)) { } CompactImage::CompactImage (shared_ptr im) : SimpleImage (im->pixel_format(), im->size(), boost::bind (round_up, _1, 1)) { assert (components() == im->components()); for (int c = 0; c < components(); ++c) { assert (line_size()[c] == im->line_size()[c]); uint8_t* t = data()[c]; uint8_t* o = im->data()[c]; for (int y = 0; y < lines(c); ++y) { memcpy (t, o, line_size()[c]); t += stride()[c]; o += im->stride()[c]; } } } FilterBufferImage::FilterBufferImage (AVPixelFormat p, AVFilterBufferRef* b) : Image (p) , _buffer (b) { } FilterBufferImage::~FilterBufferImage () { avfilter_unref_buffer (_buffer); } uint8_t ** FilterBufferImage::data () const { return _buffer->data; } int * FilterBufferImage::line_size () const { return _buffer->linesize; } int * FilterBufferImage::stride () const { /* XXX? */ return _buffer->linesize; } Size FilterBufferImage::size () const { return Size (_buffer->video->w, _buffer->video->h); } /** XXX: this could be generalised to use any format, but I don't * understand how avpicture_fill is supposed to be called with * multi-planar images. */ RGBFrameImage::RGBFrameImage (Size s) : Image (PIX_FMT_RGB24) , _size (s) { _frame = avcodec_alloc_frame (); if (_frame == 0) { throw EncodeError ("could not allocate frame"); } _data = (uint8_t *) av_malloc (size().width * size().height * 3); avpicture_fill ((AVPicture *) _frame, _data, PIX_FMT_RGB24, size().width, size().height); _frame->width = size().width; _frame->height = size().height; _frame->format = PIX_FMT_RGB24; } RGBFrameImage::~RGBFrameImage () { av_free (_data); av_free (_frame); } uint8_t ** RGBFrameImage::data () const { return _frame->data; } int * RGBFrameImage::line_size () const { return _frame->linesize; } int * RGBFrameImage::stride () const { /* XXX? */ return line_size (); } Size RGBFrameImage::size () const { return _size; }