diff options
| author | Carl Hetherington <cth@carlh.net> | 2015-08-24 16:43:38 +0100 |
|---|---|---|
| committer | Carl Hetherington <cth@carlh.net> | 2015-08-25 15:22:38 +0100 |
| commit | a978f3ac575f1af017002c861480d5203cf0a34e (patch) | |
| tree | 5cd7acdcf7b7f493f3d0047ba96adaccf8509a7e /src/lib | |
| parent | 17d4990d0fca0f38268441a73c73d9e340b4bc85 (diff) | |
Add video waveform viewer.
Diffstat (limited to 'src/lib')
| -rw-r--r-- | src/lib/dcp_video.cc | 24 | ||||
| -rw-r--r-- | src/lib/dcp_video.h | 2 | ||||
| -rw-r--r-- | src/lib/image.cc | 121 | ||||
| -rw-r--r-- | src/lib/image.h | 5 |
4 files changed, 98 insertions, 54 deletions
diff --git a/src/lib/dcp_video.cc b/src/lib/dcp_video.cc index f4c6222e2..e02bcffca 100644 --- a/src/lib/dcp_video.cc +++ b/src/lib/dcp_video.cc @@ -91,27 +91,35 @@ DCPVideo::DCPVideo (shared_ptr<const PlayerVideo> frame, shared_ptr<const cxml:: _resolution = Resolution (node->optional_number_child<int>("Resolution").get_value_or (RESOLUTION_2K)); } -/** J2K-encode this frame on the local host. - * @return Encoded data. - */ -Data -DCPVideo::encode_locally (dcp::NoteHandler note) +shared_ptr<dcp::OpenJPEGImage> +DCPVideo::convert_to_xyz (shared_ptr<const PlayerVideo> frame, dcp::NoteHandler note) { shared_ptr<dcp::OpenJPEGImage> xyz; - shared_ptr<Image> image = _frame->image (AV_PIX_FMT_RGB48LE, note); - if (_frame->colour_conversion()) { + shared_ptr<Image> image = frame->image (AV_PIX_FMT_RGB48LE, note); + if (frame->colour_conversion()) { xyz = dcp::rgb_to_xyz ( image->data()[0], image->size(), image->stride()[0], - _frame->colour_conversion().get(), + frame->colour_conversion().get(), note ); } else { xyz = dcp::xyz_to_xyz (image->data()[0], image->size(), image->stride()[0]); } + return xyz; +} + +/** J2K-encode this frame on the local host. + * @return Encoded data. + */ +Data +DCPVideo::encode_locally (dcp::NoteHandler note) +{ + shared_ptr<dcp::OpenJPEGImage> xyz = convert_to_xyz (_frame, note); + /* Set the max image and component sizes based on frame_rate */ int max_cs_len = ((float) _j2k_bandwidth) / 8 / _frames_per_second; if (_frame->eyes() == EYES_LEFT || _frame->eyes() == EYES_RIGHT) { diff --git a/src/lib/dcp_video.h b/src/lib/dcp_video.h index 8fd668b00..4c7714192 100644 --- a/src/lib/dcp_video.h +++ b/src/lib/dcp_video.h @@ -60,6 +60,8 @@ public: bool same (boost::shared_ptr<const DCPVideo> other) const; + static boost::shared_ptr<dcp::OpenJPEGImage> convert_to_xyz (boost::shared_ptr<const PlayerVideo> frame, dcp::NoteHandler note); + private: void add_metadata (xmlpp::Element *) const; diff --git a/src/lib/image.cc b/src/lib/image.cc index 55f49daf8..0c7a0ef0d 100644 --- a/src/lib/image.cc +++ b/src/lib/image.cc @@ -61,15 +61,26 @@ Image::line_factor (int n) const } /** @param n Component index. - * @return Number of lines in the image for the given component. + * @return Number of samples (i.e. pixels, unless sub-sampled) in each direction for this component. */ -int -Image::lines (int n) const +dcp::Size +Image::sample_size (int n) const { - return rint (ceil (static_cast<double>(size().height) / line_factor (n))); + int horizontal_factor = 1; + if (n > 0) { + AVPixFmtDescriptor const * d = av_pix_fmt_desc_get (_pixel_format); + if (!d) { + throw PixelFormatError ("sample_size()", _pixel_format); + } + horizontal_factor = pow (2.0f, d->log2_chroma_w); + } + + return dcp::Size ( + rint (ceil (static_cast<double>(size().width) / horizontal_factor)), + rint (ceil (static_cast<double>(size().height) / line_factor (n))) + ); } -/** @return Number of components */ int Image::components () const { @@ -78,6 +89,18 @@ Image::components () const throw PixelFormatError ("components()", _pixel_format); } + return d->nb_components; +} + +/** @return Number of planes */ +int +Image::planes () const +{ + AVPixFmtDescriptor const * d = av_pix_fmt_desc_get(_pixel_format); + if (!d) { + throw PixelFormatError ("planes()", _pixel_format); + } + if ((d->flags & PIX_FMT_PLANAR) == 0) { return 1; } @@ -136,8 +159,8 @@ Image::crop_scale_window ( } /* Prepare input data pointers with crop */ - uint8_t* scale_in_data[components()]; - for (int c = 0; c < components(); ++c) { + uint8_t* scale_in_data[planes()]; + for (int c = 0; c < planes(); ++c) { /* To work out the crop in bytes, start by multiplying the crop by the (average) bytes per pixel. Then round down so that we don't crop a subsampled pixel until @@ -150,8 +173,8 @@ Image::crop_scale_window ( /* Corner of the image within out_size */ Position<int> const corner ((out_size.width - inter_size.width) / 2, (out_size.height - inter_size.height) / 2); - uint8_t* scale_out_data[out->components()]; - for (int c = 0; c < out->components(); ++c) { + uint8_t* scale_out_data[out->planes()]; + for (int c = 0; c < out->planes(); ++c) { scale_out_data[c] = out->data()[c] + int (rint (out->bytes_per_pixel(c) * corner.x)) + out->stride()[c] * corner.y; } @@ -212,10 +235,11 @@ Image::scale (dcp::Size out_size, dcp::YUVToRGB yuv_to_rgb, AVPixelFormat out_fo void Image::yuv_16_black (uint16_t v, bool alpha) { - memset (data()[0], 0, lines(0) * stride()[0]); + memset (data()[0], 0, sample_size(0).height * stride()[0]); for (int i = 1; i < 3; ++i) { int16_t* p = reinterpret_cast<int16_t*> (data()[i]); - for (int y = 0; y < lines(i); ++y) { + int const lines = sample_size(i).height; + for (int y = 0; y < lines; ++y) { /* We divide by 2 here because we are writing 2 bytes at a time */ for (int x = 0; x < line_size()[i] / 2; ++x) { p[x] = v; @@ -225,7 +249,7 @@ Image::yuv_16_black (uint16_t v, bool alpha) } if (alpha) { - memset (data()[3], 0, lines(3) * stride()[3]); + memset (data()[3], 0, sample_size(3).height * stride()[3]); } } @@ -252,17 +276,17 @@ Image::make_black () case PIX_FMT_YUV422P: case PIX_FMT_YUV444P: case PIX_FMT_YUV411P: - memset (data()[0], 0, lines(0) * stride()[0]); - memset (data()[1], eight_bit_uv, lines(1) * stride()[1]); - memset (data()[2], eight_bit_uv, lines(2) * stride()[2]); + memset (data()[0], 0, sample_size(0).height * stride()[0]); + memset (data()[1], eight_bit_uv, sample_size(1).height * stride()[1]); + memset (data()[2], eight_bit_uv, sample_size(2).height * stride()[2]); break; case PIX_FMT_YUVJ420P: case PIX_FMT_YUVJ422P: case PIX_FMT_YUVJ444P: - memset (data()[0], 0, lines(0) * stride()[0]); - memset (data()[1], eight_bit_uv + 1, lines(1) * stride()[1]); - memset (data()[2], eight_bit_uv + 1, lines(2) * stride()[2]); + memset (data()[0], 0, sample_size(0).height * stride()[0]); + memset (data()[1], eight_bit_uv + 1, sample_size(1).height * stride()[1]); + memset (data()[2], eight_bit_uv + 1, sample_size(2).height * stride()[2]); break; case PIX_FMT_YUV422P9LE: @@ -334,12 +358,12 @@ Image::make_black () case PIX_FMT_RGB555LE: case PIX_FMT_RGB48LE: case PIX_FMT_RGB48BE: - memset (data()[0], 0, lines(0) * stride()[0]); + memset (data()[0], 0, sample_size(0).height * stride()[0]); break; case PIX_FMT_UYVY422: { - int const Y = lines(0); + int const Y = sample_size(0).height; int const X = line_size()[0]; uint8_t* p = data()[0]; for (int y = 0; y < Y; ++y) { @@ -365,7 +389,7 @@ Image::make_transparent () throw PixelFormatError ("make_transparent()", _pixel_format); } - memset (data()[0], 0, lines(0) * stride()[0]); + memset (data()[0], 0, sample_size(0).height * stride()[0]); } void @@ -471,9 +495,10 @@ Image::copy (shared_ptr<const Image> other, Position<int> position) void Image::read_from_socket (shared_ptr<Socket> socket) { - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { uint8_t* p = data()[i]; - for (int y = 0; y < lines(i); ++y) { + int const lines = sample_size(i).height; + for (int y = 0; y < lines; ++y) { socket->read (p, line_size()[i]); p += stride()[i]; } @@ -483,9 +508,10 @@ Image::read_from_socket (shared_ptr<Socket> socket) void Image::write_to_socket (shared_ptr<Socket> socket) const { - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { uint8_t* p = data()[i]; - for (int y = 0; y < lines(i); ++y) { + int const lines = sample_size(i).height; + for (int y = 0; y < lines; ++y) { socket->write (p, line_size()[i]); p += stride()[i]; } @@ -500,7 +526,7 @@ Image::bytes_per_pixel (int c) const throw PixelFormatError ("bytes_per_pixel()", _pixel_format); } - if (c >= components()) { + if (c >= planes()) { return 0; } @@ -551,7 +577,7 @@ Image::allocate () _stride = (int *) wrapped_av_malloc (4 * sizeof (int)); _stride[0] = _stride[1] = _stride[2] = _stride[3] = 0; - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { _line_size[i] = ceil (_size.width * bytes_per_pixel(i)); _stride[i] = stride_round_up (i, _line_size, _aligned ? 32 : 1); @@ -569,7 +595,7 @@ Image::allocate () so I'll just over-allocate by 32 bytes and have done with it. Empirical testing suggests that it works. */ - _data[i] = (uint8_t *) wrapped_av_malloc (_stride[i] * lines (i) + 32); + _data[i] = (uint8_t *) wrapped_av_malloc (_stride[i] * sample_size(i).height + 32); } } @@ -580,10 +606,11 @@ Image::Image (Image const & other) { allocate (); - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { uint8_t* p = _data[i]; uint8_t* q = other._data[i]; - for (int j = 0; j < lines(i); ++j) { + int const lines = sample_size(i).height; + for (int j = 0; j < lines; ++j) { memcpy (p, q, _line_size[i]); p += stride()[i]; q += other.stride()[i]; @@ -598,10 +625,11 @@ Image::Image (AVFrame* frame) { allocate (); - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { uint8_t* p = _data[i]; uint8_t* q = frame->data[i]; - for (int j = 0; j < lines(i); ++j) { + int const lines = sample_size(i).height; + for (int j = 0; j < lines; ++j) { memcpy (p, q, _line_size[i]); p += stride()[i]; /* AVFrame's linesize is what we call `stride' */ @@ -617,11 +645,12 @@ Image::Image (shared_ptr<const Image> other, bool aligned) { allocate (); - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { DCPOMATIC_ASSERT (line_size()[i] == other->line_size()[i]); uint8_t* p = _data[i]; uint8_t* q = other->data()[i]; - for (int j = 0; j < lines(i); ++j) { + int const lines = sample_size(i).height; + for (int j = 0; j < lines; ++j) { memcpy (p, q, line_size()[i]); p += stride()[i]; q += other->stride()[i]; @@ -659,7 +688,7 @@ Image::swap (Image & other) /** Destroy a Image */ Image::~Image () { - for (int i = 0; i < components(); ++i) { + for (int i = 0; i < planes(); ++i) { av_free (_data[i]); } @@ -674,7 +703,7 @@ Image::data () const return _data; } -int * +int const * Image::line_size () const { return _line_size; @@ -726,18 +755,19 @@ merge (list<PositionImage> images) bool operator== (Image const & a, Image const & b) { - if (a.components() != b.components() || a.pixel_format() != b.pixel_format() || a.aligned() != b.aligned()) { + if (a.planes() != b.planes() || a.pixel_format() != b.pixel_format() || a.aligned() != b.aligned()) { return false; } - for (int c = 0; c < a.components(); ++c) { - if (a.lines(c) != b.lines(c) || a.line_size()[c] != b.line_size()[c] || a.stride()[c] != b.stride()[c]) { + for (int c = 0; c < a.planes(); ++c) { + if (a.sample_size(c).height != b.sample_size(c).height || a.line_size()[c] != b.line_size()[c] || a.stride()[c] != b.stride()[c]) { return false; } uint8_t* p = a.data()[c]; uint8_t* q = b.data()[c]; - for (int y = 0; y < a.lines(c); ++y) { + int const lines = a.sample_size(c).height; + for (int y = 0; y < lines; ++y) { if (memcmp (p, q, a.line_size()[c]) != 0) { return false; } @@ -773,7 +803,8 @@ Image::fade (float f) /* 8-bit */ for (int c = 0; c < 3; ++c) { uint8_t* p = data()[c]; - for (int y = 0; y < lines(c); ++y) { + int const lines = sample_size(c).height; + for (int y = 0; y < lines; ++y) { uint8_t* q = p; for (int x = 0; x < line_size()[c]; ++x) { *q = int (float (*q) * f); @@ -802,7 +833,8 @@ Image::fade (float f) int const stride_pixels = stride()[c] / 2; int const line_size_pixels = line_size()[c] / 2; uint16_t* p = reinterpret_cast<uint16_t*> (data()[c]); - for (int y = 0; y < lines(c); ++y) { + int const lines = sample_size(c).height; + for (int y = 0; y < lines; ++y) { uint16_t* q = p; for (int x = 0; x < line_size_pixels; ++x) { *q = int (float (*q) * f); @@ -832,7 +864,8 @@ Image::fade (float f) int const stride_pixels = stride()[c] / 2; int const line_size_pixels = line_size()[c] / 2; uint16_t* p = reinterpret_cast<uint16_t*> (data()[c]); - for (int y = 0; y < lines(c); ++y) { + int const lines = sample_size(c).height; + for (int y = 0; y < lines; ++y) { uint16_t* q = p; for (int x = 0; x < line_size_pixels; ++x) { *q = swap_16 (int (float (swap_16 (*q)) * f)); @@ -845,7 +878,7 @@ Image::fade (float f) case PIX_FMT_UYVY422: { - int const Y = lines(0); + int const Y = sample_size(0).height; int const X = line_size()[0]; uint8_t* p = data()[0]; for (int y = 0; y < Y; ++y) { diff --git a/src/lib/image.h b/src/lib/image.h index 492f6212e..fabcb5675 100644 --- a/src/lib/image.h +++ b/src/lib/image.h @@ -49,14 +49,15 @@ public: ~Image (); uint8_t * const * data () const; - int * line_size () const; + int const * line_size () const; int const * stride () const; dcp::Size size () const; bool aligned () const; + int planes () const; int components () const; int line_factor (int) const; - int lines (int) const; + dcp::Size sample_size (int) const; boost::shared_ptr<Image> scale (dcp::Size, dcp::YUVToRGB yuv_to_rgb, AVPixelFormat, bool aligned) const; boost::shared_ptr<Image> crop_scale_window (Crop c, dcp::Size, dcp::Size, dcp::YUVToRGB yuv_to_rgb, AVPixelFormat, bool aligned) const; |
