summaryrefslogtreecommitdiff
path: root/src/lib
diff options
context:
space:
mode:
authorCarl Hetherington <cth@carlh.net>2015-08-24 16:43:38 +0100
committerCarl Hetherington <cth@carlh.net>2015-08-25 15:22:38 +0100
commita978f3ac575f1af017002c861480d5203cf0a34e (patch)
tree5cd7acdcf7b7f493f3d0047ba96adaccf8509a7e /src/lib
parent17d4990d0fca0f38268441a73c73d9e340b4bc85 (diff)
Add video waveform viewer.
Diffstat (limited to 'src/lib')
-rw-r--r--src/lib/dcp_video.cc24
-rw-r--r--src/lib/dcp_video.h2
-rw-r--r--src/lib/image.cc121
-rw-r--r--src/lib/image.h5
4 files changed, 98 insertions, 54 deletions
diff --git a/src/lib/dcp_video.cc b/src/lib/dcp_video.cc
index f4c6222e2..e02bcffca 100644
--- a/src/lib/dcp_video.cc
+++ b/src/lib/dcp_video.cc
@@ -91,27 +91,35 @@ DCPVideo::DCPVideo (shared_ptr<const PlayerVideo> frame, shared_ptr<const cxml::
_resolution = Resolution (node->optional_number_child<int>("Resolution").get_value_or (RESOLUTION_2K));
}
-/** J2K-encode this frame on the local host.
- * @return Encoded data.
- */
-Data
-DCPVideo::encode_locally (dcp::NoteHandler note)
+shared_ptr<dcp::OpenJPEGImage>
+DCPVideo::convert_to_xyz (shared_ptr<const PlayerVideo> frame, dcp::NoteHandler note)
{
shared_ptr<dcp::OpenJPEGImage> xyz;
- shared_ptr<Image> image = _frame->image (AV_PIX_FMT_RGB48LE, note);
- if (_frame->colour_conversion()) {
+ shared_ptr<Image> image = frame->image (AV_PIX_FMT_RGB48LE, note);
+ if (frame->colour_conversion()) {
xyz = dcp::rgb_to_xyz (
image->data()[0],
image->size(),
image->stride()[0],
- _frame->colour_conversion().get(),
+ frame->colour_conversion().get(),
note
);
} else {
xyz = dcp::xyz_to_xyz (image->data()[0], image->size(), image->stride()[0]);
}
+ return xyz;
+}
+
+/** J2K-encode this frame on the local host.
+ * @return Encoded data.
+ */
+Data
+DCPVideo::encode_locally (dcp::NoteHandler note)
+{
+ shared_ptr<dcp::OpenJPEGImage> xyz = convert_to_xyz (_frame, note);
+
/* Set the max image and component sizes based on frame_rate */
int max_cs_len = ((float) _j2k_bandwidth) / 8 / _frames_per_second;
if (_frame->eyes() == EYES_LEFT || _frame->eyes() == EYES_RIGHT) {
diff --git a/src/lib/dcp_video.h b/src/lib/dcp_video.h
index 8fd668b00..4c7714192 100644
--- a/src/lib/dcp_video.h
+++ b/src/lib/dcp_video.h
@@ -60,6 +60,8 @@ public:
bool same (boost::shared_ptr<const DCPVideo> other) const;
+ static boost::shared_ptr<dcp::OpenJPEGImage> convert_to_xyz (boost::shared_ptr<const PlayerVideo> frame, dcp::NoteHandler note);
+
private:
void add_metadata (xmlpp::Element *) const;
diff --git a/src/lib/image.cc b/src/lib/image.cc
index 55f49daf8..0c7a0ef0d 100644
--- a/src/lib/image.cc
+++ b/src/lib/image.cc
@@ -61,15 +61,26 @@ Image::line_factor (int n) const
}
/** @param n Component index.
- * @return Number of lines in the image for the given component.
+ * @return Number of samples (i.e. pixels, unless sub-sampled) in each direction for this component.
*/
-int
-Image::lines (int n) const
+dcp::Size
+Image::sample_size (int n) const
{
- return rint (ceil (static_cast<double>(size().height) / line_factor (n)));
+ int horizontal_factor = 1;
+ if (n > 0) {
+ AVPixFmtDescriptor const * d = av_pix_fmt_desc_get (_pixel_format);
+ if (!d) {
+ throw PixelFormatError ("sample_size()", _pixel_format);
+ }
+ horizontal_factor = pow (2.0f, d->log2_chroma_w);
+ }
+
+ return dcp::Size (
+ rint (ceil (static_cast<double>(size().width) / horizontal_factor)),
+ rint (ceil (static_cast<double>(size().height) / line_factor (n)))
+ );
}
-/** @return Number of components */
int
Image::components () const
{
@@ -78,6 +89,18 @@ Image::components () const
throw PixelFormatError ("components()", _pixel_format);
}
+ return d->nb_components;
+}
+
+/** @return Number of planes */
+int
+Image::planes () const
+{
+ AVPixFmtDescriptor const * d = av_pix_fmt_desc_get(_pixel_format);
+ if (!d) {
+ throw PixelFormatError ("planes()", _pixel_format);
+ }
+
if ((d->flags & PIX_FMT_PLANAR) == 0) {
return 1;
}
@@ -136,8 +159,8 @@ Image::crop_scale_window (
}
/* Prepare input data pointers with crop */
- uint8_t* scale_in_data[components()];
- for (int c = 0; c < components(); ++c) {
+ uint8_t* scale_in_data[planes()];
+ for (int c = 0; c < planes(); ++c) {
/* To work out the crop in bytes, start by multiplying
the crop by the (average) bytes per pixel. Then
round down so that we don't crop a subsampled pixel until
@@ -150,8 +173,8 @@ Image::crop_scale_window (
/* Corner of the image within out_size */
Position<int> const corner ((out_size.width - inter_size.width) / 2, (out_size.height - inter_size.height) / 2);
- uint8_t* scale_out_data[out->components()];
- for (int c = 0; c < out->components(); ++c) {
+ uint8_t* scale_out_data[out->planes()];
+ for (int c = 0; c < out->planes(); ++c) {
scale_out_data[c] = out->data()[c] + int (rint (out->bytes_per_pixel(c) * corner.x)) + out->stride()[c] * corner.y;
}
@@ -212,10 +235,11 @@ Image::scale (dcp::Size out_size, dcp::YUVToRGB yuv_to_rgb, AVPixelFormat out_fo
void
Image::yuv_16_black (uint16_t v, bool alpha)
{
- memset (data()[0], 0, lines(0) * stride()[0]);
+ memset (data()[0], 0, sample_size(0).height * stride()[0]);
for (int i = 1; i < 3; ++i) {
int16_t* p = reinterpret_cast<int16_t*> (data()[i]);
- for (int y = 0; y < lines(i); ++y) {
+ int const lines = sample_size(i).height;
+ for (int y = 0; y < lines; ++y) {
/* We divide by 2 here because we are writing 2 bytes at a time */
for (int x = 0; x < line_size()[i] / 2; ++x) {
p[x] = v;
@@ -225,7 +249,7 @@ Image::yuv_16_black (uint16_t v, bool alpha)
}
if (alpha) {
- memset (data()[3], 0, lines(3) * stride()[3]);
+ memset (data()[3], 0, sample_size(3).height * stride()[3]);
}
}
@@ -252,17 +276,17 @@ Image::make_black ()
case PIX_FMT_YUV422P:
case PIX_FMT_YUV444P:
case PIX_FMT_YUV411P:
- memset (data()[0], 0, lines(0) * stride()[0]);
- memset (data()[1], eight_bit_uv, lines(1) * stride()[1]);
- memset (data()[2], eight_bit_uv, lines(2) * stride()[2]);
+ memset (data()[0], 0, sample_size(0).height * stride()[0]);
+ memset (data()[1], eight_bit_uv, sample_size(1).height * stride()[1]);
+ memset (data()[2], eight_bit_uv, sample_size(2).height * stride()[2]);
break;
case PIX_FMT_YUVJ420P:
case PIX_FMT_YUVJ422P:
case PIX_FMT_YUVJ444P:
- memset (data()[0], 0, lines(0) * stride()[0]);
- memset (data()[1], eight_bit_uv + 1, lines(1) * stride()[1]);
- memset (data()[2], eight_bit_uv + 1, lines(2) * stride()[2]);
+ memset (data()[0], 0, sample_size(0).height * stride()[0]);
+ memset (data()[1], eight_bit_uv + 1, sample_size(1).height * stride()[1]);
+ memset (data()[2], eight_bit_uv + 1, sample_size(2).height * stride()[2]);
break;
case PIX_FMT_YUV422P9LE:
@@ -334,12 +358,12 @@ Image::make_black ()
case PIX_FMT_RGB555LE:
case PIX_FMT_RGB48LE:
case PIX_FMT_RGB48BE:
- memset (data()[0], 0, lines(0) * stride()[0]);
+ memset (data()[0], 0, sample_size(0).height * stride()[0]);
break;
case PIX_FMT_UYVY422:
{
- int const Y = lines(0);
+ int const Y = sample_size(0).height;
int const X = line_size()[0];
uint8_t* p = data()[0];
for (int y = 0; y < Y; ++y) {
@@ -365,7 +389,7 @@ Image::make_transparent ()
throw PixelFormatError ("make_transparent()", _pixel_format);
}
- memset (data()[0], 0, lines(0) * stride()[0]);
+ memset (data()[0], 0, sample_size(0).height * stride()[0]);
}
void
@@ -471,9 +495,10 @@ Image::copy (shared_ptr<const Image> other, Position<int> position)
void
Image::read_from_socket (shared_ptr<Socket> socket)
{
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
uint8_t* p = data()[i];
- for (int y = 0; y < lines(i); ++y) {
+ int const lines = sample_size(i).height;
+ for (int y = 0; y < lines; ++y) {
socket->read (p, line_size()[i]);
p += stride()[i];
}
@@ -483,9 +508,10 @@ Image::read_from_socket (shared_ptr<Socket> socket)
void
Image::write_to_socket (shared_ptr<Socket> socket) const
{
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
uint8_t* p = data()[i];
- for (int y = 0; y < lines(i); ++y) {
+ int const lines = sample_size(i).height;
+ for (int y = 0; y < lines; ++y) {
socket->write (p, line_size()[i]);
p += stride()[i];
}
@@ -500,7 +526,7 @@ Image::bytes_per_pixel (int c) const
throw PixelFormatError ("bytes_per_pixel()", _pixel_format);
}
- if (c >= components()) {
+ if (c >= planes()) {
return 0;
}
@@ -551,7 +577,7 @@ Image::allocate ()
_stride = (int *) wrapped_av_malloc (4 * sizeof (int));
_stride[0] = _stride[1] = _stride[2] = _stride[3] = 0;
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
_line_size[i] = ceil (_size.width * bytes_per_pixel(i));
_stride[i] = stride_round_up (i, _line_size, _aligned ? 32 : 1);
@@ -569,7 +595,7 @@ Image::allocate ()
so I'll just over-allocate by 32 bytes and have done with it. Empirical
testing suggests that it works.
*/
- _data[i] = (uint8_t *) wrapped_av_malloc (_stride[i] * lines (i) + 32);
+ _data[i] = (uint8_t *) wrapped_av_malloc (_stride[i] * sample_size(i).height + 32);
}
}
@@ -580,10 +606,11 @@ Image::Image (Image const & other)
{
allocate ();
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
uint8_t* p = _data[i];
uint8_t* q = other._data[i];
- for (int j = 0; j < lines(i); ++j) {
+ int const lines = sample_size(i).height;
+ for (int j = 0; j < lines; ++j) {
memcpy (p, q, _line_size[i]);
p += stride()[i];
q += other.stride()[i];
@@ -598,10 +625,11 @@ Image::Image (AVFrame* frame)
{
allocate ();
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
uint8_t* p = _data[i];
uint8_t* q = frame->data[i];
- for (int j = 0; j < lines(i); ++j) {
+ int const lines = sample_size(i).height;
+ for (int j = 0; j < lines; ++j) {
memcpy (p, q, _line_size[i]);
p += stride()[i];
/* AVFrame's linesize is what we call `stride' */
@@ -617,11 +645,12 @@ Image::Image (shared_ptr<const Image> other, bool aligned)
{
allocate ();
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
DCPOMATIC_ASSERT (line_size()[i] == other->line_size()[i]);
uint8_t* p = _data[i];
uint8_t* q = other->data()[i];
- for (int j = 0; j < lines(i); ++j) {
+ int const lines = sample_size(i).height;
+ for (int j = 0; j < lines; ++j) {
memcpy (p, q, line_size()[i]);
p += stride()[i];
q += other->stride()[i];
@@ -659,7 +688,7 @@ Image::swap (Image & other)
/** Destroy a Image */
Image::~Image ()
{
- for (int i = 0; i < components(); ++i) {
+ for (int i = 0; i < planes(); ++i) {
av_free (_data[i]);
}
@@ -674,7 +703,7 @@ Image::data () const
return _data;
}
-int *
+int const *
Image::line_size () const
{
return _line_size;
@@ -726,18 +755,19 @@ merge (list<PositionImage> images)
bool
operator== (Image const & a, Image const & b)
{
- if (a.components() != b.components() || a.pixel_format() != b.pixel_format() || a.aligned() != b.aligned()) {
+ if (a.planes() != b.planes() || a.pixel_format() != b.pixel_format() || a.aligned() != b.aligned()) {
return false;
}
- for (int c = 0; c < a.components(); ++c) {
- if (a.lines(c) != b.lines(c) || a.line_size()[c] != b.line_size()[c] || a.stride()[c] != b.stride()[c]) {
+ for (int c = 0; c < a.planes(); ++c) {
+ if (a.sample_size(c).height != b.sample_size(c).height || a.line_size()[c] != b.line_size()[c] || a.stride()[c] != b.stride()[c]) {
return false;
}
uint8_t* p = a.data()[c];
uint8_t* q = b.data()[c];
- for (int y = 0; y < a.lines(c); ++y) {
+ int const lines = a.sample_size(c).height;
+ for (int y = 0; y < lines; ++y) {
if (memcmp (p, q, a.line_size()[c]) != 0) {
return false;
}
@@ -773,7 +803,8 @@ Image::fade (float f)
/* 8-bit */
for (int c = 0; c < 3; ++c) {
uint8_t* p = data()[c];
- for (int y = 0; y < lines(c); ++y) {
+ int const lines = sample_size(c).height;
+ for (int y = 0; y < lines; ++y) {
uint8_t* q = p;
for (int x = 0; x < line_size()[c]; ++x) {
*q = int (float (*q) * f);
@@ -802,7 +833,8 @@ Image::fade (float f)
int const stride_pixels = stride()[c] / 2;
int const line_size_pixels = line_size()[c] / 2;
uint16_t* p = reinterpret_cast<uint16_t*> (data()[c]);
- for (int y = 0; y < lines(c); ++y) {
+ int const lines = sample_size(c).height;
+ for (int y = 0; y < lines; ++y) {
uint16_t* q = p;
for (int x = 0; x < line_size_pixels; ++x) {
*q = int (float (*q) * f);
@@ -832,7 +864,8 @@ Image::fade (float f)
int const stride_pixels = stride()[c] / 2;
int const line_size_pixels = line_size()[c] / 2;
uint16_t* p = reinterpret_cast<uint16_t*> (data()[c]);
- for (int y = 0; y < lines(c); ++y) {
+ int const lines = sample_size(c).height;
+ for (int y = 0; y < lines; ++y) {
uint16_t* q = p;
for (int x = 0; x < line_size_pixels; ++x) {
*q = swap_16 (int (float (swap_16 (*q)) * f));
@@ -845,7 +878,7 @@ Image::fade (float f)
case PIX_FMT_UYVY422:
{
- int const Y = lines(0);
+ int const Y = sample_size(0).height;
int const X = line_size()[0];
uint8_t* p = data()[0];
for (int y = 0; y < Y; ++y) {
diff --git a/src/lib/image.h b/src/lib/image.h
index 492f6212e..fabcb5675 100644
--- a/src/lib/image.h
+++ b/src/lib/image.h
@@ -49,14 +49,15 @@ public:
~Image ();
uint8_t * const * data () const;
- int * line_size () const;
+ int const * line_size () const;
int const * stride () const;
dcp::Size size () const;
bool aligned () const;
+ int planes () const;
int components () const;
int line_factor (int) const;
- int lines (int) const;
+ dcp::Size sample_size (int) const;
boost::shared_ptr<Image> scale (dcp::Size, dcp::YUVToRGB yuv_to_rgb, AVPixelFormat, bool aligned) const;
boost::shared_ptr<Image> crop_scale_window (Crop c, dcp::Size, dcp::Size, dcp::YUVToRGB yuv_to_rgb, AVPixelFormat, bool aligned) const;