*/
+
+#include "butler.h"
+#include "cross.h"
#include "ffmpeg_encoder.h"
#include "film.h"
+#include "image.h"
#include "job.h"
+#include "log.h"
#include "player.h"
#include "player_video.h"
-#include "log.h"
-#include "image.h"
-#include "cross.h"
-#include "butler.h"
#include "compose.hpp"
#include <iostream>
#include "i18n.h"
-using std::string;
-using std::runtime_error;
+
using std::cout;
-using std::pair;
using std::list;
-using std::map;
+using std::make_shared;
using std::shared_ptr;
-using boost::bind;
+using std::string;
using std::weak_ptr;
+using boost::bind;
using boost::optional;
using namespace dcpomatic;
#if BOOST_VERSION >= 106100
using namespace boost::placeholders;
#endif
-/** @param key Key to use to encrypt MP4 outputs */
+
FFmpegEncoder::FFmpegEncoder (
shared_ptr<const Film> film,
weak_ptr<Job> job,
float const overall_gain = 2 / (4 + sqrt(2));
float const minus_3dB = 1 / sqrt(2);
if (ch == 2) {
- map.set (dcp::LEFT, 0, 1);
- map.set (dcp::RIGHT, 1, 1);
+ map.set (dcp::Channel::LEFT, 0, 1);
+ map.set (dcp::Channel::RIGHT, 1, 1);
} else if (ch == 4) {
- map.set (dcp::LEFT, 0, overall_gain);
- map.set (dcp::RIGHT, 1, overall_gain);
- map.set (dcp::CENTRE, 0, overall_gain * minus_3dB);
- map.set (dcp::CENTRE, 1, overall_gain * minus_3dB);
- map.set (dcp::LS, 0, overall_gain);
+ map.set (dcp::Channel::LEFT, 0, overall_gain);
+ map.set (dcp::Channel::RIGHT, 1, overall_gain);
+ map.set (dcp::Channel::CENTRE, 0, overall_gain * minus_3dB);
+ map.set (dcp::Channel::CENTRE, 1, overall_gain * minus_3dB);
+ map.set (dcp::Channel::LS, 0, overall_gain);
} else if (ch >= 6) {
- map.set (dcp::LEFT, 0, overall_gain);
- map.set (dcp::RIGHT, 1, overall_gain);
- map.set (dcp::CENTRE, 0, overall_gain * minus_3dB);
- map.set (dcp::CENTRE, 1, overall_gain * minus_3dB);
- map.set (dcp::LS, 0, overall_gain);
- map.set (dcp::RS, 1, overall_gain);
+ map.set (dcp::Channel::LEFT, 0, overall_gain);
+ map.set (dcp::Channel::RIGHT, 1, overall_gain);
+ map.set (dcp::Channel::CENTRE, 0, overall_gain * minus_3dB);
+ map.set (dcp::Channel::CENTRE, 1, overall_gain * minus_3dB);
+ map.set (dcp::Channel::LS, 0, overall_gain);
+ map.set (dcp::Channel::RS, 1, overall_gain);
}
/* XXX: maybe we should do something better for >6 channel DCPs */
} else {
}
}
- _butler.reset (
- new Butler(_film, _player, map, _output_audio_channels, bind(&PlayerVideo::force, _1, FFmpegFileEncoder::pixel_format(format)), VIDEO_RANGE_VIDEO, true, false)
+ _butler = std::make_shared<Butler>(
+ _film,
+ _player,
+ map,
+ _output_audio_channels,
+ bind(&PlayerVideo::force, FFmpegFileEncoder::pixel_format(format)),
+ VideoRange::VIDEO,
+ Image::Alignment::PADDED,
+ false,
+ false,
+ Butler::Audio::ENABLED
);
}
FFmpegEncoder::go ()
{
{
- shared_ptr<Job> job = _job.lock ();
+ auto job = _job.lock ();
DCPOMATIC_ASSERT (job);
job->sub (_("Encoding"));
}
);
}
- list<DCPTimePeriod> reel_periods = _film->reels ();
- list<DCPTimePeriod>::const_iterator reel = reel_periods.begin ();
- list<FileEncoderSet>::iterator encoder = file_encoders.begin ();
+ auto reel_periods = _film->reels ();
+ auto reel = reel_periods.begin ();
+ auto encoder = file_encoders.begin ();
- DCPTime const video_frame = DCPTime::from_frames (1, _film->video_frame_rate ());
+ auto const video_frame = DCPTime::from_frames (1, _film->video_frame_rate ());
int const audio_frames = video_frame.frames_round(_film->audio_frame_rate());
- float* interleaved = new float[_output_audio_channels * audio_frames];
- shared_ptr<AudioBuffers> deinterleaved (new AudioBuffers (_output_audio_channels, audio_frames));
+ std::vector<float> interleaved(_output_audio_channels * audio_frames);
+ auto deinterleaved = make_shared<AudioBuffers>(_output_audio_channels, audio_frames);
int const gets_per_frame = _film->three_d() ? 2 : 1;
for (DCPTime i; i < _film->length(); i += video_frame) {
for (int j = 0; j < gets_per_frame; ++j) {
Butler::Error e;
- pair<shared_ptr<PlayerVideo>, DCPTime> v = _butler->get_video (true, &e);
+ auto v = _butler->get_video (Butler::Behaviour::BLOCKING, &e);
_butler->rethrow ();
- if (!v.first) {
- throw DecodeError(String::compose("Error during decoding: %1", e.summary()));
- }
- shared_ptr<FFmpegFileEncoder> fe = encoder->get (v.first->eyes());
- if (fe) {
- fe->video(v.first, v.second);
+ if (v.first) {
+ auto fe = encoder->get (v.first->eyes());
+ if (fe) {
+ fe->video(v.first, v.second - reel->from);
+ }
+ } else {
+ if (e.code != Butler::Error::Code::FINISHED) {
+ throw DecodeError(String::compose("Error during decoding: %1", e.summary()));
+ }
}
}
_last_time = i;
}
- shared_ptr<Job> job = _job.lock ();
+ auto job = _job.lock ();
if (job) {
job->set_progress (float(i.get()) / _film->length().get());
}
waker.nudge ();
- _butler->get_audio (interleaved, audio_frames);
+ _butler->get_audio (Butler::Behaviour::BLOCKING, interleaved.data(), audio_frames);
/* XXX: inefficient; butler interleaves and we deinterleave again */
- float* p = interleaved;
+ float* p = interleaved.data();
for (int j = 0; j < audio_frames; ++j) {
for (int k = 0; k < _output_audio_channels; ++k) {
deinterleaved->data(k)[j] = *p++;
}
encoder->audio (deinterleaved);
}
- delete[] interleaved;
for (auto i: file_encoders) {
i.flush ();
{
if (three_d) {
/// TRANSLATORS: L here is an abbreviation for "left", to indicate the left-eye part of a 3D export
- _encoders[EYES_LEFT] = shared_ptr<FFmpegFileEncoder>(
- new FFmpegFileEncoder(
- video_frame_size, video_frame_rate, audio_frame_rate, channels, format,
- audio_stream_per_channel, x264_crf, String::compose("%1_%2%3", output.string(), _("L"), extension))
+ _encoders[Eyes::LEFT] = make_shared<FFmpegFileEncoder>(
+ video_frame_size, video_frame_rate, audio_frame_rate, channels, format,
+ audio_stream_per_channel, x264_crf, String::compose("%1_%2%3", output.string(), _("L"), extension)
);
/// TRANSLATORS: R here is an abbreviation for "right", to indicate the right-eye part of a 3D export
- _encoders[EYES_RIGHT] = shared_ptr<FFmpegFileEncoder>(
- new FFmpegFileEncoder(
- video_frame_size, video_frame_rate, audio_frame_rate, channels, format,
- audio_stream_per_channel, x264_crf, String::compose("%1_%2%3", output.string(), _("R"), extension))
+ _encoders[Eyes::RIGHT] = make_shared<FFmpegFileEncoder>(
+ video_frame_size, video_frame_rate, audio_frame_rate, channels, format,
+ audio_stream_per_channel, x264_crf, String::compose("%1_%2%3", output.string(), _("R"), extension)
);
} else {
- _encoders[EYES_BOTH] = shared_ptr<FFmpegFileEncoder>(
- new FFmpegFileEncoder(
- video_frame_size, video_frame_rate, audio_frame_rate, channels, format,
- audio_stream_per_channel, x264_crf, String::compose("%1%2", output.string(), extension))
+ _encoders[Eyes::BOTH] = make_shared<FFmpegFileEncoder>(
+ video_frame_size, video_frame_rate, audio_frame_rate, channels, format,
+ audio_stream_per_channel, x264_crf, String::compose("%1%2", output.string(), extension)
);
}
}
{
if (_encoders.size() == 1) {
/* We are doing a 2D export... */
- if (eyes == EYES_LEFT) {
+ if (eyes == Eyes::LEFT) {
/* ...but we got some 3D data; put the left eye into the output... */
- eyes = EYES_BOTH;
- } else if (eyes == EYES_RIGHT) {
+ eyes = Eyes::BOTH;
+ } else if (eyes == Eyes::RIGHT) {
/* ...and ignore the right eye.*/
- return shared_ptr<FFmpegFileEncoder>();
+ return {};
}
}
- map<Eyes, std::shared_ptr<FFmpegFileEncoder> >::const_iterator i = _encoders.find (eyes);
+ auto i = _encoders.find (eyes);
DCPOMATIC_ASSERT (i != _encoders.end());
return i->second;
}
void
FFmpegEncoder::FileEncoderSet::flush ()
{
- for (map<Eyes, std::shared_ptr<FFmpegFileEncoder> >::iterator i = _encoders.begin(); i != _encoders.end(); ++i) {
- i->second->flush ();
+ for (auto& i: _encoders) {
+ i.second->flush ();
}
}
void
FFmpegEncoder::FileEncoderSet::audio (shared_ptr<AudioBuffers> a)
{
- for (map<Eyes, std::shared_ptr<FFmpegFileEncoder> >::iterator i = _encoders.begin(); i != _encoders.end(); ++i) {
- i->second->audio (a);
+ for (auto& i: _encoders) {
+ i.second->audio (a);
}
}