X-Git-Url: https://git.carlh.net/gitweb/?p=dcpomatic.git;a=blobdiff_plain;f=src%2Flib%2Faudio_content.cc;h=bd857b313d418da8fcc674ee25f8abd3ed8f3221;hp=d2d90a1f5b42ce9c0d493ea402765995f1a9512d;hb=a60668a37f05a125954ac5e03db0530995ac5769;hpb=e60bb3e51bd1508b149e6b8f6608f09b5196ae26 diff --git a/src/lib/audio_content.cc b/src/lib/audio_content.cc index d2d90a1f5..bd857b313 100644 --- a/src/lib/audio_content.cc +++ b/src/lib/audio_content.cc @@ -1,181 +1,205 @@ /* - Copyright (C) 2013-2015 Carl Hetherington + Copyright (C) 2013-2021 Carl Hetherington - This program is free software; you can redistribute it and/or modify + This file is part of DCP-o-matic. + + DCP-o-matic is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. - This program is distributed in the hope that it will be useful, + DCP-o-matic is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License - along with this program; if not, write to the Free Software - Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. + along with DCP-o-matic. If not, see . */ + #include "audio_content.h" -#include "analyse_audio_job.h" -#include "job_manager.h" -#include "film.h" -#include "exceptions.h" +#include "compose.hpp" #include "config.h" +#include "constants.h" +#include "exceptions.h" +#include "film.h" #include "frame_rate_change.h" -#include "raw_convert.h" -#include "playlist.h" +#include "maths_util.h" +#include "video_content.h" +#include #include -#include +#include +#include #include "i18n.h" -using std::string; + using std::cout; -using std::vector; -using std::stringstream; +using std::dynamic_pointer_cast; using std::fixed; +using std::list; +using std::make_shared; +using std::pair; using std::setprecision; -using boost::shared_ptr; -using boost::dynamic_pointer_cast; +using std::shared_ptr; +using std::string; +using std::vector; using boost::optional; +using dcp::raw_convert; +using namespace dcpomatic; + /** Something stream-related has changed */ -int const AudioContentProperty::AUDIO_STREAMS = 200; -int const AudioContentProperty::AUDIO_GAIN = 201; -int const AudioContentProperty::AUDIO_DELAY = 202; - -AudioContent::AudioContent (shared_ptr film) - : Content (film) - , _audio_gain (0) - , _audio_delay (Config::instance()->default_audio_delay ()) -{ +int const AudioContentProperty::STREAMS = 200; +int const AudioContentProperty::GAIN = 201; +int const AudioContentProperty::DELAY = 202; +int const AudioContentProperty::FADE_IN = 203; +int const AudioContentProperty::FADE_OUT = 204; +int const AudioContentProperty::USE_SAME_FADES_AS_VIDEO = 205; -} -AudioContent::AudioContent (shared_ptr film, DCPTime s) - : Content (film, s) - , _audio_gain (0) - , _audio_delay (Config::instance()->default_audio_delay ()) +AudioContent::AudioContent (Content* parent) + : ContentPart (parent) + , _delay (Config::instance()->default_audio_delay()) { } -AudioContent::AudioContent (shared_ptr film, boost::filesystem::path p) - : Content (film, p) - , _audio_gain (0) - , _audio_delay (Config::instance()->default_audio_delay ()) + +shared_ptr +AudioContent::from_xml (Content* parent, cxml::ConstNodePtr node, int version) { + if (version < 34) { + /* With old metadata FFmpeg content has the audio-related tags even with no + audio streams, so check for that. + */ + if (node->string_child("Type") == "FFmpeg" && node->node_children("AudioStream").empty()) { + return {}; + } + + /* Otherwise we can drop through to the newer logic */ + } + if (!node->optional_number_child ("AudioGain")) { + return {}; + } + + return make_shared(parent, node); } -AudioContent::AudioContent (shared_ptr film, cxml::ConstNodePtr node) - : Content (film, node) + +AudioContent::AudioContent (Content* parent, cxml::ConstNodePtr node) + : ContentPart (parent) { - _audio_gain = node->number_child ("AudioGain"); - _audio_delay = node->number_child ("AudioDelay"); + _gain = node->number_child ("AudioGain"); + _delay = node->number_child ("AudioDelay"); + _fade_in = ContentTime(node->optional_number_child("AudioFadeIn").get_value_or(0)); + _fade_out = ContentTime(node->optional_number_child("AudioFadeOut").get_value_or(0)); + _use_same_fades_as_video = node->optional_bool_child("AudioUseSameFadesAsVideo").get_value_or(false); } -AudioContent::AudioContent (shared_ptr film, vector > c) - : Content (film, c) + +AudioContent::AudioContent (Content* parent, vector> c) + : ContentPart (parent) { - shared_ptr ref = dynamic_pointer_cast (c[0]); + auto ref = c[0]->audio; DCPOMATIC_ASSERT (ref); - for (size_t i = 0; i < c.size(); ++i) { - shared_ptr ac = dynamic_pointer_cast (c[i]); - - if (ac->audio_gain() != ref->audio_gain()) { + for (size_t i = 1; i < c.size(); ++i) { + if (c[i]->audio->gain() != ref->gain()) { throw JoinError (_("Content to be joined must have the same audio gain.")); } - if (ac->audio_delay() != ref->audio_delay()) { + if (c[i]->audio->delay() != ref->delay()) { throw JoinError (_("Content to be joined must have the same audio delay.")); } } - _audio_gain = ref->audio_gain (); - _audio_delay = ref->audio_delay (); + _gain = ref->gain (); + _delay = ref->delay (); + _streams = ref->streams (); } + void AudioContent::as_xml (xmlpp::Node* node) const { boost::mutex::scoped_lock lm (_mutex); - node->add_child("AudioGain")->add_child_text (raw_convert (_audio_gain)); - node->add_child("AudioDelay")->add_child_text (raw_convert (_audio_delay)); + node->add_child("AudioGain")->add_child_text(raw_convert(_gain)); + node->add_child("AudioDelay")->add_child_text(raw_convert(_delay)); + node->add_child("AudioFadeIn")->add_child_text(raw_convert(_fade_in.get())); + node->add_child("AudioFadeOut")->add_child_text(raw_convert(_fade_out.get())); + node->add_child("AudioUseSameFadesAsVideo")->add_child_text(_use_same_fades_as_video ? "1" : "0"); } void -AudioContent::set_audio_gain (double g) +AudioContent::set_gain (double g) { - { - boost::mutex::scoped_lock lm (_mutex); - _audio_gain = g; - } - - signal_changed (AudioContentProperty::AUDIO_GAIN); + maybe_set (_gain, g, AudioContentProperty::GAIN); } + void -AudioContent::set_audio_delay (int d) +AudioContent::set_delay (int d) { - { - boost::mutex::scoped_lock lm (_mutex); - _audio_delay = d; - } - - signal_changed (AudioContentProperty::AUDIO_DELAY); + maybe_set (_delay, d, AudioContentProperty::DELAY); } + string AudioContent::technical_summary () const { - string s = "audio :"; - BOOST_FOREACH (AudioStreamPtr i, audio_streams ()) { - s += String::compose ("stream channels %1 rate %2", i->channels(), i->frame_rate()); + string s = "audio: "; + for (auto i: streams()) { + s += String::compose ("stream channels %1 rate %2 ", i->channels(), i->frame_rate()); } return s; } + void -AudioContent::set_audio_mapping (AudioMapping mapping) +AudioContent::set_mapping (AudioMapping mapping) { + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + int c = 0; - BOOST_FOREACH (AudioStreamPtr i, audio_streams ()) { - AudioMapping stream_mapping (i->channels (), MAX_DCP_AUDIO_CHANNELS); + for (auto i: streams()) { + AudioMapping stream_mapping (i->channels(), MAX_DCP_AUDIO_CHANNELS); for (int j = 0; j < i->channels(); ++j) { for (int k = 0; k < MAX_DCP_AUDIO_CHANNELS; ++k) { - stream_mapping.set (j, k, mapping.get (c, k)); + stream_mapping.set (j, k, mapping.get(c, k)); } ++c; } i->set_mapping (stream_mapping); } - - signal_changed (AudioContentProperty::AUDIO_STREAMS); } + AudioMapping -AudioContent::audio_mapping () const +AudioContent::mapping () const { int channels = 0; - BOOST_FOREACH (AudioStreamPtr i, audio_streams ()) { + for (auto i: streams()) { channels += i->channels (); } AudioMapping merged (channels, MAX_DCP_AUDIO_CHANNELS); + merged.make_zero (); int c = 0; int s = 0; - BOOST_FOREACH (AudioStreamPtr i, audio_streams ()) { - AudioMapping mapping = i->mapping (); + for (auto i: streams()) { + auto mapping = i->mapping (); for (int j = 0; j < mapping.input_channels(); ++j) { for (int k = 0; k < MAX_DCP_AUDIO_CHANNELS; ++k) { - merged.set (c, k, mapping.get (j, k)); + if (k < mapping.output_channels()) { + merged.set (c, k, mapping.get(j, k)); + } } ++c; } @@ -185,19 +209,16 @@ AudioContent::audio_mapping () const return merged; } + /** @return the frame rate that this content should be resampled to in order * that it is in sync with the active video content at its start time. */ int -AudioContent::resampled_audio_frame_rate () const +AudioContent::resampled_frame_rate (shared_ptr film) const { - shared_ptr film = _film.lock (); - DCPOMATIC_ASSERT (film); - - /* Resample to a DCI-approved sample rate */ - double t = has_rate_above_48k() ? 96000 : 48000; + double t = film->audio_frame_rate (); - FrameRateChange frc = film->active_frame_rate_change (position ()); + FrameRateChange frc (film, _parent); /* Compensate if the DCP is being run at a different frame rate to the source; that is, if the video is run such that it will @@ -208,14 +229,13 @@ AudioContent::resampled_audio_frame_rate () const t /= frc.speed_up; } - return rint (t); + return lrint (t); } string -AudioContent::processing_description () const +AudioContent::processing_description (shared_ptr film) const { - vector streams = audio_streams (); - if (streams.empty ()) { + if (streams().empty()) { return ""; } @@ -231,8 +251,8 @@ AudioContent::processing_description () const bool same = true; optional common_frame_rate; - BOOST_FOREACH (AudioStreamPtr i, streams) { - if (i->frame_rate() != resampled_audio_frame_rate()) { + for (auto i: streams()) { + if (i->frame_rate() != resampled_frame_rate(film)) { resampled = true; } else { not_resampled = true; @@ -249,46 +269,250 @@ AudioContent::processing_description () const } if (not_resampled && resampled) { - return String::compose (_("Some audio will be resampled to %1kHz"), resampled_audio_frame_rate ()); + return String::compose (_("Some audio will be resampled to %1Hz"), resampled_frame_rate(film)); } if (!not_resampled && resampled) { if (same) { - return String::compose (_("Audio will be resampled from %1kHz to %2kHz"), common_frame_rate.get(), resampled_audio_frame_rate ()); + return String::compose (_("Audio will be resampled from %1Hz to %2Hz"), common_frame_rate.get(), resampled_frame_rate(film)); } else { - return String::compose (_("Audio will be resampled to %1kHz"), resampled_audio_frame_rate ()); + return String::compose (_("Audio will be resampled to %1Hz"), resampled_frame_rate(film)); } } return ""; } -/** @return true if any stream in this content has a sampling rate of more than 48kHz */ -bool -AudioContent::has_rate_above_48k () const + +/** @return User-visible names of each of our audio channels */ +vector +AudioContent::channel_names () const { - BOOST_FOREACH (AudioStreamPtr i, audio_streams ()) { - if (i->frame_rate() > 48000) { - return true; + vector n; + + int index = 0; + int stream = 1; + for (auto i: streams()) { + for (int j = 0; j < i->channels(); ++j) { + n.push_back (NamedChannel(String::compose ("%1:%2", stream, j + 1), index++)); } + ++stream; } - return false; + return n; } -/** @return User-visible names of each of our audio channels */ -vector -AudioContent::audio_channel_names () const + +void +AudioContent::add_properties (shared_ptr film, list& p) const { - vector n; + shared_ptr stream; + if (streams().size() == 1) { + stream = streams().front(); + } - int t = 1; - BOOST_FOREACH (AudioStreamPtr i, audio_streams ()) { - for (int j = 0; j < i->channels(); ++j) { - n.push_back (String::compose ("%1:%2", t, j + 1)); + if (stream) { + p.push_back (UserProperty(UserProperty::AUDIO, _("Channels"), stream->channels())); + p.push_back (UserProperty(UserProperty::AUDIO, _("Content sample rate"), stream->frame_rate(), _("Hz"))); + if (auto bits = stream->bit_depth()) { + p.push_back(UserProperty(UserProperty::AUDIO, _("Content bit depth"), *bits, _("bits"))); } - ++t; } - return n; + FrameRateChange const frc (_parent->active_video_frame_rate(film), film->video_frame_rate()); + ContentTime const c (_parent->full_length(film), frc); + + p.push_back ( + UserProperty (UserProperty::LENGTH, _("Full length in video frames at content rate"), c.frames_round(frc.source)) + ); + + if (stream) { + p.push_back ( + UserProperty ( + UserProperty::LENGTH, + _("Full length in audio samples at content rate"), + c.frames_round (stream->frame_rate ()) + ) + ); + } + + p.push_back (UserProperty(UserProperty::AUDIO, _("DCP sample rate"), resampled_frame_rate(film), _("Hz"))); + p.push_back (UserProperty(UserProperty::LENGTH, _("Full length in video frames at DCP rate"), c.frames_round (frc.dcp))); + + if (stream) { + p.push_back ( + UserProperty ( + UserProperty::LENGTH, + _("Full length in audio samples at DCP rate"), + c.frames_round(resampled_frame_rate(film)) + ) + ); + } +} + + +void +AudioContent::set_streams (vector streams) +{ + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + + { + boost::mutex::scoped_lock lm (_mutex); + _streams = streams; + } +} + + +AudioStreamPtr +AudioContent::stream () const +{ + boost::mutex::scoped_lock lm (_mutex); + DCPOMATIC_ASSERT (_streams.size() == 1); + return _streams.front (); } + + +void +AudioContent::add_stream (AudioStreamPtr stream) +{ + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + + { + boost::mutex::scoped_lock lm (_mutex); + _streams.push_back (stream); + } +} + + +void +AudioContent::set_stream (AudioStreamPtr stream) +{ + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + + { + boost::mutex::scoped_lock lm (_mutex); + _streams.clear (); + _streams.push_back (stream); + } +} + + +void +AudioContent::take_settings_from (shared_ptr c) +{ + set_gain (c->_gain); + set_delay (c->_delay); + set_fade_in (c->fade_in()); + set_fade_out (c->fade_out()); + + auto const streams_to_take = std::min(_streams.size(), c->_streams.size()); + + for (auto i = 0U; i < streams_to_take; ++i) { + auto mapping = _streams[i]->mapping(); + mapping.take_from(c->_streams[i]->mapping()); + _streams[i]->set_mapping(mapping); + } +} + + +void +AudioContent::modify_position (shared_ptr film, DCPTime& pos) const +{ + pos = pos.round (film->audio_frame_rate()); +} + + +void +AudioContent::modify_trim_start(shared_ptr film, ContentTime& trim) const +{ + /* When this trim is used it the audio will have been resampled, and using the + * DCP rate here reduces the chance of rounding errors causing audio glitches + * due to errors in placement of audio frames (#2373). + */ + trim = trim.round(film ? film->audio_frame_rate() : 48000); +} + + +ContentTime +AudioContent::fade_in () const +{ + boost::mutex::scoped_lock lm (_mutex); + if (_use_same_fades_as_video && _parent->video) { + return dcpomatic::ContentTime::from_frames(_parent->video->fade_in(), _parent->video_frame_rate().get_value_or(24)); + } + + return _fade_in; +} + + +ContentTime +AudioContent::fade_out () const +{ + boost::mutex::scoped_lock lm (_mutex); + if (_use_same_fades_as_video && _parent->video) { + return dcpomatic::ContentTime::from_frames(_parent->video->fade_out(), _parent->video_frame_rate().get_value_or(24)); + } + + return _fade_out; +} + + +void +AudioContent::set_fade_in (ContentTime t) +{ + maybe_set (_fade_in, t, AudioContentProperty::FADE_IN); +} + + +void +AudioContent::set_fade_out (ContentTime t) +{ + maybe_set (_fade_out, t, AudioContentProperty::FADE_OUT); +} + + +void +AudioContent::set_use_same_fades_as_video (bool s) +{ + maybe_set (_use_same_fades_as_video, s, AudioContentProperty::USE_SAME_FADES_AS_VIDEO); +} + + +vector +AudioContent::fade (AudioStreamPtr stream, Frame frame, Frame length, int frame_rate) const +{ + auto const in = fade_in().frames_round(frame_rate); + auto const out = fade_out().frames_round(frame_rate); + + /* Where the start trim ends, at frame_rate */ + auto const trim_start = _parent->trim_start().frames_round(frame_rate); + /* Where the end trim starts within the whole length of the content, at frame_rate */ + auto const trim_end = ContentTime(ContentTime::from_frames(stream->length(), stream->frame_rate()) - _parent->trim_end()).frames_round(frame_rate); + + if ( + (in == 0 || (frame >= (trim_start + in))) && + (out == 0 || ((frame + length) < (trim_end - out))) + ) { + /* This section starts after the fade in and ends before the fade out */ + return {}; + } + + /* Start position relative to the start of the fade in */ + auto in_start = frame - trim_start; + /* Start position relative to the start of the fade out */ + auto out_start = frame - (trim_end - out); + + vector coeffs(length); + for (auto coeff = 0; coeff < length; ++coeff) { + coeffs[coeff] = 1.0; + if (in) { + coeffs[coeff] *= logarithmic_fade_in_curve(static_cast(in_start + coeff) / in); + } + if (out) { + coeffs[coeff] *= logarithmic_fade_out_curve(static_cast(out_start + coeff) / out); + } + } + + return coeffs; +} +