X-Git-Url: https://git.carlh.net/gitweb/?p=dcpomatic.git;a=blobdiff_plain;f=src%2Flib%2Faudio_content.cc;h=7083de714dcf411946c65598cdaf07d23e52ee32;hp=8458e3314c7be41b36aaccfdb8fdde437fe7b8da;hb=f4fda1d705adfca0b7c8d3748c0c9df8e9da51bc;hpb=51cc8f575b55f2a21d8503965b628c0f59f1ed33 diff --git a/src/lib/audio_content.cc b/src/lib/audio_content.cc index 8458e3314..7083de714 100644 --- a/src/lib/audio_content.cc +++ b/src/lib/audio_content.cc @@ -1,260 +1,521 @@ /* - Copyright (C) 2013-2014 Carl Hetherington + Copyright (C) 2013-2021 Carl Hetherington - This program is free software; you can redistribute it and/or modify + This file is part of DCP-o-matic. + + DCP-o-matic is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. - This program is distributed in the hope that it will be useful, + DCP-o-matic is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License - along with this program; if not, write to the Free Software - Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. + along with DCP-o-matic. If not, see . */ -#include -#include + #include "audio_content.h" -#include "analyse_audio_job.h" -#include "job_manager.h" -#include "film.h" -#include "exceptions.h" +#include "compose.hpp" #include "config.h" +#include "constants.h" +#include "exceptions.h" +#include "film.h" #include "frame_rate_change.h" -#include "audio_processor.h" +#include "maths_util.h" +#include "video_content.h" +#include +#include +#include +#include #include "i18n.h" -using std::string; + using std::cout; -using std::vector; -using std::stringstream; +using std::dynamic_pointer_cast; using std::fixed; +using std::list; +using std::make_shared; +using std::pair; using std::setprecision; -using boost::shared_ptr; -using boost::dynamic_pointer_cast; +using std::shared_ptr; +using std::string; +using std::vector; +using boost::optional; using dcp::raw_convert; +using namespace dcpomatic; -int const AudioContentProperty::AUDIO_CHANNELS = 200; -int const AudioContentProperty::AUDIO_LENGTH = 201; -int const AudioContentProperty::AUDIO_FRAME_RATE = 202; -int const AudioContentProperty::AUDIO_GAIN = 203; -int const AudioContentProperty::AUDIO_DELAY = 204; -int const AudioContentProperty::AUDIO_MAPPING = 205; -int const AudioContentProperty::AUDIO_PROCESSOR = 206; - -AudioContent::AudioContent (shared_ptr f) - : Content (f) - , _audio_gain (0) - , _audio_delay (Config::instance()->default_audio_delay ()) - , _audio_processor (0) -{ -} +/** Something stream-related has changed */ +int const AudioContentProperty::STREAMS = 200; +int const AudioContentProperty::GAIN = 201; +int const AudioContentProperty::DELAY = 202; +int const AudioContentProperty::FADE_IN = 203; +int const AudioContentProperty::FADE_OUT = 204; +int const AudioContentProperty::USE_SAME_FADES_AS_VIDEO = 205; + -AudioContent::AudioContent (shared_ptr f, DCPTime s) - : Content (f, s) - , _audio_gain (0) - , _audio_delay (Config::instance()->default_audio_delay ()) - , _audio_processor (0) +AudioContent::AudioContent (Content* parent) + : ContentPart (parent) + , _delay (Config::instance()->default_audio_delay()) { } -AudioContent::AudioContent (shared_ptr f, boost::filesystem::path p) - : Content (f, p) - , _audio_gain (0) - , _audio_delay (Config::instance()->default_audio_delay ()) - , _audio_processor (0) + +shared_ptr +AudioContent::from_xml (Content* parent, cxml::ConstNodePtr node, int version) { + if (version < 34) { + /* With old metadata FFmpeg content has the audio-related tags even with no + audio streams, so check for that. + */ + if (node->string_child("Type") == "FFmpeg" && node->node_children("AudioStream").empty()) { + return {}; + } + + /* Otherwise we can drop through to the newer logic */ + } + + if (!node->optional_number_child ("AudioGain")) { + return {}; + } + return make_shared(parent, node); } -AudioContent::AudioContent (shared_ptr f, cxml::ConstNodePtr node) - : Content (f, node) - , _audio_processor (0) + +AudioContent::AudioContent (Content* parent, cxml::ConstNodePtr node) + : ContentPart (parent) { - _audio_gain = node->number_child ("AudioGain"); - _audio_delay = node->number_child ("AudioDelay"); - if (node->optional_string_child ("AudioProcessor")) { - _audio_processor = AudioProcessor::from_id (node->string_child ("AudioProcessor")); - } + _gain = node->number_child ("AudioGain"); + _delay = node->number_child ("AudioDelay"); + _fade_in = ContentTime(node->optional_number_child("AudioFadeIn").get_value_or(0)); + _fade_out = ContentTime(node->optional_number_child("AudioFadeOut").get_value_or(0)); + _use_same_fades_as_video = node->optional_bool_child("AudioUseSameFadesAsVideo").get_value_or(false); } -AudioContent::AudioContent (shared_ptr f, vector > c) - : Content (f, c) + +AudioContent::AudioContent (Content* parent, vector> c) + : ContentPart (parent) { - shared_ptr ref = dynamic_pointer_cast (c[0]); - assert (ref); - - for (size_t i = 0; i < c.size(); ++i) { - shared_ptr ac = dynamic_pointer_cast (c[i]); + auto ref = c[0]->audio; + DCPOMATIC_ASSERT (ref); - if (ac->audio_gain() != ref->audio_gain()) { + for (size_t i = 1; i < c.size(); ++i) { + if (c[i]->audio->gain() != ref->gain()) { throw JoinError (_("Content to be joined must have the same audio gain.")); } - if (ac->audio_delay() != ref->audio_delay()) { + if (c[i]->audio->delay() != ref->delay()) { throw JoinError (_("Content to be joined must have the same audio delay.")); } } - _audio_gain = ref->audio_gain (); - _audio_delay = ref->audio_delay (); - _audio_processor = ref->audio_processor (); + _gain = ref->gain (); + _delay = ref->delay (); + _streams = ref->streams (); } + void AudioContent::as_xml (xmlpp::Node* node) const { boost::mutex::scoped_lock lm (_mutex); - node->add_child("AudioGain")->add_child_text (raw_convert (_audio_gain)); - node->add_child("AudioDelay")->add_child_text (raw_convert (_audio_delay)); - if (_audio_processor) { - node->add_child("AudioProcessor")->add_child_text (_audio_processor->id ()); + node->add_child("AudioGain")->add_child_text(raw_convert(_gain)); + node->add_child("AudioDelay")->add_child_text(raw_convert(_delay)); + node->add_child("AudioFadeIn")->add_child_text(raw_convert(_fade_in.get())); + node->add_child("AudioFadeOut")->add_child_text(raw_convert(_fade_out.get())); + node->add_child("AudioUseSameFadesAsVideo")->add_child_text(_use_same_fades_as_video ? "1" : "0"); +} + + +void +AudioContent::set_gain (double g) +{ + maybe_set (_gain, g, AudioContentProperty::GAIN); +} + + +void +AudioContent::set_delay (int d) +{ + maybe_set (_delay, d, AudioContentProperty::DELAY); +} + + +string +AudioContent::technical_summary () const +{ + string s = "audio: "; + for (auto i: streams()) { + s += String::compose ("stream channels %1 rate %2 ", i->channels(), i->frame_rate()); } + + return s; } void -AudioContent::set_audio_gain (double g) +AudioContent::set_mapping (AudioMapping mapping) { - { - boost::mutex::scoped_lock lm (_mutex); - _audio_gain = g; + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + + int c = 0; + for (auto i: streams()) { + AudioMapping stream_mapping (i->channels(), MAX_DCP_AUDIO_CHANNELS); + for (int j = 0; j < i->channels(); ++j) { + for (int k = 0; k < MAX_DCP_AUDIO_CHANNELS; ++k) { + stream_mapping.set (j, k, mapping.get(c, k)); + } + ++c; + } + i->set_mapping (stream_mapping); + } +} + + +AudioMapping +AudioContent::mapping () const +{ + int channels = 0; + for (auto i: streams()) { + channels += i->channels (); + } + + AudioMapping merged (channels, MAX_DCP_AUDIO_CHANNELS); + merged.make_zero (); + + int c = 0; + int s = 0; + for (auto i: streams()) { + auto mapping = i->mapping (); + for (int j = 0; j < mapping.input_channels(); ++j) { + for (int k = 0; k < MAX_DCP_AUDIO_CHANNELS; ++k) { + if (k < mapping.output_channels()) { + merged.set (c, k, mapping.get(j, k)); + } + } + ++c; + } + ++s; } - - signal_changed (AudioContentProperty::AUDIO_GAIN); + + return merged; } + +/** @return the frame rate that this content should be resampled to in order + * that it is in sync with the active video content at its start time. + */ +int +AudioContent::resampled_frame_rate (shared_ptr film) const +{ + double t = film->audio_frame_rate (); + + FrameRateChange frc (film, _parent); + + /* Compensate if the DCP is being run at a different frame rate + to the source; that is, if the video is run such that it will + look different in the DCP compared to the source (slower or faster). + */ + + if (frc.change_speed) { + t /= frc.speed_up; + } + + return lrint (t); +} + +string +AudioContent::processing_description (shared_ptr film) const +{ + if (streams().empty()) { + return ""; + } + + /* Possible answers are: + 1. all audio will be resampled from x to y. + 2. all audio will be resampled to y (from a variety of rates) + 3. some audio will be resampled to y (from a variety of rates) + 4. nothing will be resampled. + */ + + bool not_resampled = false; + bool resampled = false; + bool same = true; + + optional common_frame_rate; + for (auto i: streams()) { + if (i->frame_rate() != resampled_frame_rate(film)) { + resampled = true; + } else { + not_resampled = true; + } + + if (common_frame_rate && common_frame_rate != i->frame_rate ()) { + same = false; + } + common_frame_rate = i->frame_rate (); + } + + if (not_resampled && !resampled) { + return _("Audio will not be resampled"); + } + + if (not_resampled && resampled) { + return String::compose (_("Some audio will be resampled to %1Hz"), resampled_frame_rate(film)); + } + + if (!not_resampled && resampled) { + if (same) { + return String::compose (_("Audio will be resampled from %1Hz to %2Hz"), common_frame_rate.get(), resampled_frame_rate(film)); + } else { + return String::compose (_("Audio will be resampled to %1Hz"), resampled_frame_rate(film)); + } + } + + return ""; +} + + +/** @return User-visible names of each of our audio channels */ +vector +AudioContent::channel_names () const +{ + vector n; + + int index = 0; + int stream = 1; + for (auto i: streams()) { + for (int j = 0; j < i->channels(); ++j) { + n.push_back (NamedChannel(String::compose ("%1:%2", stream, j + 1), index++)); + } + ++stream; + } + + return n; +} + + void -AudioContent::set_audio_delay (int d) +AudioContent::add_properties (shared_ptr film, list& p) const { - { - boost::mutex::scoped_lock lm (_mutex); - _audio_delay = d; + shared_ptr stream; + if (streams().size() == 1) { + stream = streams().front(); + } + + if (stream) { + p.push_back (UserProperty(UserProperty::AUDIO, _("Channels"), stream->channels())); + p.push_back (UserProperty(UserProperty::AUDIO, _("Content sample rate"), stream->frame_rate(), _("Hz"))); + if (auto bits = stream->bit_depth()) { + p.push_back(UserProperty(UserProperty::AUDIO, _("Content bit depth"), *bits, _("bits"))); + } + } + + FrameRateChange const frc (_parent->active_video_frame_rate(film), film->video_frame_rate()); + ContentTime const c (_parent->full_length(film), frc); + + p.push_back ( + UserProperty (UserProperty::LENGTH, _("Full length in video frames at content rate"), c.frames_round(frc.source)) + ); + + if (stream) { + p.push_back ( + UserProperty ( + UserProperty::LENGTH, + _("Full length in audio samples at content rate"), + c.frames_round (stream->frame_rate ()) + ) + ); + } + + p.push_back (UserProperty(UserProperty::AUDIO, _("DCP sample rate"), resampled_frame_rate(film), _("Hz"))); + p.push_back (UserProperty(UserProperty::LENGTH, _("Full length in video frames at DCP rate"), c.frames_round (frc.dcp))); + + if (stream) { + p.push_back ( + UserProperty ( + UserProperty::LENGTH, + _("Full length in audio samples at DCP rate"), + c.frames_round(resampled_frame_rate(film)) + ) + ); } - - signal_changed (AudioContentProperty::AUDIO_DELAY); } + void -AudioContent::set_audio_processor (AudioProcessor const * p) +AudioContent::set_streams (vector streams) { + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + { boost::mutex::scoped_lock lm (_mutex); - _audio_processor = p; + _streams = streams; } +} - /* The channel count might have changed, so reset the mapping */ - AudioMapping m (processed_audio_channels ()); - m.make_default (); - set_audio_mapping (m); - signal_changed (AudioContentProperty::AUDIO_PROCESSOR); +AudioStreamPtr +AudioContent::stream () const +{ + boost::mutex::scoped_lock lm (_mutex); + DCPOMATIC_ASSERT (_streams.size() == 1); + return _streams.front (); } -boost::signals2::connection -AudioContent::analyse_audio (boost::function finished) + +void +AudioContent::add_stream (AudioStreamPtr stream) { - shared_ptr film = _film.lock (); - assert (film); - - shared_ptr job (new AnalyseAudioJob (film, dynamic_pointer_cast (shared_from_this()))); - boost::signals2::connection c = job->Finished.connect (finished); - JobManager::instance()->add (job); - - return c; + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + + { + boost::mutex::scoped_lock lm (_mutex); + _streams.push_back (stream); + } } -boost::filesystem::path -AudioContent::audio_analysis_path () const + +void +AudioContent::set_stream (AudioStreamPtr stream) { - shared_ptr film = _film.lock (); - if (!film) { - return boost::filesystem::path (); + ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS); + + { + boost::mutex::scoped_lock lm (_mutex); + _streams.clear (); + _streams.push_back (stream); } +} + - boost::filesystem::path p = film->audio_analysis_dir (); - p /= digest().get_value_or ("X") + "_" + audio_mapping().digest(); - return p; +void +AudioContent::take_settings_from (shared_ptr c) +{ + set_gain (c->_gain); + set_delay (c->_delay); + set_fade_in (c->fade_in()); + set_fade_out (c->fade_out()); + + size_t i = 0; + size_t j = 0; + + while (i < _streams.size() && j < c->_streams.size()) { + auto mapping = _streams[i]->mapping(); + mapping.take_from(c->_streams[j]->mapping()); + _streams[i]->set_mapping(mapping); + ++i; + ++j; + } } -string -AudioContent::technical_summary () const + +void +AudioContent::modify_position (shared_ptr film, DCPTime& pos) const { - return String::compose ( - "audio: channels %1, length %2, content rate %3, resampled rate %4", - audio_channels(), - audio_length().seconds(), - audio_frame_rate(), - resampled_audio_frame_rate() - ); + pos = pos.round (film->audio_frame_rate()); } + void -AudioContent::set_audio_mapping (AudioMapping) +AudioContent::modify_trim_start(shared_ptr film, ContentTime& trim) const { - signal_changed (AudioContentProperty::AUDIO_MAPPING); + /* When this trim is used it the audio will have been resampled, and using the + * DCP rate here reduces the chance of rounding errors causing audio glitches + * due to errors in placement of audio frames (#2373). + */ + trim = trim.round(film ? film->audio_frame_rate() : 48000); } -/** @return the frame rate that this content should be resampled to in order - * that it is in sync with the active video content at its start time. - */ -int -AudioContent::resampled_audio_frame_rate () const + +ContentTime +AudioContent::fade_in () const { - shared_ptr film = _film.lock (); - assert (film); - - /* Resample to a DCI-approved sample rate */ - double t = dcp_audio_frame_rate (audio_frame_rate ()); + boost::mutex::scoped_lock lm (_mutex); + if (_use_same_fades_as_video && _parent->video) { + return dcpomatic::ContentTime::from_frames(_parent->video->fade_in(), _parent->video_frame_rate().get_value_or(24)); + } - FrameRateChange frc = film->active_frame_rate_change (position ()); + return _fade_in; +} - /* Compensate if the DCP is being run at a different frame rate - to the source; that is, if the video is run such that it will - look different in the DCP compared to the source (slower or faster). - */ - if (frc.change_speed) { - t /= frc.speed_up; +ContentTime +AudioContent::fade_out () const +{ + boost::mutex::scoped_lock lm (_mutex); + if (_use_same_fades_as_video && _parent->video) { + return dcpomatic::ContentTime::from_frames(_parent->video->fade_out(), _parent->video_frame_rate().get_value_or(24)); } - return rint (t); + return _fade_out; } -int -AudioContent::processed_audio_channels () const + +void +AudioContent::set_fade_in (ContentTime t) { - if (!audio_processor ()) { - return audio_channels (); - } + maybe_set (_fade_in, t, AudioContentProperty::FADE_IN); +} + - return audio_processor()->out_channels (audio_channels ()); +void +AudioContent::set_fade_out (ContentTime t) +{ + maybe_set (_fade_out, t, AudioContentProperty::FADE_OUT); } -string -AudioContent::processing_description () const + +void +AudioContent::set_use_same_fades_as_video (bool s) +{ + maybe_set (_use_same_fades_as_video, s, AudioContentProperty::USE_SAME_FADES_AS_VIDEO); +} + + +vector +AudioContent::fade (AudioStreamPtr stream, Frame frame, Frame length, int frame_rate) const { - stringstream d; - - if (audio_frame_rate() != resampled_audio_frame_rate ()) { - stringstream from; - from << fixed << setprecision(3) << (audio_frame_rate() / 1000.0); - stringstream to; - to << fixed << setprecision(3) << (resampled_audio_frame_rate() / 1000.0); - - d << String::compose (_("Audio will be resampled from %1kHz to %2kHz."), from.str(), to.str()); - } else { - d << _("Audio will not be resampled."); + auto const in = fade_in().frames_round(frame_rate); + auto const out = fade_out().frames_round(frame_rate); + + /* Where the start trim ends, at frame_rate */ + auto const trim_start = _parent->trim_start().frames_round(frame_rate); + /* Where the end trim starts within the whole length of the content, at frame_rate */ + auto const trim_end = ContentTime(ContentTime::from_frames(stream->length(), stream->frame_rate()) - _parent->trim_end()).frames_round(frame_rate); + + if ( + (in == 0 || (frame >= (trim_start + in))) && + (out == 0 || ((frame + length) < (trim_end - out))) + ) { + /* This section starts after the fade in and ends before the fade out */ + return {}; + } + + /* Start position relative to the start of the fade in */ + auto in_start = frame - trim_start; + /* Start position relative to the start of the fade out */ + auto out_start = frame - (trim_end - out); + + vector coeffs(length); + for (auto coeff = 0; coeff < length; ++coeff) { + coeffs[coeff] = 1.0; + if (in) { + coeffs[coeff] *= logarithmic_fade_in_curve(static_cast(in_start + coeff) / in); + } + if (out) { + coeffs[coeff] *= logarithmic_fade_out_curve(static_cast(out_start + coeff) / out); + } } - return d.str (); + return coeffs; }