+
+/** @return User-visible names of each of our audio channels */
+vector<NamedChannel>
+AudioContent::channel_names () const
+{
+ vector<NamedChannel> n;
+
+ int index = 0;
+ int stream = 1;
+ for (auto i: streams()) {
+ for (int j = 0; j < i->channels(); ++j) {
+ n.push_back (NamedChannel(String::compose ("%1:%2", stream, j + 1), index++));
+ }
+ ++stream;
+ }
+
+ return n;
+}
+
+
+void
+AudioContent::add_properties (shared_ptr<const Film> film, list<UserProperty>& p) const
+{
+ shared_ptr<const AudioStream> stream;
+ if (streams().size() == 1) {
+ stream = streams().front();
+ }
+
+ if (stream) {
+ p.push_back (UserProperty(UserProperty::AUDIO, _("Channels"), stream->channels()));
+ p.push_back (UserProperty(UserProperty::AUDIO, _("Content audio sample rate"), stream->frame_rate(), _("Hz")));
+ }
+
+ FrameRateChange const frc (_parent->active_video_frame_rate(film), film->video_frame_rate());
+ ContentTime const c (_parent->full_length(film), frc);
+
+ p.push_back (
+ UserProperty (UserProperty::LENGTH, _("Full length in video frames at content rate"), c.frames_round(frc.source))
+ );
+
+ if (stream) {
+ p.push_back (
+ UserProperty (
+ UserProperty::LENGTH,
+ _("Full length in audio samples at content rate"),
+ c.frames_round (stream->frame_rate ())
+ )
+ );
+ }
+
+ p.push_back (UserProperty(UserProperty::AUDIO, _("DCP sample rate"), resampled_frame_rate(film), _("Hz")));
+ p.push_back (UserProperty(UserProperty::LENGTH, _("Full length in video frames at DCP rate"), c.frames_round (frc.dcp)));
+
+ if (stream) {
+ p.push_back (
+ UserProperty (
+ UserProperty::LENGTH,
+ _("Full length in audio samples at DCP rate"),
+ c.frames_round(resampled_frame_rate(film))
+ )
+ );
+ }
+}
+
+
+void
+AudioContent::set_streams (vector<AudioStreamPtr> streams)
+{
+ ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS);
+
+ {
+ boost::mutex::scoped_lock lm (_mutex);
+ _streams = streams;
+ }
+}
+
+
+AudioStreamPtr
+AudioContent::stream () const
+{
+ boost::mutex::scoped_lock lm (_mutex);
+ DCPOMATIC_ASSERT (_streams.size() == 1);
+ return _streams.front ();
+}
+
+
+void
+AudioContent::add_stream (AudioStreamPtr stream)
+{
+ ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS);
+
+ {
+ boost::mutex::scoped_lock lm (_mutex);
+ _streams.push_back (stream);
+ }
+}
+
+
+void
+AudioContent::set_stream (AudioStreamPtr stream)
+{
+ ContentChangeSignaller cc (_parent, AudioContentProperty::STREAMS);
+
+ {
+ boost::mutex::scoped_lock lm (_mutex);
+ _streams.clear ();
+ _streams.push_back (stream);
+ }
+}
+
+
+void
+AudioContent::take_settings_from (shared_ptr<const AudioContent> c)
+{
+ set_gain (c->_gain);
+ set_delay (c->_delay);
+ set_fade_in (c->fade_in());
+ set_fade_out (c->fade_out());
+
+ size_t i = 0;
+ size_t j = 0;
+
+ while (i < _streams.size() && j < c->_streams.size()) {
+ _streams[i]->set_mapping (c->_streams[j]->mapping());
+ ++i;
+ ++j;
+ }
+}
+
+
+void
+AudioContent::modify_position (shared_ptr<const Film> film, DCPTime& pos) const
+{
+ pos = pos.round (film->audio_frame_rate());
+}
+
+
+void
+AudioContent::modify_trim_start (ContentTime& trim) const
+{
+ DCPOMATIC_ASSERT (!_streams.empty());
+ /* XXX: we're in trouble if streams have different rates */
+ trim = trim.round (_streams.front()->frame_rate());
+}
+
+
+ContentTime
+AudioContent::fade_in () const
+{
+ boost::mutex::scoped_lock lm (_mutex);
+ if (_use_same_fades_as_video && _parent->video) {
+ return dcpomatic::ContentTime::from_frames(_parent->video->fade_in(), _parent->video_frame_rate().get_value_or(24));
+ }
+
+ return _fade_in;
+}
+
+
+ContentTime
+AudioContent::fade_out () const