/* Copyright (C) 2012-2016 Carl Hetherington This file is part of DCP-o-matic. DCP-o-matic is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. DCP-o-matic is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with DCP-o-matic. If not, see . */ #include "audio_decoder_stream.h" #include "audio_buffers.h" #include "audio_processor.h" #include "audio_decoder.h" #include "resampler.h" #include "util.h" #include "film.h" #include "log.h" #include "audio_content.h" #include "compose.hpp" #include #include "i18n.h" using std::list; using std::pair; using std::cout; using std::min; using std::max; using boost::optional; using boost::shared_ptr; AudioDecoderStream::AudioDecoderStream ( shared_ptr content, AudioStreamPtr stream, Decoder* decoder, AudioDecoder* audio_decoder, shared_ptr log ) : _content (content) , _stream (stream) , _decoder (decoder) , _audio_decoder (audio_decoder) , _log (log) /* We effectively start having done a seek to zero; this allows silence-padding of the first data that comes out of our decoder. */ , _seek_reference (ContentTime ()) { if (content->resampled_frame_rate() != _stream->frame_rate() && _stream->channels() > 0) { _resampler.reset (new Resampler (_stream->frame_rate(), content->resampled_frame_rate(), _stream->channels ())); } reset_decoded (); } void AudioDecoderStream::reset_decoded () { _decoded = ContentAudio (shared_ptr (new AudioBuffers (_stream->channels(), 0)), 0); } ContentAudio AudioDecoderStream::get (Frame frame, Frame length, bool accurate) { shared_ptr dec; _log->log ( String::compose ( "ADS has request for %1 %2; has %3 %4", frame, length, _decoded.frame, _decoded.audio->frames() ), LogEntry::TYPE_DEBUG_DECODE ); Frame const from = frame; Frame const to = from + length; Frame const have_from = _decoded.frame; Frame const have_to = _decoded.frame + _decoded.audio->frames(); optional missing; if (have_from > from || have_to < to) { /* We need something */ if (have_from <= from && from < have_to) { missing = have_to; } else { missing = from; } } if (missing) { optional pos = _audio_decoder->position (); _log->log ( String::compose ("ADS suggests seek to %1 (now at %2)", *missing, pos ? to_string(pos.get()) : "none"), LogEntry::TYPE_DEBUG_DECODE ); _audio_decoder->maybe_seek (ContentTime::from_frames (*missing, _content->resampled_frame_rate()), accurate); } /* Offset of the data that we want from the start of _decoded.audio (to be set up shortly) */ Frame decoded_offset = 0; /* Now enough pass() calls will either: * (a) give us what we want, or * (b) hit the end of the decoder. * * If we are being accurate, we want the right frames, * otherwise any frames will do. */ if (accurate) { /* Keep stuffing data into _decoded until we have enough data, or the subclass does not want to give us any more */ while ( (_decoded.frame > frame || (_decoded.frame + _decoded.audio->frames()) <= to) && !_decoder->pass (Decoder::PASS_REASON_AUDIO, accurate) ) {} decoded_offset = frame - _decoded.frame; _log->log ( String::compose ("Accurate ADS::get has offset %1 from request %2 and available %3", decoded_offset, frame, have_from), LogEntry::TYPE_DEBUG_DECODE ); } else { while ( _decoded.audio->frames() < length && !_decoder->pass (Decoder::PASS_REASON_AUDIO, accurate) ) {} /* Use decoded_offset of 0, as we don't really care what frames we return */ } /* The amount of data available in _decoded.audio starting from `frame'. This could be -ve if pass() returned true before we got enough data. */ Frame const available = _decoded.audio->frames() - decoded_offset; /* We will return either that, or the requested amount, whichever is smaller */ Frame const to_return = max ((Frame) 0, min (available, length)); /* Copy our data to the output */ shared_ptr out (new AudioBuffers (_decoded.audio->channels(), to_return)); out->copy_from (_decoded.audio.get(), to_return, decoded_offset, 0); Frame const remaining = max ((Frame) 0, available - to_return); /* Clean up decoded; first, move the data after what we just returned to the start of the buffer */ _decoded.audio->move (decoded_offset + to_return, 0, remaining); /* And set up the number of frames we have left */ _decoded.audio->set_frames (remaining); /* Also bump where those frames are in terms of the content */ _decoded.frame += decoded_offset + to_return; return ContentAudio (out, frame); } /** Audio timestamping is made hard by many factors, but perhaps the most entertaining is resampling. * We have to assume that we are feeding continuous data into the resampler, and so we get continuous * data out. Hence we do the timestamping here, post-resampler, just by counting samples. * * The time is passed in here so that after a seek we can set up our _position. The * time is ignored once this has been done. */ void AudioDecoderStream::audio (shared_ptr data, ContentTime time) { _log->log (String::compose ("ADS receives %1 %2", to_string(time), data->frames ()), LogEntry::TYPE_DEBUG_DECODE); if (_resampler) { data = _resampler->run (data); } Frame const frame_rate = _content->resampled_frame_rate (); if (_seek_reference) { /* We've had an accurate seek and now we're seeing some data */ ContentTime const delta = time - _seek_reference.get (); Frame const delta_frames = delta.frames_round (frame_rate); if (delta_frames > 0) { /* This data comes after the seek time. Pad the data with some silence. */ shared_ptr padded (new AudioBuffers (data->channels(), data->frames() + delta_frames)); padded->make_silent (); padded->copy_from (data.get(), data->frames(), 0, delta_frames); data = padded; time -= delta; } _seek_reference = optional (); } if (!_position) { _position = time.frames_round (frame_rate); } DCPOMATIC_ASSERT (_position.get() >= (_decoded.frame + _decoded.audio->frames())); add (data); } void AudioDecoderStream::add (shared_ptr data) { if (!_position) { /* This should only happen when there is a seek followed by a flush, but we need to cope with it. */ return; } /* Resize _decoded to fit the new data */ int new_size = 0; if (_decoded.audio->frames() == 0) { /* There's nothing in there, so just store the new data */ new_size = data->frames (); _decoded.frame = _position.get (); } else { /* Otherwise we need to extend _decoded to include the new stuff */ new_size = _position.get() + data->frames() - _decoded.frame; } _decoded.audio->ensure_size (new_size); _decoded.audio->set_frames (new_size); /* Copy new data in */ _decoded.audio->copy_from (data.get(), data->frames(), 0, _position.get() - _decoded.frame); _position = _position.get() + data->frames (); /* Limit the amount of data we keep in case nobody is asking for it */ int const max_frames = _content->resampled_frame_rate () * 10; if (_decoded.audio->frames() > max_frames) { int const to_remove = _decoded.audio->frames() - max_frames; _decoded.frame += to_remove; _decoded.audio->move (to_remove, 0, max_frames); _decoded.audio->set_frames (max_frames); } } void AudioDecoderStream::flush () { if (!_resampler) { return; } shared_ptr b = _resampler->flush (); if (b) { add (b); } } void AudioDecoderStream::seek (ContentTime t, bool accurate) { _position.reset (); reset_decoded (); if (accurate) { _seek_reference = t; } } void AudioDecoderStream::set_fast () { if (_resampler) { _resampler->set_fast (); } } optional AudioDecoderStream::position () const { if (!_position) { return optional (); } return ContentTime::from_frames (_position.get(), _content->resampled_frame_rate()); }