X-Git-Url: https://git.carlh.net/gitweb/?a=blobdiff_plain;f=src%2Flib%2Fj2k_wav_encoder.cc;h=c0c785d118d8254be4eb3489bca2806ca8f9a230;hb=44b57d623dec97a3f9955082f0b8a7a8d27b7518;hp=4deedc388994ab953c306c8e1e1d21a75cceece8;hpb=e1e7827d75df6ea11b6bfd2aabd3eb3fb1f2e701;p=dcpomatic.git diff --git a/src/lib/j2k_wav_encoder.cc b/src/lib/j2k_wav_encoder.cc index 4deedc388..c0c785d11 100644 --- a/src/lib/j2k_wav_encoder.cc +++ b/src/lib/j2k_wav_encoder.cc @@ -46,16 +46,18 @@ using namespace boost; J2KWAVEncoder::J2KWAVEncoder (shared_ptr s, shared_ptr o, Log* l) : Encoder (s, o, l) - , _deinterleave_buffer_size (8192) - , _deinterleave_buffer (0) +#ifdef HAVE_SWRESAMPLE + , _swr_context (0) +#endif + , _audio_frames_written (0) , _process_end (false) { /* Create sound output files with .tmp suffixes; we will rename them if and when we complete. */ - for (int i = 0; i < _fs->audio_channels; ++i) { + for (int i = 0; i < _fs->audio_channels(); ++i) { SF_INFO sf_info; - sf_info.samplerate = dcp_audio_sample_rate (_fs->audio_sample_rate); + sf_info.samplerate = dcp_audio_sample_rate (_fs->audio_sample_rate()); /* We write mono files */ sf_info.channels = 1; sf_info.format = SF_FORMAT_WAV | SF_FORMAT_PCM_24; @@ -65,15 +67,11 @@ J2KWAVEncoder::J2KWAVEncoder (shared_ptr s, shared_ptr yuv, int frame) +J2KWAVEncoder::process_video (shared_ptr yuv, int frame, shared_ptr sub) { boost::mutex::scoped_lock lock (_worker_mutex); /* Wait until the queue has gone down a bit */ while (_queue.size() >= _worker_threads.size() * 2 && !_process_end) { - _log->microsecond_log ("Decoder sleeps", Log::TIMING); + TIMING ("decoder sleeps with queue of %1", _queue.size()); _worker_condition.wait (lock); - _log->microsecond_log ("Decoder wakes", Log::TIMING); + TIMING ("decoder wakes with queue of %1", _queue.size()); } if (_process_end) { @@ -119,11 +117,12 @@ J2KWAVEncoder::process_video (shared_ptr yuv, int frame) /* Only do the processing if we don't already have a file for this frame */ if (!boost::filesystem::exists (_opt->frame_out_path (frame, false))) { - pair const s = Filter::ffmpeg_strings (_fs->filters); - _log->microsecond_log ("Adding to queue of " + boost::lexical_cast (_queue.size ()), Log::TIMING); + pair const s = Filter::ffmpeg_strings (_fs->filters()); + TIMING ("adding to queue of %1", _queue.size ()); _queue.push_back (boost::shared_ptr ( new DCPVideoFrame ( - yuv, _opt->out_size, _opt->padding, _fs->scaler, frame, _fs->frames_per_second, s.second, + yuv, sub, _opt->out_size, _opt->padding, _fs->subtitle_offset(), _fs->subtitle_scale(), + _fs->scaler(), frame, _fs->frames_per_second(), s.second, Config::instance()->colour_lut_index (), Config::instance()->j2k_bandwidth (), _log ) @@ -145,19 +144,20 @@ J2KWAVEncoder::encoder_thread (ServerDescription* server) int remote_backoff = 0; while (1) { + + TIMING ("encoder thread %1 sleeps", boost::this_thread::get_id()); boost::mutex::scoped_lock lock (_worker_mutex); while (_queue.empty () && !_process_end) { - _log->microsecond_log ("Encoder thread sleeps", Log::TIMING); _worker_condition.wait (lock); - _log->microsecond_log ("Encoder thread wakes", Log::TIMING); } if (_process_end) { return; } + TIMING ("encoder thread %1 wakes with queue of %2", boost::this_thread::get_id(), _queue.size()); boost::shared_ptr vf = _queue.front (); - _log->microsecond_log ("Encoder thread wakes with queue of " + boost::lexical_cast (_queue.size ()), Log::TIMING); + _log->log (String::compose ("Encoder thread %1 pops frame %2 from queue", boost::this_thread::get_id(), vf->frame())); _queue.pop_front (); lock.unlock (); @@ -169,9 +169,7 @@ J2KWAVEncoder::encoder_thread (ServerDescription* server) encoded = vf->encode_remotely (server); if (remote_backoff > 0) { - stringstream s; - s << server->host_name() << " was lost, but now she is found; removing backoff"; - _log->log (s.str ()); + _log->log (String::compose ("%1 was lost, but now she is found; removing backoff", server->host_name ())); } /* This job succeeded, so remove any backoff */ @@ -182,20 +180,20 @@ J2KWAVEncoder::encoder_thread (ServerDescription* server) /* back off more */ remote_backoff += 10; } - stringstream s; - s << "Remote encode of " << vf->frame() << " on " << server->host_name() << " failed (" << e.what() << "); thread sleeping for " << remote_backoff << "s."; - _log->log (s.str ()); + _log->log ( + String::compose ( + "Remote encode of %1 on %2 failed (%3); thread sleeping for %4s", + vf->frame(), server->host_name(), e.what(), remote_backoff) + ); } } else { try { - _log->microsecond_log ("Encoder thread begins local encode of " + lexical_cast (vf->frame ()), Log::TIMING); + TIMING ("encoder thread %1 begins local encode of %2", boost::this_thread::get_id(), vf->frame()); encoded = vf->encode_locally (); - _log->microsecond_log ("Encoder thread finishes local encode of " + lexical_cast (vf->frame ()), Log::TIMING); + TIMING ("encoder thread %1 finishes local encode of %2", boost::this_thread::get_id(), vf->frame()); } catch (std::exception& e) { - stringstream s; - s << "Local encode failed " << e.what() << "."; - _log->log (s.str ()); + _log->log (String::compose ("Local encode failed (%1)", e.what ())); } } @@ -204,6 +202,7 @@ J2KWAVEncoder::encoder_thread (ServerDescription* server) frame_done (vf->frame ()); } else { lock.lock (); + _log->log (String::compose ("Encoder thread %1 pushes frame %2 back onto queue after failure", boost::this_thread::get_id(), vf->frame())); _queue.push_front (vf); lock.unlock (); } @@ -218,8 +217,37 @@ J2KWAVEncoder::encoder_thread (ServerDescription* server) } void -J2KWAVEncoder::process_begin () +J2KWAVEncoder::process_begin (int64_t audio_channel_layout) { + if (_fs->audio_sample_rate() != _fs->target_audio_sample_rate()) { +#ifdef HAVE_SWRESAMPLE + + stringstream s; + s << "Will resample audio from " << _fs->audio_sample_rate() << " to " << _fs->target_audio_sample_rate(); + _log->log (s.str ()); + + /* We will be using planar float data when we call the resampler */ + _swr_context = swr_alloc_set_opts ( + 0, + audio_channel_layout, + AV_SAMPLE_FMT_FLTP, + _fs->target_audio_sample_rate(), + audio_channel_layout, + AV_SAMPLE_FMT_FLTP, + _fs->audio_sample_rate(), + 0, 0 + ); + + swr_init (_swr_context); +#else + throw EncodeError ("Cannot resample audio as libswresample is not present"); +#endif + } else { +#ifdef HAVE_SWRESAMPLE + _swr_context = 0; +#endif + } + for (int i = 0; i < Config::instance()->num_local_encoding_threads (); ++i) { _worker_threads.push_back (new boost::thread (boost::bind (&J2KWAVEncoder::encoder_thread, this, (ServerDescription *) 0))); } @@ -263,24 +291,49 @@ J2KWAVEncoder::process_end () */ for (list >::iterator i = _queue.begin(); i != _queue.end(); ++i) { - stringstream s; - s << "Encode left-over frame " << (*i)->frame(); - _log->log (s.str ()); + _log->log (String::compose ("Encode left-over frame %1", (*i)->frame ())); try { shared_ptr e = (*i)->encode_locally (); e->write (_opt, (*i)->frame ()); frame_done ((*i)->frame ()); } catch (std::exception& e) { - stringstream s; - s << "Local encode failed " << e.what() << "."; - _log->log (s.str ()); + _log->log (String::compose ("Local encode failed (%1)", e.what ())); + } + } + +#if HAVE_SWRESAMPLE + if (_swr_context) { + + shared_ptr out (new AudioBuffers (_fs->audio_channels(), 256)); + + while (1) { + int const frames = swr_convert (_swr_context, (uint8_t **) out->data(), 256, 0, 0); + + if (frames < 0) { + throw EncodeError ("could not run sample-rate converter"); + } + + if (frames == 0) { + break; + } + + write_audio (out); } + + swr_free (&_swr_context); } +#endif + + int const dcp_sr = dcp_audio_sample_rate (_fs->audio_sample_rate ()); + int64_t const extra_audio_frames = dcp_sr - (_audio_frames_written % dcp_sr); + shared_ptr silence (new AudioBuffers (_fs->audio_channels(), extra_audio_frames)); + silence->make_silent (); + write_audio (silence); close_sound_files (); /* Rename .wav.tmp files to .wav */ - for (int i = 0; i < _fs->audio_channels; ++i) { + for (int i = 0; i < _fs->audio_channels(); ++i) { if (boost::filesystem::exists (_opt->multichannel_audio_out_path (i, false))) { boost::filesystem::remove (_opt->multichannel_audio_out_path (i, false)); } @@ -289,43 +342,45 @@ J2KWAVEncoder::process_end () } void -J2KWAVEncoder::process_audio (uint8_t* data, int data_size) +J2KWAVEncoder::process_audio (shared_ptr audio) { - /* Size of a sample in bytes */ - int const sample_size = 2; - - /* XXX: we are assuming that sample_size is right, the _deinterleave_buffer_size is a multiple - of the sample size and that data_size is a multiple of _fs->audio_channels * sample_size. - */ - - /* XXX: this code is very tricksy and it must be possible to make it simpler ... */ + shared_ptr resampled; - /* Number of bytes left to read this time */ - int remaining = data_size; - /* Our position in the output buffers, in bytes */ - int position = 0; - while (remaining > 0) { - /* How many bytes of the deinterleaved data to do this time */ - int this_time = min (remaining / _fs->audio_channels, _deinterleave_buffer_size); - for (int i = 0; i < _fs->audio_channels; ++i) { - for (int j = 0; j < this_time; j += sample_size) { - for (int k = 0; k < sample_size; ++k) { - int const to = j + k; - int const from = position + (i * sample_size) + (j * _fs->audio_channels) + k; - _deinterleave_buffer[to] = data[from]; - } - } - - switch (_fs->audio_sample_format) { - case AV_SAMPLE_FMT_S16: - sf_write_short (_sound_files[i], (const short *) _deinterleave_buffer, this_time / sample_size); - break; - default: - throw DecodeError ("unknown audio sample format"); - } +#if HAVE_SWRESAMPLE + /* Maybe sample-rate convert */ + if (_swr_context) { + + /* Compute the resampled frames count and add 32 for luck */ + int const max_resampled_frames = ceil (audio->frames() * _fs->target_audio_sample_rate() / _fs->audio_sample_rate()) + 32; + + resampled.reset (new AudioBuffers (_fs->audio_channels(), max_resampled_frames)); + + /* Resample audio */ + int const resampled_frames = swr_convert ( + _swr_context, (uint8_t **) resampled->data(), max_resampled_frames, (uint8_t const **) audio->data(), audio->frames() + ); + + if (resampled_frames < 0) { + throw EncodeError ("could not run sample-rate converter"); } + + resampled->set_frames (resampled_frames); - position += this_time; - remaining -= this_time * _fs->audio_channels; + /* And point our variables at the resampled audio */ + audio = resampled; } +#endif + + write_audio (audio); } + +void +J2KWAVEncoder::write_audio (shared_ptr audio) +{ + for (int i = 0; i < _fs->audio_channels(); ++i) { + sf_write_float (_sound_files[i], audio->data(i), audio->frames()); + } + + _audio_frames_written += audio->frames (); +} +