X-Git-Url: https://git.carlh.net/gitweb/?a=blobdiff_plain;f=src%2Flib%2Fffmpeg_file_encoder.cc;h=44ef3f0726d6be3d5b71e0046b1389e8f4e7641d;hb=refs%2Fheads%2Fhelp;hp=af6066bfd4cfb524aefcd8ab240ea6d5b514a7a0;hpb=4d11fe7cea71b0564df9a21a3cc706509d12b0d1;p=dcpomatic.git diff --git a/src/lib/ffmpeg_file_encoder.cc b/src/lib/ffmpeg_file_encoder.cc index af6066bfd..44ef3f072 100644 --- a/src/lib/ffmpeg_file_encoder.cc +++ b/src/lib/ffmpeg_file_encoder.cc @@ -63,6 +63,7 @@ FFmpegFileEncoder::FFmpegFileEncoder ( , _video_frame_size (video_frame_size) , _video_frame_rate (video_frame_rate) , _audio_frame_rate (audio_frame_rate) + , _audio_frames (0) { _pixel_format = pixel_format (format); @@ -74,19 +75,27 @@ FFmpegFileEncoder::FFmpegFileEncoder ( av_dict_set (&_video_options, "profile", "3", 0); av_dict_set (&_video_options, "threads", "auto", 0); break; - case EXPORT_FORMAT_H264: + case EXPORT_FORMAT_H264_AAC: _sample_format = AV_SAMPLE_FMT_FLTP; _video_codec_name = "libx264"; _audio_codec_name = "aac"; av_dict_set_int (&_video_options, "crf", x264_crf, 0); break; + case EXPORT_FORMAT_H264_PCM: + _sample_format = AV_SAMPLE_FMT_S32; + _video_codec_name = "libx264"; + _audio_codec_name = "pcm_s24le"; + av_dict_set_int (&_video_options, "crf", x264_crf, 0); + break; + default: + DCPOMATIC_ASSERT (false); } setup_video (); setup_audio (); #ifdef DCPOMATIC_VARIANT_SWAROOP - int r = avformat_alloc_output_context2 (&_format_context, av_guess_format("mp4", 0, 0), 0, 0); + int r = avformat_alloc_output_context2 (&_format_context, av_guess_format("mov", 0, 0), 0, 0); #else int r = avformat_alloc_output_context2 (&_format_context, 0, 0, _output.string().c_str()); #endif @@ -121,13 +130,15 @@ FFmpegFileEncoder::FFmpegFileEncoder ( throw runtime_error (String::compose ("could not open FFmpeg audio codec (%1)", buffer)); } - if (avio_open_boost (&_format_context->pb, _output, AVIO_FLAG_WRITE) < 0) { - throw runtime_error ("could not open FFmpeg output file"); + r = avio_open_boost (&_format_context->pb, _output, AVIO_FLAG_WRITE); + if (r < 0) { + throw runtime_error (String::compose("could not open FFmpeg output file %1 (%2)", _output.string(), r)); } + AVDictionary* options = 0; + #ifdef DCPOMATIC_VARIANT_SWAROOP if (key) { - AVDictionary* options = 0; av_dict_set (&options, "encryption_key", key->hex().c_str(), 0); /* XXX: is this OK? */ av_dict_set (&options, "encryption_kid", "00000000000000000000000000000000", 0); @@ -141,7 +152,7 @@ FFmpegFileEncoder::FFmpegFileEncoder ( } #endif - if (avformat_write_header (_format_context, 0) < 0) { + if (avformat_write_header (_format_context, &options) < 0) { throw runtime_error ("could not write header to FFmpeg output file"); } @@ -154,7 +165,8 @@ FFmpegFileEncoder::pixel_format (ExportFormat format) switch (format) { case EXPORT_FORMAT_PRORES: return AV_PIX_FMT_YUV422P10; - case EXPORT_FORMAT_H264: + case EXPORT_FORMAT_H264_AAC: + case EXPORT_FORMAT_H264_PCM: return AV_PIX_FMT_YUV420P; default: DCPOMATIC_ASSERT (false); @@ -203,7 +215,7 @@ FFmpegFileEncoder::setup_audio () avcodec_get_context_defaults3 (_audio_codec_context, _audio_codec); /* XXX: configurable */ - _audio_codec_context->bit_rate = 256 * 1024; + _audio_codec_context->bit_rate = _audio_channels * 128 * 1024; _audio_codec_context->sample_fmt = _sample_format; _audio_codec_context->sample_rate = _audio_frame_rate; _audio_codec_context->channel_layout = av_get_default_channel_layout (_audio_channels); @@ -270,7 +282,11 @@ FFmpegFileEncoder::video (shared_ptr video, DCPTime time) AVFrame* frame = av_frame_alloc (); DCPOMATIC_ASSERT (frame); - _pending_images[image->data()[0]] = image; + { + boost::mutex::scoped_lock lm (_pending_images_mutex); + _pending_images[image->data()[0]] = image; + } + for (int i = 0; i < 3; ++i) { AVBufferRef* buffer = av_buffer_create(image->data()[i], image->stride()[i] * image->size().height, &buffer_free, this, 0); frame->buf[i] = av_buffer_ref (buffer); @@ -282,7 +298,8 @@ FFmpegFileEncoder::video (shared_ptr video, DCPTime time) frame->width = image->size().width; frame->height = image->size().height; frame->format = _pixel_format; - frame->pts = time.seconds() / av_q2d (_video_stream->time_base); + DCPOMATIC_ASSERT (_video_stream->time_base.num == 1); + frame->pts = time.get() * _video_stream->time_base.den / DCPTime::HZ; AVPacket packet; av_init_packet (&packet); @@ -354,6 +371,16 @@ FFmpegFileEncoder::audio_frame (int size) } break; } + case AV_SAMPLE_FMT_S32: + { + int32_t* q = reinterpret_cast (samples); + for (int i = 0; i < size; ++i) { + for (int j = 0; j < channels; ++j) { + *q++ = p[j][i] * 2147483647; + } + } + break; + } case AV_SAMPLE_FMT_FLTP: { float* q = reinterpret_cast (samples); @@ -367,6 +394,9 @@ FFmpegFileEncoder::audio_frame (int size) DCPOMATIC_ASSERT (false); } + DCPOMATIC_ASSERT (_audio_stream->time_base.num == 1); + frame->pts = _audio_frames * _audio_stream->time_base.den / _audio_frame_rate; + AVPacket packet; av_init_packet (&packet); packet.data = 0; @@ -387,6 +417,7 @@ FFmpegFileEncoder::audio_frame (int size) av_frame_free (&frame); _pending_audio->trim_start (size); + _audio_frames += size; } void @@ -404,5 +435,8 @@ FFmpegFileEncoder::buffer_free (void* opaque, uint8_t* data) void FFmpegFileEncoder::buffer_free2 (uint8_t* data) { - _pending_images.erase (data); + boost::mutex::scoped_lock lm (_pending_images_mutex); + if (_pending_images.find(data) != _pending_images.end()) { + _pending_images.erase (data); + } }