/* Copyright (C) 2014-2022 Carl Hetherington This file is part of DCP-o-matic. DCP-o-matic is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. DCP-o-matic is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with DCP-o-matic. If not, see . */ #include "atmos_decoder.h" #include "audio_content.h" #include "audio_decoder.h" #include "config.h" #include "constants.h" #include "dcp_content.h" #include "dcp_decoder.h" #include "dcpomatic_log.h" #include "digester.h" #include "ffmpeg_image_proxy.h" #include "frame_interval_checker.h" #include "image.h" #include "j2k_image_proxy.h" #include "raw_image_proxy.h" #include "text_decoder.h" #include "util.h" #include "video_decoder.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "i18n.h" using std::cout; using std::dynamic_pointer_cast; using std::list; using std::make_shared; using std::map; using std::shared_ptr; using std::string; using std::vector; using boost::optional; using namespace dcpomatic; DCPDecoder::DCPDecoder(shared_ptr film, shared_ptr content, bool fast, bool tolerant, shared_ptr old) : Decoder(film) , _dcp_content(content) { if (content->can_be_played()) { if (content->video) { video = make_shared(this, content); } if (content->has_mapped_audio()) { audio = make_shared(this, content->audio, fast); } for (auto i: content->text) { text.push_back(make_shared(this, i)); /* We should really call maybe_set_position() on this TextDecoder to set the time * of the first subtitle, but it probably doesn't matter since we'll always * have regularly occurring video (and maybe audio) content. */ } if (content->atmos) { atmos = make_shared(this, content); } } /* We try to avoid re-scanning the DCP's files every time we make a new DCPDecoder; we do this by re-using the _reels list. Before we do this we need to check that nothing too serious has changed in the DCPContent. We do this by storing a digest of the important bits of the DCPContent and then checking that's the same before we re-use _reels. */ _lazy_digest = calculate_lazy_digest(content); if (old && old->lazy_digest() == _lazy_digest) { _reels = old->_reels; } else { auto cpl_list = dcp::find_and_resolve_cpls(content->directories(), tolerant); if (cpl_list.empty()) { throw DCPError(_("No CPLs found in DCP.")); } shared_ptr cpl; for (auto i: cpl_list) { if (_dcp_content->cpl() && i->id() == _dcp_content->cpl().get()) { cpl = i; } } if (!cpl) { /* No CPL found; probably an old file that doesn't specify it; just use the first one. */ cpl = cpl_list.front(); } if (content->kdm()) { cpl->add(decrypt_kdm_with_helpful_error(content->kdm().get())); } _reels = cpl->reels(); } set_decode_referenced(false); _reel = _reels.begin(); get_readers(); _font_id_allocator.add_fonts_from_reels(_reels); _font_id_allocator.allocate(); } bool DCPDecoder::pass() { if (!_dcp_content->can_be_played()) { return true; } if (_reel == _reels.end()) { if (audio) { audio->flush(); } return true; } auto const vfr = _dcp_content->active_video_frame_rate(film()); /* Frame within the (played part of the) reel that is coming up next */ auto const frame = _next.frames_round(vfr); auto picture_asset = (*_reel)->main_picture()->asset(); DCPOMATIC_ASSERT(picture_asset); /* We must emit texts first as when we emit the video for this frame it will expect already to have the texts. */ pass_texts(_next, picture_asset->size()); if ((_j2k_mono_reader || _j2k_stereo_reader || _mpeg2_mono_reader) && (_decode_referenced || !_dcp_content->reference_video())) { auto const entry_point = (*_reel)->main_picture()->entry_point().get_value_or(0); if (_j2k_mono_reader) { video->emit( film(), std::make_shared( _j2k_mono_reader->get_frame(entry_point + frame), picture_asset->size(), AV_PIX_FMT_XYZ12LE, _forced_reduction ), ContentTime::from_frames(_offset + frame, vfr) ); } else if (_j2k_stereo_reader) { video->emit( film(), std::make_shared( _j2k_stereo_reader->get_frame(entry_point + frame), picture_asset->size(), dcp::Eye::LEFT, AV_PIX_FMT_XYZ12LE, _forced_reduction ), ContentTime::from_frames(_offset + frame, vfr) ); video->emit( film(), std::make_shared( _j2k_stereo_reader->get_frame(entry_point + frame), picture_asset->size(), dcp::Eye::RIGHT, AV_PIX_FMT_XYZ12LE, _forced_reduction ), ContentTime::from_frames(_offset + frame, vfr) ); } else if (_mpeg2_mono_reader) { /* XXX: got to flush this at some point */ try { for (auto const& image: _mpeg2_decompressor->decompress_frame(_mpeg2_mono_reader->get_frame(entry_point + frame))) { video->emit( film(), /* XXX: should this be PADDED? */ std::make_shared(std::make_shared(image.frame(), Image::Alignment::COMPACT)), /* XXX: this will be wrong */ ContentTime::from_frames(_offset + frame, vfr) ); } } catch (dcp::MPEG2DecompressionError& e) { LOG_ERROR("Failed to decompress MPEG video frame {} ({})", entry_point + frame, e.what()); } catch (dcp::ReadError& e) { LOG_ERROR("Failed to read MPEG2 video frame {} ({})", entry_point + frame, e.what()); } } } if (_sound_reader && (_decode_referenced || !_dcp_content->reference_audio())) { auto const entry_point = (*_reel)->main_sound()->entry_point().get_value_or(0); auto sf = _sound_reader->get_frame(entry_point + frame); auto from = sf->data(); int const channels = _dcp_content->audio->stream()->channels(); int const frames = sf->size() / (sf->bits() * channels / 8); auto data = make_shared(channels, frames); auto data_data = data->data(); switch (sf->bits()) { case 24: { for (int i = 0; i < frames; ++i) { for (int j = 0; j < channels; ++j) { data_data[j][i] = static_cast((from[0] << 8) | (from[1] << 16) | (from[2] << 24)) / static_cast(INT_MAX - 256); from += 3; } } break; } case 16: { for (int i = 0; i < frames; ++i) { for (int j = 0; j < channels; ++j) { data_data[j][i] = static_cast(from[0] << 16 | (from[1] << 24)) / static_cast(INT_MAX - 256); from += 2; } } break; } } audio->emit(film(), _dcp_content->audio->stream(), data, ContentTime::from_frames(_offset, vfr) + _next); } if (_atmos_reader) { DCPOMATIC_ASSERT(_atmos_metadata); auto const entry_point = (*_reel)->atmos()->entry_point().get_value_or(0); atmos->emit(film(), _atmos_reader->get_frame(entry_point + frame), _offset + frame, *_atmos_metadata); } _next += ContentTime::from_frames(1, vfr); if ((*_reel)->main_picture()) { if (_next.frames_round(vfr) >= (*_reel)->main_picture()->duration()) { next_reel(); _next = ContentTime(); } } return false; } void DCPDecoder::pass_texts(ContentTime next, dcp::Size size) { auto decoder = text.begin(); if (decoder == text.end()) { /* It's possible that there is now a main subtitle but no TextDecoders, for example if the CPL has just changed but the TextContent's texts have not been recreated yet. */ return; } if ((*_reel)->main_subtitle()) { pass_texts(next, (*_reel)->main_subtitle(), TextType::OPEN_SUBTITLE, *decoder, size); ++decoder; } if ((*_reel)->main_caption()) { pass_texts(next, (*_reel)->main_caption(), TextType::OPEN_CAPTION, *decoder, size); ++decoder; } for (auto i: (*_reel)->closed_subtitles()) { pass_texts(next, i, TextType::CLOSED_SUBTITLE, *decoder, size); ++decoder; } for (auto i: (*_reel)->closed_captions()) { pass_texts(next, i, TextType::CLOSED_CAPTION, *decoder, size); ++decoder; } } void DCPDecoder::pass_texts( ContentTime next, shared_ptr reel_asset, TextType type, shared_ptr decoder, dcp::Size size ) { auto const vfr = _dcp_content->active_video_frame_rate(film()); /* Frame within the (played part of the) reel that is coming up next */ auto const frame = next.frames_round(vfr); auto const asset = reel_asset->asset(); auto const entry_point = reel_asset->entry_point().get_value_or(0); if (_decode_referenced || !_dcp_content->reference_text(type)) { auto subs = asset->texts_during( dcp::Time(entry_point + frame, vfr, vfr), dcp::Time(entry_point + frame + 1, vfr, vfr), true ); vector strings; for (auto i: subs) { auto is = dynamic_pointer_cast(i); if (is) { if (!strings.empty() && (strings.back().in() != is->in() || strings.back().out() != is->out())) { auto b = strings.back(); decoder->emit_plain( ContentTimePeriod( ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.in().as_seconds()), ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.out().as_seconds()) ), strings, asset->subtitle_standard() ); strings.clear(); } auto is_copy = *is; if (is_copy.font()) { is_copy.set_font(_font_id_allocator.font_id(_reel - _reels.begin(), asset->id(), is_copy.font().get())); } else { is_copy.set_font(_font_id_allocator.default_font_id()); } strings.push_back(is_copy); } /* XXX: perhaps these image subs should also be collected together like the string ones are; this would need to be done both here and in DCPSubtitleDecoder. */ auto ii = dynamic_pointer_cast(i); if (ii) { emit_subtitle_image( ContentTimePeriod( ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(i->in().as_seconds()), ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(i->out().as_seconds()) ), *ii, size, decoder ); } } if (!strings.empty()) { auto b = strings.back(); decoder->emit_plain( ContentTimePeriod( ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.in().as_seconds()), ContentTime::from_frames(_offset - entry_point, vfr) + ContentTime::from_seconds(b.out().as_seconds()) ), strings, asset->subtitle_standard() ); strings.clear(); } } } void DCPDecoder::next_reel() { _offset += (*_reel)->main_picture()->actual_duration(); ++_reel; get_readers(); } void DCPDecoder::get_readers() { _j2k_mono_reader.reset(); _j2k_stereo_reader.reset(); _mpeg2_mono_reader.reset(); _sound_reader.reset(); _atmos_reader.reset(); _mpeg2_decompressor.reset(); _atmos_metadata = boost::none; if (_reel == _reels.end() || !_dcp_content->can_be_played()) { return; } if (video && !video->ignore() && (*_reel)->main_picture()) { auto asset = (*_reel)->main_picture()->asset(); auto j2k_mono = dynamic_pointer_cast(asset); auto j2k_stereo = dynamic_pointer_cast(asset); auto mpeg2_mono = dynamic_pointer_cast(asset); DCPOMATIC_ASSERT(j2k_mono || j2k_stereo || mpeg2_mono) if (j2k_mono) { _j2k_mono_reader = j2k_mono->start_read(); _j2k_mono_reader->set_check_hmac(false); } else if (j2k_stereo) { _j2k_stereo_reader = j2k_stereo->start_read(); _j2k_stereo_reader->set_check_hmac(false); } else { _mpeg2_mono_reader = mpeg2_mono->start_read(); _mpeg2_mono_reader->set_check_hmac(false); _mpeg2_decompressor = std::make_shared(); } } if (audio && !audio->ignore() && (*_reel)->main_sound()) { _sound_reader = (*_reel)->main_sound()->asset()->start_read(); _sound_reader->set_check_hmac(false); } if ((*_reel)->atmos()) { auto asset = (*_reel)->atmos()->asset(); _atmos_reader = asset->start_read(); _atmos_reader->set_check_hmac(false); _atmos_metadata = AtmosMetadata(asset); } } void DCPDecoder::seek(ContentTime t, bool accurate) { if (!_dcp_content->can_be_played()) { return; } Decoder::seek(t, accurate); _reel = _reels.begin(); _offset = 0; get_readers(); int const pre_roll_seconds = 2; /* Pre-roll for subs */ auto pre = t - ContentTime::from_seconds(pre_roll_seconds); if (pre < ContentTime()) { pre = ContentTime(); } /* Seek to pre-roll position */ while ( _reel != _reels.end() && pre >= ContentTime::from_frames((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())) ) { auto rd = ContentTime::from_frames((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())); pre -= rd; t -= rd; next_reel(); } /* Pass texts in the pre-roll */ if (_reel != _reels.end()) { auto const vfr = _dcp_content->active_video_frame_rate(film()); for (int i = 0; i < pre_roll_seconds * vfr; ++i) { pass_texts(pre, (*_reel)->main_picture()->asset()->size()); pre += ContentTime::from_frames(1, vfr); } } /* Seek to correct position */ while ( _reel != _reels.end() && t >= ContentTime::from_frames((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())) ) { t -= ContentTime::from_frames((*_reel)->main_picture()->actual_duration(), _dcp_content->active_video_frame_rate(film())); next_reel(); } _next = t; } void DCPDecoder::set_decode_referenced(bool r) { _decode_referenced = r; if (video) { video->set_ignore(_dcp_content->reference_video() && !_decode_referenced); } if (audio) { audio->set_ignore(_dcp_content->reference_audio() && !_decode_referenced); } } void DCPDecoder::set_forced_reduction(optional reduction) { _forced_reduction = reduction; } string DCPDecoder::calculate_lazy_digest(shared_ptr c) const { Digester d; for (auto i: c->paths()) { d.add(i.string()); } if (_dcp_content->kdm()) { d.add(_dcp_content->kdm()->id()); } d.add(static_cast(c->cpl())); if (c->cpl()) { d.add(c->cpl().get()); } return d.get(); } ContentTime DCPDecoder::position() const { return ContentTime::from_frames(_offset, _dcp_content->active_video_frame_rate(film())) + _next; }