Basic grunt-work, untested and unfinished, but it compiles.
[dcpomatic.git] / src / lib / dcp_decoder.cc
index e7e88d3191e9bf2a3ecfaa0f0f8108b1cf7e4613..25c805d3f2ab74d8a9c77c7d7941d18a8e7aad93 100644 (file)
@@ -28,7 +28,6 @@
 #include "image.h"
 #include "config.h"
 #include <dcp/dcp.h>
-#include <dcp/decrypted_kdm.h>
 #include <dcp/cpl.h>
 #include <dcp/reel.h>
 #include <dcp/mono_picture_asset.h>
@@ -50,39 +49,41 @@ using std::cout;
 using boost::shared_ptr;
 using boost::dynamic_pointer_cast;
 
-DCPDecoder::DCPDecoder (shared_ptr<const DCPContent> c, shared_ptr<Log> log, bool fast)
-       : _dcp_content (c)
+DCPDecoder::DCPDecoder (shared_ptr<const DCPContent> c, shared_ptr<Log> log)
+       : DCP (c)
+       , _decode_referenced (false)
 {
        video.reset (new VideoDecoder (this, c, log));
-       audio.reset (new AudioDecoder (this, c->audio, fast, log));
-
-       subtitle.reset (
-               new SubtitleDecoder (
-                       this,
-                       c->subtitle,
-                       bind (&DCPDecoder::image_subtitles_during, this, _1, _2),
-                       bind (&DCPDecoder::text_subtitles_during, this, _1, _2)
-                       )
-               );
-
-       dcp::DCP dcp (c->directory ());
-       dcp.read (false, 0, true);
-       if (c->kdm ()) {
-               dcp.add (dcp::DecryptedKDM (c->kdm().get (), Config::instance()->decryption_chain()->key().get ()));
+       audio.reset (new AudioDecoder (this, c->audio, log));
+
+       subtitle.reset (new SubtitleDecoder (this, c->subtitle, log));
+
+       shared_ptr<dcp::CPL> cpl;
+       BOOST_FOREACH (shared_ptr<dcp::CPL> i, cpls ()) {
+               if (_dcp_content->cpl() && i->id() == _dcp_content->cpl().get()) {
+                       cpl = i;
+               }
+       }
+
+       if (!cpl) {
+               /* No CPL found; probably an old file that doesn't specify it;
+                  just use the first one.
+               */
+               cpl = cpls().front ();
        }
-       DCPOMATIC_ASSERT (dcp.cpls().size() == 1);
-       _reels = dcp.cpls().front()->reels ();
+
+       _reels = cpl->reels ();
 
        _reel = _reels.begin ();
        _offset = 0;
        get_readers ();
 }
 
-bool
-DCPDecoder::pass (PassReason reason, bool)
+void
+DCPDecoder::pass ()
 {
        if (_reel == _reels.end () || !_dcp_content->can_be_played ()) {
-               return true;
+               return;
        }
 
        double const vfr = _dcp_content->active_video_frame_rate ();
@@ -90,25 +91,32 @@ DCPDecoder::pass (PassReason reason, bool)
        /* Frame within the (played part of the) reel that is coming up next */
        int64_t const frame = _next.frames_round (vfr);
 
-       if ((_mono_reader || _stereo_reader) && reason != PASS_REASON_SUBTITLE) {
+       if ((_mono_reader || _stereo_reader) && (_decode_referenced || !_dcp_content->reference_video())) {
                shared_ptr<dcp::PictureAsset> asset = (*_reel)->main_picture()->asset ();
                int64_t const entry_point = (*_reel)->main_picture()->entry_point ();
                if (_mono_reader) {
-                       video->give (shared_ptr<ImageProxy> (new J2KImageProxy (_mono_reader->get_frame (entry_point + frame), asset->size())), _offset + frame);
+                       video->emit (
+                               shared_ptr<ImageProxy> (
+                                       new J2KImageProxy (_mono_reader->get_frame (entry_point + frame), asset->size(), AV_PIX_FMT_XYZ12LE)
+                                       ),
+                               _offset + frame
+                               );
                } else {
-                       video->give (
-                               shared_ptr<ImageProxy> (new J2KImageProxy (_stereo_reader->get_frame (entry_point + frame), asset->size(), dcp::EYE_LEFT)),
+                       video->emit (
+                               shared_ptr<ImageProxy> (
+                                       new J2KImageProxy (_stereo_reader->get_frame (entry_point + frame), asset->size(), dcp::EYE_LEFT, AV_PIX_FMT_XYZ12LE)),
                                _offset + frame
                                );
 
-                       video->give (
-                               shared_ptr<ImageProxy> (new J2KImageProxy (_stereo_reader->get_frame (entry_point + frame), asset->size(), dcp::EYE_RIGHT)),
+                       video->emit (
+                               shared_ptr<ImageProxy> (
+                                       new J2KImageProxy (_stereo_reader->get_frame (entry_point + frame), asset->size(), dcp::EYE_RIGHT, AV_PIX_FMT_XYZ12LE)),
                                _offset + frame
                                );
                }
        }
 
-       if (_sound_reader && reason != PASS_REASON_SUBTITLE) {
+       if (_sound_reader && (_decode_referenced || !_dcp_content->reference_audio())) {
                int64_t const entry_point = (*_reel)->main_sound()->entry_point ();
                shared_ptr<const dcp::SoundFrame> sf = _sound_reader->get_frame (entry_point + frame);
                uint8_t const * from = sf->data ();
@@ -116,17 +124,18 @@ DCPDecoder::pass (PassReason reason, bool)
                int const channels = _dcp_content->audio->stream()->channels ();
                int const frames = sf->size() / (3 * channels);
                shared_ptr<AudioBuffers> data (new AudioBuffers (channels, frames));
+               float** data_data = data->data();
                for (int i = 0; i < frames; ++i) {
                        for (int j = 0; j < channels; ++j) {
-                               data->data()[j][i] = static_cast<int> ((from[0] << 8) | (from[1] << 16) | (from[2] << 24)) / static_cast<float> (INT_MAX - 256);
+                               data_data[j][i] = static_cast<int> ((from[0] << 8) | (from[1] << 16) | (from[2] << 24)) / static_cast<float> (INT_MAX - 256);
                                from += 3;
                        }
                }
 
-               audio->give (_dcp_content->audio->stream(), data, ContentTime::from_frames (_offset, vfr) + _next);
+               audio->emit (_dcp_content->audio->stream(), data, ContentTime::from_frames (_offset, vfr) + _next);
        }
 
-       if ((*_reel)->main_subtitle ()) {
+       if ((*_reel)->main_subtitle() && (_decode_referenced || !_dcp_content->reference_subtitle())) {
                int64_t const entry_point = (*_reel)->main_subtitle()->entry_point ();
                list<dcp::SubtitleString> subs = (*_reel)->main_subtitle()->asset()->subtitles_during (
                        dcp::Time (entry_point + frame, vfr, vfr),
@@ -136,7 +145,7 @@ DCPDecoder::pass (PassReason reason, bool)
 
                if (!subs.empty ()) {
                        /* XXX: assuming that all `subs' are at the same time; maybe this is ok */
-                       subtitle->give_text (
+                       subtitle->emit_text (
                                ContentTimePeriod (
                                        ContentTime::from_frames (_offset - entry_point, vfr) + ContentTime::from_seconds (subs.front().in().as_seconds ()),
                                        ContentTime::from_frames (_offset - entry_point, vfr) + ContentTime::from_seconds (subs.front().out().as_seconds ())
@@ -154,8 +163,6 @@ DCPDecoder::pass (PassReason reason, bool)
                        _next = ContentTime ();
                }
        }
-
-       return false;
 }
 
 void
@@ -169,7 +176,7 @@ DCPDecoder::next_reel ()
 void
 DCPDecoder::get_readers ()
 {
-       if (_reel == _reels.end()) {
+       if (_reel == _reels.end() || !_dcp_content->can_be_played ()) {
                _mono_reader.reset ();
                _stereo_reader.reset ();
                _sound_reader.reset ();
@@ -201,14 +208,12 @@ DCPDecoder::get_readers ()
 }
 
 void
-DCPDecoder::seek (ContentTime t, bool accurate)
+DCPDecoder::seek (ContentTime t, bool)
 {
-       video->seek (t, accurate);
-       audio->seek (t, accurate);
-       subtitle->seek (t, accurate);
-
-       _offset = 0;
        _reel = _reels.begin ();
+       _offset = 0;
+       get_readers ();
+
        while (_reel != _reels.end() && t >= ContentTime::from_frames ((*_reel)->main_picture()->duration(), _dcp_content->active_video_frame_rate ())) {
                t -= ContentTime::from_frames ((*_reel)->main_picture()->duration(), _dcp_content->active_video_frame_rate ());
                next_reel ();
@@ -217,43 +222,8 @@ DCPDecoder::seek (ContentTime t, bool accurate)
        _next = t;
 }
 
-
-list<ContentTimePeriod>
-DCPDecoder::image_subtitles_during (ContentTimePeriod, bool) const
-{
-       return list<ContentTimePeriod> ();
-}
-
-list<ContentTimePeriod>
-DCPDecoder::text_subtitles_during (ContentTimePeriod period, bool starting) const
+void
+DCPDecoder::set_decode_referenced ()
 {
-       /* XXX: inefficient */
-
-       list<ContentTimePeriod> ctp;
-       double const vfr = _dcp_content->active_video_frame_rate ();
-
-       BOOST_FOREACH (shared_ptr<dcp::Reel> r, _reels) {
-               if (!r->main_subtitle ()) {
-                       continue;
-               }
-
-               int64_t const entry_point = r->main_subtitle()->entry_point ();
-
-               list<dcp::SubtitleString> subs = r->main_subtitle()->asset()->subtitles_during (
-                       dcp::Time (period.from.seconds(), 1000) - dcp::Time (entry_point, vfr, vfr),
-                       dcp::Time (period.to.seconds(), 1000) - dcp::Time (entry_point, vfr, vfr),
-                       starting
-                       );
-
-               BOOST_FOREACH (dcp::SubtitleString const & s, subs) {
-                       ctp.push_back (
-                               ContentTimePeriod (
-                                       ContentTime::from_seconds (s.in().as_seconds ()),
-                                       ContentTime::from_seconds (s.out().as_seconds ())
-                                       )
-                               );
-               }
-       }
-
-       return ctp;
+       _decode_referenced = true;
 }