X-Git-Url: https://git.carlh.net/gitweb/?a=blobdiff_plain;f=src%2Flib%2Fvideo_decoder.cc;h=c628fddd983e2015a81b1bd7d13b52b0d9424a0a;hb=095f151405f0e9ab512928db0b6860fd196531d7;hp=9c80cd5c4b1ecc379b97d416f32e43ed0b6f1171;hpb=4cb18d1e0b8fdedf6bb38e1d4187a2d782957022;p=dcpomatic.git diff --git a/src/lib/video_decoder.cc b/src/lib/video_decoder.cc index 9c80cd5c4..c628fddd9 100644 --- a/src/lib/video_decoder.cc +++ b/src/lib/video_decoder.cc @@ -1,5 +1,5 @@ /* - Copyright (C) 2012-2018 Carl Hetherington + Copyright (C) 2012-2021 Carl Hetherington This file is part of DCP-o-matic. @@ -18,95 +18,52 @@ */ -#include "video_decoder.h" + +#include "compose.hpp" +#include "frame_interval_checker.h" #include "image.h" -#include "raw_image_proxy.h" -#include "film.h" +#include "j2k_image_proxy.h" #include "log.h" -#include "compose.hpp" -#include +#include "raw_image_proxy.h" +#include "video_decoder.h" #include #include "i18n.h" + using std::cout; -using std::list; -using std::max; -using std::back_inserter; -using boost::shared_ptr; -using boost::optional; +using std::dynamic_pointer_cast; +using std::shared_ptr; using namespace dcpomatic; + VideoDecoder::VideoDecoder (Decoder* parent, shared_ptr c) : DecoderPart (parent) , _content (c) + , _frame_interval_checker (new FrameIntervalChecker()) { } -/** Called by decoder classes when they have a video frame ready. - * @param frame Frame index within the content; this does not take into account 3D - * so for 3D_ALTERNATE this value goes: - * 0: frame 0 left - * 1: frame 0 right - * 2: frame 1 left - * 3: frame 1 right - * and so on. - */ + +/** Called by decoder classes when they have a video frame ready */ void -VideoDecoder::emit (shared_ptr film, shared_ptr image, Frame decoder_frame) +VideoDecoder::emit(shared_ptr film, shared_ptr image, ContentTime time) { if (ignore ()) { return; } - double const afr = _content->active_video_frame_rate(film); - - Frame frame; - if (!_position) { - /* This is the first data we have received since initialisation or seek. Set - the position based on the frame that was given. After this first time - we just cound frames, since (as with audio) it seems that ContentTimes - are unreliable from FFmpegDecoder. They are much better than audio times - but still we get the occasional one which is duplicated. In this case - ffmpeg seems to carry on regardless, processing the video frame as normal. - If we drop the frame with the duplicated timestamp we obviously lose sync. - */ - _position = ContentTime::from_frames (decoder_frame, afr); - if (_content->video->frame_type() == VIDEO_FRAME_TYPE_3D_ALTERNATE) { - frame = decoder_frame / 2; - _last_emitted_eyes = EYES_RIGHT; - } else { - frame = decoder_frame; - } - } else { - VideoFrameType const ft = _content->video->frame_type (); - if (ft == VIDEO_FRAME_TYPE_3D_ALTERNATE || ft == VIDEO_FRAME_TYPE_3D) { - DCPOMATIC_ASSERT (_last_emitted_eyes); - if (_last_emitted_eyes.get() == EYES_RIGHT) { - frame = _position->frames_round(afr) + 1; - } else { - frame = _position->frames_round(afr); - } - } else { - frame = _position->frames_round(afr) + 1; - } - } + auto const afr = _content->active_video_frame_rate(film); + auto const vft = _content->video->frame_type(); - switch (_content->video->frame_type ()) { - case VIDEO_FRAME_TYPE_2D: - Data (ContentVideo (image, frame, EYES_BOTH, PART_WHOLE)); - break; - case VIDEO_FRAME_TYPE_3D: - { - /* We should receive the same frame index twice for 3D; hence we know which - frame this one is. - */ - bool const same = (_last_emitted_frame && _last_emitted_frame.get() == frame); - if (!same && _last_emitted_eyes && *_last_emitted_eyes == EYES_LEFT) { - /* We just got a new frame index but the last frame was left-eye; it looks like - this content is not really 3D. - */ + /* Do some heuristics to try and spot the case where the user sets content to 3D + * when it is not. We try to tell this by looking at the differences in time between + * the first few frames. Real 3D content should have two frames for each timestamp. + */ + if (_frame_interval_checker) { + _frame_interval_checker->feed(time, afr); + if (_frame_interval_checker->guess() == FrameIntervalChecker::PROBABLY_NOT_3D && vft == VideoFrameType::THREE_D) { boost::throw_exception ( DecodeError( String::compose( @@ -117,45 +74,67 @@ VideoDecoder::emit (shared_ptr film, shared_ptr im ) ); } - Eyes const eyes = same ? EYES_RIGHT : EYES_LEFT; - Data (ContentVideo (image, frame, eyes, PART_WHOLE)); - _last_emitted_frame = frame; - _last_emitted_eyes = eyes; + + if (_frame_interval_checker->guess() != FrameIntervalChecker::AGAIN) { + _frame_interval_checker.reset (); + } + } + + switch (vft) { + case VideoFrameType::TWO_D: + Data(ContentVideo(image, time, Eyes::BOTH, Part::WHOLE)); + break; + case VideoFrameType::THREE_D: + { + auto eyes = Eyes::LEFT; + auto j2k = dynamic_pointer_cast(image); + if (j2k && j2k->eye()) { + eyes = *j2k->eye() == dcp::Eye::LEFT ? Eyes::LEFT : Eyes::RIGHT; + } + + Data(ContentVideo(image, time, eyes, Part::WHOLE)); break; } - case VIDEO_FRAME_TYPE_3D_ALTERNATE: + case VideoFrameType::THREE_D_ALTERNATE: { - DCPOMATIC_ASSERT (_last_emitted_eyes); - Eyes const eyes = _last_emitted_eyes.get() == EYES_LEFT ? EYES_RIGHT : EYES_LEFT; - Data (ContentVideo (image, frame, eyes, PART_WHOLE)); + Eyes eyes; + if (_last_emitted_eyes) { + eyes = _last_emitted_eyes.get() == Eyes::LEFT ? Eyes::RIGHT : Eyes::LEFT; + } else { + /* We don't know what eye this frame is, so just guess */ + auto frame = time.frames_round(_content->video_frame_rate().get_value_or(24)); + eyes = (frame % 2) ? Eyes::RIGHT : Eyes::LEFT; + } + Data(ContentVideo(image, time, eyes, Part::WHOLE)); _last_emitted_eyes = eyes; break; } - case VIDEO_FRAME_TYPE_3D_LEFT_RIGHT: - Data (ContentVideo (image, frame, EYES_LEFT, PART_LEFT_HALF)); - Data (ContentVideo (image, frame, EYES_RIGHT, PART_RIGHT_HALF)); + case VideoFrameType::THREE_D_LEFT_RIGHT: + Data(ContentVideo(image, time, Eyes::LEFT, Part::LEFT_HALF)); + Data(ContentVideo(image, time, Eyes::RIGHT, Part::RIGHT_HALF)); break; - case VIDEO_FRAME_TYPE_3D_TOP_BOTTOM: - Data (ContentVideo (image, frame, EYES_LEFT, PART_TOP_HALF)); - Data (ContentVideo (image, frame, EYES_RIGHT, PART_BOTTOM_HALF)); + case VideoFrameType::THREE_D_TOP_BOTTOM: + Data(ContentVideo(image, time, Eyes::LEFT, Part::TOP_HALF)); + Data(ContentVideo(image, time, Eyes::RIGHT, Part::BOTTOM_HALF)); break; - case VIDEO_FRAME_TYPE_3D_LEFT: - Data (ContentVideo (image, frame, EYES_LEFT, PART_WHOLE)); + case VideoFrameType::THREE_D_LEFT: + Data(ContentVideo(image, time, Eyes::LEFT, Part::WHOLE)); break; - case VIDEO_FRAME_TYPE_3D_RIGHT: - Data (ContentVideo (image, frame, EYES_RIGHT, PART_WHOLE)); + case VideoFrameType::THREE_D_RIGHT: + Data(ContentVideo(image, time, Eyes::RIGHT, Part::WHOLE)); break; default: DCPOMATIC_ASSERT (false); } - _position = ContentTime::from_frames (frame, afr); + _position = time; } + void VideoDecoder::seek () { - _position = boost::optional(); - _last_emitted_frame.reset (); + _position = boost::none; _last_emitted_eyes.reset (); + _frame_interval_checker.reset (new FrameIntervalChecker()); }