#include "video_decoder.h"
#include "image.h"
-#include "image_proxy.h"
-#include "raw_image_proxy.h"
#include "raw_image_proxy.h"
#include "film.h"
#include "log.h"
+#include "compose.hpp"
+#include <iostream>
#include "i18n.h"
, _last_seek_accurate (true)
, _ignore_video (false)
{
- _black_image.reset (new Image (PIX_FMT_RGB24, _video_content->video_size(), true));
+ _black_image.reset (new Image (AV_PIX_FMT_RGB24, _video_content->video_size(), true));
_black_image->make_black ();
}
one after the end of _decoded_video we need to seek.
*/
- _video_content->film()->log()->log (String::compose ("VD has request for %1", frame), Log::TYPE_DEBUG_DECODE);
+ _video_content->film()->log()->log (String::compose ("VD has request for %1", frame), LogEntry::TYPE_DEBUG_DECODE);
if (_decoded_video.empty() || frame < _decoded_video.front().frame || frame > (_decoded_video.back().frame + 1)) {
seek (ContentTime::from_frames (frame, _video_content->video_frame_rate()), accurate);
break;
}
- if (pass ()) {
+ if (pass (PASS_REASON_VIDEO, accurate)) {
/* The decoder has nothing more for us */
break;
}
dec = decoded_video (frame);
} else {
/* Any frame will do: use the first one that comes out of pass() */
- while (_decoded_video.empty() && !pass ()) {}
+ while (_decoded_video.empty() && !pass (PASS_REASON_VIDEO, accurate)) {}
if (!_decoded_video.empty ()) {
dec.push_back (_decoded_video.front ());
}
return dec;
}
-/** Fill _decoded_video from `from' up to, but not including, `to' */
+/** Fill _decoded_video from `from' up to, but not including, `to' with
+ * a frame for one particular Eyes value (which could be EYES_BOTH,
+ * EYES_LEFT or EYES_RIGHT)
+ */
void
-VideoDecoder::fill_2d (Frame from, Frame to)
+VideoDecoder::fill_one_eye (Frame from, Frame to, Eyes eye)
{
if (to == 0) {
/* Already OK */
}
/* Fill with black... */
- boost::shared_ptr<const ImageProxy> filler_image (new RawImageProxy (_black_image));
+ shared_ptr<const ImageProxy> filler_image (new RawImageProxy (_black_image));
Part filler_part = PART_WHOLE;
/* ...unless there's some video we can fill with */
test_gaps++;
#endif
_decoded_video.push_back (
- ContentVideo (filler_image, EYES_BOTH, filler_part, i)
+ ContentVideo (filler_image, eye, filler_part, i)
);
}
}
-/** Fill _decoded_video from `from' up to, but not including, `to' */
+/** Fill _decoded_video from `from' up to, but not including, `to'
+ * adding both left and right eye frames.
+ */
void
-VideoDecoder::fill_3d (Frame from, Frame to, Eyes eye)
+VideoDecoder::fill_both_eyes (Frame from, Frame to, Eyes eye)
{
if (to == 0 && eye == EYES_LEFT) {
/* Already OK */
}
/* Fill with black... */
- boost::shared_ptr<const ImageProxy> filler_left_image (new RawImageProxy (_black_image));
- boost::shared_ptr<const ImageProxy> filler_right_image (new RawImageProxy (_black_image));
+ shared_ptr<const ImageProxy> filler_left_image (new RawImageProxy (_black_image));
+ shared_ptr<const ImageProxy> filler_right_image (new RawImageProxy (_black_image));
Part filler_left_part = PART_WHOLE;
Part filler_right_part = PART_WHOLE;
return;
}
- _video_content->film()->log()->log (String::compose ("VD receives %1", frame), Log::TYPE_DEBUG_DECODE);
+ _video_content->film()->log()->log (String::compose ("VD receives %1", frame), LogEntry::TYPE_DEBUG_DECODE);
/* We may receive the same frame index twice for 3D, and we need to know
when that happens.
and the things we are about to push.
*/
- boost::optional<Frame> from;
- boost::optional<Frame> to;
+ optional<Frame> from;
+ optional<Frame> to;
if (_decoded_video.empty() && _last_seek_time && _last_seek_accurate) {
from = _last_seek_time->frames_round (_video_content->video_frame_rate ());
}
if (from) {
- if (_video_content->video_frame_type() == VIDEO_FRAME_TYPE_2D) {
- fill_2d (from.get(), to.get ());
- } else {
- fill_3d (from.get(), to.get(), to_push.front().eyes);
+ switch (_video_content->video_frame_type ()) {
+ case VIDEO_FRAME_TYPE_2D:
+ fill_one_eye (from.get(), to.get (), EYES_BOTH);
+ break;
+ case VIDEO_FRAME_TYPE_3D_LEFT_RIGHT:
+ case VIDEO_FRAME_TYPE_3D_TOP_BOTTOM:
+ case VIDEO_FRAME_TYPE_3D_ALTERNATE:
+ fill_both_eyes (from.get(), to.get(), to_push.front().eyes);
+ break;
+ case VIDEO_FRAME_TYPE_3D_LEFT:
+ fill_one_eye (from.get(), to.get (), EYES_LEFT);
+ break;
+ case VIDEO_FRAME_TYPE_3D_RIGHT:
+ fill_one_eye (from.get(), to.get (), EYES_RIGHT);
}
}
copy (to_push.begin(), to_push.end(), back_inserter (_decoded_video));
- /* We can't let this build up too much or we will run out of memory. We need to allow
- the most frames that can exist between blocks of sound in a multiplexed file.
+ /* We can't let this build up too much or we will run out of memory. There is a
+ `best' value for the allowed size of _decoded_video which balances memory use
+ with decoding efficiency (lack of seeks). Throwing away video frames here
+ is not a problem for correctness, so do it.
*/
- DCPOMATIC_ASSERT (_decoded_video.size() <= 96);
+ while (_decoded_video.size() > 96) {
+ _decoded_video.pop_back ();
+ }
}
void