diff options
Diffstat (limited to 'src/lib/player.cc')
| -rw-r--r-- | src/lib/player.cc | 840 |
1 files changed, 375 insertions, 465 deletions
diff --git a/src/lib/player.cc b/src/lib/player.cc index 7bf78c905..d0eb27aa3 100644 --- a/src/lib/player.cc +++ b/src/lib/player.cc @@ -18,23 +18,29 @@ */ #include <stdint.h> +#include <algorithm> #include "player.h" #include "film.h" #include "ffmpeg_decoder.h" +#include "audio_buffers.h" #include "ffmpeg_content.h" #include "image_decoder.h" #include "image_content.h" #include "sndfile_decoder.h" #include "sndfile_content.h" #include "subtitle_content.h" +#include "subrip_decoder.h" +#include "subrip_content.h" #include "playlist.h" #include "job.h" #include "image.h" #include "image_proxy.h" #include "ratio.h" -#include "resampler.h" #include "log.h" #include "scaler.h" +#include "render_subtitles.h" +#include "config.h" +#include "content_video.h" #include "player_video_frame.h" #include "frame_rate_change.h" @@ -44,23 +50,22 @@ using std::list; using std::cout; using std::min; using std::max; +using std::min; using std::vector; using std::pair; using std::map; +using std::make_pair; using boost::shared_ptr; using boost::weak_ptr; using boost::dynamic_pointer_cast; +using boost::optional; Player::Player (shared_ptr<const Film> f, shared_ptr<const Playlist> p) : _film (f) , _playlist (p) - , _video (true) - , _audio (true) , _have_valid_pieces (false) - , _video_position (0) - , _audio_position (0) - , _audio_merger (f->audio_channels(), bind (&Film::time_to_audio_frames, f.get(), _1), bind (&Film::audio_frames_to_time, f.get(), _1)) - , _last_emit_was_black (false) + , _approximate_size (false) + , _burn_subtitles (false) { _playlist_changed_connection = _playlist->Changed.connect (bind (&Player::playlist_changed, this)); _playlist_content_changed_connection = _playlist->ContentChanged.connect (bind (&Player::content_changed, this, _1, _2, _3)); @@ -69,573 +74,478 @@ Player::Player (shared_ptr<const Film> f, shared_ptr<const Playlist> p) } void -Player::disable_video () -{ - _video = false; -} - -void -Player::disable_audio () +Player::setup_pieces () { - _audio = false; -} + list<shared_ptr<Piece> > old_pieces = _pieces; + _pieces.clear (); -bool -Player::pass () -{ - if (!_have_valid_pieces) { - setup_pieces (); - } + ContentList content = _playlist->content (); - Time earliest_t = TIME_MAX; - shared_ptr<Piece> earliest; - enum { - VIDEO, - AUDIO - } type = VIDEO; + for (ContentList::iterator i = content.begin(); i != content.end(); ++i) { - for (list<shared_ptr<Piece> >::iterator i = _pieces.begin(); i != _pieces.end(); ++i) { - if ((*i)->decoder->done () || (*i)->content->length_after_trim() == 0) { + if (!(*i)->paths_valid ()) { continue; } - - shared_ptr<VideoDecoder> vd = dynamic_pointer_cast<VideoDecoder> ((*i)->decoder); - shared_ptr<AudioDecoder> ad = dynamic_pointer_cast<AudioDecoder> ((*i)->decoder); - - if (_video && vd) { - if ((*i)->video_position < earliest_t) { - earliest_t = (*i)->video_position; - earliest = *i; - type = VIDEO; + + shared_ptr<Decoder> decoder; + optional<FrameRateChange> frc; + + /* Work out a FrameRateChange for the best overlap video for this content, in case we need it below */ + DCPTime best_overlap_t; + shared_ptr<VideoContent> best_overlap; + for (ContentList::iterator j = content.begin(); j != content.end(); ++j) { + shared_ptr<VideoContent> vc = dynamic_pointer_cast<VideoContent> (*j); + if (!vc) { + continue; } - } - - if (_audio && ad && ad->has_audio ()) { - if ((*i)->audio_position < earliest_t) { - earliest_t = (*i)->audio_position; - earliest = *i; - type = AUDIO; + + DCPTime const overlap = max (vc->position(), (*i)->position()) - min (vc->end(), (*i)->end()); + if (overlap > best_overlap_t) { + best_overlap = vc; + best_overlap_t = overlap; } } - } - - if (!earliest) { - flush (); - return true; - } - switch (type) { - case VIDEO: - if (earliest_t > _video_position) { - emit_black (); + optional<FrameRateChange> best_overlap_frc; + if (best_overlap) { + best_overlap_frc = FrameRateChange (best_overlap->video_frame_rate(), _film->video_frame_rate ()); } else { - if (earliest->repeating ()) { - earliest->repeat (this); - } else { - earliest->decoder->pass (); - } + /* No video overlap; e.g. if the DCP is just audio */ + best_overlap_frc = FrameRateChange (_film->video_frame_rate(), _film->video_frame_rate ()); } - break; - case AUDIO: - if (earliest_t > _audio_position) { - emit_silence (_film->time_to_audio_frames (earliest_t - _audio_position)); - } else { - earliest->decoder->pass (); - - if (earliest->decoder->done()) { - shared_ptr<AudioContent> ac = dynamic_pointer_cast<AudioContent> (earliest->content); - assert (ac); - shared_ptr<Resampler> re = resampler (ac, false); - if (re) { - shared_ptr<const AudioBuffers> b = re->flush (); - if (b->frames ()) { - process_audio (earliest, b, ac->audio_length (), true); - } - } - } + /* FFmpeg */ + shared_ptr<const FFmpegContent> fc = dynamic_pointer_cast<const FFmpegContent> (*i); + if (fc) { + decoder.reset (new FFmpegDecoder (fc, _film->log())); + frc = FrameRateChange (fc->video_frame_rate(), _film->video_frame_rate()); } - break; - } - if (_audio) { - boost::optional<Time> audio_done_up_to; - for (list<shared_ptr<Piece> >::iterator i = _pieces.begin(); i != _pieces.end(); ++i) { - if ((*i)->decoder->done ()) { - continue; + /* ImageContent */ + shared_ptr<const ImageContent> ic = dynamic_pointer_cast<const ImageContent> (*i); + if (ic) { + /* See if we can re-use an old ImageDecoder */ + for (list<shared_ptr<Piece> >::const_iterator j = old_pieces.begin(); j != old_pieces.end(); ++j) { + shared_ptr<ImageDecoder> imd = dynamic_pointer_cast<ImageDecoder> ((*j)->decoder); + if (imd && imd->content() == ic) { + decoder = imd; + } } - shared_ptr<AudioDecoder> ad = dynamic_pointer_cast<AudioDecoder> ((*i)->decoder); - if (ad && ad->has_audio ()) { - audio_done_up_to = min (audio_done_up_to.get_value_or (TIME_MAX), (*i)->audio_position); + if (!decoder) { + decoder.reset (new ImageDecoder (ic)); } + + frc = FrameRateChange (ic->video_frame_rate(), _film->video_frame_rate()); + } + + /* SndfileContent */ + shared_ptr<const SndfileContent> sc = dynamic_pointer_cast<const SndfileContent> (*i); + if (sc) { + decoder.reset (new SndfileDecoder (sc)); + frc = best_overlap_frc; } - if (audio_done_up_to) { - TimedAudioBuffers<Time> tb = _audio_merger.pull (audio_done_up_to.get ()); - Audio (tb.audio, tb.time); - _audio_position += _film->audio_frames_to_time (tb.audio->frames ()); + /* SubRipContent */ + shared_ptr<const SubRipContent> rc = dynamic_pointer_cast<const SubRipContent> (*i); + if (rc) { + decoder.reset (new SubRipDecoder (rc)); + frc = best_overlap_frc; } + + _pieces.push_back (shared_ptr<Piece> (new Piece (*i, decoder, frc.get ()))); } - - return false; + + _have_valid_pieces = true; } -/** @param extra Amount of extra time to add to the content frame's time (for repeat) */ void -Player::process_video (weak_ptr<Piece> weak_piece, shared_ptr<const ImageProxy> image, Eyes eyes, Part part, bool same, VideoContent::Frame frame, Time extra) +Player::content_changed (weak_ptr<Content> w, int property, bool frequent) { - /* Keep a note of what came in so that we can repeat it if required */ - _last_incoming_video.weak_piece = weak_piece; - _last_incoming_video.image = image; - _last_incoming_video.eyes = eyes; - _last_incoming_video.part = part; - _last_incoming_video.same = same; - _last_incoming_video.frame = frame; - _last_incoming_video.extra = extra; - - shared_ptr<Piece> piece = weak_piece.lock (); - if (!piece) { - return; - } - - shared_ptr<VideoContent> content = dynamic_pointer_cast<VideoContent> (piece->content); - assert (content); - - FrameRateChange frc (content->video_frame_rate(), _film->video_frame_rate()); - if (frc.skip && (frame % 2) == 1) { - return; - } - - Time const relative_time = (frame * frc.factor() * TIME_HZ / _film->video_frame_rate()); - if (content->trimmed (relative_time)) { + shared_ptr<Content> c = w.lock (); + if (!c) { return; } - Time const time = content->position() + relative_time + extra - content->trim_start (); - libdcp::Size const image_size = content->scale().size (content, _video_container_size, _film->frame_size ()); - - shared_ptr<PlayerVideoFrame> pi ( - new PlayerVideoFrame ( - image, - content->crop(), - image_size, - _video_container_size, - _film->scaler(), - eyes, - part, - content->colour_conversion() - ) - ); - - if (_film->with_subtitles ()) { - for (list<Subtitle>::const_iterator i = _subtitles.begin(); i != _subtitles.end(); ++i) { - if (i->covers (time)) { - /* This may be true for more than one of _subtitles, but the last (latest-starting) - one is the one we want to use, so that's ok. - */ - Position<int> const container_offset ( - (_video_container_size.width - image_size.width) / 2, - (_video_container_size.height - image_size.width) / 2 - ); - - pi->set_subtitle (i->out_image(), i->out_position() + container_offset); - } - } - } - - /* Clear out old subtitles */ - for (list<Subtitle>::iterator i = _subtitles.begin(); i != _subtitles.end(); ) { - list<Subtitle>::iterator j = i; - ++j; + if ( + property == ContentProperty::POSITION || + property == ContentProperty::LENGTH || + property == ContentProperty::TRIM_START || + property == ContentProperty::TRIM_END || + property == ContentProperty::PATH || + property == VideoContentProperty::VIDEO_FRAME_TYPE + ) { - if (i->ends_before (time)) { - _subtitles.erase (i); - } - - i = j; - } - -#ifdef DCPOMATIC_DEBUG - _last_video = piece->content; -#endif - - Video (pi, same, time); - - _last_emit_was_black = false; - _video_position = piece->video_position = (time + TIME_HZ / _film->video_frame_rate()); + _have_valid_pieces = false; + Changed (frequent); - if (frc.repeat > 1 && !piece->repeating ()) { - piece->set_repeat (_last_incoming_video, frc.repeat - 1); + } else if ( + property == SubtitleContentProperty::SUBTITLE_X_OFFSET || + property == SubtitleContentProperty::SUBTITLE_Y_OFFSET || + property == SubtitleContentProperty::SUBTITLE_SCALE || + property == VideoContentProperty::VIDEO_CROP || + property == VideoContentProperty::VIDEO_SCALE || + property == VideoContentProperty::VIDEO_FRAME_RATE + ) { + + Changed (frequent); } } /** @param already_resampled true if this data has already been through the chain up to the resampler */ void -Player::process_audio (weak_ptr<Piece> weak_piece, shared_ptr<const AudioBuffers> audio, AudioContent::Frame frame, bool already_resampled) +Player::playlist_changed () { - shared_ptr<Piece> piece = weak_piece.lock (); - if (!piece) { - return; - } + _have_valid_pieces = false; + Changed (false); +} - shared_ptr<AudioContent> content = dynamic_pointer_cast<AudioContent> (piece->content); - assert (content); +void +Player::set_video_container_size (dcp::Size s) +{ + _video_container_size = s; - if (!already_resampled) { - /* Gain */ - if (content->audio_gain() != 0) { - shared_ptr<AudioBuffers> gain (new AudioBuffers (audio)); - gain->apply_gain (content->audio_gain ()); - audio = gain; - } - - /* Resample */ - if (content->content_audio_frame_rate() != content->output_audio_frame_rate()) { - shared_ptr<Resampler> r = resampler (content, true); - pair<shared_ptr<const AudioBuffers>, AudioContent::Frame> ro = r->run (audio, frame); - audio = ro.first; - frame = ro.second; - } - } - - Time const relative_time = _film->audio_frames_to_time (frame); + _black_image.reset (new Image (PIX_FMT_RGB24, _video_container_size, true)); + _black_image->make_black (); +} - if (content->trimmed (relative_time)) { - return; +void +Player::film_changed (Film::Property p) +{ + /* Here we should notice Film properties that affect our output, and + alert listeners that our output now would be different to how it was + last time we were run. + */ + + if (p == Film::SCALER || p == Film::WITH_SUBTITLES || p == Film::CONTAINER || p == Film::VIDEO_FRAME_RATE) { + Changed (false); } +} - Time time = content->position() + (content->audio_delay() * TIME_HZ / 1000) + relative_time - content->trim_start (); +list<PositionImage> +Player::process_content_image_subtitles (shared_ptr<SubtitleContent> content, list<shared_ptr<ContentImageSubtitle> > subs) const +{ + list<PositionImage> all; - /* Remap channels */ - shared_ptr<AudioBuffers> dcp_mapped (new AudioBuffers (_film->audio_channels(), audio->frames())); - dcp_mapped->make_silent (); - - AudioMapping map = content->audio_mapping (); - for (int i = 0; i < map.content_channels(); ++i) { - for (int j = 0; j < _film->audio_channels(); ++j) { - if (map.get (i, static_cast<libdcp::Channel> (j)) > 0) { - dcp_mapped->accumulate_channel ( - audio.get(), - i, - static_cast<libdcp::Channel> (j), - map.get (i, static_cast<libdcp::Channel> (j)) - ); - } + for (list<shared_ptr<ContentImageSubtitle> >::const_iterator i = subs.begin(); i != subs.end(); ++i) { + if (!(*i)->image) { + continue; } + + dcpomatic::Rect<double> in_rect = (*i)->rectangle; + dcp::Size scaled_size; + + in_rect.x += content->subtitle_x_offset (); + in_rect.y += content->subtitle_y_offset (); + + /* We will scale the subtitle up to fit _video_container_size, and also by the additional subtitle_scale */ + scaled_size.width = in_rect.width * _video_container_size.width * content->subtitle_scale (); + scaled_size.height = in_rect.height * _video_container_size.height * content->subtitle_scale (); + + /* Then we need a corrective translation, consisting of two parts: + * + * 1. that which is the result of the scaling of the subtitle by _video_container_size; this will be + * rect.x * _video_container_size.width and rect.y * _video_container_size.height. + * + * 2. that to shift the origin of the scale by subtitle_scale to the centre of the subtitle; this will be + * (width_before_subtitle_scale * (1 - subtitle_scale) / 2) and + * (height_before_subtitle_scale * (1 - subtitle_scale) / 2). + * + * Combining these two translations gives these expressions. + */ + + all.push_back ( + PositionImage ( + (*i)->image->scale ( + scaled_size, + Scaler::from_id ("bicubic"), + (*i)->image->pixel_format (), + true + ), + Position<int> ( + rint (_video_container_size.width * (in_rect.x + (in_rect.width * (1 - content->subtitle_scale ()) / 2))), + rint (_video_container_size.height * (in_rect.y + (in_rect.height * (1 - content->subtitle_scale ()) / 2))) + ) + ) + ); } - audio = dcp_mapped; + return all; +} - /* We must cut off anything that comes before the start of all time */ - if (time < 0) { - int const frames = - time * _film->audio_frame_rate() / TIME_HZ; - if (frames >= audio->frames ()) { - return; +list<PositionImage> +Player::process_content_text_subtitles (list<shared_ptr<ContentTextSubtitle> > sub) const +{ + list<PositionImage> all; + for (list<shared_ptr<ContentTextSubtitle> >::const_iterator i = sub.begin(); i != sub.end(); ++i) { + if (!(*i)->subs.empty ()) { + all.push_back (render_subtitles ((*i)->subs, _video_container_size)); } - - shared_ptr<AudioBuffers> trimmed (new AudioBuffers (audio->channels(), audio->frames() - frames)); - trimmed->copy_from (audio.get(), audio->frames() - frames, frames, 0); - - audio = trimmed; - time = 0; } - _audio_merger.push (audio, time); - piece->audio_position += _film->audio_frames_to_time (audio->frames ()); + return all; } void -Player::flush () +Player::set_approximate_size () { - TimedAudioBuffers<Time> tb = _audio_merger.flush (); - if (_audio && tb.audio) { - Audio (tb.audio, tb.time); - _audio_position += _film->audio_frames_to_time (tb.audio->frames ()); - } + _approximate_size = true; +} - while (_video && _video_position < _audio_position) { - emit_black (); - } +shared_ptr<PlayerVideoFrame> +Player::black_player_video_frame () const +{ + return shared_ptr<PlayerVideoFrame> ( + new PlayerVideoFrame ( + shared_ptr<const ImageProxy> (new RawImageProxy (_black_image, _film->log ())), + Crop (), + _video_container_size, + _video_container_size, + Scaler::from_id ("bicubic"), + EYES_BOTH, + PART_WHOLE, + Config::instance()->colour_conversions().front().conversion + ) + ); +} - while (_audio && _audio_position < _video_position) { - emit_silence (_film->time_to_audio_frames (_video_position - _audio_position)); +shared_ptr<PlayerVideoFrame> +Player::content_to_player_video_frame ( + shared_ptr<VideoContent> content, + ContentVideo content_video, + list<shared_ptr<Piece> > subs, + DCPTime time, + dcp::Size image_size) const +{ + shared_ptr<PlayerVideoFrame> pvf ( + new PlayerVideoFrame ( + content_video.image, + content->crop (), + image_size, + _video_container_size, + _film->scaler(), + content_video.eyes, + content_video.part, + content->colour_conversion () + ) + ); + + + /* Add subtitles */ + + list<PositionImage> sub_images; + + for (list<shared_ptr<Piece> >::const_iterator i = subs.begin(); i != subs.end(); ++i) { + shared_ptr<SubtitleDecoder> subtitle_decoder = dynamic_pointer_cast<SubtitleDecoder> ((*i)->decoder); + shared_ptr<SubtitleContent> subtitle_content = dynamic_pointer_cast<SubtitleContent> ((*i)->content); + ContentTime const from = dcp_to_content_subtitle (*i, time); + ContentTime const to = from + ContentTime::from_frames (1, content->video_frame_rate ()); + + list<shared_ptr<ContentImageSubtitle> > image_subtitles = subtitle_decoder->get_image_subtitles (ContentTimePeriod (from, to)); + if (!image_subtitles.empty ()) { + list<PositionImage> im = process_content_image_subtitles ( + subtitle_content, + image_subtitles + ); + + copy (im.begin(), im.end(), back_inserter (sub_images)); + } + + if (_burn_subtitles) { + list<shared_ptr<ContentTextSubtitle> > text_subtitles = subtitle_decoder->get_text_subtitles (ContentTimePeriod (from, to)); + if (!text_subtitles.empty ()) { + list<PositionImage> im = process_content_text_subtitles (text_subtitles); + copy (im.begin(), im.end(), back_inserter (sub_images)); + } + } } + if (!sub_images.empty ()) { + pvf->set_subtitle (merge (sub_images)); + } + + return pvf; } -/** Seek so that the next pass() will yield (approximately) the requested frame. - * Pass accurate = true to try harder to get close to the request. - * @return true on error - */ -void -Player::seek (Time t, bool accurate) +/** @return All PlayerVideoFrames at the given time (there may be two frames for 3D) */ +list<shared_ptr<PlayerVideoFrame> > +Player::get_video (DCPTime time, bool accurate) { if (!_have_valid_pieces) { setup_pieces (); } + + list<shared_ptr<Piece> > ov = overlaps<VideoContent> ( + time, + time + DCPTime::from_frames (1, _film->video_frame_rate ()) + ); - if (_pieces.empty ()) { - return; + list<shared_ptr<PlayerVideoFrame> > pvf; + + if (ov.empty ()) { + /* No video content at this time */ + pvf.push_back (black_player_video_frame ()); + return pvf; } - for (list<shared_ptr<Piece> >::iterator i = _pieces.begin(); i != _pieces.end(); ++i) { - shared_ptr<VideoContent> vc = dynamic_pointer_cast<VideoContent> ((*i)->content); - if (!vc) { - continue; - } - - /* s is the offset of t from the start position of this content */ - Time s = t - vc->position (); - s = max (static_cast<Time> (0), s); - s = min (vc->length_after_trim(), s); - - /* Hence set the piece positions to the `global' time */ - (*i)->video_position = (*i)->audio_position = vc->position() + s; + /* Create a PlayerVideoFrame from the content's video at this time */ - /* And seek the decoder */ - dynamic_pointer_cast<VideoDecoder>((*i)->decoder)->seek ( - vc->time_to_content_video_frames (s + vc->trim_start ()), accurate - ); + shared_ptr<Piece> piece = ov.back (); + shared_ptr<VideoDecoder> decoder = dynamic_pointer_cast<VideoDecoder> (piece->decoder); + assert (decoder); + shared_ptr<VideoContent> content = dynamic_pointer_cast<VideoContent> (piece->content); + assert (content); - (*i)->reset_repeat (); + list<ContentVideo> content_video = decoder->get_video (dcp_to_content_video (piece, time), accurate); + if (content_video.empty ()) { + pvf.push_back (black_player_video_frame ()); + return pvf; } - _video_position = _audio_position = t; + dcp::Size image_size = content->scale().size (content, _video_container_size, _film->frame_size ()); + if (_approximate_size) { + image_size.width &= ~3; + image_size.height &= ~3; + } - /* XXX: don't seek audio because we don't need to... */ + for (list<ContentVideo>::const_iterator i = content_video.begin(); i != content_video.end(); ++i) { + list<shared_ptr<Piece> > subs = overlaps<SubtitleContent> ( + time, + time + DCPTime::from_frames (1, _film->video_frame_rate ()) + ); + + pvf.push_back (content_to_player_video_frame (content, *i, subs, time, image_size)); + } + + return pvf; } -void -Player::setup_pieces () +shared_ptr<AudioBuffers> +Player::get_audio (DCPTime time, DCPTime length, bool accurate) { - list<shared_ptr<Piece> > old_pieces = _pieces; + if (!_have_valid_pieces) { + setup_pieces (); + } - _pieces.clear (); + AudioFrame const length_frames = length.frames (_film->audio_frame_rate ()); - ContentList content = _playlist->content (); - sort (content.begin(), content.end(), ContentSorter ()); + shared_ptr<AudioBuffers> audio (new AudioBuffers (_film->audio_channels(), length_frames)); + audio->make_silent (); + + list<shared_ptr<Piece> > ov = overlaps<AudioContent> (time, time + length); + if (ov.empty ()) { + return audio; + } - for (ContentList::iterator i = content.begin(); i != content.end(); ++i) { + for (list<shared_ptr<Piece> >::iterator i = ov.begin(); i != ov.end(); ++i) { - if (!(*i)->paths_valid ()) { + shared_ptr<AudioContent> content = dynamic_pointer_cast<AudioContent> ((*i)->content); + assert (content); + shared_ptr<AudioDecoder> decoder = dynamic_pointer_cast<AudioDecoder> ((*i)->decoder); + assert (decoder); + + if (content->audio_frame_rate() == 0) { + /* This AudioContent has no audio (e.g. if it is an FFmpegContent with no + * audio stream). + */ continue; } - shared_ptr<Piece> piece (new Piece (*i)); - - /* XXX: into content? */ - - shared_ptr<const FFmpegContent> fc = dynamic_pointer_cast<const FFmpegContent> (*i); - if (fc) { - shared_ptr<FFmpegDecoder> fd (new FFmpegDecoder (_film, fc, _video, _audio)); - - fd->Video.connect (bind (&Player::process_video, this, weak_ptr<Piece> (piece), _1, _2, _3, _4, _5, 0)); - fd->Audio.connect (bind (&Player::process_audio, this, weak_ptr<Piece> (piece), _1, _2, false)); - fd->Subtitle.connect (bind (&Player::process_subtitle, this, weak_ptr<Piece> (piece), _1, _2, _3, _4)); - - fd->seek (fc->time_to_content_video_frames (fc->trim_start ()), true); - piece->decoder = fd; + /* The time that we should request from the content */ + DCPTime request = time - DCPTime::from_seconds (content->audio_delay() / 1000.0); + DCPTime offset; + if (request < DCPTime ()) { + /* We went off the start of the content, so we will need to offset + the stuff we get back. + */ + offset = -request; + request = DCPTime (); } - - shared_ptr<const ImageContent> ic = dynamic_pointer_cast<const ImageContent> (*i); - if (ic) { - bool reusing = false; - - /* See if we can re-use an old ImageDecoder */ - for (list<shared_ptr<Piece> >::const_iterator j = old_pieces.begin(); j != old_pieces.end(); ++j) { - shared_ptr<ImageDecoder> imd = dynamic_pointer_cast<ImageDecoder> ((*j)->decoder); - if (imd && imd->content() == ic) { - piece = *j; - reusing = true; - } - } - if (!reusing) { - shared_ptr<ImageDecoder> id (new ImageDecoder (_film, ic)); - id->Video.connect (bind (&Player::process_video, this, weak_ptr<Piece> (piece), _1, _2, _3, _4, _5, 0)); - piece->decoder = id; - } - } + AudioFrame const content_frame = dcp_to_content_audio (*i, request); - shared_ptr<const SndfileContent> sc = dynamic_pointer_cast<const SndfileContent> (*i); - if (sc) { - shared_ptr<AudioDecoder> sd (new SndfileDecoder (_film, sc)); - sd->Audio.connect (bind (&Player::process_audio, this, weak_ptr<Piece> (piece), _1, _2, false)); + /* Audio from this piece's decoder (which might be more or less than what we asked for) */ + shared_ptr<ContentAudio> all = decoder->get_audio (content_frame, length_frames, accurate); - piece->decoder = sd; + /* Gain */ + if (content->audio_gain() != 0) { + shared_ptr<AudioBuffers> gain (new AudioBuffers (all->audio)); + gain->apply_gain (content->audio_gain ()); + all->audio = gain; } - _pieces.push_back (piece); - } - - _have_valid_pieces = true; -} - -void -Player::content_changed (weak_ptr<Content> w, int property, bool frequent) -{ - shared_ptr<Content> c = w.lock (); - if (!c) { - return; - } - - if ( - property == ContentProperty::POSITION || property == ContentProperty::LENGTH || - property == ContentProperty::TRIM_START || property == ContentProperty::TRIM_END || - property == VideoContentProperty::VIDEO_FRAME_TYPE - ) { - - _have_valid_pieces = false; - Changed (frequent); - - } else if ( - property == SubtitleContentProperty::SUBTITLE_X_OFFSET || - property == SubtitleContentProperty::SUBTITLE_Y_OFFSET || - property == SubtitleContentProperty::SUBTITLE_SCALE - ) { - - for (list<Subtitle>::iterator i = _subtitles.begin(); i != _subtitles.end(); ++i) { - i->update (_film, _video_container_size); + /* Remap channels */ + shared_ptr<AudioBuffers> dcp_mapped (new AudioBuffers (_film->audio_channels(), all->audio->frames())); + dcp_mapped->make_silent (); + AudioMapping map = content->audio_mapping (); + for (int i = 0; i < map.content_channels(); ++i) { + for (int j = 0; j < _film->audio_channels(); ++j) { + if (map.get (i, static_cast<dcp::Channel> (j)) > 0) { + dcp_mapped->accumulate_channel ( + all->audio.get(), + i, + j, + map.get (i, static_cast<dcp::Channel> (j)) + ); + } + } } - Changed (frequent); - - } else if ( - property == VideoContentProperty::VIDEO_CROP || property == VideoContentProperty::VIDEO_SCALE || - property == VideoContentProperty::VIDEO_FRAME_RATE - ) { - - Changed (frequent); + all->audio = dcp_mapped; - } else if (property == ContentProperty::PATH) { - - _have_valid_pieces = false; - Changed (frequent); + audio->accumulate_frames ( + all->audio.get(), + content_frame - all->frame, + offset.frames (_film->audio_frame_rate()), + min (AudioFrame (all->audio->frames()), length_frames) - offset.frames (_film->audio_frame_rate ()) + ); } } -void -Player::playlist_changed () -{ - _have_valid_pieces = false; - Changed (false); -} - -void -Player::set_video_container_size (libdcp::Size s) -{ - _video_container_size = s; - - shared_ptr<Image> im (new Image (PIX_FMT_RGB24, _video_container_size, true)); - im->make_black (); - - _black_frame.reset ( - new PlayerVideoFrame ( - shared_ptr<ImageProxy> (new RawImageProxy (im, _film->log ())), - Crop(), - _video_container_size, - _video_container_size, - Scaler::from_id ("bicubic"), - EYES_BOTH, - PART_WHOLE, - ColourConversion () - ) - ); -} - -shared_ptr<Resampler> -Player::resampler (shared_ptr<AudioContent> c, bool create) +VideoFrame +Player::dcp_to_content_video (shared_ptr<const Piece> piece, DCPTime t) const { - map<shared_ptr<AudioContent>, shared_ptr<Resampler> >::iterator i = _resamplers.find (c); - if (i != _resamplers.end ()) { - return i->second; - } - - if (!create) { - return shared_ptr<Resampler> (); - } - - LOG_GENERAL ( - "Creating new resampler for %1 to %2 with %3 channels", c->content_audio_frame_rate(), c->output_audio_frame_rate(), c->audio_channels() - ); + /* s is the offset of t from the start position of this content */ + DCPTime s = t - piece->content->position (); + s = DCPTime (max (int64_t (0), s.get ())); + s = DCPTime (min (piece->content->length_after_trim().get(), s.get())); - shared_ptr<Resampler> r (new Resampler (c->content_audio_frame_rate(), c->output_audio_frame_rate(), c->audio_channels())); - _resamplers[c] = r; - return r; + /* Convert this to the content frame */ + return DCPTime (s + piece->content->trim_start()).frames (_film->video_frame_rate()) * piece->frc.factor (); } -void -Player::emit_black () +AudioFrame +Player::dcp_to_content_audio (shared_ptr<const Piece> piece, DCPTime t) const { -#ifdef DCPOMATIC_DEBUG - _last_video.reset (); -#endif - - Video (_black_frame, _last_emit_was_black, _video_position); - _video_position += _film->video_frames_to_time (1); - _last_emit_was_black = true; -} + /* s is the offset of t from the start position of this content */ + DCPTime s = t - piece->content->position (); + s = DCPTime (max (int64_t (0), s.get ())); + s = DCPTime (min (piece->content->length_after_trim().get(), s.get())); -void -Player::emit_silence (OutputAudioFrame most) -{ - if (most == 0) { - return; - } - - OutputAudioFrame N = min (most, _film->audio_frame_rate() / 2); - shared_ptr<AudioBuffers> silence (new AudioBuffers (_film->audio_channels(), N)); - silence->make_silent (); - Audio (silence, _audio_position); - _audio_position += _film->audio_frames_to_time (N); + /* Convert this to the content frame */ + return DCPTime (s + piece->content->trim_start()).frames (_film->audio_frame_rate()); } -void -Player::film_changed (Film::Property p) +ContentTime +Player::dcp_to_content_subtitle (shared_ptr<const Piece> piece, DCPTime t) const { - /* Here we should notice Film properties that affect our output, and - alert listeners that our output now would be different to how it was - last time we were run. - */ + /* s is the offset of t from the start position of this content */ + DCPTime s = t - piece->content->position (); + s = DCPTime (max (int64_t (0), s.get ())); + s = DCPTime (min (piece->content->length_after_trim().get(), s.get())); - if (p == Film::SCALER || p == Film::WITH_SUBTITLES || p == Film::CONTAINER || p == Film::VIDEO_FRAME_RATE) { - Changed (false); - } + return ContentTime (s + piece->content->trim_start(), piece->frc); } void -Player::process_subtitle (weak_ptr<Piece> weak_piece, shared_ptr<Image> image, dcpomatic::Rect<double> rect, Time from, Time to) +PlayerStatistics::dump (shared_ptr<Log> log) const { - if (!image) { - /* A null image means that we should stop any current subtitles at `from' */ - for (list<Subtitle>::iterator i = _subtitles.begin(); i != _subtitles.end(); ++i) { - i->set_stop (from); - } - } else { - _subtitles.push_back (Subtitle (_film, _video_container_size, weak_piece, image, rect, from, to)); - } + log->log (String::compose ("Video: %1 good %2 skipped %3 black %4 repeat", video.good, video.skip, video.black, video.repeat), Log::TYPE_GENERAL); + log->log (String::compose ("Audio: %1 good %2 skipped %3 silence", audio.good, audio.skip, audio.silence.seconds()), Log::TYPE_GENERAL); } -/** Re-emit the last frame that was emitted, using current settings for crop, ratio, scaler and subtitles. - * @return false if this could not be done. - */ -bool -Player::repeat_last_video () +PlayerStatistics const & +Player::statistics () const { - if (!_last_incoming_video.image || !_have_valid_pieces) { - return false; - } - - process_video ( - _last_incoming_video.weak_piece, - _last_incoming_video.image, - _last_incoming_video.eyes, - _last_incoming_video.part, - _last_incoming_video.same, - _last_incoming_video.frame, - _last_incoming_video.extra - ); - - return true; + return _statistics; } |
