/* Copyright (C) 2013-2014 Carl Hetherington This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ #include #include #include "player.h" #include "film.h" #include "ffmpeg_decoder.h" #include "audio_buffers.h" #include "ffmpeg_content.h" #include "image_decoder.h" #include "image_content.h" #include "sndfile_decoder.h" #include "sndfile_content.h" #include "subtitle_content.h" #include "subrip_decoder.h" #include "subrip_content.h" #include "playlist.h" #include "job.h" #include "image.h" #include "image_proxy.h" #include "ratio.h" #include "log.h" #include "scaler.h" #include "render_subtitles.h" #include "config.h" #include "content_video.h" #include "player_video_frame.h" #include "frame_rate_change.h" #define LOG_GENERAL(...) _film->log()->log (String::compose (__VA_ARGS__), Log::TYPE_GENERAL); using std::list; using std::cout; using std::min; using std::max; using std::min; using std::vector; using std::pair; using std::map; using std::make_pair; using boost::shared_ptr; using boost::weak_ptr; using boost::dynamic_pointer_cast; using boost::optional; Player::Player (shared_ptr f, shared_ptr p) : _film (f) , _playlist (p) , _have_valid_pieces (false) , _approximate_size (false) , _burn_subtitles (false) { _playlist_changed_connection = _playlist->Changed.connect (bind (&Player::playlist_changed, this)); _playlist_content_changed_connection = _playlist->ContentChanged.connect (bind (&Player::content_changed, this, _1, _2, _3)); _film_changed_connection = _film->Changed.connect (bind (&Player::film_changed, this, _1)); set_video_container_size (_film->frame_size ()); } void Player::setup_pieces () { list > old_pieces = _pieces; _pieces.clear (); ContentList content = _playlist->content (); for (ContentList::iterator i = content.begin(); i != content.end(); ++i) { if (!(*i)->paths_valid ()) { continue; } shared_ptr decoder; optional frc; /* Work out a FrameRateChange for the best overlap video for this content, in case we need it below */ DCPTime best_overlap_t; shared_ptr best_overlap; for (ContentList::iterator j = content.begin(); j != content.end(); ++j) { shared_ptr vc = dynamic_pointer_cast (*j); if (!vc) { continue; } DCPTime const overlap = max (vc->position(), (*i)->position()) - min (vc->end(), (*i)->end()); if (overlap > best_overlap_t) { best_overlap = vc; best_overlap_t = overlap; } } optional best_overlap_frc; if (best_overlap) { best_overlap_frc = FrameRateChange (best_overlap->video_frame_rate(), _film->video_frame_rate ()); } else { /* No video overlap; e.g. if the DCP is just audio */ best_overlap_frc = FrameRateChange (_film->video_frame_rate(), _film->video_frame_rate ()); } /* FFmpeg */ shared_ptr fc = dynamic_pointer_cast (*i); if (fc) { decoder.reset (new FFmpegDecoder (fc, _film->log())); frc = FrameRateChange (fc->video_frame_rate(), _film->video_frame_rate()); } /* ImageContent */ shared_ptr ic = dynamic_pointer_cast (*i); if (ic) { /* See if we can re-use an old ImageDecoder */ for (list >::const_iterator j = old_pieces.begin(); j != old_pieces.end(); ++j) { shared_ptr imd = dynamic_pointer_cast ((*j)->decoder); if (imd && imd->content() == ic) { decoder = imd; } } if (!decoder) { decoder.reset (new ImageDecoder (ic)); } frc = FrameRateChange (ic->video_frame_rate(), _film->video_frame_rate()); } /* SndfileContent */ shared_ptr sc = dynamic_pointer_cast (*i); if (sc) { decoder.reset (new SndfileDecoder (sc)); frc = best_overlap_frc; } /* SubRipContent */ shared_ptr rc = dynamic_pointer_cast (*i); if (rc) { decoder.reset (new SubRipDecoder (rc)); frc = best_overlap_frc; } _pieces.push_back (shared_ptr (new Piece (*i, decoder, frc.get ()))); } _have_valid_pieces = true; } void Player::content_changed (weak_ptr w, int property, bool frequent) { shared_ptr c = w.lock (); if (!c) { return; } if ( property == ContentProperty::POSITION || property == ContentProperty::LENGTH || property == ContentProperty::TRIM_START || property == ContentProperty::TRIM_END || property == ContentProperty::PATH || property == VideoContentProperty::VIDEO_FRAME_TYPE ) { _have_valid_pieces = false; Changed (frequent); } else if ( property == SubtitleContentProperty::SUBTITLE_X_OFFSET || property == SubtitleContentProperty::SUBTITLE_Y_OFFSET || property == SubtitleContentProperty::SUBTITLE_SCALE || property == VideoContentProperty::VIDEO_CROP || property == VideoContentProperty::VIDEO_SCALE || property == VideoContentProperty::VIDEO_FRAME_RATE ) { Changed (frequent); } } /** @param already_resampled true if this data has already been through the chain up to the resampler */ void Player::playlist_changed () { _have_valid_pieces = false; Changed (false); } void Player::set_video_container_size (dcp::Size s) { _video_container_size = s; _black_image.reset (new Image (PIX_FMT_RGB24, _video_container_size, true)); _black_image->make_black (); } void Player::film_changed (Film::Property p) { /* Here we should notice Film properties that affect our output, and alert listeners that our output now would be different to how it was last time we were run. */ if (p == Film::SCALER || p == Film::WITH_SUBTITLES || p == Film::CONTAINER || p == Film::VIDEO_FRAME_RATE) { Changed (false); } } list Player::process_content_image_subtitles (shared_ptr content, list > subs) const { list all; for (list >::const_iterator i = subs.begin(); i != subs.end(); ++i) { if (!(*i)->image) { continue; } dcpomatic::Rect in_rect = (*i)->rectangle; dcp::Size scaled_size; in_rect.x += content->subtitle_x_offset (); in_rect.y += content->subtitle_y_offset (); /* We will scale the subtitle up to fit _video_container_size, and also by the additional subtitle_scale */ scaled_size.width = in_rect.width * _video_container_size.width * content->subtitle_scale (); scaled_size.height = in_rect.height * _video_container_size.height * content->subtitle_scale (); /* Then we need a corrective translation, consisting of two parts: * * 1. that which is the result of the scaling of the subtitle by _video_container_size; this will be * rect.x * _video_container_size.width and rect.y * _video_container_size.height. * * 2. that to shift the origin of the scale by subtitle_scale to the centre of the subtitle; this will be * (width_before_subtitle_scale * (1 - subtitle_scale) / 2) and * (height_before_subtitle_scale * (1 - subtitle_scale) / 2). * * Combining these two translations gives these expressions. */ all.push_back ( PositionImage ( (*i)->image->scale ( scaled_size, Scaler::from_id ("bicubic"), (*i)->image->pixel_format (), true ), Position ( rint (_video_container_size.width * (in_rect.x + (in_rect.width * (1 - content->subtitle_scale ()) / 2))), rint (_video_container_size.height * (in_rect.y + (in_rect.height * (1 - content->subtitle_scale ()) / 2))) ) ) ); } return all; } list Player::process_content_text_subtitles (list > sub) const { list all; for (list >::const_iterator i = sub.begin(); i != sub.end(); ++i) { if (!(*i)->subs.empty ()) { all.push_back (render_subtitles ((*i)->subs, _video_container_size)); } } return all; } void Player::set_approximate_size () { _approximate_size = true; } shared_ptr Player::black_player_video_frame () const { return shared_ptr ( new PlayerVideoFrame ( shared_ptr (new RawImageProxy (_black_image, _film->log ())), Crop (), _video_container_size, _video_container_size, Scaler::from_id ("bicubic"), EYES_BOTH, PART_WHOLE, Config::instance()->colour_conversions().front().conversion ) ); } shared_ptr Player::content_to_player_video_frame ( shared_ptr content, ContentVideo content_video, list > subs, DCPTime time, dcp::Size image_size) const { shared_ptr pvf ( new PlayerVideoFrame ( content_video.image, content->crop (), image_size, _video_container_size, _film->scaler(), content_video.eyes, content_video.part, content->colour_conversion () ) ); /* Add subtitles */ list sub_images; for (list >::const_iterator i = subs.begin(); i != subs.end(); ++i) { shared_ptr subtitle_decoder = dynamic_pointer_cast ((*i)->decoder); shared_ptr subtitle_content = dynamic_pointer_cast ((*i)->content); ContentTime const from = dcp_to_content_subtitle (*i, time); ContentTime const to = from + ContentTime::from_frames (1, content->video_frame_rate ()); list > image_subtitles = subtitle_decoder->get_image_subtitles (ContentTimePeriod (from, to)); if (!image_subtitles.empty ()) { list im = process_content_image_subtitles ( subtitle_content, image_subtitles ); copy (im.begin(), im.end(), back_inserter (sub_images)); } if (_burn_subtitles) { list > text_subtitles = subtitle_decoder->get_text_subtitles (ContentTimePeriod (from, to)); if (!text_subtitles.empty ()) { list im = process_content_text_subtitles (text_subtitles); copy (im.begin(), im.end(), back_inserter (sub_images)); } } } if (!sub_images.empty ()) { pvf->set_subtitle (merge (sub_images)); } return pvf; } /** @return All PlayerVideoFrames at the given time (there may be two frames for 3D) */ list > Player::get_video (DCPTime time, bool accurate) { if (!_have_valid_pieces) { setup_pieces (); } list > ov = overlaps ( time, time + DCPTime::from_frames (1, _film->video_frame_rate ()) ); list > pvf; if (ov.empty ()) { /* No video content at this time */ pvf.push_back (black_player_video_frame ()); return pvf; } /* Create a PlayerVideoFrame from the content's video at this time */ shared_ptr piece = ov.back (); shared_ptr decoder = dynamic_pointer_cast (piece->decoder); assert (decoder); shared_ptr content = dynamic_pointer_cast (piece->content); assert (content); list content_video = decoder->get_video (dcp_to_content_video (piece, time), accurate); if (content_video.empty ()) { pvf.push_back (black_player_video_frame ()); return pvf; } dcp::Size image_size = content->scale().size (content, _video_container_size, _film->frame_size ()); if (_approximate_size) { image_size.width &= ~3; image_size.height &= ~3; } for (list::const_iterator i = content_video.begin(); i != content_video.end(); ++i) { list > subs = overlaps ( time, time + DCPTime::from_frames (1, _film->video_frame_rate ()) ); pvf.push_back (content_to_player_video_frame (content, *i, subs, time, image_size)); } return pvf; } shared_ptr Player::get_audio (DCPTime time, DCPTime length, bool accurate) { if (!_have_valid_pieces) { setup_pieces (); } AudioFrame const length_frames = length.frames (_film->audio_frame_rate ()); shared_ptr audio (new AudioBuffers (_film->audio_channels(), length_frames)); audio->make_silent (); list > ov = overlaps (time, time + length); if (ov.empty ()) { return audio; } for (list >::iterator i = ov.begin(); i != ov.end(); ++i) { shared_ptr content = dynamic_pointer_cast ((*i)->content); assert (content); shared_ptr decoder = dynamic_pointer_cast ((*i)->decoder); assert (decoder); if (content->audio_frame_rate() == 0) { /* This AudioContent has no audio (e.g. if it is an FFmpegContent with no * audio stream). */ continue; } /* The time that we should request from the content */ DCPTime request = time - DCPTime::from_seconds (content->audio_delay() / 1000.0); DCPTime offset; if (request < DCPTime ()) { /* We went off the start of the content, so we will need to offset the stuff we get back. */ offset = -request; request = DCPTime (); } AudioFrame const content_frame = dcp_to_content_audio (*i, request); /* Audio from this piece's decoder (which might be more or less than what we asked for) */ shared_ptr all = decoder->get_audio (content_frame, length_frames, accurate); /* Gain */ if (content->audio_gain() != 0) { shared_ptr gain (new AudioBuffers (all->audio)); gain->apply_gain (content->audio_gain ()); all->audio = gain; } /* Remap channels */ shared_ptr dcp_mapped (new AudioBuffers (_film->audio_channels(), all->audio->frames())); dcp_mapped->make_silent (); AudioMapping map = content->audio_mapping (); for (int i = 0; i < map.content_channels(); ++i) { for (int j = 0; j < _film->audio_channels(); ++j) { if (map.get (i, static_cast (j)) > 0) { dcp_mapped->accumulate_channel ( all->audio.get(), i, j, map.get (i, static_cast (j)) ); } } } all->audio = dcp_mapped; audio->accumulate_frames ( all->audio.get(), content_frame - all->frame, offset.frames (_film->audio_frame_rate()), min (AudioFrame (all->audio->frames()), length_frames) - offset.frames (_film->audio_frame_rate ()) ); } } VideoFrame Player::dcp_to_content_video (shared_ptr piece, DCPTime t) const { /* s is the offset of t from the start position of this content */ DCPTime s = t - piece->content->position (); s = DCPTime (max (int64_t (0), s.get ())); s = DCPTime (min (piece->content->length_after_trim().get(), s.get())); /* Convert this to the content frame */ return DCPTime (s + piece->content->trim_start()).frames (_film->video_frame_rate()) * piece->frc.factor (); } AudioFrame Player::dcp_to_content_audio (shared_ptr piece, DCPTime t) const { /* s is the offset of t from the start position of this content */ DCPTime s = t - piece->content->position (); s = DCPTime (max (int64_t (0), s.get ())); s = DCPTime (min (piece->content->length_after_trim().get(), s.get())); /* Convert this to the content frame */ return DCPTime (s + piece->content->trim_start()).frames (_film->audio_frame_rate()); } ContentTime Player::dcp_to_content_subtitle (shared_ptr piece, DCPTime t) const { /* s is the offset of t from the start position of this content */ DCPTime s = t - piece->content->position (); s = DCPTime (max (int64_t (0), s.get ())); s = DCPTime (min (piece->content->length_after_trim().get(), s.get())); return ContentTime (s + piece->content->trim_start(), piece->frc); } void PlayerStatistics::dump (shared_ptr log) const { log->log (String::compose ("Video: %1 good %2 skipped %3 black %4 repeat", video.good, video.skip, video.black, video.repeat), Log::TYPE_GENERAL); log->log (String::compose ("Audio: %1 good %2 skipped %3 silence", audio.good, audio.skip, audio.silence.seconds()), Log::TYPE_GENERAL); } PlayerStatistics const & Player::statistics () const { return _statistics; }