* @brief A graph of FFmpeg filters.
*/
+#include "filter_graph.h"
+#include "filter.h"
+#include "exceptions.h"
+#include "image.h"
+#include "ffmpeg_content.h"
+#include "safe_stringstream.h"
+#include "compose.hpp"
extern "C" {
#include <libavfilter/avfiltergraph.h>
#include <libavfilter/buffersrc.h>
#include <libavfilter/buffersink.h>
#include <libavformat/avio.h>
}
-#include "decoder.h"
-#include "filter_graph.h"
-#include "filter.h"
-#include "exceptions.h"
-#include "image.h"
-#include "film.h"
-#include "ffmpeg_decoder.h"
+#include <iostream>
#include "i18n.h"
-using std::stringstream;
using std::string;
using std::list;
+using std::pair;
+using std::make_pair;
+using std::cout;
using boost::shared_ptr;
-using libdcp::Size;
+using boost::weak_ptr;
+using dcp::Size;
-/** Construct a FilterGraph for the settings in a film.
- * @param film Film.
- * @param decoder Decoder that we are using.
+/** Construct a FilterGraph for the settings in a piece of content.
+ * @param content Content.
* @param s Size of the images to process.
* @param p Pixel format of the images to process.
*/
-FilterGraph::FilterGraph (shared_ptr<Film> film, FFmpegDecoder* decoder, libdcp::Size s, AVPixelFormat p)
- : _buffer_src_context (0)
+FilterGraph::FilterGraph (shared_ptr<const FFmpegContent> content, dcp::Size s, AVPixelFormat p)
+ : _copy (false)
+ , _buffer_src_context (0)
, _buffer_sink_context (0)
, _size (s)
, _pixel_format (p)
+ , _frame (0)
{
- string filters = Filter::ffmpeg_strings (film->filters()).first;
- if (!filters.empty ()) {
- filters += N_(",");
+ string const filters = Filter::ffmpeg_string (content->filters());
+ if (filters.empty ()) {
+ _copy = true;
+ return;
}
- filters += crop_string (Position (film->crop().left, film->crop().top), film->cropped_size (decoder->native_size()));
+ _frame = av_frame_alloc ();
AVFilterGraph* graph = avfilter_graph_alloc();
if (graph == 0) {
throw DecodeError (N_("Could not create buffer sink filter"));
}
- stringstream a;
- a << _size.width << N_(":")
- << _size.height << N_(":")
- << _pixel_format << N_(":")
- << decoder->time_base_numerator() << N_(":")
- << decoder->time_base_denominator() << N_(":")
- << decoder->sample_aspect_ratio_numerator() << N_(":")
- << decoder->sample_aspect_ratio_denominator();
+ SafeStringStream a;
+ a << "video_size=" << _size.width << "x" << _size.height << ":"
+ << "pix_fmt=" << _pixel_format << ":"
+ << "time_base=1/1:"
+ << "pixel_aspect=1/1";
- int r;
-
- if ((r = avfilter_graph_create_filter (&_buffer_src_context, buffer_src, N_("in"), a.str().c_str(), 0, graph)) < 0) {
+ if (avfilter_graph_create_filter (&_buffer_src_context, buffer_src, "in", a.str().c_str(), 0, graph) < 0) {
throw DecodeError (N_("could not create buffer source"));
}
pixel_fmts[0] = _pixel_format;
pixel_fmts[1] = PIX_FMT_NONE;
sink_params->pixel_fmts = pixel_fmts;
-
+
if (avfilter_graph_create_filter (&_buffer_sink_context, buffer_sink, N_("out"), 0, sink_params, graph) < 0) {
throw DecodeError (N_("could not create buffer sink."));
}
+ av_free (sink_params);
+
AVFilterInOut* outputs = avfilter_inout_alloc ();
outputs->name = av_strdup(N_("in"));
outputs->filter_ctx = _buffer_src_context;
inputs->pad_idx = 0;
inputs->next = 0;
- if (avfilter_graph_parse (graph, filters.c_str(), &inputs, &outputs, 0) < 0) {
+ if (avfilter_graph_parse (graph, filters.c_str(), inputs, outputs, 0) < 0) {
throw DecodeError (N_("could not set up filter graph."));
}
-
+
if (avfilter_graph_config (graph, 0) < 0) {
throw DecodeError (N_("could not configure filter graph."));
}
+}
- /* XXX: leaking `inputs' / `outputs' ? */
+FilterGraph::~FilterGraph ()
+{
+ if (_frame) {
+ av_frame_free (&_frame);
+ }
}
/** Take an AVFrame and process it using our configured filters, returning a
- * set of Images.
+ * set of Images. Caller handles memory management of the input frame.
*/
-list<shared_ptr<Image> >
+list<pair<shared_ptr<Image>, int64_t> >
FilterGraph::process (AVFrame* frame)
{
- list<shared_ptr<Image> > images;
-
- if (av_buffersrc_add_frame (_buffer_src_context, frame) < 0) {
- throw DecodeError (N_("could not push buffer into filter chain."));
- }
-
- while (1) {
- AVFrame* frame = av_frame_alloc ();
- if (av_buffersink_get_frame (_buffer_sink_context, frame) < 0) {
- av_frame_free (&frame);
- break;
+ list<pair<shared_ptr<Image>, int64_t> > images;
+
+ if (_copy) {
+ images.push_back (make_pair (shared_ptr<Image> (new Image (frame)), av_frame_get_best_effort_timestamp (frame)));
+ } else {
+ int r = av_buffersrc_write_frame (_buffer_src_context, frame);
+ if (r < 0) {
+ throw DecodeError (String::compose (N_("could not push buffer into filter chain (%1)."), r));
}
- /* This takes ownership of the AVFrame */
- images.push_back (shared_ptr<Image> (new FrameImage (frame)));
+ while (true) {
+ if (av_buffersink_get_frame (_buffer_sink_context, _frame) < 0) {
+ break;
+ }
+
+ images.push_back (make_pair (shared_ptr<Image> (new Image (_frame)), av_frame_get_best_effort_timestamp (_frame)));
+ av_frame_unref (_frame);
+ }
}
-
+
return images;
}
* @return true if this chain can process images with `s' and `p', otherwise false.
*/
bool
-FilterGraph::can_process (libdcp::Size s, AVPixelFormat p) const
+FilterGraph::can_process (dcp::Size s, AVPixelFormat p) const
{
return (_size == s && _pixel_format == p);
}