/* GStreamer * Copyright (C) <2006> Wim Taymans * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-videoanalyse * * This plugin analyses every video frame and if the #GstVideoAnalyse:message * property is #TRUE, posts an element message with video statistics called * "GstVideoAnalyse". * * The message's structure contains these fields: * * * * #GstClockTime * "timestamp": * the timestamp of the buffer that triggered the message. * * * * * #GstClockTime * "stream-time": * the stream time of the buffer. * * * * * #GstClockTime * "running-time": * the running_time of the buffer. * * * * * #GstClockTime * "duration": * the duration of the buffer. * * * * * #gdouble * "brightness": * the average brightness of the frame. * * * * * #gdouble * "brightness-variance": * the brightness variance of the frame. * * * * * * Example launch line * |[ * gst-launch -m videotestsrc ! videoanalyse ! ffmpegcolorspace ! ximagesink * ]| This pipeline emits messages to the console for each frame that has been analysed. * * * Last reviewed on 2007-05-30 (0.10.5) */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstvideoanalyse.h" #include #include #include /* GstVideoAnalyse signals and args */ #define DEFAULT_MESSAGE TRUE enum { PROP_0, PROP_MESSAGE }; GST_DEBUG_CATEGORY_STATIC (video_analyse_debug); #define GST_CAT_DEFAULT video_analyse_debug static const GstElementDetails video_analyse_details = GST_ELEMENT_DETAILS ("Video analyser", "Filter/Effect/Video", "Analyse video signal", "Wim Taymans "); static GstStaticPadTemplate gst_video_analyse_src_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("{ I420, YV12 }")) ); static GstStaticPadTemplate gst_video_analyse_sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("{ I420, YV12 }")) ); static GstVideoFilterClass *parent_class = NULL; static gboolean gst_video_analyse_set_caps (GstBaseTransform * btrans, GstCaps * incaps, GstCaps * outcaps) { GstVideoAnalyse *vf; GstStructure *in_s; gboolean ret; vf = GST_VIDEO_ANALYSE (btrans); in_s = gst_caps_get_structure (incaps, 0); ret = gst_structure_get_int (in_s, "width", &vf->width); ret &= gst_structure_get_int (in_s, "height", &vf->height); return ret; } /* Useful macros */ #define GST_VIDEO_I420_Y_ROWSTRIDE(width) (GST_ROUND_UP_4(width)) #define GST_VIDEO_I420_U_ROWSTRIDE(width) (GST_ROUND_UP_8(width)/2) #define GST_VIDEO_I420_V_ROWSTRIDE(width) ((GST_ROUND_UP_8(GST_VIDEO_I420_Y_ROWSTRIDE(width)))/2) #define GST_VIDEO_I420_Y_OFFSET(w,h) (0) #define GST_VIDEO_I420_U_OFFSET(w,h) (GST_VIDEO_I420_Y_OFFSET(w,h)+(GST_VIDEO_I420_Y_ROWSTRIDE(w)*GST_ROUND_UP_2(h))) #define GST_VIDEO_I420_V_OFFSET(w,h) (GST_VIDEO_I420_U_OFFSET(w,h)+(GST_VIDEO_I420_U_ROWSTRIDE(w)*GST_ROUND_UP_2(h)/2)) #define GST_VIDEO_I420_SIZE(w,h) (GST_VIDEO_I420_V_OFFSET(w,h)+(GST_VIDEO_I420_V_ROWSTRIDE(w)*GST_ROUND_UP_2(h)/2)) static void gst_video_analyse_post_message (GstVideoAnalyse * videoanalyse, GstBuffer * buffer) { GstBaseTransform *trans; GstMessage *m; guint64 duration, timestamp, running_time, stream_time; trans = GST_BASE_TRANSFORM_CAST (videoanalyse); /* get timestamps */ timestamp = GST_BUFFER_TIMESTAMP (buffer); duration = GST_BUFFER_DURATION (buffer); running_time = gst_segment_to_running_time (&trans->segment, GST_FORMAT_TIME, timestamp); stream_time = gst_segment_to_stream_time (&trans->segment, GST_FORMAT_TIME, timestamp); m = gst_message_new_element (GST_OBJECT_CAST (videoanalyse), gst_structure_new ("GstVideoAnalyse", "timestamp", G_TYPE_UINT64, timestamp, "stream-time", G_TYPE_UINT64, stream_time, "running-time", G_TYPE_UINT64, running_time, "duration", G_TYPE_UINT64, duration, "brightness", G_TYPE_DOUBLE, videoanalyse->brightness, "brightness-variance", G_TYPE_DOUBLE, videoanalyse->brightness_var, NULL)); gst_element_post_message (GST_ELEMENT_CAST (videoanalyse), m); } static void gst_video_analyse_420 (GstVideoAnalyse * videoanalyse, guint8 * data, gint width, gint height) { guint64 sum; gint avg, diff; gint i, j; guint8 *d; d = data; sum = 0; /* do brightness as average of pixel brightness in 0.0 to 1.0 */ for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { sum += d[j]; } d += GST_VIDEO_I420_Y_ROWSTRIDE (width); } avg = sum / (width * height); videoanalyse->brightness = sum / (255.0 * width * height); d = data; sum = 0; /* do variance */ for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { diff = (avg - d[j]); sum += diff * diff; } d += GST_VIDEO_I420_Y_ROWSTRIDE (width); } videoanalyse->brightness_var = sum / (255.0 * 255.0 * width * height); } static GstFlowReturn gst_video_analyse_transform_ip (GstBaseTransform * trans, GstBuffer * buf) { GstVideoAnalyse *videoanalyse; GstFlowReturn ret = GST_FLOW_OK; guint8 *data; videoanalyse = GST_VIDEO_ANALYSE (trans); data = GST_BUFFER_DATA (buf); gst_video_analyse_420 (videoanalyse, data, videoanalyse->width, videoanalyse->height); if (videoanalyse->message) gst_video_analyse_post_message (videoanalyse, buf); return ret; } static void gst_video_analyse_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstVideoAnalyse *videoanalyse; videoanalyse = GST_VIDEO_ANALYSE (object); switch (prop_id) { case PROP_MESSAGE: videoanalyse->message = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_video_analyse_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVideoAnalyse *videoanalyse; videoanalyse = GST_VIDEO_ANALYSE (object); switch (prop_id) { case PROP_MESSAGE: g_value_set_boolean (value, videoanalyse->message); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_video_analyse_base_init (gpointer g_class) { GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); gst_element_class_set_details (element_class, &video_analyse_details); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_video_analyse_sink_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_video_analyse_src_template)); } static void gst_video_analyse_class_init (gpointer klass, gpointer class_data) { GObjectClass *gobject_class; GstBaseTransformClass *trans_class; gobject_class = (GObjectClass *) klass; trans_class = (GstBaseTransformClass *) klass; parent_class = g_type_class_peek_parent (klass); gobject_class->set_property = gst_video_analyse_set_property; gobject_class->get_property = gst_video_analyse_get_property; g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_MESSAGE, g_param_spec_boolean ("message", "Message", "Post statics messages", DEFAULT_MESSAGE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT)); trans_class->set_caps = GST_DEBUG_FUNCPTR (gst_video_analyse_set_caps); trans_class->transform_ip = GST_DEBUG_FUNCPTR (gst_video_analyse_transform_ip); trans_class->passthrough_on_same_caps = TRUE; } static void gst_video_analyse_init (GTypeInstance * instance, gpointer g_class) { GstVideoAnalyse *videoanalyse; videoanalyse = GST_VIDEO_ANALYSE (instance); GST_DEBUG_OBJECT (videoanalyse, "gst_video_analyse_init"); } GType gst_video_analyse_get_type (void) { static GType video_analyse_type = 0; if (!video_analyse_type) { static const GTypeInfo video_analyse_info = { sizeof (GstVideoAnalyseClass), gst_video_analyse_base_init, NULL, gst_video_analyse_class_init, NULL, NULL, sizeof (GstVideoAnalyse), 0, gst_video_analyse_init, }; video_analyse_type = g_type_register_static (GST_TYPE_VIDEO_FILTER, "GstVideoAnalyse", &video_analyse_info, 0); GST_DEBUG_CATEGORY_INIT (video_analyse_debug, "videoanalyse", 0, "Video Analyse element"); } return video_analyse_type; }