diff options
Diffstat (limited to 'gst/videosignal/gstvideoanalyse.c')
-rw-r--r-- | gst/videosignal/gstvideoanalyse.c | 353 |
1 files changed, 353 insertions, 0 deletions
diff --git a/gst/videosignal/gstvideoanalyse.c b/gst/videosignal/gstvideoanalyse.c new file mode 100644 index 00000000..d0cdf70c --- /dev/null +++ b/gst/videosignal/gstvideoanalyse.c @@ -0,0 +1,353 @@ +/* GStreamer + * Copyright (C) <2006> Wim Taymans <wim@fluendo.com> + * + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Library General Public + * License as published by the Free Software Foundation; either + * version 2 of the License, or (at your option) any later version. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Library General Public License for more details. + * + * You should have received a copy of the GNU Library General Public + * License along with this library; if not, write to the + * Free Software Foundation, Inc., 59 Temple Place - Suite 330, + * Boston, MA 02111-1307, USA. + */ + +/** + * SECTION:element-videoanalyse + * @short_description: Analyse and report about the video frame + * + * <refsect2> + * <para> + * This plugin analyses every video frame and if the <link + * linkend="GstVideoAnalyse--message">message property</link> is #TRUE, posts an element + * message with video statistics called <classname>"GstVideoAnalyse"</classname>. + * The message's structure contains these fields: + * <itemizedlist> + * <listitem> + * <para> + * #GstClockTime + * <classname>"timestamp"</classname>: + * the timestamp of the buffer that triggered the message. + * </para> + * </listitem> + * <listitem> + * <para> + * #GstClockTime + * <classname>"stream-time"</classname>: + * the stream time of the buffer. + * </para> + * </listitem> + * <listitem> + * <para> + * #GstClockTime + * <classname>"running-time"</classname>: + * the running_time of the buffer. + * </para> + * </listitem> + * <listitem> + * <para> + * #GstClockTime + * <classname>"duration"</classname>: + * the duration of the buffer. + * </para> + * </listitem> + * <listitem> + * <para> + * #gdouble + * <classname>"brightness"</classname>: + * the average brightness of the frame. + * </para> + * </listitem> + * <listitem> + * <para> + * #gdouble + * <classname>"brightness-variance"</classname>: + * the brightness variance of the frame. + * </para> + * </listitem> + * </itemizedlist> + * </para> + * <title>Example launch line</title> + * <para> + * <programlisting> + * gst-launch -m videotestsrc ! videoanalyse ! ffmpegcolorspace ! ximagesink + * </programlisting> + * This pipeline emits messages to the console for each frame that has been analysed. + * </para> + * </refsect2> + * + * Last reviewed on 2007-05-30 (0.10.5) + */ + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include "gstvideoanalyse.h" + +#include <string.h> +#include <math.h> + +#include <gst/video/video.h> + +/* GstVideoAnalyse signals and args */ + +#define DEFAULT_MESSAGE TRUE + +enum +{ + PROP_0, + PROP_MESSAGE +}; + +GST_DEBUG_CATEGORY_STATIC (video_analyse_debug); +#define GST_CAT_DEFAULT video_analyse_debug + +static const GstElementDetails video_analyse_details = +GST_ELEMENT_DETAILS ("Video analyser", + "Filter/Effect/Video", + "Analyse video signal", + "Wim Taymans <wim@fluendo.com>"); + +static GstStaticPadTemplate gst_video_analyse_src_template = +GST_STATIC_PAD_TEMPLATE ("src", + GST_PAD_SRC, + GST_PAD_ALWAYS, + GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("{ I420, YV12 }")) + ); + +static GstStaticPadTemplate gst_video_analyse_sink_template = +GST_STATIC_PAD_TEMPLATE ("sink", + GST_PAD_SINK, + GST_PAD_ALWAYS, + GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("{ I420, YV12 }")) + ); + +static GstVideoFilterClass *parent_class = NULL; + +static gboolean +gst_video_analyse_set_caps (GstBaseTransform * btrans, GstCaps * incaps, + GstCaps * outcaps) +{ + GstVideoAnalyse *vf; + GstStructure *in_s; + gboolean ret; + + vf = GST_VIDEO_ANALYSE (btrans); + + in_s = gst_caps_get_structure (incaps, 0); + + ret = gst_structure_get_int (in_s, "width", &vf->width); + ret &= gst_structure_get_int (in_s, "height", &vf->height); + + return ret; +} + +/* Useful macros */ +#define GST_VIDEO_I420_Y_ROWSTRIDE(width) (GST_ROUND_UP_4(width)) +#define GST_VIDEO_I420_U_ROWSTRIDE(width) (GST_ROUND_UP_8(width)/2) +#define GST_VIDEO_I420_V_ROWSTRIDE(width) ((GST_ROUND_UP_8(GST_VIDEO_I420_Y_ROWSTRIDE(width)))/2) + +#define GST_VIDEO_I420_Y_OFFSET(w,h) (0) +#define GST_VIDEO_I420_U_OFFSET(w,h) (GST_VIDEO_I420_Y_OFFSET(w,h)+(GST_VIDEO_I420_Y_ROWSTRIDE(w)*GST_ROUND_UP_2(h))) +#define GST_VIDEO_I420_V_OFFSET(w,h) (GST_VIDEO_I420_U_OFFSET(w,h)+(GST_VIDEO_I420_U_ROWSTRIDE(w)*GST_ROUND_UP_2(h)/2)) + +#define GST_VIDEO_I420_SIZE(w,h) (GST_VIDEO_I420_V_OFFSET(w,h)+(GST_VIDEO_I420_V_ROWSTRIDE(w)*GST_ROUND_UP_2(h)/2)) + +static void +gst_video_analyse_post_message (GstVideoAnalyse * videoanalyse, + GstBuffer * buffer) +{ + GstBaseTransform *trans; + GstMessage *m; + guint64 duration, timestamp, running_time, stream_time; + + trans = GST_BASE_TRANSFORM_CAST (videoanalyse); + + /* get timestamps */ + timestamp = GST_BUFFER_TIMESTAMP (buffer); + duration = GST_BUFFER_DURATION (buffer); + running_time = gst_segment_to_running_time (&trans->segment, GST_FORMAT_TIME, + timestamp); + stream_time = gst_segment_to_stream_time (&trans->segment, GST_FORMAT_TIME, + timestamp); + + m = gst_message_new_element (GST_OBJECT_CAST (videoanalyse), + gst_structure_new ("GstVideoAnalyse", + "timestamp", G_TYPE_UINT64, timestamp, + "stream-time", G_TYPE_UINT64, stream_time, + "running-time", G_TYPE_UINT64, running_time, + "duration", G_TYPE_UINT64, duration, + "brightness", G_TYPE_DOUBLE, videoanalyse->brightness, + "brightness-variance", G_TYPE_DOUBLE, videoanalyse->brightness_var, + NULL)); + + gst_element_post_message (GST_ELEMENT_CAST (videoanalyse), m); +} + +static void +gst_video_analyse_420 (GstVideoAnalyse * videoanalyse, guint8 * data, + gint width, gint height) +{ + guint64 sum; + gint avg, diff; + gint i, j; + guint8 *d; + + d = data; + sum = 0; + /* do brightness as average of pixel brightness in 0.0 to 1.0 */ + for (i = 0; i < height; i++) { + for (j = 0; j < width; j++) { + sum += data[j]; + } + d += GST_VIDEO_I420_Y_ROWSTRIDE (width); + } + avg = sum / (width * height); + videoanalyse->brightness = sum / (255.0 * width * height); + + d = data; + sum = 0; + /* do variance */ + for (i = 0; i < height; i++) { + for (j = 0; j < width; j++) { + diff = (avg - d[j]); + sum += diff * diff; + } + d += GST_VIDEO_I420_Y_ROWSTRIDE (width); + } + videoanalyse->brightness_var = sum / (255.0 * 255.0 * width * height); +} + +static GstFlowReturn +gst_video_analyse_transform_ip (GstBaseTransform * trans, GstBuffer * buf) +{ + GstVideoAnalyse *videoanalyse; + GstFlowReturn ret = GST_FLOW_OK; + guint8 *data; + + videoanalyse = GST_VIDEO_ANALYSE (trans); + + data = GST_BUFFER_DATA (buf); + + gst_video_analyse_420 (videoanalyse, data, videoanalyse->width, + videoanalyse->height); + + if (videoanalyse->message) + gst_video_analyse_post_message (videoanalyse, buf); + + return ret; +} + +static void +gst_video_analyse_set_property (GObject * object, guint prop_id, + const GValue * value, GParamSpec * pspec) +{ + GstVideoAnalyse *videoanalyse; + + videoanalyse = GST_VIDEO_ANALYSE (object); + + switch (prop_id) { + case PROP_MESSAGE: + videoanalyse->message = g_value_get_boolean (value); + break; + default: + G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); + break; + } +} + +static void +gst_video_analyse_get_property (GObject * object, guint prop_id, GValue * value, + GParamSpec * pspec) +{ + GstVideoAnalyse *videoanalyse; + + videoanalyse = GST_VIDEO_ANALYSE (object); + + switch (prop_id) { + case PROP_MESSAGE: + g_value_set_boolean (value, videoanalyse->message); + break; + default: + G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); + break; + } +} + +static void +gst_video_analyse_base_init (gpointer g_class) +{ + GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); + + gst_element_class_set_details (element_class, &video_analyse_details); + + gst_element_class_add_pad_template (element_class, + gst_static_pad_template_get (&gst_video_analyse_sink_template)); + gst_element_class_add_pad_template (element_class, + gst_static_pad_template_get (&gst_video_analyse_src_template)); +} + +static void +gst_video_analyse_class_init (gpointer klass, gpointer class_data) +{ + GObjectClass *gobject_class; + GstBaseTransformClass *trans_class; + + gobject_class = (GObjectClass *) klass; + trans_class = (GstBaseTransformClass *) klass; + + parent_class = g_type_class_peek_parent (klass); + + gobject_class->set_property = gst_video_analyse_set_property; + gobject_class->get_property = gst_video_analyse_get_property; + + g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_MESSAGE, + g_param_spec_boolean ("message", "Message", + "Post statics messages", + DEFAULT_MESSAGE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT)); + + trans_class->set_caps = GST_DEBUG_FUNCPTR (gst_video_analyse_set_caps); + trans_class->transform_ip = + GST_DEBUG_FUNCPTR (gst_video_analyse_transform_ip); + trans_class->passthrough_on_same_caps = TRUE; +} + +static void +gst_video_analyse_init (GTypeInstance * instance, gpointer g_class) +{ + GstVideoAnalyse *videoanalyse; + + videoanalyse = GST_VIDEO_ANALYSE (instance); + + GST_DEBUG_OBJECT (videoanalyse, "gst_video_analyse_init"); +} + +GType +gst_video_analyse_get_type (void) +{ + static GType video_analyse_type = 0; + + if (!video_analyse_type) { + static const GTypeInfo video_analyse_info = { + sizeof (GstVideoAnalyseClass), + gst_video_analyse_base_init, + NULL, + gst_video_analyse_class_init, + NULL, + NULL, + sizeof (GstVideoAnalyse), + 0, + gst_video_analyse_init, + }; + + video_analyse_type = g_type_register_static (GST_TYPE_VIDEO_FILTER, + "GstVideoAnalyse", &video_analyse_info, 0); + } + return video_analyse_type; +} |