/* GStreamer * Copyright (C) 2006 David A. Schleef * Copyright (C) 2007 Sebastian Dröge * * gstvideoparse.c: * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-videoparse * @short_description: parses a byte stream into video frames * * Converts a byte stream into video frames. */ #ifdef HAVE_CONFIG_H # include "config.h" #endif #include #include #include #include #define GST_TYPE_VIDEO_PARSE \ (gst_video_parse_get_type()) #define GST_VIDEO_PARSE(obj) \ (G_TYPE_CHECK_INSTANCE_CAST((obj),GST_TYPE_VIDEO_PARSE,GstVideoParse)) #define GST_VIDEO_PARSE_CLASS(klass) \ (G_TYPE_CHECK_CLASS_CAST((klass),GST_TYPE_VIDEO_PARSE,GstVideoParseClass)) #define GST_IS_VIDEO_PARSE(obj) \ (G_TYPE_CHECK_INSTANCE_TYPE((obj),GST_TYPE_VIDEO_PARSE)) #define GST_IS_VIDEO_PARSE_CLASS(klass) \ (G_TYPE_CHECK_CLASS_TYPE((klass),GST_TYPE_VIDEO_PARSE)) typedef struct _GstVideoParse GstVideoParse; typedef struct _GstVideoParseClass GstVideoParseClass; struct _GstVideoParse { GstElement parent; /* properties */ int width; int height; guint32 format; int fps_n; int fps_d; int par_n; int par_d; int bpp; int depth; int endianness; int red_mask; int blue_mask; int green_mask; int alpha_mask; /* private */ GstPad *sinkpad; GstPad *srcpad; GstAdapter *adapter; int blocksize; gboolean discont; int n_frames; GstSegment segment; gboolean negotiated; gboolean have_new_segment; }; struct _GstVideoParseClass { GstElementClass parent_class; }; typedef enum { GST_VIDEO_PARSE_FORMAT_I420, GST_VIDEO_PARSE_FORMAT_YV12, GST_VIDEO_PARSE_FORMAT_YUY2, GST_VIDEO_PARSE_FORMAT_UYVY, GST_VIDEO_PARSE_FORMAT_RGB = 10, GST_VIDEO_PARSE_FORMAT_GRAY } GstVideoParseFormat; typedef enum { GST_VIDEO_PARSE_ENDIANNESS_LITTLE = 1234, GST_VIDEO_PARSE_ENDIANNESS_BIG = 4321 } GstVideoParseEndianness; static void gst_video_parse_dispose (GObject * object); static void gst_video_parse_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_video_parse_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static GstFlowReturn gst_video_parse_chain (GstPad * pad, GstBuffer * buffer); static gboolean gst_video_parse_sink_event (GstPad * pad, GstEvent * event); static gboolean gst_video_parse_src_event (GstPad * pad, GstEvent * event); static const GstQueryType *gst_video_parse_src_query_type (GstPad * pad); static gboolean gst_video_parse_src_query (GstPad * pad, GstQuery * query); static gboolean gst_video_parse_convert (GstVideoParse * vp, GstFormat src_format, gint64 src_value, GstFormat dest_format, gint64 * dest_value); static void gst_video_parse_update_block_size (GstVideoParse * vp); static GstStaticPadTemplate gst_video_parse_src_pad_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("{ I420, YV12, YUY2, UYVY }") ";" "video/x-raw-rgb;" "video/x-raw-gray")); static GstStaticPadTemplate gst_video_parse_sink_pad_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS_ANY); GST_DEBUG_CATEGORY_STATIC (gst_video_parse_debug); #define GST_CAT_DEFAULT gst_video_parse_debug static const GstElementDetails gst_video_parse_details = GST_ELEMENT_DETAILS ("Video Parse", "Filter/Video", "Converts stream into video frames", "David Schleef "); enum { ARG_0, ARG_WIDTH, ARG_HEIGHT, ARG_FORMAT, ARG_PAR, ARG_FRAMERATE, ARG_BPP, ARG_DEPTH, ARG_ENDIANNESS, ARG_RED_MASK, ARG_GREEN_MASK, ARG_BLUE_MASK, ARG_ALPHA_MASK }; #define GST_VIDEO_PARSE_FORMAT (gst_video_parse_format_get_type ()) static GType gst_video_parse_format_get_type (void) { static GType video_parse_format_type = 0; static const GEnumValue format_types[] = { {GST_VIDEO_PARSE_FORMAT_I420, "I420", "I420"}, {GST_VIDEO_PARSE_FORMAT_YV12, "YV12", "YV12"}, {GST_VIDEO_PARSE_FORMAT_YUY2, "YUY2", "YUY2"}, {GST_VIDEO_PARSE_FORMAT_UYVY, "UYVY", "UYVY"}, {GST_VIDEO_PARSE_FORMAT_RGB, "RGB", "RGB"}, {GST_VIDEO_PARSE_FORMAT_GRAY, "GRAY", "GRAY"}, {0, NULL, NULL} }; if (!video_parse_format_type) { video_parse_format_type = g_enum_register_static ("GstVideoParseFormat", format_types); } return video_parse_format_type; } #define GST_VIDEO_PARSE_ENDIANNESS (gst_video_parse_endianness_get_type ()) static GType gst_video_parse_endianness_get_type (void) { static GType video_parse_endianness_type = 0; static const GEnumValue endian_types[] = { {GST_VIDEO_PARSE_ENDIANNESS_LITTLE, "Little Endian", "little"}, {GST_VIDEO_PARSE_ENDIANNESS_BIG, "Big Endian", "big"}, {0, NULL, NULL} }; if (!video_parse_endianness_type) { video_parse_endianness_type = g_enum_register_static ("GstVideoParseEndianness", endian_types); } return video_parse_endianness_type; } GST_BOILERPLATE (GstVideoParse, gst_video_parse, GstElement, GST_TYPE_ELEMENT); static void gst_video_parse_base_init (gpointer g_class) { GstElementClass *gstelement_class = GST_ELEMENT_CLASS (g_class); GST_DEBUG_CATEGORY_INIT (gst_video_parse_debug, "videoparse", 0, "videoparse element"); gst_element_class_add_pad_template (gstelement_class, gst_static_pad_template_get (&gst_video_parse_src_pad_template)); gst_element_class_add_pad_template (gstelement_class, gst_static_pad_template_get (&gst_video_parse_sink_pad_template)); gst_element_class_set_details (gstelement_class, &gst_video_parse_details); } static void gst_video_parse_class_init (GstVideoParseClass * klass) { GObjectClass *gobject_class = G_OBJECT_CLASS (klass); //GstElementClass *gstelement_class = GST_ELEMENT_CLASS (klass); gobject_class->set_property = gst_video_parse_set_property; gobject_class->get_property = gst_video_parse_get_property; gobject_class->dispose = gst_video_parse_dispose; g_object_class_install_property (gobject_class, ARG_WIDTH, g_param_spec_int ("width", "Width", "Width of images in raw stream", 0, INT_MAX, 320, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_HEIGHT, g_param_spec_int ("height", "Height", "Height of images in raw stream", 0, INT_MAX, 240, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_FORMAT, g_param_spec_enum ("format", "Format", "Format of images in raw stream", GST_VIDEO_PARSE_FORMAT, GST_VIDEO_PARSE_FORMAT_I420, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_FRAMERATE, gst_param_spec_fraction ("framerate", "Frame Rate", "Frame rate of images in raw stream", 0, 1, 100, 1, 25, 1, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_PAR, gst_param_spec_fraction ("pixel_aspect_ratio", "Pixel Aspect Ratio", "Pixel aspect ratio of images in raw stream", 1, 100, 100, 1, 1, 1, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_BPP, g_param_spec_int ("bpp", "Bpp", "Bits per pixel of images in raw stream", 0, INT_MAX, 24, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_DEPTH, g_param_spec_int ("depth", "Depth", "Depth of images in raw stream", 0, INT_MAX, 24, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_ENDIANNESS, g_param_spec_enum ("endianness", "Endianness", "Endianness of images in raw stream", GST_VIDEO_PARSE_ENDIANNESS, G_BYTE_ORDER, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_RED_MASK, g_param_spec_int ("red-mask", "Red mask", "Red mask of images in raw stream", 0, INT_MAX, GST_VIDEO_BYTE1_MASK_24_INT, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_GREEN_MASK, g_param_spec_int ("green-mask", "Green mask", "Green mask of images in raw stream", 0, INT_MAX, GST_VIDEO_BYTE2_MASK_24_INT, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_BLUE_MASK, g_param_spec_int ("blue-mask", "Blue mask", "Blue mask of images in raw stream", 0, INT_MAX, GST_VIDEO_BYTE3_MASK_24_INT, G_PARAM_READWRITE)); g_object_class_install_property (gobject_class, ARG_ALPHA_MASK, g_param_spec_int ("alpha-mask", "Alpha mask", "Alpha mask of images in raw stream", 0, INT_MAX, 0, G_PARAM_READWRITE)); } static void gst_video_parse_init (GstVideoParse * vp, GstVideoParseClass * g_class) { vp->sinkpad = gst_pad_new_from_static_template (&gst_video_parse_sink_pad_template, "sink"); gst_element_add_pad (GST_ELEMENT (vp), vp->sinkpad); gst_pad_set_chain_function (vp->sinkpad, gst_video_parse_chain); gst_pad_set_event_function (vp->sinkpad, gst_video_parse_sink_event); vp->srcpad = gst_pad_new_from_static_template (&gst_video_parse_src_pad_template, "src"); gst_element_add_pad (GST_ELEMENT (vp), vp->srcpad); gst_pad_set_event_function (vp->srcpad, gst_video_parse_src_event); if (1) { gst_pad_set_query_type_function (vp->srcpad, gst_video_parse_src_query_type); gst_pad_set_query_function (vp->srcpad, gst_video_parse_src_query); } vp->adapter = gst_adapter_new (); vp->width = 320; vp->height = 240; vp->format = GST_VIDEO_PARSE_FORMAT_I420; vp->par_n = 1; vp->par_d = 1; vp->fps_n = 25; vp->fps_d = 1; vp->bpp = 24; vp->depth = 24; vp->endianness = G_BYTE_ORDER; vp->red_mask = GST_VIDEO_BYTE1_MASK_24_INT; vp->green_mask = GST_VIDEO_BYTE2_MASK_24_INT; vp->blue_mask = GST_VIDEO_BYTE3_MASK_24_INT; vp->alpha_mask = 0; gst_video_parse_update_block_size (vp); } static void gst_video_parse_dispose (GObject * object) { GstVideoParse *vp = GST_VIDEO_PARSE (object); if (vp->adapter) { g_object_unref (vp->adapter); vp->adapter = NULL; } G_OBJECT_CLASS (parent_class)->dispose (object); } static void gst_video_parse_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstVideoParse *vp = GST_VIDEO_PARSE (object); g_return_if_fail (!vp->negotiated); switch (prop_id) { case ARG_WIDTH: vp->width = g_value_get_int (value); break; case ARG_HEIGHT: vp->height = g_value_get_int (value); break; case ARG_FORMAT: vp->format = g_value_get_enum (value); break; case ARG_FRAMERATE: vp->fps_n = gst_value_get_fraction_numerator (value); vp->fps_d = gst_value_get_fraction_denominator (value); break; case ARG_PAR: vp->par_n = gst_value_get_fraction_numerator (value); vp->par_d = gst_value_get_fraction_denominator (value); break; case ARG_BPP: vp->bpp = g_value_get_int (value); break; case ARG_DEPTH: vp->depth = g_value_get_int (value); break; case ARG_ENDIANNESS: vp->endianness = g_value_get_enum (value); break; case ARG_RED_MASK: vp->red_mask = g_value_get_int (value); break; case ARG_GREEN_MASK: vp->green_mask = g_value_get_int (value); break; case ARG_BLUE_MASK: vp->blue_mask = g_value_get_int (value); break; case ARG_ALPHA_MASK: vp->alpha_mask = g_value_get_int (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } gst_video_parse_update_block_size (vp); } static void gst_video_parse_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVideoParse *vp = GST_VIDEO_PARSE (object); switch (prop_id) { case ARG_WIDTH: g_value_set_int (value, vp->width); break; case ARG_HEIGHT: g_value_set_int (value, vp->height); break; case ARG_FORMAT: g_value_set_enum (value, vp->format); break; case ARG_FRAMERATE: gst_value_set_fraction (value, vp->fps_n, vp->fps_d); break; case ARG_PAR: gst_value_set_fraction (value, vp->par_n, vp->par_d); break; case ARG_BPP: g_value_set_int (value, vp->bpp); break; case ARG_DEPTH: g_value_set_int (value, vp->depth); break; case ARG_ENDIANNESS: g_value_set_enum (value, vp->endianness); break; case ARG_RED_MASK: g_value_set_int (value, vp->red_mask); break; case ARG_GREEN_MASK: g_value_set_int (value, vp->green_mask); break; case ARG_BLUE_MASK: g_value_set_int (value, vp->blue_mask); break; case ARG_ALPHA_MASK: g_value_set_int (value, vp->alpha_mask); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static guint32 gst_video_parse_format_to_fourcc (GstVideoParseFormat format) { switch (format) { case GST_VIDEO_PARSE_FORMAT_I420: return GST_MAKE_FOURCC ('I', '4', '2', '0'); case GST_VIDEO_PARSE_FORMAT_YV12: return GST_MAKE_FOURCC ('Y', 'V', '1', '2'); case GST_VIDEO_PARSE_FORMAT_YUY2: return GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'); case GST_VIDEO_PARSE_FORMAT_UYVY: return GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'); default: g_assert_not_reached (); } return 0; } void gst_video_parse_update_block_size (GstVideoParse * vp) { if (vp->format == GST_VIDEO_PARSE_FORMAT_I420 || vp->format == GST_VIDEO_PARSE_FORMAT_YV12) { vp->blocksize = GST_ROUND_UP_4 (vp->width) * GST_ROUND_UP_2 (vp->height) + 2 * (GST_ROUND_UP_8 (vp->width) / 2) * (GST_ROUND_UP_2 (vp->height) / 2); } else if (vp->format == GST_VIDEO_PARSE_FORMAT_YUY2 || vp->format == GST_VIDEO_PARSE_FORMAT_UYVY) { vp->blocksize = GST_ROUND_UP_4 (vp->width * 2) * vp->height; } else if (vp->format == GST_VIDEO_PARSE_FORMAT_RGB) { vp->blocksize = GST_ROUND_UP_4 (vp->width * vp->bpp) * vp->height; } else { vp->blocksize = GST_ROUND_UP_4 (vp->width * vp->bpp) * vp->height; } } static void gst_video_parse_reset (GstVideoParse * vp) { vp->n_frames = 0; vp->discont = TRUE; gst_segment_init (&vp->segment, GST_FORMAT_TIME); gst_adapter_clear (vp->adapter); } static GstFlowReturn gst_video_parse_chain (GstPad * pad, GstBuffer * buffer) { GstVideoParse *vp = GST_VIDEO_PARSE (gst_pad_get_parent (pad)); GstFlowReturn ret = GST_FLOW_OK; if (G_UNLIKELY (GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_DISCONT))) { GST_DEBUG_OBJECT (vp, "received DISCONT buffer"); vp->discont = TRUE; } if (!vp->negotiated) { GstCaps *caps; g_print ("%d\n", vp->format); if (vp->format < GST_VIDEO_PARSE_FORMAT_RGB) { caps = gst_caps_new_simple ("video/x-raw-yuv", "width", G_TYPE_INT, vp->width, "height", G_TYPE_INT, vp->height, "format", GST_TYPE_FOURCC, gst_video_parse_format_to_fourcc (vp->format), "framerate", GST_TYPE_FRACTION, vp->fps_n, vp->fps_d, "pixel_aspect_ratio", GST_TYPE_FRACTION, vp->par_n, vp->par_d, NULL); } else if (vp->format == GST_VIDEO_PARSE_FORMAT_RGB) { caps = gst_caps_new_simple ("video/x-raw-rgb", "width", G_TYPE_INT, vp->width, "height", G_TYPE_INT, vp->height, "bpp", G_TYPE_INT, vp->bpp, "depth", G_TYPE_INT, vp->depth, "framerate", GST_TYPE_FRACTION, vp->fps_n, vp->fps_d, "pixel_aspect_ratio", GST_TYPE_FRACTION, vp->par_n, vp->par_d, "red_mask", G_TYPE_INT, vp->red_mask, "green_mask", G_TYPE_INT, vp->green_mask, "blue_mask", G_TYPE_INT, vp->blue_mask, "alpha_mask", G_TYPE_INT, vp->alpha_mask, "endianness", G_TYPE_INT, vp->endianness, NULL); } else { caps = gst_caps_new_simple ("video/x-raw-gray", "width", G_TYPE_INT, vp->width, "height", G_TYPE_INT, vp->height, "bpp", G_TYPE_INT, vp->bpp, "depth", G_TYPE_INT, vp->depth, "framerate", GST_TYPE_FRACTION, vp->fps_n, vp->fps_d, "pixel_aspect_ratio", GST_TYPE_FRACTION, vp->par_n, vp->par_d, NULL); } gst_pad_set_caps (vp->srcpad, caps); vp->negotiated = TRUE; } gst_adapter_push (vp->adapter, buffer); while (gst_adapter_available (vp->adapter) >= vp->blocksize) { buffer = gst_adapter_take_buffer (vp->adapter, vp->blocksize); if (vp->fps_n) { GST_BUFFER_TIMESTAMP (buffer) = vp->segment.start + gst_util_uint64_scale (vp->n_frames, GST_SECOND * vp->fps_d, vp->fps_n); GST_BUFFER_DURATION (buffer) = gst_util_uint64_scale (GST_SECOND, vp->fps_d, vp->fps_n); } else { GST_BUFFER_TIMESTAMP (buffer) = vp->segment.start; GST_BUFFER_DURATION (buffer) = GST_CLOCK_TIME_NONE; } gst_buffer_set_caps (buffer, GST_PAD_CAPS (vp->srcpad)); if (vp->discont) { GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_DISCONT); vp->discont = FALSE; } vp->n_frames++; ret = gst_pad_push (vp->srcpad, buffer); if (ret != GST_FLOW_OK) break; } gst_object_unref (vp); return ret; } static gboolean gst_video_parse_convert (GstVideoParse * vp, GstFormat src_format, gint64 src_value, GstFormat dest_format, gint64 * dest_value) { gboolean ret = FALSE; GST_DEBUG ("converting value %" G_GINT64_FORMAT " from %s to %s", src_value, gst_format_get_name (src_format), gst_format_get_name (dest_format)); if (src_format == dest_format) { *dest_value = src_value; ret = TRUE; goto done; } if (src_value == -1) { *dest_value = -1; ret = TRUE; goto done; } /* bytes to frames */ if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_DEFAULT) { if (vp->blocksize != 0) { *dest_value = gst_util_uint64_scale_int (src_value, 1, vp->blocksize); } else { GST_ERROR ("blocksize is 0"); *dest_value = 0; } ret = TRUE; goto done; } /* frames to bytes */ if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_BYTES) { *dest_value = gst_util_uint64_scale_int (src_value, vp->blocksize, 1); ret = TRUE; goto done; } /* time to frames */ if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_DEFAULT) { if (vp->fps_d != 0) { *dest_value = gst_util_uint64_scale (src_value, vp->fps_n, GST_SECOND * vp->fps_d); } else { GST_ERROR ("framerate denominator is 0"); *dest_value = 0; } ret = TRUE; goto done; } /* frames to time */ if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_TIME) { if (vp->fps_n != 0) { *dest_value = gst_util_uint64_scale (src_value, GST_SECOND * vp->fps_d, vp->fps_n); } else { GST_ERROR ("framerate numerator is 0"); *dest_value = 0; } ret = TRUE; goto done; } /* time to bytes */ if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_BYTES) { if (vp->fps_d != 0) { *dest_value = gst_util_uint64_scale (src_value, vp->fps_n * vp->blocksize, GST_SECOND * vp->fps_d); } else { GST_ERROR ("framerate denominator is 0"); *dest_value = 0; } ret = TRUE; goto done; } /* bytes to time */ if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_TIME) { if (vp->fps_n != 0 && vp->blocksize != 0) { *dest_value = gst_util_uint64_scale (src_value, GST_SECOND * vp->fps_d, vp->fps_n * vp->blocksize); } else { GST_ERROR ("framerate denominator and/or blocksize is 0"); *dest_value = 0; } ret = TRUE; } done: GST_DEBUG ("ret=%d result %" G_GINT64_FORMAT, ret, *dest_value); return ret; } static gboolean gst_video_parse_sink_event (GstPad * pad, GstEvent * event) { GstVideoParse *vp = GST_VIDEO_PARSE (gst_pad_get_parent (pad)); gboolean ret; switch (GST_EVENT_TYPE (event)) { case GST_EVENT_FLUSH_STOP: gst_video_parse_reset (vp); ret = gst_pad_push_event (vp->srcpad, event); break; case GST_EVENT_NEWSEGMENT: { GstClockTimeDiff start, stop, time; gdouble rate, arate; gboolean update; GstFormat format; gst_event_parse_new_segment_full (event, &update, &rate, &arate, &format, &start, &stop, &time); if (format == GST_FORMAT_TIME) { ret = gst_pad_push_event (vp->srcpad, event); gst_segment_set_newsegment_full (&vp->segment, update, rate, arate, GST_FORMAT_TIME, start, stop, time); } else { gst_event_unref (event); ret = gst_video_parse_convert (vp, format, start, GST_FORMAT_TIME, &start); ret &= gst_video_parse_convert (vp, format, time, GST_FORMAT_TIME, &time); ret &= gst_video_parse_convert (vp, format, stop, GST_FORMAT_TIME, &stop); if (!ret) { GST_ERROR_OBJECT (vp, "Failed converting to GST_FORMAT_TIME format (%d)", format); break; } gst_segment_set_newsegment_full (&vp->segment, update, rate, arate, GST_FORMAT_TIME, start, stop, time); event = gst_event_new_new_segment (FALSE, vp->segment.rate, GST_FORMAT_TIME, start, stop, time); ret = gst_pad_push_event (vp->srcpad, event); } vp->n_frames = 0; vp->discont = TRUE; gst_adapter_clear (vp->adapter); break; } default: ret = gst_pad_event_default (vp->sinkpad, event); break; } gst_object_unref (vp); return ret; } static gboolean gst_video_parse_src_event (GstPad * pad, GstEvent * event) { GstVideoParse *vp = GST_VIDEO_PARSE (gst_pad_get_parent (pad)); gboolean ret; switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK:{ GstFormat format; gdouble rate; GstSeekFlags flags; GstSeekType start_type, stop_type; gint64 start, stop; gst_event_parse_seek (event, &rate, &format, &flags, &start_type, &start, &stop_type, &stop); /* First try if upstream handles the seek */ ret = gst_pad_push_event (vp->sinkpad, event); if (ret) goto done; /* Otherwise convert to bytes and push upstream */ if (format == GST_FORMAT_TIME || format == GST_FORMAT_DEFAULT) { gst_event_unref (event); ret = gst_video_parse_convert (vp, format, start, GST_FORMAT_BYTES, &start); ret &= gst_video_parse_convert (vp, format, stop, GST_FORMAT_BYTES, &stop); if (ret) { event = gst_event_new_seek (rate, GST_FORMAT_BYTES, flags, start_type, start, stop_type, stop); ret = gst_pad_push_event (vp->sinkpad, event); } } break; } default: ret = gst_pad_event_default (vp->srcpad, event); break; } done: gst_object_unref (vp); return ret; } static const GstQueryType * gst_video_parse_src_query_type (GstPad * pad) { static const GstQueryType types[] = { GST_QUERY_POSITION, GST_QUERY_DURATION, GST_QUERY_CONVERT, 0 }; return types; } static gboolean gst_video_parse_src_query (GstPad * pad, GstQuery * query) { GstVideoParse *vp = GST_VIDEO_PARSE (gst_pad_get_parent (pad)); gboolean ret = FALSE; GST_DEBUG ("src_query %s", gst_query_type_get_name (GST_QUERY_TYPE (query))); switch (GST_QUERY_TYPE (query)) { case GST_QUERY_POSITION: { GstFormat format; gint64 time, value; GST_LOG ("query position"); gst_query_parse_position (query, &format, NULL); time = gst_util_uint64_scale (vp->n_frames, GST_SECOND * vp->fps_d, vp->fps_n); ret = gst_video_parse_convert (vp, GST_FORMAT_TIME, time, format, &value); gst_query_set_position (query, format, value); break; } case GST_QUERY_DURATION:{ gint64 duration; GstFormat format; GstQuery *bquery; GST_LOG ("query duration"); ret = gst_pad_peer_query (vp->srcpad, query); if (ret) goto done; gst_query_parse_duration (query, &format, NULL); /* We only handle TIME and DEFAULT format */ if (format != GST_FORMAT_TIME && format != GST_FORMAT_DEFAULT) goto error; bquery = gst_query_new_duration (GST_FORMAT_BYTES); ret = gst_pad_peer_query (vp->srcpad, bquery); if (!ret) { gst_query_unref (bquery); goto error; } gst_query_parse_duration (bquery, NULL, &duration); gst_query_unref (bquery); ret = gst_video_parse_convert (vp, GST_FORMAT_BYTES, duration, format, &duration); if (ret) gst_query_set_duration (query, format, duration); break; } case GST_QUERY_CONVERT: { GstFormat src_fmt, dest_fmt; gint64 src_val, dest_val; GST_LOG ("query convert"); gst_query_parse_convert (query, &src_fmt, &src_val, &dest_fmt, &dest_val); ret = gst_video_parse_convert (vp, src_fmt, src_val, dest_fmt, &dest_val); if (!ret) goto error; gst_query_set_convert (query, src_fmt, src_val, dest_fmt, dest_val); break; } default: /* else forward upstream */ ret = gst_pad_peer_query (vp->sinkpad, query); break; } done: gst_object_unref (vp); return ret; error: GST_DEBUG_OBJECT (vp, "query failed"); goto done; } static gboolean plugin_init (GstPlugin * plugin) { gst_element_register (plugin, "videoparse", GST_RANK_NONE, gst_video_parse_get_type ()); return TRUE; } GST_PLUGIN_DEFINE (GST_VERSION_MAJOR, GST_VERSION_MINOR, "videoparse", "Parses byte streams into video frames", plugin_init, VERSION, "LGPL", GST_PACKAGE_NAME, GST_PACKAGE_ORIGIN);