/* * GStreamer * Copyright (C) 2009 Carl-Anton Ingmarsson * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include "gstvdpauvideobuffer.h" #include "gstvdpauvideoyuv.h" GST_DEBUG_CATEGORY_STATIC (gst_vdpau_video_yuv_debug); #define GST_CAT_DEFAULT gst_vdpau_video_yuv_debug /* Filter signals and args */ enum { /* FILL ME */ LAST_SIGNAL }; enum { PROP_0, PROP_SILENT }; static GstStaticPadTemplate sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VDPAU_VIDEO_CAPS)); static GstStaticPadTemplate src_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS ("video/x-raw-yuv, " "framerate = (fraction) [ 0, MAX ], " "width = (int) [ 1, MAX ], " "height = (int) [ 1, MAX ]")); #define DEBUG_INIT(bla) \ GST_DEBUG_CATEGORY_INIT (gst_vdpau_video_yuv_debug, "vdpauvideoyuv", 0, "VDPAU VdpSurface to YUV"); GST_BOILERPLATE_FULL (GstVdpauVideoYUV, gst_vdpau_video_yuv, GstElement, GST_TYPE_ELEMENT, DEBUG_INIT); static void gst_vdpau_video_yuv_finalize (GObject * object); static void gst_vdpau_video_yuv_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_vdpau_video_yuv_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); GstFlowReturn gst_vdpau_video_yuv_chain (GstPad * pad, GstBuffer * buffer) { GstVdpauVideoYUV *video_yuv; GstVdpauDevice *device; VdpVideoSurface surface; GstBuffer *outbuf = NULL; video_yuv = GST_VDPAU_VIDEO_YUV (GST_OBJECT_PARENT (pad)); device = GST_VDPAU_VIDEO_BUFFER (buffer)->device; surface = GST_VDPAU_VIDEO_BUFFER (buffer)->surface; switch (video_yuv->format) { case GST_MAKE_FOURCC ('Y', 'V', '1', '2'): { gint size; GstFlowReturn result; VdpStatus status; guint8 *data[3]; guint32 stride[3]; size = gst_video_format_get_size (GST_VIDEO_FORMAT_YV12, video_yuv->width, video_yuv->height); result = gst_pad_alloc_buffer_and_set_caps (video_yuv->src, GST_BUFFER_OFFSET_NONE, size, GST_PAD_CAPS (video_yuv->src), &outbuf); if (G_UNLIKELY (result != GST_FLOW_OK)) return result; data[0] = GST_BUFFER_DATA (outbuf) + gst_video_format_get_component_offset (GST_VIDEO_FORMAT_YV12, 0, video_yuv->width, video_yuv->height); data[1] = data[0] + gst_video_format_get_component_offset (GST_VIDEO_FORMAT_YV12, 2, video_yuv->width, video_yuv->height); data[2] = data[0] + gst_video_format_get_component_offset (GST_VIDEO_FORMAT_YV12, 1, video_yuv->width, video_yuv->height); stride[0] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_YV12, 0, video_yuv->width); stride[1] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_YV12, 2, video_yuv->width); stride[2] = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_YV12, 1, video_yuv->width); status = device->vdp_video_surface_get_bits_ycbcr (surface, VDP_YCBCR_FORMAT_YV12, (void *) data, stride); if (G_UNLIKELY (status != VDP_STATUS_OK)) { GST_ELEMENT_ERROR (video_yuv, RESOURCE, READ, ("Couldn't get data from vdpau"), ("Error returned from vdpau was: %s", device->vdp_get_error_string (status))); break; } break; } case GST_MAKE_FOURCC ('N', 'V', '1', '2'): { gint size; GstFlowReturn result; VdpStatus status; guint8 *data[2]; guint32 stride[2]; size = video_yuv->width * video_yuv->height + video_yuv->width * video_yuv->height / 2; result = gst_pad_alloc_buffer_and_set_caps (video_yuv->src, GST_BUFFER_OFFSET_NONE, size, GST_PAD_CAPS (video_yuv->src), &outbuf); if (G_UNLIKELY (result != GST_FLOW_OK)) return result; data[0] = GST_BUFFER_DATA (outbuf); data[1] = GST_BUFFER_DATA (outbuf) + video_yuv->width * video_yuv->height; stride[0] = video_yuv->width; stride[1] = video_yuv->width; status = device->vdp_video_surface_get_bits_ycbcr (surface, VDP_YCBCR_FORMAT_NV12, (void *) data, stride); if (G_UNLIKELY (status != VDP_STATUS_OK)) { GST_ELEMENT_ERROR (video_yuv, RESOURCE, READ, ("Couldn't get data from vdpau"), ("Error returned from vdpau was: %s", device->vdp_get_error_string (status))); break; } break; } default: break; } gst_buffer_unref (buffer); if (outbuf) { gst_buffer_copy_metadata (outbuf, buffer, GST_BUFFER_COPY_TIMESTAMPS); return gst_pad_push (video_yuv->src, outbuf); } return GST_FLOW_ERROR; } typedef struct { VdpChromaType chroma_type; VdpYCbCrFormat format; guint32 fourcc; } VdpauFormats; static VdpauFormats formats[6] = { { VDP_CHROMA_TYPE_420, VDP_YCBCR_FORMAT_NV12, GST_MAKE_FOURCC ('N', 'V', '1', '2') }, { VDP_CHROMA_TYPE_422, VDP_YCBCR_FORMAT_UYVY, GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y') }, { VDP_CHROMA_TYPE_444, VDP_YCBCR_FORMAT_V8U8Y8A8, GST_MAKE_FOURCC ('A', 'Y', 'U', 'V') }, { VDP_CHROMA_TYPE_444, VDP_YCBCR_FORMAT_Y8U8V8A8, GST_MAKE_FOURCC ('A', 'V', 'U', 'Y') }, { VDP_CHROMA_TYPE_422, VDP_YCBCR_FORMAT_YUYV, GST_MAKE_FOURCC ('Y', 'U', 'Y', 'V') }, { VDP_CHROMA_TYPE_420, VDP_YCBCR_FORMAT_YV12, GST_MAKE_FOURCC ('Y', 'V', '1', '2') } }; static GstCaps * gst_vdpau_video_yuv_get_caps (GstVdpauVideoYUV * video_yuv, GstVdpauDevice * device, gint chroma_type, gint width, gint height, gint framerate_numerator, gint framerate_denominator, gint par_numerator, gint par_denominator) { GstCaps *caps; gint i; caps = gst_caps_new_empty (); for (i = 0; i < 6; i++) { VdpStatus status; VdpBool is_supported; if (formats[i].chroma_type != chroma_type) continue; status = device->vdp_video_surface_query_ycbcr_capabilities (device->device, chroma_type, formats[i].format, &is_supported); if (status != VDP_STATUS_OK && status != VDP_STATUS_INVALID_Y_CB_CR_FORMAT) { GST_ELEMENT_ERROR (video_yuv, RESOURCE, READ, ("Could not query VDPAU YCbCr capabilites"), ("Error returned from vdpau was: %s", device->vdp_get_error_string (status))); return NULL; } if (is_supported) { GstCaps *format_caps; format_caps = gst_caps_new_simple ("video/x-raw-yuv", "format", GST_TYPE_FOURCC, formats[i].fourcc, "width", G_TYPE_INT, width, "height", G_TYPE_INT, height, "framerate", GST_TYPE_FRACTION, framerate_numerator, framerate_denominator, "pixel-aspect-ratio", GST_TYPE_FRACTION, par_numerator, par_denominator, NULL); gst_caps_append (caps, format_caps); } } if (gst_caps_is_empty (caps)) { gst_caps_unref (caps); return NULL; } return caps; } static gboolean gst_vdpau_video_yuv_sink_set_caps (GstPad * pad, GstCaps * caps) { GstVdpauVideoYUV *video_yuv = GST_VDPAU_VIDEO_YUV (GST_OBJECT_PARENT (pad)); GstCaps *src_caps, *new_caps; GstStructure *structure; const GValue *value; GstVdpauDevice *device; gint chroma_type; gint width, height; gint framerate_numerator, framerate_denominator; gint par_numerator, par_denominator; guint32 fourcc_format; gboolean res; structure = gst_caps_get_structure (caps, 0); value = gst_structure_get_value (structure, "device"); device = g_value_get_object (value); gst_structure_get_int (structure, "chroma-type", &chroma_type); gst_structure_get_int (structure, "width", &width); gst_structure_get_int (structure, "height", &height); gst_structure_get_fraction (structure, "framerate", &framerate_numerator, &framerate_denominator); gst_structure_get_fraction (structure, "pixel-aspect-ratio", &par_numerator, &par_denominator); src_caps = gst_vdpau_video_yuv_get_caps (video_yuv, device, chroma_type, width, height, framerate_numerator, framerate_denominator, par_numerator, par_denominator); if (G_UNLIKELY (!src_caps)) return FALSE; video_yuv->src_caps = src_caps; src_caps = gst_pad_get_allowed_caps (video_yuv->src); if (G_UNLIKELY (!src_caps || !gst_caps_get_size (src_caps))) return FALSE; new_caps = gst_caps_copy_nth (src_caps, 0); gst_caps_unref (src_caps); if (G_UNLIKELY (!new_caps)) return FALSE; structure = gst_caps_get_structure (new_caps, 0); gst_structure_get_fourcc (structure, "format", &fourcc_format); gst_pad_fixate_caps (video_yuv->src, new_caps); res = gst_pad_set_caps (video_yuv->src, new_caps); gst_caps_unref (new_caps); if (G_UNLIKELY (!res)) return FALSE; video_yuv->width = width; video_yuv->height = height; video_yuv->framerate_numerator = framerate_numerator; video_yuv->framerate_denominator = framerate_denominator; video_yuv->format = fourcc_format; return TRUE; } static GstCaps * gst_vdpau_video_yuv_src_getcaps (GstPad * pad) { GstVdpauVideoYUV *video_yuv; video_yuv = GST_VDPAU_VIDEO_YUV (GST_OBJECT_PARENT (pad)); if (video_yuv->src_caps) return gst_caps_copy (video_yuv->src_caps); if (GST_PAD_CAPS (video_yuv->src)) return gst_caps_copy (GST_PAD_CAPS (video_yuv->src)); return gst_caps_copy (gst_pad_get_pad_template_caps (video_yuv->src)); } /* GObject vmethod implementations */ static void gst_vdpau_video_yuv_base_init (gpointer klass) { GstElementClass *element_class = GST_ELEMENT_CLASS (klass); gst_element_class_set_details_simple (element_class, "VdpauVideoYUV", "Covideo_yuv/Decoder/Video", "VDPAU video surface to YUV", "Carl-Anton Ingmarsson "); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&sink_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&src_template)); } static void gst_vdpau_video_yuv_class_init (GstVdpauVideoYUVClass * klass) { GObjectClass *gobject_class; GstElementClass *gstelement_class; gobject_class = (GObjectClass *) klass; gstelement_class = (GstElementClass *) klass; gobject_class->finalize = gst_vdpau_video_yuv_finalize; gobject_class->set_property = gst_vdpau_video_yuv_set_property; gobject_class->get_property = gst_vdpau_video_yuv_get_property; g_object_class_install_property (gobject_class, PROP_SILENT, g_param_spec_boolean ("silent", "Silent", "Produce verbose output ?", FALSE, G_PARAM_READWRITE | GST_PARAM_CONTROLLABLE)); } static void gst_vdpau_video_yuv_init (GstVdpauVideoYUV * video_yuv, GstVdpauVideoYUVClass * klass) { video_yuv->silent = FALSE; video_yuv->src_caps = NULL; video_yuv->height = 0; video_yuv->width = 0; video_yuv->framerate_numerator = 0; video_yuv->framerate_denominator = 0; video_yuv->par_numerator = 1; video_yuv->par_denominator = 1; video_yuv->src = gst_pad_new_from_static_template (&src_template, "src"); gst_pad_set_getcaps_function (video_yuv->src, gst_vdpau_video_yuv_src_getcaps); gst_element_add_pad (GST_ELEMENT (video_yuv), video_yuv->src); video_yuv->sink = gst_pad_new_from_static_template (&sink_template, "sink"); gst_pad_set_setcaps_function (video_yuv->sink, gst_vdpau_video_yuv_sink_set_caps); gst_pad_set_chain_function (video_yuv->sink, gst_vdpau_video_yuv_chain); gst_element_add_pad (GST_ELEMENT (video_yuv), video_yuv->sink); } static void gst_vdpau_video_yuv_finalize (GObject * object) { GstVdpauVideoYUV *video_yuv = (GstVdpauVideoYUV *) object; if (video_yuv->src_caps) gst_caps_unref (video_yuv->src_caps); } static void gst_vdpau_video_yuv_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstVdpauVideoYUV *video_yuv = GST_VDPAU_VIDEO_YUV (object); switch (prop_id) { case PROP_SILENT: video_yuv->silent = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_vdpau_video_yuv_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVdpauVideoYUV *video_yuv = GST_VDPAU_VIDEO_YUV (object); switch (prop_id) { case PROP_SILENT: g_value_set_boolean (value, video_yuv->silent); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } }