/* GStreamer
 * Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
 *
 * This library is free software; you can redistribute it and/or
 * modify it under the terms of the GNU Library General Public
 * License as published by the Free Software Foundation; either
 * version 2 of the License, or (at your option) any later version.
 *
 * This library is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * Library General Public License for more details.
 *
 * You should have received a copy of the GNU Library General Public
 * License along with this library; if not, write to the
 * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
 * Boston, MA 02111-1307, USA.
 */

#ifdef HAVE_CONFIG_H
#include "config.h"
#endif

/*#define GST_DEBUG_ENABLED */
#include "gstmp1videoparse.h"

/* Start codes. */
#define SEQ_START_CODE 0x000001b3
#define GOP_START_CODE 0x000001b8
#define PICTURE_START_CODE 0x00000100
#define SLICE_MIN_START_CODE 0x00000101
#define SLICE_MAX_START_CODE 0x000001af
#define EXT_START_CODE 0x000001b5
#define USER_START_CODE 0x000001b2
#define SEQUENCE_ERROR_CODE 0x000001b4
#define SEQ_END_CODE 0x000001b7

/* elementfactory information */
static GstElementDetails mpeg1videoparse_details = GST_ELEMENT_DETAILS (
  "MPEG 1 Video Parser",
  "Codec/Parser/Video",
  "Parses and frames MPEG 1 video streams, provides seek",
  "Wim Taymans <wim.taymans@chello.be>"
);

static GstStaticPadTemplate src_factory =
GST_STATIC_PAD_TEMPLATE (
  "src",
  GST_PAD_SRC,
  GST_PAD_ALWAYS,
  GST_STATIC_CAPS ("video/mpeg, "
    "mpegversion = (int) 1, "
    "systemstream = (boolean) false, "
    "width = (int) [ 16, 4096 ], "
    "height = (int) [ 16, 4096 ], "
    "pixel_width = (int) [ 1, 255 ], "
    "pixel_height = (int) [ 1, 255 ], "
    "framerate = (double) [ 0, MAX ]"
  )
);

static GstStaticPadTemplate sink_factory =
GST_STATIC_PAD_TEMPLATE (
  "sink",
  GST_PAD_SINK,
  GST_PAD_ALWAYS,
  GST_STATIC_CAPS ("video/mpeg, "
    "mpegversion = (int) 1, "
    "systemstream = (boolean) false"
  )
);

/* Mp1VideoParse signals and args */
enum {
  /* FILL ME */
  LAST_SIGNAL
};

enum {
  ARG_0,
  /* FILL ME */
};

static void	gst_mp1videoparse_class_init	(Mp1VideoParseClass *klass);
static void	gst_mp1videoparse_base_init	(Mp1VideoParseClass *klass);
static void	gst_mp1videoparse_init		(Mp1VideoParse *mp1videoparse);

static void	gst_mp1videoparse_chain		(GstPad *pad, GstData *_data);
static void	gst_mp1videoparse_real_chain	(Mp1VideoParse *mp1videoparse, GstBuffer *buf, GstPad *outpad);
static void	gst_mp1videoparse_flush		(Mp1VideoParse *mp1videoparse);
static GstElementStateReturn
		gst_mp1videoparse_change_state	(GstElement *element);

static GstElementClass *parent_class = NULL;
/*static guint gst_mp1videoparse_signals[LAST_SIGNAL] = { 0 }; */

GType
mp1videoparse_get_type (void)
{
  static GType mp1videoparse_type = 0;

  if (!mp1videoparse_type) {
    static const GTypeInfo mp1videoparse_info = {
      sizeof(Mp1VideoParseClass),
      (GBaseInitFunc)gst_mp1videoparse_base_init,
      NULL,
      (GClassInitFunc)gst_mp1videoparse_class_init,
      NULL,
      NULL,
      sizeof(Mp1VideoParse),
      0,
      (GInstanceInitFunc)gst_mp1videoparse_init,
    };
    mp1videoparse_type = g_type_register_static(GST_TYPE_ELEMENT, "Mp1VideoParse", &mp1videoparse_info, 0);
  }
  return mp1videoparse_type;
}

static void
gst_mp1videoparse_base_init (Mp1VideoParseClass *klass)
{
  GstElementClass *element_class = GST_ELEMENT_CLASS (klass);

  gst_element_class_add_pad_template (element_class,
		gst_static_pad_template_get (&src_factory));
  gst_element_class_add_pad_template (element_class,
		gst_static_pad_template_get (&sink_factory));
  gst_element_class_set_details (element_class, &mpeg1videoparse_details);
}

static void
gst_mp1videoparse_class_init (Mp1VideoParseClass *klass)
{
  GstElementClass *gstelement_class;

  gstelement_class = (GstElementClass*)klass;

  parent_class = g_type_class_ref(GST_TYPE_ELEMENT);

  gstelement_class->change_state = gst_mp1videoparse_change_state;
}

static void
gst_mp1videoparse_init (Mp1VideoParse *mp1videoparse)
{
  mp1videoparse->sinkpad = gst_pad_new_from_template (
	gst_static_pad_template_get (&sink_factory), "sink");
  gst_element_add_pad(GST_ELEMENT(mp1videoparse),mp1videoparse->sinkpad);
  gst_pad_set_chain_function(mp1videoparse->sinkpad,gst_mp1videoparse_chain);

  mp1videoparse->srcpad = gst_pad_new_from_template (
	gst_static_pad_template_get (&src_factory), "src");
  gst_element_add_pad(GST_ELEMENT(mp1videoparse),mp1videoparse->srcpad);
  gst_pad_use_explicit_caps (mp1videoparse->srcpad);

  mp1videoparse->partialbuf = NULL;
  mp1videoparse->need_resync = FALSE;
  mp1videoparse->last_pts = GST_CLOCK_TIME_NONE;
  mp1videoparse->picture_in_buffer = 0;
  mp1videoparse->width = mp1videoparse->height = -1;
  mp1videoparse->fps = mp1videoparse->asr = 0.;
}

static void
mp1videoparse_parse_seq (Mp1VideoParse *mp1videoparse, GstBuffer *buf)
{
  gint width, height, asr_idx, fps_idx;
  gfloat asr_table[] = { 0., 1.,
			 0.6735, 0.7031, 0.7615, 0.8055, 0.8437, 
			 0.8935, 0.9157, 0.9815, 1.0255, 1.0695,
			 1.0950, 1.1575, 1.2015 };
  gfloat fps_table[] = { 0., 24./1.001, 24., 25.,
			 30./1.001, 30.,
			 50., 60./1.001, 60. };
  guint32 n = GUINT32_FROM_BE (*(guint32 *) GST_BUFFER_DATA (buf));

  width   = (n & 0xfff00000) >> 20;
  height  = (n & 0x000fff00) >>  8;
  asr_idx = (n & 0x000000f0) >>  4;
  fps_idx = (n & 0x0000000f) >>  0;

  if (fps_idx >= 9 || fps_idx <= 0)
    fps_idx = 3; /* well, we need a default */
  if (asr_idx >= 15 || asr_idx <= 0)
    asr_idx = 1; /* no aspect ratio */

  if (asr_table[asr_idx] != mp1videoparse->asr    ||
      fps_table[fps_idx] != mp1videoparse->fps    ||
      width              != mp1videoparse->width  ||
      height             != mp1videoparse->height) {
    GstCaps *caps;
    gint p_w, p_h;

    mp1videoparse->asr    = asr_table[asr_idx];
    mp1videoparse->fps    = fps_table[fps_idx];
    mp1videoparse->width  = width;
    mp1videoparse->height = height;

    p_w = (asr_table[asr_idx] < 1.0) ? (100 / asr_table[asr_idx]) : 1;
    p_h = (asr_table[asr_idx] > 1.0) ? (100 * asr_table[asr_idx]) : 1;

    caps = gst_caps_new_simple ("video/mpeg",
	"systemstream", G_TYPE_BOOLEAN, FALSE,
	"mpegversion",  G_TYPE_INT, 1,
	"width",        G_TYPE_INT, width,
	"height",       G_TYPE_INT, height,
	"framerate",    G_TYPE_DOUBLE, fps_table[fps_idx],
	"pixel_width",  G_TYPE_INT, p_w,
	"pixel_height", G_TYPE_INT, p_h, NULL);

    GST_DEBUG ("New mpeg1videoparse caps: " GST_PTR_FORMAT, caps);

    gst_pad_set_explicit_caps (mp1videoparse->srcpad, caps);
  }
}

static gboolean
mp1videoparse_valid_sync (Mp1VideoParse *mp1videoparse, guint32 head, GstBuffer *buf)
{
  switch (head) {
    case SEQ_START_CODE: {
      GstBuffer *subbuf = gst_buffer_create_sub (buf, 4,
		      				 GST_BUFFER_SIZE (buf) - 4);
      mp1videoparse_parse_seq(mp1videoparse, subbuf);
      gst_buffer_unref (subbuf);
      return TRUE;
    }
    case GOP_START_CODE:
    case PICTURE_START_CODE:
    case USER_START_CODE:
    case EXT_START_CODE:
      return TRUE;
    default:
      if (head >= SLICE_MIN_START_CODE &&
	  head <= SLICE_MAX_START_CODE)
        return TRUE;
  }

  return FALSE;
}

static gint
mp1videoparse_find_next_gop (Mp1VideoParse *mp1videoparse, GstBuffer *buf)
{
  guchar *data = GST_BUFFER_DATA(buf);
  gulong size = GST_BUFFER_SIZE(buf);
  gulong offset = 0;
  gint sync_zeros = 0;
  gboolean have_sync = FALSE;

  while (offset < size) {
    guchar byte = *(data+offset);
    offset++;
    if (byte == 0) {
      sync_zeros++;
    }
    else if (byte == 1 && sync_zeros >=2 ) {
      sync_zeros = 0;
      have_sync = TRUE;
    }
    else if (have_sync) {
      if (byte == (SEQ_START_CODE & 0xff) || byte == (GOP_START_CODE & 0xff)) {
        return offset - 4;
      } else {
        sync_zeros = 0;
	have_sync = FALSE;
      }
    }
    else {
      sync_zeros = 0;
    }
  }

  return -1;
}

static guint64
gst_mp1videoparse_time_code (guchar *gop,
			     gfloat  fps)
{
  guint32 data = GUINT32_FROM_BE (* (guint32 *) gop);

  return ((((data & 0xfc000000) >> 26) * 3600 * GST_SECOND) + /* hours */
	  (((data & 0x03f00000) >> 20) *   60 * GST_SECOND) + /* minutes */
	  (((data & 0x0007e000) >> 13) *        GST_SECOND) + /* seconds */
	  (((data & 0x00001f80) >>  7) * GST_SECOND / fps));  /* frames */
}

static void
gst_mp1videoparse_flush (Mp1VideoParse *mp1videoparse)
{
  GST_DEBUG ("mp1videoparse: flushing");
  if (mp1videoparse->partialbuf) {
    gst_buffer_unref(mp1videoparse->partialbuf);
    mp1videoparse->partialbuf= NULL;
  }
  mp1videoparse->need_resync = TRUE;
  mp1videoparse->in_flush = TRUE;
  mp1videoparse->picture_in_buffer = 0;
}

static void
gst_mp1videoparse_chain (GstPad *pad,GstData *_data)
{
  GstBuffer *buf = GST_BUFFER (_data);
  Mp1VideoParse *mp1videoparse;

  g_return_if_fail (pad != NULL);
  g_return_if_fail (GST_IS_PAD (pad));
  g_return_if_fail (buf != NULL);

  mp1videoparse = GST_MP1VIDEOPARSE (GST_OBJECT_PARENT (pad));

  gst_mp1videoparse_real_chain(mp1videoparse, buf, mp1videoparse->srcpad);
}

static void
gst_mp1videoparse_real_chain (Mp1VideoParse *mp1videoparse, GstBuffer *buf, GstPad *outpad)
{
  guchar *data;
  gulong size, offset = 0;
  GstBuffer *outbuf;
  gint sync_state;
  gboolean have_sync;
  guchar sync_byte;
  guint32 head;
  gint sync_pos;
  guint64 time_stamp;
  GstBuffer *temp;

  time_stamp = GST_BUFFER_TIMESTAMP(buf);

  if (GST_IS_EVENT (buf)) {
    GstEvent *event = GST_EVENT (buf);

    switch (GST_EVENT_TYPE (event)) {
      case GST_EVENT_FLUSH:
      case GST_EVENT_DISCONTINUOUS:
        gst_mp1videoparse_flush(mp1videoparse);
        break;
      case GST_EVENT_EOS:
        gst_mp1videoparse_flush(mp1videoparse);
        gst_event_ref(event);
        gst_pad_push(outpad, GST_DATA (event));
        gst_element_set_eos (GST_ELEMENT (mp1videoparse));
        break;
      default:
        GST_DEBUG ("Unhandled event type %d",
		   GST_EVENT_TYPE (event));
        break;
    }
 
    gst_event_unref (event);
    return;
  }
 

  if (mp1videoparse->partialbuf) {
    GstBuffer *merge;

    offset = GST_BUFFER_SIZE(mp1videoparse->partialbuf);
    merge = gst_buffer_merge(mp1videoparse->partialbuf, buf);

    gst_buffer_unref(mp1videoparse->partialbuf);
    gst_buffer_unref(buf);

    mp1videoparse->partialbuf = merge;
  }
  else {
    mp1videoparse->partialbuf = buf;
    offset = 0;
  }

  data = GST_BUFFER_DATA(mp1videoparse->partialbuf);
  size = GST_BUFFER_SIZE(mp1videoparse->partialbuf);

  GST_DEBUG ("mp1videoparse: received buffer of %ld bytes %" G_GINT64_FORMAT,size, GST_BUFFER_TIMESTAMP(buf));

  do {
    data = GST_BUFFER_DATA(mp1videoparse->partialbuf);
    size = GST_BUFFER_SIZE(mp1videoparse->partialbuf);

    head = GUINT32_FROM_BE(*((guint32 *)data));

    GST_DEBUG ("mp1videoparse: head is %08x", (unsigned int)head);

    if (!mp1videoparse_valid_sync(mp1videoparse, head,
				  mp1videoparse->partialbuf) ||
        mp1videoparse->need_resync) {
      sync_pos = mp1videoparse_find_next_gop(mp1videoparse, mp1videoparse->partialbuf);
      if (sync_pos >= 0) {
        mp1videoparse->need_resync = FALSE;
        GST_DEBUG ("mp1videoparse: found new gop at %d", sync_pos);

        if (sync_pos != 0) {
          temp = gst_buffer_create_sub(mp1videoparse->partialbuf, sync_pos, size-sync_pos);
	  g_assert(temp != NULL);
	  gst_buffer_unref(mp1videoparse->partialbuf);
	  mp1videoparse->partialbuf = temp;
          data = GST_BUFFER_DATA(mp1videoparse->partialbuf);
          size = GST_BUFFER_SIZE(mp1videoparse->partialbuf);
	  offset = 0;
        }

        head = GUINT32_FROM_BE(*((guint32 *)data));
        /* re-call this function so that if we hadn't already, we can
         * now read the sequence header and parse video properties,
         * set caps, stream data, be happy, bla, bla, bla... */
        if (!mp1videoparse_valid_sync (mp1videoparse, head,
				       mp1videoparse->partialbuf))
          g_error ("Found sync but no valid sync point at pos 0x0");
      }
      else {
        GST_DEBUG ("mp1videoparse: could not sync");
        gst_buffer_unref(mp1videoparse->partialbuf);
        mp1videoparse->partialbuf = NULL;
        return;
      }
    }

    if (mp1videoparse->picture_in_buffer == 1 &&
        time_stamp != GST_CLOCK_TIME_NONE) {
      mp1videoparse->last_pts = time_stamp;
    }

    sync_state = 0;
    have_sync = FALSE;

    GST_DEBUG ("mp1videoparse: searching sync");

    while (offset < size-1) {
      sync_byte = *(data + offset);
      if (sync_byte == 0) {
        sync_state++;
      }
      else if ((sync_byte == 1) && (sync_state >=2)) {
        GST_DEBUG ("mp1videoparse: code 0x000001%02x",data[offset+1]);
        if (data[offset+1] == (PICTURE_START_CODE & 0xff)) {
	  mp1videoparse->picture_in_buffer++;
	  if (mp1videoparse->picture_in_buffer == 1) {
	    if (time_stamp != GST_CLOCK_TIME_NONE) {
              mp1videoparse->last_pts = time_stamp;
	    }
	    sync_state = 0;
	  }
	  else if (mp1videoparse->picture_in_buffer == 2) {
            have_sync = TRUE;
            break;
	  }
	  else {
            GST_DEBUG ("mp1videoparse: %d in buffer", mp1videoparse->picture_in_buffer);
            g_assert_not_reached();
	  }
        }
        /* A new sequence (or GOP) is a valid sync too. Note that the
         * sequence header should be put in the next buffer, not here. */
        else if (data[offset+1] == (SEQ_START_CODE & 0xFF) ||
	         data[offset+1] == (GOP_START_CODE & 0xFF)) {
          if (mp1videoparse->picture_in_buffer == 0 &&
	      data[offset+1] == (GOP_START_CODE & 0xFF)) {
	    mp1videoparse->last_pts = gst_mp1videoparse_time_code (&data[2],
			  			mp1videoparse->fps);
	  }
          else if (mp1videoparse->picture_in_buffer == 1) {
	    have_sync = TRUE;
	    break;
	  } else {
	    g_assert (mp1videoparse->picture_in_buffer == 0);
	  }
        }
        /* end-of-sequence is a valid sync point and should be included
         * in the current picture, not the next. */
        else if (data[offset+1] == (SEQ_END_CODE & 0xFF)) {
          if (mp1videoparse->picture_in_buffer == 1) {
            offset += 4;
	    have_sync = TRUE;
	    break;
	  } else {
	    g_assert (mp1videoparse->picture_in_buffer == 0);
	  }
        }
        else sync_state = 0;
      }
      /* something else... */
      else sync_state = 0;
      /* go down the buffer */
      offset++;
    }

    if (have_sync) {
      offset -= 2;
      GST_DEBUG ("mp1videoparse: synced at %ld code 0x000001%02x",offset,data[offset+3]);

      outbuf = gst_buffer_create_sub(mp1videoparse->partialbuf, 0, offset+4);
      g_assert(outbuf != NULL);
      GST_BUFFER_TIMESTAMP(outbuf) = mp1videoparse->last_pts;
      GST_BUFFER_DURATION(outbuf) = GST_SECOND / mp1videoparse->fps;
      mp1videoparse->last_pts += GST_BUFFER_DURATION (outbuf);

      if (mp1videoparse->in_flush) {
        /* FIXME, send a flush event here */
        mp1videoparse->in_flush = FALSE;
      }

      if (GST_PAD_CAPS (outpad) != NULL) {
        GST_DEBUG ("mp1videoparse: pushing  %d bytes %" G_GUINT64_FORMAT, GST_BUFFER_SIZE(outbuf), GST_BUFFER_TIMESTAMP(outbuf));
	gst_pad_push(outpad, GST_DATA (outbuf));
        GST_DEBUG ("mp1videoparse: pushing  done");
      } else {
        GST_DEBUG ("No capsnego yet, delaying buffer push");
        gst_buffer_unref (outbuf);
      }
      mp1videoparse->picture_in_buffer = 0;

      temp = gst_buffer_create_sub(mp1videoparse->partialbuf, offset, size-offset);
      gst_buffer_unref(mp1videoparse->partialbuf);
      mp1videoparse->partialbuf = temp;
      offset = 0;
    }
    else {
      if (time_stamp != GST_CLOCK_TIME_NONE) {
        mp1videoparse->last_pts = time_stamp;
	break;
      }
    }
  } while (1);
}

static GstElementStateReturn 
gst_mp1videoparse_change_state (GstElement *element) 
{
  Mp1VideoParse *mp1videoparse;
  g_return_val_if_fail(GST_IS_MP1VIDEOPARSE(element),GST_STATE_FAILURE);

  mp1videoparse = GST_MP1VIDEOPARSE(element);

  switch (GST_STATE_TRANSITION (element)) {
    case GST_STATE_PAUSED_TO_READY:
      gst_mp1videoparse_flush(mp1videoparse);
      mp1videoparse->width = mp1videoparse->height = -1;
      mp1videoparse->fps   = mp1videoparse->asr = 0.;
      break;
    default:
      break;
  }

  if (GST_ELEMENT_CLASS(parent_class)->change_state)
    return GST_ELEMENT_CLASS(parent_class)->change_state(element);

  return GST_STATE_SUCCESS;
}

static gboolean
plugin_init (GstPlugin *plugin)
{
  return gst_element_register (plugin, "mpeg1videoparse",
			       GST_RANK_NONE, GST_TYPE_MP1VIDEOPARSE);
}

GST_PLUGIN_DEFINE (
  GST_VERSION_MAJOR,
  GST_VERSION_MINOR,
  "mpeg1videoparse",
  "MPEG-1 video parser",
  plugin_init,
  VERSION,
  GST_LICENSE,
  GST_PACKAGE,
  GST_ORIGIN
)