ORC_SOURCE=videoblendorc
include $(top_srcdir)/common/orc.mak
-glib_enum_headers = video.h colorbalance.h navigation.h
+glib_enum_headers = video.h video-format.h video-color.h video-info.h \
+ colorbalance.h navigation.h
glib_enum_define = GST_VIDEO
glib_gen_prefix = gst_video
glib_gen_basename = video
colorbalancechannel.c \
navigation.c \
video.c \
+ video-event.c \
+ video-format.c \
+ video-color.c \
+ video-info.c \
+ video-frame.c \
gstvideosink.c \
gstvideofilter.c \
convertframe.c \
colorbalancechannel.h \
navigation.h \
video.h \
+ video-event.h \
+ video-format.h \
+ video-color.h \
+ video-info.h \
+ video-frame.h \
gstvideosink.h \
gstvideofilter.h \
gstvideometa.h \
--- /dev/null
+/* GStreamer
+ * Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
+ * Library <2002> Ronald Bultje <rbultje@ronald.bitfreak.net>
+ * Copyright (C) 2007 David A. Schleef <ds@schleef.org>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifdef HAVE_CONFIG_H
+# include "config.h"
+#endif
+
+#include <string.h>
+#include <stdio.h>
+
+#include "video-color.h"
+
+typedef struct
+{
+ const gchar *name;
+ GstVideoColorimetry color;
+} ColorimetryInfo;
+
+#define MAKE_COLORIMETRY(n,r,m,t,p) { GST_VIDEO_COLORIMETRY_ ##n, \
+ { GST_VIDEO_COLOR_RANGE ##r, GST_VIDEO_COLOR_MATRIX_ ##m, \
+ GST_VIDEO_TRANSFER_ ##t, GST_VIDEO_COLOR_PRIMARIES_ ##p } }
+
+#define GST_VIDEO_COLORIMETRY_NONAME NULL
+
+#define DEFAULT_YUV_SD 0
+#define DEFAULT_YUV_HD 1
+#define DEFAULT_RGB 3
+#define DEFAULT_GRAY 4
+#define DEFAULT_UNKNOWN 5
+
+static const ColorimetryInfo colorimetry[] = {
+ MAKE_COLORIMETRY (BT601, _16_235, BT601, BT709, BT470M),
+ MAKE_COLORIMETRY (BT709, _16_235, BT709, BT709, BT709),
+ MAKE_COLORIMETRY (SMPTE240M, _16_235, SMPTE240M, SMPTE240M, SMPTE240M),
+ MAKE_COLORIMETRY (NONAME, _0_255, RGB, UNKNOWN, UNKNOWN),
+ MAKE_COLORIMETRY (NONAME, _0_255, BT601, UNKNOWN, UNKNOWN),
+ MAKE_COLORIMETRY (NONAME, _UNKNOWN, UNKNOWN, UNKNOWN, UNKNOWN),
+};
+
+static const ColorimetryInfo *
+gst_video_get_colorimetry (const gchar * s)
+{
+ gint i;
+
+ for (i = 0; colorimetry[i].name; i++) {
+ if (g_str_equal (colorimetry[i].name, s))
+ return &colorimetry[i];
+ }
+ return NULL;
+}
+
+#define IS_EQUAL(ci,i) (((ci)->color.range == (i)->range) && \
+ ((ci)->color.matrix == (i)->matrix) && \
+ ((ci)->color.transfer == (i)->transfer) && \
+ ((ci)->color.primaries == (i)->primaries))
+
+#define IS_UNKNOWN(ci) (IS_EQUAL (&colorimetry[DEFAULT_UNKNOWN], ci))
+
+/**
+ * gst_video_colorimetry_from_string
+ * @cinfo: a #GstVideoColorimetry
+ * @color: a colorimetry string
+ *
+ * Parse the colorimetry string and update @cinfo with the parsed
+ * values.
+ *
+ * Returns: #TRUE if @color points to valid colorimetry info.
+ */
+gboolean
+gst_video_colorimetry_from_string (GstVideoColorimetry * cinfo,
+ const gchar * color)
+{
+ const ColorimetryInfo *ci;
+
+ if ((ci = gst_video_get_colorimetry (color))) {
+ *cinfo = ci->color;
+ } else {
+ gint r, m, t, p;
+
+ if (sscanf (color, "%d:%d:%d:%d", &r, &m, &t, &p) == 4) {
+ cinfo->range = r;
+ cinfo->matrix = m;
+ cinfo->transfer = t;
+ cinfo->primaries = p;
+ }
+ }
+ return TRUE;
+}
+
+/**
+ * gst_video_colorimetry_to_string
+ * @cinfo: a #GstVideoColorimetry
+ *
+ * Make a string representation of @cinfo.
+ *
+ * Returns: a string representation of @cinfo.
+ */
+gchar *
+gst_video_colorimetry_to_string (GstVideoColorimetry * cinfo)
+{
+ gint i;
+
+ for (i = 0; colorimetry[i].name; i++) {
+ if (IS_EQUAL (&colorimetry[i], cinfo)) {
+ return g_strdup (colorimetry[i].name);
+ }
+ }
+ if (!IS_UNKNOWN (cinfo)) {
+ return g_strdup_printf ("%d:%d:%d:%d", cinfo->range, cinfo->matrix,
+ cinfo->transfer, cinfo->primaries);
+ }
+ return NULL;
+}
+
+/**
+ * gst_video_colorimetry_matches:
+ * @info: a #GstVideoInfo
+ * @color: a colorimetry string
+ *
+ * Check if the colorimetry information in @info matches that of the
+ * string @color.
+ *
+ * Returns: #TRUE if @color conveys the same colorimetry info as the color
+ * information in @info.
+ */
+gboolean
+gst_video_colorimetry_matches (GstVideoColorimetry * cinfo, const gchar * color)
+{
+ const ColorimetryInfo *ci;
+
+ if ((ci = gst_video_get_colorimetry (color)))
+ return IS_EQUAL (ci, cinfo);
+
+ return FALSE;
+}
+
+#if 0
+typedef struct
+{
+ GstVideoColorPrimaries primaries;
+ gdouble xW, yW;
+ gdouble xR, yR;
+ gdouble xG, yG;
+ gdouble xB, yB;
+} PrimariesInfo;
+
+#define WP_C 0.31006, 0.31616
+#define WP_D65 0.31271, 0.32902
+
+static const PrimariesInfo primaries[] = {
+ {GST_VIDEO_COLOR_PRIMARIES_UNKNOWN, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0},
+ {GST_VIDEO_COLOR_PRIMARIES_BT709, WP_D65, 0.64, 0.33, 0.30, 0.60, 0.15, 0.06},
+ {GST_VIDEO_COLOR_PRIMARIES_BT470M, WP_C, 0.67, 0.33, 0.21, 0.71, 0.14, 0.08},
+ {GST_VIDEO_COLOR_PRIMARIES_BT470BG, WP_D65, 0.64, 0.33, 0.29, 0.60, 0.15,
+ 0.06},
+ {GST_VIDEO_COLOR_PRIMARIES_SMPTE170M, WP_D65, 0.63, 0.34, 0.31, 0.595, 0.155,
+ 0.07},
+ {GST_VIDEO_COLOR_PRIMARIES_SMPTE240M, WP_D65, 0.63, 0.34, 0.31, 0.595, 0.155,
+ 0.07},
+ {GST_VIDEO_COLOR_PRIMARIES_FILM, WP_C, 0.681, 0.319, 0.243, 0.692, 0.145,
+ 0.049}
+};
+#endif
--- /dev/null
+/* GStreamer
+ * Copyright (C) <2011> Wim Taymans <wim.taymans@gmail.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifndef __GST_VIDEO_COLOR_H__
+#define __GST_VIDEO_COLOR_H__
+
+#include <gst/gst.h>
+
+G_BEGIN_DECLS
+
+/**
+ * GstVideoColorRange:
+ * @GST_VIDEO_COLOR_RANGE_UNKNOWN: unknown range
+ * @GST_VIDEO_COLOR_RANGE_0_255: [0..255] for 8 bit components
+ * @GST_VIDEO_COLOR_RANGE_16_235: [16..235] for 8 bit components. Chroma has
+ * [16..240] range.
+ *
+ * Possible color range values. These constants are defined for 8 bit color
+ * values and can be scaled for other bit depths.
+ */
+typedef enum {
+ GST_VIDEO_COLOR_RANGE_UNKNOWN = 0,
+ GST_VIDEO_COLOR_RANGE_0_255,
+ GST_VIDEO_COLOR_RANGE_16_235
+} GstVideoColorRange;
+
+/**
+ * GstVideoColorMatrix:
+ * @GST_VIDEO_COLOR_MATRIX_UNKNOWN: unknown matrix
+ * @GST_VIDEO_COLOR_MATRIX_RGB: identity matrix
+ * @GST_VIDEO_COLOR_MATRIX_FCC: FCC color matrix
+ * @GST_VIDEO_COLOR_MATRIX_BT709: ITU-R BT.709 color matrix
+ * @GST_VIDEO_COLOR_MATRIX_BT601: ITU-R BT.601 color matrix
+ * @GST_VIDEO_COLOR_MATRIX_SMPTE240M: SMPTE 240M color matrix
+ *
+ * The color matrix is used to convert between Y'PbPr and
+ * non-linear RGB (R'G'B')
+ */
+typedef enum {
+ GST_VIDEO_COLOR_MATRIX_UNKNOWN = 0,
+ GST_VIDEO_COLOR_MATRIX_RGB,
+ GST_VIDEO_COLOR_MATRIX_FCC,
+ GST_VIDEO_COLOR_MATRIX_BT709,
+ GST_VIDEO_COLOR_MATRIX_BT601,
+ GST_VIDEO_COLOR_MATRIX_SMPTE240M
+} GstVideoColorMatrix;
+
+/**
+ * GstVideoTransferFunction:
+ * @GST_VIDEO_TRANSFER_UNKNOWN: unknown transfer function
+ * @GST_VIDEO_TRANSFER_GAMMA10: linear RGB, gamma 1.0 curve
+ * @GST_VIDEO_TRANSFER_GAMMA18: Gamma 1.8 curve
+ * @GST_VIDEO_TRANSFER_GAMMA20: Gamma 2.0 curve
+ * @GST_VIDEO_TRANSFER_GAMMA22: Gamma 2.2 curve
+ * @GST_VIDEO_TRANSFER_BT709: Gamma 2.2 curve with a linear segment in the lower
+ * range
+ * @GST_VIDEO_TRANSFER_SMPTE240M: Gamma 2.2 curve with a linear segment in the
+ * lower range
+ * @GST_VIDEO_TRANSFER_SRGB: Gamma 2.4 curve with a linear segment in the lower
+ * range
+ * @GST_VIDEO_TRANSFER_GAMMA28: Gamma 2.8 curve
+ * @GST_VIDEO_TRANSFER_LOG100: Logarithmic transfer characteristic
+ * 100:1 range
+ * @GST_VIDEO_TRANSFER_LOG316: Logarithmic transfer characteristic
+ * 316.22777:1 range
+ *
+ * The video transfer function defines the formula for converting between
+ * non-linear RGB (R'G'B') and linear RGB
+ */
+typedef enum {
+ GST_VIDEO_TRANSFER_UNKNOWN = 0,
+ GST_VIDEO_TRANSFER_GAMMA10,
+ GST_VIDEO_TRANSFER_GAMMA18,
+ GST_VIDEO_TRANSFER_GAMMA20,
+ GST_VIDEO_TRANSFER_GAMMA22,
+ GST_VIDEO_TRANSFER_BT709,
+ GST_VIDEO_TRANSFER_SMPTE240M,
+ GST_VIDEO_TRANSFER_SRGB,
+ GST_VIDEO_TRANSFER_GAMMA28,
+ GST_VIDEO_TRANSFER_LOG100,
+ GST_VIDEO_TRANSFER_LOG316
+} GstVideoTransferFunction;
+
+/**
+ * GstVideoColorPrimaries:
+ * @GST_VIDEO_COLOR_PRIMARIES_UNKNOWN: unknown color primaries
+ * @GST_VIDEO_COLOR_PRIMARIES_BT709: BT709 primaries
+ * @GST_VIDEO_COLOR_PRIMARIES_BT470M: BT470M primaries
+ * @GST_VIDEO_COLOR_PRIMARIES_BT470BG: BT470BG primaries
+ * @GST_VIDEO_COLOR_PRIMARIES_SMPTE170M: SMPTE170M primaries
+ * @GST_VIDEO_COLOR_PRIMARIES_SMPTE240M: SMPTE240M primaries
+ * @GST_VIDEO_COLOR_PRIMARIES_FILM: Generic film
+ *
+ * The color primaries define the how to transform linear RGB values to and from
+ * the CIE XYZ colorspace.
+ */
+typedef enum {
+ GST_VIDEO_COLOR_PRIMARIES_UNKNOWN = 0,
+ GST_VIDEO_COLOR_PRIMARIES_BT709,
+ GST_VIDEO_COLOR_PRIMARIES_BT470M,
+ GST_VIDEO_COLOR_PRIMARIES_BT470BG,
+ GST_VIDEO_COLOR_PRIMARIES_SMPTE170M,
+ GST_VIDEO_COLOR_PRIMARIES_SMPTE240M,
+ GST_VIDEO_COLOR_PRIMARIES_FILM
+} GstVideoColorPrimaries;
+
+/**
+ * GstVideoColorimetry:
+ * @range: the color range. This is the valid range for the samples.
+ * It is used to convert the samples to Y'PbPr values.
+ * @matrix: the color matrix. Used to convert between Y'PbPr and
+ * non-linear RGB (R'G'B')
+ * @transfer: the transfer function. used to convert between R'G'B' and RGB
+ * @primaries: color primaries. used to convert between R'G'B' and CIE XYZ
+ *
+ * Structure describing the color info.
+ */
+typedef struct {
+ GstVideoColorRange range;
+ GstVideoColorMatrix matrix;
+ GstVideoTransferFunction transfer;
+ GstVideoColorPrimaries primaries;
+} GstVideoColorimetry;
+
+/* predefined colorimetry */
+#define GST_VIDEO_COLORIMETRY_BT601 "bt601"
+#define GST_VIDEO_COLORIMETRY_BT709 "bt709"
+#define GST_VIDEO_COLORIMETRY_SMPTE240M "smpte240m"
+
+gboolean gst_video_colorimetry_matches (GstVideoColorimetry *cinfo, const gchar *color);
+gboolean gst_video_colorimetry_from_string (GstVideoColorimetry *cinfo, const gchar *color);
+gchar * gst_video_colorimetry_to_string (GstVideoColorimetry *cinfo);
+
+G_END_DECLS
+
+#endif /* __GST_VIDEO_COLOR_H__ */
--- /dev/null
+/* GStreamer
+ * Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
+ * Library <2002> Ronald Bultje <rbultje@ronald.bitfreak.net>
+ * Copyright (C) 2007 David A. Schleef <ds@schleef.org>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#include "video-event.h"
+
+#define GST_VIDEO_EVENT_STILL_STATE_NAME "GstEventStillFrame"
+
+/**
+ * gst_video_event_new_still_frame:
+ * @in_still: boolean value for the still-frame state of the event.
+ *
+ * Creates a new Still Frame event. If @in_still is %TRUE, then the event
+ * represents the start of a still frame sequence. If it is %FALSE, then
+ * the event ends a still frame sequence.
+ *
+ * To parse an event created by gst_video_event_new_still_frame() use
+ * gst_video_event_parse_still_frame().
+ *
+ * Returns: The new GstEvent
+ * Since: 0.10.26
+ */
+GstEvent *
+gst_video_event_new_still_frame (gboolean in_still)
+{
+ GstEvent *still_event;
+ GstStructure *s;
+
+ s = gst_structure_new (GST_VIDEO_EVENT_STILL_STATE_NAME,
+ "still-state", G_TYPE_BOOLEAN, in_still, NULL);
+ still_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s);
+
+ return still_event;
+}
+
+/**
+ * gst_video_event_parse_still_frame:
+ * @event: A #GstEvent to parse
+ * @in_still: A boolean to receive the still-frame status from the event, or NULL
+ *
+ * Parse a #GstEvent, identify if it is a Still Frame event, and
+ * return the still-frame state from the event if it is.
+ * If the event represents the start of a still frame, the in_still
+ * variable will be set to TRUE, otherwise FALSE. It is OK to pass NULL for the
+ * in_still variable order to just check whether the event is a valid still-frame
+ * event.
+ *
+ * Create a still frame event using gst_video_event_new_still_frame()
+ *
+ * Returns: %TRUE if the event is a valid still-frame event. %FALSE if not
+ * Since: 0.10.26
+ */
+gboolean
+gst_video_event_parse_still_frame (GstEvent * event, gboolean * in_still)
+{
+ const GstStructure *s;
+ gboolean ev_still_state;
+
+ g_return_val_if_fail (event != NULL, FALSE);
+
+ if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM)
+ return FALSE; /* Not a still frame event */
+
+ s = gst_event_get_structure (event);
+ if (s == NULL
+ || !gst_structure_has_name (s, GST_VIDEO_EVENT_STILL_STATE_NAME))
+ return FALSE; /* Not a still frame event */
+ if (!gst_structure_get_boolean (s, "still-state", &ev_still_state))
+ return FALSE; /* Not a still frame event */
+ if (in_still)
+ *in_still = ev_still_state;
+ return TRUE;
+}
+
+#define GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME "GstForceKeyUnit"
+
+/**
+ * gst_video_event_new_downstream_force_key_unit:
+ * @timestamp: the timestamp of the buffer that starts a new key unit
+ * @stream_time: the stream_time of the buffer that starts a new key unit
+ * @running_time: the running_time of the buffer that starts a new key unit
+ * @all_headers: %TRUE to produce headers when starting a new key unit
+ * @count: integer that can be used to number key units
+ *
+ * Creates a new downstream force key unit event. A downstream force key unit
+ * event can be sent down the pipeline to request downstream elements to produce
+ * a key unit. A downstream force key unit event must also be sent when handling
+ * an upstream force key unit event to notify downstream that the latter has been
+ * handled.
+ *
+ * To parse an event created by gst_video_event_new_downstream_force_key_unit() use
+ * gst_video_event_parse_downstream_force_key_unit().
+ *
+ * Returns: The new GstEvent
+ * Since: 0.10.36
+ */
+GstEvent *
+gst_video_event_new_downstream_force_key_unit (GstClockTime timestamp,
+ GstClockTime stream_time, GstClockTime running_time, gboolean all_headers,
+ guint count)
+{
+ GstEvent *force_key_unit_event;
+ GstStructure *s;
+
+ s = gst_structure_new (GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME,
+ "timestamp", G_TYPE_UINT64, timestamp,
+ "stream-time", G_TYPE_UINT64, stream_time,
+ "running-time", G_TYPE_UINT64, running_time,
+ "all-headers", G_TYPE_BOOLEAN, all_headers,
+ "count", G_TYPE_UINT, count, NULL);
+ force_key_unit_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s);
+
+ return force_key_unit_event;
+}
+
+/**
+ * gst_video_event_new_upstream_force_key_unit:
+ * @running_time: the running_time at which a new key unit should be produced
+ * @all_headers: %TRUE to produce headers when starting a new key unit
+ * @count: integer that can be used to number key units
+ *
+ * Creates a new upstream force key unit event. An upstream force key unit event
+ * can be sent to request upstream elements to produce a key unit.
+ *
+ * @running_time can be set to request a new key unit at a specific
+ * running_time. If set to GST_CLOCK_TIME_NONE, upstream elements will produce a
+ * new key unit as soon as possible.
+ *
+ * To parse an event created by gst_video_event_new_downstream_force_key_unit() use
+ * gst_video_event_parse_downstream_force_key_unit().
+ *
+ * Returns: The new GstEvent
+ * Since: 0.10.36
+ */
+GstEvent *
+gst_video_event_new_upstream_force_key_unit (GstClockTime running_time,
+ gboolean all_headers, guint count)
+{
+ GstEvent *force_key_unit_event;
+ GstStructure *s;
+
+ s = gst_structure_new (GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME,
+ "running-time", GST_TYPE_CLOCK_TIME, running_time,
+ "all-headers", G_TYPE_BOOLEAN, all_headers,
+ "count", G_TYPE_UINT, count, NULL);
+ force_key_unit_event = gst_event_new_custom (GST_EVENT_CUSTOM_UPSTREAM, s);
+
+ return force_key_unit_event;
+}
+
+/**
+ * gst_video_event_is_force_key_unit:
+ * @event: A #GstEvent to check
+ *
+ * Checks if an event is a force key unit event. Returns true for both upstream
+ * and downstream force key unit events.
+ *
+ * Returns: %TRUE if the event is a valid force key unit event
+ * Since: 0.10.36
+ */
+gboolean
+gst_video_event_is_force_key_unit (GstEvent * event)
+{
+ const GstStructure *s;
+
+ g_return_val_if_fail (event != NULL, FALSE);
+
+ if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM &&
+ GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_UPSTREAM)
+ return FALSE; /* Not a force key unit event */
+
+ s = gst_event_get_structure (event);
+ if (s == NULL
+ || !gst_structure_has_name (s, GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME))
+ return FALSE;
+
+ return TRUE;
+}
+
+/**
+ * gst_video_event_parse_downstream_force_key_unit:
+ * @event: A #GstEvent to parse
+ * @timestamp: (out): A pointer to the timestamp in the event
+ * @stream_time: (out): A pointer to the stream-time in the event
+ * @running_time: (out): A pointer to the running-time in the event
+ * @all_headers: (out): A pointer to the all_headers flag in the event
+ * @count: (out): A pointer to the count field of the event
+ *
+ * Get timestamp, stream-time, running-time, all-headers and count in the force
+ * key unit event. See gst_video_event_new_downstream_force_key_unit() for a
+ * full description of the downstream force key unit event.
+ *
+ * Returns: %TRUE if the event is a valid downstream force key unit event.
+ * Since: 0.10.36
+ */
+gboolean
+gst_video_event_parse_downstream_force_key_unit (GstEvent * event,
+ GstClockTime * timestamp, GstClockTime * stream_time,
+ GstClockTime * running_time, gboolean * all_headers, guint * count)
+{
+ const GstStructure *s;
+ GstClockTime ev_timestamp, ev_stream_time, ev_running_time;
+ gboolean ev_all_headers;
+ guint ev_count;
+
+ g_return_val_if_fail (event != NULL, FALSE);
+
+ if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM)
+ return FALSE; /* Not a force key unit event */
+
+ s = gst_event_get_structure (event);
+ if (s == NULL
+ || !gst_structure_has_name (s, GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME))
+ return FALSE;
+
+ if (!gst_structure_get_clock_time (s, "timestamp", &ev_timestamp))
+ ev_timestamp = GST_CLOCK_TIME_NONE;
+ if (!gst_structure_get_clock_time (s, "stream-time", &ev_stream_time))
+ ev_stream_time = GST_CLOCK_TIME_NONE;
+ if (!gst_structure_get_clock_time (s, "running-time", &ev_running_time))
+ ev_running_time = GST_CLOCK_TIME_NONE;
+ if (!gst_structure_get_boolean (s, "all-headers", &ev_all_headers))
+ ev_all_headers = FALSE;
+ if (!gst_structure_get_uint (s, "count", &ev_count))
+ ev_count = 0;
+
+ if (timestamp)
+ *timestamp = ev_timestamp;
+
+ if (stream_time)
+ *stream_time = ev_stream_time;
+
+ if (running_time)
+ *running_time = ev_running_time;
+
+ if (all_headers)
+ *all_headers = ev_all_headers;
+
+ if (count)
+ *count = ev_count;
+
+ return TRUE;
+}
+
+/**
+ * gst_video_event_parse_upstream_force_key_unit:
+ * @event: A #GstEvent to parse
+ * @running_time: (out): A pointer to the running_time in the event
+ * @all_headers: (out): A pointer to the all_headers flag in the event
+ * @count: (out): A pointer to the count field in the event
+ *
+ * Get running-time, all-headers and count in the force key unit event. See
+ * gst_video_event_new_upstream_force_key_unit() for a full description of the
+ * upstream force key unit event.
+ *
+ * Create an upstream force key unit event using gst_video_event_new_upstream_force_key_unit()
+ *
+ * Returns: %TRUE if the event is a valid upstream force-key-unit event. %FALSE if not
+ * Since: 0.10.36
+ */
+gboolean
+gst_video_event_parse_upstream_force_key_unit (GstEvent * event,
+ GstClockTime * running_time, gboolean * all_headers, guint * count)
+{
+ const GstStructure *s;
+ GstClockTime ev_running_time;
+ gboolean ev_all_headers;
+ guint ev_count;
+
+ g_return_val_if_fail (event != NULL, FALSE);
+
+ if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_UPSTREAM)
+ return FALSE; /* Not a force key unit event */
+
+ s = gst_event_get_structure (event);
+ if (s == NULL
+ || !gst_structure_has_name (s, GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME))
+ return FALSE;
+
+ if (!gst_structure_get_clock_time (s, "running-time", &ev_running_time))
+ ev_running_time = GST_CLOCK_TIME_NONE;
+ if (!gst_structure_get_boolean (s, "all-headers", &ev_all_headers))
+ ev_all_headers = FALSE;
+ if (!gst_structure_get_uint (s, "count", &ev_count))
+ ev_count = 0;
+
+ if (running_time)
+ *running_time = ev_running_time;
+
+ if (all_headers)
+ *all_headers = ev_all_headers;
+
+ if (count)
+ *count = ev_count;
+
+ return TRUE;
+}
--- /dev/null
+/* GStreamer
+ * Copyright (C) <2011> Wim Taymans <wim.taymans@gmail.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifndef __GST_VIDEO_EVENT_H__
+#define __GST_VIDEO_EVENT_H__
+
+#include <gst/gst.h>
+
+G_BEGIN_DECLS
+
+/* video still frame event creation and parsing */
+
+GstEvent * gst_video_event_new_still_frame (gboolean in_still);
+
+gboolean gst_video_event_parse_still_frame (GstEvent * event, gboolean * in_still);
+
+/* video force key unit event creation and parsing */
+
+GstEvent * gst_video_event_new_downstream_force_key_unit (GstClockTime timestamp,
+ GstClockTime stream_time,
+ GstClockTime running_time,
+ gboolean all_headers,
+ guint count);
+
+gboolean gst_video_event_parse_downstream_force_key_unit (GstEvent * event,
+ GstClockTime * timestamp,
+ GstClockTime * stream_time,
+ GstClockTime * running_time,
+ gboolean * all_headers,
+ guint * count);
+
+GstEvent * gst_video_event_new_upstream_force_key_unit (GstClockTime running_time,
+ gboolean all_headers,
+ guint count);
+
+gboolean gst_video_event_parse_upstream_force_key_unit (GstEvent * event,
+ GstClockTime * running_time,
+ gboolean * all_headers,
+ guint * count);
+
+gboolean gst_video_event_is_force_key_unit(GstEvent *event);
+
+G_END_DECLS
+
+#endif /* __GST_VIDEO_EVENT_H__ */
--- /dev/null
+/* GStreamer
+ * Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
+ * Library <2002> Ronald Bultje <rbultje@ronald.bitfreak.net>
+ * Copyright (C) 2007 David A. Schleef <ds@schleef.org>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifdef HAVE_CONFIG_H
+# include "config.h"
+#endif
+
+#include <string.h>
+#include <stdio.h>
+
+#include "video-format.h"
+
+/**
+ * SECTION:gstvideo
+ * @short_description: Support library for video operations
+ *
+ * <refsect2>
+ * <para>
+ * This library contains some helper functions and includes the
+ * videosink and videofilter base classes.
+ * </para>
+ * </refsect2>
+ */
+
+#include "videoblendorc.h"
+
+/* Line conversion to AYUV */
+
+#define GET_PLANE_STRIDE(plane) (stride(plane))
+#define GET_PLANE_LINE(plane, line) \
+ (gpointer)(((guint8*)(data[plane])) + stride[plane] * (line))
+
+#define GET_COMP_STRIDE(comp) \
+ GST_VIDEO_FORMAT_INFO_STRIDE (info, stride, comp)
+#define GET_COMP_DATA(comp) \
+ GST_VIDEO_FORMAT_INFO_DATA (info, data, comp)
+
+#define GET_COMP_LINE(comp, line) \
+ (gpointer)(((guint8*)GET_COMP_DATA (comp)) + \
+ GET_COMP_STRIDE(comp) * (line))
+
+#define GET_STRIDE() GET_PLANE_STRIDE (0)
+#define GET_LINE(line) GET_PLANE_LINE (0, line)
+
+#define GET_Y_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_Y, line)
+#define GET_U_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_U, line)
+#define GET_V_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_V, line)
+#define GET_A_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_A, line)
+
+#define GET_Y_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_Y)
+#define GET_U_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_U)
+#define GET_V_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_V)
+#define GET_A_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_A)
+
+#define PACK_420 GST_VIDEO_FORMAT_AYUV, unpack_planar_420, pack_planar_420
+static void
+unpack_planar_420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_I420 (dest,
+ GET_Y_LINE (y), GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), width);
+}
+
+static void
+pack_planar_420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_I420 (GET_Y_LINE (y),
+ GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), src, width / 2);
+}
+
+#define PACK_YUY2 GST_VIDEO_FORMAT_AYUV, unpack_YUY2, pack_YUY2
+static void
+unpack_YUY2 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_YUY2 (dest, GET_LINE (y), width / 2);
+}
+
+static void
+pack_YUY2 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_YUY2 (GET_LINE (y), src, width / 2);
+}
+
+#define PACK_UYVY GST_VIDEO_FORMAT_AYUV, unpack_UYVY, pack_UYVY
+static void
+unpack_UYVY (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_UYVY (dest, GET_LINE (y), width / 2);
+}
+
+static void
+pack_UYVY (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_UYVY (GET_LINE (y), src, width / 2);
+}
+
+#define PACK_YVYU GST_VIDEO_FORMAT_AYUV, unpack_YVYU, pack_YVYU
+static void
+unpack_YVYU (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_YVYU (dest, GET_LINE (y), width / 2);
+}
+
+static void
+pack_YVYU (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_YVYU (GET_LINE (y), src, width / 2);
+}
+
+#define PACK_v308 GST_VIDEO_FORMAT_AYUV, unpack_v308, pack_v308
+static void
+unpack_v308 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 1] = s[i * 3 + 0];
+ d[i * 4 + 2] = s[i * 3 + 1];
+ d[i * 4 + 3] = s[i * 3 + 2];
+ }
+}
+
+static void
+pack_v308 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i * 3 + 0] = s[i * 4 + 1];
+ d[i * 3 + 1] = s[i * 4 + 2];
+ d[i * 3 + 2] = s[i * 4 + 3];
+ }
+}
+
+#define PACK_AYUV GST_VIDEO_FORMAT_AYUV, unpack_copy4, pack_copy4
+#define PACK_ARGB GST_VIDEO_FORMAT_ARGB, unpack_copy4, pack_copy4
+static void
+unpack_copy4 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ memcpy (dest, GET_LINE (y), width * 4);
+}
+
+static void
+pack_copy4 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ memcpy (GET_LINE (y), src, width * 4);
+}
+
+#define PACK_v210 GST_VIDEO_FORMAT_AYUV64, unpack_v210, pack_v210
+static void
+unpack_v210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i += 6) {
+ guint32 a0, a1, a2, a3;
+ guint16 y0, y1, y2, y3, y4, y5;
+ guint16 u0, u2, u4;
+ guint16 v0, v2, v4;
+
+ a0 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 0);
+ a1 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 4);
+ a2 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 8);
+ a3 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 12);
+
+ u0 = ((a0 >> 0) & 0x3ff) << 6;
+ y0 = ((a0 >> 10) & 0x3ff) << 6;
+ v0 = ((a0 >> 20) & 0x3ff) << 6;
+ y1 = ((a1 >> 0) & 0x3ff) << 6;
+
+ u2 = ((a1 >> 10) & 0x3ff) << 6;
+ y2 = ((a1 >> 20) & 0x3ff) << 6;
+ v2 = ((a2 >> 0) & 0x3ff) << 6;
+ y3 = ((a2 >> 10) & 0x3ff) << 6;
+
+ u4 = ((a2 >> 20) & 0x3ff) << 6;
+ y4 = ((a3 >> 0) & 0x3ff) << 6;
+ v4 = ((a3 >> 10) & 0x3ff) << 6;
+ y5 = ((a3 >> 20) & 0x3ff) << 6;
+
+ d[4 * (i + 0) + 0] = 0xffff;
+ d[4 * (i + 0) + 1] = y0;
+ d[4 * (i + 0) + 2] = u0;
+ d[4 * (i + 0) + 3] = v0;
+
+ d[4 * (i + 1) + 0] = 0xffff;
+ d[4 * (i + 1) + 1] = y1;
+ d[4 * (i + 1) + 2] = u0;
+ d[4 * (i + 1) + 3] = v0;
+
+ d[4 * (i + 2) + 0] = 0xffff;
+ d[4 * (i + 2) + 1] = y2;
+ d[4 * (i + 2) + 2] = u2;
+ d[4 * (i + 2) + 3] = v2;
+
+ d[4 * (i + 3) + 0] = 0xffff;
+ d[4 * (i + 3) + 1] = y3;
+ d[4 * (i + 3) + 2] = u2;
+ d[4 * (i + 3) + 3] = v2;
+
+ d[4 * (i + 4) + 0] = 0xffff;
+ d[4 * (i + 4) + 1] = y4;
+ d[4 * (i + 4) + 2] = u4;
+ d[4 * (i + 4) + 3] = v4;
+
+ d[4 * (i + 5) + 0] = 0xffff;
+ d[4 * (i + 5) + 1] = y5;
+ d[4 * (i + 5) + 2] = u4;
+ d[4 * (i + 5) + 3] = v4;
+ }
+}
+
+static void
+pack_v210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint16 *s = src;
+
+ for (i = 0; i < width; i += 6) {
+ guint32 a0, a1, a2, a3;
+ guint16 y0, y1, y2, y3, y4, y5;
+ guint16 u0, u1, u2;
+ guint16 v0, v1, v2;
+
+ y0 = s[4 * (i + 0) + 1] >> 6;
+ y1 = s[4 * (i + 1) + 1] >> 6;
+ y2 = s[4 * (i + 2) + 1] >> 6;
+ y3 = s[4 * (i + 3) + 1] >> 6;
+ y4 = s[4 * (i + 4) + 1] >> 6;
+ y5 = s[4 * (i + 5) + 1] >> 6;
+
+ u0 = (s[4 * (i + 0) + 2] + s[4 * (i + 1) + 2] + 1) >> 7;
+ u1 = (s[4 * (i + 2) + 2] + s[4 * (i + 3) + 2] + 1) >> 7;
+ u2 = (s[4 * (i + 4) + 2] + s[4 * (i + 5) + 2] + 1) >> 7;
+
+ v0 = (s[4 * (i + 0) + 3] + s[4 * (i + 1) + 3] + 1) >> 7;
+ v1 = (s[4 * (i + 2) + 3] + s[4 * (i + 3) + 3] + 1) >> 7;
+ v2 = (s[4 * (i + 4) + 3] + s[4 * (i + 5) + 3] + 1) >> 7;
+
+ a0 = u0 | (y0 << 10) | (v0 << 20);
+ a1 = y1 | (u1 << 10) | (y2 << 20);
+ a2 = v1 | (y3 << 10) | (u2 << 20);
+ a3 = y4 | (v2 << 10) | (y5 << 20);
+
+ GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 0, a0);
+ GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 4, a1);
+ GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 8, a2);
+ GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 12, a3);
+ }
+}
+
+#define PACK_v216 GST_VIDEO_FORMAT_AYUV64, unpack_v216, pack_v216
+static void
+unpack_v216 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xffff;
+ d[i * 4 + 1] = GST_READ_UINT16_LE (s + i * 4 + 2);
+ d[i * 4 + 2] = GST_READ_UINT16_LE (s + (i >> 1) * 8 + 0);
+ d[i * 4 + 3] = GST_READ_UINT16_LE (s + (i >> 1) * 8 + 4);
+ }
+}
+
+static void
+pack_v216 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint16 *s = src;
+
+ for (i = 0; i < width / 2; i++) {
+ GST_WRITE_UINT16_LE (d + i * 8 + 0, s[(i * 2 + 0) * 4 + 2]);
+ GST_WRITE_UINT16_LE (d + i * 8 + 2, s[(i * 2 + 0) * 4 + 1]);
+ GST_WRITE_UINT16_LE (d + i * 8 + 4, s[(i * 2 + 0) * 4 + 3]);
+ GST_WRITE_UINT16_LE (d + i * 8 + 6, s[(i * 2 + 1) * 4 + 1]);
+ }
+}
+
+#define PACK_Y41B GST_VIDEO_FORMAT_AYUV, unpack_Y41B, pack_Y41B
+static void
+unpack_Y41B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_YUV9 (dest,
+ GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), width / 2);
+}
+
+static void
+pack_Y41B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *destY = GET_Y_LINE (y);
+ guint8 *destU = GET_U_LINE (y);
+ guint8 *destV = GET_V_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width - 3; i += 4) {
+ destY[i] = s[i * 4 + 1];
+ destY[i + 1] = s[i * 4 + 5];
+ destY[i + 2] = s[i * 4 + 9];
+ destY[i + 3] = s[i * 4 + 13];
+
+ destU[i >> 2] =
+ (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + s[i * 4 + 14] + 2) >> 2;
+ destV[i >> 2] =
+ (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + s[i * 4 + 15] + 2) >> 2;
+ }
+
+ if (i == width - 3) {
+ destY[i] = s[i * 4 + 1];
+ destY[i + 1] = s[i * 4 + 5];
+ destY[i + 2] = s[i * 4 + 9];
+
+ destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + 1) / 3;
+ destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + 1) / 3;
+ } else if (i == width - 2) {
+ destY[i] = s[i * 4 + 1];
+ destY[i + 1] = s[i * 4 + 5];
+
+ destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6] + 1) >> 1;
+ destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7] + 1) >> 1;
+ } else if (i == width - 1) {
+ destY[i + 1] = s[i * 4 + 5];
+
+ destU[i >> 2] = s[i * 4 + 2];
+ destV[i >> 2] = s[i * 4 + 3];
+ }
+}
+
+#define PACK_Y42B GST_VIDEO_FORMAT_AYUV, unpack_Y42B, pack_Y42B
+static void
+unpack_Y42B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_Y42B (dest,
+ GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), width / 2);
+}
+
+static void
+pack_Y42B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_Y42B (GET_Y_LINE (y),
+ GET_U_LINE (y), GET_V_LINE (y), src, width / 2);
+}
+
+#define PACK_Y444 GST_VIDEO_FORMAT_AYUV, unpack_Y444, pack_Y444
+static void
+unpack_Y444 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_Y444 (dest, GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), width);
+}
+
+static void
+pack_Y444 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_Y444 (GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), src, width);
+}
+
+#define PACK_GRAY8 GST_VIDEO_FORMAT_AYUV, unpack_GRAY8, pack_GRAY8
+static void
+unpack_GRAY8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_GRAY8 (dest, GET_LINE (y), width);
+}
+
+static void
+pack_GRAY8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_GRAY8 (GET_LINE (y), src, width);
+}
+
+#define PACK_GRAY16_BE GST_VIDEO_FORMAT_AYUV64, unpack_GRAY16_BE, pack_GRAY16_BE
+static void
+unpack_GRAY16_BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint16 *s = GET_LINE (y);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xffff;
+ d[i * 4 + 1] = GST_READ_UINT16_BE (s + i);
+ d[i * 4 + 2] = 0x8000;
+ d[i * 4 + 3] = 0x8000;
+ }
+}
+
+static void
+pack_GRAY16_BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *d = GET_LINE (y);
+ const guint16 *s = src;
+
+ for (i = 0; i < width; i++) {
+ GST_WRITE_UINT16_BE (d + i, s[i * 4 + 1]);
+ }
+}
+
+#define PACK_GRAY16_LE GST_VIDEO_FORMAT_AYUV64, unpack_GRAY16_LE, pack_GRAY16_LE
+static void
+unpack_GRAY16_LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint16 *s = GET_LINE (y);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xffff;
+ d[i * 4 + 1] = GST_READ_UINT16_LE (s + i);
+ d[i * 4 + 2] = 0x8000;
+ d[i * 4 + 3] = 0x8000;
+ }
+}
+
+static void
+pack_GRAY16_LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *d = GET_LINE (y);
+ const guint16 *s = src;
+
+ for (i = 0; i < width; i++) {
+ GST_WRITE_UINT16_LE (d + i, s[i * 4 + 1]);
+ }
+}
+
+#define PACK_RGB16 GST_VIDEO_FORMAT_ARGB, unpack_RGB16, pack_RGB16
+static void
+unpack_RGB16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint16 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 1] = ((s[i] >> 11) & 0x1f) << 3;
+ d[i * 4 + 2] = ((s[i] >> 5) & 0x3f) << 2;
+ d[i * 4 + 3] = ((s[i]) & 0x1f) << 3;
+ }
+}
+
+static void
+pack_RGB16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i] = ((s[i * 4 + 1] >> 3) << 11) |
+ ((s[i * 4 + 2] >> 2) << 5) | (s[i * 4 + 3] >> 3);
+ }
+}
+
+#define PACK_BGR16 GST_VIDEO_FORMAT_ARGB, unpack_BGR16, pack_BGR16
+static void
+unpack_BGR16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint16 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 3] = ((s[i] >> 11) & 0x1f) << 3;
+ d[i * 4 + 2] = ((s[i] >> 5) & 0x3f) << 2;
+ d[i * 4 + 1] = ((s[i]) & 0x1f) << 3;
+ }
+}
+
+static void
+pack_BGR16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i] = ((s[i * 4 + 3] >> 3) << 11) |
+ ((s[i * 4 + 2] >> 2) << 5) | (s[i * 4 + 1] >> 3);
+ }
+}
+
+#define PACK_RGB15 GST_VIDEO_FORMAT_ARGB, unpack_RGB15, pack_RGB15
+static void
+unpack_RGB15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint16 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 1] = ((s[i] >> 10) & 0x1f) << 3;
+ d[i * 4 + 2] = ((s[i] >> 5) & 0x1f) << 3;
+ d[i * 4 + 3] = ((s[i]) & 0x1f) << 3;
+ }
+}
+
+static void
+pack_RGB15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i] = ((s[i * 4 + 1] >> 3) << 10) |
+ ((s[i * 4 + 2] >> 3) << 5) | (s[i * 4 + 3] >> 3);
+ }
+}
+
+#define PACK_BGR15 GST_VIDEO_FORMAT_ARGB, unpack_BGR15, pack_BGR15
+static void
+unpack_BGR15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint16 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 3] = ((s[i] >> 10) & 0x1f) << 3;
+ d[i * 4 + 2] = ((s[i] >> 5) & 0x1f) << 3;
+ d[i * 4 + 1] = ((s[i]) & 0x1f) << 3;
+ }
+}
+
+static void
+pack_BGR15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i] = ((s[i * 4 + 3] >> 3) << 10) |
+ ((s[i * 4 + 2] >> 3) << 5) | (s[i * 4 + 1] >> 3);
+ }
+}
+
+#define PACK_BGRA GST_VIDEO_FORMAT_ARGB, unpack_BGRA, pack_BGRA
+static void
+unpack_BGRA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_BGRA (dest, GET_LINE (y), width);
+}
+
+static void
+pack_BGRA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_BGRA (GET_LINE (y), src, width);
+}
+
+#define PACK_ABGR GST_VIDEO_FORMAT_ARGB, unpack_ABGR, pack_ABGR
+static void
+unpack_ABGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_ABGR (dest, GET_LINE (y), width);
+}
+
+static void
+pack_ABGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_ABGR (GET_LINE (y), src, width);
+}
+
+#define PACK_RGBA GST_VIDEO_FORMAT_ARGB, unpack_RGBA, pack_RGBA
+static void
+unpack_RGBA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_RGBA (dest, GET_LINE (y), width);
+}
+
+static void
+pack_RGBA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_RGBA (GET_LINE (y), src, width);
+}
+
+#define PACK_RGB GST_VIDEO_FORMAT_ARGB, unpack_RGB, pack_RGB
+static void
+unpack_RGB (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 1] = s[i * 3 + 0];
+ d[i * 4 + 2] = s[i * 3 + 1];
+ d[i * 4 + 3] = s[i * 3 + 2];
+ }
+}
+
+static void
+pack_RGB (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i * 3 + 0] = s[i * 4 + 1];
+ d[i * 3 + 1] = s[i * 4 + 2];
+ d[i * 3 + 2] = s[i * 4 + 3];
+ }
+}
+
+#define PACK_BGR GST_VIDEO_FORMAT_ARGB, unpack_BGR, pack_BGR
+static void
+unpack_BGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 1] = s[i * 3 + 2];
+ d[i * 4 + 2] = s[i * 3 + 1];
+ d[i * 4 + 3] = s[i * 3 + 0];
+ }
+}
+
+static void
+pack_BGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width; i++) {
+ d[i * 3 + 0] = s[i * 4 + 3];
+ d[i * 3 + 1] = s[i * 4 + 2];
+ d[i * 3 + 2] = s[i * 4 + 1];
+ }
+}
+
+#define PACK_NV12 GST_VIDEO_FORMAT_AYUV, unpack_NV12, pack_NV12
+static void
+unpack_NV12 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_NV12 (dest,
+ GET_PLANE_LINE (0, y), GET_PLANE_LINE (1, y >> 1), width / 2);
+}
+
+static void
+pack_NV12 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_NV12 (GET_PLANE_LINE (0, y),
+ GET_PLANE_LINE (1, y >> 1), src, width / 2);
+}
+
+#define PACK_NV21 GST_VIDEO_FORMAT_AYUV, unpack_NV21, pack_NV21
+static void
+unpack_NV21 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_NV21 (dest,
+ GET_PLANE_LINE (0, y), GET_PLANE_LINE (1, y >> 1), width / 2);
+}
+
+static void
+pack_NV21 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_NV21 (GET_PLANE_LINE (0, y),
+ GET_PLANE_LINE (1, y >> 1), src, width / 2);
+}
+
+#define PACK_UYVP GST_VIDEO_FORMAT_AYUV64, unpack_UYVP, pack_UYVP
+static void
+unpack_UYVP (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i += 2) {
+ guint16 y0, y1;
+ guint16 u0;
+ guint16 v0;
+
+ u0 = (s[(i / 2) * 5 + 0] << 2) | (s[(i / 2) * 5 + 1] >> 6);
+ y0 = ((s[(i / 2) * 5 + 1] & 0x3f) << 4) | (s[(i / 2) * 5 + 2] >> 4);
+ v0 = ((s[(i / 2) * 5 + 2] & 0x0f) << 6) | (s[(i / 2) * 5 + 3] >> 2);
+ y1 = ((s[(i / 2) * 5 + 3] & 0x03) << 8) | s[(i / 2) * 5 + 4];
+
+ d[i * 4 + 0] = 0xffff;
+ d[i * 4 + 1] = y0 << 6;
+ d[i * 4 + 2] = u0 << 6;
+ d[i * 4 + 3] = v0 << 6;
+ d[i * 4 + 4] = 0xffff;
+ d[i * 4 + 5] = y1 << 6;
+ d[i * 4 + 6] = u0 << 6;
+ d[i * 4 + 7] = v0 << 6;
+ }
+}
+
+static void
+pack_UYVP (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint16 *s = src;
+
+ for (i = 0; i < width; i += 2) {
+ guint16 y0, y1;
+ guint16 u0;
+ guint16 v0;
+
+ y0 = s[4 * (i + 0) + 1];
+ y1 = s[4 * (i + 1) + 1];
+ u0 = (s[4 * (i + 0) + 2] + s[4 * (i + 1) + 2] + 1) >> 1;
+ v0 = (s[4 * (i + 0) + 3] + s[4 * (i + 1) + 3] + 1) >> 1;
+
+ d[(i / 2) * 5 + 0] = u0 >> 8;
+ d[(i / 2) * 5 + 1] = (u0 & 0xc0) | y0 >> 10;
+ d[(i / 2) * 5 + 2] = ((y0 & 0x3c0) >> 2) | (v0 >> 12);
+ d[(i / 2) * 5 + 3] = ((v0 & 0xfc0) >> 4) | (y1 >> 14);
+ d[(i / 2) * 5 + 4] = (y1 >> 6);
+ }
+}
+
+#define PACK_A420 GST_VIDEO_FORMAT_AYUV, unpack_A420, pack_A420
+static void
+unpack_A420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_A420 (dest,
+ GET_Y_LINE (y), GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), GET_A_LINE (y),
+ width);
+}
+
+static void
+pack_A420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ orc_pack_A420 (GET_Y_LINE (y),
+ GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), GET_A_LINE (y), src, width / 2);
+}
+
+#define PACK_RGB8P GST_VIDEO_FORMAT_ARGB, unpack_RGB8P, pack_RGB8P
+static void
+unpack_RGB8P (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ const guint32 *p = data[1];
+ guint8 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ guint32 v = p[s[i]];
+ d[i * 4 + 0] = (v >> 24) & 0xff;
+ d[i * 4 + 1] = (v >> 16) & 0xff;
+ d[i * 4 + 2] = (v >> 8) & 0xff;
+ d[i * 4 + 3] = (v) & 0xff;
+ }
+}
+
+static void
+pack_RGB8P (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ /* Use our poor man's palette, taken from ffmpegcolorspace too */
+ for (i = 0; i < width; i++) {
+ /* crude approximation for alpha ! */
+ if (s[i * 4 + 0] < 0x80)
+ d[i] = 6 * 6 * 6;
+ else
+ d[i] =
+ ((((s[i * 4 + 1]) / 47) % 6) * 6 * 6 + (((s[i * 4 +
+ 2]) / 47) % 6) * 6 + (((s[i * 4 + 3]) / 47) % 6));
+ }
+}
+
+#define PACK_410 GST_VIDEO_FORMAT_AYUV, unpack_410, pack_410
+static void
+unpack_410 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ orc_unpack_YUV9 (dest,
+ GET_Y_LINE (y), GET_U_LINE (y >> 2), GET_V_LINE (y >> 2), width / 2);
+}
+
+static void
+pack_410 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *destY = GET_Y_LINE (y);
+ guint8 *destU = GET_U_LINE (y >> 2);
+ guint8 *destV = GET_V_LINE (y >> 2);
+ const guint8 *s = src;
+
+ for (i = 0; i < width - 3; i += 4) {
+ destY[i] = s[i * 4 + 1];
+ destY[i + 1] = s[i * 4 + 5];
+ destY[i + 2] = s[i * 4 + 9];
+ destY[i + 3] = s[i * 4 + 13];
+ if (y % 4 == 0) {
+ destU[i >> 2] =
+ (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + s[i * 4 + 14]) >> 2;
+ destV[i >> 2] =
+ (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + s[i * 4 + 15]) >> 2;
+ }
+ }
+
+ if (i == width - 3) {
+ destY[i] = s[i * 4 + 1];
+ destY[i + 1] = s[i * 4 + 5];
+ destY[i + 2] = s[i * 4 + 9];
+ if (y % 4 == 0) {
+ destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10]) / 3;
+ destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11]) / 3;
+ }
+ } else if (i == width - 2) {
+ destY[i] = s[i * 4 + 1];
+ destY[i + 1] = s[i * 4 + 5];
+ if (y % 4 == 0) {
+ destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6]) >> 1;
+ destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7]) >> 1;
+ }
+ } else if (i == width - 1) {
+ destY[i] = s[i * 4 + 1];
+ destU[i >> 2] = s[i * 4 + 2];
+ destV[i >> 2] = s[i * 4 + 3];
+ }
+}
+
+#define PACK_IYU1 GST_VIDEO_FORMAT_AYUV, unpack_IYU1, pack_IYU1
+static void
+unpack_IYU1 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint8 *d = dest;
+
+ for (i = 0; i < width - 3; i += 4) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 4] = 0xff;
+ d[i * 4 + 8] = 0xff;
+ d[i * 4 + 12] = 0xff;
+ d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
+ d[i * 4 + 5] = s[(i >> 2) * 6 + 2];
+ d[i * 4 + 9] = s[(i >> 2) * 6 + 4];
+ d[i * 4 + 13] = s[(i >> 2) * 6 + 5];
+ d[i * 4 + 2] = d[i * 4 + 6] = d[i * 4 + 10] = d[i * 4 + 14] =
+ s[(i >> 2) * 6 + 0];
+ d[i * 4 + 3] = d[i * 4 + 7] = d[i * 4 + 11] = d[i * 4 + 15] =
+ s[(i >> 2) * 6 + 3];
+ }
+
+ if (i == width - 3) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 4] = 0xff;
+ d[i * 4 + 8] = 0xff;
+ d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
+ d[i * 4 + 5] = s[(i >> 2) * 6 + 2];
+ d[i * 4 + 9] = s[(i >> 2) * 6 + 4];
+ d[i * 4 + 2] = d[i * 4 + 6] = d[i * 4 + 10] = s[(i >> 2) * 6 + 0];
+ d[i * 4 + 3] = d[i * 4 + 7] = d[i * 4 + 11] = s[(i >> 2) * 6 + 3];
+ } else if (i == width - 2) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 4] = 0xff;
+ d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
+ d[i * 4 + 5] = s[(i >> 2) * 6 + 2];
+ d[i * 4 + 2] = d[i * 4 + 6] = s[(i >> 2) * 6 + 0];
+ d[i * 4 + 3] = d[i * 4 + 7] = s[(i >> 2) * 6 + 3];
+ } else if (i == width - 1) {
+ d[i * 4 + 0] = 0xff;
+ d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
+ d[i * 4 + 2] = s[(i >> 2) * 6 + 0];
+ d[i * 4 + 3] = s[(i >> 2) * 6 + 3];
+ }
+}
+
+static void
+pack_IYU1 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint8 *s = src;
+
+ for (i = 0; i < width - 3; i += 4) {
+ d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
+ d[(i >> 2) * 6 + 2] = s[i * 4 + 5];
+ d[(i >> 2) * 6 + 4] = s[i * 4 + 9];
+ d[(i >> 2) * 6 + 5] = s[i * 4 + 13];
+ d[(i >> 2) * 6 + 0] =
+ (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + s[i * 4 + 14]) >> 2;
+ d[(i >> 2) * 6 + 3] =
+ (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + s[i * 4 + 15]) >> 2;
+ }
+
+ if (i == width - 3) {
+ d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
+ d[(i >> 2) * 6 + 2] = s[i * 4 + 5];
+ d[(i >> 2) * 6 + 4] = s[i * 4 + 9];
+ d[(i >> 2) * 6 + 0] = (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10]) / 3;
+ d[(i >> 2) * 6 + 3] = (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11]) / 3;
+ } else if (i == width - 2) {
+ d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
+ d[(i >> 2) * 6 + 2] = s[i * 4 + 5];
+ d[(i >> 2) * 6 + 0] = (s[i * 4 + 2] + s[i * 4 + 6]) >> 1;
+ d[(i >> 2) * 6 + 3] = (s[i * 4 + 3] + s[i * 4 + 7]) >> 1;
+ } else if (i == width - 1) {
+ d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
+ d[(i >> 2) * 6 + 0] = s[i * 4 + 2];
+ d[(i >> 2) * 6 + 3] = s[i * 4 + 3];
+ }
+}
+
+#define PACK_ARGB64 GST_VIDEO_FORMAT_ARGB64, unpack_copy8, pack_copy8
+#define PACK_AYUV64 GST_VIDEO_FORMAT_AYUV64, unpack_copy8, pack_copy8
+static void
+unpack_copy8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ memcpy (dest, GET_LINE (y), width * 8);
+}
+
+static void
+pack_copy8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ memcpy (GET_LINE (y), src, width * 8);
+}
+
+#define PACK_r210 GST_VIDEO_FORMAT_AYUV64, unpack_r210, pack_r210
+static void
+unpack_r210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ const guint8 *s = GET_LINE (y);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ guint32 x;
+ d[i * 4 + 0] = 0xffff;
+ x = GST_READ_UINT32_BE (s + i * 4);
+ d[i * 4 + 1] = ((x >> 14) & 0xffc0) | (x >> 24);
+ d[i * 4 + 2] = ((x >> 4) & 0xffc0) | ((x >> 14) & 0x3f);
+ d[i * 4 + 3] = ((x << 6) & 0xffc0) | ((x >> 4) & 0x3f);
+ }
+}
+
+static void
+pack_r210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint8 *d = GET_LINE (y);
+ const guint16 *s = src;
+
+ for (i = 0; i < width; i++) {
+ guint32 x = 0;
+ x |= (s[i * 4 + 1] & 0xffc0) << 14;
+ x |= (s[i * 4 + 2] & 0xffc0) << 4;
+ x |= (s[i * 4 + 3] & 0xffc0) >> 6;
+ GST_WRITE_UINT32_BE (d + i * 4, x);
+ }
+}
+
+#define PACK_I420_10LE GST_VIDEO_FORMAT_AYUV64, unpack_I420_10LE, pack_I420_10LE
+static void
+unpack_I420_10LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ guint16 *srcY = GET_Y_LINE (y);
+ guint16 *srcU = GET_U_LINE (y >> 1);
+ guint16 *srcV = GET_V_LINE (y >> 1);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xffff;
+ d[i * 4 + 1] = GST_READ_UINT16_LE (srcY + i) << 6;
+ d[i * 4 + 2] = GST_READ_UINT16_LE (srcU + (i >> 1)) << 6;
+ d[i * 4 + 3] = GST_READ_UINT16_LE (srcV + (i >> 1)) << 6;
+ }
+}
+
+static void
+pack_I420_10LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *destY = GET_Y_LINE (y);
+ guint16 *destU = GET_U_LINE (y >> 1);
+ guint16 *destV = GET_V_LINE (y >> 1);
+ guint16 Y0, Y1, U, V;
+ const guint16 *s = src;
+
+ for (i = 0; i < width - 1; i += 2) {
+ Y0 = (s[i * 4 + 1]) >> 6;
+ Y1 = (s[i * 4 + 5]) >> 6;
+ U = ((s[i * 4 + 2] + s[i * 4 + 6] + 1) >> 1) >> 6;
+ V = ((s[i * 4 + 3] + s[i * 4 + 7] + 1) >> 1) >> 6;
+
+ GST_WRITE_UINT16_LE (destY + i + 0, Y0);
+ GST_WRITE_UINT16_LE (destY + i + 1, Y1);
+ GST_WRITE_UINT16_LE (destU + (i >> 1), U);
+ GST_WRITE_UINT16_LE (destV + (i >> 1), V);
+ }
+ if (i == width - 1) {
+ Y0 = s[i * 4 + 1] >> 6;
+ U = s[i * 4 + 2] >> 6;
+ V = s[i * 4 + 3] >> 6;
+
+ GST_WRITE_UINT16_LE (destY + i, Y0);
+ GST_WRITE_UINT16_LE (destU + (i >> 1), U);
+ GST_WRITE_UINT16_LE (destV + (i >> 1), V);
+ }
+}
+
+#define PACK_I420_10BE GST_VIDEO_FORMAT_AYUV64, unpack_I420_10BE, pack_I420_10BE
+static void
+unpack_I420_10BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
+{
+ int i;
+ guint16 *srcY = GET_Y_LINE (y);
+ guint16 *srcU = GET_U_LINE (y >> 1);
+ guint16 *srcV = GET_V_LINE (y >> 1);
+ guint16 *d = dest;
+
+ for (i = 0; i < width; i++) {
+ d[i * 4 + 0] = 0xffff;
+ d[i * 4 + 1] = GST_READ_UINT16_BE (srcY + i) << 6;
+ d[i * 4 + 2] = GST_READ_UINT16_BE (srcU + (i >> 1)) << 6;
+ d[i * 4 + 3] = GST_READ_UINT16_BE (srcV + (i >> 1)) << 6;
+ }
+}
+
+static void
+pack_I420_10BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
+ const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
+ gint y, gint width)
+{
+ int i;
+ guint16 *destY = GET_Y_LINE (y);
+ guint16 *destU = GET_U_LINE (y >> 1);
+ guint16 *destV = GET_V_LINE (y >> 1);
+ guint16 Y0, Y1, U, V;
+ const guint16 *s = src;
+
+ for (i = 0; i < width - 1; i += 2) {
+ Y0 = s[i * 4 + 1] >> 6;
+ Y1 = s[i * 4 + 5] >> 6;
+ U = ((s[i * 4 + 2] + s[i * 4 + 6] + 1) >> 1) >> 6;
+ V = ((s[i * 4 + 3] + s[i * 4 + 7] + 1) >> 1) >> 6;
+
+ GST_WRITE_UINT16_BE (destY + i + 0, Y0);
+ GST_WRITE_UINT16_BE (destY + i + 1, Y1);
+ GST_WRITE_UINT16_BE (destU + (i >> 1), U);
+ GST_WRITE_UINT16_BE (destV + (i >> 1), V);
+ }
+ if (i == width - 1) {
+ Y0 = s[i * 4 + 1] >> 6;
+ U = s[i * 4 + 2] >> 6;
+ V = s[i * 4 + 3] >> 6;
+
+ GST_WRITE_UINT16_BE (destY + i, Y0);
+ GST_WRITE_UINT16_BE (destU + (i >> 1), U);
+ GST_WRITE_UINT16_BE (destV + (i >> 1), V);
+ }
+}
+
+typedef struct
+{
+ guint32 fourcc;
+ GstVideoFormatInfo info;
+} VideoFormat;
+
+/* depths: bits, n_components, shift, depth */
+#define DPTH0 0, 0, { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
+#define DPTH8 8, 1, { 0, 0, 0, 0 }, { 8, 0, 0, 0 }
+#define DPTH8_32 8, 2, { 0, 0, 0, 0 }, { 8, 32, 0, 0 }
+#define DPTH888 8, 3, { 0, 0, 0, 0 }, { 8, 8, 8, 0 }
+#define DPTH8888 8, 4, { 0, 0, 0, 0 }, { 8, 8, 8, 8 }
+#define DPTH10_10_10 10, 3, { 0, 0, 0, 0 }, { 10, 10, 10, 0 }
+#define DPTH16 16, 1, { 0, 0, 0, 0 }, { 16, 0, 0, 0 }
+#define DPTH16_16_16 16, 3, { 0, 0, 0, 0 }, { 16, 16, 16, 0 }
+#define DPTH16_16_16_16 16, 4, { 0, 0, 0, 0 }, { 16, 16, 16, 16 }
+#define DPTH555 16, 3, { 10, 5, 0, 0 }, { 5, 5, 5, 0 }
+#define DPTH565 16, 3, { 11, 5, 0, 0 }, { 5, 6, 5, 0 }
+
+/* pixel strides */
+#define PSTR0 { 0, 0, 0, 0 }
+#define PSTR1 { 1, 0, 0, 0 }
+#define PSTR14 { 1, 4, 0, 0 }
+#define PSTR111 { 1, 1, 1, 0 }
+#define PSTR1111 { 1, 1, 1, 1 }
+#define PSTR122 { 1, 2, 2, 0 }
+#define PSTR2 { 2, 0, 0, 0 }
+#define PSTR222 { 2, 2, 2, 0 }
+#define PSTR244 { 2, 4, 4, 0 }
+#define PSTR444 { 4, 4, 4, 0 }
+#define PSTR4444 { 4, 4, 4, 4 }
+#define PSTR333 { 3, 3, 3, 0 }
+#define PSTR488 { 4, 8, 8, 0 }
+#define PSTR8888 { 8, 8, 8, 8 }
+
+/* planes */
+#define PLANE_NA 0, { 0, 0, 0, 0 }
+#define PLANE0 1, { 0, 0, 0, 0 }
+#define PLANE01 2, { 0, 1, 0, 0 }
+#define PLANE011 2, { 0, 1, 1, 0 }
+#define PLANE012 3, { 0, 1, 2, 0 }
+#define PLANE0123 4, { 0, 1, 2, 3 }
+#define PLANE021 3, { 0, 2, 1, 0 }
+
+/* offsets */
+#define OFFS0 { 0, 0, 0, 0 }
+#define OFFS013 { 0, 1, 3, 0 }
+#define OFFS102 { 1, 0, 2, 0 }
+#define OFFS1230 { 1, 2, 3, 0 }
+#define OFFS012 { 0, 1, 2, 0 }
+#define OFFS210 { 2, 1, 0, 0 }
+#define OFFS123 { 1, 2, 3, 0 }
+#define OFFS321 { 3, 2, 1, 0 }
+#define OFFS0123 { 0, 1, 2, 3 }
+#define OFFS2103 { 2, 1, 0, 3 }
+#define OFFS3210 { 3, 2, 1, 0 }
+#define OFFS031 { 0, 3, 1, 0 }
+#define OFFS204 { 2, 0, 4, 0 }
+#define OFFS001 { 0, 0, 1, 0 }
+#define OFFS010 { 0, 1, 0, 0 }
+#define OFFS104 { 1, 0, 4, 0 }
+#define OFFS2460 { 2, 4, 6, 0 }
+
+/* subsampling */
+#define SUB410 { 0, 2, 2, 0 }, { 0, 2, 2, 0 }
+#define SUB411 { 0, 2, 2, 0 }, { 0, 0, 0, 0 }
+#define SUB420 { 0, 1, 1, 0 }, { 0, 1, 1, 0 }
+#define SUB422 { 0, 1, 1, 0 }, { 0, 0, 0, 0 }
+#define SUB4 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
+#define SUB44 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
+#define SUB444 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
+#define SUB4444 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
+#define SUB4204 { 0, 1, 1, 0 }, { 0, 1, 1, 0 }
+
+#define MAKE_YUV_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack ) \
+ { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_YUV_LE_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack ) \
+ { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV | GST_VIDEO_FORMAT_FLAG_LE, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_YUVA_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack) \
+ { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV | GST_VIDEO_FORMAT_FLAG_ALPHA, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_YUV_C_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack) \
+ { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV | GST_VIDEO_FORMAT_FLAG_COMPLEX, depth, pstride, plane, offs, sub, pack } }
+
+#define MAKE_RGB_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
+ { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_RGB_LE_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
+ { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB | GST_VIDEO_FORMAT_FLAG_LE, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_RGBA_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
+ { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB | GST_VIDEO_FORMAT_FLAG_ALPHA, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_RGBAP_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
+ { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB | GST_VIDEO_FORMAT_FLAG_ALPHA | GST_VIDEO_FORMAT_FLAG_PALETTE, depth, pstride, plane, offs, sub, pack } }
+
+#define MAKE_GRAY_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
+ { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_GRAY, depth, pstride, plane, offs, sub, pack } }
+#define MAKE_GRAY_LE_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
+ { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_GRAY | GST_VIDEO_FORMAT_FLAG_LE, depth, pstride, plane, offs, sub, pack } }
+
+static VideoFormat formats[] = {
+ {0x00000000, {GST_VIDEO_FORMAT_UNKNOWN, "UNKNOWN", "unknown video", 0, DPTH0,
+ PSTR0, PLANE_NA,
+ OFFS0}},
+ {0x00000000, {GST_VIDEO_FORMAT_ENCODED, "ENCODED", "encoded video",
+ GST_VIDEO_FORMAT_FLAG_COMPLEX, DPTH0, PSTR0, PLANE_NA, OFFS0}},
+
+ MAKE_YUV_FORMAT (I420, "raw video", GST_MAKE_FOURCC ('I', '4', '2', '0'),
+ DPTH888, PSTR111,
+ PLANE012, OFFS0, SUB420, PACK_420),
+ MAKE_YUV_FORMAT (YV12, "raw video", GST_MAKE_FOURCC ('Y', 'V', '1', '2'),
+ DPTH888, PSTR111,
+ PLANE021, OFFS0, SUB420, PACK_420),
+ MAKE_YUV_FORMAT (YUY2, "raw video", GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'),
+ DPTH888, PSTR244,
+ PLANE0, OFFS013, SUB422, PACK_YUY2),
+ MAKE_YUV_FORMAT (UYVY, "raw video", GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'),
+ DPTH888, PSTR244,
+ PLANE0, OFFS102, SUB422, PACK_UYVY),
+ MAKE_YUVA_FORMAT (AYUV, "raw video", GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'),
+ DPTH8888,
+ PSTR4444, PLANE0, OFFS1230, SUB4444, PACK_AYUV),
+ MAKE_RGB_FORMAT (RGBx, "raw video", DPTH888, PSTR444, PLANE0, OFFS012,
+ SUB444, PACK_RGBA),
+ MAKE_RGB_FORMAT (BGRx, "raw video", DPTH888, PSTR444, PLANE0, OFFS210,
+ SUB444, PACK_BGRA),
+ MAKE_RGB_FORMAT (xRGB, "raw video", DPTH888, PSTR444, PLANE0, OFFS123,
+ SUB444, PACK_ARGB),
+ MAKE_RGB_FORMAT (xBGR, "raw video", DPTH888, PSTR444, PLANE0, OFFS321,
+ SUB444, PACK_ABGR),
+ MAKE_RGBA_FORMAT (RGBA, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS0123,
+ SUB4444, PACK_RGBA),
+ MAKE_RGBA_FORMAT (BGRA, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS2103,
+ SUB4444, PACK_BGRA),
+ MAKE_RGBA_FORMAT (ARGB, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS1230,
+ SUB4444, PACK_ARGB),
+ MAKE_RGBA_FORMAT (ABGR, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS3210,
+ SUB4444, PACK_ABGR),
+ MAKE_RGB_FORMAT (RGB, "raw video", DPTH888, PSTR333, PLANE0, OFFS012, SUB444,
+ PACK_RGB),
+ MAKE_RGB_FORMAT (BGR, "raw video", DPTH888, PSTR333, PLANE0, OFFS210, SUB444,
+ PACK_BGR),
+
+ MAKE_YUV_FORMAT (Y41B, "raw video", GST_MAKE_FOURCC ('Y', '4', '1', 'B'),
+ DPTH888, PSTR111,
+ PLANE012, OFFS0, SUB411, PACK_Y41B),
+ MAKE_YUV_FORMAT (Y42B, "raw video", GST_MAKE_FOURCC ('Y', '4', '2', 'B'),
+ DPTH888, PSTR111,
+ PLANE012, OFFS0, SUB422, PACK_Y42B),
+ MAKE_YUV_FORMAT (YVYU, "raw video", GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'),
+ DPTH888, PSTR244,
+ PLANE0, OFFS031, SUB422, PACK_YVYU),
+ MAKE_YUV_FORMAT (Y444, "raw video", GST_MAKE_FOURCC ('Y', '4', '4', '4'),
+ DPTH888, PSTR111,
+ PLANE012, OFFS0, SUB444, PACK_Y444),
+ MAKE_YUV_C_FORMAT (v210, "raw video", GST_MAKE_FOURCC ('v', '2', '1', '0'),
+ DPTH10_10_10,
+ PSTR0, PLANE0, OFFS0, SUB422, PACK_v210),
+ MAKE_YUV_FORMAT (v216, "raw video", GST_MAKE_FOURCC ('v', '2', '1', '6'),
+ DPTH16_16_16,
+ PSTR488, PLANE0, OFFS204, SUB422, PACK_v216),
+ MAKE_YUV_FORMAT (NV12, "raw video", GST_MAKE_FOURCC ('N', 'V', '1', '2'),
+ DPTH888, PSTR122,
+ PLANE011, OFFS001, SUB420, PACK_NV12),
+ MAKE_YUV_FORMAT (NV21, "raw video", GST_MAKE_FOURCC ('N', 'V', '2', '1'),
+ DPTH888, PSTR122,
+ PLANE011, OFFS010, SUB420, PACK_NV21),
+
+ MAKE_GRAY_FORMAT (GRAY8, "raw video", DPTH8, PSTR1, PLANE0, OFFS0, SUB4,
+ PACK_GRAY8),
+ MAKE_GRAY_FORMAT (GRAY16_BE, "raw video", DPTH16, PSTR2, PLANE0, OFFS0, SUB4,
+ PACK_GRAY16_BE),
+ MAKE_GRAY_LE_FORMAT (GRAY16_LE, "raw video", DPTH16, PSTR2, PLANE0, OFFS0,
+ SUB4, PACK_GRAY16_LE),
+
+ MAKE_YUV_FORMAT (v308, "raw video", GST_MAKE_FOURCC ('v', '3', '0', '8'),
+ DPTH888, PSTR333,
+ PLANE0, OFFS012, SUB444, PACK_v308),
+
+#if G_BYTE_ORDER == G_LITTLE_ENDIAN
+ MAKE_RGB_LE_FORMAT (RGB16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0,
+ SUB444, PACK_RGB16),
+ MAKE_RGB_LE_FORMAT (BGR16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0,
+ SUB444, PACK_BGR16),
+ MAKE_RGB_LE_FORMAT (RGB15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0,
+ SUB444, PACK_RGB15),
+ MAKE_RGB_LE_FORMAT (BGR15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0,
+ SUB444, PACK_BGR15),
+#else
+ MAKE_RGB_FORMAT (RGB16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0, SUB444,
+ PACK_RGB16),
+ MAKE_RGB_FORMAT (BGR16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0, SUB444,
+ PACK_BGR16),
+ MAKE_RGB_FORMAT (RGB15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0, SUB444,
+ PACK_RGB15),
+ MAKE_RGB_FORMAT (BGR15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0, SUB444,
+ PACK_BGR15),
+#endif
+
+ MAKE_YUV_C_FORMAT (UYVP, "raw video", GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'),
+ DPTH10_10_10,
+ PSTR0, PLANE0, OFFS0, SUB422, PACK_UYVP),
+ MAKE_YUVA_FORMAT (A420, "raw video", GST_MAKE_FOURCC ('A', '4', '2', '0'),
+ DPTH8888,
+ PSTR1111, PLANE0123, OFFS0, SUB4204, PACK_A420),
+ MAKE_RGBAP_FORMAT (RGB8P, "raw video", DPTH8_32, PSTR14, PLANE01,
+ OFFS0, SUB44, PACK_RGB8P),
+ MAKE_YUV_FORMAT (YUV9, "raw video", GST_MAKE_FOURCC ('Y', 'U', 'V', '9'),
+ DPTH888, PSTR111,
+ PLANE012, OFFS0, SUB410, PACK_410),
+ MAKE_YUV_FORMAT (YVU9, "raw video", GST_MAKE_FOURCC ('Y', 'V', 'U', '9'),
+ DPTH888, PSTR111,
+ PLANE021, OFFS0, SUB410, PACK_410),
+ MAKE_YUV_FORMAT (IYU1, "raw video", GST_MAKE_FOURCC ('I', 'Y', 'U', '1'),
+ DPTH888, PSTR0,
+ PLANE0, OFFS104, SUB411, PACK_IYU1),
+ MAKE_RGBA_FORMAT (ARGB64, "raw video", DPTH16_16_16_16, PSTR8888, PLANE0,
+ OFFS2460,
+ SUB444, PACK_ARGB64),
+ MAKE_YUVA_FORMAT (AYUV64, "raw video", 0x00000000, DPTH16_16_16_16, PSTR8888,
+ PLANE0,
+ OFFS2460, SUB444, PACK_AYUV64),
+ MAKE_YUV_FORMAT (r210, "raw video", GST_MAKE_FOURCC ('r', '2', '1', '0'),
+ DPTH10_10_10,
+ PSTR444, PLANE0, OFFS0, SUB444, PACK_r210),
+ MAKE_YUV_FORMAT (I420_10BE, "raw video", 0x00000000, DPTH10_10_10,
+ PSTR222, PLANE012, OFFS0, SUB420, PACK_I420_10BE),
+ MAKE_YUV_LE_FORMAT (I420_10LE, "raw video", 0x00000000, DPTH10_10_10,
+ PSTR222, PLANE012, OFFS0, SUB420, PACK_I420_10LE),
+};
+
+static GstVideoFormat
+gst_video_format_from_rgb32_masks (int red_mask, int green_mask, int blue_mask)
+{
+ if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
+ blue_mask == 0x0000ff00) {
+ return GST_VIDEO_FORMAT_RGBx;
+ }
+ if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
+ blue_mask == 0xff000000) {
+ return GST_VIDEO_FORMAT_BGRx;
+ }
+ if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
+ blue_mask == 0x000000ff) {
+ return GST_VIDEO_FORMAT_xRGB;
+ }
+ if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
+ blue_mask == 0x00ff0000) {
+ return GST_VIDEO_FORMAT_xBGR;
+ }
+
+ return GST_VIDEO_FORMAT_UNKNOWN;
+}
+
+static GstVideoFormat
+gst_video_format_from_rgba32_masks (int red_mask, int green_mask,
+ int blue_mask, int alpha_mask)
+{
+ if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
+ blue_mask == 0x0000ff00 && alpha_mask == 0x000000ff) {
+ return GST_VIDEO_FORMAT_RGBA;
+ }
+ if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
+ blue_mask == 0xff000000 && alpha_mask == 0x000000ff) {
+ return GST_VIDEO_FORMAT_BGRA;
+ }
+ if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
+ blue_mask == 0x000000ff && alpha_mask == 0xff000000) {
+ return GST_VIDEO_FORMAT_ARGB;
+ }
+ if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
+ blue_mask == 0x00ff0000 && alpha_mask == 0xff000000) {
+ return GST_VIDEO_FORMAT_ABGR;
+ }
+ return GST_VIDEO_FORMAT_UNKNOWN;
+}
+
+static GstVideoFormat
+gst_video_format_from_rgb24_masks (int red_mask, int green_mask, int blue_mask)
+{
+ if (red_mask == 0xff0000 && green_mask == 0x00ff00 && blue_mask == 0x0000ff) {
+ return GST_VIDEO_FORMAT_RGB;
+ }
+ if (red_mask == 0x0000ff && green_mask == 0x00ff00 && blue_mask == 0xff0000) {
+ return GST_VIDEO_FORMAT_BGR;
+ }
+
+ return GST_VIDEO_FORMAT_UNKNOWN;
+}
+
+#define GST_VIDEO_COMP1_MASK_16_INT 0xf800
+#define GST_VIDEO_COMP2_MASK_16_INT 0x07e0
+#define GST_VIDEO_COMP3_MASK_16_INT 0x001f
+
+#define GST_VIDEO_COMP1_MASK_15_INT 0x7c00
+#define GST_VIDEO_COMP2_MASK_15_INT 0x03e0
+#define GST_VIDEO_COMP3_MASK_15_INT 0x001f
+
+static GstVideoFormat
+gst_video_format_from_rgb16_masks (int red_mask, int green_mask, int blue_mask)
+{
+ if (red_mask == GST_VIDEO_COMP1_MASK_16_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_16_INT
+ && blue_mask == GST_VIDEO_COMP3_MASK_16_INT) {
+ return GST_VIDEO_FORMAT_RGB16;
+ }
+ if (red_mask == GST_VIDEO_COMP3_MASK_16_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_16_INT
+ && blue_mask == GST_VIDEO_COMP1_MASK_16_INT) {
+ return GST_VIDEO_FORMAT_BGR16;
+ }
+ if (red_mask == GST_VIDEO_COMP1_MASK_15_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_15_INT
+ && blue_mask == GST_VIDEO_COMP3_MASK_15_INT) {
+ return GST_VIDEO_FORMAT_RGB15;
+ }
+ if (red_mask == GST_VIDEO_COMP3_MASK_15_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_15_INT
+ && blue_mask == GST_VIDEO_COMP1_MASK_15_INT) {
+ return GST_VIDEO_FORMAT_BGR15;
+ }
+ return GST_VIDEO_FORMAT_UNKNOWN;
+}
+
+/**
+ * gst_video_format_from_masks:
+ * @depth: the amount of bits used for a pixel
+ * @bpp: the amount of bits used to store a pixel. This value is bigger than
+ * @depth
+ * @endianness: the endianness of the masks
+ * @red_mask: the red mask
+ * @green_mask: the green mask
+ * @blue_mask: the blue mask
+ * @alpha_mask: the optional alpha mask
+ *
+ * Find the #GstVideoFormat for the given parameters.
+ *
+ * Returns: a #GstVideoFormat or GST_VIDEO_FORMAT_UNKNOWN when the parameters to
+ * not specify a known format.
+ */
+GstVideoFormat
+gst_video_format_from_masks (gint depth, gint bpp, gint endianness,
+ gint red_mask, gint green_mask, gint blue_mask, gint alpha_mask)
+{
+ GstVideoFormat format;
+
+ /* our caps system handles 24/32bpp RGB as big-endian. */
+ if ((bpp == 24 || bpp == 32) && endianness == G_LITTLE_ENDIAN) {
+ red_mask = GUINT32_TO_BE (red_mask);
+ green_mask = GUINT32_TO_BE (green_mask);
+ blue_mask = GUINT32_TO_BE (blue_mask);
+ endianness = G_BIG_ENDIAN;
+ if (bpp == 24) {
+ red_mask >>= 8;
+ green_mask >>= 8;
+ blue_mask >>= 8;
+ }
+ }
+
+ if (depth == 30 && bpp == 32) {
+ format = GST_VIDEO_FORMAT_r210;
+ } else if (depth == 24 && bpp == 32) {
+ format = gst_video_format_from_rgb32_masks (red_mask, green_mask,
+ blue_mask);
+ } else if (depth == 32 && bpp == 32 && alpha_mask) {
+ format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
+ blue_mask, alpha_mask);
+ } else if (depth == 24 && bpp == 24) {
+ format = gst_video_format_from_rgb24_masks (red_mask, green_mask,
+ blue_mask);
+ } else if ((depth == 15 || depth == 16) && bpp == 16 &&
+ endianness == G_BYTE_ORDER) {
+ format = gst_video_format_from_rgb16_masks (red_mask, green_mask,
+ blue_mask);
+ } else if (depth == 8 && bpp == 8) {
+ format = GST_VIDEO_FORMAT_RGB8P;
+ } else if (depth == 64 && bpp == 64) {
+ format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
+ blue_mask, alpha_mask);
+ if (format == GST_VIDEO_FORMAT_ARGB) {
+ format = GST_VIDEO_FORMAT_ARGB64;
+ } else {
+ format = GST_VIDEO_FORMAT_UNKNOWN;
+ }
+ } else {
+ format = GST_VIDEO_FORMAT_UNKNOWN;
+ }
+ return format;
+}
+
+/**
+ * gst_video_format_from_fourcc:
+ * @fourcc: a FOURCC value representing raw YUV video
+ *
+ * Converts a FOURCC value into the corresponding #GstVideoFormat.
+ * If the FOURCC cannot be represented by #GstVideoFormat,
+ * #GST_VIDEO_FORMAT_UNKNOWN is returned.
+ *
+ * Since: 0.10.16
+ *
+ * Returns: the #GstVideoFormat describing the FOURCC value
+ */
+GstVideoFormat
+gst_video_format_from_fourcc (guint32 fourcc)
+{
+ switch (fourcc) {
+ case GST_MAKE_FOURCC ('I', '4', '2', '0'):
+ return GST_VIDEO_FORMAT_I420;
+ case GST_MAKE_FOURCC ('Y', 'V', '1', '2'):
+ return GST_VIDEO_FORMAT_YV12;
+ case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'):
+ return GST_VIDEO_FORMAT_YUY2;
+ case GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'):
+ return GST_VIDEO_FORMAT_YVYU;
+ case GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'):
+ return GST_VIDEO_FORMAT_UYVY;
+ case GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'):
+ return GST_VIDEO_FORMAT_AYUV;
+ case GST_MAKE_FOURCC ('Y', '4', '1', 'B'):
+ return GST_VIDEO_FORMAT_Y41B;
+ case GST_MAKE_FOURCC ('Y', '4', '2', 'B'):
+ return GST_VIDEO_FORMAT_Y42B;
+ case GST_MAKE_FOURCC ('Y', '4', '4', '4'):
+ return GST_VIDEO_FORMAT_Y444;
+ case GST_MAKE_FOURCC ('v', '2', '1', '0'):
+ return GST_VIDEO_FORMAT_v210;
+ case GST_MAKE_FOURCC ('v', '2', '1', '6'):
+ return GST_VIDEO_FORMAT_v216;
+ case GST_MAKE_FOURCC ('N', 'V', '1', '2'):
+ return GST_VIDEO_FORMAT_NV12;
+ case GST_MAKE_FOURCC ('N', 'V', '2', '1'):
+ return GST_VIDEO_FORMAT_NV21;
+ case GST_MAKE_FOURCC ('v', '3', '0', '8'):
+ return GST_VIDEO_FORMAT_v308;
+ case GST_MAKE_FOURCC ('Y', '8', '0', '0'):
+ case GST_MAKE_FOURCC ('Y', '8', ' ', ' '):
+ case GST_MAKE_FOURCC ('G', 'R', 'E', 'Y'):
+ return GST_VIDEO_FORMAT_GRAY8;
+ case GST_MAKE_FOURCC ('Y', '1', '6', ' '):
+ return GST_VIDEO_FORMAT_GRAY16_LE;
+ case GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'):
+ return GST_VIDEO_FORMAT_UYVP;
+ case GST_MAKE_FOURCC ('A', '4', '2', '0'):
+ return GST_VIDEO_FORMAT_A420;
+ case GST_MAKE_FOURCC ('Y', 'U', 'V', '9'):
+ return GST_VIDEO_FORMAT_YUV9;
+ case GST_MAKE_FOURCC ('Y', 'V', 'U', '9'):
+ return GST_VIDEO_FORMAT_YVU9;
+ case GST_MAKE_FOURCC ('I', 'Y', 'U', '1'):
+ return GST_VIDEO_FORMAT_IYU1;
+ case GST_MAKE_FOURCC ('A', 'Y', '6', '4'):
+ return GST_VIDEO_FORMAT_AYUV64;
+ default:
+ return GST_VIDEO_FORMAT_UNKNOWN;
+ }
+}
+
+/**
+ * gst_video_format_from_string:
+ * @format: a format string
+ *
+ * Convert the @format string to its #GstVideoFormat.
+ *
+ * Returns: the #GstVideoFormat for @format or GST_VIDEO_FORMAT_UNKNOWN when the
+ * string is not a known format.
+ */
+GstVideoFormat
+gst_video_format_from_string (const gchar * format)
+{
+ guint i;
+
+ g_return_val_if_fail (format != NULL, GST_VIDEO_FORMAT_UNKNOWN);
+
+ for (i = 0; i < G_N_ELEMENTS (formats); i++) {
+ if (strcmp (GST_VIDEO_FORMAT_INFO_NAME (&formats[i].info), format) == 0)
+ return GST_VIDEO_FORMAT_INFO_FORMAT (&formats[i].info);
+ }
+ return GST_VIDEO_FORMAT_UNKNOWN;
+}
+
+
+/**
+ * gst_video_format_to_fourcc:
+ * @format: a #GstVideoFormat video format
+ *
+ * Converts a #GstVideoFormat value into the corresponding FOURCC. Only
+ * a few YUV formats have corresponding FOURCC values. If @format has
+ * no corresponding FOURCC value, 0 is returned.
+ *
+ * Since: 0.10.16
+ *
+ * Returns: the FOURCC corresponding to @format
+ */
+guint32
+gst_video_format_to_fourcc (GstVideoFormat format)
+{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return 0;
+
+ return formats[format].fourcc;
+}
+
+const gchar *
+gst_video_format_to_string (GstVideoFormat format)
+{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return NULL;
+
+ return GST_VIDEO_FORMAT_INFO_NAME (&formats[format].info);
+}
+
+/**
+ * gst_video_format_get_info:
+ * @format: a #GstVideoFormat
+ *
+ * Get the #GstVideoFormatInfo for @format
+ *
+ * Returns: The #GstVideoFormatInfo for @format.
+ */
+const GstVideoFormatInfo *
+gst_video_format_get_info (GstVideoFormat format)
+{
+ g_return_val_if_fail (format < G_N_ELEMENTS (formats), NULL);
+
+ return &formats[format].info;
+}
+
+typedef struct
+{
+ const gchar *name;
+ GstVideoChromaSite site;
+} ChromaSiteInfo;
+
+static const ChromaSiteInfo chromasite[] = {
+ {"jpeg", GST_VIDEO_CHROMA_SITE_JPEG},
+ {"mpeg2", GST_VIDEO_CHROMA_SITE_MPEG2},
+ {"dv", GST_VIDEO_CHROMA_SITE_DV}
+};
+
+/**
+ * gst_video_chroma_from_string:
+ * @s: a chromasite string
+ *
+ * Convert @s to a #GstVideoChromaSite
+ *
+ * Returns: a #GstVideoChromaSite or %GST_VIDEO_CHROMA_SITE_UNKNOWN when @s does
+ * not contain a valid chroma description.
+ */
+GstVideoChromaSite
+gst_video_chroma_from_string (const gchar * s)
+{
+ gint i;
+ for (i = 0; i < G_N_ELEMENTS (chromasite); i++) {
+ if (g_str_equal (chromasite[i].name, s))
+ return chromasite[i].site;
+ }
+ return GST_VIDEO_CHROMA_SITE_UNKNOWN;
+}
+
+/**
+ * gst_video_chroma_to_string:
+ * @site: a #GstVideoChromaSite
+ *
+ * Converts @site to its string representation.
+ *
+ * Returns: a string describing @site.
+ */
+const gchar *
+gst_video_chroma_to_string (GstVideoChromaSite site)
+{
+ gint i;
+ for (i = 0; i < G_N_ELEMENTS (chromasite); i++) {
+ if (chromasite[i].site == site)
+ return chromasite[i].name;
+ }
+ return NULL;
+}
--- /dev/null
+/* GStreamer
+ * Copyright (C) <2011> Wim Taymans <wim.taymans@gmail.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifndef __GST_VIDEO_FORMAT_H__
+#define __GST_VIDEO_FORMAT_H__
+
+#include <gst/gst.h>
+
+G_BEGIN_DECLS
+
+#include <gst/video/video-enumtypes.h>
+
+/**
+ * GstVideoFormat:
+ * @GST_VIDEO_FORMAT_UNKNOWN: Unknown or unset video format id
+ * @GST_VIDEO_FORMAT_ENCODED: Encoded video format
+ * @GST_VIDEO_FORMAT_I420: planar 4:2:0 YUV
+ * @GST_VIDEO_FORMAT_YV12: planar 4:2:0 YVU (like I420 but UV planes swapped)
+ * @GST_VIDEO_FORMAT_YUY2: packed 4:2:2 YUV (Y0-U0-Y1-V0 Y2-U2-Y3-V2 Y4 ...)
+ * @GST_VIDEO_FORMAT_UYVY: packed 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...)
+ * @GST_VIDEO_FORMAT_AYUV: packed 4:4:4 YUV with alpha channel (A0-Y0-U0-V0 ...)
+ * @GST_VIDEO_FORMAT_RGBx: sparse rgb packed into 32 bit, space last
+ * @GST_VIDEO_FORMAT_BGRx: sparse reverse rgb packed into 32 bit, space last
+ * @GST_VIDEO_FORMAT_xRGB: sparse rgb packed into 32 bit, space first
+ * @GST_VIDEO_FORMAT_xBGR: sparse reverse rgb packed into 32 bit, space first
+ * @GST_VIDEO_FORMAT_RGBA: rgb with alpha channel last
+ * @GST_VIDEO_FORMAT_BGRA: reverse rgb with alpha channel last
+ * @GST_VIDEO_FORMAT_ARGB: rgb with alpha channel first
+ * @GST_VIDEO_FORMAT_ABGR: reverse rgb with alpha channel first
+ * @GST_VIDEO_FORMAT_RGB: rgb
+ * @GST_VIDEO_FORMAT_BGR: reverse rgb
+ * @GST_VIDEO_FORMAT_Y41B: planar 4:1:1 YUV (Since: 0.10.18)
+ * @GST_VIDEO_FORMAT_Y42B: planar 4:2:2 YUV (Since: 0.10.18)
+ * @GST_VIDEO_FORMAT_YVYU: packed 4:2:2 YUV (Y0-V0-Y1-U0 Y2-V2-Y3-U2 Y4 ...) (Since: 0.10.23)
+ * @GST_VIDEO_FORMAT_Y444: planar 4:4:4 YUV (Since: 0.10.24)
+ * @GST_VIDEO_FORMAT_v210: packed 4:2:2 10-bit YUV, complex format (Since: 0.10.24)
+ * @GST_VIDEO_FORMAT_v216: packed 4:2:2 16-bit YUV, Y0-U0-Y1-V1 order (Since: 0.10.24)
+ * @GST_VIDEO_FORMAT_NV12: planar 4:2:0 YUV with interleaved UV plane (Since: 0.10.26)
+ * @GST_VIDEO_FORMAT_NV21: planar 4:2:0 YUV with interleaved VU plane (Since: 0.10.26)
+ * @GST_VIDEO_FORMAT_GRAY8: 8-bit grayscale (Since: 0.10.29)
+ * @GST_VIDEO_FORMAT_GRAY16_BE: 16-bit grayscale, most significant byte first (Since: 0.10.29)
+ * @GST_VIDEO_FORMAT_GRAY16_LE: 16-bit grayscale, least significant byte first (Since: 0.10.29)
+ * @GST_VIDEO_FORMAT_v308: packed 4:4:4 YUV (Since: 0.10.29)
+ * @GST_VIDEO_FORMAT_RGB16: rgb 5-6-5 bits per component (Since: 0.10.30)
+ * @GST_VIDEO_FORMAT_BGR16: reverse rgb 5-6-5 bits per component (Since: 0.10.30)
+ * @GST_VIDEO_FORMAT_RGB15: rgb 5-5-5 bits per component (Since: 0.10.30)
+ * @GST_VIDEO_FORMAT_BGR15: reverse rgb 5-5-5 bits per component (Since: 0.10.30)
+ * @GST_VIDEO_FORMAT_UYVP: packed 10-bit 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...) (Since: 0.10.31)
+ * @GST_VIDEO_FORMAT_A420: planar 4:4:2:0 AYUV (Since: 0.10.31)
+ * @GST_VIDEO_FORMAT_RGB8P: 8-bit paletted RGB (Since: 0.10.32)
+ * @GST_VIDEO_FORMAT_YUV9: planar 4:1:0 YUV (Since: 0.10.32)
+ * @GST_VIDEO_FORMAT_YVU9: planar 4:1:0 YUV (like YUV9 but UV planes swapped) (Since: 0.10.32)
+ * @GST_VIDEO_FORMAT_IYU1: packed 4:1:1 YUV (Cb-Y0-Y1-Cr-Y2-Y3 ...) (Since: 0.10.32)
+ * @GST_VIDEO_FORMAT_ARGB64: rgb with alpha channel first, 16 bits per channel (Since: 0.10.33)
+ * @GST_VIDEO_FORMAT_AYUV64: packed 4:4:4 YUV with alpha channel, 16 bits per channel (A0-Y0-U0-V0 ...) (Since: 0.10.33)
+ * @GST_VIDEO_FORMAT_r210: packed 4:4:4 RGB, 10 bits per channel (Since: 0.10.33)
+ * @GST_VIDEO_FORMAT_I420_10BE: planar 4:2:0 YUV, 10 bits per channel
+ * @GST_VIDEO_FORMAT_I420_10LE: planar 4:2:0 YUV, 10 bits per channel
+ *
+ * Enum value describing the most common video formats.
+ */
+typedef enum {
+ GST_VIDEO_FORMAT_UNKNOWN,
+ GST_VIDEO_FORMAT_ENCODED,
+ GST_VIDEO_FORMAT_I420,
+ GST_VIDEO_FORMAT_YV12,
+ GST_VIDEO_FORMAT_YUY2,
+ GST_VIDEO_FORMAT_UYVY,
+ GST_VIDEO_FORMAT_AYUV,
+ GST_VIDEO_FORMAT_RGBx,
+ GST_VIDEO_FORMAT_BGRx,
+ GST_VIDEO_FORMAT_xRGB,
+ GST_VIDEO_FORMAT_xBGR,
+ GST_VIDEO_FORMAT_RGBA,
+ GST_VIDEO_FORMAT_BGRA,
+ GST_VIDEO_FORMAT_ARGB,
+ GST_VIDEO_FORMAT_ABGR,
+ GST_VIDEO_FORMAT_RGB,
+ GST_VIDEO_FORMAT_BGR,
+ GST_VIDEO_FORMAT_Y41B,
+ GST_VIDEO_FORMAT_Y42B,
+ GST_VIDEO_FORMAT_YVYU,
+ GST_VIDEO_FORMAT_Y444,
+ GST_VIDEO_FORMAT_v210,
+ GST_VIDEO_FORMAT_v216,
+ GST_VIDEO_FORMAT_NV12,
+ GST_VIDEO_FORMAT_NV21,
+ GST_VIDEO_FORMAT_GRAY8,
+ GST_VIDEO_FORMAT_GRAY16_BE,
+ GST_VIDEO_FORMAT_GRAY16_LE,
+ GST_VIDEO_FORMAT_v308,
+ GST_VIDEO_FORMAT_RGB16,
+ GST_VIDEO_FORMAT_BGR16,
+ GST_VIDEO_FORMAT_RGB15,
+ GST_VIDEO_FORMAT_BGR15,
+ GST_VIDEO_FORMAT_UYVP,
+ GST_VIDEO_FORMAT_A420,
+ GST_VIDEO_FORMAT_RGB8P,
+ GST_VIDEO_FORMAT_YUV9,
+ GST_VIDEO_FORMAT_YVU9,
+ GST_VIDEO_FORMAT_IYU1,
+ GST_VIDEO_FORMAT_ARGB64,
+ GST_VIDEO_FORMAT_AYUV64,
+ GST_VIDEO_FORMAT_r210,
+ GST_VIDEO_FORMAT_I420_10BE,
+ GST_VIDEO_FORMAT_I420_10LE
+} GstVideoFormat;
+
+#define GST_VIDEO_MAX_PLANES 4
+#define GST_VIDEO_MAX_COMPONENTS 4
+
+typedef struct _GstVideoFormatInfo GstVideoFormatInfo;
+
+/**
+ * GstVideoChromaSite:
+ * @GST_VIDEO_CHROMA_SITE_UNKNOWN: unknown cositing
+ * @GST_VIDEO_CHROMA_SITE_NONE: no cositing
+ * @GST_VIDEO_CHROMA_SITE_H_COSITED: chroma is horizontally cosited
+ * @GST_VIDEO_CHROMA_SITE_V_COSITED: chroma is vertically cosited
+ * @GST_VIDEO_CHROMA_SITE_ALT_LINE: choma samples are sited on alternate lines
+ * @GST_VIDEO_CHROMA_SITE_COSITED: chroma samples cosited with luma samples
+ * @GST_VIDEO_CHROMA_SITE_JPEG: jpeg style cositing, also for mpeg1 and mjpeg
+ * @GST_VIDEO_CHROMA_SITE_MPEG2: mpeg2 style cositing
+ * @GST_VIDEO_CHROMA_SITE_DV: DV style cositing
+ *
+ * Various Chroma sitings.
+ */
+typedef enum {
+ GST_VIDEO_CHROMA_SITE_UNKNOWN = 0,
+ GST_VIDEO_CHROMA_SITE_NONE = (1 << 0),
+ GST_VIDEO_CHROMA_SITE_H_COSITED = (1 << 1),
+ GST_VIDEO_CHROMA_SITE_V_COSITED = (1 << 2),
+ GST_VIDEO_CHROMA_SITE_ALT_LINE = (1 << 3),
+ /* some common chroma cositing */
+ GST_VIDEO_CHROMA_SITE_COSITED = (GST_VIDEO_CHROMA_SITE_H_COSITED | GST_VIDEO_CHROMA_SITE_V_COSITED),
+ GST_VIDEO_CHROMA_SITE_JPEG = (GST_VIDEO_CHROMA_SITE_NONE),
+ GST_VIDEO_CHROMA_SITE_MPEG2 = (GST_VIDEO_CHROMA_SITE_H_COSITED),
+ GST_VIDEO_CHROMA_SITE_DV = (GST_VIDEO_CHROMA_SITE_COSITED | GST_VIDEO_CHROMA_SITE_ALT_LINE),
+} GstVideoChromaSite;
+
+GstVideoChromaSite gst_video_chroma_from_string (const gchar * s);
+const gchar * gst_video_chroma_to_string (GstVideoChromaSite site);
+
+/**
+ * GstVideoFormatFlags:
+ * @GST_VIDEO_FORMAT_FLAG_YUV: The video format is YUV, components are numbered
+ * 0=Y, 1=U, 2=V.
+ * @GST_VIDEO_FORMAT_FLAG_RGB: The video format is RGB, components are numbered
+ * 0=R, 1=G, 2=B.
+ * @GST_VIDEO_FORMAT_FLAG_GRAY: The video is gray, there is one gray component
+ * with index 0.
+ * @GST_VIDEO_FORMAT_FLAG_ALPHA: The video format has an alpha components with
+ * the number 3.
+ * @GST_VIDEO_FORMAT_FLAG_LE: The video format has data stored in little
+ * endianness.
+ * @GST_VIDEO_FORMAT_FLAG_PALETTE: The video format has a palette. The palette
+ * is stored in the second plane and indexes are stored in the first plane.
+ * @GST_VIDEO_FORMAT_FLAG_COMPLEX: The video format has a complex layout that
+ * can't be described with the usual information in the #GstVideoFormatInfo.
+ *
+ * The different video flags that a format info can have.
+ */
+typedef enum
+{
+ GST_VIDEO_FORMAT_FLAG_YUV = (1 << 0),
+ GST_VIDEO_FORMAT_FLAG_RGB = (1 << 1),
+ GST_VIDEO_FORMAT_FLAG_GRAY = (1 << 2),
+ GST_VIDEO_FORMAT_FLAG_ALPHA = (1 << 3),
+ GST_VIDEO_FORMAT_FLAG_LE = (1 << 4),
+ GST_VIDEO_FORMAT_FLAG_PALETTE = (1 << 5),
+ GST_VIDEO_FORMAT_FLAG_COMPLEX = (1 << 6)
+} GstVideoFormatFlags;
+
+/* YUV components */
+#define GST_VIDEO_COMP_Y 0
+#define GST_VIDEO_COMP_U 1
+#define GST_VIDEO_COMP_V 2
+
+/* RGB components */
+#define GST_VIDEO_COMP_R 0
+#define GST_VIDEO_COMP_G 1
+#define GST_VIDEO_COMP_B 2
+
+/* alpha component */
+#define GST_VIDEO_COMP_A 3
+
+/* palette components */
+#define GST_VIDEO_COMP_INDEX 0
+#define GST_VIDEO_COMP_PALETTE 1
+
+/**
+ * GstVideoPackFlags:
+ * @GST_VIDEO_PACK_FLAG_NONE: No flag
+ *
+ * The different flags that can be used when packing and unpacking.
+ */
+typedef enum
+{
+ GST_VIDEO_PACK_FLAG_NONE = 0
+} GstVideoPackFlags;
+
+/**
+ * GstVideoFormatUnpack:
+ * @info: a #GstVideoFormatInfo
+ * @flags: flags to control the unpacking
+ * @dest: a destination array
+ * @data: pointers to the data planes
+ * @stride: strides of the planes
+ * @x: the x position in the image to start from
+ * @y: the y position in the image to start from
+ * @width: the amount of pixels to unpack.
+ *
+ * Unpacks @width pixels from the given planes and strides containing data of
+ * format @info. The pixels will be unpacked into @dest which each component
+ * interleaved. @dest should at least be big enough to hold @width *
+ * n_components * size(unpack_format) bytes.
+ */
+typedef void (*GstVideoFormatUnpack) (const GstVideoFormatInfo *info,
+ GstVideoPackFlags flags, gpointer dest,
+ const gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES],
+ gint x, gint y, gint width);
+/**
+ * GstVideoFormatPack:
+ * @info: a #GstVideoFormatInfo
+ * @flags: flags to control the packing
+ * @src: a source array
+ * @sstride: the source array stride
+ * @data: pointers to the destination data planes
+ * @stride: strides of the destination planes
+ * @chroma_site: the chroma siting of the target when subsampled
+ * @y: the y position in the image to pack to
+ * @width: the amount of pixels to pack.
+ *
+ * Packs @width pixels from @src to the given planes and strides in the
+ * format @info. The pixels from source have each component interleaved
+ * and will be packed into the planes in @data.
+ *
+ * When @info refers to a format with a h_sub != 0, this function operates on
+ * (1 << h_sub) lines meaning that @src should contain at least (1 << h_sub)
+ * lines with a stride of @sstride and @y should be a multiple of (1 << h_sub).
+ */
+typedef void (*GstVideoFormatPack) (const GstVideoFormatInfo *info,
+ GstVideoPackFlags flags,
+ const gpointer src, gint sstride,
+ gpointer data[GST_VIDEO_MAX_PLANES],
+ const gint stride[GST_VIDEO_MAX_PLANES],
+ GstVideoChromaSite chroma_site,
+ gint y, gint width);
+
+/**
+ * GstVideoFormatInfo:
+ * @format: #GstVideoFormat
+ * @name: string representation of the format
+ * @description: use readable description of the format
+ * @flags: #GstVideoFormatFlags
+ * @bits: The number of bits used to pack data items. This can be less than 8
+ * when multiple pixels are stored in a byte. for values > 8 multiple bytes
+ * should be read according to the endianness flag before applying the shift
+ * and mask.
+ * @n_components: the number of components in the video format.
+ * @shift: the number of bits to shift away to get the component data
+ * @depth: the depth in bits for each component
+ * @pixel_stride: the pixel stride of each component. This is the amount of
+ * bytes to the pixel immediately to the right. When bits < 8, the stride is
+ * expressed in bits.
+ * @n_planes: the number of planes for this format. The number of planes can be
+ * less than the amount of components when multiple components are packed into
+ * one plane.
+ * @plane: the plane number where a component can be found
+ * @poffset: the offset in the plane where the first pixel of the components
+ * can be found. If bits < 8 the amount is specified in bits.
+ * @w_sub: subsampling factor of the width for the component. Use
+ * GST_VIDEO_SUB_SCALE to scale a width.
+ * @h_sub: subsampling factor of the height for the component. Use
+ * GST_VIDEO_SUB_SCALE to scale a height.
+ * @unpack_format: the format of the unpacked pixels.
+ * @unpack_func: an unpack function for this format
+ * @pack_func: an pack function for this format
+ *
+ * Information for a video format.
+ */
+struct _GstVideoFormatInfo {
+ GstVideoFormat format;
+ const gchar *name;
+ const gchar *description;
+ GstVideoFormatFlags flags;
+ guint bits;
+ guint n_components;
+ guint shift[GST_VIDEO_MAX_COMPONENTS];
+ guint depth[GST_VIDEO_MAX_COMPONENTS];
+ gint pixel_stride[GST_VIDEO_MAX_COMPONENTS];
+ guint n_planes;
+ guint plane[GST_VIDEO_MAX_COMPONENTS];
+ guint poffset[GST_VIDEO_MAX_COMPONENTS];
+ guint w_sub[GST_VIDEO_MAX_COMPONENTS];
+ guint h_sub[GST_VIDEO_MAX_COMPONENTS];
+
+ GstVideoFormat unpack_format;
+ GstVideoFormatUnpack unpack_func;
+ GstVideoFormatPack pack_func;
+
+ gpointer _gst_reserved[GST_PADDING];
+};
+
+#define GST_VIDEO_FORMAT_INFO_FORMAT(info) ((info)->format)
+#define GST_VIDEO_FORMAT_INFO_NAME(info) ((info)->name)
+#define GST_VIDEO_FORMAT_INFO_FLAGS(info) ((info)->flags)
+
+#define GST_VIDEO_FORMAT_INFO_IS_YUV(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_YUV)
+#define GST_VIDEO_FORMAT_INFO_IS_RGB(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_RGB)
+#define GST_VIDEO_FORMAT_INFO_IS_GRAY(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_GRAY)
+#define GST_VIDEO_FORMAT_INFO_HAS_ALPHA(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_ALPHA)
+#define GST_VIDEO_FORMAT_INFO_IS_LE(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_LE)
+#define GST_VIDEO_FORMAT_INFO_HAS_PALETTE(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_PALETTE)
+#define GST_VIDEO_FORMAT_INFO_IS_COMPLEX(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_COMPLEX)
+
+#define GST_VIDEO_FORMAT_INFO_BITS(info) ((info)->bits)
+#define GST_VIDEO_FORMAT_INFO_N_COMPONENTS(info) ((info)->n_components)
+#define GST_VIDEO_FORMAT_INFO_SHIFT(info,c) ((info)->shift[c])
+#define GST_VIDEO_FORMAT_INFO_DEPTH(info,c) ((info)->depth[c])
+#define GST_VIDEO_FORMAT_INFO_PSTRIDE(info,c) ((info)->pixel_stride[c])
+#define GST_VIDEO_FORMAT_INFO_N_PLANES(info) ((info)->n_planes)
+#define GST_VIDEO_FORMAT_INFO_PLANE(info,c) ((info)->plane[c])
+#define GST_VIDEO_FORMAT_INFO_POFFSET(info,c) ((info)->poffset[c])
+#define GST_VIDEO_FORMAT_INFO_W_SUB(info,c) ((info)->w_sub[c])
+#define GST_VIDEO_FORMAT_INFO_H_SUB(info,c) ((info)->h_sub[c])
+
+/* rounds up */
+#define GST_VIDEO_SUB_SCALE(scale,val) (-((-((gint)(val)))>>(scale)))
+
+#define GST_VIDEO_FORMAT_INFO_SCALE_WIDTH(info,c,w) GST_VIDEO_SUB_SCALE ((info)->w_sub[c],(w))
+#define GST_VIDEO_FORMAT_INFO_SCALE_HEIGHT(info,c,h) GST_VIDEO_SUB_SCALE ((info)->h_sub[c],(h))
+
+#define GST_VIDEO_FORMAT_INFO_DATA(info,planes,comp) \
+ (((guint8*)(planes)[(info)->plane[comp]]) + (info)->poffset[comp])
+#define GST_VIDEO_FORMAT_INFO_STRIDE(info,strides,comp) ((strides)[(info)->plane[comp]])
+#define GST_VIDEO_FORMAT_INFO_OFFSET(info,offsets,comp) \
+ (((offsets)[(info)->plane[comp]]) + (info)->poffset[comp])
+
+/* format properties */
+GstVideoFormat gst_video_format_from_masks (gint depth, gint bpp, gint endianness,
+ gint red_mask, gint green_mask,
+ gint blue_mask, gint alpha_mask) G_GNUC_CONST;
+
+GstVideoFormat gst_video_format_from_fourcc (guint32 fourcc) G_GNUC_CONST;
+GstVideoFormat gst_video_format_from_string (const gchar *format) G_GNUC_CONST;
+
+guint32 gst_video_format_to_fourcc (GstVideoFormat format) G_GNUC_CONST;
+const gchar * gst_video_format_to_string (GstVideoFormat format) G_GNUC_CONST;
+
+const GstVideoFormatInfo *
+ gst_video_format_get_info (GstVideoFormat format) G_GNUC_CONST;
+
+#define GST_VIDEO_SIZE_RANGE "(int) [ 1, max ]"
+#define GST_VIDEO_FPS_RANGE "(fraction) [ 0, max ]"
+
+#if G_BYTE_ORDER == G_LITTLE_ENDIAN
+# define GST_VIDEO_NE(s) G_STRINGIFY(s)"_LE"
+# define GST_VIDEO_OE(s) G_STRINGIFY(s)"_BE"
+#else
+# define GST_VIDEO_NE(s) G_STRINGIFY(s)"_BE"
+# define GST_VIDEO_OE(s) G_STRINGIFY(s)"_LE"
+#endif
+
+#define GST_VIDEO_FORMATS_ALL "{ I420, YV12, YUY2, UYVY, AYUV, RGBx, " \
+ "BGRx, xRGB, xBGR, RGBA, BGRA, ARGB, ABGR, RGB, BGR, Y41B, Y42B, " \
+ "YVYU, Y444, v210, v216, NV12, NV21, GRAY8, GRAY16_BE, GRAY16_LE, " \
+ "v308, RGB16, BGR16, RGB15, BGR15, UYVP, A420, RGB8P, YUV9, YVU9, " \
+ "IYU1, ARGB64, AYUV64, r210, I420_10LE, I420_10BE }"
+
+/**
+ * GST_VIDEO_CAPS_MAKE:
+ * @format: string format that describes the pixel layout, as string
+ * (e.g. "I420", "RGB", "YV12", "YUY2", "AYUV", etc.)
+ *
+ * Generic caps string for video, for use in pad templates.
+ */
+#define GST_VIDEO_CAPS_MAKE(format) \
+ "video/x-raw, " \
+ "format = (string) " format ", " \
+ "width = " GST_VIDEO_SIZE_RANGE ", " \
+ "height = " GST_VIDEO_SIZE_RANGE ", " \
+ "framerate = " GST_VIDEO_FPS_RANGE
+
+G_END_DECLS
+
+#endif /* __GST_VIDEO_FORMAT_H__ */
--- /dev/null
+/* GStreamer
+ * Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
+ * Library <2002> Ronald Bultje <rbultje@ronald.bitfreak.net>
+ * Copyright (C) 2007 David A. Schleef <ds@schleef.org>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifdef HAVE_CONFIG_H
+# include "config.h"
+#endif
+
+#include <string.h>
+#include <stdio.h>
+
+#include "video-frame.h"
+#include "gstvideometa.h"
+
+GST_DEBUG_CATEGORY_EXTERN (GST_CAT_PERFORMANCE);
+
+/**
+ * gst_video_frame_map_id:
+ * @frame: pointer to #GstVideoFrame
+ * @info: a #GstVideoInfo
+ * @buffer: the buffer to map
+ * @id: the frame id to map
+ * @flags: #GstMapFlags
+ *
+ * Use @info and @buffer to fill in the values of @frame with the video frame
+ * information of frame @id.
+ *
+ * When @id is -1, the default frame is mapped. When @id != -1, this function
+ * will return %FALSE when there is no GstVideoMeta with that id.
+ *
+ * All video planes of @buffer will be mapped and the pointers will be set in
+ * @frame->data.
+ *
+ * Returns: %TRUE on success.
+ */
+gboolean
+gst_video_frame_map_id (GstVideoFrame * frame, GstVideoInfo * info,
+ GstBuffer * buffer, gint id, GstMapFlags flags)
+{
+ GstVideoMeta *meta;
+ gint i;
+
+ g_return_val_if_fail (frame != NULL, FALSE);
+ g_return_val_if_fail (info != NULL, FALSE);
+ g_return_val_if_fail (GST_IS_BUFFER (buffer), FALSE);
+
+ if (id == -1)
+ meta = gst_buffer_get_video_meta (buffer);
+ else
+ meta = gst_buffer_get_video_meta_id (buffer, id);
+
+ /* copy the info */
+ frame->info = *info;
+
+ if (meta) {
+ frame->info.finfo = gst_video_format_get_info (meta->format);
+ frame->info.width = meta->width;
+ frame->info.height = meta->height;
+ frame->id = meta->id;
+ frame->flags = meta->flags;
+
+ for (i = 0; i < info->finfo->n_planes; i++)
+ gst_video_meta_map (meta, i, &frame->map[i], &frame->data[i],
+ &frame->info.stride[i], flags);
+ } else {
+ /* no metadata, we really need to have the metadata when the id is
+ * specified. */
+ if (id != -1)
+ goto no_metadata;
+
+ frame->id = id;
+ frame->flags = 0;
+
+ if (GST_VIDEO_INFO_IS_INTERLACED (info)) {
+ if (GST_BUFFER_FLAG_IS_SET (buffer, GST_VIDEO_BUFFER_FLAG_TFF))
+ frame->flags |= GST_VIDEO_FRAME_FLAG_TFF;
+ if (GST_BUFFER_FLAG_IS_SET (buffer, GST_VIDEO_BUFFER_FLAG_RFF))
+ frame->flags |= GST_VIDEO_FRAME_FLAG_RFF;
+ if (GST_BUFFER_FLAG_IS_SET (buffer, GST_VIDEO_BUFFER_FLAG_ONEFIELD))
+ frame->flags |= GST_VIDEO_FRAME_FLAG_ONEFIELD;
+ }
+
+ gst_buffer_map (buffer, &frame->map[0], flags);
+
+ /* do some sanity checks */
+ if (frame->map[0].size < info->size)
+ goto invalid_size;
+
+ /* set up pointers */
+ for (i = 0; i < info->finfo->n_planes; i++) {
+ frame->data[i] = frame->map[0].data + info->offset[i];
+ }
+ }
+ frame->buffer = gst_buffer_ref (buffer);
+ frame->meta = meta;
+
+ return TRUE;
+
+ /* ERRORS */
+no_metadata:
+ {
+ GST_ERROR ("no GstVideoMeta for id %d", id);
+ return FALSE;
+ }
+invalid_size:
+ {
+ GST_ERROR ("invalid buffer size %" G_GSIZE_FORMAT " < %" G_GSIZE_FORMAT,
+ frame->map[0].size, info->size);
+ gst_buffer_unmap (buffer, &frame->map[0]);
+ return FALSE;
+ }
+}
+
+/**
+ * gst_video_frame_map:
+ * @frame: pointer to #GstVideoFrame
+ * @info: a #GstVideoInfo
+ * @buffer: the buffer to map
+ * @flags: #GstMapFlags
+ *
+ * Use @info and @buffer to fill in the values of @frame.
+ *
+ * All video planes of @buffer will be mapped and the pointers will be set in
+ * @frame->data.
+ *
+ * Returns: %TRUE on success.
+ */
+gboolean
+gst_video_frame_map (GstVideoFrame * frame, GstVideoInfo * info,
+ GstBuffer * buffer, GstMapFlags flags)
+{
+ return gst_video_frame_map_id (frame, info, buffer, -1, flags);
+}
+
+/**
+ * gst_video_frame_unmap:
+ * @frame: a #GstVideoFrame
+ *
+ * Unmap the memory previously mapped with gst_video_frame_map.
+ */
+void
+gst_video_frame_unmap (GstVideoFrame * frame)
+{
+ GstBuffer *buffer;
+ GstVideoMeta *meta;
+ gint i;
+
+ g_return_if_fail (frame != NULL);
+
+ buffer = frame->buffer;
+ meta = frame->meta;
+
+ if (meta) {
+ for (i = 0; i < frame->info.finfo->n_planes; i++) {
+ gst_video_meta_unmap (meta, i, &frame->map[i]);
+ }
+ } else {
+ gst_buffer_unmap (buffer, &frame->map[0]);
+ }
+ gst_buffer_unref (buffer);
+}
+
+/**
+ * gst_video_frame_copy:
+ * @dest: a #GstVideoFrame
+ * @src: a #GstVideoFrame
+ * @plane: a plane
+ *
+ * Copy the plane with index @plane from @src to @dest.
+ *
+ * Returns: TRUE if the contents could be copied.
+ */
+gboolean
+gst_video_frame_copy_plane (GstVideoFrame * dest, const GstVideoFrame * src,
+ guint plane)
+{
+ const GstVideoInfo *sinfo;
+ GstVideoInfo *dinfo;
+ guint w, h, j;
+ guint8 *sp, *dp;
+ gint ss, ds;
+
+ g_return_val_if_fail (dest != NULL, FALSE);
+ g_return_val_if_fail (src != NULL, FALSE);
+
+ sinfo = &src->info;
+ dinfo = &dest->info;
+
+ g_return_val_if_fail (dinfo->finfo->format == sinfo->finfo->format, FALSE);
+ g_return_val_if_fail (dinfo->width == sinfo->width
+ && dinfo->height == sinfo->height, FALSE);
+ g_return_val_if_fail (dinfo->finfo->n_planes > plane, FALSE);
+
+ sp = src->data[plane];
+ dp = dest->data[plane];
+
+ ss = sinfo->stride[plane];
+ ds = dinfo->stride[plane];
+
+ w = MIN (ABS (ss), ABS (ds));
+ h = GST_VIDEO_FRAME_COMP_HEIGHT (dest, plane);
+
+ GST_CAT_DEBUG (GST_CAT_PERFORMANCE, "copy plane %d, w:%d h:%d ", plane, w, h);
+
+ for (j = 0; j < h; j++) {
+ memcpy (dp, sp, w);
+ dp += ds;
+ sp += ss;
+ }
+ return TRUE;
+}
+
+/**
+ * gst_video_frame_copy:
+ * @dest: a #GstVideoFrame
+ * @src: a #GstVideoFrame
+ *
+ * Copy the contents from @src to @dest.
+ *
+ * Returns: TRUE if the contents could be copied.
+ */
+gboolean
+gst_video_frame_copy (GstVideoFrame * dest, const GstVideoFrame * src)
+{
+ guint i, n_planes;
+ const GstVideoInfo *sinfo;
+ GstVideoInfo *dinfo;
+
+ g_return_val_if_fail (dest != NULL, FALSE);
+ g_return_val_if_fail (src != NULL, FALSE);
+
+ sinfo = &src->info;
+ dinfo = &dest->info;
+
+ g_return_val_if_fail (dinfo->finfo->format == sinfo->finfo->format, FALSE);
+ g_return_val_if_fail (dinfo->width == sinfo->width
+ && dinfo->height == sinfo->height, FALSE);
+
+ n_planes = dinfo->finfo->n_planes;
+ if (GST_VIDEO_FORMAT_INFO_HAS_PALETTE (sinfo->finfo)) {
+ memcpy (dest->data[1], src->data[1], 256 * 4);
+ n_planes = 1;
+ }
+
+ for (i = 0; i < n_planes; i++)
+ gst_video_frame_copy_plane (dest, src, i);
+
+ return TRUE;
+}
--- /dev/null
+/* GStreamer
+ * Copyright (C) <2011> Wim Taymans <wim.taymans@gmail.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifndef __GST_VIDEO_FRAME_H__
+#define __GST_VIDEO_FRAME_H__
+
+#include <gst/gst.h>
+#include <gst/video/video-info.h>
+
+G_BEGIN_DECLS
+
+#include <gst/video/video-enumtypes.h>
+
+typedef struct _GstVideoFrame GstVideoFrame;
+
+/**
+ * GstVideoFrameFlags:
+ * @GST_VIDEO_FRAME_FLAG_NONE: no flags
+ * @GST_VIDEO_FRAME_FLAG_TFF: The video frame has the top field first
+ * @GST_VIDEO_FRAME_FLAG_RFF: The video frame has the repeat flag
+ * @GST_VIDEO_FRAME_FLAG_ONEFIELD: The video frame has one field
+ *
+ * Extra video frame flags
+ */
+typedef enum {
+ GST_VIDEO_FRAME_FLAG_NONE = 0,
+ GST_VIDEO_FRAME_FLAG_TFF = (1 << 0),
+ GST_VIDEO_FRAME_FLAG_RFF = (1 << 1),
+ GST_VIDEO_FRAME_FLAG_ONEFIELD = (1 << 2)
+} GstVideoFrameFlags;
+
+/**
+ * GstVideoFrame:
+ * @info: the #GstVideoInfo
+ * @buffer: the mapped buffer
+ * @meta: pointer to metadata if any
+ * @id: id of the mapped frame. the id can for example be used to
+ * indentify the frame in case of multiview video.
+ * @data: pointers to the plane data
+ * @map: mappings of the planes
+ *
+ * A video frame obtained from gst_video_frame_map()
+ */
+struct _GstVideoFrame {
+ GstVideoInfo info;
+ GstVideoFrameFlags flags;
+
+ GstBuffer *buffer;
+ gpointer meta;
+ gint id;
+
+ gpointer data[GST_VIDEO_MAX_PLANES];
+ GstMapInfo map[GST_VIDEO_MAX_PLANES];
+};
+
+gboolean gst_video_frame_map (GstVideoFrame *frame, GstVideoInfo *info,
+ GstBuffer *buffer, GstMapFlags flags);
+gboolean gst_video_frame_map_id (GstVideoFrame *frame, GstVideoInfo *info,
+ GstBuffer *buffer, gint id, GstMapFlags flags);
+void gst_video_frame_unmap (GstVideoFrame *frame);
+
+gboolean gst_video_frame_copy (GstVideoFrame *dest, const GstVideoFrame *src);
+gboolean gst_video_frame_copy_plane (GstVideoFrame *dest, const GstVideoFrame *src,
+ guint plane);
+
+/* general info */
+#define GST_VIDEO_FRAME_FORMAT(f) (GST_VIDEO_INFO_FORMAT(&(f)->info))
+#define GST_VIDEO_FRAME_WIDTH(f) (GST_VIDEO_INFO_WIDTH(&(f)->info))
+#define GST_VIDEO_FRAME_HEIGHT(f) (GST_VIDEO_INFO_HEIGHT(&(f)->info))
+#define GST_VIDEO_FRAME_SIZE(f) (GST_VIDEO_INFO_SIZE(&(f)->info))
+
+/* dealing with planes */
+#define GST_VIDEO_FRAME_N_PLANES(f) (GST_VIDEO_INFO_N_PLANES(&(f)->info))
+#define GST_VIDEO_FRAME_PLANE_DATA(f,p) ((f)->data[p])
+#define GST_VIDEO_FRAME_PLANE_OFFSET(f,p) (GST_VIDEO_INFO_PLANE_OFFSET(&(f)->info,(p)))
+#define GST_VIDEO_FRAME_PLANE_STRIDE(f,p) (GST_VIDEO_INFO_PLANE_STRIDE(&(f)->info,(p)))
+
+/* dealing with components */
+#define GST_VIDEO_FRAME_N_COMPONENTS(f) GST_VIDEO_INFO_N_COMPONENTS(&(f)->info)
+#define GST_VIDEO_FRAME_COMP_DATA(f,c) GST_VIDEO_INFO_COMP_DATA(&(f)->info,(f)->data,(c))
+#define GST_VIDEO_FRAME_COMP_STRIDE(f,c) GST_VIDEO_INFO_COMP_STRIDE(&(f)->info,(c))
+#define GST_VIDEO_FRAME_COMP_OFFSET(f,c) GST_VIDEO_INFO_COMP_OFFSET(&(f)->info,(c))
+#define GST_VIDEO_FRAME_COMP_WIDTH(f,c) GST_VIDEO_INFO_COMP_WIDTH(&(f)->info,(c))
+#define GST_VIDEO_FRAME_COMP_HEIGHT(f,c) GST_VIDEO_INFO_COMP_HEIGHT(&(f)->info,(c))
+#define GST_VIDEO_FRAME_COMP_PLANE(f,c) GST_VIDEO_INFO_COMP_PLANE(&(f)->info,(c))
+#define GST_VIDEO_FRAME_COMP_PSTRIDE(f,c) GST_VIDEO_INFO_COMP_PSTRIDE(&(f)->info,(c))
+#define GST_VIDEO_FRAME_COMP_POFFSET(f,c) GST_VIDEO_INFO_COMP_POFFSET(&(f)->info,(c))
+
+/* buffer flags */
+
+/**
+ * GstVideoBufferFlags:
+ * @GST_VIDEO_BUFFER_FLAG_TFF: If the #GstBuffer is interlaced, then the first field
+ * in the video frame is the top field. If unset, the
+ * bottom field is first.
+ * @GST_VIDEO_BUFFER_FLAG_RFF: If the #GstBuffer is interlaced, then the first field
+ * (as defined by the %GST_VIDEO_BUFFER_TFF flag setting)
+ * is repeated.
+ * @GST_VIDEO_BUFFER_FLAG_ONEFIELD: If the #GstBuffer is interlaced, then only the
+ * first field (as defined by the %GST_VIDEO_BUFFER_TFF
+ * flag setting) is to be displayed.
+ *
+ * Additional video buffer flags.
+ */
+typedef enum {
+ GST_VIDEO_BUFFER_FLAG_TFF = (GST_BUFFER_FLAG_LAST << 0),
+ GST_VIDEO_BUFFER_FLAG_RFF = (GST_BUFFER_FLAG_LAST << 1),
+ GST_VIDEO_BUFFER_FLAG_ONEFIELD = (GST_BUFFER_FLAG_LAST << 2),
+
+ GST_VIDEO_BUFFER_FLAG_LAST = (GST_BUFFER_FLAG_LAST << 8)
+} GstVideoBufferFlags;
+
+G_END_DECLS
+
+#endif /* __GST_VIDEO_FRAME_H__ */
--- /dev/null
+/* GStreamer
+ * Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
+ * Library <2002> Ronald Bultje <rbultje@ronald.bitfreak.net>
+ * Copyright (C) 2007 David A. Schleef <ds@schleef.org>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifdef HAVE_CONFIG_H
+# include "config.h"
+#endif
+
+#include <string.h>
+#include <stdio.h>
+
+#include "video-info.h"
+
+static int fill_planes (GstVideoInfo * info);
+
+/**
+ * gst_video_info_init:
+ * @info: a #GstVideoInfo
+ *
+ * Initialize @info with default values.
+ */
+void
+gst_video_info_init (GstVideoInfo * info)
+{
+ g_return_if_fail (info != NULL);
+
+ memset (info, 0, sizeof (GstVideoInfo));
+
+ info->finfo = gst_video_format_get_info (GST_VIDEO_FORMAT_UNKNOWN);
+
+ info->views = 1;
+ /* arrange for sensible defaults, e.g. if turned into caps */
+ info->fps_n = 0;
+ info->fps_d = 1;
+ info->par_n = 1;
+ info->par_d = 1;
+}
+
+#define MAKE_COLORIMETRY(r,m,t,p) { \
+ GST_VIDEO_COLOR_RANGE ##r, GST_VIDEO_COLOR_MATRIX_ ##m, \
+ GST_VIDEO_TRANSFER_ ##t, GST_VIDEO_COLOR_PRIMARIES_ ##p }
+
+#define DEFAULT_YUV_SD 0
+#define DEFAULT_YUV_HD 1
+#define DEFAULT_RGB 2
+#define DEFAULT_GRAY 3
+#define DEFAULT_UNKNOWN 4
+
+static const GstVideoColorimetry default_color[] = {
+ MAKE_COLORIMETRY (_16_235, BT601, BT709, BT470M),
+ MAKE_COLORIMETRY (_16_235, BT709, BT709, BT709),
+ MAKE_COLORIMETRY (_0_255, RGB, UNKNOWN, UNKNOWN),
+ MAKE_COLORIMETRY (_0_255, BT601, UNKNOWN, UNKNOWN),
+ MAKE_COLORIMETRY (_UNKNOWN, UNKNOWN, UNKNOWN, UNKNOWN),
+};
+
+/**
+ * gst_video_info_set_format:
+ * @info: a #GstVideoInfo
+ * @format: the format
+ * @width: a width
+ * @height: a height
+ *
+ * Set the default info for a video frame of @format and @width and @height.
+ */
+void
+gst_video_info_set_format (GstVideoInfo * info, GstVideoFormat format,
+ guint width, guint height)
+{
+ const GstVideoFormatInfo *finfo;
+
+ g_return_if_fail (info != NULL);
+ g_return_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN);
+
+ finfo = gst_video_format_get_info (format);
+
+ info->flags = 0;
+ info->finfo = finfo;
+ info->width = width;
+ info->height = height;
+
+ if (GST_VIDEO_FORMAT_INFO_IS_YUV (finfo)) {
+ if (height > 576)
+ info->colorimetry = default_color[DEFAULT_YUV_HD];
+ else
+ info->colorimetry = default_color[DEFAULT_YUV_SD];
+ } else if (GST_VIDEO_FORMAT_INFO_IS_GRAY (finfo)) {
+ info->colorimetry = default_color[DEFAULT_GRAY];
+ } else if (GST_VIDEO_FORMAT_INFO_IS_RGB (finfo)) {
+ info->colorimetry = default_color[DEFAULT_RGB];
+ } else {
+ info->colorimetry = default_color[DEFAULT_UNKNOWN];
+ }
+
+ fill_planes (info);
+}
+
+static const gchar *interlace_mode[] = {
+ "progressive",
+ "interleaved",
+ "mixed",
+ "fields"
+};
+
+static const gchar *
+gst_interlace_mode_to_string (GstVideoInterlaceMode mode)
+{
+ if (((guint) mode) >= G_N_ELEMENTS (interlace_mode))
+ return NULL;
+
+ return interlace_mode[mode];
+}
+
+static GstVideoInterlaceMode
+gst_interlace_mode_from_string (const gchar * mode)
+{
+ gint i;
+ for (i = 0; i < G_N_ELEMENTS (interlace_mode); i++) {
+ if (g_str_equal (interlace_mode[i], mode))
+ return i;
+ }
+ return GST_VIDEO_INTERLACE_MODE_PROGRESSIVE;
+}
+
+/**
+ * gst_video_info_from_caps:
+ * @info: a #GstVideoInfo
+ * @caps: a #GstCaps
+ *
+ * Parse @caps and update @info.
+ *
+ * Returns: TRUE if @caps could be parsed
+ */
+gboolean
+gst_video_info_from_caps (GstVideoInfo * info, const GstCaps * caps)
+{
+ GstStructure *structure;
+ const gchar *s;
+ GstVideoFormat format = GST_VIDEO_FORMAT_UNKNOWN;
+ gint width = 0, height = 0, views;
+ gint fps_n, fps_d;
+ gint par_n, par_d;
+
+ g_return_val_if_fail (info != NULL, FALSE);
+ g_return_val_if_fail (caps != NULL, FALSE);
+ g_return_val_if_fail (gst_caps_is_fixed (caps), FALSE);
+
+ GST_DEBUG ("parsing caps %" GST_PTR_FORMAT, caps);
+
+ structure = gst_caps_get_structure (caps, 0);
+
+ if (gst_structure_has_name (structure, "video/x-raw")) {
+ if (!(s = gst_structure_get_string (structure, "format")))
+ goto no_format;
+
+ format = gst_video_format_from_string (s);
+ if (format == GST_VIDEO_FORMAT_UNKNOWN)
+ goto unknown_format;
+
+ } else if (g_str_has_prefix (gst_structure_get_name (structure), "video/") ||
+ g_str_has_prefix (gst_structure_get_name (structure), "image/")) {
+ format = GST_VIDEO_FORMAT_ENCODED;
+ } else {
+ goto wrong_name;
+ }
+
+ /* width and height are mandatory, except for non-raw-formats */
+ if (!gst_structure_get_int (structure, "width", &width) &&
+ format != GST_VIDEO_FORMAT_ENCODED)
+ goto no_width;
+ if (!gst_structure_get_int (structure, "height", &height) &&
+ format != GST_VIDEO_FORMAT_ENCODED)
+ goto no_height;
+
+ gst_video_info_set_format (info, format, width, height);
+
+ if (gst_structure_get_fraction (structure, "framerate", &fps_n, &fps_d)) {
+ if (fps_n == 0) {
+ /* variable framerate */
+ info->flags |= GST_VIDEO_FLAG_VARIABLE_FPS;
+ /* see if we have a max-framerate */
+ gst_structure_get_fraction (structure, "max-framerate", &fps_n, &fps_d);
+ }
+ info->fps_n = fps_n;
+ info->fps_d = fps_d;
+ } else {
+ /* unspecified is variable framerate */
+ info->fps_n = 0;
+ info->fps_d = 1;
+ }
+
+ if ((s = gst_structure_get_string (structure, "interlace-mode")))
+ info->interlace_mode = gst_interlace_mode_from_string (s);
+ else
+ info->interlace_mode = GST_VIDEO_INTERLACE_MODE_PROGRESSIVE;
+
+ if (gst_structure_get_int (structure, "views", &views))
+ info->views = views;
+ else
+ info->views = 1;
+
+ if ((s = gst_structure_get_string (structure, "chroma-site")))
+ info->chroma_site = gst_video_chroma_from_string (s);
+ else
+ info->chroma_site = GST_VIDEO_CHROMA_SITE_UNKNOWN;
+
+ if ((s = gst_structure_get_string (structure, "colorimetry")))
+ gst_video_colorimetry_from_string (&info->colorimetry, s);
+
+ if (gst_structure_get_fraction (structure, "pixel-aspect-ratio",
+ &par_n, &par_d)) {
+ info->par_n = par_n;
+ info->par_d = par_d;
+ } else {
+ info->par_n = 1;
+ info->par_d = 1;
+ }
+ return TRUE;
+
+ /* ERROR */
+wrong_name:
+ {
+ GST_ERROR ("wrong name '%s', expected video/ or image/",
+ gst_structure_get_name (structure));
+ return FALSE;
+ }
+no_format:
+ {
+ GST_ERROR ("no format given");
+ return FALSE;
+ }
+unknown_format:
+ {
+ GST_ERROR ("unknown format '%s' given", s);
+ return FALSE;
+ }
+no_width:
+ {
+ GST_ERROR ("no width property given");
+ return FALSE;
+ }
+no_height:
+ {
+ GST_ERROR ("no height property given");
+ return FALSE;
+ }
+}
+
+/**
+ * gst_video_info_is_equal:
+ * @info: a #GstVideoInfo
+ * @other: a #GstVideoInfo
+ *
+ * Compares two #GstVideoInfo and returns whether they are equal or not
+ *
+ * Returns: %TRUE if @info and @other are equal, else %FALSE.
+ */
+gboolean
+gst_video_info_is_equal (const GstVideoInfo * info, const GstVideoInfo * other)
+{
+ if (GST_VIDEO_INFO_FORMAT (info) != GST_VIDEO_INFO_FORMAT (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_INTERLACE_MODE (info) !=
+ GST_VIDEO_INFO_INTERLACE_MODE (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_FLAGS (info) != GST_VIDEO_INFO_FLAGS (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_WIDTH (info) != GST_VIDEO_INFO_WIDTH (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_HEIGHT (info) != GST_VIDEO_INFO_HEIGHT (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_SIZE (info) != GST_VIDEO_INFO_SIZE (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_PAR_N (info) != GST_VIDEO_INFO_PAR_N (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_PAR_D (info) != GST_VIDEO_INFO_PAR_D (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_FPS_N (info) != GST_VIDEO_INFO_FPS_N (other))
+ return FALSE;
+ if (GST_VIDEO_INFO_FPS_D (info) != GST_VIDEO_INFO_FPS_D (other))
+ return FALSE;
+ return TRUE;
+}
+
+/**
+ * gst_video_info_to_caps:
+ * @info: a #GstVideoInfo
+ *
+ * Convert the values of @info into a #GstCaps.
+ *
+ * Returns: a new #GstCaps containing the info of @info.
+ */
+GstCaps *
+gst_video_info_to_caps (GstVideoInfo * info)
+{
+ GstCaps *caps;
+ const gchar *format;
+ gchar *color;
+
+ g_return_val_if_fail (info != NULL, NULL);
+ g_return_val_if_fail (info->finfo != NULL, NULL);
+ g_return_val_if_fail (info->finfo->format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
+
+ format = gst_video_format_to_string (info->finfo->format);
+ g_return_val_if_fail (format != NULL, NULL);
+
+ caps = gst_caps_new_simple ("video/x-raw",
+ "format", G_TYPE_STRING, format,
+ "width", G_TYPE_INT, info->width,
+ "height", G_TYPE_INT, info->height,
+ "pixel-aspect-ratio", GST_TYPE_FRACTION, info->par_n, info->par_d, NULL);
+
+ gst_caps_set_simple (caps, "interlace-mode", G_TYPE_STRING,
+ gst_interlace_mode_to_string (info->interlace_mode), NULL);
+
+ if (info->chroma_site != GST_VIDEO_CHROMA_SITE_UNKNOWN)
+ gst_caps_set_simple (caps, "chroma-site", G_TYPE_STRING,
+ gst_video_chroma_to_string (info->chroma_site), NULL);
+
+ if ((color = gst_video_colorimetry_to_string (&info->colorimetry))) {
+ gst_caps_set_simple (caps, "colorimetry", G_TYPE_STRING, color, NULL);
+ g_free (color);
+ }
+
+ if (info->views > 1)
+ gst_caps_set_simple (caps, "views", G_TYPE_INT, info->views, NULL);
+
+ if (info->flags & GST_VIDEO_FLAG_VARIABLE_FPS && info->fps_n != 0) {
+ /* variable fps with a max-framerate */
+ gst_caps_set_simple (caps, "framerate", GST_TYPE_FRACTION, 0, 1,
+ "max-framerate", GST_TYPE_FRACTION, info->fps_n, info->fps_d, NULL);
+ } else {
+ /* no variable fps or no max-framerate */
+ gst_caps_set_simple (caps, "framerate", GST_TYPE_FRACTION,
+ info->fps_n, info->fps_d, NULL);
+ }
+
+ return caps;
+}
+
+static int
+fill_planes (GstVideoInfo * info)
+{
+ gint width, height;
+
+ width = info->width;
+ height = info->height;
+
+ switch (info->finfo->format) {
+ case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
+ case GST_VIDEO_FORMAT_UYVY:
+ info->stride[0] = GST_ROUND_UP_4 (width * 2);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_AYUV:
+ case GST_VIDEO_FORMAT_RGBx:
+ case GST_VIDEO_FORMAT_RGBA:
+ case GST_VIDEO_FORMAT_BGRx:
+ case GST_VIDEO_FORMAT_BGRA:
+ case GST_VIDEO_FORMAT_xRGB:
+ case GST_VIDEO_FORMAT_ARGB:
+ case GST_VIDEO_FORMAT_xBGR:
+ case GST_VIDEO_FORMAT_ABGR:
+ case GST_VIDEO_FORMAT_r210:
+ info->stride[0] = width * 4;
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_RGB16:
+ case GST_VIDEO_FORMAT_BGR16:
+ case GST_VIDEO_FORMAT_RGB15:
+ case GST_VIDEO_FORMAT_BGR15:
+ info->stride[0] = GST_ROUND_UP_4 (width * 2);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_RGB:
+ case GST_VIDEO_FORMAT_BGR:
+ case GST_VIDEO_FORMAT_v308:
+ info->stride[0] = GST_ROUND_UP_4 (width * 3);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_v210:
+ info->stride[0] = ((width + 47) / 48) * 128;
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_v216:
+ info->stride[0] = GST_ROUND_UP_8 (width * 4);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_GRAY8:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ info->stride[0] = GST_ROUND_UP_4 (width * 2);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_UYVP:
+ info->stride[0] = GST_ROUND_UP_4 ((width * 2 * 5 + 3) / 4);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_RGB8P:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = 4;
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * height;
+ info->size = info->offset[1] + (4 * 256);
+ break;
+ case GST_VIDEO_FORMAT_IYU1:
+ info->stride[0] = GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) +
+ GST_ROUND_UP_4 (width) / 2);
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ info->stride[0] = width * 8;
+ info->offset[0] = 0;
+ info->size = info->stride[0] * height;
+ break;
+ case GST_VIDEO_FORMAT_I420:
+ case GST_VIDEO_FORMAT_YV12: /* same as I420, but plane 1+2 swapped */
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
+ info->stride[2] = info->stride[1];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
+ info->offset[2] = info->offset[1] +
+ info->stride[1] * (GST_ROUND_UP_2 (height) / 2);
+ info->size = info->offset[2] +
+ info->stride[2] * (GST_ROUND_UP_2 (height) / 2);
+ break;
+ case GST_VIDEO_FORMAT_Y41B:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = GST_ROUND_UP_16 (width) / 4;
+ info->stride[2] = info->stride[1];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * height;
+ info->offset[2] = info->offset[1] + info->stride[1] * height;
+ /* simplification of ROUNDUP4(w)*h + 2*((ROUNDUP16(w)/4)*h */
+ info->size = (info->stride[0] + (GST_ROUND_UP_16 (width) / 2)) * height;
+ break;
+ case GST_VIDEO_FORMAT_Y42B:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = GST_ROUND_UP_8 (width) / 2;
+ info->stride[2] = info->stride[1];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * height;
+ info->offset[2] = info->offset[1] + info->stride[1] * height;
+ /* simplification of ROUNDUP4(w)*h + 2*(ROUNDUP8(w)/2)*h */
+ info->size = (info->stride[0] + GST_ROUND_UP_8 (width)) * height;
+ break;
+ case GST_VIDEO_FORMAT_Y444:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = info->stride[0];
+ info->stride[2] = info->stride[0];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * height;
+ info->offset[2] = info->offset[1] * 2;
+ info->size = info->stride[0] * height * 3;
+ break;
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = info->stride[0];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
+ info->size = info->stride[0] * GST_ROUND_UP_2 (height) * 3 / 2;
+ break;
+ case GST_VIDEO_FORMAT_A420:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
+ info->stride[2] = info->stride[1];
+ info->stride[3] = info->stride[0];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
+ info->offset[2] = info->offset[1] +
+ info->stride[1] * (GST_ROUND_UP_2 (height) / 2);
+ info->offset[3] = info->offset[2] +
+ info->stride[2] * (GST_ROUND_UP_2 (height) / 2);
+ info->size = info->offset[3] + info->stride[0];
+ break;
+ case GST_VIDEO_FORMAT_YUV9:
+ case GST_VIDEO_FORMAT_YVU9:
+ info->stride[0] = GST_ROUND_UP_4 (width);
+ info->stride[1] = GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4);
+ info->stride[2] = info->stride[1];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * height;
+ info->offset[2] = info->offset[1] +
+ info->stride[1] * (GST_ROUND_UP_4 (height) / 4);
+ info->size = info->offset[2] +
+ info->stride[2] * (GST_ROUND_UP_4 (height) / 4);
+ break;
+ case GST_VIDEO_FORMAT_I420_10LE:
+ case GST_VIDEO_FORMAT_I420_10BE:
+ info->stride[0] = GST_ROUND_UP_4 (width * 2);
+ info->stride[1] = GST_ROUND_UP_4 (width);
+ info->stride[2] = info->stride[1];
+ info->offset[0] = 0;
+ info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
+ info->offset[2] = info->offset[1] +
+ info->stride[1] * (GST_ROUND_UP_2 (height) / 2);
+ info->size = info->offset[2] +
+ info->stride[2] * (GST_ROUND_UP_2 (height) / 2);
+ break;
+ case GST_VIDEO_FORMAT_ENCODED:
+ break;
+ case GST_VIDEO_FORMAT_UNKNOWN:
+ GST_ERROR ("invalid format");
+ g_warning ("invalid format");
+ break;
+ }
+ return 0;
+}
+
+/**
+ * gst_video_info_convert:
+ * @info: a #GstVideoInfo
+ * @src_format: #GstFormat of the @src_value
+ * @src_value: value to convert
+ * @dest_format: #GstFormat of the @dest_value
+ * @dest_value: pointer to destination value
+ *
+ * Converts among various #GstFormat types. This function handles
+ * GST_FORMAT_BYTES, GST_FORMAT_TIME, and GST_FORMAT_DEFAULT. For
+ * raw video, GST_FORMAT_DEFAULT corresponds to video frames. This
+ * function can be used to handle pad queries of the type GST_QUERY_CONVERT.
+ *
+ * Since: 0.10.16
+ *
+ * Returns: TRUE if the conversion was successful.
+ */
+gboolean
+gst_video_info_convert (GstVideoInfo * info,
+ GstFormat src_format, gint64 src_value,
+ GstFormat dest_format, gint64 * dest_value)
+{
+ gboolean ret = FALSE;
+ int size, fps_n, fps_d;
+
+ g_return_val_if_fail (info != NULL, 0);
+ g_return_val_if_fail (info->finfo != NULL, 0);
+ g_return_val_if_fail (info->finfo->format != GST_VIDEO_FORMAT_UNKNOWN, 0);
+ g_return_val_if_fail (info->size > 0, 0);
+
+ size = info->size;
+ fps_n = info->fps_n;
+ fps_d = info->fps_d;
+
+ GST_DEBUG ("converting value %" G_GINT64_FORMAT " from %s to %s",
+ src_value, gst_format_get_name (src_format),
+ gst_format_get_name (dest_format));
+
+ if (src_format == dest_format) {
+ *dest_value = src_value;
+ ret = TRUE;
+ goto done;
+ }
+
+ if (src_value == -1) {
+ *dest_value = -1;
+ ret = TRUE;
+ goto done;
+ }
+
+ /* bytes to frames */
+ if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_DEFAULT) {
+ if (size != 0) {
+ *dest_value = gst_util_uint64_scale_int (src_value, 1, size);
+ } else {
+ GST_ERROR ("blocksize is 0");
+ *dest_value = 0;
+ }
+ ret = TRUE;
+ goto done;
+ }
+
+ /* frames to bytes */
+ if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_BYTES) {
+ *dest_value = gst_util_uint64_scale_int (src_value, size, 1);
+ ret = TRUE;
+ goto done;
+ }
+
+ /* time to frames */
+ if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_DEFAULT) {
+ if (fps_d != 0) {
+ *dest_value = gst_util_uint64_scale (src_value,
+ fps_n, GST_SECOND * fps_d);
+ } else {
+ GST_ERROR ("framerate denominator is 0");
+ *dest_value = 0;
+ }
+ ret = TRUE;
+ goto done;
+ }
+
+ /* frames to time */
+ if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_TIME) {
+ if (fps_n != 0) {
+ *dest_value = gst_util_uint64_scale (src_value,
+ GST_SECOND * fps_d, fps_n);
+ } else {
+ GST_ERROR ("framerate numerator is 0");
+ *dest_value = 0;
+ }
+ ret = TRUE;
+ goto done;
+ }
+
+ /* time to bytes */
+ if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_BYTES) {
+ if (fps_d != 0) {
+ *dest_value = gst_util_uint64_scale (src_value,
+ fps_n * size, GST_SECOND * fps_d);
+ } else {
+ GST_ERROR ("framerate denominator is 0");
+ *dest_value = 0;
+ }
+ ret = TRUE;
+ goto done;
+ }
+
+ /* bytes to time */
+ if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_TIME) {
+ if (fps_n != 0 && size != 0) {
+ *dest_value = gst_util_uint64_scale (src_value,
+ GST_SECOND * fps_d, fps_n * size);
+ } else {
+ GST_ERROR ("framerate denominator and/or blocksize is 0");
+ *dest_value = 0;
+ }
+ ret = TRUE;
+ }
+
+done:
+
+ GST_DEBUG ("ret=%d result %" G_GINT64_FORMAT, ret, *dest_value);
+
+ return ret;
+}
--- /dev/null
+/* GStreamer
+ * Copyright (C) <2011> Wim Taymans <wim.taymans@gmail.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
+ * Boston, MA 02111-1307, USA.
+ */
+
+#ifndef __GST_VIDEO_INFO_H__
+#define __GST_VIDEO_INFO_H__
+
+#include <gst/gst.h>
+#include <gst/video/video-format.h>
+#include <gst/video/video-color.h>
+
+G_BEGIN_DECLS
+
+#include <gst/video/video-enumtypes.h>
+
+typedef struct _GstVideoInfo GstVideoInfo;
+
+/**
+ * GstVideoInterlaceMode:
+ * @GST_VIDEO_INTERLACE_MODE_PROGRESSIVE: all frames are progressive
+ * @GST_VIDEO_INTERLACE_MODE_INTERLEAVED: 2 fields are interleaved in one video
+ * frame. Extra buffer flags describe the field order.
+ * @GST_VIDEO_INTERLACE_MODE_MIXED: frames contains both interlaced and
+ * progressive video, the buffer flags describe the frame and fields.
+ * @GST_VIDEO_INTERLACE_MODE_FIELDS: 2 fields are stored in one buffer, use the
+ * frame ID to get access to the required field. For multiview (the
+ * 'views' property > 1) the fields of view N can be found at frame ID
+ * (N * 2) and (N * 2) + 1.
+ * Each field has only half the amount of lines as noted in the
+ * height property. This mode requires multiple GstVideoMeta metadata
+ * to describe the fields.
+ *
+ * The possible values of the #GstVideoInterlaceMode describing the interlace
+ * mode of the stream.
+ */
+typedef enum {
+ GST_VIDEO_INTERLACE_MODE_PROGRESSIVE = 0,
+ GST_VIDEO_INTERLACE_MODE_INTERLEAVED,
+ GST_VIDEO_INTERLACE_MODE_MIXED,
+ GST_VIDEO_INTERLACE_MODE_FIELDS
+} GstVideoInterlaceMode;
+
+/**
+ * GstVideoFlags:
+ * @GST_VIDEO_FLAG_NONE: no flags
+ * @GST_VIDEO_FLAG_VARIABLE_FPS: a variable fps is selected, fps_n and fps_d
+ * denote the maximum fps of the video
+ * @GST_VIDEO_FLAG_PREMULTIPLIED_ALPHA: Each color has been scaled by the alpha
+ * value.
+ *
+ * Extra video flags
+ */
+typedef enum {
+ GST_VIDEO_FLAG_NONE = 0,
+ GST_VIDEO_FLAG_VARIABLE_FPS = (1 << 0),
+ GST_VIDEO_FLAG_PREMULTIPLIED_ALPHA = (1 << 1)
+} GstVideoFlags;
+
+/**
+ * GstVideoInfo:
+ * @finfo: the format info of the video
+ * @interlace_mode: the interlace mode
+ * @flags: additional video flags
+ * @width: the width of the video
+ * @height: the height of the video
+ * @views: the number of views for multiview video
+ * @size: the default size of one frame
+ * @chroma_site: a #GstVideoChromaSite.
+ * @colorimetry: the colorimetry info
+ * @par_n: the pixel-aspect-ratio numerator
+ * @par_d: the pixel-aspect-ratio demnominator
+ * @fps_n: the framerate numerator
+ * @fps_d: the framerate demnominator
+ * @offset: offsets of the planes
+ * @stride: strides of the planes
+ *
+ * Information describing image properties. This information can be filled
+ * in from GstCaps with gst_video_info_from_caps(). The information is also used
+ * to store the specific video info when mapping a video frame with
+ * gst_video_frame_map().
+ *
+ * Use the provided macros to access the info in this structure.
+ */
+struct _GstVideoInfo {
+ const GstVideoFormatInfo *finfo;
+
+ GstVideoInterlaceMode interlace_mode;
+ GstVideoFlags flags;
+ gint width;
+ gint height;
+ gsize size;
+ gint views;
+
+ GstVideoChromaSite chroma_site;
+ GstVideoColorimetry colorimetry;
+
+ gint par_n;
+ gint par_d;
+ gint fps_n;
+ gint fps_d;
+
+ gsize offset[GST_VIDEO_MAX_PLANES];
+ gint stride[GST_VIDEO_MAX_PLANES];
+};
+
+/* general info */
+#define GST_VIDEO_INFO_FORMAT(i) (GST_VIDEO_FORMAT_INFO_FORMAT((i)->finfo))
+#define GST_VIDEO_INFO_NAME(i) (GST_VIDEO_FORMAT_INFO_NAME((i)->finfo))
+#define GST_VIDEO_INFO_IS_YUV(i) (GST_VIDEO_FORMAT_INFO_IS_YUV((i)->finfo))
+#define GST_VIDEO_INFO_IS_RGB(i) (GST_VIDEO_FORMAT_INFO_IS_RGB((i)->finfo))
+#define GST_VIDEO_INFO_IS_GRAY(i) (GST_VIDEO_FORMAT_INFO_IS_GRAY((i)->finfo))
+#define GST_VIDEO_INFO_HAS_ALPHA(i) (GST_VIDEO_FORMAT_INFO_HAS_ALPHA((i)->finfo))
+
+#define GST_VIDEO_INFO_INTERLACE_MODE(i) ((i)->interlace_mode)
+#define GST_VIDEO_INFO_IS_INTERLACED(i) ((i)->interlace_mode != GST_VIDEO_INTERLACE_MODE_PROGRESSIVE)
+#define GST_VIDEO_INFO_FLAGS(i) ((i)->flags)
+#define GST_VIDEO_INFO_WIDTH(i) ((i)->width)
+#define GST_VIDEO_INFO_HEIGHT(i) ((i)->height)
+#define GST_VIDEO_INFO_SIZE(i) ((i)->size)
+#define GST_VIDEO_INFO_VIEWS(i) ((i)->views)
+#define GST_VIDEO_INFO_PAR_N(i) ((i)->par_n)
+#define GST_VIDEO_INFO_PAR_D(i) ((i)->par_d)
+#define GST_VIDEO_INFO_FPS_N(i) ((i)->fps_n)
+#define GST_VIDEO_INFO_FPS_D(i) ((i)->fps_d)
+
+/* dealing with GstVideoInfo flags */
+#define GST_VIDEO_INFO_FLAG_IS_SET(i,flag) ((GST_VIDEO_INFO_FLAGS(i) & (flag)) == (flag))
+#define GST_VIDEO_INFO_FLAG_SET(i,flag) (GST_VIDEO_INFO_FLAGS(i) |= (flag))
+#define GST_VIDEO_INFO_FLAG_UNSET(i,flag) (GST_VIDEO_INFO_FLAGS(i) &= ~(flag))
+
+/* dealing with planes */
+#define GST_VIDEO_INFO_N_PLANES(i) (GST_VIDEO_FORMAT_INFO_N_PLANES((i)->finfo))
+#define GST_VIDEO_INFO_PLANE_OFFSET(i,p) ((i)->offset[p])
+#define GST_VIDEO_INFO_PLANE_STRIDE(i,p) ((i)->stride[p])
+
+/* dealing with components */
+#define GST_VIDEO_INFO_N_COMPONENTS(i) GST_VIDEO_FORMAT_INFO_N_COMPONENTS((i)->finfo)
+#define GST_VIDEO_INFO_COMP_DEPTH(i,c) GST_VIDEO_FORMAT_INFO_DEPTH((i)->finfo,(c))
+#define GST_VIDEO_INFO_COMP_DATA(i,d,c) GST_VIDEO_FORMAT_INFO_DATA((i)->finfo,d,(c))
+#define GST_VIDEO_INFO_COMP_OFFSET(i,c) GST_VIDEO_FORMAT_INFO_OFFSET((i)->finfo,(i)->offset,(c))
+#define GST_VIDEO_INFO_COMP_STRIDE(i,c) GST_VIDEO_FORMAT_INFO_STRIDE((i)->finfo,(i)->stride,(c))
+#define GST_VIDEO_INFO_COMP_WIDTH(i,c) GST_VIDEO_FORMAT_INFO_SCALE_WIDTH((i)->finfo,(c),(i)->width)
+#define GST_VIDEO_INFO_COMP_HEIGHT(i,c) GST_VIDEO_FORMAT_INFO_SCALE_HEIGHT((i)->finfo,(c),(i)->height)
+#define GST_VIDEO_INFO_COMP_PLANE(i,c) GST_VIDEO_FORMAT_INFO_PLANE((i)->finfo,(c))
+#define GST_VIDEO_INFO_COMP_PSTRIDE(i,c) GST_VIDEO_FORMAT_INFO_PSTRIDE((i)->finfo,(c))
+#define GST_VIDEO_INFO_COMP_POFFSET(i,c) GST_VIDEO_FORMAT_INFO_POFFSET((i)->finfo,(c))
+
+void gst_video_info_init (GstVideoInfo *info);
+
+void gst_video_info_set_format (GstVideoInfo *info, GstVideoFormat format,
+ guint width, guint height);
+
+gboolean gst_video_info_from_caps (GstVideoInfo *info, const GstCaps * caps);
+
+GstCaps * gst_video_info_to_caps (GstVideoInfo *info);
+
+gboolean gst_video_info_convert (GstVideoInfo *info,
+ GstFormat src_format,
+ gint64 src_value,
+ GstFormat dest_format,
+ gint64 *dest_value);
+gboolean gst_video_info_is_equal (const GstVideoInfo *info,
+ const GstVideoInfo *other);
+
+G_END_DECLS
+
+#endif /* __GST_VIDEO_INFO_H__ */
* </refsect2>
*/
-GST_DEBUG_CATEGORY_EXTERN (GST_CAT_PERFORMANCE);
-
-static int fill_planes (GstVideoInfo * info);
-
-#include "videoblendorc.h"
-
-/* Line conversion to AYUV */
-
-#define GET_PLANE_STRIDE(plane) (stride(plane))
-#define GET_PLANE_LINE(plane, line) \
- (gpointer)(((guint8*)(data[plane])) + stride[plane] * (line))
-
-#define GET_COMP_STRIDE(comp) \
- GST_VIDEO_FORMAT_INFO_STRIDE (info, stride, comp)
-#define GET_COMP_DATA(comp) \
- GST_VIDEO_FORMAT_INFO_DATA (info, data, comp)
-
-#define GET_COMP_LINE(comp, line) \
- (gpointer)(((guint8*)GET_COMP_DATA (comp)) + \
- GET_COMP_STRIDE(comp) * (line))
-
-#define GET_STRIDE() GET_PLANE_STRIDE (0)
-#define GET_LINE(line) GET_PLANE_LINE (0, line)
-
-#define GET_Y_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_Y, line)
-#define GET_U_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_U, line)
-#define GET_V_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_V, line)
-#define GET_A_LINE(line) GET_COMP_LINE(GST_VIDEO_COMP_A, line)
-
-#define GET_Y_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_Y)
-#define GET_U_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_U)
-#define GET_V_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_V)
-#define GET_A_STRIDE() GET_COMP_STRIDE(GST_VIDEO_COMP_A)
-
-#define PACK_420 GST_VIDEO_FORMAT_AYUV, unpack_planar_420, pack_planar_420
-static void
-unpack_planar_420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_I420 (dest,
- GET_Y_LINE (y), GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), width);
-}
-
-static void
-pack_planar_420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_I420 (GET_Y_LINE (y),
- GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), src, width / 2);
-}
-
-#define PACK_YUY2 GST_VIDEO_FORMAT_AYUV, unpack_YUY2, pack_YUY2
-static void
-unpack_YUY2 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_YUY2 (dest, GET_LINE (y), width / 2);
-}
-
-static void
-pack_YUY2 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_YUY2 (GET_LINE (y), src, width / 2);
-}
-
-#define PACK_UYVY GST_VIDEO_FORMAT_AYUV, unpack_UYVY, pack_UYVY
-static void
-unpack_UYVY (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_UYVY (dest, GET_LINE (y), width / 2);
-}
-
-static void
-pack_UYVY (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_UYVY (GET_LINE (y), src, width / 2);
-}
-
-#define PACK_YVYU GST_VIDEO_FORMAT_AYUV, unpack_YVYU, pack_YVYU
-static void
-unpack_YVYU (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_YVYU (dest, GET_LINE (y), width / 2);
-}
-
-static void
-pack_YVYU (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_YVYU (GET_LINE (y), src, width / 2);
-}
-
-#define PACK_v308 GST_VIDEO_FORMAT_AYUV, unpack_v308, pack_v308
-static void
-unpack_v308 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 1] = s[i * 3 + 0];
- d[i * 4 + 2] = s[i * 3 + 1];
- d[i * 4 + 3] = s[i * 3 + 2];
- }
-}
-
-static void
-pack_v308 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i * 3 + 0] = s[i * 4 + 1];
- d[i * 3 + 1] = s[i * 4 + 2];
- d[i * 3 + 2] = s[i * 4 + 3];
- }
-}
-
-#define PACK_AYUV GST_VIDEO_FORMAT_AYUV, unpack_copy4, pack_copy4
-#define PACK_ARGB GST_VIDEO_FORMAT_ARGB, unpack_copy4, pack_copy4
-static void
-unpack_copy4 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- memcpy (dest, GET_LINE (y), width * 4);
-}
-
-static void
-pack_copy4 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- memcpy (GET_LINE (y), src, width * 4);
-}
-
-#define PACK_v210 GST_VIDEO_FORMAT_AYUV64, unpack_v210, pack_v210
-static void
-unpack_v210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint16 *d = dest;
-
- for (i = 0; i < width; i += 6) {
- guint32 a0, a1, a2, a3;
- guint16 y0, y1, y2, y3, y4, y5;
- guint16 u0, u2, u4;
- guint16 v0, v2, v4;
-
- a0 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 0);
- a1 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 4);
- a2 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 8);
- a3 = GST_READ_UINT32_LE (s + (i / 6) * 16 + 12);
-
- u0 = ((a0 >> 0) & 0x3ff) << 6;
- y0 = ((a0 >> 10) & 0x3ff) << 6;
- v0 = ((a0 >> 20) & 0x3ff) << 6;
- y1 = ((a1 >> 0) & 0x3ff) << 6;
-
- u2 = ((a1 >> 10) & 0x3ff) << 6;
- y2 = ((a1 >> 20) & 0x3ff) << 6;
- v2 = ((a2 >> 0) & 0x3ff) << 6;
- y3 = ((a2 >> 10) & 0x3ff) << 6;
-
- u4 = ((a2 >> 20) & 0x3ff) << 6;
- y4 = ((a3 >> 0) & 0x3ff) << 6;
- v4 = ((a3 >> 10) & 0x3ff) << 6;
- y5 = ((a3 >> 20) & 0x3ff) << 6;
-
- d[4 * (i + 0) + 0] = 0xffff;
- d[4 * (i + 0) + 1] = y0;
- d[4 * (i + 0) + 2] = u0;
- d[4 * (i + 0) + 3] = v0;
-
- d[4 * (i + 1) + 0] = 0xffff;
- d[4 * (i + 1) + 1] = y1;
- d[4 * (i + 1) + 2] = u0;
- d[4 * (i + 1) + 3] = v0;
-
- d[4 * (i + 2) + 0] = 0xffff;
- d[4 * (i + 2) + 1] = y2;
- d[4 * (i + 2) + 2] = u2;
- d[4 * (i + 2) + 3] = v2;
-
- d[4 * (i + 3) + 0] = 0xffff;
- d[4 * (i + 3) + 1] = y3;
- d[4 * (i + 3) + 2] = u2;
- d[4 * (i + 3) + 3] = v2;
-
- d[4 * (i + 4) + 0] = 0xffff;
- d[4 * (i + 4) + 1] = y4;
- d[4 * (i + 4) + 2] = u4;
- d[4 * (i + 4) + 3] = v4;
-
- d[4 * (i + 5) + 0] = 0xffff;
- d[4 * (i + 5) + 1] = y5;
- d[4 * (i + 5) + 2] = u4;
- d[4 * (i + 5) + 3] = v4;
- }
-}
-
-static void
-pack_v210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint16 *s = src;
-
- for (i = 0; i < width; i += 6) {
- guint32 a0, a1, a2, a3;
- guint16 y0, y1, y2, y3, y4, y5;
- guint16 u0, u1, u2;
- guint16 v0, v1, v2;
-
- y0 = s[4 * (i + 0) + 1] >> 6;
- y1 = s[4 * (i + 1) + 1] >> 6;
- y2 = s[4 * (i + 2) + 1] >> 6;
- y3 = s[4 * (i + 3) + 1] >> 6;
- y4 = s[4 * (i + 4) + 1] >> 6;
- y5 = s[4 * (i + 5) + 1] >> 6;
-
- u0 = (s[4 * (i + 0) + 2] + s[4 * (i + 1) + 2] + 1) >> 7;
- u1 = (s[4 * (i + 2) + 2] + s[4 * (i + 3) + 2] + 1) >> 7;
- u2 = (s[4 * (i + 4) + 2] + s[4 * (i + 5) + 2] + 1) >> 7;
-
- v0 = (s[4 * (i + 0) + 3] + s[4 * (i + 1) + 3] + 1) >> 7;
- v1 = (s[4 * (i + 2) + 3] + s[4 * (i + 3) + 3] + 1) >> 7;
- v2 = (s[4 * (i + 4) + 3] + s[4 * (i + 5) + 3] + 1) >> 7;
-
- a0 = u0 | (y0 << 10) | (v0 << 20);
- a1 = y1 | (u1 << 10) | (y2 << 20);
- a2 = v1 | (y3 << 10) | (u2 << 20);
- a3 = y4 | (v2 << 10) | (y5 << 20);
-
- GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 0, a0);
- GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 4, a1);
- GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 8, a2);
- GST_WRITE_UINT32_LE (d + (i / 6) * 16 + 12, a3);
- }
-}
-
-#define PACK_v216 GST_VIDEO_FORMAT_AYUV64, unpack_v216, pack_v216
-static void
-unpack_v216 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint16 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xffff;
- d[i * 4 + 1] = GST_READ_UINT16_LE (s + i * 4 + 2);
- d[i * 4 + 2] = GST_READ_UINT16_LE (s + (i >> 1) * 8 + 0);
- d[i * 4 + 3] = GST_READ_UINT16_LE (s + (i >> 1) * 8 + 4);
- }
-}
-
-static void
-pack_v216 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint16 *s = src;
-
- for (i = 0; i < width / 2; i++) {
- GST_WRITE_UINT16_LE (d + i * 8 + 0, s[(i * 2 + 0) * 4 + 2]);
- GST_WRITE_UINT16_LE (d + i * 8 + 2, s[(i * 2 + 0) * 4 + 1]);
- GST_WRITE_UINT16_LE (d + i * 8 + 4, s[(i * 2 + 0) * 4 + 3]);
- GST_WRITE_UINT16_LE (d + i * 8 + 6, s[(i * 2 + 1) * 4 + 1]);
- }
-}
-
-#define PACK_Y41B GST_VIDEO_FORMAT_AYUV, unpack_Y41B, pack_Y41B
-static void
-unpack_Y41B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_YUV9 (dest,
- GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), width / 2);
-}
-
-static void
-pack_Y41B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *destY = GET_Y_LINE (y);
- guint8 *destU = GET_U_LINE (y);
- guint8 *destV = GET_V_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width - 3; i += 4) {
- destY[i] = s[i * 4 + 1];
- destY[i + 1] = s[i * 4 + 5];
- destY[i + 2] = s[i * 4 + 9];
- destY[i + 3] = s[i * 4 + 13];
-
- destU[i >> 2] =
- (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + s[i * 4 + 14] + 2) >> 2;
- destV[i >> 2] =
- (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + s[i * 4 + 15] + 2) >> 2;
- }
-
- if (i == width - 3) {
- destY[i] = s[i * 4 + 1];
- destY[i + 1] = s[i * 4 + 5];
- destY[i + 2] = s[i * 4 + 9];
-
- destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + 1) / 3;
- destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + 1) / 3;
- } else if (i == width - 2) {
- destY[i] = s[i * 4 + 1];
- destY[i + 1] = s[i * 4 + 5];
-
- destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6] + 1) >> 1;
- destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7] + 1) >> 1;
- } else if (i == width - 1) {
- destY[i + 1] = s[i * 4 + 5];
-
- destU[i >> 2] = s[i * 4 + 2];
- destV[i >> 2] = s[i * 4 + 3];
- }
-}
-
-#define PACK_Y42B GST_VIDEO_FORMAT_AYUV, unpack_Y42B, pack_Y42B
-static void
-unpack_Y42B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_Y42B (dest,
- GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), width / 2);
-}
-
-static void
-pack_Y42B (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_Y42B (GET_Y_LINE (y),
- GET_U_LINE (y), GET_V_LINE (y), src, width / 2);
-}
-
-#define PACK_Y444 GST_VIDEO_FORMAT_AYUV, unpack_Y444, pack_Y444
-static void
-unpack_Y444 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_Y444 (dest, GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), width);
-}
-
-static void
-pack_Y444 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_Y444 (GET_Y_LINE (y), GET_U_LINE (y), GET_V_LINE (y), src, width);
-}
-
-#define PACK_GRAY8 GST_VIDEO_FORMAT_AYUV, unpack_GRAY8, pack_GRAY8
-static void
-unpack_GRAY8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_GRAY8 (dest, GET_LINE (y), width);
-}
-
-static void
-pack_GRAY8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_GRAY8 (GET_LINE (y), src, width);
-}
-
-#define PACK_GRAY16_BE GST_VIDEO_FORMAT_AYUV64, unpack_GRAY16_BE, pack_GRAY16_BE
-static void
-unpack_GRAY16_BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint16 *s = GET_LINE (y);
- guint16 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xffff;
- d[i * 4 + 1] = GST_READ_UINT16_BE (s + i);
- d[i * 4 + 2] = 0x8000;
- d[i * 4 + 3] = 0x8000;
- }
-}
-
-static void
-pack_GRAY16_BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *d = GET_LINE (y);
- const guint16 *s = src;
-
- for (i = 0; i < width; i++) {
- GST_WRITE_UINT16_BE (d + i, s[i * 4 + 1]);
- }
-}
-
-#define PACK_GRAY16_LE GST_VIDEO_FORMAT_AYUV64, unpack_GRAY16_LE, pack_GRAY16_LE
-static void
-unpack_GRAY16_LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint16 *s = GET_LINE (y);
- guint16 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xffff;
- d[i * 4 + 1] = GST_READ_UINT16_LE (s + i);
- d[i * 4 + 2] = 0x8000;
- d[i * 4 + 3] = 0x8000;
- }
-}
-
-static void
-pack_GRAY16_LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *d = GET_LINE (y);
- const guint16 *s = src;
-
- for (i = 0; i < width; i++) {
- GST_WRITE_UINT16_LE (d + i, s[i * 4 + 1]);
- }
-}
-
-#define PACK_RGB16 GST_VIDEO_FORMAT_ARGB, unpack_RGB16, pack_RGB16
-static void
-unpack_RGB16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint16 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 1] = ((s[i] >> 11) & 0x1f) << 3;
- d[i * 4 + 2] = ((s[i] >> 5) & 0x3f) << 2;
- d[i * 4 + 3] = ((s[i]) & 0x1f) << 3;
- }
-}
-
-static void
-pack_RGB16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i] = ((s[i * 4 + 1] >> 3) << 11) |
- ((s[i * 4 + 2] >> 2) << 5) | (s[i * 4 + 3] >> 3);
- }
-}
-
-#define PACK_BGR16 GST_VIDEO_FORMAT_ARGB, unpack_BGR16, pack_BGR16
-static void
-unpack_BGR16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint16 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 3] = ((s[i] >> 11) & 0x1f) << 3;
- d[i * 4 + 2] = ((s[i] >> 5) & 0x3f) << 2;
- d[i * 4 + 1] = ((s[i]) & 0x1f) << 3;
- }
-}
-
-static void
-pack_BGR16 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i] = ((s[i * 4 + 3] >> 3) << 11) |
- ((s[i * 4 + 2] >> 2) << 5) | (s[i * 4 + 1] >> 3);
- }
-}
-
-#define PACK_RGB15 GST_VIDEO_FORMAT_ARGB, unpack_RGB15, pack_RGB15
-static void
-unpack_RGB15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint16 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 1] = ((s[i] >> 10) & 0x1f) << 3;
- d[i * 4 + 2] = ((s[i] >> 5) & 0x1f) << 3;
- d[i * 4 + 3] = ((s[i]) & 0x1f) << 3;
- }
-}
-
-static void
-pack_RGB15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i] = ((s[i * 4 + 1] >> 3) << 10) |
- ((s[i * 4 + 2] >> 3) << 5) | (s[i * 4 + 3] >> 3);
- }
-}
-
-#define PACK_BGR15 GST_VIDEO_FORMAT_ARGB, unpack_BGR15, pack_BGR15
-static void
-unpack_BGR15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint16 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 3] = ((s[i] >> 10) & 0x1f) << 3;
- d[i * 4 + 2] = ((s[i] >> 5) & 0x1f) << 3;
- d[i * 4 + 1] = ((s[i]) & 0x1f) << 3;
- }
-}
-
-static void
-pack_BGR15 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i] = ((s[i * 4 + 3] >> 3) << 10) |
- ((s[i * 4 + 2] >> 3) << 5) | (s[i * 4 + 1] >> 3);
- }
-}
-
-#define PACK_BGRA GST_VIDEO_FORMAT_ARGB, unpack_BGRA, pack_BGRA
-static void
-unpack_BGRA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_BGRA (dest, GET_LINE (y), width);
-}
-
-static void
-pack_BGRA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_BGRA (GET_LINE (y), src, width);
-}
-
-#define PACK_ABGR GST_VIDEO_FORMAT_ARGB, unpack_ABGR, pack_ABGR
-static void
-unpack_ABGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_ABGR (dest, GET_LINE (y), width);
-}
-
-static void
-pack_ABGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_ABGR (GET_LINE (y), src, width);
-}
-
-#define PACK_RGBA GST_VIDEO_FORMAT_ARGB, unpack_RGBA, pack_RGBA
-static void
-unpack_RGBA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_RGBA (dest, GET_LINE (y), width);
-}
-
-static void
-pack_RGBA (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_RGBA (GET_LINE (y), src, width);
-}
-
-#define PACK_RGB GST_VIDEO_FORMAT_ARGB, unpack_RGB, pack_RGB
-static void
-unpack_RGB (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 1] = s[i * 3 + 0];
- d[i * 4 + 2] = s[i * 3 + 1];
- d[i * 4 + 3] = s[i * 3 + 2];
- }
-}
-
-static void
-pack_RGB (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i * 3 + 0] = s[i * 4 + 1];
- d[i * 3 + 1] = s[i * 4 + 2];
- d[i * 3 + 2] = s[i * 4 + 3];
- }
-}
-
-#define PACK_BGR GST_VIDEO_FORMAT_ARGB, unpack_BGR, pack_BGR
-static void
-unpack_BGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 1] = s[i * 3 + 2];
- d[i * 4 + 2] = s[i * 3 + 1];
- d[i * 4 + 3] = s[i * 3 + 0];
- }
-}
-
-static void
-pack_BGR (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width; i++) {
- d[i * 3 + 0] = s[i * 4 + 3];
- d[i * 3 + 1] = s[i * 4 + 2];
- d[i * 3 + 2] = s[i * 4 + 1];
- }
-}
-
-#define PACK_NV12 GST_VIDEO_FORMAT_AYUV, unpack_NV12, pack_NV12
-static void
-unpack_NV12 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_NV12 (dest,
- GET_PLANE_LINE (0, y), GET_PLANE_LINE (1, y >> 1), width / 2);
-}
-
-static void
-pack_NV12 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_NV12 (GET_PLANE_LINE (0, y),
- GET_PLANE_LINE (1, y >> 1), src, width / 2);
-}
-
-#define PACK_NV21 GST_VIDEO_FORMAT_AYUV, unpack_NV21, pack_NV21
-static void
-unpack_NV21 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_NV21 (dest,
- GET_PLANE_LINE (0, y), GET_PLANE_LINE (1, y >> 1), width / 2);
-}
-
-static void
-pack_NV21 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_NV21 (GET_PLANE_LINE (0, y),
- GET_PLANE_LINE (1, y >> 1), src, width / 2);
-}
-
-#define PACK_UYVP GST_VIDEO_FORMAT_AYUV64, unpack_UYVP, pack_UYVP
-static void
-unpack_UYVP (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint16 *d = dest;
-
- for (i = 0; i < width; i += 2) {
- guint16 y0, y1;
- guint16 u0;
- guint16 v0;
-
- u0 = (s[(i / 2) * 5 + 0] << 2) | (s[(i / 2) * 5 + 1] >> 6);
- y0 = ((s[(i / 2) * 5 + 1] & 0x3f) << 4) | (s[(i / 2) * 5 + 2] >> 4);
- v0 = ((s[(i / 2) * 5 + 2] & 0x0f) << 6) | (s[(i / 2) * 5 + 3] >> 2);
- y1 = ((s[(i / 2) * 5 + 3] & 0x03) << 8) | s[(i / 2) * 5 + 4];
-
- d[i * 4 + 0] = 0xffff;
- d[i * 4 + 1] = y0 << 6;
- d[i * 4 + 2] = u0 << 6;
- d[i * 4 + 3] = v0 << 6;
- d[i * 4 + 4] = 0xffff;
- d[i * 4 + 5] = y1 << 6;
- d[i * 4 + 6] = u0 << 6;
- d[i * 4 + 7] = v0 << 6;
- }
-}
-
-static void
-pack_UYVP (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint16 *s = src;
-
- for (i = 0; i < width; i += 2) {
- guint16 y0, y1;
- guint16 u0;
- guint16 v0;
-
- y0 = s[4 * (i + 0) + 1];
- y1 = s[4 * (i + 1) + 1];
- u0 = (s[4 * (i + 0) + 2] + s[4 * (i + 1) + 2] + 1) >> 1;
- v0 = (s[4 * (i + 0) + 3] + s[4 * (i + 1) + 3] + 1) >> 1;
-
- d[(i / 2) * 5 + 0] = u0 >> 8;
- d[(i / 2) * 5 + 1] = (u0 & 0xc0) | y0 >> 10;
- d[(i / 2) * 5 + 2] = ((y0 & 0x3c0) >> 2) | (v0 >> 12);
- d[(i / 2) * 5 + 3] = ((v0 & 0xfc0) >> 4) | (y1 >> 14);
- d[(i / 2) * 5 + 4] = (y1 >> 6);
- }
-}
-
-#define PACK_A420 GST_VIDEO_FORMAT_AYUV, unpack_A420, pack_A420
-static void
-unpack_A420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_A420 (dest,
- GET_Y_LINE (y), GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), GET_A_LINE (y),
- width);
-}
-
-static void
-pack_A420 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- orc_pack_A420 (GET_Y_LINE (y),
- GET_U_LINE (y >> 1), GET_V_LINE (y >> 1), GET_A_LINE (y), src, width / 2);
-}
-
-#define PACK_RGB8P GST_VIDEO_FORMAT_ARGB, unpack_RGB8P, pack_RGB8P
-static void
-unpack_RGB8P (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- const guint32 *p = data[1];
- guint8 *d = dest;
-
- for (i = 0; i < width; i++) {
- guint32 v = p[s[i]];
- d[i * 4 + 0] = (v >> 24) & 0xff;
- d[i * 4 + 1] = (v >> 16) & 0xff;
- d[i * 4 + 2] = (v >> 8) & 0xff;
- d[i * 4 + 3] = (v) & 0xff;
- }
-}
-
-static void
-pack_RGB8P (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint8 *s = src;
-
- /* Use our poor man's palette, taken from ffmpegcolorspace too */
- for (i = 0; i < width; i++) {
- /* crude approximation for alpha ! */
- if (s[i * 4 + 0] < 0x80)
- d[i] = 6 * 6 * 6;
- else
- d[i] =
- ((((s[i * 4 + 1]) / 47) % 6) * 6 * 6 + (((s[i * 4 +
- 2]) / 47) % 6) * 6 + (((s[i * 4 + 3]) / 47) % 6));
- }
-}
-
-#define PACK_410 GST_VIDEO_FORMAT_AYUV, unpack_410, pack_410
-static void
-unpack_410 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- orc_unpack_YUV9 (dest,
- GET_Y_LINE (y), GET_U_LINE (y >> 2), GET_V_LINE (y >> 2), width / 2);
-}
-
-static void
-pack_410 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *destY = GET_Y_LINE (y);
- guint8 *destU = GET_U_LINE (y >> 2);
- guint8 *destV = GET_V_LINE (y >> 2);
- const guint8 *s = src;
-
- for (i = 0; i < width - 3; i += 4) {
- destY[i] = s[i * 4 + 1];
- destY[i + 1] = s[i * 4 + 5];
- destY[i + 2] = s[i * 4 + 9];
- destY[i + 3] = s[i * 4 + 13];
- if (y % 4 == 0) {
- destU[i >> 2] =
- (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + s[i * 4 + 14]) >> 2;
- destV[i >> 2] =
- (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + s[i * 4 + 15]) >> 2;
- }
- }
-
- if (i == width - 3) {
- destY[i] = s[i * 4 + 1];
- destY[i + 1] = s[i * 4 + 5];
- destY[i + 2] = s[i * 4 + 9];
- if (y % 4 == 0) {
- destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10]) / 3;
- destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11]) / 3;
- }
- } else if (i == width - 2) {
- destY[i] = s[i * 4 + 1];
- destY[i + 1] = s[i * 4 + 5];
- if (y % 4 == 0) {
- destU[i >> 2] = (s[i * 4 + 2] + s[i * 4 + 6]) >> 1;
- destV[i >> 2] = (s[i * 4 + 3] + s[i * 4 + 7]) >> 1;
- }
- } else if (i == width - 1) {
- destY[i] = s[i * 4 + 1];
- destU[i >> 2] = s[i * 4 + 2];
- destV[i >> 2] = s[i * 4 + 3];
- }
-}
-
-#define PACK_IYU1 GST_VIDEO_FORMAT_AYUV, unpack_IYU1, pack_IYU1
-static void
-unpack_IYU1 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint8 *d = dest;
-
- for (i = 0; i < width - 3; i += 4) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 4] = 0xff;
- d[i * 4 + 8] = 0xff;
- d[i * 4 + 12] = 0xff;
- d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
- d[i * 4 + 5] = s[(i >> 2) * 6 + 2];
- d[i * 4 + 9] = s[(i >> 2) * 6 + 4];
- d[i * 4 + 13] = s[(i >> 2) * 6 + 5];
- d[i * 4 + 2] = d[i * 4 + 6] = d[i * 4 + 10] = d[i * 4 + 14] =
- s[(i >> 2) * 6 + 0];
- d[i * 4 + 3] = d[i * 4 + 7] = d[i * 4 + 11] = d[i * 4 + 15] =
- s[(i >> 2) * 6 + 3];
- }
-
- if (i == width - 3) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 4] = 0xff;
- d[i * 4 + 8] = 0xff;
- d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
- d[i * 4 + 5] = s[(i >> 2) * 6 + 2];
- d[i * 4 + 9] = s[(i >> 2) * 6 + 4];
- d[i * 4 + 2] = d[i * 4 + 6] = d[i * 4 + 10] = s[(i >> 2) * 6 + 0];
- d[i * 4 + 3] = d[i * 4 + 7] = d[i * 4 + 11] = s[(i >> 2) * 6 + 3];
- } else if (i == width - 2) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 4] = 0xff;
- d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
- d[i * 4 + 5] = s[(i >> 2) * 6 + 2];
- d[i * 4 + 2] = d[i * 4 + 6] = s[(i >> 2) * 6 + 0];
- d[i * 4 + 3] = d[i * 4 + 7] = s[(i >> 2) * 6 + 3];
- } else if (i == width - 1) {
- d[i * 4 + 0] = 0xff;
- d[i * 4 + 1] = s[(i >> 2) * 6 + 1];
- d[i * 4 + 2] = s[(i >> 2) * 6 + 0];
- d[i * 4 + 3] = s[(i >> 2) * 6 + 3];
- }
-}
-
-static void
-pack_IYU1 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint8 *s = src;
-
- for (i = 0; i < width - 3; i += 4) {
- d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
- d[(i >> 2) * 6 + 2] = s[i * 4 + 5];
- d[(i >> 2) * 6 + 4] = s[i * 4 + 9];
- d[(i >> 2) * 6 + 5] = s[i * 4 + 13];
- d[(i >> 2) * 6 + 0] =
- (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10] + s[i * 4 + 14]) >> 2;
- d[(i >> 2) * 6 + 3] =
- (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11] + s[i * 4 + 15]) >> 2;
- }
-
- if (i == width - 3) {
- d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
- d[(i >> 2) * 6 + 2] = s[i * 4 + 5];
- d[(i >> 2) * 6 + 4] = s[i * 4 + 9];
- d[(i >> 2) * 6 + 0] = (s[i * 4 + 2] + s[i * 4 + 6] + s[i * 4 + 10]) / 3;
- d[(i >> 2) * 6 + 3] = (s[i * 4 + 3] + s[i * 4 + 7] + s[i * 4 + 11]) / 3;
- } else if (i == width - 2) {
- d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
- d[(i >> 2) * 6 + 2] = s[i * 4 + 5];
- d[(i >> 2) * 6 + 0] = (s[i * 4 + 2] + s[i * 4 + 6]) >> 1;
- d[(i >> 2) * 6 + 3] = (s[i * 4 + 3] + s[i * 4 + 7]) >> 1;
- } else if (i == width - 1) {
- d[(i >> 2) * 6 + 1] = s[i * 4 + 1];
- d[(i >> 2) * 6 + 0] = s[i * 4 + 2];
- d[(i >> 2) * 6 + 3] = s[i * 4 + 3];
- }
-}
-
-#define PACK_ARGB64 GST_VIDEO_FORMAT_ARGB64, unpack_copy8, pack_copy8
-#define PACK_AYUV64 GST_VIDEO_FORMAT_AYUV64, unpack_copy8, pack_copy8
-static void
-unpack_copy8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- memcpy (dest, GET_LINE (y), width * 8);
-}
-
-static void
-pack_copy8 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- memcpy (GET_LINE (y), src, width * 8);
-}
-
-#define PACK_r210 GST_VIDEO_FORMAT_AYUV64, unpack_r210, pack_r210
-static void
-unpack_r210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- const guint8 *s = GET_LINE (y);
- guint16 *d = dest;
-
- for (i = 0; i < width; i++) {
- guint32 x;
- d[i * 4 + 0] = 0xffff;
- x = GST_READ_UINT32_BE (s + i * 4);
- d[i * 4 + 1] = ((x >> 14) & 0xffc0) | (x >> 24);
- d[i * 4 + 2] = ((x >> 4) & 0xffc0) | ((x >> 14) & 0x3f);
- d[i * 4 + 3] = ((x << 6) & 0xffc0) | ((x >> 4) & 0x3f);
- }
-}
-
-static void
-pack_r210 (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint8 *d = GET_LINE (y);
- const guint16 *s = src;
-
- for (i = 0; i < width; i++) {
- guint32 x = 0;
- x |= (s[i * 4 + 1] & 0xffc0) << 14;
- x |= (s[i * 4 + 2] & 0xffc0) << 4;
- x |= (s[i * 4 + 3] & 0xffc0) >> 6;
- GST_WRITE_UINT32_BE (d + i * 4, x);
- }
-}
-
-#define PACK_I420_10LE GST_VIDEO_FORMAT_AYUV64, unpack_I420_10LE, pack_I420_10LE
-static void
-unpack_I420_10LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- guint16 *srcY = GET_Y_LINE (y);
- guint16 *srcU = GET_U_LINE (y >> 1);
- guint16 *srcV = GET_V_LINE (y >> 1);
- guint16 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xffff;
- d[i * 4 + 1] = GST_READ_UINT16_LE (srcY + i) << 6;
- d[i * 4 + 2] = GST_READ_UINT16_LE (srcU + (i >> 1)) << 6;
- d[i * 4 + 3] = GST_READ_UINT16_LE (srcV + (i >> 1)) << 6;
- }
-}
-
-static void
-pack_I420_10LE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *destY = GET_Y_LINE (y);
- guint16 *destU = GET_U_LINE (y >> 1);
- guint16 *destV = GET_V_LINE (y >> 1);
- guint16 Y0, Y1, U, V;
- const guint16 *s = src;
-
- for (i = 0; i < width - 1; i += 2) {
- Y0 = (s[i * 4 + 1]) >> 6;
- Y1 = (s[i * 4 + 5]) >> 6;
- U = ((s[i * 4 + 2] + s[i * 4 + 6] + 1) >> 1) >> 6;
- V = ((s[i * 4 + 3] + s[i * 4 + 7] + 1) >> 1) >> 6;
-
- GST_WRITE_UINT16_LE (destY + i + 0, Y0);
- GST_WRITE_UINT16_LE (destY + i + 1, Y1);
- GST_WRITE_UINT16_LE (destU + (i >> 1), U);
- GST_WRITE_UINT16_LE (destV + (i >> 1), V);
- }
- if (i == width - 1) {
- Y0 = s[i * 4 + 1] >> 6;
- U = s[i * 4 + 2] >> 6;
- V = s[i * 4 + 3] >> 6;
-
- GST_WRITE_UINT16_LE (destY + i, Y0);
- GST_WRITE_UINT16_LE (destU + (i >> 1), U);
- GST_WRITE_UINT16_LE (destV + (i >> 1), V);
- }
-}
-
-#define PACK_I420_10BE GST_VIDEO_FORMAT_AYUV64, unpack_I420_10BE, pack_I420_10BE
-static void
-unpack_I420_10BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- gpointer dest, const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], gint x, gint y, gint width)
-{
- int i;
- guint16 *srcY = GET_Y_LINE (y);
- guint16 *srcU = GET_U_LINE (y >> 1);
- guint16 *srcV = GET_V_LINE (y >> 1);
- guint16 *d = dest;
-
- for (i = 0; i < width; i++) {
- d[i * 4 + 0] = 0xffff;
- d[i * 4 + 1] = GST_READ_UINT16_BE (srcY + i) << 6;
- d[i * 4 + 2] = GST_READ_UINT16_BE (srcU + (i >> 1)) << 6;
- d[i * 4 + 3] = GST_READ_UINT16_BE (srcV + (i >> 1)) << 6;
- }
-}
-
-static void
-pack_I420_10BE (const GstVideoFormatInfo * info, GstVideoPackFlags flags,
- const gpointer src, gint sstride, gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES], GstVideoChromaSite chroma_site,
- gint y, gint width)
-{
- int i;
- guint16 *destY = GET_Y_LINE (y);
- guint16 *destU = GET_U_LINE (y >> 1);
- guint16 *destV = GET_V_LINE (y >> 1);
- guint16 Y0, Y1, U, V;
- const guint16 *s = src;
-
- for (i = 0; i < width - 1; i += 2) {
- Y0 = s[i * 4 + 1] >> 6;
- Y1 = s[i * 4 + 5] >> 6;
- U = ((s[i * 4 + 2] + s[i * 4 + 6] + 1) >> 1) >> 6;
- V = ((s[i * 4 + 3] + s[i * 4 + 7] + 1) >> 1) >> 6;
-
- GST_WRITE_UINT16_BE (destY + i + 0, Y0);
- GST_WRITE_UINT16_BE (destY + i + 1, Y1);
- GST_WRITE_UINT16_BE (destU + (i >> 1), U);
- GST_WRITE_UINT16_BE (destV + (i >> 1), V);
- }
- if (i == width - 1) {
- Y0 = s[i * 4 + 1] >> 6;
- U = s[i * 4 + 2] >> 6;
- V = s[i * 4 + 3] >> 6;
-
- GST_WRITE_UINT16_BE (destY + i, Y0);
- GST_WRITE_UINT16_BE (destU + (i >> 1), U);
- GST_WRITE_UINT16_BE (destV + (i >> 1), V);
- }
-}
-
-typedef struct
-{
- guint32 fourcc;
- GstVideoFormatInfo info;
-} VideoFormat;
-
-/* depths: bits, n_components, shift, depth */
-#define DPTH0 0, 0, { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
-#define DPTH8 8, 1, { 0, 0, 0, 0 }, { 8, 0, 0, 0 }
-#define DPTH8_32 8, 2, { 0, 0, 0, 0 }, { 8, 32, 0, 0 }
-#define DPTH888 8, 3, { 0, 0, 0, 0 }, { 8, 8, 8, 0 }
-#define DPTH8888 8, 4, { 0, 0, 0, 0 }, { 8, 8, 8, 8 }
-#define DPTH10_10_10 10, 3, { 0, 0, 0, 0 }, { 10, 10, 10, 0 }
-#define DPTH16 16, 1, { 0, 0, 0, 0 }, { 16, 0, 0, 0 }
-#define DPTH16_16_16 16, 3, { 0, 0, 0, 0 }, { 16, 16, 16, 0 }
-#define DPTH16_16_16_16 16, 4, { 0, 0, 0, 0 }, { 16, 16, 16, 16 }
-#define DPTH555 16, 3, { 10, 5, 0, 0 }, { 5, 5, 5, 0 }
-#define DPTH565 16, 3, { 11, 5, 0, 0 }, { 5, 6, 5, 0 }
-
-/* pixel strides */
-#define PSTR0 { 0, 0, 0, 0 }
-#define PSTR1 { 1, 0, 0, 0 }
-#define PSTR14 { 1, 4, 0, 0 }
-#define PSTR111 { 1, 1, 1, 0 }
-#define PSTR1111 { 1, 1, 1, 1 }
-#define PSTR122 { 1, 2, 2, 0 }
-#define PSTR2 { 2, 0, 0, 0 }
-#define PSTR222 { 2, 2, 2, 0 }
-#define PSTR244 { 2, 4, 4, 0 }
-#define PSTR444 { 4, 4, 4, 0 }
-#define PSTR4444 { 4, 4, 4, 4 }
-#define PSTR333 { 3, 3, 3, 0 }
-#define PSTR488 { 4, 8, 8, 0 }
-#define PSTR8888 { 8, 8, 8, 8 }
-
-/* planes */
-#define PLANE_NA 0, { 0, 0, 0, 0 }
-#define PLANE0 1, { 0, 0, 0, 0 }
-#define PLANE01 2, { 0, 1, 0, 0 }
-#define PLANE011 2, { 0, 1, 1, 0 }
-#define PLANE012 3, { 0, 1, 2, 0 }
-#define PLANE0123 4, { 0, 1, 2, 3 }
-#define PLANE021 3, { 0, 2, 1, 0 }
-
-/* offsets */
-#define OFFS0 { 0, 0, 0, 0 }
-#define OFFS013 { 0, 1, 3, 0 }
-#define OFFS102 { 1, 0, 2, 0 }
-#define OFFS1230 { 1, 2, 3, 0 }
-#define OFFS012 { 0, 1, 2, 0 }
-#define OFFS210 { 2, 1, 0, 0 }
-#define OFFS123 { 1, 2, 3, 0 }
-#define OFFS321 { 3, 2, 1, 0 }
-#define OFFS0123 { 0, 1, 2, 3 }
-#define OFFS2103 { 2, 1, 0, 3 }
-#define OFFS3210 { 3, 2, 1, 0 }
-#define OFFS031 { 0, 3, 1, 0 }
-#define OFFS204 { 2, 0, 4, 0 }
-#define OFFS001 { 0, 0, 1, 0 }
-#define OFFS010 { 0, 1, 0, 0 }
-#define OFFS104 { 1, 0, 4, 0 }
-#define OFFS2460 { 2, 4, 6, 0 }
-
-/* subsampling */
-#define SUB410 { 0, 2, 2, 0 }, { 0, 2, 2, 0 }
-#define SUB411 { 0, 2, 2, 0 }, { 0, 0, 0, 0 }
-#define SUB420 { 0, 1, 1, 0 }, { 0, 1, 1, 0 }
-#define SUB422 { 0, 1, 1, 0 }, { 0, 0, 0, 0 }
-#define SUB4 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
-#define SUB44 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
-#define SUB444 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
-#define SUB4444 { 0, 0, 0, 0 }, { 0, 0, 0, 0 }
-#define SUB4204 { 0, 1, 1, 0 }, { 0, 1, 1, 0 }
-
-#define MAKE_YUV_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack ) \
- { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_YUV_LE_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack ) \
- { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV | GST_VIDEO_FORMAT_FLAG_LE, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_YUVA_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack) \
- { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV | GST_VIDEO_FORMAT_FLAG_ALPHA, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_YUV_C_FORMAT(name, desc, fourcc, depth, pstride, plane, offs, sub, pack) \
- { fourcc, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_YUV | GST_VIDEO_FORMAT_FLAG_COMPLEX, depth, pstride, plane, offs, sub, pack } }
-
-#define MAKE_RGB_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
- { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_RGB_LE_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
- { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB | GST_VIDEO_FORMAT_FLAG_LE, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_RGBA_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
- { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB | GST_VIDEO_FORMAT_FLAG_ALPHA, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_RGBAP_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
- { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_RGB | GST_VIDEO_FORMAT_FLAG_ALPHA | GST_VIDEO_FORMAT_FLAG_PALETTE, depth, pstride, plane, offs, sub, pack } }
-
-#define MAKE_GRAY_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
- { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_GRAY, depth, pstride, plane, offs, sub, pack } }
-#define MAKE_GRAY_LE_FORMAT(name, desc, depth, pstride, plane, offs, sub, pack) \
- { 0x00000000, {GST_VIDEO_FORMAT_ ##name, G_STRINGIFY(name), desc, GST_VIDEO_FORMAT_FLAG_GRAY | GST_VIDEO_FORMAT_FLAG_LE, depth, pstride, plane, offs, sub, pack } }
-
-static VideoFormat formats[] = {
- {0x00000000, {GST_VIDEO_FORMAT_UNKNOWN, "UNKNOWN", "unknown video", 0, DPTH0,
- PSTR0, PLANE_NA,
- OFFS0}},
- {0x00000000, {GST_VIDEO_FORMAT_ENCODED, "ENCODED", "encoded video",
- GST_VIDEO_FORMAT_FLAG_COMPLEX, DPTH0, PSTR0, PLANE_NA, OFFS0}},
-
- MAKE_YUV_FORMAT (I420, "raw video", GST_MAKE_FOURCC ('I', '4', '2', '0'),
- DPTH888, PSTR111,
- PLANE012, OFFS0, SUB420, PACK_420),
- MAKE_YUV_FORMAT (YV12, "raw video", GST_MAKE_FOURCC ('Y', 'V', '1', '2'),
- DPTH888, PSTR111,
- PLANE021, OFFS0, SUB420, PACK_420),
- MAKE_YUV_FORMAT (YUY2, "raw video", GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'),
- DPTH888, PSTR244,
- PLANE0, OFFS013, SUB422, PACK_YUY2),
- MAKE_YUV_FORMAT (UYVY, "raw video", GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'),
- DPTH888, PSTR244,
- PLANE0, OFFS102, SUB422, PACK_UYVY),
- MAKE_YUVA_FORMAT (AYUV, "raw video", GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'),
- DPTH8888,
- PSTR4444, PLANE0, OFFS1230, SUB4444, PACK_AYUV),
- MAKE_RGB_FORMAT (RGBx, "raw video", DPTH888, PSTR444, PLANE0, OFFS012,
- SUB444, PACK_RGBA),
- MAKE_RGB_FORMAT (BGRx, "raw video", DPTH888, PSTR444, PLANE0, OFFS210,
- SUB444, PACK_BGRA),
- MAKE_RGB_FORMAT (xRGB, "raw video", DPTH888, PSTR444, PLANE0, OFFS123,
- SUB444, PACK_ARGB),
- MAKE_RGB_FORMAT (xBGR, "raw video", DPTH888, PSTR444, PLANE0, OFFS321,
- SUB444, PACK_ABGR),
- MAKE_RGBA_FORMAT (RGBA, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS0123,
- SUB4444, PACK_RGBA),
- MAKE_RGBA_FORMAT (BGRA, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS2103,
- SUB4444, PACK_BGRA),
- MAKE_RGBA_FORMAT (ARGB, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS1230,
- SUB4444, PACK_ARGB),
- MAKE_RGBA_FORMAT (ABGR, "raw video", DPTH8888, PSTR4444, PLANE0, OFFS3210,
- SUB4444, PACK_ABGR),
- MAKE_RGB_FORMAT (RGB, "raw video", DPTH888, PSTR333, PLANE0, OFFS012, SUB444,
- PACK_RGB),
- MAKE_RGB_FORMAT (BGR, "raw video", DPTH888, PSTR333, PLANE0, OFFS210, SUB444,
- PACK_BGR),
-
- MAKE_YUV_FORMAT (Y41B, "raw video", GST_MAKE_FOURCC ('Y', '4', '1', 'B'),
- DPTH888, PSTR111,
- PLANE012, OFFS0, SUB411, PACK_Y41B),
- MAKE_YUV_FORMAT (Y42B, "raw video", GST_MAKE_FOURCC ('Y', '4', '2', 'B'),
- DPTH888, PSTR111,
- PLANE012, OFFS0, SUB422, PACK_Y42B),
- MAKE_YUV_FORMAT (YVYU, "raw video", GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'),
- DPTH888, PSTR244,
- PLANE0, OFFS031, SUB422, PACK_YVYU),
- MAKE_YUV_FORMAT (Y444, "raw video", GST_MAKE_FOURCC ('Y', '4', '4', '4'),
- DPTH888, PSTR111,
- PLANE012, OFFS0, SUB444, PACK_Y444),
- MAKE_YUV_C_FORMAT (v210, "raw video", GST_MAKE_FOURCC ('v', '2', '1', '0'),
- DPTH10_10_10,
- PSTR0, PLANE0, OFFS0, SUB422, PACK_v210),
- MAKE_YUV_FORMAT (v216, "raw video", GST_MAKE_FOURCC ('v', '2', '1', '6'),
- DPTH16_16_16,
- PSTR488, PLANE0, OFFS204, SUB422, PACK_v216),
- MAKE_YUV_FORMAT (NV12, "raw video", GST_MAKE_FOURCC ('N', 'V', '1', '2'),
- DPTH888, PSTR122,
- PLANE011, OFFS001, SUB420, PACK_NV12),
- MAKE_YUV_FORMAT (NV21, "raw video", GST_MAKE_FOURCC ('N', 'V', '2', '1'),
- DPTH888, PSTR122,
- PLANE011, OFFS010, SUB420, PACK_NV21),
-
- MAKE_GRAY_FORMAT (GRAY8, "raw video", DPTH8, PSTR1, PLANE0, OFFS0, SUB4,
- PACK_GRAY8),
- MAKE_GRAY_FORMAT (GRAY16_BE, "raw video", DPTH16, PSTR2, PLANE0, OFFS0, SUB4,
- PACK_GRAY16_BE),
- MAKE_GRAY_LE_FORMAT (GRAY16_LE, "raw video", DPTH16, PSTR2, PLANE0, OFFS0,
- SUB4, PACK_GRAY16_LE),
-
- MAKE_YUV_FORMAT (v308, "raw video", GST_MAKE_FOURCC ('v', '3', '0', '8'),
- DPTH888, PSTR333,
- PLANE0, OFFS012, SUB444, PACK_v308),
-
-#if G_BYTE_ORDER == G_LITTLE_ENDIAN
- MAKE_RGB_LE_FORMAT (RGB16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0,
- SUB444, PACK_RGB16),
- MAKE_RGB_LE_FORMAT (BGR16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0,
- SUB444, PACK_BGR16),
- MAKE_RGB_LE_FORMAT (RGB15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0,
- SUB444, PACK_RGB15),
- MAKE_RGB_LE_FORMAT (BGR15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0,
- SUB444, PACK_BGR15),
-#else
- MAKE_RGB_FORMAT (RGB16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0, SUB444,
- PACK_RGB16),
- MAKE_RGB_FORMAT (BGR16, "raw video", DPTH565, PSTR222, PLANE0, OFFS0, SUB444,
- PACK_BGR16),
- MAKE_RGB_FORMAT (RGB15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0, SUB444,
- PACK_RGB15),
- MAKE_RGB_FORMAT (BGR15, "raw video", DPTH555, PSTR222, PLANE0, OFFS0, SUB444,
- PACK_BGR15),
-#endif
-
- MAKE_YUV_C_FORMAT (UYVP, "raw video", GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'),
- DPTH10_10_10,
- PSTR0, PLANE0, OFFS0, SUB422, PACK_UYVP),
- MAKE_YUVA_FORMAT (A420, "raw video", GST_MAKE_FOURCC ('A', '4', '2', '0'),
- DPTH8888,
- PSTR1111, PLANE0123, OFFS0, SUB4204, PACK_A420),
- MAKE_RGBAP_FORMAT (RGB8P, "raw video", DPTH8_32, PSTR14, PLANE01,
- OFFS0, SUB44, PACK_RGB8P),
- MAKE_YUV_FORMAT (YUV9, "raw video", GST_MAKE_FOURCC ('Y', 'U', 'V', '9'),
- DPTH888, PSTR111,
- PLANE012, OFFS0, SUB410, PACK_410),
- MAKE_YUV_FORMAT (YVU9, "raw video", GST_MAKE_FOURCC ('Y', 'V', 'U', '9'),
- DPTH888, PSTR111,
- PLANE021, OFFS0, SUB410, PACK_410),
- MAKE_YUV_FORMAT (IYU1, "raw video", GST_MAKE_FOURCC ('I', 'Y', 'U', '1'),
- DPTH888, PSTR0,
- PLANE0, OFFS104, SUB411, PACK_IYU1),
- MAKE_RGBA_FORMAT (ARGB64, "raw video", DPTH16_16_16_16, PSTR8888, PLANE0,
- OFFS2460,
- SUB444, PACK_ARGB64),
- MAKE_YUVA_FORMAT (AYUV64, "raw video", 0x00000000, DPTH16_16_16_16, PSTR8888,
- PLANE0,
- OFFS2460, SUB444, PACK_AYUV64),
- MAKE_YUV_FORMAT (r210, "raw video", GST_MAKE_FOURCC ('r', '2', '1', '0'),
- DPTH10_10_10,
- PSTR444, PLANE0, OFFS0, SUB444, PACK_r210),
- MAKE_YUV_FORMAT (I420_10BE, "raw video", 0x00000000, DPTH10_10_10,
- PSTR222, PLANE012, OFFS0, SUB420, PACK_I420_10BE),
- MAKE_YUV_LE_FORMAT (I420_10LE, "raw video", 0x00000000, DPTH10_10_10,
- PSTR222, PLANE012, OFFS0, SUB420, PACK_I420_10LE),
-};
-
-/**
- * gst_video_calculate_display_ratio:
- * @dar_n: Numerator of the calculated display_ratio
- * @dar_d: Denominator of the calculated display_ratio
- * @video_width: Width of the video frame in pixels
- * @video_height: Height of the video frame in pixels
- * @video_par_n: Numerator of the pixel aspect ratio of the input video.
- * @video_par_d: Denominator of the pixel aspect ratio of the input video.
- * @display_par_n: Numerator of the pixel aspect ratio of the display device
- * @display_par_d: Denominator of the pixel aspect ratio of the display device
- *
- * Given the Pixel Aspect Ratio and size of an input video frame, and the
- * pixel aspect ratio of the intended display device, calculates the actual
- * display ratio the video will be rendered with.
- *
- * Returns: A boolean indicating success and a calculated Display Ratio in the
- * dar_n and dar_d parameters.
- * The return value is FALSE in the case of integer overflow or other error.
- *
- * Since: 0.10.7
- */
-gboolean
-gst_video_calculate_display_ratio (guint * dar_n, guint * dar_d,
- guint video_width, guint video_height,
- guint video_par_n, guint video_par_d,
- guint display_par_n, guint display_par_d)
-{
- gint num, den;
- gint tmp_n, tmp_d;
-
- g_return_val_if_fail (dar_n != NULL, FALSE);
- g_return_val_if_fail (dar_d != NULL, FALSE);
-
- /* Calculate (video_width * video_par_n * display_par_d) /
- * (video_height * video_par_d * display_par_n) */
- if (!gst_util_fraction_multiply (video_width, video_height, video_par_n,
- video_par_d, &tmp_n, &tmp_d))
- goto error_overflow;
-
- if (!gst_util_fraction_multiply (tmp_n, tmp_d, display_par_d, display_par_n,
- &num, &den))
- goto error_overflow;
-
- g_return_val_if_fail (num > 0, FALSE);
- g_return_val_if_fail (den > 0, FALSE);
-
- *dar_n = num;
- *dar_d = den;
-
- return TRUE;
-
- /* ERRORS */
-error_overflow:
- {
- GST_WARNING ("overflow in multiply");
- return FALSE;
- }
-}
-
-static GstVideoFormat
-gst_video_format_from_rgb32_masks (int red_mask, int green_mask, int blue_mask)
-{
- if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
- blue_mask == 0x0000ff00) {
- return GST_VIDEO_FORMAT_RGBx;
- }
- if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
- blue_mask == 0xff000000) {
- return GST_VIDEO_FORMAT_BGRx;
- }
- if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
- blue_mask == 0x000000ff) {
- return GST_VIDEO_FORMAT_xRGB;
- }
- if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
- blue_mask == 0x00ff0000) {
- return GST_VIDEO_FORMAT_xBGR;
- }
-
- return GST_VIDEO_FORMAT_UNKNOWN;
-}
-
-static GstVideoFormat
-gst_video_format_from_rgba32_masks (int red_mask, int green_mask,
- int blue_mask, int alpha_mask)
-{
- if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
- blue_mask == 0x0000ff00 && alpha_mask == 0x000000ff) {
- return GST_VIDEO_FORMAT_RGBA;
- }
- if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
- blue_mask == 0xff000000 && alpha_mask == 0x000000ff) {
- return GST_VIDEO_FORMAT_BGRA;
- }
- if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
- blue_mask == 0x000000ff && alpha_mask == 0xff000000) {
- return GST_VIDEO_FORMAT_ARGB;
- }
- if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
- blue_mask == 0x00ff0000 && alpha_mask == 0xff000000) {
- return GST_VIDEO_FORMAT_ABGR;
- }
- return GST_VIDEO_FORMAT_UNKNOWN;
-}
-
-static GstVideoFormat
-gst_video_format_from_rgb24_masks (int red_mask, int green_mask, int blue_mask)
-{
- if (red_mask == 0xff0000 && green_mask == 0x00ff00 && blue_mask == 0x0000ff) {
- return GST_VIDEO_FORMAT_RGB;
- }
- if (red_mask == 0x0000ff && green_mask == 0x00ff00 && blue_mask == 0xff0000) {
- return GST_VIDEO_FORMAT_BGR;
- }
-
- return GST_VIDEO_FORMAT_UNKNOWN;
-}
-
-#define GST_VIDEO_COMP1_MASK_16_INT 0xf800
-#define GST_VIDEO_COMP2_MASK_16_INT 0x07e0
-#define GST_VIDEO_COMP3_MASK_16_INT 0x001f
-
-#define GST_VIDEO_COMP1_MASK_15_INT 0x7c00
-#define GST_VIDEO_COMP2_MASK_15_INT 0x03e0
-#define GST_VIDEO_COMP3_MASK_15_INT 0x001f
-
-static GstVideoFormat
-gst_video_format_from_rgb16_masks (int red_mask, int green_mask, int blue_mask)
-{
- if (red_mask == GST_VIDEO_COMP1_MASK_16_INT
- && green_mask == GST_VIDEO_COMP2_MASK_16_INT
- && blue_mask == GST_VIDEO_COMP3_MASK_16_INT) {
- return GST_VIDEO_FORMAT_RGB16;
- }
- if (red_mask == GST_VIDEO_COMP3_MASK_16_INT
- && green_mask == GST_VIDEO_COMP2_MASK_16_INT
- && blue_mask == GST_VIDEO_COMP1_MASK_16_INT) {
- return GST_VIDEO_FORMAT_BGR16;
- }
- if (red_mask == GST_VIDEO_COMP1_MASK_15_INT
- && green_mask == GST_VIDEO_COMP2_MASK_15_INT
- && blue_mask == GST_VIDEO_COMP3_MASK_15_INT) {
- return GST_VIDEO_FORMAT_RGB15;
- }
- if (red_mask == GST_VIDEO_COMP3_MASK_15_INT
- && green_mask == GST_VIDEO_COMP2_MASK_15_INT
- && blue_mask == GST_VIDEO_COMP1_MASK_15_INT) {
- return GST_VIDEO_FORMAT_BGR15;
- }
- return GST_VIDEO_FORMAT_UNKNOWN;
-}
-
-/**
- * gst_video_format_from_masks:
- * @depth: the amount of bits used for a pixel
- * @bpp: the amount of bits used to store a pixel. This value is bigger than
- * @depth
- * @endianness: the endianness of the masks
- * @red_mask: the red mask
- * @green_mask: the green mask
- * @blue_mask: the blue mask
- * @alpha_mask: the optional alpha mask
- *
- * Find the #GstVideoFormat for the given parameters.
- *
- * Returns: a #GstVideoFormat or GST_VIDEO_FORMAT_UNKNOWN when the parameters to
- * not specify a known format.
- */
-GstVideoFormat
-gst_video_format_from_masks (gint depth, gint bpp, gint endianness,
- gint red_mask, gint green_mask, gint blue_mask, gint alpha_mask)
-{
- GstVideoFormat format;
-
- /* our caps system handles 24/32bpp RGB as big-endian. */
- if ((bpp == 24 || bpp == 32) && endianness == G_LITTLE_ENDIAN) {
- red_mask = GUINT32_TO_BE (red_mask);
- green_mask = GUINT32_TO_BE (green_mask);
- blue_mask = GUINT32_TO_BE (blue_mask);
- endianness = G_BIG_ENDIAN;
- if (bpp == 24) {
- red_mask >>= 8;
- green_mask >>= 8;
- blue_mask >>= 8;
- }
- }
-
- if (depth == 30 && bpp == 32) {
- format = GST_VIDEO_FORMAT_r210;
- } else if (depth == 24 && bpp == 32) {
- format = gst_video_format_from_rgb32_masks (red_mask, green_mask,
- blue_mask);
- } else if (depth == 32 && bpp == 32 && alpha_mask) {
- format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
- blue_mask, alpha_mask);
- } else if (depth == 24 && bpp == 24) {
- format = gst_video_format_from_rgb24_masks (red_mask, green_mask,
- blue_mask);
- } else if ((depth == 15 || depth == 16) && bpp == 16 &&
- endianness == G_BYTE_ORDER) {
- format = gst_video_format_from_rgb16_masks (red_mask, green_mask,
- blue_mask);
- } else if (depth == 8 && bpp == 8) {
- format = GST_VIDEO_FORMAT_RGB8P;
- } else if (depth == 64 && bpp == 64) {
- format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
- blue_mask, alpha_mask);
- if (format == GST_VIDEO_FORMAT_ARGB) {
- format = GST_VIDEO_FORMAT_ARGB64;
- } else {
- format = GST_VIDEO_FORMAT_UNKNOWN;
- }
- } else {
- format = GST_VIDEO_FORMAT_UNKNOWN;
- }
- return format;
-}
-
-/**
- * gst_video_format_from_fourcc:
- * @fourcc: a FOURCC value representing raw YUV video
- *
- * Converts a FOURCC value into the corresponding #GstVideoFormat.
- * If the FOURCC cannot be represented by #GstVideoFormat,
- * #GST_VIDEO_FORMAT_UNKNOWN is returned.
- *
- * Since: 0.10.16
- *
- * Returns: the #GstVideoFormat describing the FOURCC value
- */
-GstVideoFormat
-gst_video_format_from_fourcc (guint32 fourcc)
-{
- switch (fourcc) {
- case GST_MAKE_FOURCC ('I', '4', '2', '0'):
- return GST_VIDEO_FORMAT_I420;
- case GST_MAKE_FOURCC ('Y', 'V', '1', '2'):
- return GST_VIDEO_FORMAT_YV12;
- case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'):
- return GST_VIDEO_FORMAT_YUY2;
- case GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'):
- return GST_VIDEO_FORMAT_YVYU;
- case GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'):
- return GST_VIDEO_FORMAT_UYVY;
- case GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'):
- return GST_VIDEO_FORMAT_AYUV;
- case GST_MAKE_FOURCC ('Y', '4', '1', 'B'):
- return GST_VIDEO_FORMAT_Y41B;
- case GST_MAKE_FOURCC ('Y', '4', '2', 'B'):
- return GST_VIDEO_FORMAT_Y42B;
- case GST_MAKE_FOURCC ('Y', '4', '4', '4'):
- return GST_VIDEO_FORMAT_Y444;
- case GST_MAKE_FOURCC ('v', '2', '1', '0'):
- return GST_VIDEO_FORMAT_v210;
- case GST_MAKE_FOURCC ('v', '2', '1', '6'):
- return GST_VIDEO_FORMAT_v216;
- case GST_MAKE_FOURCC ('N', 'V', '1', '2'):
- return GST_VIDEO_FORMAT_NV12;
- case GST_MAKE_FOURCC ('N', 'V', '2', '1'):
- return GST_VIDEO_FORMAT_NV21;
- case GST_MAKE_FOURCC ('v', '3', '0', '8'):
- return GST_VIDEO_FORMAT_v308;
- case GST_MAKE_FOURCC ('Y', '8', '0', '0'):
- case GST_MAKE_FOURCC ('Y', '8', ' ', ' '):
- case GST_MAKE_FOURCC ('G', 'R', 'E', 'Y'):
- return GST_VIDEO_FORMAT_GRAY8;
- case GST_MAKE_FOURCC ('Y', '1', '6', ' '):
- return GST_VIDEO_FORMAT_GRAY16_LE;
- case GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'):
- return GST_VIDEO_FORMAT_UYVP;
- case GST_MAKE_FOURCC ('A', '4', '2', '0'):
- return GST_VIDEO_FORMAT_A420;
- case GST_MAKE_FOURCC ('Y', 'U', 'V', '9'):
- return GST_VIDEO_FORMAT_YUV9;
- case GST_MAKE_FOURCC ('Y', 'V', 'U', '9'):
- return GST_VIDEO_FORMAT_YVU9;
- case GST_MAKE_FOURCC ('I', 'Y', 'U', '1'):
- return GST_VIDEO_FORMAT_IYU1;
- case GST_MAKE_FOURCC ('A', 'Y', '6', '4'):
- return GST_VIDEO_FORMAT_AYUV64;
- default:
- return GST_VIDEO_FORMAT_UNKNOWN;
- }
-}
-
-/**
- * gst_video_format_from_string:
- * @format: a format string
- *
- * Convert the @format string to its #GstVideoFormat.
- *
- * Returns: the #GstVideoFormat for @format or GST_VIDEO_FORMAT_UNKNOWN when the
- * string is not a known format.
- */
-GstVideoFormat
-gst_video_format_from_string (const gchar * format)
-{
- guint i;
-
- g_return_val_if_fail (format != NULL, GST_VIDEO_FORMAT_UNKNOWN);
-
- for (i = 0; i < G_N_ELEMENTS (formats); i++) {
- if (strcmp (GST_VIDEO_FORMAT_INFO_NAME (&formats[i].info), format) == 0)
- return GST_VIDEO_FORMAT_INFO_FORMAT (&formats[i].info);
- }
- return GST_VIDEO_FORMAT_UNKNOWN;
-}
-
-
-/**
- * gst_video_format_to_fourcc:
- * @format: a #GstVideoFormat video format
- *
- * Converts a #GstVideoFormat value into the corresponding FOURCC. Only
- * a few YUV formats have corresponding FOURCC values. If @format has
- * no corresponding FOURCC value, 0 is returned.
- *
- * Since: 0.10.16
- *
- * Returns: the FOURCC corresponding to @format
- */
-guint32
-gst_video_format_to_fourcc (GstVideoFormat format)
-{
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
-
- if (format >= G_N_ELEMENTS (formats))
- return 0;
-
- return formats[format].fourcc;
-}
-
-const gchar *
-gst_video_format_to_string (GstVideoFormat format)
-{
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
-
- if (format >= G_N_ELEMENTS (formats))
- return NULL;
-
- return GST_VIDEO_FORMAT_INFO_NAME (&formats[format].info);
-}
-
-/**
- * gst_video_format_get_info:
- * @format: a #GstVideoFormat
- *
- * Get the #GstVideoFormatInfo for @format
- *
- * Returns: The #GstVideoFormatInfo for @format.
- */
-const GstVideoFormatInfo *
-gst_video_format_get_info (GstVideoFormat format)
-{
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
- g_return_val_if_fail (format < G_N_ELEMENTS (formats), NULL);
-
- return &formats[format].info;
-}
-
-typedef struct
-{
- const gchar *name;
- GstVideoColorimetry color;
-} ColorimetryInfo;
-
-#define MAKE_COLORIMETRY(n,r,m,t,p) { GST_VIDEO_COLORIMETRY_ ##n, \
- { GST_VIDEO_COLOR_RANGE ##r, GST_VIDEO_COLOR_MATRIX_ ##m, \
- GST_VIDEO_TRANSFER_ ##t, GST_VIDEO_COLOR_PRIMARIES_ ##p } }
-
-#define GST_VIDEO_COLORIMETRY_NONAME NULL
-
-#define DEFAULT_YUV_SD 0
-#define DEFAULT_YUV_HD 1
-#define DEFAULT_RGB 3
-#define DEFAULT_GRAY 4
-#define DEFAULT_UNKNOWN 5
-
-static const ColorimetryInfo colorimetry[] = {
- MAKE_COLORIMETRY (BT601, _16_235, BT601, BT709, BT470M),
- MAKE_COLORIMETRY (BT709, _16_235, BT709, BT709, BT709),
- MAKE_COLORIMETRY (SMPTE240M, _16_235, SMPTE240M, SMPTE240M, SMPTE240M),
- MAKE_COLORIMETRY (NONAME, _0_255, RGB, UNKNOWN, UNKNOWN),
- MAKE_COLORIMETRY (NONAME, _0_255, BT601, UNKNOWN, UNKNOWN),
- MAKE_COLORIMETRY (NONAME, _UNKNOWN, UNKNOWN, UNKNOWN, UNKNOWN),
-};
-
-/**
- * gst_video_info_init:
- * @info: a #GstVideoInfo
- *
- * Initialize @info with default values.
- */
-void
-gst_video_info_init (GstVideoInfo * info)
-{
- g_return_if_fail (info != NULL);
-
- memset (info, 0, sizeof (GstVideoInfo));
-
- info->finfo = &formats[GST_VIDEO_FORMAT_UNKNOWN].info;
-
- info->views = 1;
- /* arrange for sensible defaults, e.g. if turned into caps */
- info->fps_n = 0;
- info->fps_d = 1;
- info->par_n = 1;
- info->par_d = 1;
-}
-
-/**
- * gst_video_info_set_format:
- * @info: a #GstVideoInfo
- * @format: the format
- * @width: a width
- * @height: a height
- *
- * Set the default info for a video frame of @format and @width and @height.
- */
-void
-gst_video_info_set_format (GstVideoInfo * info, GstVideoFormat format,
- guint width, guint height)
-{
- const GstVideoFormatInfo *finfo;
-
- g_return_if_fail (info != NULL);
- g_return_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN);
-
- finfo = &formats[format].info;
-
- info->flags = 0;
- info->finfo = finfo;
- info->width = width;
- info->height = height;
-
- if (GST_VIDEO_FORMAT_INFO_IS_YUV (finfo)) {
- if (height > 576)
- info->colorimetry = colorimetry[DEFAULT_YUV_HD].color;
- else
- info->colorimetry = colorimetry[DEFAULT_YUV_SD].color;
- } else if (GST_VIDEO_FORMAT_INFO_IS_GRAY (finfo)) {
- info->colorimetry = colorimetry[DEFAULT_GRAY].color;
- } else if (GST_VIDEO_FORMAT_INFO_IS_RGB (finfo)) {
- info->colorimetry = colorimetry[DEFAULT_RGB].color;
- } else {
- info->colorimetry = colorimetry[DEFAULT_UNKNOWN].color;
- }
-
- fill_planes (info);
-}
-
-static const gchar *interlace_mode[] = {
- "progressive",
- "interleaved",
- "mixed",
- "fields"
-};
-
-static const gchar *
-gst_interlace_mode_to_string (GstVideoInterlaceMode mode)
-{
- if (((guint) mode) >= G_N_ELEMENTS (interlace_mode))
- return NULL;
-
- return interlace_mode[mode];
-}
-
-static GstVideoInterlaceMode
-gst_interlace_mode_from_string (const gchar * mode)
-{
- gint i;
- for (i = 0; i < G_N_ELEMENTS (interlace_mode); i++) {
- if (g_str_equal (interlace_mode[i], mode))
- return i;
- }
- return GST_VIDEO_INTERLACE_MODE_PROGRESSIVE;
-}
-
-typedef struct
-{
- const gchar *name;
- GstVideoChromaSite site;
-} ChromaSiteInfo;
-
-static const ChromaSiteInfo chromasite[] = {
- {"jpeg", GST_VIDEO_CHROMA_SITE_JPEG},
- {"mpeg2", GST_VIDEO_CHROMA_SITE_MPEG2},
- {"dv", GST_VIDEO_CHROMA_SITE_DV}
-};
-
-static GstVideoChromaSite
-gst_video_chroma_from_string (const gchar * s)
-{
- gint i;
- for (i = 0; i < G_N_ELEMENTS (chromasite); i++) {
- if (g_str_equal (chromasite[i].name, s))
- return chromasite[i].site;
- }
- return GST_VIDEO_CHROMA_SITE_UNKNOWN;
-}
-
-static const gchar *
-gst_video_chroma_to_string (GstVideoChromaSite site)
-{
- gint i;
- for (i = 0; i < G_N_ELEMENTS (chromasite); i++) {
- if (chromasite[i].site == site)
- return chromasite[i].name;
- }
- return NULL;
-}
-
-static const ColorimetryInfo *
-gst_video_get_colorimetry (const gchar * s)
-{
- gint i;
-
- for (i = 0; colorimetry[i].name; i++) {
- if (g_str_equal (colorimetry[i].name, s))
- return &colorimetry[i];
- }
- return NULL;
-}
-
-#define IS_EQUAL(ci,i) (((ci)->color.range == (i)->range) && \
- ((ci)->color.matrix == (i)->matrix) && \
- ((ci)->color.transfer == (i)->transfer) && \
- ((ci)->color.primaries == (i)->primaries))
-
-#define IS_UNKNOWN(ci) (IS_EQUAL (&colorimetry[DEFAULT_UNKNOWN], ci))
-
-/**
- * gst_video_colorimetry_from_string
- * @cinfo: a #GstVideoColorimetry
- * @color: a colorimetry string
- *
- * Parse the colorimetry string and update @cinfo with the parsed
- * values.
- *
- * Returns: #TRUE if @color points to valid colorimetry info.
- */
-gboolean
-gst_video_colorimetry_from_string (GstVideoColorimetry * cinfo,
- const gchar * color)
-{
- const ColorimetryInfo *ci;
-
- if ((ci = gst_video_get_colorimetry (color))) {
- *cinfo = ci->color;
- } else {
- gint r, m, t, p;
-
- if (sscanf (color, "%d:%d:%d:%d", &r, &m, &t, &p) == 4) {
- cinfo->range = r;
- cinfo->matrix = m;
- cinfo->transfer = t;
- cinfo->primaries = p;
- }
- }
- return TRUE;
-}
-
-static void
-gst_video_caps_set_colorimetry (GstCaps * caps, GstVideoColorimetry * cinfo)
-{
- gint i;
-
- for (i = 0; colorimetry[i].name; i++) {
- if (IS_EQUAL (&colorimetry[i], cinfo)) {
- gst_caps_set_simple (caps, "colorimetry", G_TYPE_STRING,
- colorimetry[i].name, NULL);
- return;
- }
- }
- if (!IS_UNKNOWN (cinfo)) {
- gchar *str;
-
- str =
- g_strdup_printf ("%d:%d:%d:%d", cinfo->range, cinfo->matrix,
- cinfo->transfer, cinfo->primaries);
- gst_caps_set_simple (caps, "colorimetry", G_TYPE_STRING, str, NULL);
- g_free (str);
- }
-}
-
-/**
- * gst_video_colorimetry_matches:
- * @info: a #GstVideoInfo
- * @color: a colorimetry string
- *
- * Check if the colorimetry information in @info matches that of the
- * string @color.
- *
- * Returns: #TRUE if @color conveys the same colorimetry info as the color
- * information in @info.
- */
-gboolean
-gst_video_colorimetry_matches (GstVideoColorimetry * cinfo, const gchar * color)
-{
- const ColorimetryInfo *ci;
-
- if ((ci = gst_video_get_colorimetry (color)))
- return IS_EQUAL (ci, cinfo);
-
- return FALSE;
-}
-
-/**
- * gst_video_info_from_caps:
- * @info: a #GstVideoInfo
- * @caps: a #GstCaps
- *
- * Parse @caps and update @info.
- *
- * Returns: TRUE if @caps could be parsed
- */
-gboolean
-gst_video_info_from_caps (GstVideoInfo * info, const GstCaps * caps)
-{
- GstStructure *structure;
- const gchar *s;
- GstVideoFormat format = GST_VIDEO_FORMAT_UNKNOWN;
- gint width = 0, height = 0, views;
- gint fps_n, fps_d;
- gint par_n, par_d;
-
- g_return_val_if_fail (info != NULL, FALSE);
- g_return_val_if_fail (caps != NULL, FALSE);
- g_return_val_if_fail (gst_caps_is_fixed (caps), FALSE);
-
- GST_DEBUG ("parsing caps %" GST_PTR_FORMAT, caps);
-
- structure = gst_caps_get_structure (caps, 0);
-
- if (gst_structure_has_name (structure, "video/x-raw")) {
- if (!(s = gst_structure_get_string (structure, "format")))
- goto no_format;
-
- format = gst_video_format_from_string (s);
- if (format == GST_VIDEO_FORMAT_UNKNOWN)
- goto unknown_format;
-
- } else if (g_str_has_prefix (gst_structure_get_name (structure), "video/") ||
- g_str_has_prefix (gst_structure_get_name (structure), "image/")) {
- format = GST_VIDEO_FORMAT_ENCODED;
- } else {
- goto wrong_name;
- }
-
- /* width and height are mandatory, except for non-raw-formats */
- if (!gst_structure_get_int (structure, "width", &width) &&
- format != GST_VIDEO_FORMAT_ENCODED)
- goto no_width;
- if (!gst_structure_get_int (structure, "height", &height) &&
- format != GST_VIDEO_FORMAT_ENCODED)
- goto no_height;
-
- gst_video_info_set_format (info, format, width, height);
-
- if (gst_structure_get_fraction (structure, "framerate", &fps_n, &fps_d)) {
- if (fps_n == 0) {
- /* variable framerate */
- info->flags |= GST_VIDEO_FLAG_VARIABLE_FPS;
- /* see if we have a max-framerate */
- gst_structure_get_fraction (structure, "max-framerate", &fps_n, &fps_d);
- }
- info->fps_n = fps_n;
- info->fps_d = fps_d;
- } else {
- /* unspecified is variable framerate */
- info->fps_n = 0;
- info->fps_d = 1;
- }
-
- if ((s = gst_structure_get_string (structure, "interlace-mode")))
- info->interlace_mode = gst_interlace_mode_from_string (s);
- else
- info->interlace_mode = GST_VIDEO_INTERLACE_MODE_PROGRESSIVE;
-
- if (gst_structure_get_int (structure, "views", &views))
- info->views = views;
- else
- info->views = 1;
-
- if ((s = gst_structure_get_string (structure, "chroma-site")))
- info->chroma_site = gst_video_chroma_from_string (s);
- else
- info->chroma_site = GST_VIDEO_CHROMA_SITE_UNKNOWN;
-
- if ((s = gst_structure_get_string (structure, "colorimetry")))
- gst_video_colorimetry_from_string (&info->colorimetry, s);
-
- if (gst_structure_get_fraction (structure, "pixel-aspect-ratio",
- &par_n, &par_d)) {
- info->par_n = par_n;
- info->par_d = par_d;
- } else {
- info->par_n = 1;
- info->par_d = 1;
- }
- return TRUE;
-
- /* ERROR */
-wrong_name:
- {
- GST_ERROR ("wrong name '%s', expected video/ or image/",
- gst_structure_get_name (structure));
- return FALSE;
- }
-no_format:
- {
- GST_ERROR ("no format given");
- return FALSE;
- }
-unknown_format:
- {
- GST_ERROR ("unknown format '%s' given", s);
- return FALSE;
- }
-no_width:
- {
- GST_ERROR ("no width property given");
- return FALSE;
- }
-no_height:
- {
- GST_ERROR ("no height property given");
- return FALSE;
- }
-}
-
-/**
- * gst_video_info_is_equal:
- * @info: a #GstVideoInfo
- * @other: a #GstVideoInfo
- *
- * Compares two #GstVideoInfo and returns whether they are equal or not
- *
- * Returns: %TRUE if @info and @other are equal, else %FALSE.
- */
-gboolean
-gst_video_info_is_equal (const GstVideoInfo * info, const GstVideoInfo * other)
-{
- if (GST_VIDEO_INFO_FORMAT (info) != GST_VIDEO_INFO_FORMAT (other))
- return FALSE;
- if (GST_VIDEO_INFO_INTERLACE_MODE (info) !=
- GST_VIDEO_INFO_INTERLACE_MODE (other))
- return FALSE;
- if (GST_VIDEO_INFO_FLAGS (info) != GST_VIDEO_INFO_FLAGS (other))
- return FALSE;
- if (GST_VIDEO_INFO_WIDTH (info) != GST_VIDEO_INFO_WIDTH (other))
- return FALSE;
- if (GST_VIDEO_INFO_HEIGHT (info) != GST_VIDEO_INFO_HEIGHT (other))
- return FALSE;
- if (GST_VIDEO_INFO_SIZE (info) != GST_VIDEO_INFO_SIZE (other))
- return FALSE;
- if (GST_VIDEO_INFO_PAR_N (info) != GST_VIDEO_INFO_PAR_N (other))
- return FALSE;
- if (GST_VIDEO_INFO_PAR_D (info) != GST_VIDEO_INFO_PAR_D (other))
- return FALSE;
- if (GST_VIDEO_INFO_FPS_N (info) != GST_VIDEO_INFO_FPS_N (other))
- return FALSE;
- if (GST_VIDEO_INFO_FPS_D (info) != GST_VIDEO_INFO_FPS_D (other))
- return FALSE;
- return TRUE;
-}
-
-/**
- * gst_video_info_to_caps:
- * @info: a #GstVideoInfo
- *
- * Convert the values of @info into a #GstCaps.
- *
- * Returns: a new #GstCaps containing the info of @info.
- */
-GstCaps *
-gst_video_info_to_caps (GstVideoInfo * info)
-{
- GstCaps *caps;
- const gchar *format;
-
- g_return_val_if_fail (info != NULL, NULL);
- g_return_val_if_fail (info->finfo != NULL, NULL);
- g_return_val_if_fail (info->finfo->format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
-
- format = gst_video_format_to_string (info->finfo->format);
- g_return_val_if_fail (format != NULL, NULL);
-
- caps = gst_caps_new_simple ("video/x-raw",
- "format", G_TYPE_STRING, format,
- "width", G_TYPE_INT, info->width,
- "height", G_TYPE_INT, info->height,
- "pixel-aspect-ratio", GST_TYPE_FRACTION, info->par_n, info->par_d, NULL);
-
- gst_caps_set_simple (caps, "interlace-mode", G_TYPE_STRING,
- gst_interlace_mode_to_string (info->interlace_mode), NULL);
-
- if (info->chroma_site != GST_VIDEO_CHROMA_SITE_UNKNOWN)
- gst_caps_set_simple (caps, "chroma-site", G_TYPE_STRING,
- gst_video_chroma_to_string (info->chroma_site), NULL);
-
- gst_video_caps_set_colorimetry (caps, &info->colorimetry);
-
- if (info->views > 1)
- gst_caps_set_simple (caps, "views", G_TYPE_INT, info->views, NULL);
-
- if (info->flags & GST_VIDEO_FLAG_VARIABLE_FPS && info->fps_n != 0) {
- /* variable fps with a max-framerate */
- gst_caps_set_simple (caps, "framerate", GST_TYPE_FRACTION, 0, 1,
- "max-framerate", GST_TYPE_FRACTION, info->fps_n, info->fps_d, NULL);
- } else {
- /* no variable fps or no max-framerate */
- gst_caps_set_simple (caps, "framerate", GST_TYPE_FRACTION,
- info->fps_n, info->fps_d, NULL);
- }
-
- return caps;
-}
-
/**
- * gst_video_frame_map_id:
- * @frame: pointer to #GstVideoFrame
- * @info: a #GstVideoInfo
- * @buffer: the buffer to map
- * @id: the frame id to map
- * @flags: #GstMapFlags
- *
- * Use @info and @buffer to fill in the values of @frame with the video frame
- * information of frame @id.
+ * gst_video_calculate_display_ratio:
+ * @dar_n: Numerator of the calculated display_ratio
+ * @dar_d: Denominator of the calculated display_ratio
+ * @video_width: Width of the video frame in pixels
+ * @video_height: Height of the video frame in pixels
+ * @video_par_n: Numerator of the pixel aspect ratio of the input video.
+ * @video_par_d: Denominator of the pixel aspect ratio of the input video.
+ * @display_par_n: Numerator of the pixel aspect ratio of the display device
+ * @display_par_d: Denominator of the pixel aspect ratio of the display device
*
- * When @id is -1, the default frame is mapped. When @id != -1, this function
- * will return %FALSE when there is no GstVideoMeta with that id.
+ * Given the Pixel Aspect Ratio and size of an input video frame, and the
+ * pixel aspect ratio of the intended display device, calculates the actual
+ * display ratio the video will be rendered with.
*
- * All video planes of @buffer will be mapped and the pointers will be set in
- * @frame->data.
+ * Returns: A boolean indicating success and a calculated Display Ratio in the
+ * dar_n and dar_d parameters.
+ * The return value is FALSE in the case of integer overflow or other error.
*
- * Returns: %TRUE on success.
+ * Since: 0.10.7
*/
gboolean
-gst_video_frame_map_id (GstVideoFrame * frame, GstVideoInfo * info,
- GstBuffer * buffer, gint id, GstMapFlags flags)
+gst_video_calculate_display_ratio (guint * dar_n, guint * dar_d,
+ guint video_width, guint video_height,
+ guint video_par_n, guint video_par_d,
+ guint display_par_n, guint display_par_d)
{
- GstVideoMeta *meta;
- gint i;
-
- g_return_val_if_fail (frame != NULL, FALSE);
- g_return_val_if_fail (info != NULL, FALSE);
- g_return_val_if_fail (GST_IS_BUFFER (buffer), FALSE);
-
- if (id == -1)
- meta = gst_buffer_get_video_meta (buffer);
- else
- meta = gst_buffer_get_video_meta_id (buffer, id);
-
- /* copy the info */
- frame->info = *info;
-
- if (meta) {
- frame->info.finfo = &formats[meta->format].info;
- frame->info.width = meta->width;
- frame->info.height = meta->height;
- frame->id = meta->id;
- frame->flags = meta->flags;
-
- for (i = 0; i < info->finfo->n_planes; i++)
- gst_video_meta_map (meta, i, &frame->map[i], &frame->data[i],
- &frame->info.stride[i], flags);
- } else {
- /* no metadata, we really need to have the metadata when the id is
- * specified. */
- if (id != -1)
- goto no_metadata;
+ gint num, den;
+ gint tmp_n, tmp_d;
- frame->id = id;
- frame->flags = 0;
+ g_return_val_if_fail (dar_n != NULL, FALSE);
+ g_return_val_if_fail (dar_d != NULL, FALSE);
- if (GST_VIDEO_INFO_IS_INTERLACED (info)) {
- if (GST_BUFFER_FLAG_IS_SET (buffer, GST_VIDEO_BUFFER_FLAG_TFF))
- frame->flags |= GST_VIDEO_FRAME_FLAG_TFF;
- if (GST_BUFFER_FLAG_IS_SET (buffer, GST_VIDEO_BUFFER_FLAG_RFF))
- frame->flags |= GST_VIDEO_FRAME_FLAG_RFF;
- if (GST_BUFFER_FLAG_IS_SET (buffer, GST_VIDEO_BUFFER_FLAG_ONEFIELD))
- frame->flags |= GST_VIDEO_FRAME_FLAG_ONEFIELD;
- }
+ /* Calculate (video_width * video_par_n * display_par_d) /
+ * (video_height * video_par_d * display_par_n) */
+ if (!gst_util_fraction_multiply (video_width, video_height, video_par_n,
+ video_par_d, &tmp_n, &tmp_d))
+ goto error_overflow;
- gst_buffer_map (buffer, &frame->map[0], flags);
+ if (!gst_util_fraction_multiply (tmp_n, tmp_d, display_par_d, display_par_n,
+ &num, &den))
+ goto error_overflow;
- /* do some sanity checks */
- if (frame->map[0].size < info->size)
- goto invalid_size;
+ g_return_val_if_fail (num > 0, FALSE);
+ g_return_val_if_fail (den > 0, FALSE);
- /* set up pointers */
- for (i = 0; i < info->finfo->n_planes; i++) {
- frame->data[i] = frame->map[0].data + info->offset[i];
- }
- }
- frame->buffer = gst_buffer_ref (buffer);
- frame->meta = meta;
+ *dar_n = num;
+ *dar_d = den;
return TRUE;
/* ERRORS */
-no_metadata:
- {
- GST_ERROR ("no GstVideoMeta for id %d", id);
- return FALSE;
- }
-invalid_size:
+error_overflow:
{
- GST_ERROR ("invalid buffer size %" G_GSIZE_FORMAT " < %" G_GSIZE_FORMAT,
- frame->map[0].size, info->size);
- gst_buffer_unmap (buffer, &frame->map[0]);
+ GST_WARNING ("overflow in multiply");
return FALSE;
}
}
-
-/**
- * gst_video_frame_map:
- * @frame: pointer to #GstVideoFrame
- * @info: a #GstVideoInfo
- * @buffer: the buffer to map
- * @flags: #GstMapFlags
- *
- * Use @info and @buffer to fill in the values of @frame.
- *
- * All video planes of @buffer will be mapped and the pointers will be set in
- * @frame->data.
- *
- * Returns: %TRUE on success.
- */
-gboolean
-gst_video_frame_map (GstVideoFrame * frame, GstVideoInfo * info,
- GstBuffer * buffer, GstMapFlags flags)
-{
- return gst_video_frame_map_id (frame, info, buffer, -1, flags);
-}
-
-/**
- * gst_video_frame_unmap:
- * @frame: a #GstVideoFrame
- *
- * Unmap the memory previously mapped with gst_video_frame_map.
- */
-void
-gst_video_frame_unmap (GstVideoFrame * frame)
-{
- GstBuffer *buffer;
- GstVideoMeta *meta;
- gint i;
-
- g_return_if_fail (frame != NULL);
-
- buffer = frame->buffer;
- meta = frame->meta;
-
- if (meta) {
- for (i = 0; i < frame->info.finfo->n_planes; i++) {
- gst_video_meta_unmap (meta, i, &frame->map[i]);
- }
- } else {
- gst_buffer_unmap (buffer, &frame->map[0]);
- }
- gst_buffer_unref (buffer);
-}
-
-/**
- * gst_video_frame_copy:
- * @dest: a #GstVideoFrame
- * @src: a #GstVideoFrame
- * @plane: a plane
- *
- * Copy the plane with index @plane from @src to @dest.
- *
- * Returns: TRUE if the contents could be copied.
- */
-gboolean
-gst_video_frame_copy_plane (GstVideoFrame * dest, const GstVideoFrame * src,
- guint plane)
-{
- const GstVideoInfo *sinfo;
- GstVideoInfo *dinfo;
- guint w, h, j;
- guint8 *sp, *dp;
- gint ss, ds;
-
- g_return_val_if_fail (dest != NULL, FALSE);
- g_return_val_if_fail (src != NULL, FALSE);
-
- sinfo = &src->info;
- dinfo = &dest->info;
-
- g_return_val_if_fail (dinfo->finfo->format == sinfo->finfo->format, FALSE);
- g_return_val_if_fail (dinfo->width == sinfo->width
- && dinfo->height == sinfo->height, FALSE);
- g_return_val_if_fail (dinfo->finfo->n_planes > plane, FALSE);
-
- sp = src->data[plane];
- dp = dest->data[plane];
-
- ss = sinfo->stride[plane];
- ds = dinfo->stride[plane];
-
- w = MIN (ABS (ss), ABS (ds));
- h = GST_VIDEO_FRAME_COMP_HEIGHT (dest, plane);
-
- GST_CAT_DEBUG (GST_CAT_PERFORMANCE, "copy plane %d, w:%d h:%d ", plane, w, h);
-
- for (j = 0; j < h; j++) {
- memcpy (dp, sp, w);
- dp += ds;
- sp += ss;
- }
- return TRUE;
-}
-
-/**
- * gst_video_frame_copy:
- * @dest: a #GstVideoFrame
- * @src: a #GstVideoFrame
- *
- * Copy the contents from @src to @dest.
- *
- * Returns: TRUE if the contents could be copied.
- */
-gboolean
-gst_video_frame_copy (GstVideoFrame * dest, const GstVideoFrame * src)
-{
- guint i, n_planes;
- const GstVideoInfo *sinfo;
- GstVideoInfo *dinfo;
-
- g_return_val_if_fail (dest != NULL, FALSE);
- g_return_val_if_fail (src != NULL, FALSE);
-
- sinfo = &src->info;
- dinfo = &dest->info;
-
- g_return_val_if_fail (dinfo->finfo->format == sinfo->finfo->format, FALSE);
- g_return_val_if_fail (dinfo->width == sinfo->width
- && dinfo->height == sinfo->height, FALSE);
-
- n_planes = dinfo->finfo->n_planes;
- if (GST_VIDEO_FORMAT_INFO_HAS_PALETTE (sinfo->finfo)) {
- memcpy (dest->data[1], src->data[1], 256 * 4);
- n_planes = 1;
- }
-
- for (i = 0; i < n_planes; i++)
- gst_video_frame_copy_plane (dest, src, i);
-
- return TRUE;
-}
-
-static int
-fill_planes (GstVideoInfo * info)
-{
- gint width, height;
-
- width = info->width;
- height = info->height;
-
- switch (info->finfo->format) {
- case GST_VIDEO_FORMAT_YUY2:
- case GST_VIDEO_FORMAT_YVYU:
- case GST_VIDEO_FORMAT_UYVY:
- info->stride[0] = GST_ROUND_UP_4 (width * 2);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_AYUV:
- case GST_VIDEO_FORMAT_RGBx:
- case GST_VIDEO_FORMAT_RGBA:
- case GST_VIDEO_FORMAT_BGRx:
- case GST_VIDEO_FORMAT_BGRA:
- case GST_VIDEO_FORMAT_xRGB:
- case GST_VIDEO_FORMAT_ARGB:
- case GST_VIDEO_FORMAT_xBGR:
- case GST_VIDEO_FORMAT_ABGR:
- case GST_VIDEO_FORMAT_r210:
- info->stride[0] = width * 4;
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_RGB16:
- case GST_VIDEO_FORMAT_BGR16:
- case GST_VIDEO_FORMAT_RGB15:
- case GST_VIDEO_FORMAT_BGR15:
- info->stride[0] = GST_ROUND_UP_4 (width * 2);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_RGB:
- case GST_VIDEO_FORMAT_BGR:
- case GST_VIDEO_FORMAT_v308:
- info->stride[0] = GST_ROUND_UP_4 (width * 3);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_v210:
- info->stride[0] = ((width + 47) / 48) * 128;
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_v216:
- info->stride[0] = GST_ROUND_UP_8 (width * 4);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_GRAY8:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_GRAY16_BE:
- case GST_VIDEO_FORMAT_GRAY16_LE:
- info->stride[0] = GST_ROUND_UP_4 (width * 2);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_UYVP:
- info->stride[0] = GST_ROUND_UP_4 ((width * 2 * 5 + 3) / 4);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_RGB8P:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = 4;
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * height;
- info->size = info->offset[1] + (4 * 256);
- break;
- case GST_VIDEO_FORMAT_IYU1:
- info->stride[0] = GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) +
- GST_ROUND_UP_4 (width) / 2);
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_ARGB64:
- case GST_VIDEO_FORMAT_AYUV64:
- info->stride[0] = width * 8;
- info->offset[0] = 0;
- info->size = info->stride[0] * height;
- break;
- case GST_VIDEO_FORMAT_I420:
- case GST_VIDEO_FORMAT_YV12: /* same as I420, but plane 1+2 swapped */
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
- info->stride[2] = info->stride[1];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
- info->offset[2] = info->offset[1] +
- info->stride[1] * (GST_ROUND_UP_2 (height) / 2);
- info->size = info->offset[2] +
- info->stride[2] * (GST_ROUND_UP_2 (height) / 2);
- break;
- case GST_VIDEO_FORMAT_Y41B:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = GST_ROUND_UP_16 (width) / 4;
- info->stride[2] = info->stride[1];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * height;
- info->offset[2] = info->offset[1] + info->stride[1] * height;
- /* simplification of ROUNDUP4(w)*h + 2*((ROUNDUP16(w)/4)*h */
- info->size = (info->stride[0] + (GST_ROUND_UP_16 (width) / 2)) * height;
- break;
- case GST_VIDEO_FORMAT_Y42B:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = GST_ROUND_UP_8 (width) / 2;
- info->stride[2] = info->stride[1];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * height;
- info->offset[2] = info->offset[1] + info->stride[1] * height;
- /* simplification of ROUNDUP4(w)*h + 2*(ROUNDUP8(w)/2)*h */
- info->size = (info->stride[0] + GST_ROUND_UP_8 (width)) * height;
- break;
- case GST_VIDEO_FORMAT_Y444:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = info->stride[0];
- info->stride[2] = info->stride[0];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * height;
- info->offset[2] = info->offset[1] * 2;
- info->size = info->stride[0] * height * 3;
- break;
- case GST_VIDEO_FORMAT_NV12:
- case GST_VIDEO_FORMAT_NV21:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = info->stride[0];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
- info->size = info->stride[0] * GST_ROUND_UP_2 (height) * 3 / 2;
- break;
- case GST_VIDEO_FORMAT_A420:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
- info->stride[2] = info->stride[1];
- info->stride[3] = info->stride[0];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
- info->offset[2] = info->offset[1] +
- info->stride[1] * (GST_ROUND_UP_2 (height) / 2);
- info->offset[3] = info->offset[2] +
- info->stride[2] * (GST_ROUND_UP_2 (height) / 2);
- info->size = info->offset[3] + info->stride[0];
- break;
- case GST_VIDEO_FORMAT_YUV9:
- case GST_VIDEO_FORMAT_YVU9:
- info->stride[0] = GST_ROUND_UP_4 (width);
- info->stride[1] = GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4);
- info->stride[2] = info->stride[1];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * height;
- info->offset[2] = info->offset[1] +
- info->stride[1] * (GST_ROUND_UP_4 (height) / 4);
- info->size = info->offset[2] +
- info->stride[2] * (GST_ROUND_UP_4 (height) / 4);
- break;
- case GST_VIDEO_FORMAT_I420_10LE:
- case GST_VIDEO_FORMAT_I420_10BE:
- info->stride[0] = GST_ROUND_UP_4 (width * 2);
- info->stride[1] = GST_ROUND_UP_4 (width);
- info->stride[2] = info->stride[1];
- info->offset[0] = 0;
- info->offset[1] = info->stride[0] * GST_ROUND_UP_2 (height);
- info->offset[2] = info->offset[1] +
- info->stride[1] * (GST_ROUND_UP_2 (height) / 2);
- info->size = info->offset[2] +
- info->stride[2] * (GST_ROUND_UP_2 (height) / 2);
- break;
- case GST_VIDEO_FORMAT_ENCODED:
- break;
- case GST_VIDEO_FORMAT_UNKNOWN:
- GST_ERROR ("invalid format");
- g_warning ("invalid format");
- break;
- }
- return 0;
-}
-
-/**
- * gst_video_format_convert:
- * @info: a #GstVideoInfo
- * @src_format: #GstFormat of the @src_value
- * @src_value: value to convert
- * @dest_format: #GstFormat of the @dest_value
- * @dest_value: pointer to destination value
- *
- * Converts among various #GstFormat types. This function handles
- * GST_FORMAT_BYTES, GST_FORMAT_TIME, and GST_FORMAT_DEFAULT. For
- * raw video, GST_FORMAT_DEFAULT corresponds to video frames. This
- * function can be used to handle pad queries of the type GST_QUERY_CONVERT.
- *
- * Since: 0.10.16
- *
- * Returns: TRUE if the conversion was successful.
- */
-gboolean
-gst_video_info_convert (GstVideoInfo * info,
- GstFormat src_format, gint64 src_value,
- GstFormat dest_format, gint64 * dest_value)
-{
- gboolean ret = FALSE;
- int size, fps_n, fps_d;
-
- g_return_val_if_fail (info != NULL, 0);
- g_return_val_if_fail (info->finfo != NULL, 0);
- g_return_val_if_fail (info->finfo->format != GST_VIDEO_FORMAT_UNKNOWN, 0);
- g_return_val_if_fail (info->size > 0, 0);
-
- size = info->size;
- fps_n = info->fps_n;
- fps_d = info->fps_d;
-
- GST_DEBUG ("converting value %" G_GINT64_FORMAT " from %s to %s",
- src_value, gst_format_get_name (src_format),
- gst_format_get_name (dest_format));
-
- if (src_format == dest_format) {
- *dest_value = src_value;
- ret = TRUE;
- goto done;
- }
-
- if (src_value == -1) {
- *dest_value = -1;
- ret = TRUE;
- goto done;
- }
-
- /* bytes to frames */
- if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_DEFAULT) {
- if (size != 0) {
- *dest_value = gst_util_uint64_scale_int (src_value, 1, size);
- } else {
- GST_ERROR ("blocksize is 0");
- *dest_value = 0;
- }
- ret = TRUE;
- goto done;
- }
-
- /* frames to bytes */
- if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_BYTES) {
- *dest_value = gst_util_uint64_scale_int (src_value, size, 1);
- ret = TRUE;
- goto done;
- }
-
- /* time to frames */
- if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_DEFAULT) {
- if (fps_d != 0) {
- *dest_value = gst_util_uint64_scale (src_value,
- fps_n, GST_SECOND * fps_d);
- } else {
- GST_ERROR ("framerate denominator is 0");
- *dest_value = 0;
- }
- ret = TRUE;
- goto done;
- }
-
- /* frames to time */
- if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_TIME) {
- if (fps_n != 0) {
- *dest_value = gst_util_uint64_scale (src_value,
- GST_SECOND * fps_d, fps_n);
- } else {
- GST_ERROR ("framerate numerator is 0");
- *dest_value = 0;
- }
- ret = TRUE;
- goto done;
- }
-
- /* time to bytes */
- if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_BYTES) {
- if (fps_d != 0) {
- *dest_value = gst_util_uint64_scale (src_value,
- fps_n * size, GST_SECOND * fps_d);
- } else {
- GST_ERROR ("framerate denominator is 0");
- *dest_value = 0;
- }
- ret = TRUE;
- goto done;
- }
-
- /* bytes to time */
- if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_TIME) {
- if (fps_n != 0 && size != 0) {
- *dest_value = gst_util_uint64_scale (src_value,
- GST_SECOND * fps_d, fps_n * size);
- } else {
- GST_ERROR ("framerate denominator and/or blocksize is 0");
- *dest_value = 0;
- }
- ret = TRUE;
- }
-
-done:
-
- GST_DEBUG ("ret=%d result %" G_GINT64_FORMAT, ret, *dest_value);
-
- return ret;
-}
-
-#define GST_VIDEO_EVENT_STILL_STATE_NAME "GstEventStillFrame"
-
-/**
- * gst_video_event_new_still_frame:
- * @in_still: boolean value for the still-frame state of the event.
- *
- * Creates a new Still Frame event. If @in_still is %TRUE, then the event
- * represents the start of a still frame sequence. If it is %FALSE, then
- * the event ends a still frame sequence.
- *
- * To parse an event created by gst_video_event_new_still_frame() use
- * gst_video_event_parse_still_frame().
- *
- * Returns: The new GstEvent
- * Since: 0.10.26
- */
-GstEvent *
-gst_video_event_new_still_frame (gboolean in_still)
-{
- GstEvent *still_event;
- GstStructure *s;
-
- s = gst_structure_new (GST_VIDEO_EVENT_STILL_STATE_NAME,
- "still-state", G_TYPE_BOOLEAN, in_still, NULL);
- still_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s);
-
- return still_event;
-}
-
-/**
- * gst_video_event_parse_still_frame:
- * @event: A #GstEvent to parse
- * @in_still: A boolean to receive the still-frame status from the event, or NULL
- *
- * Parse a #GstEvent, identify if it is a Still Frame event, and
- * return the still-frame state from the event if it is.
- * If the event represents the start of a still frame, the in_still
- * variable will be set to TRUE, otherwise FALSE. It is OK to pass NULL for the
- * in_still variable order to just check whether the event is a valid still-frame
- * event.
- *
- * Create a still frame event using gst_video_event_new_still_frame()
- *
- * Returns: %TRUE if the event is a valid still-frame event. %FALSE if not
- * Since: 0.10.26
- */
-gboolean
-gst_video_event_parse_still_frame (GstEvent * event, gboolean * in_still)
-{
- const GstStructure *s;
- gboolean ev_still_state;
-
- g_return_val_if_fail (event != NULL, FALSE);
-
- if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM)
- return FALSE; /* Not a still frame event */
-
- s = gst_event_get_structure (event);
- if (s == NULL
- || !gst_structure_has_name (s, GST_VIDEO_EVENT_STILL_STATE_NAME))
- return FALSE; /* Not a still frame event */
- if (!gst_structure_get_boolean (s, "still-state", &ev_still_state))
- return FALSE; /* Not a still frame event */
- if (in_still)
- *in_still = ev_still_state;
- return TRUE;
-}
-
-#define GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME "GstForceKeyUnit"
-
-/**
- * gst_video_event_new_downstream_force_key_unit:
- * @timestamp: the timestamp of the buffer that starts a new key unit
- * @stream_time: the stream_time of the buffer that starts a new key unit
- * @running_time: the running_time of the buffer that starts a new key unit
- * @all_headers: %TRUE to produce headers when starting a new key unit
- * @count: integer that can be used to number key units
- *
- * Creates a new downstream force key unit event. A downstream force key unit
- * event can be sent down the pipeline to request downstream elements to produce
- * a key unit. A downstream force key unit event must also be sent when handling
- * an upstream force key unit event to notify downstream that the latter has been
- * handled.
- *
- * To parse an event created by gst_video_event_new_downstream_force_key_unit() use
- * gst_video_event_parse_downstream_force_key_unit().
- *
- * Returns: The new GstEvent
- * Since: 0.10.36
- */
-GstEvent *
-gst_video_event_new_downstream_force_key_unit (GstClockTime timestamp,
- GstClockTime stream_time, GstClockTime running_time, gboolean all_headers,
- guint count)
-{
- GstEvent *force_key_unit_event;
- GstStructure *s;
-
- s = gst_structure_new (GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME,
- "timestamp", G_TYPE_UINT64, timestamp,
- "stream-time", G_TYPE_UINT64, stream_time,
- "running-time", G_TYPE_UINT64, running_time,
- "all-headers", G_TYPE_BOOLEAN, all_headers,
- "count", G_TYPE_UINT, count, NULL);
- force_key_unit_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s);
-
- return force_key_unit_event;
-}
-
-/**
- * gst_video_event_new_upstream_force_key_unit:
- * @running_time: the running_time at which a new key unit should be produced
- * @all_headers: %TRUE to produce headers when starting a new key unit
- * @count: integer that can be used to number key units
- *
- * Creates a new upstream force key unit event. An upstream force key unit event
- * can be sent to request upstream elements to produce a key unit.
- *
- * @running_time can be set to request a new key unit at a specific
- * running_time. If set to GST_CLOCK_TIME_NONE, upstream elements will produce a
- * new key unit as soon as possible.
- *
- * To parse an event created by gst_video_event_new_downstream_force_key_unit() use
- * gst_video_event_parse_downstream_force_key_unit().
- *
- * Returns: The new GstEvent
- * Since: 0.10.36
- */
-GstEvent *
-gst_video_event_new_upstream_force_key_unit (GstClockTime running_time,
- gboolean all_headers, guint count)
-{
- GstEvent *force_key_unit_event;
- GstStructure *s;
-
- s = gst_structure_new (GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME,
- "running-time", GST_TYPE_CLOCK_TIME, running_time,
- "all-headers", G_TYPE_BOOLEAN, all_headers,
- "count", G_TYPE_UINT, count, NULL);
- force_key_unit_event = gst_event_new_custom (GST_EVENT_CUSTOM_UPSTREAM, s);
-
- return force_key_unit_event;
-}
-
-/**
- * gst_video_event_is_force_key_unit:
- * @event: A #GstEvent to check
- *
- * Checks if an event is a force key unit event. Returns true for both upstream
- * and downstream force key unit events.
- *
- * Returns: %TRUE if the event is a valid force key unit event
- * Since: 0.10.36
- */
-gboolean
-gst_video_event_is_force_key_unit (GstEvent * event)
-{
- const GstStructure *s;
-
- g_return_val_if_fail (event != NULL, FALSE);
-
- if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM &&
- GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_UPSTREAM)
- return FALSE; /* Not a force key unit event */
-
- s = gst_event_get_structure (event);
- if (s == NULL
- || !gst_structure_has_name (s, GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME))
- return FALSE;
-
- return TRUE;
-}
-
-/**
- * gst_video_event_parse_downstream_force_key_unit:
- * @event: A #GstEvent to parse
- * @timestamp: (out): A pointer to the timestamp in the event
- * @stream_time: (out): A pointer to the stream-time in the event
- * @running_time: (out): A pointer to the running-time in the event
- * @all_headers: (out): A pointer to the all_headers flag in the event
- * @count: (out): A pointer to the count field of the event
- *
- * Get timestamp, stream-time, running-time, all-headers and count in the force
- * key unit event. See gst_video_event_new_downstream_force_key_unit() for a
- * full description of the downstream force key unit event.
- *
- * Returns: %TRUE if the event is a valid downstream force key unit event.
- * Since: 0.10.36
- */
-gboolean
-gst_video_event_parse_downstream_force_key_unit (GstEvent * event,
- GstClockTime * timestamp, GstClockTime * stream_time,
- GstClockTime * running_time, gboolean * all_headers, guint * count)
-{
- const GstStructure *s;
- GstClockTime ev_timestamp, ev_stream_time, ev_running_time;
- gboolean ev_all_headers;
- guint ev_count;
-
- g_return_val_if_fail (event != NULL, FALSE);
-
- if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM)
- return FALSE; /* Not a force key unit event */
-
- s = gst_event_get_structure (event);
- if (s == NULL
- || !gst_structure_has_name (s, GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME))
- return FALSE;
-
- if (!gst_structure_get_clock_time (s, "timestamp", &ev_timestamp))
- ev_timestamp = GST_CLOCK_TIME_NONE;
- if (!gst_structure_get_clock_time (s, "stream-time", &ev_stream_time))
- ev_stream_time = GST_CLOCK_TIME_NONE;
- if (!gst_structure_get_clock_time (s, "running-time", &ev_running_time))
- ev_running_time = GST_CLOCK_TIME_NONE;
- if (!gst_structure_get_boolean (s, "all-headers", &ev_all_headers))
- ev_all_headers = FALSE;
- if (!gst_structure_get_uint (s, "count", &ev_count))
- ev_count = 0;
-
- if (timestamp)
- *timestamp = ev_timestamp;
-
- if (stream_time)
- *stream_time = ev_stream_time;
-
- if (running_time)
- *running_time = ev_running_time;
-
- if (all_headers)
- *all_headers = ev_all_headers;
-
- if (count)
- *count = ev_count;
-
- return TRUE;
-}
-
-/**
- * gst_video_event_parse_upstream_force_key_unit:
- * @event: A #GstEvent to parse
- * @running_time: (out): A pointer to the running_time in the event
- * @all_headers: (out): A pointer to the all_headers flag in the event
- * @count: (out): A pointer to the count field in the event
- *
- * Get running-time, all-headers and count in the force key unit event. See
- * gst_video_event_new_upstream_force_key_unit() for a full description of the
- * upstream force key unit event.
- *
- * Create an upstream force key unit event using gst_video_event_new_upstream_force_key_unit()
- *
- * Returns: %TRUE if the event is a valid upstream force-key-unit event. %FALSE if not
- * Since: 0.10.36
- */
-gboolean
-gst_video_event_parse_upstream_force_key_unit (GstEvent * event,
- GstClockTime * running_time, gboolean * all_headers, guint * count)
-{
- const GstStructure *s;
- GstClockTime ev_running_time;
- gboolean ev_all_headers;
- guint ev_count;
-
- g_return_val_if_fail (event != NULL, FALSE);
-
- if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_UPSTREAM)
- return FALSE; /* Not a force key unit event */
-
- s = gst_event_get_structure (event);
- if (s == NULL
- || !gst_structure_has_name (s, GST_VIDEO_EVENT_FORCE_KEY_UNIT_NAME))
- return FALSE;
-
- if (!gst_structure_get_clock_time (s, "running-time", &ev_running_time))
- ev_running_time = GST_CLOCK_TIME_NONE;
- if (!gst_structure_get_boolean (s, "all-headers", &ev_all_headers))
- ev_all_headers = FALSE;
- if (!gst_structure_get_uint (s, "count", &ev_count))
- ev_count = 0;
-
- if (running_time)
- *running_time = ev_running_time;
-
- if (all_headers)
- *all_headers = ev_all_headers;
-
- if (count)
- *count = ev_count;
-
- return TRUE;
-}
#define __GST_VIDEO_H__
#include <gst/gst.h>
-
-G_BEGIN_DECLS
-
+#include <gst/video/video-event.h>
+#include <gst/video/video-format.h>
+#include <gst/video/video-color.h>
+#include <gst/video/video-info.h>
+#include <gst/video/video-frame.h>
#include <gst/video/video-enumtypes.h>
-/**
- * GstVideoFormat:
- * @GST_VIDEO_FORMAT_UNKNOWN: Unknown or unset video format id
- * @GST_VIDEO_FORMAT_ENCODED: Encoded video format
- * @GST_VIDEO_FORMAT_I420: planar 4:2:0 YUV
- * @GST_VIDEO_FORMAT_YV12: planar 4:2:0 YVU (like I420 but UV planes swapped)
- * @GST_VIDEO_FORMAT_YUY2: packed 4:2:2 YUV (Y0-U0-Y1-V0 Y2-U2-Y3-V2 Y4 ...)
- * @GST_VIDEO_FORMAT_UYVY: packed 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...)
- * @GST_VIDEO_FORMAT_AYUV: packed 4:4:4 YUV with alpha channel (A0-Y0-U0-V0 ...)
- * @GST_VIDEO_FORMAT_RGBx: sparse rgb packed into 32 bit, space last
- * @GST_VIDEO_FORMAT_BGRx: sparse reverse rgb packed into 32 bit, space last
- * @GST_VIDEO_FORMAT_xRGB: sparse rgb packed into 32 bit, space first
- * @GST_VIDEO_FORMAT_xBGR: sparse reverse rgb packed into 32 bit, space first
- * @GST_VIDEO_FORMAT_RGBA: rgb with alpha channel last
- * @GST_VIDEO_FORMAT_BGRA: reverse rgb with alpha channel last
- * @GST_VIDEO_FORMAT_ARGB: rgb with alpha channel first
- * @GST_VIDEO_FORMAT_ABGR: reverse rgb with alpha channel first
- * @GST_VIDEO_FORMAT_RGB: rgb
- * @GST_VIDEO_FORMAT_BGR: reverse rgb
- * @GST_VIDEO_FORMAT_Y41B: planar 4:1:1 YUV (Since: 0.10.18)
- * @GST_VIDEO_FORMAT_Y42B: planar 4:2:2 YUV (Since: 0.10.18)
- * @GST_VIDEO_FORMAT_YVYU: packed 4:2:2 YUV (Y0-V0-Y1-U0 Y2-V2-Y3-U2 Y4 ...) (Since: 0.10.23)
- * @GST_VIDEO_FORMAT_Y444: planar 4:4:4 YUV (Since: 0.10.24)
- * @GST_VIDEO_FORMAT_v210: packed 4:2:2 10-bit YUV, complex format (Since: 0.10.24)
- * @GST_VIDEO_FORMAT_v216: packed 4:2:2 16-bit YUV, Y0-U0-Y1-V1 order (Since: 0.10.24)
- * @GST_VIDEO_FORMAT_NV12: planar 4:2:0 YUV with interleaved UV plane (Since: 0.10.26)
- * @GST_VIDEO_FORMAT_NV21: planar 4:2:0 YUV with interleaved VU plane (Since: 0.10.26)
- * @GST_VIDEO_FORMAT_GRAY8: 8-bit grayscale (Since: 0.10.29)
- * @GST_VIDEO_FORMAT_GRAY16_BE: 16-bit grayscale, most significant byte first (Since: 0.10.29)
- * @GST_VIDEO_FORMAT_GRAY16_LE: 16-bit grayscale, least significant byte first (Since: 0.10.29)
- * @GST_VIDEO_FORMAT_v308: packed 4:4:4 YUV (Since: 0.10.29)
- * @GST_VIDEO_FORMAT_RGB16: rgb 5-6-5 bits per component (Since: 0.10.30)
- * @GST_VIDEO_FORMAT_BGR16: reverse rgb 5-6-5 bits per component (Since: 0.10.30)
- * @GST_VIDEO_FORMAT_RGB15: rgb 5-5-5 bits per component (Since: 0.10.30)
- * @GST_VIDEO_FORMAT_BGR15: reverse rgb 5-5-5 bits per component (Since: 0.10.30)
- * @GST_VIDEO_FORMAT_UYVP: packed 10-bit 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...) (Since: 0.10.31)
- * @GST_VIDEO_FORMAT_A420: planar 4:4:2:0 AYUV (Since: 0.10.31)
- * @GST_VIDEO_FORMAT_RGB8P: 8-bit paletted RGB (Since: 0.10.32)
- * @GST_VIDEO_FORMAT_YUV9: planar 4:1:0 YUV (Since: 0.10.32)
- * @GST_VIDEO_FORMAT_YVU9: planar 4:1:0 YUV (like YUV9 but UV planes swapped) (Since: 0.10.32)
- * @GST_VIDEO_FORMAT_IYU1: packed 4:1:1 YUV (Cb-Y0-Y1-Cr-Y2-Y3 ...) (Since: 0.10.32)
- * @GST_VIDEO_FORMAT_ARGB64: rgb with alpha channel first, 16 bits per channel (Since: 0.10.33)
- * @GST_VIDEO_FORMAT_AYUV64: packed 4:4:4 YUV with alpha channel, 16 bits per channel (A0-Y0-U0-V0 ...) (Since: 0.10.33)
- * @GST_VIDEO_FORMAT_r210: packed 4:4:4 RGB, 10 bits per channel (Since: 0.10.33)
- * @GST_VIDEO_FORMAT_I420_10BE: planar 4:2:0 YUV, 10 bits per channel
- * @GST_VIDEO_FORMAT_I420_10LE: planar 4:2:0 YUV, 10 bits per channel
- *
- * Enum value describing the most common video formats.
- */
-typedef enum {
- GST_VIDEO_FORMAT_UNKNOWN,
- GST_VIDEO_FORMAT_ENCODED,
- GST_VIDEO_FORMAT_I420,
- GST_VIDEO_FORMAT_YV12,
- GST_VIDEO_FORMAT_YUY2,
- GST_VIDEO_FORMAT_UYVY,
- GST_VIDEO_FORMAT_AYUV,
- GST_VIDEO_FORMAT_RGBx,
- GST_VIDEO_FORMAT_BGRx,
- GST_VIDEO_FORMAT_xRGB,
- GST_VIDEO_FORMAT_xBGR,
- GST_VIDEO_FORMAT_RGBA,
- GST_VIDEO_FORMAT_BGRA,
- GST_VIDEO_FORMAT_ARGB,
- GST_VIDEO_FORMAT_ABGR,
- GST_VIDEO_FORMAT_RGB,
- GST_VIDEO_FORMAT_BGR,
- GST_VIDEO_FORMAT_Y41B,
- GST_VIDEO_FORMAT_Y42B,
- GST_VIDEO_FORMAT_YVYU,
- GST_VIDEO_FORMAT_Y444,
- GST_VIDEO_FORMAT_v210,
- GST_VIDEO_FORMAT_v216,
- GST_VIDEO_FORMAT_NV12,
- GST_VIDEO_FORMAT_NV21,
- GST_VIDEO_FORMAT_GRAY8,
- GST_VIDEO_FORMAT_GRAY16_BE,
- GST_VIDEO_FORMAT_GRAY16_LE,
- GST_VIDEO_FORMAT_v308,
- GST_VIDEO_FORMAT_RGB16,
- GST_VIDEO_FORMAT_BGR16,
- GST_VIDEO_FORMAT_RGB15,
- GST_VIDEO_FORMAT_BGR15,
- GST_VIDEO_FORMAT_UYVP,
- GST_VIDEO_FORMAT_A420,
- GST_VIDEO_FORMAT_RGB8P,
- GST_VIDEO_FORMAT_YUV9,
- GST_VIDEO_FORMAT_YVU9,
- GST_VIDEO_FORMAT_IYU1,
- GST_VIDEO_FORMAT_ARGB64,
- GST_VIDEO_FORMAT_AYUV64,
- GST_VIDEO_FORMAT_r210,
- GST_VIDEO_FORMAT_I420_10BE,
- GST_VIDEO_FORMAT_I420_10LE
-} GstVideoFormat;
-
-#define GST_VIDEO_MAX_PLANES 4
-#define GST_VIDEO_MAX_COMPONENTS 4
-
-typedef struct _GstVideoFormatInfo GstVideoFormatInfo;
-typedef struct _GstVideoColorimetry GstVideoColorimetry;
-
-/**
- * GstVideoChromaSite:
- * @GST_VIDEO_CHROMA_SITE_UNKNOWN: unknown cositing
- * @GST_VIDEO_CHROMA_SITE_NONE: no cositing
- * @GST_VIDEO_CHROMA_SITE_H_COSITED: chroma is horizontally cosited
- * @GST_VIDEO_CHROMA_SITE_V_COSITED: chroma is vertically cosited
- * @GST_VIDEO_CHROMA_SITE_ALT_LINE: choma samples are sited on alternate lines
- * @GST_VIDEO_CHROMA_SITE_COSITED: chroma samples cosited with luma samples
- * @GST_VIDEO_CHROMA_SITE_JPEG: jpeg style cositing, also for mpeg1 and mjpeg
- * @GST_VIDEO_CHROMA_SITE_MPEG2: mpeg2 style cositing
- * @GST_VIDEO_CHROMA_SITE_DV: DV style cositing
- *
- * Various Chroma sitings.
- */
-typedef enum {
- GST_VIDEO_CHROMA_SITE_UNKNOWN = 0,
- GST_VIDEO_CHROMA_SITE_NONE = (1 << 0),
- GST_VIDEO_CHROMA_SITE_H_COSITED = (1 << 1),
- GST_VIDEO_CHROMA_SITE_V_COSITED = (1 << 2),
- GST_VIDEO_CHROMA_SITE_ALT_LINE = (1 << 3),
- /* some common chroma cositing */
- GST_VIDEO_CHROMA_SITE_COSITED = (GST_VIDEO_CHROMA_SITE_H_COSITED | GST_VIDEO_CHROMA_SITE_V_COSITED),
- GST_VIDEO_CHROMA_SITE_JPEG = (GST_VIDEO_CHROMA_SITE_NONE),
- GST_VIDEO_CHROMA_SITE_MPEG2 = (GST_VIDEO_CHROMA_SITE_H_COSITED),
- GST_VIDEO_CHROMA_SITE_DV = (GST_VIDEO_CHROMA_SITE_COSITED | GST_VIDEO_CHROMA_SITE_ALT_LINE),
-} GstVideoChromaSite;
-
-/**
- * GstVideoFormatFlags:
- * @GST_VIDEO_FORMAT_FLAG_YUV: The video format is YUV, components are numbered
- * 0=Y, 1=U, 2=V.
- * @GST_VIDEO_FORMAT_FLAG_RGB: The video format is RGB, components are numbered
- * 0=R, 1=G, 2=B.
- * @GST_VIDEO_FORMAT_FLAG_GRAY: The video is gray, there is one gray component
- * with index 0.
- * @GST_VIDEO_FORMAT_FLAG_ALPHA: The video format has an alpha components with
- * the number 3.
- * @GST_VIDEO_FORMAT_FLAG_LE: The video format has data stored in little
- * endianness.
- * @GST_VIDEO_FORMAT_FLAG_PALETTE: The video format has a palette. The palette
- * is stored in the second plane and indexes are stored in the first plane.
- * @GST_VIDEO_FORMAT_FLAG_COMPLEX: The video format has a complex layout that
- * can't be described with the usual information in the #GstVideoFormatInfo.
- *
- * The different video flags that a format info can have.
- */
-typedef enum
-{
- GST_VIDEO_FORMAT_FLAG_YUV = (1 << 0),
- GST_VIDEO_FORMAT_FLAG_RGB = (1 << 1),
- GST_VIDEO_FORMAT_FLAG_GRAY = (1 << 2),
- GST_VIDEO_FORMAT_FLAG_ALPHA = (1 << 3),
- GST_VIDEO_FORMAT_FLAG_LE = (1 << 4),
- GST_VIDEO_FORMAT_FLAG_PALETTE = (1 << 5),
- GST_VIDEO_FORMAT_FLAG_COMPLEX = (1 << 6)
-} GstVideoFormatFlags;
-
-/* YUV components */
-#define GST_VIDEO_COMP_Y 0
-#define GST_VIDEO_COMP_U 1
-#define GST_VIDEO_COMP_V 2
-
-/* RGB components */
-#define GST_VIDEO_COMP_R 0
-#define GST_VIDEO_COMP_G 1
-#define GST_VIDEO_COMP_B 2
-
-/* alpha component */
-#define GST_VIDEO_COMP_A 3
-
-/* palette components */
-#define GST_VIDEO_COMP_INDEX 0
-#define GST_VIDEO_COMP_PALETTE 1
-
-/**
- * GstVideoPackFlags:
- * @GST_VIDEO_PACK_FLAG_NONE: No flag
- *
- * The different flags that can be used when packing and unpacking.
- */
-typedef enum
-{
- GST_VIDEO_PACK_FLAG_NONE = 0
-} GstVideoPackFlags;
-
-/**
- * GstVideoFormatUnpack:
- * @info: a #GstVideoFormatInfo
- * @flags: flags to control the unpacking
- * @dest: a destination array
- * @data: pointers to the data planes
- * @stride: strides of the planes
- * @x: the x position in the image to start from
- * @y: the y position in the image to start from
- * @width: the amount of pixels to unpack.
- *
- * Unpacks @width pixels from the given planes and strides containing data of
- * format @info. The pixels will be unpacked into @dest which each component
- * interleaved. @dest should at least be big enough to hold @width *
- * n_components * size(unpack_format) bytes.
- */
-typedef void (*GstVideoFormatUnpack) (const GstVideoFormatInfo *info,
- GstVideoPackFlags flags, gpointer dest,
- const gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES],
- gint x, gint y, gint width);
-/**
- * GstVideoFormatPack:
- * @info: a #GstVideoFormatInfo
- * @flags: flags to control the packing
- * @src: a source array
- * @sstride: the source array stride
- * @data: pointers to the destination data planes
- * @stride: strides of the destination planes
- * @chroma_site: the chroma siting of the target when subsampled
- * @y: the y position in the image to pack to
- * @width: the amount of pixels to pack.
- *
- * Packs @width pixels from @src to the given planes and strides in the
- * format @info. The pixels from source have each component interleaved
- * and will be packed into the planes in @data.
- *
- * When @info refers to a format with a h_sub != 0, this function operates on
- * (1 << h_sub) lines meaning that @src should contain at least (1 << h_sub)
- * lines with a stride of @sstride and @y should be a multiple of (1 << h_sub).
- */
-typedef void (*GstVideoFormatPack) (const GstVideoFormatInfo *info,
- GstVideoPackFlags flags,
- const gpointer src, gint sstride,
- gpointer data[GST_VIDEO_MAX_PLANES],
- const gint stride[GST_VIDEO_MAX_PLANES],
- GstVideoChromaSite chroma_site,
- gint y, gint width);
-
-/**
- * GstVideoFormatInfo:
- * @format: #GstVideoFormat
- * @name: string representation of the format
- * @description: use readable description of the format
- * @flags: #GstVideoFormatFlags
- * @bits: The number of bits used to pack data items. This can be less than 8
- * when multiple pixels are stored in a byte. for values > 8 multiple bytes
- * should be read according to the endianness flag before applying the shift
- * and mask.
- * @n_components: the number of components in the video format.
- * @shift: the number of bits to shift away to get the component data
- * @depth: the depth in bits for each component
- * @pixel_stride: the pixel stride of each component. This is the amount of
- * bytes to the pixel immediately to the right. When bits < 8, the stride is
- * expressed in bits.
- * @n_planes: the number of planes for this format. The number of planes can be
- * less than the amount of components when multiple components are packed into
- * one plane.
- * @plane: the plane number where a component can be found
- * @poffset: the offset in the plane where the first pixel of the components
- * can be found. If bits < 8 the amount is specified in bits.
- * @w_sub: subsampling factor of the width for the component. Use
- * GST_VIDEO_SUB_SCALE to scale a width.
- * @h_sub: subsampling factor of the height for the component. Use
- * GST_VIDEO_SUB_SCALE to scale a height.
- * @unpack_format: the format of the unpacked pixels.
- * @unpack_func: an unpack function for this format
- * @pack_func: an pack function for this format
- *
- * Information for a video format.
- */
-struct _GstVideoFormatInfo {
- GstVideoFormat format;
- const gchar *name;
- const gchar *description;
- GstVideoFormatFlags flags;
- guint bits;
- guint n_components;
- guint shift[GST_VIDEO_MAX_COMPONENTS];
- guint depth[GST_VIDEO_MAX_COMPONENTS];
- gint pixel_stride[GST_VIDEO_MAX_COMPONENTS];
- guint n_planes;
- guint plane[GST_VIDEO_MAX_COMPONENTS];
- guint poffset[GST_VIDEO_MAX_COMPONENTS];
- guint w_sub[GST_VIDEO_MAX_COMPONENTS];
- guint h_sub[GST_VIDEO_MAX_COMPONENTS];
-
- GstVideoFormat unpack_format;
- GstVideoFormatUnpack unpack_func;
- GstVideoFormatPack pack_func;
-
- gpointer _gst_reserved[GST_PADDING];
-};
-
-#define GST_VIDEO_FORMAT_INFO_FORMAT(info) ((info)->format)
-#define GST_VIDEO_FORMAT_INFO_NAME(info) ((info)->name)
-#define GST_VIDEO_FORMAT_INFO_FLAGS(info) ((info)->flags)
-
-#define GST_VIDEO_FORMAT_INFO_IS_YUV(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_YUV)
-#define GST_VIDEO_FORMAT_INFO_IS_RGB(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_RGB)
-#define GST_VIDEO_FORMAT_INFO_IS_GRAY(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_GRAY)
-#define GST_VIDEO_FORMAT_INFO_HAS_ALPHA(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_ALPHA)
-#define GST_VIDEO_FORMAT_INFO_IS_LE(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_LE)
-#define GST_VIDEO_FORMAT_INFO_HAS_PALETTE(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_PALETTE)
-#define GST_VIDEO_FORMAT_INFO_IS_COMPLEX(info) ((info)->flags & GST_VIDEO_FORMAT_FLAG_COMPLEX)
-
-#define GST_VIDEO_FORMAT_INFO_BITS(info) ((info)->bits)
-#define GST_VIDEO_FORMAT_INFO_N_COMPONENTS(info) ((info)->n_components)
-#define GST_VIDEO_FORMAT_INFO_SHIFT(info,c) ((info)->shift[c])
-#define GST_VIDEO_FORMAT_INFO_DEPTH(info,c) ((info)->depth[c])
-#define GST_VIDEO_FORMAT_INFO_PSTRIDE(info,c) ((info)->pixel_stride[c])
-#define GST_VIDEO_FORMAT_INFO_N_PLANES(info) ((info)->n_planes)
-#define GST_VIDEO_FORMAT_INFO_PLANE(info,c) ((info)->plane[c])
-#define GST_VIDEO_FORMAT_INFO_POFFSET(info,c) ((info)->poffset[c])
-#define GST_VIDEO_FORMAT_INFO_W_SUB(info,c) ((info)->w_sub[c])
-#define GST_VIDEO_FORMAT_INFO_H_SUB(info,c) ((info)->h_sub[c])
-
-/* rounds up */
-#define GST_VIDEO_SUB_SCALE(scale,val) (-((-((gint)(val)))>>(scale)))
-
-#define GST_VIDEO_FORMAT_INFO_SCALE_WIDTH(info,c,w) GST_VIDEO_SUB_SCALE ((info)->w_sub[c],(w))
-#define GST_VIDEO_FORMAT_INFO_SCALE_HEIGHT(info,c,h) GST_VIDEO_SUB_SCALE ((info)->h_sub[c],(h))
-
-#define GST_VIDEO_FORMAT_INFO_DATA(info,planes,comp) \
- (((guint8*)(planes)[(info)->plane[comp]]) + (info)->poffset[comp])
-#define GST_VIDEO_FORMAT_INFO_STRIDE(info,strides,comp) ((strides)[(info)->plane[comp]])
-#define GST_VIDEO_FORMAT_INFO_OFFSET(info,offsets,comp) \
- (((offsets)[(info)->plane[comp]]) + (info)->poffset[comp])
-
-/* format properties */
-GstVideoFormat gst_video_format_from_masks (gint depth, gint bpp, gint endianness,
- gint red_mask, gint green_mask,
- gint blue_mask, gint alpha_mask) G_GNUC_CONST;
-
-GstVideoFormat gst_video_format_from_fourcc (guint32 fourcc) G_GNUC_CONST;
-GstVideoFormat gst_video_format_from_string (const gchar *format) G_GNUC_CONST;
-
-guint32 gst_video_format_to_fourcc (GstVideoFormat format) G_GNUC_CONST;
-const gchar * gst_video_format_to_string (GstVideoFormat format) G_GNUC_CONST;
-
-const GstVideoFormatInfo *
- gst_video_format_get_info (GstVideoFormat format) G_GNUC_CONST;
-
-typedef struct _GstVideoInfo GstVideoInfo;
-typedef struct _GstVideoFrame GstVideoFrame;
-
-/**
- * GstVideoInterlaceMode:
- * @GST_VIDEO_INTERLACE_MODE_PROGRESSIVE: all frames are progressive
- * @GST_VIDEO_INTERLACE_MODE_INTERLEAVED: 2 fields are interleaved in one video
- * frame. Extra buffer flags describe the field order.
- * @GST_VIDEO_INTERLACE_MODE_MIXED: frames contains both interlaced and
- * progressive video, the buffer flags describe the frame and fields.
- * @GST_VIDEO_INTERLACE_MODE_FIELDS: 2 fields are stored in one buffer, use the
- * frame ID to get access to the required field. For multiview (the
- * 'views' property > 1) the fields of view N can be found at frame ID
- * (N * 2) and (N * 2) + 1.
- * Each field has only half the amount of lines as noted in the
- * height property. This mode requires multiple GstVideoMeta metadata
- * to describe the fields.
- *
- * The possible values of the #GstVideoInterlaceMode describing the interlace
- * mode of the stream.
- */
-typedef enum {
- GST_VIDEO_INTERLACE_MODE_PROGRESSIVE = 0,
- GST_VIDEO_INTERLACE_MODE_INTERLEAVED,
- GST_VIDEO_INTERLACE_MODE_MIXED,
- GST_VIDEO_INTERLACE_MODE_FIELDS
-} GstVideoInterlaceMode;
-
-/**
- * GstVideoFlags:
- * @GST_VIDEO_FLAG_NONE: no flags
- * @GST_VIDEO_FLAG_VARIABLE_FPS: a variable fps is selected, fps_n and fps_d
- * denote the maximum fps of the video
- * @GST_VIDEO_FLAG_PREMULTIPLIED_ALPHA: Each color has been scaled by the alpha
- * value.
- *
- * Extra video flags
- */
-typedef enum {
- GST_VIDEO_FLAG_NONE = 0,
- GST_VIDEO_FLAG_VARIABLE_FPS = (1 << 0),
- GST_VIDEO_FLAG_PREMULTIPLIED_ALPHA = (1 << 1)
-} GstVideoFlags;
-
-/**
- * GstVideoColorRange:
- * @GST_VIDEO_COLOR_RANGE_UNKNOWN: unknown range
- * @GST_VIDEO_COLOR_RANGE_0_255: [0..255] for 8 bit components
- * @GST_VIDEO_COLOR_RANGE_16_235: [16..235] for 8 bit components. Chroma has
- * [16..240] range.
- *
- * Possible color range values. These constants are defined for 8 bit color
- * values and can be scaled for other bit depths.
- */
-typedef enum {
- GST_VIDEO_COLOR_RANGE_UNKNOWN = 0,
- GST_VIDEO_COLOR_RANGE_0_255,
- GST_VIDEO_COLOR_RANGE_16_235
-} GstVideoColorRange;
-
-/**
- * GstVideoColorMatrix:
- * @GST_VIDEO_COLOR_MATRIX_UNKNOWN: unknown matrix
- * @GST_VIDEO_COLOR_MATRIX_RGB: identity matrix
- * @GST_VIDEO_COLOR_MATRIX_FCC: FCC color matrix
- * @GST_VIDEO_COLOR_MATRIX_BT709: ITU-R BT.709 color matrix
- * @GST_VIDEO_COLOR_MATRIX_BT601: ITU-R BT.601 color matrix
- * @GST_VIDEO_COLOR_MATRIX_SMPTE240M: SMPTE 240M color matrix
- *
- * The color matrix is used to convert between Y'PbPr and
- * non-linear RGB (R'G'B')
- */
-typedef enum {
- GST_VIDEO_COLOR_MATRIX_UNKNOWN = 0,
- GST_VIDEO_COLOR_MATRIX_RGB,
- GST_VIDEO_COLOR_MATRIX_FCC,
- GST_VIDEO_COLOR_MATRIX_BT709,
- GST_VIDEO_COLOR_MATRIX_BT601,
- GST_VIDEO_COLOR_MATRIX_SMPTE240M
-} GstVideoColorMatrix;
-
-/**
- * GstVideoTransferFunction:
- * @GST_VIDEO_TRANSFER_UNKNOWN: unknown transfer function
- * @GST_VIDEO_TRANSFER_GAMMA10: linear RGB, gamma 1.0 curve
- * @GST_VIDEO_TRANSFER_GAMMA18: Gamma 1.8 curve
- * @GST_VIDEO_TRANSFER_GAMMA20: Gamma 2.0 curve
- * @GST_VIDEO_TRANSFER_GAMMA22: Gamma 2.2 curve
- * @GST_VIDEO_TRANSFER_BT709: Gamma 2.2 curve with a linear segment in the lower
- * range
- * @GST_VIDEO_TRANSFER_SMPTE240M: Gamma 2.2 curve with a linear segment in the
- * lower range
- * @GST_VIDEO_TRANSFER_SRGB: Gamma 2.4 curve with a linear segment in the lower
- * range
- * @GST_VIDEO_TRANSFER_GAMMA28: Gamma 2.8 curve
- * @GST_VIDEO_TRANSFER_LOG100: Logarithmic transfer characteristic
- * 100:1 range
- * @GST_VIDEO_TRANSFER_LOG316: Logarithmic transfer characteristic
- * 316.22777:1 range
- *
- * The video transfer function defines the formula for converting between
- * non-linear RGB (R'G'B') and linear RGB
- */
-typedef enum {
- GST_VIDEO_TRANSFER_UNKNOWN = 0,
- GST_VIDEO_TRANSFER_GAMMA10,
- GST_VIDEO_TRANSFER_GAMMA18,
- GST_VIDEO_TRANSFER_GAMMA20,
- GST_VIDEO_TRANSFER_GAMMA22,
- GST_VIDEO_TRANSFER_BT709,
- GST_VIDEO_TRANSFER_SMPTE240M,
- GST_VIDEO_TRANSFER_SRGB,
- GST_VIDEO_TRANSFER_GAMMA28,
- GST_VIDEO_TRANSFER_LOG100,
- GST_VIDEO_TRANSFER_LOG316
-} GstVideoTransferFunction;
-
-/**
- * GstVideoColorPrimaries:
- * @GST_VIDEO_COLOR_PRIMARIES_UNKNOWN: unknown color primaries
- * @GST_VIDEO_COLOR_PRIMARIES_BT709: BT709 primaries
- * @GST_VIDEO_COLOR_PRIMARIES_BT470M: BT470M primaries
- * @GST_VIDEO_COLOR_PRIMARIES_BT470BG: BT470BG primaries
- * @GST_VIDEO_COLOR_PRIMARIES_SMPTE170M: SMPTE170M primaries
- * @GST_VIDEO_COLOR_PRIMARIES_SMPTE240M: SMPTE240M primaries
- * @GST_VIDEO_COLOR_PRIMARIES_FILM: Generic film
- *
- * The color primaries define the how to transform linear RGB values to and from
- * the CIE XYZ colorspace.
- */
-typedef enum {
- GST_VIDEO_COLOR_PRIMARIES_UNKNOWN = 0,
- GST_VIDEO_COLOR_PRIMARIES_BT709,
- GST_VIDEO_COLOR_PRIMARIES_BT470M,
- GST_VIDEO_COLOR_PRIMARIES_BT470BG,
- GST_VIDEO_COLOR_PRIMARIES_SMPTE170M,
- GST_VIDEO_COLOR_PRIMARIES_SMPTE240M,
- GST_VIDEO_COLOR_PRIMARIES_FILM
-} GstVideoColorPrimaries;
-
-/**
- * GstVideoColorimetry:
- * @range: the color range. This is the valid range for the samples.
- * It is used to convert the samples to Y'PbPr values.
- * @matrix: the color matrix. Used to convert between Y'PbPr and
- * non-linear RGB (R'G'B')
- * @transfer: the transfer function. used to convert between R'G'B' and RGB
- * @primaries: color primaries. used to convert between R'G'B' and CIE XYZ
- *
- * Structure describing the color info.
- */
-struct _GstVideoColorimetry {
- GstVideoColorRange range;
- GstVideoColorMatrix matrix;
- GstVideoTransferFunction transfer;
- GstVideoColorPrimaries primaries;
-};
-
-/* predefined colorimetry */
-#define GST_VIDEO_COLORIMETRY_BT601 "bt601"
-#define GST_VIDEO_COLORIMETRY_BT709 "bt709"
-#define GST_VIDEO_COLORIMETRY_SMPTE240M "smpte240m"
-
-gboolean gst_video_colorimetry_matches (GstVideoColorimetry *cinfo, const gchar *color);
-gboolean gst_video_colorimetry_from_string (GstVideoColorimetry *cinfo, const gchar *color);
-gchar * gst_video_colorimetry_to_string (GstVideoColorimetry *cinfo);
-
-/**
- * GstVideoInfo:
- * @finfo: the format info of the video
- * @interlace_mode: the interlace mode
- * @flags: additional video flags
- * @width: the width of the video
- * @height: the height of the video
- * @views: the number of views for multiview video
- * @size: the default size of one frame
- * @chroma_site: a #GstVideoChromaSite.
- * @colorimetry: the colorimetry info
- * @par_n: the pixel-aspect-ratio numerator
- * @par_d: the pixel-aspect-ratio demnominator
- * @fps_n: the framerate numerator
- * @fps_d: the framerate demnominator
- * @offset: offsets of the planes
- * @stride: strides of the planes
- *
- * Information describing image properties. This information can be filled
- * in from GstCaps with gst_video_info_from_caps(). The information is also used
- * to store the specific video info when mapping a video frame with
- * gst_video_frame_map().
- *
- * Use the provided macros to access the info in this structure.
- */
-struct _GstVideoInfo {
- const GstVideoFormatInfo *finfo;
-
- GstVideoInterlaceMode interlace_mode;
- GstVideoFlags flags;
- gint width;
- gint height;
- gsize size;
- gint views;
-
- GstVideoChromaSite chroma_site;
- GstVideoColorimetry colorimetry;
-
- gint par_n;
- gint par_d;
- gint fps_n;
- gint fps_d;
-
- gsize offset[GST_VIDEO_MAX_PLANES];
- gint stride[GST_VIDEO_MAX_PLANES];
-};
-
-/* general info */
-#define GST_VIDEO_INFO_FORMAT(i) (GST_VIDEO_FORMAT_INFO_FORMAT((i)->finfo))
-#define GST_VIDEO_INFO_NAME(i) (GST_VIDEO_FORMAT_INFO_NAME((i)->finfo))
-#define GST_VIDEO_INFO_IS_YUV(i) (GST_VIDEO_FORMAT_INFO_IS_YUV((i)->finfo))
-#define GST_VIDEO_INFO_IS_RGB(i) (GST_VIDEO_FORMAT_INFO_IS_RGB((i)->finfo))
-#define GST_VIDEO_INFO_IS_GRAY(i) (GST_VIDEO_FORMAT_INFO_IS_GRAY((i)->finfo))
-#define GST_VIDEO_INFO_HAS_ALPHA(i) (GST_VIDEO_FORMAT_INFO_HAS_ALPHA((i)->finfo))
-
-#define GST_VIDEO_INFO_INTERLACE_MODE(i) ((i)->interlace_mode)
-#define GST_VIDEO_INFO_IS_INTERLACED(i) ((i)->interlace_mode != GST_VIDEO_INTERLACE_MODE_PROGRESSIVE)
-#define GST_VIDEO_INFO_FLAGS(i) ((i)->flags)
-#define GST_VIDEO_INFO_WIDTH(i) ((i)->width)
-#define GST_VIDEO_INFO_HEIGHT(i) ((i)->height)
-#define GST_VIDEO_INFO_SIZE(i) ((i)->size)
-#define GST_VIDEO_INFO_VIEWS(i) ((i)->views)
-#define GST_VIDEO_INFO_PAR_N(i) ((i)->par_n)
-#define GST_VIDEO_INFO_PAR_D(i) ((i)->par_d)
-#define GST_VIDEO_INFO_FPS_N(i) ((i)->fps_n)
-#define GST_VIDEO_INFO_FPS_D(i) ((i)->fps_d)
-
-/* dealing with GstVideoInfo flags */
-#define GST_VIDEO_INFO_FLAG_IS_SET(i,flag) ((GST_VIDEO_INFO_FLAGS(i) & (flag)) == (flag))
-#define GST_VIDEO_INFO_FLAG_SET(i,flag) (GST_VIDEO_INFO_FLAGS(i) |= (flag))
-#define GST_VIDEO_INFO_FLAG_UNSET(i,flag) (GST_VIDEO_INFO_FLAGS(i) &= ~(flag))
-
-/* dealing with planes */
-#define GST_VIDEO_INFO_N_PLANES(i) (GST_VIDEO_FORMAT_INFO_N_PLANES((i)->finfo))
-#define GST_VIDEO_INFO_PLANE_OFFSET(i,p) ((i)->offset[p])
-#define GST_VIDEO_INFO_PLANE_STRIDE(i,p) ((i)->stride[p])
-
-/* dealing with components */
-#define GST_VIDEO_INFO_N_COMPONENTS(i) GST_VIDEO_FORMAT_INFO_N_COMPONENTS((i)->finfo)
-#define GST_VIDEO_INFO_COMP_DEPTH(i,c) GST_VIDEO_FORMAT_INFO_DEPTH((i)->finfo,(c))
-#define GST_VIDEO_INFO_COMP_DATA(i,d,c) GST_VIDEO_FORMAT_INFO_DATA((i)->finfo,d,(c))
-#define GST_VIDEO_INFO_COMP_OFFSET(i,c) GST_VIDEO_FORMAT_INFO_OFFSET((i)->finfo,(i)->offset,(c))
-#define GST_VIDEO_INFO_COMP_STRIDE(i,c) GST_VIDEO_FORMAT_INFO_STRIDE((i)->finfo,(i)->stride,(c))
-#define GST_VIDEO_INFO_COMP_WIDTH(i,c) GST_VIDEO_FORMAT_INFO_SCALE_WIDTH((i)->finfo,(c),(i)->width)
-#define GST_VIDEO_INFO_COMP_HEIGHT(i,c) GST_VIDEO_FORMAT_INFO_SCALE_HEIGHT((i)->finfo,(c),(i)->height)
-#define GST_VIDEO_INFO_COMP_PLANE(i,c) GST_VIDEO_FORMAT_INFO_PLANE((i)->finfo,(c))
-#define GST_VIDEO_INFO_COMP_PSTRIDE(i,c) GST_VIDEO_FORMAT_INFO_PSTRIDE((i)->finfo,(c))
-#define GST_VIDEO_INFO_COMP_POFFSET(i,c) GST_VIDEO_FORMAT_INFO_POFFSET((i)->finfo,(c))
-
-void gst_video_info_init (GstVideoInfo *info);
-
-void gst_video_info_set_format (GstVideoInfo *info, GstVideoFormat format,
- guint width, guint height);
-
-gboolean gst_video_info_from_caps (GstVideoInfo *info, const GstCaps * caps);
-
-GstCaps * gst_video_info_to_caps (GstVideoInfo *info);
-
-gboolean gst_video_info_convert (GstVideoInfo *info,
- GstFormat src_format,
- gint64 src_value,
- GstFormat dest_format,
- gint64 *dest_value);
-gboolean gst_video_info_is_equal (const GstVideoInfo *info,
- const GstVideoInfo *other);
-
-/**
- * GstVideoFrameFlags:
- * @GST_VIDEO_FRAME_FLAG_NONE: no flags
- * @GST_VIDEO_FRAME_FLAG_TFF: The video frame has the top field first
- * @GST_VIDEO_FRAME_FLAG_RFF: The video frame has the repeat flag
- * @GST_VIDEO_FRAME_FLAG_ONEFIELD: The video frame has one field
- *
- * Extra video frame flags
- */
-typedef enum {
- GST_VIDEO_FRAME_FLAG_NONE = 0,
- GST_VIDEO_FRAME_FLAG_TFF = (1 << 0),
- GST_VIDEO_FRAME_FLAG_RFF = (1 << 1),
- GST_VIDEO_FRAME_FLAG_ONEFIELD = (1 << 2)
-} GstVideoFrameFlags;
-
-/**
- * GstVideoFrame:
- * @info: the #GstVideoInfo
- * @buffer: the mapped buffer
- * @meta: pointer to metadata if any
- * @id: id of the mapped frame. the id can for example be used to
- * indentify the frame in case of multiview video.
- * @data: pointers to the plane data
- * @map: mappings of the planes
- *
- * A video frame obtained from gst_video_frame_map()
- */
-struct _GstVideoFrame {
- GstVideoInfo info;
- GstVideoFrameFlags flags;
-
- GstBuffer *buffer;
- gpointer meta;
- gint id;
-
- gpointer data[GST_VIDEO_MAX_PLANES];
- GstMapInfo map[GST_VIDEO_MAX_PLANES];
-};
-
-gboolean gst_video_frame_map (GstVideoFrame *frame, GstVideoInfo *info,
- GstBuffer *buffer, GstMapFlags flags);
-gboolean gst_video_frame_map_id (GstVideoFrame *frame, GstVideoInfo *info,
- GstBuffer *buffer, gint id, GstMapFlags flags);
-void gst_video_frame_unmap (GstVideoFrame *frame);
-
-gboolean gst_video_frame_copy (GstVideoFrame *dest, const GstVideoFrame *src);
-gboolean gst_video_frame_copy_plane (GstVideoFrame *dest, const GstVideoFrame *src,
- guint plane);
-
-/* general info */
-#define GST_VIDEO_FRAME_FORMAT(f) (GST_VIDEO_INFO_FORMAT(&(f)->info))
-#define GST_VIDEO_FRAME_WIDTH(f) (GST_VIDEO_INFO_WIDTH(&(f)->info))
-#define GST_VIDEO_FRAME_HEIGHT(f) (GST_VIDEO_INFO_HEIGHT(&(f)->info))
-#define GST_VIDEO_FRAME_SIZE(f) (GST_VIDEO_INFO_SIZE(&(f)->info))
-
-/* dealing with planes */
-#define GST_VIDEO_FRAME_N_PLANES(f) (GST_VIDEO_INFO_N_PLANES(&(f)->info))
-#define GST_VIDEO_FRAME_PLANE_DATA(f,p) ((f)->data[p])
-#define GST_VIDEO_FRAME_PLANE_OFFSET(f,p) (GST_VIDEO_INFO_PLANE_OFFSET(&(f)->info,(p)))
-#define GST_VIDEO_FRAME_PLANE_STRIDE(f,p) (GST_VIDEO_INFO_PLANE_STRIDE(&(f)->info,(p)))
-
-/* dealing with components */
-#define GST_VIDEO_FRAME_N_COMPONENTS(f) GST_VIDEO_INFO_N_COMPONENTS(&(f)->info)
-#define GST_VIDEO_FRAME_COMP_DATA(f,c) GST_VIDEO_INFO_COMP_DATA(&(f)->info,(f)->data,(c))
-#define GST_VIDEO_FRAME_COMP_STRIDE(f,c) GST_VIDEO_INFO_COMP_STRIDE(&(f)->info,(c))
-#define GST_VIDEO_FRAME_COMP_OFFSET(f,c) GST_VIDEO_INFO_COMP_OFFSET(&(f)->info,(c))
-#define GST_VIDEO_FRAME_COMP_WIDTH(f,c) GST_VIDEO_INFO_COMP_WIDTH(&(f)->info,(c))
-#define GST_VIDEO_FRAME_COMP_HEIGHT(f,c) GST_VIDEO_INFO_COMP_HEIGHT(&(f)->info,(c))
-#define GST_VIDEO_FRAME_COMP_PLANE(f,c) GST_VIDEO_INFO_COMP_PLANE(&(f)->info,(c))
-#define GST_VIDEO_FRAME_COMP_PSTRIDE(f,c) GST_VIDEO_INFO_COMP_PSTRIDE(&(f)->info,(c))
-#define GST_VIDEO_FRAME_COMP_POFFSET(f,c) GST_VIDEO_INFO_COMP_POFFSET(&(f)->info,(c))
-
-#define GST_VIDEO_SIZE_RANGE "(int) [ 1, max ]"
-#define GST_VIDEO_FPS_RANGE "(fraction) [ 0, max ]"
-
-#if G_BYTE_ORDER == G_LITTLE_ENDIAN
-# define GST_VIDEO_NE(s) G_STRINGIFY(s)"_LE"
-# define GST_VIDEO_OE(s) G_STRINGIFY(s)"_BE"
-#else
-# define GST_VIDEO_NE(s) G_STRINGIFY(s)"_BE"
-# define GST_VIDEO_OE(s) G_STRINGIFY(s)"_LE"
-#endif
-
-#define GST_VIDEO_FORMATS_ALL "{ I420, YV12, YUY2, UYVY, AYUV, RGBx, " \
- "BGRx, xRGB, xBGR, RGBA, BGRA, ARGB, ABGR, RGB, BGR, Y41B, Y42B, " \
- "YVYU, Y444, v210, v216, NV12, NV21, GRAY8, GRAY16_BE, GRAY16_LE, " \
- "v308, RGB16, BGR16, RGB15, BGR15, UYVP, A420, RGB8P, YUV9, YVU9, " \
- "IYU1, ARGB64, AYUV64, r210, I420_10LE, I420_10BE }"
-
-/**
- * GST_VIDEO_CAPS_MAKE:
- * @format: string format that describes the pixel layout, as string
- * (e.g. "I420", "RGB", "YV12", "YUY2", "AYUV", etc.)
- *
- * Generic caps string for video, for use in pad templates.
- */
-#define GST_VIDEO_CAPS_MAKE(format) \
- "video/x-raw, " \
- "format = (string) " format ", " \
- "width = " GST_VIDEO_SIZE_RANGE ", " \
- "height = " GST_VIDEO_SIZE_RANGE ", " \
- "framerate = " GST_VIDEO_FPS_RANGE
-
-/* buffer flags */
-
-/**
- * GstVideoBufferFlags:
- * @GST_VIDEO_BUFFER_FLAG_TFF: If the #GstBuffer is interlaced, then the first field
- * in the video frame is the top field. If unset, the
- * bottom field is first.
- * @GST_VIDEO_BUFFER_FLAG_RFF: If the #GstBuffer is interlaced, then the first field
- * (as defined by the %GST_VIDEO_BUFFER_TFF flag setting)
- * is repeated.
- * @GST_VIDEO_BUFFER_FLAG_ONEFIELD: If the #GstBuffer is interlaced, then only the
- * first field (as defined by the %GST_VIDEO_BUFFER_TFF
- * flag setting) is to be displayed.
- *
- * Additional video buffer flags.
- */
-typedef enum {
- GST_VIDEO_BUFFER_FLAG_TFF = (GST_BUFFER_FLAG_LAST << 0),
- GST_VIDEO_BUFFER_FLAG_RFF = (GST_BUFFER_FLAG_LAST << 1),
- GST_VIDEO_BUFFER_FLAG_ONEFIELD = (GST_BUFFER_FLAG_LAST << 2),
-
- GST_VIDEO_BUFFER_FLAG_LAST = (GST_BUFFER_FLAG_LAST << 8)
-} GstVideoBufferFlags;
+G_BEGIN_DECLS
/* some helper functions */
gboolean gst_video_calculate_display_ratio (guint * dar_n,
guint display_par_n,
guint display_par_d);
-/* video still frame event creation and parsing */
-
-GstEvent * gst_video_event_new_still_frame (gboolean in_still);
-
-gboolean gst_video_event_parse_still_frame (GstEvent * event, gboolean * in_still);
-
-/* video force key unit event creation and parsing */
-
-GstEvent * gst_video_event_new_downstream_force_key_unit (GstClockTime timestamp,
- GstClockTime stream_time,
- GstClockTime running_time,
- gboolean all_headers,
- guint count);
-
-gboolean gst_video_event_parse_downstream_force_key_unit (GstEvent * event,
- GstClockTime * timestamp,
- GstClockTime * stream_time,
- GstClockTime * running_time,
- gboolean * all_headers,
- guint * count);
-
-GstEvent * gst_video_event_new_upstream_force_key_unit (GstClockTime running_time,
- gboolean all_headers,
- guint count);
-
-gboolean gst_video_event_parse_upstream_force_key_unit (GstEvent * event,
- GstClockTime * running_time,
- gboolean * all_headers,
- guint * count);
-
-gboolean gst_video_event_is_force_key_unit(GstEvent *event);
-
/* convert/encode video sample from one format to another */
typedef void (*GstVideoConvertSampleCallback) (GstSample * sample, GError *error, gpointer user_data);